mirror of
https://github.com/jaypyles/Scraperr.git
synced 2026-05-04 00:10:44 +00:00
Update README.md
This commit is contained in:
@@ -6,14 +6,17 @@ From the table, users can download a csv of the job's results, along with an opt
|
||||
|
||||
## Features
|
||||
|
||||
- Submit URLs for web scraping
|
||||
- Submit/Queue URLs for web scraping
|
||||
- Add and manage elements to scrape using XPath
|
||||
- Scrape all pages within same domain
|
||||
- Add custom json headers to send in requests to URLs
|
||||
- Display results of scraped data
|
||||
|
||||

|
||||
|
||||
- Download csv containing results
|
||||
- Rerun jobs
|
||||
- View status of queued jobs
|
||||
|
||||

|
||||
|
||||
@@ -21,6 +24,10 @@ From the table, users can download a csv of the job's results, along with an opt
|
||||
|
||||

|
||||
|
||||
- View app logs inside of web ui
|
||||
|
||||

|
||||
|
||||
## Installation
|
||||
|
||||
1. Clone the repository:
|
||||
@@ -56,8 +63,8 @@ The app provides its own `traefik` configuration to use independently, but can e
|
||||
1. Open the application in your browser at `http://localhost`.
|
||||
2. Enter the URL you want to scrape in the URL field.
|
||||
3. Add elements to scrape by specifying a name and the corresponding XPath.
|
||||
4. Click the "Submit" button to start the scraping process.
|
||||
5. The results will be displayed in the "Results" section.
|
||||
4. Click the "Submit" button to queue URL to be scraped.
|
||||
5. View queue in the "Previous Jobs" section.
|
||||
|
||||
## API Endpoints
|
||||
|
||||
|
||||
Reference in New Issue
Block a user