| ▲ | zachperkitny 7 hours ago | |||||||
Hello! I wanted to share my recent project: Tadpole. It is a custom DSL built on top of KDL specifically for web scraping and browser automation. I wanted there to be a standardized way of writing scrapers and reusing existing scraper logic. This was my solution. Why?
ExampleThis is a fully running example, @tadpole/cli is published on npm: tadpole run redfin.kdl --input '{"text": "Seattle, WA"}' --auto --output output.json
Roadmap?
Planned for 0.2.0
0.3.0
Github Repository: https://github.com/tadpolehq/tadpoleI've also created a community repository for sharing scraper logic: https://github.com/tadpolehq/community Feedback would be greatly appreciated! | ||||||||
| ▲ | bobajeff 6 hours ago | parent [-] | |||||||
I like the idea of a DSL for scraping but my scrapers do more than extract text. I also download files (+monitor download progress) and intercept images (+ check for partial or failed to load images). So it seems my use case isn't really covered with this. | ||||||||
| ||||||||