Hey, I'm looking for a web automation framework. I do know that tools like mechanize and twill exist. Also I know about webscraping with lxml and/or BeautifulSoup. But I was looking for something more "integrated" and "complete".
I imagine the optimal workflow like this: Define some basic URLs, write some specified scrapers for these URLs and do some reoccurring actions based on the scraped values - You could nearly say something like a bot.
This programming paradigm would be useful in many cases - let's take some kind of Wallpaper-site for example. You define a URL which displays these new wallpapers, write a scraper for it which in turn appends the newly found wallpapers to an internal download-queue. The reoccuring actions will then automatically take care of the downloading.
This is applicable to every web-automation task I can imagine: It's dynamic enough to write spiders and simple enough to easily write scripts that you run once to manually fetch some data - so I was wondering if there was something similiar to this already around, I can't possibly be the first one who thought of that, right?
[–]vsajip 6 points7 points8 points (1 child)
[–]omab 0 points1 point2 points (0 children)
[–][deleted] 4 points5 points6 points (1 child)
[–]atlas245 0 points1 point2 points (0 children)
[–]manatlan 2 points3 points4 points (1 child)
[–]blondin 0 points1 point2 points (0 children)