{ "info": { "author": "Alex Sieusahai", "author_email": "alexsieu14@gmail.com", "bugtrack_url": null, "classifiers": [], "description": "# Librarian\n\nThe goal of this package is to almost be like a training wheels setup for web scraping.\n\nA good example is recursively trying to visit all of the links on a site such as:\n\nhttp://web.archive.org/web/20080827084856/http://www.nanowerk.com:80/nanotechnology/nanomaterial/commercial_all.php?page=2\n\nIf you mess up your scrape and you have done no preemptive measures, you lose all of your work done thus far. What `Librarian` aims to do is to save that html for later, so you don't have to redo scrapes that you have done before, letting you be much nicer to where you're requesting to and saving you time, letting you go through a much smoother scraping experience.\n\nLet's outline an example:\n\nTake a look at the site above's html via Inspect Element; You will see that all of the names and links are under