Skip to content
GitLab
Projects Groups Snippets
  • /
  • Help
    • Help
    • Support
    • Community forum
    • Submit feedback
    • Contribute to GitLab
  • Sign in / Register
  • A awesome-python
  • Project information
    • Project information
    • Activity
    • Labels
    • Members
  • Repository
    • Repository
    • Files
    • Commits
    • Branches
    • Tags
    • Contributors
    • Graph
    • Compare
  • Issues 13
    • Issues 13
    • List
    • Boards
    • Service Desk
    • Milestones
  • Merge requests 317
    • Merge requests 317
  • CI/CD
    • CI/CD
    • Pipelines
    • Jobs
    • Schedules
  • Deployments
    • Deployments
    • Environments
    • Releases
  • Packages and registries
    • Packages and registries
    • Package Registry
    • Infrastructure Registry
  • Monitor
    • Monitor
    • Incidents
  • Analytics
    • Analytics
    • Value stream
    • CI/CD
    • Repository
  • Wiki
    • Wiki
  • Snippets
    • Snippets
  • Activity
  • Graph
  • Create a new issue
  • Jobs
  • Commits
  • Issue Boards
Collapse sidebar
  • Vinta Chen
  • awesome-python
  • Merge requests
  • !977

Adding weboob in the Web Crawling section

  • Review changes

  • Download
  • Email patches
  • Plain diff
Closed Administrator requested to merge github/fork/Mistress-Anna/patch-1 into master Nov 15, 2017
  • Overview 0
  • Commits 1
  • Pipelines 0
  • Changes 1

Created by: Mistress-Anna

What is this Python project?

WebOOB is a framework for scraping websites and aggregating data from multiple websites.

What's the difference between this Python project and similar ones?

  • Routing model of URL patterns to multiple class of Page with all the parsing associated to each of those Pages, for cleaner code
  • Scraping is made easy thanks to "declarative parsing": each Page can have a few XPaths, configure a few "filters" to apply on those XPaths (like parsing int, apply regex, etc.), and you're set!
  • Like every high-level feature in WebOOB, this declarative parsing can be disabled locally, when it doesn't fit for a particular site, and it's always possible to fallback to plain-old procedural parsing code
  • Pagination handling, supports infinite iterators
  • Typed data models to ensure clean scraped data
  • Can handle HTML/XML, JSON, and even XLS or PDF
  • (Optional) Can aggregate data from multiple websites by grouping them in categories (for example "video sites", "banking sites", "public transport sites", "event sites", etc.)
  • Comes builtin with a ~250 pre-existing website crawling backends
  • Has a few graphical and command-line apps to explore and search the scraped data
Assignee
Assign to
Reviewers
Request review from
Time tracking
Source branch: github/fork/Mistress-Anna/patch-1