Skip to content

Automated scraper build with that scrapes data from the running event Zevenheuvelenloop.

Notifications You must be signed in to change notification settings

Frituurpanda/Zevenheuvelenloop-Scraper

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

2 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Zevenheuvelenloop Uitslag Scraper

Automated scraper build with Scrapy that scrapes data from the running event Zevenheuvelenloop. The data gets scraped from http://evenementen.uitslagen.nl/.


The scraper process works as follows:

  • The allowed domain gets set to evenementen.uitslagen.nl
  • The max_id value should be changed to the maximum amount of pages that you want to have scraped.
  • An example of the final url can be ~uitslag01233.html"
  • The scraper will scrape the set amount of pages and add those to the csv file until done.

Example with Airbnb's Superset

Example1

Why???

After competing in the event myself I wanted to have some answers based on the data that resulted from the competition. This included questions like:

  • What was the average finish time of a certain category.
  • How much people did a person overtake in his own category.
  • How much people did a person overtake given all categories
  • Etc etc..

How to use

You can use the scraper by using the command

scrapy runspider ZevenHeuvelSpider_spider.py -o zevenheuvel.csv

After completion, you can open the zevenheuvel.csv for inspection.

Resources Used

About

Automated scraper build with that scrapes data from the running event Zevenheuvelenloop.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages