Skip to content

Python spider for corona virus data from worldometer / worldometers

Notifications You must be signed in to change notification settings

KDKHD/CoronaDataScraper

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

6 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

CoronaDataScraper

Elastic Search Corona Dashboard

ES dashboard

Getting started

The spider stores all of its data in elastic search. Elastic search has been dockerized so all you have to do to download it is downloader docker desktop, and run "docker-compose up" in the root directory.

After this install requrements.txt (if you dont want to use a virtual enviroment, skip to next step). This should be done by creating a virtualenv using "virtualenv -p python3 env". Then activate the env with "source env/bin/activate".

Install requirements. "pip3 install -r requirements.txt"

Now cd into "crawlers/crawlers/spiders" and run the spider "scrapy crawl worldMeter".

If you dont want to use elastic search, comment out these lines in the spider

def __init__(self):
        try:
            self.es = elasticSearchApi.es()
        except Exception as e:
            quit()
self.es.store_data(rowData, index = "corona_daily_worldometer_table", doc_id = dataHash)
self.es.store_data(data, index = "corona_country_worldometer_past", doc_id = dataHash)

About

Python spider for corona virus data from worldometer / worldometers

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages