diff options
| author | Jules Laplace <julescarbon@gmail.com> | 2018-12-16 15:02:59 +0100 |
|---|---|---|
| committer | Jules Laplace <julescarbon@gmail.com> | 2018-12-16 15:02:59 +0100 |
| commit | 110f3a34f1f36d0ea999d4aa34bbe66d5f2a01da (patch) | |
| tree | f21fbeccb6a7e8d3af5d5c537ed5931ecbd62d7e /scraper/README.md | |
| parent | 2cb31d4999649a22a0ac659a59a0aa0a0f7a241e (diff) | |
skip empty, pull citations again
Diffstat (limited to 'scraper/README.md')
| -rw-r--r-- | scraper/README.md | 4 |
1 files changed, 2 insertions, 2 deletions
diff --git a/scraper/README.md b/scraper/README.md index 318bba9a..4399abd3 100644 --- a/scraper/README.md +++ b/scraper/README.md @@ -70,9 +70,9 @@ Included in the content-script folder is a Chrome extension which scrapes Google Once you have the data from S2, you can scrape all the PDFs (and other URLs) you find, and then extract institutions from those and geocode them. -### s2-dump-pdf-urls.py +### s2-dump-db-pdf-urls.py -Dump PDF urls (and also IEEE urls etc) to CSV files. +Dump PDF urls (and also DOI urls etc) to CSV files. ### s2-fetch-pdf.py |
