summaryrefslogtreecommitdiff
path: root/scraper/README.md
diff options
context:
space:
mode:
authorJules Laplace <julescarbon@gmail.com>2018-12-16 15:02:59 +0100
committerJules Laplace <julescarbon@gmail.com>2018-12-16 15:02:59 +0100
commit110f3a34f1f36d0ea999d4aa34bbe66d5f2a01da (patch)
treef21fbeccb6a7e8d3af5d5c537ed5931ecbd62d7e /scraper/README.md
parent2cb31d4999649a22a0ac659a59a0aa0a0f7a241e (diff)
skip empty, pull citations again
Diffstat (limited to 'scraper/README.md')
-rw-r--r--scraper/README.md4
1 files changed, 2 insertions, 2 deletions
diff --git a/scraper/README.md b/scraper/README.md
index 318bba9a..4399abd3 100644
--- a/scraper/README.md
+++ b/scraper/README.md
@@ -70,9 +70,9 @@ Included in the content-script folder is a Chrome extension which scrapes Google
Once you have the data from S2, you can scrape all the PDFs (and other URLs) you find, and then extract institutions from those and geocode them.
-### s2-dump-pdf-urls.py
+### s2-dump-db-pdf-urls.py
-Dump PDF urls (and also IEEE urls etc) to CSV files.
+Dump PDF urls (and also DOI urls etc) to CSV files.
### s2-fetch-pdf.py