diff options
Diffstat (limited to 'site')
22 files changed, 177 insertions, 77 deletions
diff --git a/site/assets/css/applets.css b/site/assets/css/applets.css index 41d04783..7fac3e27 100644 --- a/site/assets/css/applets.css +++ b/site/assets/css/applets.css @@ -3,6 +3,7 @@ .applet_container { min-height: 340px; clear: left; + margin: 20px auto 40px auto; } .applet_container.autosize { min-height: 0; diff --git a/site/assets/css/css.css b/site/assets/css/css.css index c3800315..0ee8a4f3 100644 --- a/site/assets/css/css.css +++ b/site/assets/css/css.css @@ -4,7 +4,8 @@ html, body { padding: 0; width: 100%; min-height: 100%; - font-family: 'Roboto Mono', sans-serif; + /*font-family: 'Roboto Mono', sans-serif;*/ + font-family: 'Roboto', sans-serif; color: #eee; overflow-x: hidden; } @@ -163,33 +164,33 @@ h1 { margin: 75px 0 10px; padding: 0; transition: color 0.1s cubic-bezier(0,0,1,1); - font-family: 'Roboto'; + font-family: 'Roboto Mono', monospace; } h2 { color: #eee; font-weight: 400; - font-size: 32pt; - line-height: 43pt; - margin: 20px 0 10px; + font-size: 32px; + line-height: 43px; + margin: 20px 0 20px; padding: 0; transition: color 0.1s cubic-bezier(0,0,1,1); - font-family: 'Roboto'; + font-family: 'Roboto Mono', monospace; } h3 { margin: 0 0 20px 0; padding: 20px 0 0 0; font-size: 22pt; - font-weight: 500; + font-weight: 400; transition: color 0.1s cubic-bezier(0,0,1,1); - font-family: 'Roboto'; + font-family: 'Roboto Mono', monospace; } h4 { margin: 0 0 10px 0; padding: 0; font-size: 11pt; - font-weight: 500; + font-weight: 400; transition: color 0.1s cubic-bezier(0,0,1,1); - font-family: 'Roboto'; + font-family: 'Roboto Mono', monospace; } .content h3 a { color: #888; @@ -212,11 +213,11 @@ h4 { border-bottom: 0; } th, .gray { - font-family: 'Roboto Mono', monospace; + font-family: 'Roboto', monospace; font-weight: 500; text-transform: uppercase; letter-spacing: .15rem; - color: #999; + color: #777; } th, .gray { font-size: 9pt; @@ -248,8 +249,9 @@ section { p { margin: 0 10px 20px 0; line-height: 2; - font-size: 16px; + font-size: 18px; font-weight: 300; + color: #dedede; } p.subp{ font-size: 14px; @@ -272,18 +274,19 @@ p.subp{ flex-direction: row; justify-content: flex-start; align-items: flex-start; - font-size: 14px; + font-size: 12px; + color: #ccc; margin-bottom: 20px; font-family: 'Roboto', sans-serif; } .meta > div { margin-right: 20px; - line-height: 19px + line-height: 17px /*font-size:11px;*/ } .meta .gray { font-size: 9pt; - padding-bottom: 4px; + padding-bottom: 5px; line-height: 14px } .right-sidebar { @@ -303,7 +306,7 @@ p.subp{ padding-top: 10px; padding-right: 20px; /*margin-right: 20px;*/ - margin-bottom: 30px; + margin-bottom: 10px; /*border-right: 1px solid #444;*/ font-family: 'Roboto'; font-size: 14px; @@ -881,7 +884,7 @@ ul.map-legend li.source:before { font-family: Roboto, sans-serif; font-weight: 400; background: #202020; - padding: 15px; + padding: 20px; margin: 10px; } .columns .column:first-of-type { @@ -934,7 +937,7 @@ ul.map-legend li.source:before { margin:0 0 0 40px; } .content-about .team-member p{ - font-size:14px; + font-size:16px; } .content-about .team-member img{ margin:0; diff --git a/site/content/pages/datasets/brainwash/assets/00818000_640x480.jpg b/site/content/pages/datasets/brainwash/assets/00818000_640x480.jpg Binary files differdeleted file mode 100644 index 30c0fcb1..00000000 --- a/site/content/pages/datasets/brainwash/assets/00818000_640x480.jpg +++ /dev/null diff --git a/site/content/pages/datasets/brainwash/assets/background_540.jpg b/site/content/pages/datasets/brainwash/assets/background_540.jpg Binary files differdeleted file mode 100644 index 5c8c0ad4..00000000 --- a/site/content/pages/datasets/brainwash/assets/background_540.jpg +++ /dev/null diff --git a/site/content/pages/datasets/brainwash/assets/background_600.jpg b/site/content/pages/datasets/brainwash/assets/background_600.jpg Binary files differdeleted file mode 100755 index 8f2de697..00000000 --- a/site/content/pages/datasets/brainwash/assets/background_600.jpg +++ /dev/null diff --git a/site/content/pages/datasets/brainwash/assets/brainwash_mean_overlay.jpg b/site/content/pages/datasets/brainwash/assets/brainwash_mean_overlay.jpg Binary files differnew file mode 100755 index 00000000..2f5917e3 --- /dev/null +++ b/site/content/pages/datasets/brainwash/assets/brainwash_mean_overlay.jpg diff --git a/site/content/pages/datasets/brainwash/assets/brainwash_mean_overlay_wm.jpg b/site/content/pages/datasets/brainwash/assets/brainwash_mean_overlay_wm.jpg Binary files differnew file mode 100755 index 00000000..790dbb79 --- /dev/null +++ b/site/content/pages/datasets/brainwash/assets/brainwash_mean_overlay_wm.jpg diff --git a/site/content/pages/datasets/brainwash/index.md b/site/content/pages/datasets/brainwash/index.md index 0bf67455..6d90e78f 100644 --- a/site/content/pages/datasets/brainwash/index.md +++ b/site/content/pages/datasets/brainwash/index.md @@ -2,8 +2,8 @@ status: published title: Brainwash -desc: Brainwash is a dataset of webcam images taken from the Brainwash Cafe in San Francisco -subdesc: The Brainwash dataset includes 11,918 images of "everyday life of a busy downtown cafe" and is used for training head detection algorithms +desc: Brainwash is a dataset of webcam images taken from the Brainwash Cafe in San Francisco in 2014 +subdesc: The Brainwash dataset includes 11,918 images of "everyday life of a busy downtown cafe" and is used for training head detection surveillance algorithms slug: brainwash cssclass: dataset image: assets/background.jpg @@ -19,28 +19,23 @@ authors: Adam Harvey + Published: 2015 + Images: 11,918 + Faces: 91,146 -+ Created by: Stanford Department of Computer Science ++ Created by: Stanford University (US)<br>Max Planck Institute for Informatics (DE) + Funded by: Max Planck Center for Visual Computing and Communication -+ Location: Brainwash Cafe, San Franscisco -+ Purpose: Training face detection ++ Purpose: Head detection ++ Download Size: 4.1GB + Website: <a href="https://exhibits.stanford.edu/data/catalog/sx925dc9385">stanford.edu</a> -+ Paper: <a href="http://arxiv.org/abs/1506.04878">End-to-End People Detection in Crowded Scenes</a> -+ Explicit Consent: No ## Brainwash Dataset -(PAGE UNDER DEVELOPMENT) +*Brainwash* is a head detection dataset created from San Francisco's Brainwash Cafe livecam footage. It includes 11,918 images of "everyday life of a busy downtown cafe"[^readme] captured at 100 second intervals throught the entire day. Brainwash dataset was captured during 3 days in 2014: October 27, November 13, and November 24. According the author's reserach paper introducing the dataset, the images were acquired with the help of Angelcam.com [cite orig paper]. -*Brainwash* is a face detection dataset created from the Brainwash Cafe's livecam footage including 11,918 images of "everyday life of a busy downtown cafe[^readme]". The images are used to develop face detection algorithms for the "challenging task of detecting people in crowded scenes" and tracking them. +Brainwash is not a widely used dataset but since its publication by Stanford University in 2015, it has notably appeared in several research papers from the National University of Defense Technology in Changsha, China. In 2016 and in 2017 researchers there conducted studies on detecting people's heads in crowded scenes for the purpose of surveillance [^localized_region_context] [^replacement_algorithm]. -Before closing in 2017, Brainwash Cafe was a "cafe and laundromat" located in San Francisco's SoMA district. The cafe published a publicy available livestream from the cafe with a view of the cash register, performance stage, and seating area. +If you happen to have been at Brainwash cafe in San Franscisco at any time on October 26, November 13, or November 24 in 2014 you are most likely included in the Brainwash dataset. -Since it's publication by Stanford in 2015, the Brainwash dataset has appeared in several notable research papers. In September 2016 four researchers from the National University of Defense Technology in Changsha, China used the Brainwash dataset for a research study on "people head detection in crowded scenes", concluding that their algorithm "achieves superior head detection performance on the crowded scenes dataset[^localized_region_context]". And again in 2017 three researchers at the National University of Defense Technology used Brainwash for a study on object detection noting "the data set used in our experiment is shown in Table 1, which includes one scene of the brainwash dataset[^replacement_algorithm]". + - - - {% include 'chart.html' %} @@ -48,19 +43,27 @@ Since it's publication by Stanford in 2015, the Brainwash dataset has appeared i {% include 'map.html' %} -Add more analysis here - +{% include 'citations.html' %} {% include 'supplementary_header.html' %} -{% include 'citations.html' %} + + -### Additional Information +#### Additional Resources - The dataset author spoke about his research at the CVPR conference in 2016 <https://www.youtube.com/watch?v=Nl2fBKxwusQ> +TODO + +- add bounding boxes to the header image +- remake montage with randomized images, with bboxes +- clean up intro text +- verify quote citations + + ### Footnotes [^readme]: "readme.txt" https://exhibits.stanford.edu/data/catalog/sx925dc9385. diff --git a/site/content/pages/datasets/duke_mtmc/assets/duke_mtmc_cam5_average_comp.jpg b/site/content/pages/datasets/duke_mtmc/assets/duke_mtmc_cam5_average_comp.jpg Binary files differnew file mode 100755 index 00000000..3cd64df1 --- /dev/null +++ b/site/content/pages/datasets/duke_mtmc/assets/duke_mtmc_cam5_average_comp.jpg diff --git a/site/content/pages/datasets/duke_mtmc/index.md b/site/content/pages/datasets/duke_mtmc/index.md index de1fa14c..c626ef4e 100644 --- a/site/content/pages/datasets/duke_mtmc/index.md +++ b/site/content/pages/datasets/duke_mtmc/index.md @@ -2,8 +2,8 @@ status: published title: Duke Multi-Target, Multi-Camera Tracking -desc: <span class="dataset-name">Duke MTMC</span> is a dataset of CCTV footage of students at Duke University -subdesc: Duke MTMC contains over 2 million video frames and 2,000 unique identities collected from 8 cameras at Duke University campus in March 2014 +desc: <span class="dataset-name">Duke MTMC</span> is a dataset of surveillance camera footage of students on Duke University campus +subdesc: Duke MTMC contains over 2 million video frames and 2,000 unique identities collected from 8 HD cameras at Duke University campus in March 2014 slug: duke_mtmc cssclass: dataset image: assets/background.jpg @@ -15,17 +15,27 @@ authors: Adam Harvey ### sidebar -+ Collected: March 19, 2014 -+ Cameras: 8 -+ Video Frames: 2,000,000 -+ Identities: Over 2,000 -+ Used for: Person re-identification, <br>face recognition -+ Sector: Academic ++ Created: 2014 ++ Identities: Over 2,700 ++ Used for: Face recognition, person re-identification ++ Created by: Computer Science Department, Duke University, Durham, US + Website: <a href="http://vision.cs.duke.edu/DukeMTMC/">duke.edu</a> ## Duke Multi-Target, Multi-Camera Tracking Dataset (Duke MTMC) -(PAGE UNDER DEVELOPMENT) +[ PAGE UNDER DEVELOPMENT ] + +Duke MTMC is a dataset of video recorded on Duke University campus during for the purpose of training, evaluating, and improving *multi-target multi-camera tracking*. The videos were recorded during February and March 2014 and cinclude + +Includes a total of 888.8 minutes of video (ind. verified) + +"We make available a new data set that has more than 2 million frames and more than 2,700 identities. It consists of 8×85 minutes of 1080p video recorded at 60 frames per second from 8 static cameras deployed on the Duke University campus during periods between lectures, when pedestrian traffic is heavy." + +The dataset includes approximately 2,000 annotated identities appearing in 85 hours of video from 8 cameras located throughout Duke University's campus. + + + +According to the dataset authors, {% include 'map.html' %} diff --git a/site/content/pages/datasets/index.md b/site/content/pages/datasets/index.md index 2e943fbe..c0373d60 100644 --- a/site/content/pages/datasets/index.md +++ b/site/content/pages/datasets/index.md @@ -13,4 +13,4 @@ sync: false # Facial Recognition Datasets -### Survey +Explore publicly available facial recognition datasets. More datasets will be added throughout 2019. diff --git a/site/content/pages/datasets/uccs/assets/uccs_bboxes_clr_fill.jpg b/site/content/pages/datasets/uccs/assets/uccs_bboxes_clr_fill.jpg Binary files differdeleted file mode 100644 index c8002bb9..00000000 --- a/site/content/pages/datasets/uccs/assets/uccs_bboxes_clr_fill.jpg +++ /dev/null diff --git a/site/content/pages/datasets/uccs/assets/uccs_bboxes_grayscale.jpg b/site/content/pages/datasets/uccs/assets/uccs_bboxes_grayscale.jpg Binary files differdeleted file mode 100644 index 6e2833dd..00000000 --- a/site/content/pages/datasets/uccs/assets/uccs_bboxes_grayscale.jpg +++ /dev/null diff --git a/site/content/pages/datasets/uccs/assets/uccs_mean_bboxes_comp.jpg b/site/content/pages/datasets/uccs/assets/uccs_mean_bboxes_comp.jpg Binary files differnew file mode 100644 index 00000000..18f4c5ec --- /dev/null +++ b/site/content/pages/datasets/uccs/assets/uccs_mean_bboxes_comp.jpg diff --git a/site/content/pages/datasets/uccs/index.md b/site/content/pages/datasets/uccs/index.md index 092638c0..e0925e07 100644 --- a/site/content/pages/datasets/uccs/index.md +++ b/site/content/pages/datasets/uccs/index.md @@ -2,9 +2,8 @@ status: published title: Unconstrained College Students -desc: <span class="dataset-name">Unconstrained College Students (UCCS)</span> is a dataset of images ... -subdesc: The UCCS dataset includes ... -slug: uccs +desc: <span class="dataset-name">Unconstrained College Students (UCCS)</span> is a dataset of long-range surveillance photos of students taken without their knowledge +subdesc: The UCCS dataset includes 16,149 images and 1,732 identities of students at University of Colorado Colorado Springs campus and is used for face recognition and face detection cssclass: dataset image: assets/background.jpg published: 2019-2-23 @@ -15,16 +14,22 @@ authors: Adam Harvey ### sidebar -+ Collected: TBD -+ Published: TBD -+ Images: TBD -+ Faces: TBD ++ Published: 2018 ++ Images: 16,149 ++ Identities: 1,732 ++ Used for: Face recognition, face detection ++ Created by: Unviversity of Colorado Colorado Springs (US) ++ Funded by: ODNI, IARPA, ONR MURI, Amry SBIR, SOCOM SBIR ++ Website: <a href="https://vast.uccs.edu/Opensetface/">vast.uccs.edu</a> ## Unconstrained College Students ... (PAGE UNDER DEVELOPMENT) + + + {% include 'map.html' %} {% include 'chart.html' %} @@ -36,7 +41,6 @@ authors: Adam Harvey {% include 'citations.html' %} - ### Research Notes @@ -55,4 +59,15 @@ The more recent UCCS version of the dataset received funding from [^funding_uccs [^funding_sb]: Sapkota, Archana and Boult, Terrance. "Large Scale Unconstrained Open Set Face Database." 2013. -[^funding_uccs]: Günther, M. et. al. "Unconstrained Face Detection and Open-Set Face Recognition Challenge," 2018. Arxiv 1708.02337v3.
\ No newline at end of file +[^funding_uccs]: Günther, M. et. al. "Unconstrained Face Detection and Open-Set Face Recognition Challenge," 2018. Arxiv 1708.02337v3. + + +" In most face detection/recognition datasets, the majority of images are “posed”, i.e. the subjects know they are being photographed, and/or the images are selected for publication in public media. Hence, blurry, occluded and badly illuminated images are generally uncommon in these datasets. In addition, most of these challenges are close-set, i.e. the list of subjects in the gallery is the same as the one used for testing. + +This challenge explores more unconstrained data, by introducing the new UnConstrained College Students (UCCS) dataset, where subjects are photographed using a long-range high-resolution surveillance camera without their knowledge. Faces inside these images are of various poses, and varied levels of blurriness and occlusion. The challenge also creates an open set recognition problem, where unknown people will be seen during testing and must be rejected. + +With this challenge, we hope to foster face detection and recognition research towards surveillance applications that are becoming more popular and more required nowadays, and where no automatic recognition algorithm has proven to be useful yet. + +UnConstrained College Students (UCCS) Dataset + +The UCCS dataset was collected over several months using Canon 7D camera fitted with Sigma 800mm F5.6 EX APO DG HSM lens, taking images at one frame per second, during times when many students were walking on the sidewalk. "
\ No newline at end of file diff --git a/site/content/pages/research/01_from_1_to_100_pixels/index.md b/site/content/pages/research/01_from_1_to_100_pixels/index.md index a7b863a9..b219dffb 100644 --- a/site/content/pages/research/01_from_1_to_100_pixels/index.md +++ b/site/content/pages/research/01_from_1_to_100_pixels/index.md @@ -56,3 +56,55 @@ Ideas: - "Note that we only keep the images with a minimal side length of 80 pixels." and "a face will be labeled as “Ignore” if it is very difficult to be detected due to blurring, severe deformation and unrecognizable eyes, or the side length of its bounding box is less than 32 pixels." Ge_Detecting_Masked_Faces_CVPR_2017_paper.pdf - IBM DiF: "Faces with region size less than 50x50 or inter-ocular distance of less than 30 pixels were discarded. Faces with non-frontal pose, or anything beyond being slightly tilted to the left or the right, were also discarded." + + + + +As the resolution +formatted as rectangular databases of 16 bit RGB-tuples or 8 bit grayscale values + + +To consider how visual privacy applies to real world surveillance situations, the first + +A single 8-bit grayscale pixel with 256 values is enough to represent the entire alphabet `a-Z0-9` with room to spare. + +A 2x2 pixels contains + +Using no more than a 42 pixel (6x7 image) face image researchers [cite] were able to correctly distinguish between a group of 50 people. Yet + +The likely outcome of face recognition research is that more data is needed to improve. Indeed, resolution is the determining factor for all biometric systems, both as training data to increase + +Pixels, typically considered the buiding blocks of images and vidoes, can also be plotted as a graph of sensor values corresponding to the intensity of RGB-calibrated sensors. + + +Wi-Fi and cameras presents elevated risks for transmitting videos and image documentation from conflict zones, high-risk situations, or even sharing on social media. How can new developments in computer vision also be used in reverse, as a counter-forensic tool, to minimize an individual's privacy risk? + +As the global Internet becomes increasingly effecient at turning the Internet into a giant dataset for machine learning, forensics, and data analysing, it would be prudent to also consider tools for decreasing the resolution. The Visual Defense module is just that. What are new ways to minimize the adverse effects of surveillance by dulling the blade. For example, a researcher paper showed that by decreasing a face size to 12x16 it was possible to do 98% accuracy with 50 people. This is clearly an example of + +This research module, tentatively called Visual Defense Tools, aims to explore the + + +### Prior Research + +- MPI visual privacy advisor +- NIST: super resolution +- YouTube blur tool +- WITNESS: blur tool +- Pixellated text +- CV Dazzle +- Bellingcat guide to geolocation +- Peng! magic passport + +### Notes + +- In China, out of the approximately 200 million surveillance cameras only about 15% have enough resolution for face recognition. +- In Apple's FaceID security guide, the probability of someone else's face unlocking your phone is 1 out of 1,000,000. +- In England, the Metropolitan Police reported a false-positive match rate of 98% when attempting to use face recognition to locate wanted criminals. +- In a face recognition trial at Berlin's Sudkreuz station, the false-match rate was 20%. + + +What all 3 examples illustrate is that face recognition is anything but absolute. In a 2017 talk, Jason Matheny the former directory of IARPA, admitted the face recognition is so brittle it can be subverted by using a magic marker and drawing "a few dots on your forehead". In fact face recognition is a misleading term. Face recognition is search engine for faces that can only ever show you the mos likely match. This presents real a real threat to privacy and lends + + +Globally, iPhone users unwittingly agree to 1/1,000,000 probably +relying on FaceID and TouchID to protect their information agree to a
\ No newline at end of file diff --git a/site/content/pages/research/02_what_computers_can_see/index.md b/site/content/pages/research/02_what_computers_can_see/index.md index ab4c7884..51621f46 100644 --- a/site/content/pages/research/02_what_computers_can_see/index.md +++ b/site/content/pages/research/02_what_computers_can_see/index.md @@ -100,6 +100,7 @@ A list of 100 things computer vision can see, eg: - Wearing Necktie - Wearing Necklace +for i in {1..9};do wget http://visiond1.cs.umbc.edu/webpage/codedata/ADLdataset/ADL_videos/P_0$i.MP4;done;for i in {10..20}; do wget http://visiond1.cs.umbc.edu/webpage/codedata/ADLdataset/ADL_videos/P_$i.MP4;done ## From Market 1501 @@ -149,4 +150,26 @@ Visibility boolean for each keypoint Region annotations (upper clothes, lower clothes, dress, socks, shoes, hands, gloves, neck, face, hair, hat, sunglasses, bag, occluder) Body type (male, female or child) -source: https://www2.eecs.berkeley.edu/Research/Projects/CS/vision/shape/h3d/
\ No newline at end of file +source: https://www2.eecs.berkeley.edu/Research/Projects/CS/vision/shape/h3d/ + +## From Leeds Sports Pose + +=INDEX(A2:A9,MATCH(datasets!D1,B2:B9,0)) +=VLOOKUP(A2, datasets!A:J, 7, FALSE) + +Right ankle +Right knee +Right hip +Left hip +Left knee +Left ankle +Right wrist +Right elbow +Right shoulder +Left shoulder +Left elbow +Left wrist +Neck +Head top + +source: http://web.archive.org/web/20170915023005/sam.johnson.io/research/lsp.html
\ No newline at end of file diff --git a/site/includes/chart.html b/site/includes/chart.html index 45c13493..01c2e83b 100644 --- a/site/includes/chart.html +++ b/site/includes/chart.html @@ -2,8 +2,7 @@ <h3>Who used {{ metadata.meta.dataset.name_display }}?</h3> <p> - This bar chart presents a ranking of the top countries where citations originated. Mouse over individual columns - to see yearly totals. These charts show at most the top 10 countries. + This bar chart presents a ranking of the top countries where dataset citations originated. Mouse over individual columns to see yearly totals. These charts show at most the top 10 countries. </p> </section> diff --git a/site/includes/citations.html b/site/includes/citations.html index ebd37d61..32558d4a 100644 --- a/site/includes/citations.html +++ b/site/includes/citations.html @@ -1,9 +1,8 @@ <section class="applet_container"> - <h3>Citations</h3> + <h3>Dataset Citations</h3> <p> - Citations were collected from <a href="https://www.semanticscholar.org">Semantic Scholar</a>, a website which aggregates - and indexes research papers. The citations were geocoded using names of institutions found in the PDF front matter, or as listed on other resources. These papers have been manually verified to show that researchers downloaded and used the dataset to train or test machine learning algorithms. + The dataset citations used in the visualizations were collected from <a href="https://www.semanticscholar.org">Semantic Scholar</a>, a website which aggregates and indexes research papers. Each citation was geocoded using names of institutions found in the PDF front matter, or as listed on other resources. These papers have been manually verified to show that researchers downloaded and used the dataset to train or test machine learning algorithms. </p> <p> Add [button/link] to download CSV. Add search input field to filter. diff --git a/site/includes/map.html b/site/includes/map.html index 74771768..30c248a6 100644 --- a/site/includes/map.html +++ b/site/includes/map.html @@ -1,6 +1,6 @@ <section> - <h3>Information Supply Chain</h3> + <h3>Biometric Trade Routes</h3> <!-- <div class="map-sidebar right-sidebar"> <h3>Legend</h3> @@ -12,27 +12,28 @@ </div> --> <p> - To understand how {{ metadata.meta.dataset.name_display }} has been used around the world... - affected global research on computer vision, surveillance, defense, and consumer technology, the and where this dataset has been used the locations of each organization that used or referenced the datast + To help understand how {{ metadata.meta.dataset.name_display }} has been used around the world for commercial, military and academic research; publicly available research citing {{ metadata.meta.dataset.name_full} is collected, verified, and geocoded to show the biometric trade routes of people appearing in the images. Click on the markers to reveal reserach projects at that location. </p> </section> -<section class="applet_container"> +<section class="applet_container fullwidth"> <div class="applet" data-payload="{"command": "map"}"></div> + </section> <div class="caption"> <ul class="map-legend"> <li class="edu">Academic</li> - <li class="com">Industry</li> - <li class="gov">Government / Military</li> + <li class="com">Commercial</li> + <li class="gov">Military / Government</li> <li class="source">Citation data is collected using <a href="https://semanticscholar.org" target="_blank">SemanticScholar.org</a> then dataset usage verified and geolocated.</li> </ul> </div> -<section> +<!-- <section> <p class='subp'> [section under development] {{ metadata.meta.dataset.name_display }} ... Standardized paragraph of text about the map. Sed ut perspiciatis, unde omnis iste natus error sit voluptatem accusantium doloremque laudantium, totam rem aperiam eaque ipsa, quae ab illo inventore veritatis et quasi architecto beatae vitae dicta sunt, explicabo. </p> </section> + -->
\ No newline at end of file diff --git a/site/includes/piechart.html b/site/includes/piechart.html index e739bb28..94c8aae7 100644 --- a/site/includes/piechart.html +++ b/site/includes/piechart.html @@ -1,10 +1,3 @@ -<section> - <p> - These pie charts show overall totals based on country and institution type. - </p> - - </section> - <section class="applet_container"> <div class="applet" data-payload="{"command": "piechart"}"></div> </section> diff --git a/site/includes/supplementary_header.html b/site/includes/supplementary_header.html index 5fd4b2b4..bcd84223 100644 --- a/site/includes/supplementary_header.html +++ b/site/includes/supplementary_header.html @@ -6,5 +6,6 @@ <div class="hr-wave-line hr-wave-line2"></div> </div> - <h2>Supplementary Information</h2> + <h3>Supplementary Information</h3> + </section> |
