diff options
Diffstat (limited to 'site/public/datasets')
22 files changed, 837 insertions, 444 deletions
diff --git a/site/public/datasets/50_people_one_question/index.html b/site/public/datasets/50_people_one_question/index.html index bded7fbd..796af8d6 100644 --- a/site/public/datasets/50_people_one_question/index.html +++ b/site/public/datasets/50_people_one_question/index.html @@ -8,7 +8,6 @@ <meta name="referrer" content="no-referrer" /> <meta name="viewport" content="width=device-width, initial-scale=1.0, user-scalable=yes" /> <link rel='stylesheet' href='/assets/css/fonts.css' /> - <link rel='stylesheet' href='/assets/css/tabulator.css' /> <link rel='stylesheet' href='/assets/css/css.css' /> <link rel='stylesheet' href='/assets/css/leaflet.css' /> <link rel='stylesheet' href='/assets/css/applets.css' /> @@ -18,6 +17,7 @@ <a class='slogan' href="/"> <div class='logo'></div> <div class='site_name'>MegaPixels</div> + <div class='splash'>50 People One Question</div> </a> <div class='links'> <a href="/datasets/">Datasets</a> @@ -27,13 +27,25 @@ <div class="content content-dataset"> <section class='intro_section' style='background-image: url(https://nyc3.digitaloceanspaces.com/megapixels/v1/datasets/50_people_one_question/assets/background.jpg)'><div class='inner'><div class='hero_desc'><span class='bgpad'><span style="color:#ffaa00">People One Question</span> is a dataset of people from an online video series on YouTube and Vimeo used for building facial recogntion algorithms</span></div><div class='hero_subdesc'><span class='bgpad'>People One Question dataset includes ... -</span></div></div></section><section><div class='left-sidebar'><div class='meta'><div><div class='gray'>Collected</div><div>TBD</div></div><div><div class='gray'>Published</div><div>TBD</div></div><div><div class='gray'>Images</div><div>TBD</div></div><div><div class='gray'>Faces</div><div>TBD</div></div></div></div><h2>50 People 1 Question</h2> +</span></div></div></section><section><div class='left-sidebar'><div class='meta'> + <div class='gray'>Published</div> + <div>2013</div> + </div><div class='meta'> + <div class='gray'>Videos</div> + <div>33 </div> + </div><div class='meta'> + <div class='gray'>Purpose</div> + <div>Facial landmark estimation in the wild</div> + </div><div class='meta'> + <div class='gray'>Website</div> + <div><a href='http://www.vision.caltech.edu/~dhall/projects/MergingPoseEstimates/' target='_blank' rel='nofollow noopener'>caltech.edu</a></div> + </div><div class='meta'><div><div class='gray'>Collected</div><div>TBD</div></div><div><div class='gray'>Published</div><div>TBD</div></div><div><div class='gray'>Images</div><div>TBD</div></div><div><div class='gray'>Faces</div><div>TBD</div></div></div></div><h2>50 People 1 Question</h2> <p>(PAGE UNDER DEVELOPMENT)</p> <p>At vero eos et accusamus et iusto odio dignissimos ducimus, qui blanditiis praesentium voluptatum deleniti atque corrupti, quos dolores et quas molestias excepturi sint, obcaecati cupiditate non-provident, similique sunt in culpa, qui officia deserunt mollitia animi, id est laborum et dolorum fuga. Et harum quidem rerum facilis est et expedita distinctio.</p> <p>Nam libero tempore, cum soluta nobis est eligendi optio, cumque nihil impedit, quo minus id, quod maxime placeat, facere possimus, omnis voluptas assumenda est, omnis dolor repellendus. Temporibus autem quibusdam et aut officiis debitis aut rerum necessitatibus saepe eveniet, ut et voluptates repudiandae sint et molestiae non-recusandae. Itaque earum rerum hic tenetur a sapiente delectus, ut aut reiciendis voluptatibus maiores alias consequatur aut perferendis doloribus asperiores repellat</p> </section><section> - <h3>Information Supply Chain</h3> + <h3>Biometric Trade Routes</h3> <!-- <div class="map-sidebar right-sidebar"> <h3>Legend</h3> @@ -45,45 +57,44 @@ </div> --> <p> - To understand how 50 People One Question Dataset has been used around the world... - affected global research on computer vision, surveillance, defense, and consumer technology, the and where this dataset has been used the locations of each organization that used or referenced the datast + To help understand how 50 People One Question Dataset has been used around the world for commercial, military and academic research; publicly available research citing 50 People One Question is collected, verified, and geocoded to show the biometric trade routes of people appearing in the images. Click on the markers to reveal reserach projects at that location. </p> </section> -<section class="applet_container"> +<section class="applet_container fullwidth"> <div class="applet" data-payload="{"command": "map"}"></div> + </section> <div class="caption"> - <div class="map-legend-item edu">Academic</div> - <div class="map-legend-item com">Industry</div> - <div class="map-legend-item gov">Government</div> - Data is compiled from <a href="https://www.semanticscholar.org">Semantic Scholar</a> and has been manually verified to show usage of 50 People One Question Dataset. + <ul class="map-legend"> + <li class="edu">Academic</li> + <li class="com">Commercial</li> + <li class="gov">Military / Government</li> + <li class="source">Citation data is collected using <a href="https://semanticscholar.org" target="_blank">SemanticScholar.org</a> then dataset usage verified and geolocated.</li> + </ul> </div> -<section> +<!-- <section> <p class='subp'> - Standardized paragraph of text about the map. Sed ut perspiciatis, unde omnis iste natus error sit voluptatem accusantium doloremque laudantium, totam rem aperiam eaque ipsa, quae ab illo inventore veritatis et quasi architecto beatae vitae dicta sunt, explicabo. + [section under development] 50 People One Question Dataset ... Standardized paragraph of text about the map. Sed ut perspiciatis, unde omnis iste natus error sit voluptatem accusantium doloremque laudantium, totam rem aperiam eaque ipsa, quae ab illo inventore veritatis et quasi architecto beatae vitae dicta sunt, explicabo. </p> -</section><section> - +</section> + --><section> <div class="hr-wave-holder"> <div class="hr-wave-line hr-wave-line1"></div> <div class="hr-wave-line hr-wave-line2"></div> </div> - <h2>Supplementary Information</h2> + <h3>Supplementary Information</h3> + </section><section class="applet_container"> - <h3>Citations</h3> - <p> - Citations were collected from <a href="https://www.semanticscholar.org">Semantic Scholar</a>, a website which aggregates - and indexes research papers. The citations were geocoded using names of institutions found in the PDF front matter, or as listed on other resources. These papers have been manually verified to show that researchers downloaded and used the dataset to train and/or test machine learning algorithms. - </p> + <h3>Dataset Citations</h3> <p> - Add button/link to download CSV + The dataset citations used in the visualizations were collected from <a href="https://www.semanticscholar.org">Semantic Scholar</a>, a website which aggregates and indexes research papers. Each citation was geocoded using names of institutions found in the PDF front matter, or as listed on other resources. These papers have been manually verified to show that researchers downloaded and used the dataset to train or test machine learning algorithms. </p> <div class="applet" data-payload="{"command": "citations"}"></div> diff --git a/site/public/datasets/afad/index.html b/site/public/datasets/afad/index.html index a0aea1a6..ac025a80 100644 --- a/site/public/datasets/afad/index.html +++ b/site/public/datasets/afad/index.html @@ -8,7 +8,6 @@ <meta name="referrer" content="no-referrer" /> <meta name="viewport" content="width=device-width, initial-scale=1.0, user-scalable=yes" /> <link rel='stylesheet' href='/assets/css/fonts.css' /> - <link rel='stylesheet' href='/assets/css/tabulator.css' /> <link rel='stylesheet' href='/assets/css/css.css' /> <link rel='stylesheet' href='/assets/css/leaflet.css' /> <link rel='stylesheet' href='/assets/css/applets.css' /> @@ -18,6 +17,7 @@ <a class='slogan' href="/"> <div class='logo'></div> <div class='site_name'>MegaPixels</div> + <div class='splash'>AFAD</div> </a> <div class='links'> <a href="/datasets/">Datasets</a> diff --git a/site/public/datasets/aflw/index.html b/site/public/datasets/aflw/index.html index 7aaa9af0..476f390c 100644 --- a/site/public/datasets/aflw/index.html +++ b/site/public/datasets/aflw/index.html @@ -8,7 +8,6 @@ <meta name="referrer" content="no-referrer" /> <meta name="viewport" content="width=device-width, initial-scale=1.0, user-scalable=yes" /> <link rel='stylesheet' href='/assets/css/fonts.css' /> - <link rel='stylesheet' href='/assets/css/tabulator.css' /> <link rel='stylesheet' href='/assets/css/css.css' /> <link rel='stylesheet' href='/assets/css/leaflet.css' /> <link rel='stylesheet' href='/assets/css/applets.css' /> @@ -18,6 +17,7 @@ <a class='slogan' href="/"> <div class='logo'></div> <div class='site_name'>MegaPixels</div> + </a> <div class='links'> <a href="/datasets/">Datasets</a> diff --git a/site/public/datasets/brainwash/index.html b/site/public/datasets/brainwash/index.html index 41484257..ec5ee434 100644 --- a/site/public/datasets/brainwash/index.html +++ b/site/public/datasets/brainwash/index.html @@ -4,11 +4,10 @@ <title>MegaPixels</title> <meta charset="utf-8" /> <meta name="author" content="Adam Harvey" /> - <meta name="description" content="Brainwash is a dataset of webcam images taken from the Brainwash Cafe in San Francisco" /> + <meta name="description" content="Brainwash is a dataset of webcam images taken from the Brainwash Cafe in San Francisco in 2014" /> <meta name="referrer" content="no-referrer" /> <meta name="viewport" content="width=device-width, initial-scale=1.0, user-scalable=yes" /> <link rel='stylesheet' href='/assets/css/fonts.css' /> - <link rel='stylesheet' href='/assets/css/tabulator.css' /> <link rel='stylesheet' href='/assets/css/css.css' /> <link rel='stylesheet' href='/assets/css/leaflet.css' /> <link rel='stylesheet' href='/assets/css/applets.css' /> @@ -18,6 +17,7 @@ <a class='slogan' href="/"> <div class='logo'></div> <div class='site_name'>MegaPixels</div> + <div class='splash'>Brainwash</div> </a> <div class='links'> <a href="/datasets/">Datasets</a> @@ -26,36 +26,50 @@ </header> <div class="content content-dataset"> - <section class='intro_section' style='background-image: url(https://nyc3.digitaloceanspaces.com/megapixels/v1/datasets/brainwash/assets/background.jpg)'><div class='inner'><div class='hero_desc'><span class='bgpad'>Brainwash is a dataset of webcam images taken from the Brainwash Cafe in San Francisco</span></div><div class='hero_subdesc'><span class='bgpad'>The Brainwash dataset includes 11,918 images of "everyday life of a busy downtown cafe" and is used for training head detection algorithms -</span></div></div></section><section><div class='left-sidebar'><div class='meta'><div><div class='gray'>Published</div><div>2015</div></div><div><div class='gray'>Images</div><div>11,918</div></div><div><div class='gray'>Faces</div><div>91,146</div></div><div><div class='gray'>Created by</div><div>Stanford Department of Computer Science</div></div><div><div class='gray'>Funded by</div><div>Max Planck Center for Visual Computing and Communication</div></div><div><div class='gray'>Location</div><div>Brainwash Cafe, San Franscisco</div></div><div><div class='gray'>Purpose</div><div>Training face detection</div></div><div><div class='gray'>Website</div><div><a href="https://exhibits.stanford.edu/data/catalog/sx925dc9385">stanford.edu</a></div></div><div><div class='gray'>Paper</div><div><a href="http://arxiv.org/abs/1506.04878">End-to-End People Detection in Crowded Scenes</a></div></div><div><div class='gray'>Explicit Consent</div><div>No</div></div></div></div><h2>Brainwash Dataset</h2> -<p>(PAGE UNDER DEVELOPMENT)</p> -<p><em>Brainwash</em> is a face detection dataset created from the Brainwash Cafe's livecam footage including 11,918 images of "everyday life of a busy downtown cafe<a class="footnote_shim" name="[^readme]_1"> </a><a href="#[^readme]" class="footnote" title="Footnote 1">1</a>". The images are used to develop face detection algorithms for the "challenging task of detecting people in crowded scenes" and tracking them.</p> -<p>Before closing in 2017, Brainwash Cafe was a "cafe and laundromat" located in San Francisco's SoMA district. The cafe published a publicy available livestream from the cafe with a view of the cash register, performance stage, and seating area.</p> -<p>Since it's publication by Stanford in 2015, the Brainwash dataset has appeared in several notable research papers. In September 2016 four researchers from the National University of Defense Technology in Changsha, China used the Brainwash dataset for a research study on "people head detection in crowded scenes", concluding that their algorithm "achieves superior head detection performance on the crowded scenes dataset<a class="footnote_shim" name="[^localized_region_context]_1"> </a><a href="#[^localized_region_context]" class="footnote" title="Footnote 2">2</a>". And again in 2017 three researchers at the National University of Defense Technology used Brainwash for a study on object detection noting "the data set used in our experiment is shown in Table 1, which includes one scene of the brainwash dataset<a class="footnote_shim" name="[^replacement_algorithm]_1"> </a><a href="#[^replacement_algorithm]" class="footnote" title="Footnote 3">3</a>".</p> -</section><section class='images'><div class='image'><img src='https://nyc3.digitaloceanspaces.com/megapixels/v1/datasets/brainwash/assets/00425000_960.jpg' alt=' An sample image from the Brainwash dataset used for training face and head detection algorithms for surveillance. The datset contains about 12,000 images. License: Open Data Commons Public Domain Dedication (PDDL)'><div class='caption'> An sample image from the Brainwash dataset used for training face and head detection algorithms for surveillance. The datset contains about 12,000 images. License: Open Data Commons Public Domain Dedication (PDDL)</div></div></section><section class='images'><div class='image'><img src='https://nyc3.digitaloceanspaces.com/megapixels/v1/datasets/brainwash/assets/brainwash_montage.jpg' alt=' 49 of the 11,918 images included in the Brainwash dataset. License: Open Data Commons Public Domain Dedication (PDDL)'><div class='caption'> 49 of the 11,918 images included in the Brainwash dataset. License: Open Data Commons Public Domain Dedication (PDDL)</div></div></section><section> + <section class='intro_section' style='background-image: url(https://nyc3.digitaloceanspaces.com/megapixels/v1/datasets/brainwash/assets/background.jpg)'><div class='inner'><div class='hero_desc'><span class='bgpad'>Brainwash is a dataset of webcam images taken from the Brainwash Cafe in San Francisco in 2014</span></div><div class='hero_subdesc'><span class='bgpad'>The Brainwash dataset includes 11,918 images of "everyday life of a busy downtown cafe" and is used for training head detection surveillance algorithms +</span></div></div></section><section><div class='left-sidebar'><div class='meta'> + <div class='gray'>Published</div> + <div>2015</div> + </div><div class='meta'> + <div class='gray'>Images</div> + <div>11,917 </div> + </div><div class='meta'> + <div class='gray'>Purpose</div> + <div>Head detection</div> + </div><div class='meta'> + <div class='gray'>Created by</div> + <div>Stanford University (US), Max Planck Institute for Informatics (DE)</div> + </div><div class='meta'> + <div class='gray'>Funded by</div> + <div>Max Planck Center for Visual Computing and Communication</div> + </div><div class='meta'> + <div class='gray'>Download Size</div> + <div>4.1 GB</div> + </div><div class='meta'> + <div class='gray'>Website</div> + <div><a href='https://purl.stanford.edu/sx925dc9385' target='_blank' rel='nofollow noopener'>stanford.edu</a></div> + </div></div><h2>Brainwash Dataset</h2> +<p><em>Brainwash</em> is a head detection dataset created from San Francisco's Brainwash Cafe livecam footage. It includes 11,918 images of "everyday life of a busy downtown cafe"<a class="footnote_shim" name="[^readme]_1"> </a><a href="#[^readme]" class="footnote" title="Footnote 1">1</a> captured at 100 second intervals throught the entire day. Brainwash dataset was captured during 3 days in 2014: October 27, November 13, and November 24. According the author's reserach paper introducing the dataset, the images were acquired with the help of Angelcam.com [cite orig paper].</p> +<p>Brainwash is not a widely used dataset but since its publication by Stanford University in 2015, it has notably appeared in several research papers from the National University of Defense Technology in Changsha, China. In 2016 and in 2017 researchers there conducted studies on detecting people's heads in crowded scenes for the purpose of surveillance <a class="footnote_shim" name="[^localized_region_context]_1"> </a><a href="#[^localized_region_context]" class="footnote" title="Footnote 2">2</a> <a class="footnote_shim" name="[^replacement_algorithm]_1"> </a><a href="#[^replacement_algorithm]" class="footnote" title="Footnote 3">3</a>.</p> +<p>If you happen to have been at Brainwash cafe in San Franscisco at any time on October 26, November 13, or November 24 in 2014 you are most likely included in the Brainwash dataset.</p> +</section><section class='images'><div class='image'><img src='https://nyc3.digitaloceanspaces.com/megapixels/v1/datasets/brainwash/assets/brainwash_mean_overlay.jpg' alt=' The pixel-averaged image of all Brainwash dataset images is shown with 81,973 head annotations drawn from the Brainwash training partition. (c) Adam Harvey'><div class='caption'> The pixel-averaged image of all Brainwash dataset images is shown with 81,973 head annotations drawn from the Brainwash training partition. (c) Adam Harvey</div></div></section><section> <h3>Who used Brainwash Dataset?</h3> <p> - This bar chart presents a ranking of the top countries where citations originated. Mouse over individual columns - to see yearly totals. These charts show at most the top 10 countries. + This bar chart presents a ranking of the top countries where dataset citations originated. Mouse over individual columns to see yearly totals. These charts show at most the top 10 countries. </p> </section> <section class="applet_container"> +<!-- <div style="position: absolute;top: 0px;right: -55px;width: 180px;font-size: 14px;">Labeled Faces in the Wild Dataset<br><span class="numc" style="font-size: 11px;">20 citations</span> +</div> --> <div class="applet" data-payload="{"command": "chart"}"></div> -</section><section> - <p> - These pie charts show overall totals based on country and institution type. - </p> - - </section> - -<section class="applet_container"> +</section><section class="applet_container"> <div class="applet" data-payload="{"command": "piechart"}"></div> </section><section> - <h3>Information Supply Chain</h3> + <h3>Biometric Trade Routes</h3> <!-- <div class="map-sidebar right-sidebar"> <h3>Legend</h3> @@ -67,53 +81,58 @@ </div> --> <p> - To understand how Brainwash Dataset has been used around the world... - affected global research on computer vision, surveillance, defense, and consumer technology, the and where this dataset has been used the locations of each organization that used or referenced the datast + To help understand how Brainwash Dataset has been used around the world for commercial, military and academic research; publicly available research citing Brainwash Dataset is collected, verified, and geocoded to show the biometric trade routes of people appearing in the images. Click on the markers to reveal reserach projects at that location. </p> </section> -<section class="applet_container"> +<section class="applet_container fullwidth"> <div class="applet" data-payload="{"command": "map"}"></div> + </section> <div class="caption"> - <div class="map-legend-item edu">Academic</div> - <div class="map-legend-item com">Industry</div> - <div class="map-legend-item gov">Government</div> - Data is compiled from <a href="https://www.semanticscholar.org">Semantic Scholar</a> and has been manually verified to show usage of Brainwash Dataset. + <ul class="map-legend"> + <li class="edu">Academic</li> + <li class="com">Commercial</li> + <li class="gov">Military / Government</li> + <li class="source">Citation data is collected using <a href="https://semanticscholar.org" target="_blank">SemanticScholar.org</a> then dataset usage verified and geolocated.</li> + </ul> </div> -<section> +<!-- <section> <p class='subp'> - Standardized paragraph of text about the map. Sed ut perspiciatis, unde omnis iste natus error sit voluptatem accusantium doloremque laudantium, totam rem aperiam eaque ipsa, quae ab illo inventore veritatis et quasi architecto beatae vitae dicta sunt, explicabo. + [section under development] Brainwash Dataset ... Standardized paragraph of text about the map. Sed ut perspiciatis, unde omnis iste natus error sit voluptatem accusantium doloremque laudantium, totam rem aperiam eaque ipsa, quae ab illo inventore veritatis et quasi architecto beatae vitae dicta sunt, explicabo. </p> -</section><section><p>Add more analysis here</p> -</section><section> +</section> + --><section class="applet_container"> + <h3>Dataset Citations</h3> + <p> + The dataset citations used in the visualizations were collected from <a href="https://www.semanticscholar.org">Semantic Scholar</a>, a website which aggregates and indexes research papers. Each citation was geocoded using names of institutions found in the PDF front matter, or as listed on other resources. These papers have been manually verified to show that researchers downloaded and used the dataset to train or test machine learning algorithms. + </p> + + <div class="applet" data-payload="{"command": "citations"}"></div> +</section><section> <div class="hr-wave-holder"> <div class="hr-wave-line hr-wave-line1"></div> <div class="hr-wave-line hr-wave-line2"></div> </div> - <h2>Supplementary Information</h2> -</section><section class="applet_container"> - - <h3>Citations</h3> - <p> - Citations were collected from <a href="https://www.semanticscholar.org">Semantic Scholar</a>, a website which aggregates - and indexes research papers. The citations were geocoded using names of institutions found in the PDF front matter, or as listed on other resources. These papers have been manually verified to show that researchers downloaded and used the dataset to train and/or test machine learning algorithms. - </p> - <p> - Add button/link to download CSV - </p> - - <div class="applet" data-payload="{"command": "citations"}"></div> -</section><section><h3>Additional Information</h3> + <h3>Supplementary Information</h3> + +</section><section class='images'><div class='image'><img src='https://nyc3.digitaloceanspaces.com/megapixels/v1/datasets/brainwash/assets/00425000_960.jpg' alt=' An sample image from the Brainwash dataset used for training face and head detection algorithms for surveillance. The datset contains about 12,000 images. License: Open Data Commons Public Domain Dedication (PDDL)'><div class='caption'> An sample image from the Brainwash dataset used for training face and head detection algorithms for surveillance. The datset contains about 12,000 images. License: Open Data Commons Public Domain Dedication (PDDL)</div></div></section><section class='images'><div class='image'><img src='https://nyc3.digitaloceanspaces.com/megapixels/v1/datasets/brainwash/assets/brainwash_montage.jpg' alt=' 49 of the 11,918 images included in the Brainwash dataset. License: Open Data Commons Public Domain Dedication (PDDL)'><div class='caption'> 49 of the 11,918 images included in the Brainwash dataset. License: Open Data Commons Public Domain Dedication (PDDL)</div></div></section><section><h4>Additional Resources</h4> <ul> <li>The dataset author spoke about his research at the CVPR conference in 2016 <a href="https://www.youtube.com/watch?v=Nl2fBKxwusQ">https://www.youtube.com/watch?v=Nl2fBKxwusQ</a></li> </ul> +<p>TODO</p> +<ul> +<li>add bounding boxes to the header image</li> +<li>remake montage with randomized images, with bboxes</li> +<li>clean up intro text</li> +<li>verify quote citations</li> +</ul> </section><section><ul class="footnotes"><li><a name="[^readme]" class="footnote_shim"></a><span class="backlinks"><a href="#[^readme]_1">a</a></span><p>"readme.txt" <a href="https://exhibits.stanford.edu/data/catalog/sx925dc9385">https://exhibits.stanford.edu/data/catalog/sx925dc9385</a>.</p> </li><li><a name="[^localized_region_context]" class="footnote_shim"></a><span class="backlinks"><a href="#[^localized_region_context]_1">a</a></span><p>Li, Y. and Dou, Y. and Liu, X. and Li, T. Localized Region Context and Object Feature Fusion for People Head Detection. ICIP16 Proceedings. 2016. Pages 594-598.</p> </li><li><a name="[^replacement_algorithm]" class="footnote_shim"></a><span class="backlinks"><a href="#[^replacement_algorithm]_1">a</a></span><p>Zhao. X, Wang Y, Dou, Y. A Replacement Algorithm of Non-Maximum Suppression Base on Graph Clustering.</p> diff --git a/site/public/datasets/caltech_10k/index.html b/site/public/datasets/caltech_10k/index.html index 6615bb1a..9aa0b2c3 100644 --- a/site/public/datasets/caltech_10k/index.html +++ b/site/public/datasets/caltech_10k/index.html @@ -8,7 +8,6 @@ <meta name="referrer" content="no-referrer" /> <meta name="viewport" content="width=device-width, initial-scale=1.0, user-scalable=yes" /> <link rel='stylesheet' href='/assets/css/fonts.css' /> - <link rel='stylesheet' href='/assets/css/tabulator.css' /> <link rel='stylesheet' href='/assets/css/css.css' /> <link rel='stylesheet' href='/assets/css/leaflet.css' /> <link rel='stylesheet' href='/assets/css/applets.css' /> @@ -18,6 +17,7 @@ <a class='slogan' href="/"> <div class='logo'></div> <div class='site_name'>MegaPixels</div> + </a> <div class='links'> <a href="/datasets/">Datasets</a> diff --git a/site/public/datasets/celeba/index.html b/site/public/datasets/celeba/index.html index 09347f10..26a43803 100644 --- a/site/public/datasets/celeba/index.html +++ b/site/public/datasets/celeba/index.html @@ -8,7 +8,6 @@ <meta name="referrer" content="no-referrer" /> <meta name="viewport" content="width=device-width, initial-scale=1.0, user-scalable=yes" /> <link rel='stylesheet' href='/assets/css/fonts.css' /> - <link rel='stylesheet' href='/assets/css/tabulator.css' /> <link rel='stylesheet' href='/assets/css/css.css' /> <link rel='stylesheet' href='/assets/css/leaflet.css' /> <link rel='stylesheet' href='/assets/css/applets.css' /> @@ -18,6 +17,7 @@ <a class='slogan' href="/"> <div class='logo'></div> <div class='site_name'>MegaPixels</div> + <div class='splash'>CelebA</div> </a> <div class='links'> <a href="/datasets/">Datasets</a> @@ -27,13 +27,31 @@ <div class="content content-dataset"> <section class='intro_section' style='background-image: url(https://nyc3.digitaloceanspaces.com/megapixels/v1/datasets/celeba/assets/background.jpg)'><div class='inner'><div class='hero_desc'><span class='bgpad'><span style="color:#ffaa00">CelebA</span> is a dataset of people...</span></div><div class='hero_subdesc'><span class='bgpad'>CelebA includes... -</span></div></div></section><section><div class='left-sidebar'><div class='meta'><div><div class='gray'>Collected</div><div>TBD</div></div><div><div class='gray'>Published</div><div>TBD</div></div><div><div class='gray'>Images</div><div>TBD</div></div><div><div class='gray'>Faces</div><div>TBD</div></div></div></div><h2>CelebA</h2> +</span></div></div></section><section><div class='left-sidebar'><div class='meta'> + <div class='gray'>Published</div> + <div>2015</div> + </div><div class='meta'> + <div class='gray'>Images</div> + <div>202,599 </div> + </div><div class='meta'> + <div class='gray'>Identities</div> + <div>10,177 </div> + </div><div class='meta'> + <div class='gray'>Purpose</div> + <div>face attribute recognition, face detection, and landmark (or facial part) localization</div> + </div><div class='meta'> + <div class='gray'>Download Size</div> + <div>1.4 GB</div> + </div><div class='meta'> + <div class='gray'>Website</div> + <div><a href='http://mmlab.ie.cuhk.edu.hk/projects/CelebA.html' target='_blank' rel='nofollow noopener'>edu.hk</a></div> + </div><div class='meta'><div><div class='gray'>Collected</div><div>TBD</div></div><div><div class='gray'>Published</div><div>TBD</div></div><div><div class='gray'>Images</div><div>TBD</div></div><div><div class='gray'>Faces</div><div>TBD</div></div></div></div><h2>CelebA</h2> <p>(PAGE UNDER DEVELOPMENT)</p> <p>At vero eos et accusamus et iusto odio dignissimos ducimus, qui blanditiis praesentium voluptatum deleniti atque corrupti, quos dolores et quas molestias excepturi sint, obcaecati cupiditate non-provident, similique sunt in culpa, qui officia deserunt mollitia animi, id est laborum et dolorum fuga. Et harum quidem rerum facilis est et expedita distinctio.</p> <p>Nam libero tempore, cum soluta nobis est eligendi optio, cumque nihil impedit, quo minus id, quod maxime placeat, facere possimus, omnis voluptas assumenda est, omnis dolor repellendus. Temporibus autem quibusdam et aut officiis debitis aut rerum necessitatibus saepe eveniet, ut et voluptates repudiandae sint et molestiae non-recusandae. Itaque earum rerum hic tenetur a sapiente delectus, ut aut reiciendis voluptatibus maiores alias consequatur aut perferendis doloribus asperiores repellat</p> </section><section> - <h3>Information Supply Chain</h3> + <h3>Biometric Trade Routes</h3> <!-- <div class="map-sidebar right-sidebar"> <h3>Legend</h3> @@ -45,45 +63,44 @@ </div> --> <p> - To understand how CelebA Dataset has been used around the world... - affected global research on computer vision, surveillance, defense, and consumer technology, the and where this dataset has been used the locations of each organization that used or referenced the datast + To help understand how CelebA Dataset has been used around the world for commercial, military and academic research; publicly available research citing Large-scale CelebFaces Attributes Dataset is collected, verified, and geocoded to show the biometric trade routes of people appearing in the images. Click on the markers to reveal reserach projects at that location. </p> </section> -<section class="applet_container"> +<section class="applet_container fullwidth"> <div class="applet" data-payload="{"command": "map"}"></div> + </section> <div class="caption"> - <div class="map-legend-item edu">Academic</div> - <div class="map-legend-item com">Industry</div> - <div class="map-legend-item gov">Government</div> - Data is compiled from <a href="https://www.semanticscholar.org">Semantic Scholar</a> and has been manually verified to show usage of CelebA Dataset. + <ul class="map-legend"> + <li class="edu">Academic</li> + <li class="com">Commercial</li> + <li class="gov">Military / Government</li> + <li class="source">Citation data is collected using <a href="https://semanticscholar.org" target="_blank">SemanticScholar.org</a> then dataset usage verified and geolocated.</li> + </ul> </div> -<section> +<!-- <section> <p class='subp'> - Standardized paragraph of text about the map. Sed ut perspiciatis, unde omnis iste natus error sit voluptatem accusantium doloremque laudantium, totam rem aperiam eaque ipsa, quae ab illo inventore veritatis et quasi architecto beatae vitae dicta sunt, explicabo. + [section under development] CelebA Dataset ... Standardized paragraph of text about the map. Sed ut perspiciatis, unde omnis iste natus error sit voluptatem accusantium doloremque laudantium, totam rem aperiam eaque ipsa, quae ab illo inventore veritatis et quasi architecto beatae vitae dicta sunt, explicabo. </p> -</section><section> - +</section> + --><section> <div class="hr-wave-holder"> <div class="hr-wave-line hr-wave-line1"></div> <div class="hr-wave-line hr-wave-line2"></div> </div> - <h2>Supplementary Information</h2> + <h3>Supplementary Information</h3> + </section><section class="applet_container"> - <h3>Citations</h3> - <p> - Citations were collected from <a href="https://www.semanticscholar.org">Semantic Scholar</a>, a website which aggregates - and indexes research papers. The citations were geocoded using names of institutions found in the PDF front matter, or as listed on other resources. These papers have been manually verified to show that researchers downloaded and used the dataset to train and/or test machine learning algorithms. - </p> + <h3>Dataset Citations</h3> <p> - Add button/link to download CSV + The dataset citations used in the visualizations were collected from <a href="https://www.semanticscholar.org">Semantic Scholar</a>, a website which aggregates and indexes research papers. Each citation was geocoded using names of institutions found in the PDF front matter, or as listed on other resources. These papers have been manually verified to show that researchers downloaded and used the dataset to train or test machine learning algorithms. </p> <div class="applet" data-payload="{"command": "citations"}"></div> diff --git a/site/public/datasets/cofw/index.html b/site/public/datasets/cofw/index.html index eac1f7a6..8925d4b8 100644 --- a/site/public/datasets/cofw/index.html +++ b/site/public/datasets/cofw/index.html @@ -8,7 +8,6 @@ <meta name="referrer" content="no-referrer" /> <meta name="viewport" content="width=device-width, initial-scale=1.0, user-scalable=yes" /> <link rel='stylesheet' href='/assets/css/fonts.css' /> - <link rel='stylesheet' href='/assets/css/tabulator.css' /> <link rel='stylesheet' href='/assets/css/css.css' /> <link rel='stylesheet' href='/assets/css/leaflet.css' /> <link rel='stylesheet' href='/assets/css/applets.css' /> @@ -18,6 +17,7 @@ <a class='slogan' href="/"> <div class='logo'></div> <div class='site_name'>MegaPixels</div> + <div class='splash'>COFW</div> </a> <div class='links'> <a href="/datasets/">Datasets</a> @@ -26,7 +26,19 @@ </header> <div class="content content-"> - <section><div class='left-sidebar'><div class='meta'><div><div class='gray'>Years</div><div>1993-1996</div></div><div><div class='gray'>Images</div><div>14,126</div></div><div><div class='gray'>Identities</div><div>1,199 </div></div><div><div class='gray'>Origin</div><div>Web Searches</div></div><div><div class='gray'>Funded by</div><div>ODNI, IARPA, Microsoft</div></div></div></div><h2>Caltech Occluded Faces in the Wild</h2> + <section><div class='left-sidebar'><div class='meta'> + <div class='gray'>Published</div> + <div>2013</div> + </div><div class='meta'> + <div class='gray'>Images</div> + <div>1,007 </div> + </div><div class='meta'> + <div class='gray'>Purpose</div> + <div>challenging dataset (sunglasses, hats, interaction with objects)</div> + </div><div class='meta'> + <div class='gray'>Website</div> + <div><a href='http://www.vision.caltech.edu/xpburgos/ICCV13/' target='_blank' rel='nofollow noopener'>caltech.edu</a></div> + </div><div class='meta'><div><div class='gray'>Years</div><div>1993-1996</div></div><div><div class='gray'>Images</div><div>14,126</div></div><div><div class='gray'>Identities</div><div>1,199 </div></div><div><div class='gray'>Origin</div><div>Web Searches</div></div><div><div class='gray'>Funded by</div><div>ODNI, IARPA, Microsoft</div></div></div></div><h2>Caltech Occluded Faces in the Wild</h2> <p>(PAGE UNDER DEVELOPMENT)</p> <p>COFW is "is designed to benchmark face landmark algorithms in realistic conditions, which include heavy occlusions and large shape variations" [Robust face landmark estimation under occlusion].</p> <p>RESEARCH below this line</p> @@ -43,7 +55,7 @@ To increase the number of training images, and since COFW has the exact same la <p><a href="https://www.cs.cmu.edu/~peiyunh/topdown/">https://www.cs.cmu.edu/~peiyunh/topdown/</a></p> </section><section> - <h3>Information Supply Chain</h3> + <h3>Biometric Trade Routes</h3> <!-- <div class="map-sidebar right-sidebar"> <h3>Legend</h3> @@ -55,45 +67,44 @@ To increase the number of training images, and since COFW has the exact same la </div> --> <p> - To understand how COFW Dataset has been used around the world... - affected global research on computer vision, surveillance, defense, and consumer technology, the and where this dataset has been used the locations of each organization that used or referenced the datast + To help understand how COFW Dataset has been used around the world for commercial, military and academic research; publicly available research citing Caltech Occluded Faces in the Wild is collected, verified, and geocoded to show the biometric trade routes of people appearing in the images. Click on the markers to reveal reserach projects at that location. </p> </section> -<section class="applet_container"> +<section class="applet_container fullwidth"> <div class="applet" data-payload="{"command": "map"}"></div> + </section> <div class="caption"> - <div class="map-legend-item edu">Academic</div> - <div class="map-legend-item com">Industry</div> - <div class="map-legend-item gov">Government</div> - Data is compiled from <a href="https://www.semanticscholar.org">Semantic Scholar</a> and has been manually verified to show usage of COFW Dataset. + <ul class="map-legend"> + <li class="edu">Academic</li> + <li class="com">Commercial</li> + <li class="gov">Military / Government</li> + <li class="source">Citation data is collected using <a href="https://semanticscholar.org" target="_blank">SemanticScholar.org</a> then dataset usage verified and geolocated.</li> + </ul> </div> -<section> +<!-- <section> <p class='subp'> - Standardized paragraph of text about the map. Sed ut perspiciatis, unde omnis iste natus error sit voluptatem accusantium doloremque laudantium, totam rem aperiam eaque ipsa, quae ab illo inventore veritatis et quasi architecto beatae vitae dicta sunt, explicabo. + [section under development] COFW Dataset ... Standardized paragraph of text about the map. Sed ut perspiciatis, unde omnis iste natus error sit voluptatem accusantium doloremque laudantium, totam rem aperiam eaque ipsa, quae ab illo inventore veritatis et quasi architecto beatae vitae dicta sunt, explicabo. </p> -</section><section> - +</section> + --><section> <div class="hr-wave-holder"> <div class="hr-wave-line hr-wave-line1"></div> <div class="hr-wave-line hr-wave-line2"></div> </div> - <h2>Supplementary Information</h2> + <h3>Supplementary Information</h3> + </section><section class="applet_container"> - <h3>Citations</h3> - <p> - Citations were collected from <a href="https://www.semanticscholar.org">Semantic Scholar</a>, a website which aggregates - and indexes research papers. The citations were geocoded using names of institutions found in the PDF front matter, or as listed on other resources. These papers have been manually verified to show that researchers downloaded and used the dataset to train and/or test machine learning algorithms. - </p> + <h3>Dataset Citations</h3> <p> - Add button/link to download CSV + The dataset citations used in the visualizations were collected from <a href="https://www.semanticscholar.org">Semantic Scholar</a>, a website which aggregates and indexes research papers. Each citation was geocoded using names of institutions found in the PDF front matter, or as listed on other resources. These papers have been manually verified to show that researchers downloaded and used the dataset to train or test machine learning algorithms. </p> <div class="applet" data-payload="{"command": "citations"}"></div> @@ -101,13 +112,14 @@ To increase the number of training images, and since COFW has the exact same la <h3>Who used COFW Dataset?</h3> <p> - This bar chart presents a ranking of the top countries where citations originated. Mouse over individual columns - to see yearly totals. These charts show at most the top 10 countries. + This bar chart presents a ranking of the top countries where dataset citations originated. Mouse over individual columns to see yearly totals. These charts show at most the top 10 countries. </p> </section> <section class="applet_container"> +<!-- <div style="position: absolute;top: 0px;right: -55px;width: 180px;font-size: 14px;">Labeled Faces in the Wild Dataset<br><span class="numc" style="font-size: 11px;">20 citations</span> +</div> --> <div class="applet" data-payload="{"command": "chart"}"></div> </section><section><p>TODO</p> <h2>- replace graphic</h2> diff --git a/site/public/datasets/duke_mtmc/index.html b/site/public/datasets/duke_mtmc/index.html index 299331d7..37de48ad 100644 --- a/site/public/datasets/duke_mtmc/index.html +++ b/site/public/datasets/duke_mtmc/index.html @@ -4,11 +4,10 @@ <title>MegaPixels</title> <meta charset="utf-8" /> <meta name="author" content="Adam Harvey" /> - <meta name="description" content="Duke MTMC is a dataset of CCTV footage of students at Duke University" /> + <meta name="description" content="Duke MTMC is a dataset of surveillance camera footage of students on Duke University campus" /> <meta name="referrer" content="no-referrer" /> <meta name="viewport" content="width=device-width, initial-scale=1.0, user-scalable=yes" /> <link rel='stylesheet' href='/assets/css/fonts.css' /> - <link rel='stylesheet' href='/assets/css/tabulator.css' /> <link rel='stylesheet' href='/assets/css/css.css' /> <link rel='stylesheet' href='/assets/css/leaflet.css' /> <link rel='stylesheet' href='/assets/css/applets.css' /> @@ -18,6 +17,7 @@ <a class='slogan' href="/"> <div class='logo'></div> <div class='site_name'>MegaPixels</div> + <div class='splash'>Duke MTMC</div> </a> <div class='links'> <a href="/datasets/">Datasets</a> @@ -26,12 +26,35 @@ </header> <div class="content content-dataset"> - <section class='intro_section' style='background-image: url(https://nyc3.digitaloceanspaces.com/megapixels/v1/datasets/duke_mtmc/assets/background.jpg)'><div class='inner'><div class='hero_desc'><span class='bgpad'><span class="dataset-name">Duke MTMC</span> is a dataset of CCTV footage of students at Duke University</span></div><div class='hero_subdesc'><span class='bgpad'>Duke MTMC contains over 2 million video frames and 2,000 unique identities collected from 8 cameras at Duke University campus in March 2014 -</span></div></div></section><section><div class='left-sidebar'><div class='meta'><div><div class='gray'>Collected</div><div>March 19, 2014</div></div><div><div class='gray'>Cameras</div><div>8</div></div><div><div class='gray'>Video Frames</div><div>2,000,000</div></div><div><div class='gray'>Identities</div><div>Over 2,000</div></div><div><div class='gray'>Used for</div><div>Person re-identification, <br>face recognition</div></div><div><div class='gray'>Sector</div><div>Academic</div></div><div><div class='gray'>Website</div><div><a href="http://vision.cs.duke.edu/DukeMTMC/">duke.edu</a></div></div></div></div><h2>Duke Multi-Target, Multi-Camera Tracking Dataset (Duke MTMC)</h2> -<p>(PAGE UNDER DEVELOPMENT)</p> + <section class='intro_section' style='background-image: url(https://nyc3.digitaloceanspaces.com/megapixels/v1/datasets/duke_mtmc/assets/background.jpg)'><div class='inner'><div class='hero_desc'><span class='bgpad'><span class="dataset-name">Duke MTMC</span> is a dataset of surveillance camera footage of students on Duke University campus</span></div><div class='hero_subdesc'><span class='bgpad'>Duke MTMC contains over 2 million video frames and 2,000 unique identities collected from 8 HD cameras at Duke University campus in March 2014 +</span></div></div></section><section><div class='left-sidebar'><div class='meta'> + <div class='gray'>Published</div> + <div>2016</div> + </div><div class='meta'> + <div class='gray'>Images</div> + <div>2,000,000 </div> + </div><div class='meta'> + <div class='gray'>Identities</div> + <div>1,812 </div> + </div><div class='meta'> + <div class='gray'>Purpose</div> + <div>Person re-identification and multi-camera tracking</div> + </div><div class='meta'> + <div class='gray'>Created by</div> + <div>Computer Science Department, Duke University, Durham, US</div> + </div><div class='meta'> + <div class='gray'>Website</div> + <div><a href='http://vision.cs.duke.edu/DukeMTMC/' target='_blank' rel='nofollow noopener'>duke.edu</a></div> + </div><div class='meta'><div><div class='gray'>Created</div><div>2014</div></div><div><div class='gray'>Identities</div><div>Over 2,700</div></div><div><div class='gray'>Used for</div><div>Face recognition, person re-identification</div></div><div><div class='gray'>Created by</div><div>Computer Science Department, Duke University, Durham, US</div></div><div><div class='gray'>Website</div><div><a href="http://vision.cs.duke.edu/DukeMTMC/">duke.edu</a></div></div></div></div><h2>Duke Multi-Target, Multi-Camera Tracking Dataset (Duke MTMC)</h2> +<p>[ PAGE UNDER DEVELOPMENT ]</p> +<p>Duke MTMC is a dataset of video recorded on Duke University campus during for the purpose of training, evaluating, and improving <em>multi-target multi-camera tracking</em>. The videos were recorded during February and March 2014 and cinclude</p> +<p>Includes a total of 888.8 minutes of video (ind. verified)</p> +<p>"We make available a new data set that has more than 2 million frames and more than 2,700 identities. It consists of 8×85 minutes of 1080p video recorded at 60 frames per second from 8 static cameras deployed on the Duke University campus during periods between lectures, when pedestrian traffic is heavy."</p> +<p>The dataset includes approximately 2,000 annotated identities appearing in 85 hours of video from 8 cameras located throughout Duke University's campus.</p> +</section><section class='images'><div class='image'><img src='https://nyc3.digitaloceanspaces.com/megapixels/v1/datasets/duke_mtmc/assets/duke_mtmc_cam5_average_comp.jpg' alt=' Duke MTMC pixel-averaged image of camera #5 is shown with the bounding boxes for each student drawn in white. (c) Adam Harvey'><div class='caption'> Duke MTMC pixel-averaged image of camera #5 is shown with the bounding boxes for each student drawn in white. (c) Adam Harvey</div></div></section><section><p>According to the dataset authors,</p> </section><section> - <h3>Information Supply Chain</h3> + <h3>Biometric Trade Routes</h3> <!-- <div class="map-sidebar right-sidebar"> <h3>Legend</h3> @@ -43,71 +66,66 @@ </div> --> <p> - To understand how Duke MTMC Dataset has been used around the world... - affected global research on computer vision, surveillance, defense, and consumer technology, the and where this dataset has been used the locations of each organization that used or referenced the datast + To help understand how Duke MTMC Dataset has been used around the world for commercial, military and academic research; publicly available research citing Duke Multi-Target, Multi-Camera Tracking Project is collected, verified, and geocoded to show the biometric trade routes of people appearing in the images. Click on the markers to reveal reserach projects at that location. </p> </section> -<section class="applet_container"> +<section class="applet_container fullwidth"> <div class="applet" data-payload="{"command": "map"}"></div> + </section> <div class="caption"> - <div class="map-legend-item edu">Academic</div> - <div class="map-legend-item com">Industry</div> - <div class="map-legend-item gov">Government</div> - Data is compiled from <a href="https://www.semanticscholar.org">Semantic Scholar</a> and has been manually verified to show usage of Duke MTMC Dataset. + <ul class="map-legend"> + <li class="edu">Academic</li> + <li class="com">Commercial</li> + <li class="gov">Military / Government</li> + <li class="source">Citation data is collected using <a href="https://semanticscholar.org" target="_blank">SemanticScholar.org</a> then dataset usage verified and geolocated.</li> + </ul> </div> -<section> +<!-- <section> <p class='subp'> - Standardized paragraph of text about the map. Sed ut perspiciatis, unde omnis iste natus error sit voluptatem accusantium doloremque laudantium, totam rem aperiam eaque ipsa, quae ab illo inventore veritatis et quasi architecto beatae vitae dicta sunt, explicabo. + [section under development] Duke MTMC Dataset ... Standardized paragraph of text about the map. Sed ut perspiciatis, unde omnis iste natus error sit voluptatem accusantium doloremque laudantium, totam rem aperiam eaque ipsa, quae ab illo inventore veritatis et quasi architecto beatae vitae dicta sunt, explicabo. </p> -</section><section> +</section> + --><section> <h3>Who used Duke MTMC Dataset?</h3> <p> - This bar chart presents a ranking of the top countries where citations originated. Mouse over individual columns - to see yearly totals. These charts show at most the top 10 countries. + This bar chart presents a ranking of the top countries where dataset citations originated. Mouse over individual columns to see yearly totals. These charts show at most the top 10 countries. </p> </section> <section class="applet_container"> +<!-- <div style="position: absolute;top: 0px;right: -55px;width: 180px;font-size: 14px;">Labeled Faces in the Wild Dataset<br><span class="numc" style="font-size: 11px;">20 citations</span> +</div> --> <div class="applet" data-payload="{"command": "chart"}"></div> -</section><section> - <p> - These pie charts show overall totals based on country and institution type. - </p> - - </section> - -<section class="applet_container"> +</section><section class="applet_container"> <div class="applet" data-payload="{"command": "piechart"}"></div> </section><section> - <div class="hr-wave-holder"> <div class="hr-wave-line hr-wave-line1"></div> <div class="hr-wave-line hr-wave-line2"></div> </div> - <h2>Supplementary Information</h2> + <h3>Supplementary Information</h3> + </section><section class="applet_container"> - <h3>Citations</h3> - <p> - Citations were collected from <a href="https://www.semanticscholar.org">Semantic Scholar</a>, a website which aggregates - and indexes research papers. The citations were geocoded using names of institutions found in the PDF front matter, or as listed on other resources. These papers have been manually verified to show that researchers downloaded and used the dataset to train and/or test machine learning algorithms. - </p> + <h3>Dataset Citations</h3> <p> - Add button/link to download CSV + The dataset citations used in the visualizations were collected from <a href="https://www.semanticscholar.org">Semantic Scholar</a>, a website which aggregates and indexes research papers. Each citation was geocoded using names of institutions found in the PDF front matter, or as listed on other resources. These papers have been manually verified to show that researchers downloaded and used the dataset to train or test machine learning algorithms. </p> <div class="applet" data-payload="{"command": "citations"}"></div> </section><section><h2>Research Notes</h2> <ul> +<li>"We make available a new data set that has more than 2 million frames and more than 2,700 identities. It consists of 8×85 minutes of 1080p video recorded at 60 frames per second from 8 static cameras deployed on the Duke University campus during periods between lectures, when pedestrian traffic is heavy." - 27a2fad58dd8727e280f97036e0d2bc55ef5424c</li> +<li>"This work was supported in part by the EPSRC Programme Grant (FACER2VM) EP/N007743/1, EPSRC/dstl/MURI project EP/R018456/1, the National Natural Science Foundation of China (61373055, 61672265, 61602390, 61532009, 61571313), Chinese Ministry of Education (Z2015101), Science and Technology Department of Sichuan Province (2017RZ0009 and 2017FZ0029), Education Department of Sichuan Province (15ZB0130), the Open Research Fund from Province Key Laboratory of Xihua University (szjj2015-056) and the NVIDIA GPU Grant Program." - ec9c20ed6cce15e9b63ac96bb5a6d55e69661e0b</li> <li>"DukeMTMC aims to accelerate advances in multi-target multi-camera tracking. It provides a tracking system that works within and across cameras, a new large scale HD video data set recorded by 8 synchronized cameras with more than 7,000 single camera trajectories and over 2,000 unique identities, and a new performance evaluation method that measures how often a system is correct about who is where"</li> <li><p>DukeMTMC is a new, manually annotated, calibrated, multi-camera data set recorded outdoors on the Duke University campus with 8 synchronized cameras. It consists of:</p> <p>8 static cameras x 85 minutes of 1080p 60 fps video diff --git a/site/public/datasets/facebook/index.html b/site/public/datasets/facebook/index.html index 7fb1901a..b2943e1f 100644 --- a/site/public/datasets/facebook/index.html +++ b/site/public/datasets/facebook/index.html @@ -8,7 +8,6 @@ <meta name="referrer" content="no-referrer" /> <meta name="viewport" content="width=device-width, initial-scale=1.0, user-scalable=yes" /> <link rel='stylesheet' href='/assets/css/fonts.css' /> - <link rel='stylesheet' href='/assets/css/tabulator.css' /> <link rel='stylesheet' href='/assets/css/css.css' /> <link rel='stylesheet' href='/assets/css/leaflet.css' /> <link rel='stylesheet' href='/assets/css/applets.css' /> @@ -18,6 +17,7 @@ <a class='slogan' href="/"> <div class='logo'></div> <div class='site_name'>MegaPixels</div> + </a> <div class='links'> <a href="/datasets/">Datasets</a> diff --git a/site/public/datasets/feret/index.html b/site/public/datasets/feret/index.html index ce60f3de..45510f64 100644 --- a/site/public/datasets/feret/index.html +++ b/site/public/datasets/feret/index.html @@ -8,7 +8,6 @@ <meta name="referrer" content="no-referrer" /> <meta name="viewport" content="width=device-width, initial-scale=1.0, user-scalable=yes" /> <link rel='stylesheet' href='/assets/css/fonts.css' /> - <link rel='stylesheet' href='/assets/css/tabulator.css' /> <link rel='stylesheet' href='/assets/css/css.css' /> <link rel='stylesheet' href='/assets/css/leaflet.css' /> <link rel='stylesheet' href='/assets/css/applets.css' /> @@ -18,6 +17,7 @@ <a class='slogan' href="/"> <div class='logo'></div> <div class='site_name'>MegaPixels</div> + <div class='splash'>LFW</div> </a> <div class='links'> <a href="/datasets/">Datasets</a> diff --git a/site/public/datasets/hrt_transgender/index.html b/site/public/datasets/hrt_transgender/index.html index e38e134b..80a4f40b 100644 --- a/site/public/datasets/hrt_transgender/index.html +++ b/site/public/datasets/hrt_transgender/index.html @@ -8,7 +8,6 @@ <meta name="referrer" content="no-referrer" /> <meta name="viewport" content="width=device-width, initial-scale=1.0, user-scalable=yes" /> <link rel='stylesheet' href='/assets/css/fonts.css' /> - <link rel='stylesheet' href='/assets/css/tabulator.css' /> <link rel='stylesheet' href='/assets/css/css.css' /> <link rel='stylesheet' href='/assets/css/leaflet.css' /> <link rel='stylesheet' href='/assets/css/applets.css' /> @@ -18,6 +17,7 @@ <a class='slogan' href="/"> <div class='logo'></div> <div class='site_name'>MegaPixels</div> + <div class='splash'>HRT Transgender</div> </a> <div class='links'> <a href="/datasets/">Datasets</a> @@ -27,31 +27,40 @@ <div class="content content-dataset"> <section class='intro_section' style='background-image: url(https://nyc3.digitaloceanspaces.com/megapixels/v1/datasets/hrt_transgender/assets/background.jpg)'><div class='inner'><div class='hero_desc'><span class='bgpad'>TBD</span></div><div class='hero_subdesc'><span class='bgpad'>TBD -</span></div></div></section><section><div class='left-sidebar'><div class='meta'><div><div class='gray'>Published</div><div>TBD</div></div><div><div class='gray'>Images</div><div>TBD</div></div></div></div><h2>HRT Transgender Dataset</h2> +</span></div></div></section><section><div class='left-sidebar'><div class='meta'> + <div class='gray'>Published</div> + <div>2013</div> + </div><div class='meta'> + <div class='gray'>Images</div> + <div>10,564 </div> + </div><div class='meta'> + <div class='gray'>Identities</div> + <div>38 </div> + </div><div class='meta'> + <div class='gray'>Purpose</div> + <div>gender transition and facial recognition</div> + </div><div class='meta'> + <div class='gray'>Website</div> + <div><a href='http://www.faceaginggroup.com/hrt-transgender/' target='_blank' rel='nofollow noopener'>faceaginggroup.com</a></div> + </div><div class='meta'><div><div class='gray'>Published</div><div>TBD</div></div><div><div class='gray'>Images</div><div>TBD</div></div></div></div><h2>HRT Transgender Dataset</h2> </section><section> <h3>Who used HRT Transgender?</h3> <p> - This bar chart presents a ranking of the top countries where citations originated. Mouse over individual columns - to see yearly totals. These charts show at most the top 10 countries. + This bar chart presents a ranking of the top countries where dataset citations originated. Mouse over individual columns to see yearly totals. These charts show at most the top 10 countries. </p> </section> <section class="applet_container"> +<!-- <div style="position: absolute;top: 0px;right: -55px;width: 180px;font-size: 14px;">Labeled Faces in the Wild Dataset<br><span class="numc" style="font-size: 11px;">20 citations</span> +</div> --> <div class="applet" data-payload="{"command": "chart"}"></div> -</section><section> - <p> - These pie charts show overall totals based on country and institution type. - </p> - - </section> - -<section class="applet_container"> +</section><section class="applet_container"> <div class="applet" data-payload="{"command": "piechart"}"></div> </section><section> - <h3>Information Supply Chain</h3> + <h3>Biometric Trade Routes</h3> <!-- <div class="map-sidebar right-sidebar"> <h3>Legend</h3> @@ -63,45 +72,44 @@ </div> --> <p> - To understand how HRT Transgender has been used around the world... - affected global research on computer vision, surveillance, defense, and consumer technology, the and where this dataset has been used the locations of each organization that used or referenced the datast + To help understand how HRT Transgender has been used around the world for commercial, military and academic research; publicly available research citing HRT Transgender Dataset is collected, verified, and geocoded to show the biometric trade routes of people appearing in the images. Click on the markers to reveal reserach projects at that location. </p> </section> -<section class="applet_container"> +<section class="applet_container fullwidth"> <div class="applet" data-payload="{"command": "map"}"></div> + </section> <div class="caption"> - <div class="map-legend-item edu">Academic</div> - <div class="map-legend-item com">Industry</div> - <div class="map-legend-item gov">Government</div> - Data is compiled from <a href="https://www.semanticscholar.org">Semantic Scholar</a> and has been manually verified to show usage of HRT Transgender. + <ul class="map-legend"> + <li class="edu">Academic</li> + <li class="com">Commercial</li> + <li class="gov">Military / Government</li> + <li class="source">Citation data is collected using <a href="https://semanticscholar.org" target="_blank">SemanticScholar.org</a> then dataset usage verified and geolocated.</li> + </ul> </div> -<section> +<!-- <section> <p class='subp'> - Standardized paragraph of text about the map. Sed ut perspiciatis, unde omnis iste natus error sit voluptatem accusantium doloremque laudantium, totam rem aperiam eaque ipsa, quae ab illo inventore veritatis et quasi architecto beatae vitae dicta sunt, explicabo. + [section under development] HRT Transgender ... Standardized paragraph of text about the map. Sed ut perspiciatis, unde omnis iste natus error sit voluptatem accusantium doloremque laudantium, totam rem aperiam eaque ipsa, quae ab illo inventore veritatis et quasi architecto beatae vitae dicta sunt, explicabo. </p> -</section><section> - +</section> + --><section> <div class="hr-wave-holder"> <div class="hr-wave-line hr-wave-line1"></div> <div class="hr-wave-line hr-wave-line2"></div> </div> - <h2>Supplementary Information</h2> + <h3>Supplementary Information</h3> + </section><section class="applet_container"> - <h3>Citations</h3> - <p> - Citations were collected from <a href="https://www.semanticscholar.org">Semantic Scholar</a>, a website which aggregates - and indexes research papers. The citations were geocoded using names of institutions found in the PDF front matter, or as listed on other resources. These papers have been manually verified to show that researchers downloaded and used the dataset to train and/or test machine learning algorithms. - </p> + <h3>Dataset Citations</h3> <p> - Add button/link to download CSV + The dataset citations used in the visualizations were collected from <a href="https://www.semanticscholar.org">Semantic Scholar</a>, a website which aggregates and indexes research papers. Each citation was geocoded using names of institutions found in the PDF front matter, or as listed on other resources. These papers have been manually verified to show that researchers downloaded and used the dataset to train or test machine learning algorithms. </p> <div class="applet" data-payload="{"command": "citations"}"></div> diff --git a/site/public/datasets/index.html b/site/public/datasets/index.html index f618e86b..f4776f6a 100644 --- a/site/public/datasets/index.html +++ b/site/public/datasets/index.html @@ -8,7 +8,6 @@ <meta name="referrer" content="no-referrer" /> <meta name="viewport" content="width=device-width, initial-scale=1.0, user-scalable=yes" /> <link rel='stylesheet' href='/assets/css/fonts.css' /> - <link rel='stylesheet' href='/assets/css/tabulator.css' /> <link rel='stylesheet' href='/assets/css/css.css' /> <link rel='stylesheet' href='/assets/css/leaflet.css' /> <link rel='stylesheet' href='/assets/css/applets.css' /> @@ -18,6 +17,7 @@ <a class='slogan' href="/"> <div class='logo'></div> <div class='site_name'>MegaPixels</div> + </a> <div class='links'> <a href="/datasets/">Datasets</a> @@ -28,7 +28,7 @@ <section><h1>Facial Recognition Datasets</h1> -<h3>Survey</h3> +<p>Explore publicly available facial recognition datasets. More datasets will be added throughout 2019.</p> </section> <section class='applet_container autosize'><div class='applet' data-payload='{"command":"dataset_list"}'></div></section> @@ -42,8 +42,8 @@ <span class='title'>Brainwash</span> <div class='fields'> <div class='year visible'><span>2015</span></div> - <div class='purpose'><span>Decoding image into set of people detections.</span></div> - <div class='images'><span>11,918 images</span></div> + <div class='purpose'><span>Head detection</span></div> + <div class='images'><span>11,917 images</span></div> <div class='identities'><span></span></div> </div> </div> @@ -85,6 +85,18 @@ </div> </a> + <a href="/datasets/msceleb/" style="background-image: url(https://nyc3.digitaloceanspaces.com/megapixels/v1/datasets/msceleb/assets/index.jpg)"> + <div class="dataset"> + <span class='title'>MS Celeb</span> + <div class='fields'> + <div class='year visible'><span>2016</span></div> + <div class='purpose'><span>Large-scale face recognition</span></div> + <div class='images'><span>1,000,000 images</span></div> + <div class='identities'><span>100,000 </span></div> + </div> + </div> + </a> + <a href="/datasets/pipa/" style="background-image: url(https://nyc3.digitaloceanspaces.com/megapixels/v1/datasets/pipa/assets/index.jpg)"> <div class="dataset"> <span class='title'>People in Photo Albums</span> @@ -102,9 +114,9 @@ <span class='title'>Unconstrained College Students</span> <div class='fields'> <div class='year visible'><span>2018</span></div> - <div class='purpose'><span>Unconstrained face recognition</span></div> + <div class='purpose'><span>Face recognition, face detection</span></div> <div class='images'><span>16,149 images</span></div> - <div class='identities'><span>4,362 </span></div> + <div class='identities'><span>1,732 </span></div> </div> </div> </a> @@ -114,7 +126,7 @@ <span class='title'>VIPeR</span> <div class='fields'> <div class='year visible'><span>2007</span></div> - <div class='purpose'><span>pedestrian re-identification</span></div> + <div class='purpose'><span>Person re-identification</span></div> <div class='images'><span>1,264 images</span></div> <div class='identities'><span>632 </span></div> </div> diff --git a/site/public/datasets/lfpw/index.html b/site/public/datasets/lfpw/index.html index 087d8b1d..77189ce7 100644 --- a/site/public/datasets/lfpw/index.html +++ b/site/public/datasets/lfpw/index.html @@ -8,7 +8,6 @@ <meta name="referrer" content="no-referrer" /> <meta name="viewport" content="width=device-width, initial-scale=1.0, user-scalable=yes" /> <link rel='stylesheet' href='/assets/css/fonts.css' /> - <link rel='stylesheet' href='/assets/css/tabulator.css' /> <link rel='stylesheet' href='/assets/css/css.css' /> <link rel='stylesheet' href='/assets/css/leaflet.css' /> <link rel='stylesheet' href='/assets/css/applets.css' /> @@ -18,6 +17,7 @@ <a class='slogan' href="/"> <div class='logo'></div> <div class='site_name'>MegaPixels</div> + <div class='splash'>LFWP</div> </a> <div class='links'> <a href="/datasets/">Datasets</a> diff --git a/site/public/datasets/lfw/index.html b/site/public/datasets/lfw/index.html index b4923877..d451d0cd 100644 --- a/site/public/datasets/lfw/index.html +++ b/site/public/datasets/lfw/index.html @@ -8,7 +8,6 @@ <meta name="referrer" content="no-referrer" /> <meta name="viewport" content="width=device-width, initial-scale=1.0, user-scalable=yes" /> <link rel='stylesheet' href='/assets/css/fonts.css' /> - <link rel='stylesheet' href='/assets/css/tabulator.css' /> <link rel='stylesheet' href='/assets/css/css.css' /> <link rel='stylesheet' href='/assets/css/leaflet.css' /> <link rel='stylesheet' href='/assets/css/applets.css' /> @@ -18,6 +17,7 @@ <a class='slogan' href="/"> <div class='logo'></div> <div class='site_name'>MegaPixels</div> + <div class='splash'>LFW</div> </a> <div class='links'> <a href="/datasets/">Datasets</a> @@ -27,7 +27,22 @@ <div class="content content-"> <section class='intro_section' style='background-image: url(https://nyc3.digitaloceanspaces.com/megapixels/v1/datasets/lfw/assets/background.jpg)'><div class='inner'><div class='hero_desc'><span class='bgpad'><span class="dataset-name">Labeled Faces in The Wild (LFW)</span> is the first facial recognition dataset created entirely from online photos</span></div><div class='hero_subdesc'><span class='bgpad'>It includes 13,456 images of 4,432 people's images copied from the Internet during 2002-2004 and is the most frequently used dataset in the world for benchmarking face recognition algorithms. -</span></div></div></section><section><div class='left-sidebar'><div class='meta'><div><div class='gray'>Created</div><div>2002 – 2004</div></div><div><div class='gray'>Images</div><div>13,233</div></div><div><div class='gray'>Identities</div><div>5,749</div></div><div><div class='gray'>Origin</div><div>Yahoo! News Images</div></div><div><div class='gray'>Used by</div><div>Facebook, Google, Microsoft, Baidu, Tencent, SenseTime, Face++, CIA, NSA, IARPA</div></div><div><div class='gray'>Website</div><div><a href="http://vis-www.cs.umass.edu/lfw">umass.edu</a></div></div></div><ul> +</span></div></div></section><section><div class='left-sidebar'><div class='meta'> + <div class='gray'>Published</div> + <div>2007</div> + </div><div class='meta'> + <div class='gray'>Images</div> + <div>13,233 </div> + </div><div class='meta'> + <div class='gray'>Identities</div> + <div>5,749 </div> + </div><div class='meta'> + <div class='gray'>Purpose</div> + <div>face recognition</div> + </div><div class='meta'> + <div class='gray'>Website</div> + <div><a href='http://vis-www.cs.umass.edu/lfw/' target='_blank' rel='nofollow noopener'>umass.edu</a></div> + </div><div class='meta'><div><div class='gray'>Created</div><div>2002 – 2004</div></div><div><div class='gray'>Images</div><div>13,233</div></div><div><div class='gray'>Identities</div><div>5,749</div></div><div><div class='gray'>Origin</div><div>Yahoo! News Images</div></div><div><div class='gray'>Used by</div><div>Facebook, Google, Microsoft, Baidu, Tencent, SenseTime, Face++, CIA, NSA, IARPA</div></div><div><div class='gray'>Website</div><div><a href="http://vis-www.cs.umass.edu/lfw">umass.edu</a></div></div></div><ul> <li>There are about 3 men for every 1 woman in the LFW dataset<a class="footnote_shim" name="[^lfw_www]_1"> </a><a href="#[^lfw_www]" class="footnote" title="Footnote 1">1</a></li> <li>The person with the most images is <a href="http://vis-www.cs.umass.edu/lfw/person/George_W_Bush_comp.html">George W. Bush</a> with 530</li> <li>There are about 3 George W. Bush's for every 1 <a href="http://vis-www.cs.umass.edu/lfw/person/Tony_Blair.html">Tony Blair</a></li> @@ -46,7 +61,7 @@ <p>The <em>Names and Faces</em> dataset was the first face recognition dataset created entire from online photos. However, <em>Names and Faces</em> and <em>LFW</em> are not the first face recognition dataset created entirely "in the wild". That title belongs to the <a href="/datasets/ucd_faces/">UCD dataset</a>. Images obtained "in the wild" means using an image without explicit consent or awareness from the subject or photographer.</p> </section><section> - <h3>Information Supply Chain</h3> + <h3>Biometric Trade Routes</h3> <!-- <div class="map-sidebar right-sidebar"> <h3>Legend</h3> @@ -58,66 +73,59 @@ </div> --> <p> - To understand how LFW has been used around the world... - affected global research on computer vision, surveillance, defense, and consumer technology, the and where this dataset has been used the locations of each organization that used or referenced the datast + To help understand how LFW has been used around the world for commercial, military and academic research; publicly available research citing Labeled Faces in the Wild is collected, verified, and geocoded to show the biometric trade routes of people appearing in the images. Click on the markers to reveal reserach projects at that location. </p> </section> -<section class="applet_container"> +<section class="applet_container fullwidth"> <div class="applet" data-payload="{"command": "map"}"></div> + </section> <div class="caption"> - <div class="map-legend-item edu">Academic</div> - <div class="map-legend-item com">Industry</div> - <div class="map-legend-item gov">Government</div> - Data is compiled from <a href="https://www.semanticscholar.org">Semantic Scholar</a> and has been manually verified to show usage of LFW. + <ul class="map-legend"> + <li class="edu">Academic</li> + <li class="com">Commercial</li> + <li class="gov">Military / Government</li> + <li class="source">Citation data is collected using <a href="https://semanticscholar.org" target="_blank">SemanticScholar.org</a> then dataset usage verified and geolocated.</li> + </ul> </div> -<section> +<!-- <section> <p class='subp'> - Standardized paragraph of text about the map. Sed ut perspiciatis, unde omnis iste natus error sit voluptatem accusantium doloremque laudantium, totam rem aperiam eaque ipsa, quae ab illo inventore veritatis et quasi architecto beatae vitae dicta sunt, explicabo. + [section under development] LFW ... Standardized paragraph of text about the map. Sed ut perspiciatis, unde omnis iste natus error sit voluptatem accusantium doloremque laudantium, totam rem aperiam eaque ipsa, quae ab illo inventore veritatis et quasi architecto beatae vitae dicta sunt, explicabo. </p> -</section><section> +</section> + --><section> <h3>Who used LFW?</h3> <p> - This bar chart presents a ranking of the top countries where citations originated. Mouse over individual columns - to see yearly totals. These charts show at most the top 10 countries. + This bar chart presents a ranking of the top countries where dataset citations originated. Mouse over individual columns to see yearly totals. These charts show at most the top 10 countries. </p> </section> <section class="applet_container"> +<!-- <div style="position: absolute;top: 0px;right: -55px;width: 180px;font-size: 14px;">Labeled Faces in the Wild Dataset<br><span class="numc" style="font-size: 11px;">20 citations</span> +</div> --> <div class="applet" data-payload="{"command": "chart"}"></div> -</section><section> - <p> - These pie charts show overall totals based on country and institution type. - </p> - - </section> - -<section class="applet_container"> +</section><section class="applet_container"> <div class="applet" data-payload="{"command": "piechart"}"></div> </section><section> - <div class="hr-wave-holder"> <div class="hr-wave-line hr-wave-line1"></div> <div class="hr-wave-line hr-wave-line2"></div> </div> - <h2>Supplementary Information</h2> + <h3>Supplementary Information</h3> + </section><section class="applet_container"> - <h3>Citations</h3> - <p> - Citations were collected from <a href="https://www.semanticscholar.org">Semantic Scholar</a>, a website which aggregates - and indexes research papers. The citations were geocoded using names of institutions found in the PDF front matter, or as listed on other resources. These papers have been manually verified to show that researchers downloaded and used the dataset to train and/or test machine learning algorithms. - </p> + <h3>Dataset Citations</h3> <p> - Add button/link to download CSV + The dataset citations used in the visualizations were collected from <a href="https://www.semanticscholar.org">Semantic Scholar</a>, a website which aggregates and indexes research papers. Each citation was geocoded using names of institutions found in the PDF front matter, or as listed on other resources. These papers have been manually verified to show that researchers downloaded and used the dataset to train or test machine learning algorithms. </p> <div class="applet" data-payload="{"command": "citations"}"></div> diff --git a/site/public/datasets/market_1501/index.html b/site/public/datasets/market_1501/index.html index a80c23fa..1ffd7e6c 100644 --- a/site/public/datasets/market_1501/index.html +++ b/site/public/datasets/market_1501/index.html @@ -8,7 +8,6 @@ <meta name="referrer" content="no-referrer" /> <meta name="viewport" content="width=device-width, initial-scale=1.0, user-scalable=yes" /> <link rel='stylesheet' href='/assets/css/fonts.css' /> - <link rel='stylesheet' href='/assets/css/tabulator.css' /> <link rel='stylesheet' href='/assets/css/css.css' /> <link rel='stylesheet' href='/assets/css/leaflet.css' /> <link rel='stylesheet' href='/assets/css/applets.css' /> @@ -18,6 +17,7 @@ <a class='slogan' href="/"> <div class='logo'></div> <div class='site_name'>MegaPixels</div> + <div class='splash'>Market 1501</div> </a> <div class='links'> <a href="/datasets/">Datasets</a> @@ -27,11 +27,26 @@ <div class="content content-dataset"> <section class='intro_section' style='background-image: url(https://nyc3.digitaloceanspaces.com/megapixels/v1/datasets/market_1501/assets/background.jpg)'><div class='inner'><div class='hero_desc'><span class='bgpad'><span class="dataset-name">Market-1501</span> is a dataset is collection of CCTV footage from ...</span></div><div class='hero_subdesc'><span class='bgpad'>The Market-1501 dataset includes ... -</span></div></div></section><section><div class='left-sidebar'><div class='meta'><div><div class='gray'>Collected</div><div>TBD</div></div><div><div class='gray'>Published</div><div>TBD</div></div><div><div class='gray'>Images</div><div>TBD</div></div><div><div class='gray'>Faces</div><div>TBD</div></div></div></div><h2>Market-1501 ...</h2> +</span></div></div></section><section><div class='left-sidebar'><div class='meta'> + <div class='gray'>Published</div> + <div>2015</div> + </div><div class='meta'> + <div class='gray'>Images</div> + <div>32,668 </div> + </div><div class='meta'> + <div class='gray'>Identities</div> + <div>1,501 </div> + </div><div class='meta'> + <div class='gray'>Purpose</div> + <div>Person re-identification</div> + </div><div class='meta'> + <div class='gray'>Website</div> + <div><a href='http://www.liangzheng.org/Project/project_reid.html' target='_blank' rel='nofollow noopener'>liangzheng.org</a></div> + </div><div class='meta'><div><div class='gray'>Collected</div><div>TBD</div></div><div><div class='gray'>Published</div><div>TBD</div></div><div><div class='gray'>Images</div><div>TBD</div></div><div><div class='gray'>Faces</div><div>TBD</div></div></div></div><h2>Market-1501 ...</h2> <p>(PAGE UNDER DEVELOPMENT)</p> </section><section> - <h3>Information Supply Chain</h3> + <h3>Biometric Trade Routes</h3> <!-- <div class="map-sidebar right-sidebar"> <h3>Legend</h3> @@ -43,45 +58,44 @@ </div> --> <p> - To understand how Market 1501 has been used around the world... - affected global research on computer vision, surveillance, defense, and consumer technology, the and where this dataset has been used the locations of each organization that used or referenced the datast + To help understand how Market 1501 has been used around the world for commercial, military and academic research; publicly available research citing Market 1501 Dataset is collected, verified, and geocoded to show the biometric trade routes of people appearing in the images. Click on the markers to reveal reserach projects at that location. </p> </section> -<section class="applet_container"> +<section class="applet_container fullwidth"> <div class="applet" data-payload="{"command": "map"}"></div> + </section> <div class="caption"> - <div class="map-legend-item edu">Academic</div> - <div class="map-legend-item com">Industry</div> - <div class="map-legend-item gov">Government</div> - Data is compiled from <a href="https://www.semanticscholar.org">Semantic Scholar</a> and has been manually verified to show usage of Market 1501. + <ul class="map-legend"> + <li class="edu">Academic</li> + <li class="com">Commercial</li> + <li class="gov">Military / Government</li> + <li class="source">Citation data is collected using <a href="https://semanticscholar.org" target="_blank">SemanticScholar.org</a> then dataset usage verified and geolocated.</li> + </ul> </div> -<section> +<!-- <section> <p class='subp'> - Standardized paragraph of text about the map. Sed ut perspiciatis, unde omnis iste natus error sit voluptatem accusantium doloremque laudantium, totam rem aperiam eaque ipsa, quae ab illo inventore veritatis et quasi architecto beatae vitae dicta sunt, explicabo. + [section under development] Market 1501 ... Standardized paragraph of text about the map. Sed ut perspiciatis, unde omnis iste natus error sit voluptatem accusantium doloremque laudantium, totam rem aperiam eaque ipsa, quae ab illo inventore veritatis et quasi architecto beatae vitae dicta sunt, explicabo. </p> -</section><section> - +</section> + --><section> <div class="hr-wave-holder"> <div class="hr-wave-line hr-wave-line1"></div> <div class="hr-wave-line hr-wave-line2"></div> </div> - <h2>Supplementary Information</h2> + <h3>Supplementary Information</h3> + </section><section class="applet_container"> - <h3>Citations</h3> - <p> - Citations were collected from <a href="https://www.semanticscholar.org">Semantic Scholar</a>, a website which aggregates - and indexes research papers. The citations were geocoded using names of institutions found in the PDF front matter, or as listed on other resources. These papers have been manually verified to show that researchers downloaded and used the dataset to train and/or test machine learning algorithms. - </p> + <h3>Dataset Citations</h3> <p> - Add button/link to download CSV + The dataset citations used in the visualizations were collected from <a href="https://www.semanticscholar.org">Semantic Scholar</a>, a website which aggregates and indexes research papers. Each citation was geocoded using names of institutions found in the PDF front matter, or as listed on other resources. These papers have been manually verified to show that researchers downloaded and used the dataset to train or test machine learning algorithms. </p> <div class="applet" data-payload="{"command": "citations"}"></div> diff --git a/site/public/datasets/mars/index.html b/site/public/datasets/mars/index.html deleted file mode 100644 index b053b456..00000000 --- a/site/public/datasets/mars/index.html +++ /dev/null @@ -1,110 +0,0 @@ -<!doctype html> -<html> -<head> - <title>MegaPixels</title> - <meta charset="utf-8" /> - <meta name="author" content="Adam Harvey" /> - <meta name="description" content="Motion Analysis and Re-identification Set (MARS) is a dataset is collection of CCTV footage " /> - <meta name="referrer" content="no-referrer" /> - <meta name="viewport" content="width=device-width, initial-scale=1.0, user-scalable=yes" /> - <link rel='stylesheet' href='/assets/css/fonts.css' /> - <link rel='stylesheet' href='/assets/css/tabulator.css' /> - <link rel='stylesheet' href='/assets/css/css.css' /> - <link rel='stylesheet' href='/assets/css/leaflet.css' /> - <link rel='stylesheet' href='/assets/css/applets.css' /> -</head> -<body> - <header> - <a class='slogan' href="/"> - <div class='logo'></div> - <div class='site_name'>MegaPixels</div> - </a> - <div class='links'> - <a href="/datasets/">Datasets</a> - <a href="/about/">About</a> - </div> - </header> - <div class="content content-dataset"> - - <section class='intro_section' style='background-image: url(https://nyc3.digitaloceanspaces.com/megapixels/v1/datasets/mars/assets/background.jpg)'><div class='inner'><div class='hero_desc'><span class='bgpad'><span style="color:#99ccee">Motion Analysis and Re-identification Set (MARS)</span> is a dataset is collection of CCTV footage </span></div><div class='hero_subdesc'><span class='bgpad'>The MARS dataset includes 1,191,003 of people used for training person re-identification algorithms -</span></div></div></section><section><div class='left-sidebar'><div class='meta'><div><div class='gray'>Collected</div><div>TBD</div></div><div><div class='gray'>Published</div><div>TBD</div></div><div><div class='gray'>Images</div><div>TBD</div></div><div><div class='gray'>Faces</div><div>TBD</div></div></div></div><h2>Motion Analysis and Re-identification Set (MARS)</h2> -<p>(PAGE UNDER DEVELOPMENT)</p> -<p>At vero eos et accusamus et iusto odio dignissimos ducimus, qui blanditiis praesentium voluptatum deleniti atque corrupti, quos dolores et quas molestias excepturi sint, obcaecati cupiditate non-provident, similique sunt in culpa, qui officia deserunt mollitia animi, id est laborum et dolorum fuga. Et harum quidem rerum facilis est et expedita distinctio.</p> -<p>Nam libero tempore, cum soluta nobis est eligendi optio, cumque nihil impedit, quo minus id, quod maxime placeat, facere possimus, omnis voluptas assumenda est, omnis dolor repellendus. Temporibus autem quibusdam et aut officiis debitis aut rerum necessitatibus saepe eveniet, ut et voluptates repudiandae sint et molestiae non-recusandae. Itaque earum rerum hic tenetur a sapiente delectus, ut aut reiciendis voluptatibus maiores alias consequatur aut perferendis doloribus asperiores repellat</p> -</section><section> - - <h3>Biometric Trade Routes (beta)</h3> -<!-- - <div class="map-sidebar right-sidebar"> - <h3>Legend</h3> - <ul> - <li><span style="color: #f2f293">■</span> Industry</li> - <li><span style="color: #f30000">■</span> Academic</li> - <li><span style="color: #3264f6">■</span> Government</li> - </ul> - </div> - --> - <p> - To understand how MARS has been used around the world... - affected global research on computer vision, surveillance, defense, and consumer technology, the and where this dataset has been used the locations of each organization that used or referenced the datast - </p> - - </section> - -<section class="applet_container"> - <div class="applet" data-payload="{"command": "map"}"></div> -</section> - -<div class="caption"> - <div class="map-legend-item edu">Academic</div> - <div class="map-legend-item com">Industry</div> - <div class="map-legend-item gov">Government</div> - Data is compiled from <a href="https://www.semanticscholar.org">Semantic Scholar</a> and not yet manually verified. -</div> - -<section> - <p class='subp'> - The data is generated by collecting all citations for all original research papers associated with the dataset. Then the PDFs are then converted to text and the organization names are extracted and geocoded. Because of the automated approach to extracting data, actual use of the dataset can not yet be confirmed. This visualization is provided to help locate and confirm usage and will be updated as data noise is reduced. - </p> -</section><section> - - - <div class="hr-wave-holder"> - <div class="hr-wave-line hr-wave-line1"></div> - <div class="hr-wave-line hr-wave-line2"></div> - </div> - - <h2>Supplementary Information</h2> -</section><section class="applet_container"> - - <h3>Citations</h3> - <p> - Citations were collected from <a href="https://www.semanticscholar.org">Semantic Scholar</a>, a website which aggregates - and indexes research papers. Metadata was extracted from these papers, including extracting names of institutions automatically from PDFs, and then the addresses were geocoded. Data is not yet manually verified, and reflects anytime the paper was cited. Some papers may only mention the dataset in passing, while others use it as part of their research methodology. - </p> - <p> - Add button/link to download CSV - </p> - - <div class="applet" data-payload="{"command": "citations"}"></div> -</section> - - </div> - <footer> - <div> - <a href="/">MegaPixels.cc</a> - <a href="/about/disclaimer/">Disclaimer</a> - <a href="/about/terms/">Terms of Use</a> - <a href="/about/privacy/">Privacy</a> - <a href="/about/">About</a> - <a href="/about/team/">Team</a> - </div> - <div> - MegaPixels ©2017-19 Adam R. Harvey / - <a href="https://ahprojects.com">ahprojects.com</a> - </div> - </footer> -</body> - -<script src="/assets/js/dist/index.js"></script> -</html>
\ No newline at end of file diff --git a/site/public/datasets/msceleb/index.html b/site/public/datasets/msceleb/index.html new file mode 100644 index 00000000..fb08c737 --- /dev/null +++ b/site/public/datasets/msceleb/index.html @@ -0,0 +1,153 @@ +<!doctype html> +<html> +<head> + <title>MegaPixels</title> + <meta charset="utf-8" /> + <meta name="author" content="Adam Harvey" /> + <meta name="description" content="MS Celeb is a dataset of web images used for training and evaluating face recognition algorithms" /> + <meta name="referrer" content="no-referrer" /> + <meta name="viewport" content="width=device-width, initial-scale=1.0, user-scalable=yes" /> + <link rel='stylesheet' href='/assets/css/fonts.css' /> + <link rel='stylesheet' href='/assets/css/css.css' /> + <link rel='stylesheet' href='/assets/css/leaflet.css' /> + <link rel='stylesheet' href='/assets/css/applets.css' /> +</head> +<body> + <header> + <a class='slogan' href="/"> + <div class='logo'></div> + <div class='site_name'>MegaPixels</div> + <div class='splash'>MsCeleb</div> + </a> + <div class='links'> + <a href="/datasets/">Datasets</a> + <a href="/about/">About</a> + </div> + </header> + <div class="content content-dataset"> + + <section class='intro_section' style='background-image: url(https://nyc3.digitaloceanspaces.com/megapixels/v1/datasets/msceleb/assets/background.jpg)'><div class='inner'><div class='hero_desc'><span class='bgpad'>MS Celeb is a dataset of web images used for training and evaluating face recognition algorithms</span></div><div class='hero_subdesc'><span class='bgpad'>The MS Celeb dataset includes over 10,000,000 images and 93,000 identities of semi-public figures collected using the Bing search engine +</span></div></div></section><section><div class='left-sidebar'><div class='meta'> + <div class='gray'>Published</div> + <div>2016</div> + </div><div class='meta'> + <div class='gray'>Images</div> + <div>1,000,000 </div> + </div><div class='meta'> + <div class='gray'>Identities</div> + <div>100,000 </div> + </div><div class='meta'> + <div class='gray'>Purpose</div> + <div>Large-scale face recognition</div> + </div><div class='meta'> + <div class='gray'>Created by</div> + <div>Microsoft Research</div> + </div><div class='meta'> + <div class='gray'>Funded by</div> + <div>Microsoft Research</div> + </div><div class='meta'> + <div class='gray'>Website</div> + <div><a href='http://www.msceleb.org/' target='_blank' rel='nofollow noopener'>msceleb.org</a></div> + </div><div class='meta'><div><div class='gray'>Published</div><div>TBD</div></div><div><div class='gray'>Images</div><div>TBD</div></div><div><div class='gray'>Faces</div><div>TBD</div></div><div><div class='gray'>Created by</div><div>TBD</div></div></div></div><h2>Microsoft Celeb Dataset (MS Celeb)</h2> +<p>(PAGE UNDER DEVELOPMENT)</p> +<p>At vero eos et accusamus et iusto odio dignissimos ducimus, qui blanditiis praesentium voluptatum deleniti atque corrupti, quos dolores et quas molestias excepturi sint, obcaecati cupiditate non-provident, similique sunt in culpa, qui officia deserunt mollitia animi, id est laborum et dolorum fuga. Et harum quidem rerum facilis est et expedita distinctio.</p> +<p>Nam libero tempore, cum soluta nobis est eligendi optio, cumque nihil impedit, quo minus id, quod maxime placeat, facere possimus, omnis voluptas assumenda est, omnis dolor repellendus. Temporibus autem quibusdam et aut officiis debitis aut rerum necessitatibus saepe eveniet, ut et voluptates repudiandae sint et molestiae non-recusandae. Itaque earum rerum hic tenetur a sapiente delectus, ut aut reiciendis voluptatibus maiores alias consequatur aut perferendis doloribus asperiores repellat</p> +</section><section> + <h3>Who used MsCeleb?</h3> + + <p> + This bar chart presents a ranking of the top countries where dataset citations originated. Mouse over individual columns to see yearly totals. These charts show at most the top 10 countries. + </p> + + </section> + +<section class="applet_container"> +<!-- <div style="position: absolute;top: 0px;right: -55px;width: 180px;font-size: 14px;">Labeled Faces in the Wild Dataset<br><span class="numc" style="font-size: 11px;">20 citations</span> +</div> --> + <div class="applet" data-payload="{"command": "chart"}"></div> +</section><section class="applet_container"> + <div class="applet" data-payload="{"command": "piechart"}"></div> +</section><section> + + <h3>Biometric Trade Routes</h3> +<!-- + <div class="map-sidebar right-sidebar"> + <h3>Legend</h3> + <ul> + <li><span style="color: #f2f293">■</span> Industry</li> + <li><span style="color: #f30000">■</span> Academic</li> + <li><span style="color: #3264f6">■</span> Government</li> + </ul> + </div> + --> + <p> + To help understand how MsCeleb has been used around the world for commercial, military and academic research; publicly available research citing Microsoft Celebrity Dataset is collected, verified, and geocoded to show the biometric trade routes of people appearing in the images. Click on the markers to reveal reserach projects at that location. + </p> + + </section> + +<section class="applet_container fullwidth"> + <div class="applet" data-payload="{"command": "map"}"></div> + +</section> + +<div class="caption"> + <ul class="map-legend"> + <li class="edu">Academic</li> + <li class="com">Commercial</li> + <li class="gov">Military / Government</li> + <li class="source">Citation data is collected using <a href="https://semanticscholar.org" target="_blank">SemanticScholar.org</a> then dataset usage verified and geolocated.</li> + </ul> +</div> + +<!-- <section> + <p class='subp'> + [section under development] MsCeleb ... Standardized paragraph of text about the map. Sed ut perspiciatis, unde omnis iste natus error sit voluptatem accusantium doloremque laudantium, totam rem aperiam eaque ipsa, quae ab illo inventore veritatis et quasi architecto beatae vitae dicta sunt, explicabo. + </p> +</section> + --><section><p>Add more analysis here</p> +</section><section> + + <div class="hr-wave-holder"> + <div class="hr-wave-line hr-wave-line1"></div> + <div class="hr-wave-line hr-wave-line2"></div> + </div> + + <h3>Supplementary Information</h3> + +</section><section class="applet_container"> + + <h3>Dataset Citations</h3> + <p> + The dataset citations used in the visualizations were collected from <a href="https://www.semanticscholar.org">Semantic Scholar</a>, a website which aggregates and indexes research papers. Each citation was geocoded using names of institutions found in the PDF front matter, or as listed on other resources. These papers have been manually verified to show that researchers downloaded and used the dataset to train or test machine learning algorithms. + </p> + + <div class="applet" data-payload="{"command": "citations"}"></div> +</section><section><h3>Additional Information</h3> +<ul> +<li>The dataset author spoke about his research at the CVPR conference in 2016 <a href="https://www.youtube.com/watch?v=Nl2fBKxwusQ">https://www.youtube.com/watch?v=Nl2fBKxwusQ</a></li> +</ul> +</section><section><ul class="footnotes"><li><a name="[^readme]" class="footnote_shim"></a><span class="backlinks"></span><p>"readme.txt" <a href="https://exhibits.stanford.edu/data/catalog/sx925dc9385">https://exhibits.stanford.edu/data/catalog/sx925dc9385</a>.</p> +</li><li><a name="[^localized_region_context]" class="footnote_shim"></a><span class="backlinks"></span><p>Li, Y. and Dou, Y. and Liu, X. and Li, T. Localized Region Context and Object Feature Fusion for People Head Detection. ICIP16 Proceedings. 2016. Pages 594-598.</p> +</li><li><a name="[^replacement_algorithm]" class="footnote_shim"></a><span class="backlinks"></span><p>Zhao. X, Wang Y, Dou, Y. A Replacement Algorithm of Non-Maximum Suppression Base on Graph Clustering.</p> +</li></ul></section> + + </div> + <footer> + <div> + <a href="/">MegaPixels.cc</a> + <a href="/about/disclaimer/">Disclaimer</a> + <a href="/about/terms/">Terms of Use</a> + <a href="/about/privacy/">Privacy</a> + <a href="/about/">About</a> + <a href="/about/team/">Team</a> + </div> + <div> + MegaPixels ©2017-19 Adam R. Harvey / + <a href="https://ahprojects.com">ahprojects.com</a> + </div> + </footer> +</body> + +<script src="/assets/js/dist/index.js"></script> +</html>
\ No newline at end of file diff --git a/site/public/datasets/pipa/index.html b/site/public/datasets/pipa/index.html index 62754070..27168c5c 100644 --- a/site/public/datasets/pipa/index.html +++ b/site/public/datasets/pipa/index.html @@ -8,7 +8,6 @@ <meta name="referrer" content="no-referrer" /> <meta name="viewport" content="width=device-width, initial-scale=1.0, user-scalable=yes" /> <link rel='stylesheet' href='/assets/css/fonts.css' /> - <link rel='stylesheet' href='/assets/css/tabulator.css' /> <link rel='stylesheet' href='/assets/css/css.css' /> <link rel='stylesheet' href='/assets/css/leaflet.css' /> <link rel='stylesheet' href='/assets/css/applets.css' /> @@ -18,6 +17,7 @@ <a class='slogan' href="/"> <div class='logo'></div> <div class='site_name'>MegaPixels</div> + <div class='splash'>PIPA</div> </a> <div class='links'> <a href="/datasets/">Datasets</a> @@ -27,11 +27,29 @@ <div class="content content-dataset"> <section class='intro_section' style='background-image: url(https://nyc3.digitaloceanspaces.com/megapixels/v1/datasets/pipa/assets/background.jpg)'><div class='inner'><div class='hero_desc'><span class='bgpad'><span class="dataset-name"> is a dataset...</span></div><div class='hero_subdesc'><span class='bgpad'>PIPA subdescription -</span></div></div></section><section><div class='left-sidebar'><div class='meta'><div><div class='gray'>Collected</div><div>TBD</div></div><div><div class='gray'>Published</div><div>TBD</div></div><div><div class='gray'>Images</div><div>TBD</div></div><div><div class='gray'>Faces</div><div>TBD</div></div></div></div><h2>Dataset Title TBD</h2> +</span></div></div></section><section><div class='left-sidebar'><div class='meta'> + <div class='gray'>Published</div> + <div>2015</div> + </div><div class='meta'> + <div class='gray'>Images</div> + <div>37,107 </div> + </div><div class='meta'> + <div class='gray'>Identities</div> + <div>2,356 </div> + </div><div class='meta'> + <div class='gray'>Purpose</div> + <div>Face recognition</div> + </div><div class='meta'> + <div class='gray'>Download Size</div> + <div>12 GB</div> + </div><div class='meta'> + <div class='gray'>Website</div> + <div><a href='https://people.eecs.berkeley.edu/~nzhang/piper.html' target='_blank' rel='nofollow noopener'>berkeley.edu</a></div> + </div><div class='meta'><div><div class='gray'>Collected</div><div>TBD</div></div><div><div class='gray'>Published</div><div>TBD</div></div><div><div class='gray'>Images</div><div>TBD</div></div><div><div class='gray'>Faces</div><div>TBD</div></div></div></div><h2>Dataset Title TBD</h2> <p>(PAGE UNDER DEVELOPMENT)</p> </section><section> - <h3>Information Supply Chain</h3> + <h3>Biometric Trade Routes</h3> <!-- <div class="map-sidebar right-sidebar"> <h3>Legend</h3> @@ -43,45 +61,44 @@ </div> --> <p> - To understand how PIPA Dataset has been used around the world... - affected global research on computer vision, surveillance, defense, and consumer technology, the and where this dataset has been used the locations of each organization that used or referenced the datast + To help understand how PIPA Dataset has been used around the world for commercial, military and academic research; publicly available research citing People in Photo Albums Dataset is collected, verified, and geocoded to show the biometric trade routes of people appearing in the images. Click on the markers to reveal reserach projects at that location. </p> </section> -<section class="applet_container"> +<section class="applet_container fullwidth"> <div class="applet" data-payload="{"command": "map"}"></div> + </section> <div class="caption"> - <div class="map-legend-item edu">Academic</div> - <div class="map-legend-item com">Industry</div> - <div class="map-legend-item gov">Government</div> - Data is compiled from <a href="https://www.semanticscholar.org">Semantic Scholar</a> and has been manually verified to show usage of PIPA Dataset. + <ul class="map-legend"> + <li class="edu">Academic</li> + <li class="com">Commercial</li> + <li class="gov">Military / Government</li> + <li class="source">Citation data is collected using <a href="https://semanticscholar.org" target="_blank">SemanticScholar.org</a> then dataset usage verified and geolocated.</li> + </ul> </div> -<section> +<!-- <section> <p class='subp'> - Standardized paragraph of text about the map. Sed ut perspiciatis, unde omnis iste natus error sit voluptatem accusantium doloremque laudantium, totam rem aperiam eaque ipsa, quae ab illo inventore veritatis et quasi architecto beatae vitae dicta sunt, explicabo. + [section under development] PIPA Dataset ... Standardized paragraph of text about the map. Sed ut perspiciatis, unde omnis iste natus error sit voluptatem accusantium doloremque laudantium, totam rem aperiam eaque ipsa, quae ab illo inventore veritatis et quasi architecto beatae vitae dicta sunt, explicabo. </p> -</section><section> - +</section> + --><section> <div class="hr-wave-holder"> <div class="hr-wave-line hr-wave-line1"></div> <div class="hr-wave-line hr-wave-line2"></div> </div> - <h2>Supplementary Information</h2> + <h3>Supplementary Information</h3> + </section><section class="applet_container"> - <h3>Citations</h3> - <p> - Citations were collected from <a href="https://www.semanticscholar.org">Semantic Scholar</a>, a website which aggregates - and indexes research papers. The citations were geocoded using names of institutions found in the PDF front matter, or as listed on other resources. These papers have been manually verified to show that researchers downloaded and used the dataset to train and/or test machine learning algorithms. - </p> + <h3>Dataset Citations</h3> <p> - Add button/link to download CSV + The dataset citations used in the visualizations were collected from <a href="https://www.semanticscholar.org">Semantic Scholar</a>, a website which aggregates and indexes research papers. Each citation was geocoded using names of institutions found in the PDF front matter, or as listed on other resources. These papers have been manually verified to show that researchers downloaded and used the dataset to train or test machine learning algorithms. </p> <div class="applet" data-payload="{"command": "citations"}"></div> diff --git a/site/public/datasets/uccs/index.html b/site/public/datasets/uccs/index.html index 08000c6e..593ac498 100644 --- a/site/public/datasets/uccs/index.html +++ b/site/public/datasets/uccs/index.html @@ -4,11 +4,10 @@ <title>MegaPixels</title> <meta charset="utf-8" /> <meta name="author" content="Adam Harvey" /> - <meta name="description" content="Unconstrained College Students (UCCS) is a dataset of images ..." /> + <meta name="description" content="Unconstrained College Students (UCCS) is a dataset of long-range surveillance photos of students taken without their knowledge" /> <meta name="referrer" content="no-referrer" /> <meta name="viewport" content="width=device-width, initial-scale=1.0, user-scalable=yes" /> <link rel='stylesheet' href='/assets/css/fonts.css' /> - <link rel='stylesheet' href='/assets/css/tabulator.css' /> <link rel='stylesheet' href='/assets/css/css.css' /> <link rel='stylesheet' href='/assets/css/leaflet.css' /> <link rel='stylesheet' href='/assets/css/applets.css' /> @@ -18,6 +17,7 @@ <a class='slogan' href="/"> <div class='logo'></div> <div class='site_name'>MegaPixels</div> + <div class='splash'>UCCS</div> </a> <div class='links'> <a href="/datasets/">Datasets</a> @@ -26,12 +26,215 @@ </header> <div class="content content-dataset"> - <section class='intro_section' style='background-image: url(https://nyc3.digitaloceanspaces.com/megapixels/v1/datasets/uccs/assets/background.jpg)'><div class='inner'><div class='hero_desc'><span class='bgpad'><span class="dataset-name">Unconstrained College Students (UCCS)</span> is a dataset of images ...</span></div><div class='hero_subdesc'><span class='bgpad'>The UCCS dataset includes ... -</span></div></div></section><section><div class='left-sidebar'><div class='meta'><div><div class='gray'>Collected</div><div>TBD</div></div><div><div class='gray'>Published</div><div>TBD</div></div><div><div class='gray'>Images</div><div>TBD</div></div><div><div class='gray'>Faces</div><div>TBD</div></div></div></div><h2>Unconstrained College Students ...</h2> + <section class='intro_section' style='background-image: url(https://nyc3.digitaloceanspaces.com/megapixels/v1/datasets/uccs/assets/background.jpg)'><div class='inner'><div class='hero_desc'><span class='bgpad'><span class="dataset-name">Unconstrained College Students (UCCS)</span> is a dataset of long-range surveillance photos of students taken without their knowledge</span></div><div class='hero_subdesc'><span class='bgpad'>The UCCS dataset includes 16,149 images and 1,732 identities of students at University of Colorado Colorado Springs campus and is used for face recognition and face detection +</span></div></div></section><section><div class='left-sidebar'><div class='meta'> + <div class='gray'>Published</div> + <div>2018</div> + </div><div class='meta'> + <div class='gray'>Images</div> + <div>16,149 </div> + </div><div class='meta'> + <div class='gray'>Identities</div> + <div>1,732 </div> + </div><div class='meta'> + <div class='gray'>Purpose</div> + <div>Face recognition, face detection</div> + </div><div class='meta'> + <div class='gray'>Created by</div> + <div>University of Colorado Colorado Springs (US)</div> + </div><div class='meta'> + <div class='gray'>Funded by</div> + <div>ODNI, IARPA, ONR MURI, Amry SBIR, SOCOM SBIR</div> + </div><div class='meta'> + <div class='gray'>Website</div> + <div><a href='http://vast.uccs.edu/Opensetface/' target='_blank' rel='nofollow noopener'>uccs.edu</a></div> + </div><div class='meta'><div><div class='gray'>Published</div><div>2018</div></div><div><div class='gray'>Images</div><div>16,149</div></div><div><div class='gray'>Identities</div><div>1,732</div></div><div><div class='gray'>Used for</div><div>Face recognition, face detection</div></div><div><div class='gray'>Created by</div><div>Unviversity of Colorado Colorado Springs (US)</div></div><div><div class='gray'>Funded by</div><div>ODNI, IARPA, ONR MURI, Amry SBIR, SOCOM SBIR</div></div><div><div class='gray'>Website</div><div><a href="https://vast.uccs.edu/Opensetface/">vast.uccs.edu</a></div></div></div></div><h2>Unconstrained College Students ...</h2> <p>(PAGE UNDER DEVELOPMENT)</p> -</section><section> +<p>Unconstrained College Students (UCCS) is a dataset of long-range surveillance photos captured at University of Colorado Colorado Springs. According to the authors of two papers associated with the dataset, subjects were "photographed using a long-range high-resolution surveillance camera without their knowledge" [^funding_sb]. The images were captured using a Canon 7D digital camera fitted with a Sigma 800mm telephoto lens pointed out the window of an office.</p> +<p>The UCCS dataset was funded by ODNI (Office of Director of National Intelligence), IARPA (Intelligence Advance Research Projects Activity), ONR MURI Office of Naval Research and The Department of Defense Multidisciplinary University Research Initiative, Army SBIR (Small Business Innovation Research), SOCOM SBIR (Special Operations Command and Small Business Innovation Research), and the National Science Foundation.</p> +<p>The images in UCCS include students walking between classes on campus over 19 days in 2012 - 2013. The dates include:</p> +<table> +<thead><tr> +<th>Year</th> +<th>Month</th> +<th>Day</th> +<th>Date</th> +<th>Time Range</th> +<th>Photos</th> +</tr> +</thead> +<tbody> +<tr> +<td>2012</td> +<td>Februay</td> +<td>---</td> +<td>23</td> +<td>-</td> +<td>132</td> +</tr> +<tr> +<td>2012</td> +<td>March</td> +<td>---</td> +<td>6</td> +<td>-</td> +<td>-</td> +</tr> +<tr> +<td>2012</td> +<td>March</td> +<td>---</td> +<td>8</td> +<td>-</td> +<td>-</td> +</tr> +<tr> +<td>2012</td> +<td>March</td> +<td>---</td> +<td>13</td> +<td>-</td> +<td>-</td> +</tr> +<tr> +<td>2012</td> +<td>Februay</td> +<td>---</td> +<td>23</td> +<td>-</td> +<td>132</td> +</tr> +<tr> +<td>2012</td> +<td>March</td> +<td>---</td> +<td>6</td> +<td>-</td> +<td>-</td> +</tr> +<tr> +<td>2012</td> +<td>March</td> +<td>---</td> +<td>8</td> +<td>-</td> +<td>-</td> +</tr> +<tr> +<td>2012</td> +<td>March</td> +<td>---</td> +<td>13</td> +<td>-</td> +<td>-</td> +</tr> +<tr> +<td>2012</td> +<td>Februay</td> +<td>---</td> +<td>23</td> +<td>-</td> +<td>132</td> +</tr> +<tr> +<td>2012</td> +<td>March</td> +<td>---</td> +<td>6</td> +<td>-</td> +<td>-</td> +</tr> +<tr> +<td>2012</td> +<td>March</td> +<td>---</td> +<td>8</td> +<td>-</td> +<td>-</td> +</tr> +<tr> +<td>2012</td> +<td>March</td> +<td>---</td> +<td>13</td> +<td>-</td> +<td>-</td> +</tr> +<tr> +<td>2012</td> +<td>Februay</td> +<td>---</td> +<td>23</td> +<td>-</td> +<td>132</td> +</tr> +<tr> +<td>2012</td> +<td>March</td> +<td>---</td> +<td>6</td> +<td>-</td> +<td>-</td> +</tr> +<tr> +<td>2012</td> +<td>March</td> +<td>---</td> +<td>8</td> +<td>-</td> +<td>-</td> +</tr> +<tr> +<td>2012</td> +<td>March</td> +<td>---</td> +<td>13</td> +<td>-</td> +<td>-</td> +</tr> +<tr> +<td>2012</td> +<td>Februay</td> +<td>---</td> +<td>23</td> +<td>-</td> +<td>132</td> +</tr> +<tr> +<td>2012</td> +<td>March</td> +<td>---</td> +<td>6</td> +<td>-</td> +<td>-</td> +</tr> +<tr> +<td>2012</td> +<td>March</td> +<td>---</td> +<td>8</td> +<td>-</td> +<td>-</td> +</tr> +</tbody> +</table> +<p>2012-03-20 +2012-03-22 +2012-04-03 +2012-04-12 +2012-04-17 +2012-04-24 +2012-04-25 +2012-04-26 +2013-01-28 +2013-01-29 +2013-02-13 +2013-02-19 +2013-02-20 +2013-02-26</p> +</section><section class='images'><div class='image'><img src='https://nyc3.digitaloceanspaces.com/megapixels/v1/datasets/uccs/assets/uccs_mean_bboxes_comp.jpg' alt=' The pixel-average of all Uconstrained College Students images is shown with all 51,838 face annotations. (c) Adam Harvey'><div class='caption'> The pixel-average of all Uconstrained College Students images is shown with all 51,838 face annotations. (c) Adam Harvey</div></div></section><section> - <h3>Information Supply Chain</h3> + <h3>Biometric Trade Routes</h3> <!-- <div class="map-sidebar right-sidebar"> <h3>Legend</h3> @@ -43,71 +246,63 @@ </div> --> <p> - To understand how UCCS has been used around the world... - affected global research on computer vision, surveillance, defense, and consumer technology, the and where this dataset has been used the locations of each organization that used or referenced the datast + To help understand how UCCS has been used around the world for commercial, military and academic research; publicly available research citing UnConstrained College Students Dataset is collected, verified, and geocoded to show the biometric trade routes of people appearing in the images. Click on the markers to reveal reserach projects at that location. </p> </section> -<section class="applet_container"> +<section class="applet_container fullwidth"> <div class="applet" data-payload="{"command": "map"}"></div> + </section> <div class="caption"> - <div class="map-legend-item edu">Academic</div> - <div class="map-legend-item com">Industry</div> - <div class="map-legend-item gov">Government</div> - Data is compiled from <a href="https://www.semanticscholar.org">Semantic Scholar</a> and has been manually verified to show usage of UCCS. + <ul class="map-legend"> + <li class="edu">Academic</li> + <li class="com">Commercial</li> + <li class="gov">Military / Government</li> + <li class="source">Citation data is collected using <a href="https://semanticscholar.org" target="_blank">SemanticScholar.org</a> then dataset usage verified and geolocated.</li> + </ul> </div> -<section> +<!-- <section> <p class='subp'> - Standardized paragraph of text about the map. Sed ut perspiciatis, unde omnis iste natus error sit voluptatem accusantium doloremque laudantium, totam rem aperiam eaque ipsa, quae ab illo inventore veritatis et quasi architecto beatae vitae dicta sunt, explicabo. + [section under development] UCCS ... Standardized paragraph of text about the map. Sed ut perspiciatis, unde omnis iste natus error sit voluptatem accusantium doloremque laudantium, totam rem aperiam eaque ipsa, quae ab illo inventore veritatis et quasi architecto beatae vitae dicta sunt, explicabo. </p> -</section><section> +</section> + --><section> <h3>Who used UCCS?</h3> <p> - This bar chart presents a ranking of the top countries where citations originated. Mouse over individual columns - to see yearly totals. These charts show at most the top 10 countries. + This bar chart presents a ranking of the top countries where dataset citations originated. Mouse over individual columns to see yearly totals. These charts show at most the top 10 countries. </p> </section> <section class="applet_container"> +<!-- <div style="position: absolute;top: 0px;right: -55px;width: 180px;font-size: 14px;">Labeled Faces in the Wild Dataset<br><span class="numc" style="font-size: 11px;">20 citations</span> +</div> --> <div class="applet" data-payload="{"command": "chart"}"></div> -</section><section> +</section><section class="applet_container"> + <div class="applet" data-payload="{"command": "piechart"}"></div> +</section><section class="applet_container"> + + <h3>Dataset Citations</h3> <p> - These pie charts show overall totals based on country and institution type. + The dataset citations used in the visualizations were collected from <a href="https://www.semanticscholar.org">Semantic Scholar</a>, a website which aggregates and indexes research papers. Each citation was geocoded using names of institutions found in the PDF front matter, or as listed on other resources. These papers have been manually verified to show that researchers downloaded and used the dataset to train or test machine learning algorithms. </p> - - </section> -<section class="applet_container"> - <div class="applet" data-payload="{"command": "piechart"}"></div> + <div class="applet" data-payload="{"command": "citations"}"></div> </section><section> - <div class="hr-wave-holder"> <div class="hr-wave-line hr-wave-line1"></div> <div class="hr-wave-line hr-wave-line2"></div> </div> - <h2>Supplementary Information</h2> -</section><section class="applet_container"> - - <h3>Citations</h3> - <p> - Citations were collected from <a href="https://www.semanticscholar.org">Semantic Scholar</a>, a website which aggregates - and indexes research papers. The citations were geocoded using names of institutions found in the PDF front matter, or as listed on other resources. These papers have been manually verified to show that researchers downloaded and used the dataset to train and/or test machine learning algorithms. - </p> - <p> - Add button/link to download CSV - </p> - - <div class="applet" data-payload="{"command": "citations"}"></div> -</section><section><h3>Research Notes</h3> -<p>The original Sapkota and Boult dataset, from which UCCS is derived, received funding from<sup class="footnote-ref" id="fnref-funding_sb"><a href="#fn-funding_sb">1</a></sup>:</p> + <h3>Supplementary Information</h3> + +</section><section><p>The original Sapkota and Boult dataset, from which UCCS is derived, received funding from<sup class="footnote-ref" id="fnref-funding_sb"><a href="#fn-funding_sb">1</a></sup>:</p> <ul> <li>ONR (Office of Naval Research) MURI (The Department of Defense Multidisciplinary University Research Initiative) grant N00014-08-1-0638</li> <li>Army SBIR (Small Business Innovation Research) grant W15P7T-12-C-A210</li> @@ -119,6 +314,14 @@ <li>ODNI (Office of Director of National Intelligence)</li> <li>IARPA (Intelligence Advance Research Projects Activity) R&D contract 2014-14071600012</li> </ul> +<h3>TODO</h3> +<ul> +<li>add tabulator module for dates</li> +<li>parse dates into CSV using Python</li> +<li>get google image showing line of sight?</li> +<li>fix up quote/citations</li> +</ul> +<h3>footnotes</h3> <div class="footnotes"> <hr> <ol><li id="fn-funding_sb"><p>Sapkota, Archana and Boult, Terrance. "Large Scale Unconstrained Open Set Face Database." 2013.<a href="#fnref-funding_sb" class="footnote">↩</a></p></li> diff --git a/site/public/datasets/vgg_face2/index.html b/site/public/datasets/vgg_face2/index.html index 42e3b961..d5c1d98c 100644 --- a/site/public/datasets/vgg_face2/index.html +++ b/site/public/datasets/vgg_face2/index.html @@ -8,7 +8,6 @@ <meta name="referrer" content="no-referrer" /> <meta name="viewport" content="width=device-width, initial-scale=1.0, user-scalable=yes" /> <link rel='stylesheet' href='/assets/css/fonts.css' /> - <link rel='stylesheet' href='/assets/css/tabulator.css' /> <link rel='stylesheet' href='/assets/css/css.css' /> <link rel='stylesheet' href='/assets/css/leaflet.css' /> <link rel='stylesheet' href='/assets/css/applets.css' /> @@ -18,6 +17,7 @@ <a class='slogan' href="/"> <div class='logo'></div> <div class='site_name'>MegaPixels</div> + </a> <div class='links'> <a href="/datasets/">Datasets</a> diff --git a/site/public/datasets/viper/index.html b/site/public/datasets/viper/index.html index 5acd0845..6d27b15b 100644 --- a/site/public/datasets/viper/index.html +++ b/site/public/datasets/viper/index.html @@ -8,7 +8,6 @@ <meta name="referrer" content="no-referrer" /> <meta name="viewport" content="width=device-width, initial-scale=1.0, user-scalable=yes" /> <link rel='stylesheet' href='/assets/css/fonts.css' /> - <link rel='stylesheet' href='/assets/css/tabulator.css' /> <link rel='stylesheet' href='/assets/css/css.css' /> <link rel='stylesheet' href='/assets/css/leaflet.css' /> <link rel='stylesheet' href='/assets/css/applets.css' /> @@ -18,6 +17,7 @@ <a class='slogan' href="/"> <div class='logo'></div> <div class='site_name'>MegaPixels</div> + <div class='splash'>VIPeR</div> </a> <div class='links'> <a href="/datasets/">Datasets</a> @@ -27,7 +27,25 @@ <div class="content content-dataset"> <section class='intro_section' style='background-image: url(https://nyc3.digitaloceanspaces.com/megapixels/v1/datasets/viper/assets/background.jpg)'><div class='inner'><div class='hero_desc'><span class='bgpad'><span class="dataset-name">VIPeR</span> is a person re-identification dataset of images captured at UC Santa Cruz in 2007</span></div><div class='hero_subdesc'><span class='bgpad'>VIPeR contains 1,264 images and 632 persons on the UC Santa Cruz campus and is used to train person re-identification algorithms for surveillance -</span></div></div></section><section><div class='left-sidebar'><div class='meta'><div><div class='gray'>Published</div><div>2007</div></div><div><div class='gray'>Images</div><div>1,264</div></div><div><div class='gray'>Persons</div><div>632</div></div><div><div class='gray'>Created by</div><div>UC Santa Cruz</div></div></div></div><h2>VIPeR Dataset</h2> +</span></div></div></section><section><div class='left-sidebar'><div class='meta'> + <div class='gray'>Published</div> + <div>2007</div> + </div><div class='meta'> + <div class='gray'>Images</div> + <div>1,264 </div> + </div><div class='meta'> + <div class='gray'>Identities</div> + <div>632 </div> + </div><div class='meta'> + <div class='gray'>Purpose</div> + <div>Person re-identification</div> + </div><div class='meta'> + <div class='gray'>Created by</div> + <div>University of California Santa Cruz</div> + </div><div class='meta'> + <div class='gray'>Website</div> + <div><a href='https://vision.soe.ucsc.edu/node/178' target='_blank' rel='nofollow noopener'>ucsc.edu</a></div> + </div><div class='meta'><div><div class='gray'>Published</div><div>2007</div></div><div><div class='gray'>Images</div><div>1,264</div></div><div><div class='gray'>Persons</div><div>632</div></div><div><div class='gray'>Created by</div><div>UC Santa Cruz</div></div></div></div><h2>VIPeR Dataset</h2> <p>(PAGE UNDER DEVELOPMENT)</p> <p><em>VIPeR (Viewpoint Invariant Pedestrian Recognition)</em> is a dataset of pedestrian images captured at University of California Santa Cruz in 2007. Accoriding to the reserachers 2 "cameras were placed in different locations in an academic setting and subjects were notified of the presence of cameras, but were not coached or instructed in any way."</p> <p>VIPeR is amongst the most widely used publicly available person re-identification datasets. In 2017 the VIPeR dataset was combined into a larger person re-identification created by the Chinese University of Hong Kong called PETA (PEdesTrian Attribute).</p> @@ -35,26 +53,20 @@ <h3>Who used VIPeR?</h3> <p> - This bar chart presents a ranking of the top countries where citations originated. Mouse over individual columns - to see yearly totals. These charts show at most the top 10 countries. + This bar chart presents a ranking of the top countries where dataset citations originated. Mouse over individual columns to see yearly totals. These charts show at most the top 10 countries. </p> </section> <section class="applet_container"> +<!-- <div style="position: absolute;top: 0px;right: -55px;width: 180px;font-size: 14px;">Labeled Faces in the Wild Dataset<br><span class="numc" style="font-size: 11px;">20 citations</span> +</div> --> <div class="applet" data-payload="{"command": "chart"}"></div> -</section><section> - <p> - These pie charts show overall totals based on country and institution type. - </p> - - </section> - -<section class="applet_container"> +</section><section class="applet_container"> <div class="applet" data-payload="{"command": "piechart"}"></div> </section><section> - <h3>Information Supply Chain</h3> + <h3>Biometric Trade Routes</h3> <!-- <div class="map-sidebar right-sidebar"> <h3>Legend</h3> @@ -66,45 +78,44 @@ </div> --> <p> - To understand how VIPeR has been used around the world... - affected global research on computer vision, surveillance, defense, and consumer technology, the and where this dataset has been used the locations of each organization that used or referenced the datast + To help understand how VIPeR has been used around the world for commercial, military and academic research; publicly available research citing Viewpoint Invariant Pedestrian Recognition is collected, verified, and geocoded to show the biometric trade routes of people appearing in the images. Click on the markers to reveal reserach projects at that location. </p> </section> -<section class="applet_container"> +<section class="applet_container fullwidth"> <div class="applet" data-payload="{"command": "map"}"></div> + </section> <div class="caption"> - <div class="map-legend-item edu">Academic</div> - <div class="map-legend-item com">Industry</div> - <div class="map-legend-item gov">Government</div> - Data is compiled from <a href="https://www.semanticscholar.org">Semantic Scholar</a> and has been manually verified to show usage of VIPeR. + <ul class="map-legend"> + <li class="edu">Academic</li> + <li class="com">Commercial</li> + <li class="gov">Military / Government</li> + <li class="source">Citation data is collected using <a href="https://semanticscholar.org" target="_blank">SemanticScholar.org</a> then dataset usage verified and geolocated.</li> + </ul> </div> -<section> +<!-- <section> <p class='subp'> - Standardized paragraph of text about the map. Sed ut perspiciatis, unde omnis iste natus error sit voluptatem accusantium doloremque laudantium, totam rem aperiam eaque ipsa, quae ab illo inventore veritatis et quasi architecto beatae vitae dicta sunt, explicabo. + [section under development] VIPeR ... Standardized paragraph of text about the map. Sed ut perspiciatis, unde omnis iste natus error sit voluptatem accusantium doloremque laudantium, totam rem aperiam eaque ipsa, quae ab illo inventore veritatis et quasi architecto beatae vitae dicta sunt, explicabo. </p> -</section><section> - +</section> + --><section> <div class="hr-wave-holder"> <div class="hr-wave-line hr-wave-line1"></div> <div class="hr-wave-line hr-wave-line2"></div> </div> - <h2>Supplementary Information</h2> + <h3>Supplementary Information</h3> + </section><section class="applet_container"> - <h3>Citations</h3> - <p> - Citations were collected from <a href="https://www.semanticscholar.org">Semantic Scholar</a>, a website which aggregates - and indexes research papers. The citations were geocoded using names of institutions found in the PDF front matter, or as listed on other resources. These papers have been manually verified to show that researchers downloaded and used the dataset to train and/or test machine learning algorithms. - </p> + <h3>Dataset Citations</h3> <p> - Add button/link to download CSV + The dataset citations used in the visualizations were collected from <a href="https://www.semanticscholar.org">Semantic Scholar</a>, a website which aggregates and indexes research papers. Each citation was geocoded using names of institutions found in the PDF front matter, or as listed on other resources. These papers have been manually verified to show that researchers downloaded and used the dataset to train or test machine learning algorithms. </p> <div class="applet" data-payload="{"command": "citations"}"></div> diff --git a/site/public/datasets/youtube_celebrities/index.html b/site/public/datasets/youtube_celebrities/index.html index dd230926..ee11f3c0 100644 --- a/site/public/datasets/youtube_celebrities/index.html +++ b/site/public/datasets/youtube_celebrities/index.html @@ -8,7 +8,6 @@ <meta name="referrer" content="no-referrer" /> <meta name="viewport" content="width=device-width, initial-scale=1.0, user-scalable=yes" /> <link rel='stylesheet' href='/assets/css/fonts.css' /> - <link rel='stylesheet' href='/assets/css/tabulator.css' /> <link rel='stylesheet' href='/assets/css/css.css' /> <link rel='stylesheet' href='/assets/css/leaflet.css' /> <link rel='stylesheet' href='/assets/css/applets.css' /> @@ -18,6 +17,7 @@ <a class='slogan' href="/"> <div class='logo'></div> <div class='site_name'>MegaPixels</div> + <div class='splash'>YouTube Celebrities</div> </a> <div class='links'> <a href="/datasets/">Datasets</a> |
