summaryrefslogtreecommitdiff
path: root/site/public/research
diff options
context:
space:
mode:
Diffstat (limited to 'site/public/research')
-rw-r--r--site/public/research/_from_1_to_100_pixels/index.html20
-rw-r--r--site/public/research/_introduction/index.html22
-rw-r--r--site/public/research/_what_computers_can_see/index.html20
-rw-r--r--site/public/research/index.html13
-rw-r--r--site/public/research/munich_security_conference/index.html55
5 files changed, 51 insertions, 79 deletions
diff --git a/site/public/research/_from_1_to_100_pixels/index.html b/site/public/research/_from_1_to_100_pixels/index.html
index 74f334cc..a978b264 100644
--- a/site/public/research/_from_1_to_100_pixels/index.html
+++ b/site/public/research/_from_1_to_100_pixels/index.html
@@ -50,27 +50,13 @@
<div class='links'>
<a href="/datasets/">Datasets</a>
<a href="/about/">About</a>
- <a href="/about/news">News</a>
+ <a href="/research">Research</a>
</div>
</header>
<div class="content content-">
- <section>
- <h1>From 1 to 100 Pixels</h1>
- <div class='meta'>
- <div>
- <div class='gray'>Posted</div>
- <div>2018-12-04</div>
- </div>
- <div>
- <div class='gray'>By</div>
- <div>Adam Harvey</div>
- </div>
-
- </div>
- </section>
-
- <section><h3>High resolution insights from low resolution data</h3>
+ <section><h1>From 1 to 100 Pixels</h1>
+<h3>High resolution insights from low resolution data</h3>
<p>This post will be about the meaning of "face". How do people define it? How to biometrics researchers define it? How has it changed during the last decade.</p>
<p>What can you know from a very small amount of information?</p>
<ul>
diff --git a/site/public/research/_introduction/index.html b/site/public/research/_introduction/index.html
index 66905247..8b17c016 100644
--- a/site/public/research/_introduction/index.html
+++ b/site/public/research/_introduction/index.html
@@ -50,27 +50,13 @@
<div class='links'>
<a href="/datasets/">Datasets</a>
<a href="/about/">About</a>
- <a href="/about/news">News</a>
+ <a href="/research">Research</a>
</div>
</header>
<div class="content content-dataset">
- <section>
- <h1>Introducing MegaPixels</h1>
- <div class='meta'>
- <div>
- <div class='gray'>Posted</div>
- <div>2018-12-15</div>
- </div>
- <div>
- <div class='gray'>By</div>
- <div>Adam Harvey</div>
- </div>
-
- </div>
- </section>
-
- <section><p>Face recognition has become the focal point for ...</p>
+ <section><h1>Introduction</h1>
+<p>Face recognition has become the focal point for ...</p>
<p>Add 68pt landmarks animation</p>
<p>But biometric currency is ...</p>
<p>Add rotation 3D head</p>
@@ -82,7 +68,7 @@
<li>Posted: Dec. 15</li>
<li>Author: Adam Harvey</li>
</ul>
-</section><section class='applet_container'><div class='applet' data-payload='{"command": "load_file /site/research/00_introduction/assets/summary_countries_top.csv", "fields": ["country, Xcitations"]}'></div></section><section><p>Paragraph text to test css formatting. Paragraph text to test css formatting. Paragraph text to test css formatting. Paragraph text to test css formatting. Paragraph text to test css formatting.</p>
+</section><section class='applet_container'><div class='applet' data-payload='{"command": "load_file /site/research/00_introduction/assets/summary_countries_top.csv", "fields": ["Headings: country, Xcitations"]}'></div></section><section><p>Paragraph text to test css formatting. Paragraph text to test css formatting. Paragraph text to test css formatting. Paragraph text to test css formatting. Paragraph text to test css formatting.</p>
<p>[ page under development ]</p>
</section><section class='images'><div class='image'><img src='https://nyc3.digitaloceanspaces.com/megapixels/v1/site/research/_introduction/assets/test.png' alt=' This is the caption'><div class='caption'> This is the caption</div></div></section>
diff --git a/site/public/research/_what_computers_can_see/index.html b/site/public/research/_what_computers_can_see/index.html
index 003dd733..35f6d47d 100644
--- a/site/public/research/_what_computers_can_see/index.html
+++ b/site/public/research/_what_computers_can_see/index.html
@@ -50,27 +50,13 @@
<div class='links'>
<a href="/datasets/">Datasets</a>
<a href="/about/">About</a>
- <a href="/about/news">News</a>
+ <a href="/research">Research</a>
</div>
</header>
<div class="content content-">
- <section>
- <h1>What Computers Can See</h1>
- <div class='meta'>
- <div>
- <div class='gray'>Posted</div>
- <div>2018-12-15</div>
- </div>
- <div>
- <div class='gray'>By</div>
- <div>Adam Harvey</div>
- </div>
-
- </div>
- </section>
-
- <section><p>Rosalind Picard on Affective Computing Podcast with Lex Fridman</p>
+ <section><h1>What Computers Can See About Your Face</h1>
+<p>Rosalind Picard on Affective Computing Podcast with Lex Fridman</p>
<ul>
<li>we can read with an ordinary camera on your phone, from a neutral face if</li>
<li>your heart is racing</li>
diff --git a/site/public/research/index.html b/site/public/research/index.html
index 571b8230..f4f90531 100644
--- a/site/public/research/index.html
+++ b/site/public/research/index.html
@@ -50,13 +50,22 @@
<div class='links'>
<a href="/datasets/">Datasets</a>
<a href="/about/">About</a>
- <a href="/about/news">News</a>
+ <a href="/research">Research</a>
</div>
</header>
<div class="content content-">
<section><h1>Research Blog</h1>
-</section><div class='research_index'><a href='/research/_introduction/'><section class='wide'><img src='data:image/gif;base64,R0lGODlhAQABAAAAACH5BAEKAAEALAAAAAABAAEAAAICTAEAOw==' alt='Research post' /><section><h1>Introducing MegaPixels</h1><h2></h2></section></section></a><a href='/research/munich_security_conference/'><section class='wide'><img src='https://nyc3.digitaloceanspaces.com/megapixels/v1/site/research/munich_security_conference/assets/background.jpg' alt='Research post' /><section><h1>Transnational Data Analysis of Publicly Available Face Recognition Training Datasets</h1><h2></h2></section></section></a></div>
+</section><div class='research_index'>
+ <a href='/research/munich_security_conference/'><section class='wide' style='background-image: url(https://nyc3.digitaloceanspaces.com/megapixels/v1/site/research/munich_security_conference/assets/background.jpg);' />
+ <section>
+ <h4><span class='bgpad'>28 June 2019</span></h4>
+ <h2><span class='bgpad'>Analyzing Transnational Flows of Face Recognition Image Training Data</span></h2>
+ <h3><span class='bgpad'>Where does face data originate and who's using it?</span></h3>
+ <h4 class='readmore'><span class='bgpad'>Read more...</span></h4>
+ </section>
+ </section></a>
+ </div>
</div>
<footer>
diff --git a/site/public/research/munich_security_conference/index.html b/site/public/research/munich_security_conference/index.html
index 499d8e9f..b0503f84 100644
--- a/site/public/research/munich_security_conference/index.html
+++ b/site/public/research/munich_security_conference/index.html
@@ -4,7 +4,7 @@
<title>MegaPixels: MSC</title>
<meta charset="utf-8" />
<meta name="author" content="Adam Harvey" />
- <meta name="description" content="Analyzing the Transnational Flow of Facial Recognition Data" />
+ <meta name="description" content="Analyzing Transnational Flows of Face Recognition Image Training Data" />
<meta property="og:title" content="MegaPixels: MSC"/>
<meta property="og:type" content="website"/>
<meta property="og:summary" content="MegaPixels is an art and research project about face recognition datasets created \"in the wild\"/>
@@ -50,31 +50,36 @@
<div class='links'>
<a href="/datasets/">Datasets</a>
<a href="/about/">About</a>
- <a href="/about/news">News</a>
+ <a href="/research">Research</a>
</div>
</header>
<div class="content content-dataset">
- <section>
- <h1>MSC</h1>
- <div class='meta'>
- <div>
- <div class='gray'>Posted</div>
- <div>2019-4-18</div>
- </div>
- <div>
- <div class='gray'>By</div>
- <div>Adam Harvey</div>
- </div>
-
- </div>
- </section>
-
- <section class='intro_section' style='background-image: url(https://nyc3.digitaloceanspaces.com/megapixels/v1/site/research/munich_security_conference/assets/background.jpg)'><div class='inner'><div class='hero_desc'><span class='bgpad'>Analyzing the Transnational Flow of Facial Recognition Data</span></div><div class='hero_subdesc'><span class='bgpad'>Where does face data originate and who's using it?
-</span></div></div></section><section><p>[page under devlopment]</p>
-<p>Intro paragraph.</p>
-<p>[ add montage of extracted faces here]</p>
-</section><section class='images'><div class='image'><img src='https://nyc3.digitaloceanspaces.com/megapixels/v1/site/research/munich_security_conference/assets/montage_placeholder.jpg' alt=' Placeholder caption'><div class='caption'> Placeholder caption</div></div></section><section class='images'><div class='image'><img src='https://nyc3.digitaloceanspaces.com/megapixels/v1/site/research/munich_security_conference/assets/bar_placeholder.png' alt=' Placeholder caption'><div class='caption'> Placeholder caption</div></div></section><section><div class='columns columns-2'><section class='images'><div class='image'><img src='https://nyc3.digitaloceanspaces.com/megapixels/v1/site/research/munich_security_conference/assets/pie_placeholder.png' alt=' Placeholder caption'><div class='caption'> Placeholder caption</div></div></section><section class='images'><div class='image'><img src='https://nyc3.digitaloceanspaces.com/megapixels/v1/site/research/munich_security_conference/assets/pie_placeholder.png' alt=' Placeholder caption'><div class='caption'> Placeholder caption</div></div></section></div></section><section>
+ <section class='intro_section' style='background-image: url(https://nyc3.digitaloceanspaces.com/megapixels/v1/site/research/munich_security_conference/assets/background.jpg)'><div class='inner'><div class='hero_desc'><span class='bgpad'>Analyzing Transnational Flows of Face Recognition Image Training Data</span></div><div class='hero_subdesc'><span class='bgpad'>Where does face data originate and who's using it?
+</span></div></div></section><section><h2>Face Datasets and Information Supply Chains</h2>
+</section><section><div class='right-sidebar'><div class='meta'><div class='gray'>Images Analyzed</div><div>24,302,637</div></div><div class='meta'><div class='gray'>Datasets Analyzed</div><div>30</div></div><div class='meta'><div class='gray'>Years</div><div>2006 - 2018</div></div><div class='meta'><div class='gray'>Status</div><div>Ongoing Investigation</div></div><div class='meta'><div class='gray'>Last Updated</div><div>June 28, 2019</div></div></div><p>National AI strategies often rely on transnational data sources to capitalize on recent advancements in deep learning and neural networks. Researchers benefiting from these transnational data flows can yield quick and significant gains across diverse sectors from health care to biometrics. But new challenges emerge when national AI strategies collide with national interests.</p>
+<p>Our earlier research on the <a href="/datsets">MS Celeb</a> and <a href="/datsets/duke_mtmc">Duke</a> datasets published with the Financial Times revealed that several computer vision image datasets created by US companies and universities were unexpectedly also used for research by the National University of Defense Technology in China, along with top Chinese surveillance firms including SenseTime, SenseNets, CloudWalk, Hikvision, and Megvii/Face++ which have all been linked to the oppressive surveillance of Uighur Muslims in Xinjiang.</p>
+<p>In this new research for the Munich Security Conference's Transnational Security Report we provide summary statistics about the origins and endpoints of facial recognition information supply chains. To make it more personal, we gathered additional data on the number of public photos from Embassies that are currently being used in facial recognition datasets.</p>
+<h3>24 Million Non-Cooperative Faces</h3>
+<p>In total, we analyzed 30 publicly available face recognition and face analysis datasets that collectively include over 24 million non-cooperative images. Of these 24 million images, over 15 million face images are from Internet search engines, over 5.8 million from Flickr.com, over 2.5 million from the Internet Movie Database (IMDb.com), and nearly 500,000 from CCTV footage. All 24 million images were collected without any explicit consent, a type of face image researchers call "in the wild".</p>
+<p>Next we manually verified 1,134 publicly available research papers that cite these datasets to determine who was using the data and where it was being used. Even though all of the images originated in the United States, the publicly available research citations show that only about 25% citations are from the country of the origin while the majority of citations are from China.</p>
+</section><section><div class='columns columns-2'><section class='applet_container'><div class='applet' data-payload='{"command": "single_pie_chart /site/research/munich_security_conference/assets/megapixels_origins_top.csv", "fields": ["Caption: Sources of Publicly Available Non-Cooperative Face Image Training Data 2006 - 2018", "Top: 10", "OtherLabel: Other"]}'></div></section><section class='applet_container'><div class='applet' data-payload='{"command": "single_pie_chart /site/research/munich_security_conference/assets/summary_countries.csv", "fields": ["Caption: Locations Where Face Data Is Used Based on Public Research Citations", "Top: 14", "OtherLabel: Other"]}'></div></section></div></section><section><h3>6,000 Embassy Photos Being Used To Train Facial Recognition</h3>
+<p>Of the 5.8 million Flickr images we found over 6,000 public photos from Embassy Flickr accounts were used to train facial recognition technologies. These images were used in the MegaFace, IBM Diversity in Faces datasets. Over 2,000 more images were used in the Who Goes There datasets used for facial ethnicity analysis research. A few of the embassy images found in facial recognition datasets are shown below.</p>
+</section><section><div class='columns columns-2'><section class='applet_container'><div class='applet' data-payload='{"command": "single_pie_chart /site/research/munich_security_conference/assets/country_counts.csv", "fields": ["Caption: Photos from these embassies are being used to train face recognition software", "Top: 4", "OtherLabel: Other", "Colors: categoryRainbow"]}'></div></section><section class='applet_container'><div class='applet' data-payload='{"command": "single_pie_chart /site/research/munich_security_conference/assets/embassy_counts_summary_dataset.csv", "fields": ["Caption: Embassy images were found in these datasets", "Top: 4", "OtherLabel: Other", "Colors: categoryRainbow"]}'></div></section></div></section><section class='images'><div class='image'><img src='https://nyc3.digitaloceanspaces.com/megapixels/v1/site/research/munich_security_conference/assets/4606260362.jpg' alt=' An image in the MegaFace dataset obtained from United Kingdoms Embassy in Italy'><div class='caption'> An image in the MegaFace dataset obtained from United Kingdom's Embassy in Italy</div></div>
+<div class='image'><img src='https://nyc3.digitaloceanspaces.com/megapixels/v1/site/research/munich_security_conference/assets/4749096858.jpg' alt=' An image in the MegaFace dataset obtained from the Flickr account of the United States Embassy in Kabul, Afghanistan'><div class='caption'> An image in the MegaFace dataset obtained from the Flickr account of the United States Embassy in Kabul, Afghanistan</div></div></section><section class='images'><div class='image'><img src='https://nyc3.digitaloceanspaces.com/megapixels/v1/site/research/munich_security_conference/assets/4730007024.jpg' alt=' An image in the MegaFace dataset obtained from U.S. Embassy Canberra'><div class='caption'> An image in the MegaFace dataset obtained from U.S. Embassy Canberra</div></div></section><section><p>This brief research aims to shed light on the emerging politics of data. A photo is no longer just a photo when it can also be surveillance training data, and datasets can no longer be separated from the development of software when software is now built with data. "Our relationship to computers has changed", says Geoffrey Hinton, one of the founders of modern day neural networks and deep learning. "Instead of programming them, we now show them and they figure it out."<a class="footnote_shim" name="[^hinton]_1"> </a><a href="#[^hinton]" class="footnote" title="Footnote 1">1</a>.</p>
+<p>National AI strategies might also want to include transnational dataset strategies.</p>
+<p><em>This research post is going and will updated during July and August, 2019.</em></p>
+<h3>Further Reading</h3>
+<ul>
+<li><a href="/datasets/msceleb">MS Celeb Dataset Analysis</a></li>
+<li><a href="/datasets/brainwash">Brainwash Dataset Analysis</a></li>
+<li><a href="/datasets/duke_mtmc">Duke MTMC Dataset Analysis</a></li>
+<li><a href="/datasets/uccs">Unconstrained College Students Dataset Analysis</a></li>
+<li><a href="https://www.dukechronicle.com/article/2019/06/duke-university-facial-recognition-data-set-study-surveillance-video-students-china-uyghur">Duke MTMC dataset author apologies to students</a></li>
+<li><a href="https://www.bbc.com/news/technology-48555149">BBC coverage of MS Celeb dataset takedown</a></li>
+<li><a href="https://www.spiegel.de/netzwelt/web/microsoft-gesichtserkennung-datenbank-mit-zehn-millionen-fotos-geloescht-a-1271221.html">Spiegel coverage of MS Celeb dataset takdown</a></li>
+</ul>
+</section><section>
<div class="hr-wave-holder">
<div class="hr-wave-line hr-wave-line1"></div>
@@ -83,8 +88,7 @@
<h2>Supplementary Information</h2>
-</section><section><p>[ add a download button for CSV data ]</p>
-</section><section class='applet_container'><div class='applet' data-payload='{"command": "load_file /site/research/munich_security_conference/assets/embassy_counts_public.csv", "fields": ["Images, Dataset, Embassy, Flickr ID, URL, Guest, Host"]}'></div></section><section>
+</section><section class='applet_container'><div class='applet' data-payload='{"command": "load_file /site/research/munich_security_conference/assets/embassy_counts_public.csv", "fields": ["Headings: Images, Dataset, Embassy, Flickr ID, URL, Guest, Host"]}'></div></section><section>
<h4>Cite Our Work</h4>
<p>
@@ -101,7 +105,8 @@
}</pre>
</p>
-</section>
+</section><section><h3>References</h3><section><ul class="footnotes"><li>1 <a name="[^hinton]" class="footnote_shim"></a><span class="backlinks"><a href="#[^hinton]_1">a</a></span>"Heroes of Deep Learning: Andrew Ng interviews Geoffrey Hinton". Published on Aug 8, 2017. <a href="https://www.youtube.com/watch?v=-eyhCTvrEtE">https://www.youtube.com/watch?v=-eyhCTvrEtE</a>
+</li></ul></section></section>
</div>
<footer>