diff options
Diffstat (limited to 'scraper/samples')
| -rw-r--r-- | scraper/samples/s2-orc-paper.json | 131 | ||||
| -rw-r--r-- | scraper/samples/s2-paper-detail.json | 3114 | ||||
| -rw-r--r-- | scraper/samples/s2-papers-api.json | 2855 | ||||
| -rw-r--r-- | scraper/samples/s2-search-api.json | 270 |
4 files changed, 6370 insertions, 0 deletions
diff --git a/scraper/samples/s2-orc-paper.json b/scraper/samples/s2-orc-paper.json new file mode 100644 index 00000000..1fb29126 --- /dev/null +++ b/scraper/samples/s2-orc-paper.json @@ -0,0 +1,131 @@ +{ + "entities": [ + "Database", + "Experiment", + "Facial recognition system", + "Optimization problem", + "Program optimization", + "Simultaneous localization and mapping", + "Taxicab geometry", + "The Matrix", + "Video game localization" + ], + "journalVolume": "", + "journalPages": "3871-3879", + "pmid": "", + "year": 2015, + "outCitations": [ + "bd5accc96a772f2bbb7c64af0aa0b600b0b8594b", + "0e3fcfe63b7b6620e3c47e9751fe3456e85cc52f", + "023f6fc69fe1f6498e35dbf85932ecb549d36ca4", + "54aafe33c4a32fb5875d04fd4a6e6da50920c263", + "9f87e3212ab1d89c5d46924c925d6bb1da02f92b", + "0f0fcf041559703998abf310e56f8a2f90ee6f21", + "53b919c994b3658cca8178e455681ca244a7afad", + "084bd02d171e36458f108f07265386f22b34a1ae", + "14ce7635ff18318e7094417d0f92acbec6669f1c", + "e42998bbebddeeb4b2bedf5da23fa5c4efc976fa", + "0cb48f543c4bf329a16c0408c4d2c198679a6057", + "0de91641f37b0a81a892e4c914b46d05d33fd36e", + "3f204a413d9c8c16f146c306c8d96b91839fed0c", + "3393459600368be2c4c9878a3f65a57dcc0c2cfa", + "1824b1ccace464ba275ccc86619feaa89018c0ad", + "70eeacf9f86ba08fceb3dd703cf015016dac1930", + "5040f7f261872a30eec88788f98326395a44db03", + "95f12d27c3b4914e0668a268360948bce92f7db3", + "2ee56d6c29072abd576ef16bcb26360c42caaf5f", + "370b5757a5379b15e30d619e4d3fb9e8e13f3256", + "0296fc4d042ca8657a7d9dd02df7eb7c0a0017ad", + "129388116bc3229546a84b9bc3d11fdac8b93201", + "63f9f3f0e1daede934d6dde1a84fb7994f8929f0", + "03f98c175b4230960ac347b1100fbfc10c100d0c", + "32d6ec2810b52d1df128be51464bc43b53702232", + "1a1a60fd4dc88a14c016b95789385801c6b80574", + "800683c891b9c5934246ce2931d1d28e0a364fbf", + "0e986f51fe45b00633de9fd0c94d082d2be51406", + "0dccc881cb9b474186a01fd60eb3a3e061fa6546", + "f731b6745d829241941307c3ebf163e90e200318", + "07119bc66e256f88b7436e62a4ac3384365e4e9b", + "0a5bf19c1b297713a3267bc670b252f9e51c2b78", + "013909077ad843eb6df7a3e8e290cfd5575999d2", + "177bc509dd0c7b8d388bb47403f28d6228c14b5c", + "b2d9877443ec7da2490027ccc932468f05c7bf85", + "b960984881665f692764069015764d25974c8d1b", + "4998462014c907c519ae801af39cf58a4c538bc9", + "5d1c4e93e32ee686234c5aae7f38025523993c8c", + "64d5772f44efe32eb24c9968a3085bc0786bfca7", + "1c08824da1383051fba69384a3edf135ba58e7fd", + "b5a3cc76ceee9489e66c5929c66aec1cf5210241", + "4068574b8678a117d9a434360e9c12fe6232dae0", + "830e5b1043227fe189b3f93619ef4c58868758a7", + "044d9a8c61383312cdafbcc44b9d00d650b21c70", + "3957b51b44f8727fe008162ea8a142a8c7917dea", + "55b4b1168c734eeb42882082bd131206dbfedd5b", + "140438a77a771a8fb656b39a78ff488066eb6b50", + "31a38fd2d9d4f34d2b54318021209fe5565b8f7f" + ], + "s2Url": "https://semanticscholar.org/paper/788a7b59ea72e23ef4f86dc9abb4450efefeca41", + "s2PdfUrl": "", + "id": "788a7b59ea72e23ef4f86dc9abb4450efefeca41", + "authors": [ + { + "name": "Christos Sagonas", + "ids": [ + "3320415" + ] + }, + { + "name": "Yannis Panagakis", + "ids": [ + "1780393" + ] + }, + { + "name": "Stefanos Zafeiriou", + "ids": [ + "1776444" + ] + }, + { + "name": "Maja Pantic", + "ids": [ + "1694605" + ] + } + ], + "journalName": "2015 IEEE International Conference on Computer Vision (ICCV)", + "paperAbstract": "Recently, it has been shown that excellent resultscan be achieved in both facial landmark localization and pose-invariant face recognition. These breakthroughs are attributed to the efforts of the community to manually annotate facial images in many different poses and to collect 3D facial data. In this paper, we propose a novel method for joint frontal view reconstruction and landmark localization using a small set of frontal images only. By observing that the frontal facial image is the one having the minimum rank of all different poses, an appropriate model which is able to jointly recover the frontalized version of the face as well as the facial landmarks is devised. To this end, a suitable optimization problem, involving the minimization of the nuclear norm and the matrix l1 norm is solved. The proposed method is assessed in frontal face reconstruction, face landmark localization, pose-invariant face recognition, and face verification in unconstrained conditions. The relevant experiments have been conducted on 8 databases. The experimental results demonstrate the effectiveness of the proposed method in comparison to the state-of-the-art methods for the target problems.", + "inCitations": [ + "0e2365254f7833a1a902a6439f42ece6e5326332", + "368d59cf1733af511ed8abbcbeb4fb47afd4da1c", + "0db8e6eb861ed9a70305c1839eaef34f2c85bbaf", + "fd46586ff7d1cdedc184051268bb07a6630bc72e", + "239401dc0dc4c64aa741a75f7324768ef7557889", + "ef21738c40f78fd5834cf5910fc58f45f88cf698", + "566c104c9a494ae385ab253e3c69864f69d53b52", + "5cf7ac83f2b63f2c090757bcf08df4bd2dd63b81", + "c62c910264658709e9bf0e769e011e7944c45c90", + "b730f01e47151ccfaf568285acf21374efc53f88", + "0822cf1f041b2e572a038667474e79acab264ebf", + "566a2ede36a6493010ea42a7df49916739e00c9d", + "b1a3b19700b8738b4510eecf78a35ff38406df22", + "6e9a8a34ab5b7cdc12ea52d94e3462225af2c32c", + "a06b6d30e2b31dc600f622ab15afe5e2929581a7", + "52dfa36d756a46cf11e6537463add3bdd9bb4011" + ], + "pdfUrls": [ + "http://eprints.eemcs.utwente.nl/26840/01/Pantic_Robust_Statistical_Face_Frontalization.pdf", + "http://doi.ieeecomputersociety.org/10.1109/ICCV.2015.441", + "http://openaccess.thecvf.com/content_iccv_2015/papers/Sagonas_Robust_Statistical_Face_ICCV_2015_paper.pdf", + "http://ibug.doc.ic.ac.uk/media/uploads/documents/robust_frontalization.pdf", + "http://www.cv-foundation.org/openaccess/content_iccv_2015/papers/Sagonas_Robust_Statistical_Face_ICCV_2015_paper.pdf", + "https://ibug.doc.ic.ac.uk/media/uploads/documents/robust_frontalization.pdf" + ], + "title": "Robust Statistical Face Frontalization", + "doi": "10.1109/ICCV.2015.441", + "sources": [ + "DBLP" + ], + "doiUrl": "https://doi.org/10.1109/ICCV.2015.441", + "venue": "2015 IEEE International Conference on Computer Vision (ICCV)" +}
\ No newline at end of file diff --git a/scraper/samples/s2-paper-detail.json b/scraper/samples/s2-paper-detail.json new file mode 100644 index 00000000..5273fd54 --- /dev/null +++ b/scraper/samples/s2-paper-detail.json @@ -0,0 +1,3114 @@ +{ + "responseType": "PAPER_DETAIL", + "paper": { + "id": "e4754afaa15b1b53e70743880484b8d0736990ff", + "title": { + "text": "300 Faces In-The-Wild Challenge: database and results", + "fragments": [] + }, + "slug": "300-Faces-In-The-Wild-Challenge:-database-and-Sagonas-Antonakos", + "paperAbstract": { + "text": "Computer Vision has recently witnessed great research advance towards automatic facial points detection. Numerous methodologies have been proposed during the last few years that achieve accurate and efficient performance. However, fair comparison between these methodologies is infeasible mainly due to two issues. (a) Most existing databases, captured under both constrained and unconstrained (in-the-wild) conditions have been annotated using different mark-ups and, in most cases, the accuracy of the annotations is low. (b) Most published works report experimental results using different training/testing sets, different error metrics and, of course, landmark points with semantically different locations. In this paper, we aim to overcome the aforementioned problems by (a) proposing a semi-automatic annotation technique that was employed to re-annotate most existing facial databases under a unified protocol, and (b) presenting the 300 Faces In-The-Wild Challenge (300-W), the first facial landmark localization challenge that was organized twice, in 2013 and 2015. To the best of our knowledge, this is the first effort towards a unified annotation scheme of massive databases and a fair experimental comparison of existing facial landmark localization systems. The images and annotations of the new testing database that was used in the 300-W challenge are available from http://ibug.doc.ic.ac.uk/resources/facial-point-annotations/.", + "fragments": [] + }, + "authors": [ + [ + { + "name": "Christos Sagonas", + "ids": [ + "3320415" + ], + "slug": "Christos-Sagonas" + }, + { + "text": "Christos Sagonas", + "fragments": [] + } + ], + [ + { + "name": "Epameinondas Antonakos", + "ids": [ + "2788012" + ], + "slug": "Epameinondas-Antonakos" + }, + { + "text": "Epameinondas Antonakos", + "fragments": [] + } + ], + [ + { + "name": "Georgios Tzimiropoulos", + "ids": [ + "2610880" + ], + "slug": "Georgios-Tzimiropoulos" + }, + { + "text": "Georgios Tzimiropoulos", + "fragments": [] + } + ], + [ + { + "name": "Stefanos P. Zafeiriou", + "ids": [ + "1776444" + ], + "slug": "Stefanos-P.-Zafeiriou" + }, + { + "text": "Stefanos P. Zafeiriou", + "fragments": [] + } + ], + [ + { + "name": "Maja Pantic", + "ids": [ + "1694605" + ], + "slug": "Maja-Pantic" + }, + { + "text": "Maja Pantic", + "fragments": [] + } + ] + ], + "structuredAuthors": [ + { + "firstName": "Christos", + "middleNames": [], + "lastName": "Sagonas" + }, + { + "firstName": "Epameinondas", + "middleNames": [], + "lastName": "Antonakos" + }, + { + "firstName": "Georgios", + "middleNames": [], + "lastName": "Tzimiropoulos" + }, + { + "firstName": "Stefanos", + "middleNames": [ + "P." + ], + "lastName": "Zafeiriou" + }, + { + "firstName": "Maja", + "middleNames": [], + "lastName": "Pantic" + } + ], + "year": { + "text": "2016", + "fragments": [] + }, + "venue": { + "text": "Image Vision Comput.", + "fragments": [] + }, + "citationContexts": [], + "citationStats": { + "citedByBuckets": [ + { + "startKey": 2015, + "endKey": 2015, + "count": 2, + "estimate": { + "min": 2.2728870038174662, + "value": 2.7194889455650393, + "max": 3.3020132174096997, + "confidence": 0.9 + } + }, + { + "startKey": 2016, + "endKey": 2016, + "count": 17, + "estimate": { + "min": 19.319539532448463, + "value": 23.115656037302834, + "max": 28.067112347982448, + "confidence": 0.9 + } + }, + { + "startKey": 2017, + "endKey": 2017, + "count": 45, + "estimate": { + "min": 51.139957585892994, + "value": 61.18850127521338, + "max": 74.29529739171824, + "confidence": 0.9 + } + }, + { + "startKey": 2018, + "endKey": 2018, + "count": 32, + "estimate": { + "min": 43.63943047329535, + "value": 52.21418775484875, + "max": 63.39865377426623, + "confidence": 0.9 + } + } + ], + "keyCitedByBuckets": [ + { + "startKey": 2016, + "endKey": 2016, + "count": 2 + }, + { + "startKey": 2017, + "endKey": 2017, + "count": 6 + }, + { + "startKey": 2018, + "endKey": 2018, + "count": 2 + } + ], + "numCitations": 103, + "estNumCitations": { + "min": 117.05368069659951, + "value": 140.0536806965995, + "max": 170.0536806965995, + "confidence": 0.9 + }, + "numReferences": 49, + "numKeyCitations": 11, + "numKeyReferences": 19, + "numViewableReferences": 49, + "keyCitationRate": 0.10679611650485436, + "estCitationVelocity": { + "estimate": { + "min": 42.65391605341533, + "value": 47.439973828190126, + "max": 53.68265788224421, + "confidence": 0.9 + }, + "estCitationsByRange": [ + { + "value": { + "min": 19.319539532448463, + "value": 23.115656037302834, + "max": 28.067112347982448, + "confidence": 0.9 + }, + "start": [ + 2016, + 1, + 1 + ], + "end": [ + 2016, + 12, + 31 + ] + }, + { + "value": { + "min": 51.139957585892994, + "value": 61.18850127521338, + "max": 74.29529739171824, + "confidence": 0.9 + }, + "start": [ + 2017, + 1, + 1 + ], + "end": [ + 2017, + 12, + 31 + ] + }, + { + "value": { + "min": 48.488256081439275, + "value": 58.01576417205417, + "max": 70.44294863807359, + "confidence": 0.9 + }, + "start": [ + 2018, + 1, + 1 + ], + "end": [ + 2018, + 12, + 31 + ] + } + ] + }, + "estCitationAcceleration": { + "estimate": { + "min": -0.026615969581749135, + "value": -0.026615969581749065, + "max": -0.026615969581749045, + "confidence": 0.9 + }, + "estCitationsByRange": [ + { + "value": { + "min": 51.139957585892994, + "value": 61.18850127521338, + "max": 74.29529739171824, + "confidence": 0.9 + }, + "start": [ + 2017, + 1, + 1 + ], + "end": [ + 2017, + 12, + 31 + ] + }, + { + "value": { + "min": 48.488256081439275, + "value": 58.01576417205417, + "max": 70.44294863807359, + "confidence": 0.9 + }, + "start": [ + 2018, + 1, + 1 + ], + "end": [ + 2018, + 12, + 31 + ] + } + ] + } + }, + "sources": [ + "DBLP", + "Grobid", + "Anansi", + "Crawler", + "ScienceParse", + "SPv2" + ], + "journal": { + "name": "Image Vision Comput.", + "volume": "47", + "pages": "3-18" + }, + "socialLinks": [], + "presentationUrls": [], + "doiInfo": { + "doi": "10.1016/j.imavis.2016.01.002", + "doiUrl": "http://doi.org/10.1016/j.imavis.2016.01.002" + }, + "links": [ + { + "url": "http://doi.org/10.1016/j.imavis.2016.01.002", + "linkType": "doi" + } + ], + "primaryPaperLink": { + "url": "http://doi.org/10.1016/j.imavis.2016.01.002", + "linkType": "doi" + }, + "alternatePaperLinks": [ + { + "url": "http://ibug.doc.ic.ac.uk/media/uploads/documents/1-s2.0-s0262885616000147-main.pdf", + "linkType": "crawler" + }, + { + "url": "https://doi.org/10.1016/j.imavis.2016.01.002", + "linkType": "dblp" + }, + { + "url": "https://spiral.imperial.ac.uk:8443/bitstream/10044/1/32322/2/300w.pdf", + "linkType": "anansi" + } + ], + "entities": [ + { + "id": "5332", + "name": "Computer vision", + "slug": "Computer-vision" + }, + { + "id": "55238", + "name": "Emoticon", + "slug": "Emoticon" + }, + { + "id": "539730", + "name": "Landmark point", + "slug": "Landmark-point" + }, + { + "id": "76540", + "name": "Semiconductor industry", + "slug": "Semiconductor-industry" + }, + { + "id": "468", + "name": "Malignant Fibrous Histiocytoma", + "slug": "Malignant-Fibrous-Histiocytoma" + } + ], + "entityRelations": [], + "blogs": [], + "videos": [], + "githubReferences": [], + "faqs": [], + "scorecardStats": [ + { + "typeKey": "highly_influential", + "score": 50, + "keyCitationCount": 11 + }, + { + "typeKey": "cited_by", + "citationRankPercent": 99, + "citationCount": 140, + "score": 10 + } + ], + "hasPdf": false + }, + "citingPapers": { + "citationType": "citingPapers", + "citations": [ + { + "id": "6742c0a26315d7354ab6b1fa62a5fffaea06da14", + "title": { + "text": "What does 2 D geometric information really tell us about 3 D face shape ?", + "fragments": [] + }, + "slug": "What-does-2-D-geometric-information-really-tell-us-Bas-Smith", + "venue": { + "text": "", + "fragments": [] + }, + "year": 2017, + "authors": [ + [ + { + "name": "Anil Bas", + "ids": [ + "39180407" + ], + "slug": "Anil-Bas" + }, + { + "text": "Anil Bas", + "fragments": [] + } + ], + [ + { + "name": "William A. P. Smith", + "ids": [ + "1687021" + ], + "slug": "William-A.-P.-Smith" + }, + { + "text": "William A. P. Smith", + "fragments": [] + } + ] + ], + "citationContexts": [ + { + "text": "For this reason, there has been sustained interest in building feature detectors capable of accurately labelling face landmarks in uncontrolled images (Sagonas et al. 2016).", + "fragments": [ + { + "start": 152, + "end": 171 + } + ] + }, + { + "text": "Despite the small sample size, we find a pair of faces whose mean landmark error is 2.48% (i.e. they are within the expected accuracy of a landmark detector (Sagonas et al. 2016)) when they are viewed at 61cm and 488cm respectively (second and fourth image in the figure).", + "fragments": [ + { + "start": 158, + "end": 177 + } + ] + }, + { + "text": "Landmark detection on highly uncontrolled face images is now a mature research field with benchmarks (Sagonas et al. 2016) providing an indication of likely accuracy.", + "fragments": [ + { + "start": 102, + "end": 121 + } + ] + }, + { + "text": "they are within the expected accuracy of a landmark detector (Sagonas et al. 2016)) when they are viewed at 61cm and 488cm respectively (second and fourth image in the figure).", + "fragments": [ + { + "start": 61, + "end": 82 + } + ] + }, + { + "text": "Similarly, the 300 faces in the wild challenge (Sagonas et al. 2016) found that even the best methods did not obtain better than 5% accuracy for more than 50% of the landmarks.", + "fragments": [ + { + "start": 48, + "end": 67 + } + ] + }, + { + "text": "For example, state-of-the-art automatic face landmarking provides a mean landmark error under 4.5% of interocular distance for only 50% of images (according to the second conduct of the 300 Faces in the Wild challenge (Sagonas et al. 2016)).", + "fragments": [ + { + "start": 219, + "end": 238 + } + ] + }, + { + "text": "5% of interocular distance for only 50% of images (according to the second conduct of the 300 Faces in the Wild challenge (Sagonas et al. 2016)).", + "fragments": [ + { + "start": 122, + "end": 143 + } + ] + } + ], + "isKey": true + }, + { + "id": "014e3d0fa5248e6f4634dc237e2398160294edce", + "title": { + "text": "What does 2D geometric information really tell us about 3D face shape?", + "fragments": [] + }, + "slug": "What-does-2D-geometric-information-really-tell-us-Bas-Smith", + "venue": { + "text": "ArXiv", + "fragments": [] + }, + "year": 2017, + "authors": [ + [ + { + "name": "Anil Bas", + "ids": [ + "39180407" + ], + "slug": "Anil-Bas" + }, + { + "text": "Anil Bas", + "fragments": [] + } + ], + [ + { + "name": "William A. P. Smith", + "ids": [ + "1687021" + ], + "slug": "William-A.-P.-Smith" + }, + { + "text": "William A. P. Smith", + "fragments": [] + } + ] + ], + "citationContexts": [ + { + "text": "For this reason, there has been sustained interest in building feature detectors capable of accurately labelling face landmarks in uncontrolled images (Sagonas et al. 2016).", + "fragments": [ + { + "start": 152, + "end": 171 + } + ] + }, + { + "text": "Despite the small sample size, we find a pair of faces whose mean landmark error is 2.48% (i.e. they are within the expected accuracy of a landmark detector (Sagonas et al. 2016)) when they are viewed at 61cm and 488cm respectively (second and fourth image in the figure).", + "fragments": [ + { + "start": 158, + "end": 177 + } + ] + }, + { + "text": "Landmark detection on highly uncontrolled face images is now a mature research field with benchmarks (Sagonas et al. 2016) providing an indication of likely accuracy.", + "fragments": [ + { + "start": 102, + "end": 121 + } + ] + }, + { + "text": "they are within the expected accuracy of a landmark detector (Sagonas et al. 2016)) when they are viewed at 61cm and 488cm respectively (second and fourth image in the figure).", + "fragments": [ + { + "start": 61, + "end": 82 + } + ] + }, + { + "text": "Similarly, the 300 faces in the wild challenge (Sagonas et al. 2016) found that even the best methods did not obtain better than 5% accuracy for more than 50% of the landmarks.", + "fragments": [ + { + "start": 48, + "end": 67 + } + ] + }, + { + "text": "For example, state-of-the-art automatic face landmarking provides a mean landmark error under 4.5% of interocular distance for only 50% of images (according to the second conduct of the 300 Faces in the Wild challenge (Sagonas et al. 2016)).", + "fragments": [ + { + "start": 219, + "end": 238 + } + ] + }, + { + "text": "5% of interocular distance for only 50% of images (according to the second conduct of the 300 Faces in the Wild challenge (Sagonas et al. 2016)).", + "fragments": [ + { + "start": 122, + "end": 143 + } + ] + } + ], + "isKey": true + }, + { + "id": "8f772d9ce324b2ef5857d6e0b2a420bc93961196", + "title": { + "text": "Facial Landmark Point Localization using Coarse-to-Fine Deep Recurrent Neural Network", + "fragments": [] + }, + "slug": "Facial-Landmark-Point-Localization-using-Deep-Mahpod-Das", + "venue": { + "text": "ArXiv", + "fragments": [] + }, + "year": 2018, + "authors": [ + [ + { + "name": "Shahar Mahpod", + "ids": [ + "2748312" + ], + "slug": "Shahar-Mahpod" + }, + { + "text": "Shahar Mahpod", + "fragments": [] + } + ], + [ + { + "name": "Rig Das", + "ids": [ + "3001038" + ], + "slug": "Rig-Das" + }, + { + "text": "Rig Das", + "fragments": [] + } + ], + [ + { + "name": "Emanuele Maiorana", + "ids": [ + "1767715" + ], + "slug": "Emanuele-Maiorana" + }, + { + "text": "Emanuele Maiorana", + "fragments": [] + } + ], + [ + { + "name": "Yosi Keller", + "ids": [ + "1926432" + ], + "slug": "Yosi-Keller" + }, + { + "text": "Yosi Keller", + "fragments": [] + } + ], + [ + { + "name": "Patrizio Campisi", + "ids": [ + "1682433" + ], + "slug": "Patrizio-Campisi" + }, + { + "text": "Patrizio Campisi", + "fragments": [] + } + ] + ], + "citationContexts": [ + { + "text": "In order to evaluate the performance of our proposed CFDRNN framework we perform some exhaustive experiments on the data released for the 300-W competition [23].", + "fragments": [ + { + "start": 156, + "end": 160 + } + ] + }, + { + "text": "300-W, MENPO 300-W private test set [23] 3.", + "fragments": [ + { + "start": 36, + "end": 40 + } + ] + }, + { + "text": "The authors have also tested their proposed framework’s performance against the 300-W private test set [23], which consists of 300 indoor and 300 outdoor images.", + "fragments": [ + { + "start": 103, + "end": 107 + } + ] + }, + { + "text": "This dataset [23] contains 135 complicated facial images with different facial expression, poses, illumination, and multiple faces in a single image.", + "fragments": [ + { + "start": 13, + "end": 17 + } + ] + }, + { + "text": "bug [21], [22], 300-W [23], and Menpo [19], [24] leaves a significant scope for further improvement.", + "fragments": [ + { + "start": 22, + "end": 26 + } + ] + } + ], + "isKey": true + }, + { + "id": "7789a5d87884f8bafec8a82085292e87d4e2866f", + "title": { + "text": "A Unified Tensor-based Active Appearance Face Model", + "fragments": [] + }, + "slug": "A-Unified-Tensor-based-Active-Appearance-Face-Model-Feng-Kittler", + "venue": { + "text": "ArXiv", + "fragments": [] + }, + "year": 2016, + "authors": [ + [ + { + "name": "Zhen-Hua Feng", + "ids": [ + "2976854" + ], + "slug": "Zhen-Hua-Feng" + }, + { + "text": "Zhen-Hua Feng", + "fragments": [] + } + ], + [ + { + "name": "Josef Kittler", + "ids": [ + "1748684" + ], + "slug": "Josef-Kittler" + }, + { + "text": "Josef Kittler", + "fragments": [] + } + ], + [ + { + "name": "William J. Christmas", + "ids": [ + "1942955" + ], + "slug": "William-J.-Christmas" + }, + { + "text": "William J. Christmas", + "fragments": [] + } + ], + [ + { + "name": "Xiaojun Wu", + "ids": [ + "50171811" + ], + "slug": "Xiaojun-Wu" + }, + { + "text": "Xiaojun Wu", + "fragments": [] + } + ] + ], + "citationContexts": [ + { + "text": "W face dataset (Sagonas et al., 2016).", + "fragments": [ + { + "start": 16, + "end": 36 + } + ] + }, + { + "text": "The 300-W dataset has been widely used for benchmarking a facial landmark detection algorithm (Sagonas et al., 2016).", + "fragments": [ + { + "start": 94, + "end": 116 + } + ] + }, + { + "text": "Then we demonstrate the capacity of the proposed UT-AAM to synthesise a large number of virtual faces and examine how these synthesised faces can improve the training of a facial landmark detector, using the 300-W face dataset (Sagonas et al., 2016).", + "fragments": [ + { + "start": 227, + "end": 249 + } + ] + }, + { + "text": "W dataset has been widely used for benchmarking a facial landmark detection algorithm (Sagonas et al., 2016).", + "fragments": [ + { + "start": 87, + "end": 107 + } + ] + }, + { + "text": "7 A comparison of the cumulative error distribution curves of SDM and CCR, as well as a set of state-of-the-art methods from Baltrusaitis, Hasan, Jaiswal, Miborrow, Yan and Zhou (Sagonas et al., 2016), on the 300-W face dataset: (a) results on the 300 outdoor face images; (b) results on the 300 indoor faces.", + "fragments": [ + { + "start": 178, + "end": 200 + } + ] + } + ], + "isKey": true + }, + { + "id": "0f21a39fa4c0a19c4a5b4733579e393cb1d04f71", + "title": { + "text": "Evaluation of optimization components of a 3D to 2D landmark fitting algorithm for head pose estimation", + "fragments": [] + }, + "slug": "Evaluation-of-optimization-components-of-a-3D-to-2D-Haan", + "venue": { + "text": "", + "fragments": [] + }, + "year": 2018, + "authors": [ + [ + { + "name": "Tim de Haan", + "ids": [], + "slug": "Tim-de-Haan" + }, + { + "text": "Tim de Haan", + "fragments": [] + } + ] + ], + "citationContexts": [ + { + "text": "Landmark Localization: Landmark localization has been extensively researched [1], [17], [18], which has led to the development of Dlib [7], a machine learning toolkit that has its usage in many complex applications, such as image processing.", + "fragments": [ + { + "start": 77, + "end": 80 + } + ] + }, + { + "text": "This thesis uses the new 300-W dataset [1, 17, 18], this set consists of 600 photographs that can be used to project the 3DMM onto the landmarks detected by dlib.", + "fragments": [ + { + "start": 39, + "end": 50 + } + ] + }, + { + "text": "histogram of oriented gradients combined with a linear classifier [7], trained on the iBUG 300-W dataset from [1], the landmark detection is an implementation of [19].", + "fragments": [ + { + "start": 110, + "end": 113 + } + ] + }, + { + "text": "Table 8 shows how Dlib compares with other landmarks detection algorithms, which are the winners of the two contests that were organized by [1].", + "fragments": [ + { + "start": 140, + "end": 143 + } + ] + }, + { + "text": "Additionally, the Oracle curve is added, which is the minimum error that can be achieved [1].", + "fragments": [ + { + "start": 89, + "end": 92 + } + ] + } + ], + "isKey": true + }, + { + "id": "1885acea0d24e7b953485f78ec57b2f04e946eaf", + "title": { + "text": "Combining Local and Global Features for 3D Face Tracking", + "fragments": [] + }, + "slug": "Combining-Local-and-Global-Features-for-3D-Face-Xiong-Li", + "venue": { + "text": "2017 IEEE International Conference on Computer Vision Workshops (ICCVW)", + "fragments": [] + }, + "year": 2017, + "authors": [ + [ + { + "name": "Pengfei Xiong", + "ids": [ + "40448951" + ], + "slug": "Pengfei-Xiong" + }, + { + "text": "Pengfei Xiong", + "fragments": [] + } + ], + [ + { + "name": "Guoqing Li", + "ids": [ + "1775836" + ], + "slug": "Guoqing-Li" + }, + { + "text": "Guoqing Li", + "fragments": [] + } + ], + [ + { + "name": "Yuhang Sun", + "ids": [ + "48186289" + ], + "slug": "Yuhang-Sun" + }, + { + "text": "Yuhang Sun", + "fragments": [] + } + ] + ], + "citationContexts": [ + { + "text": "In the past few decades, extensive studies [39, 40, 31, 11, 26, 41, 33, 43] have been proposed and singnificant improvements have been achieved, especially since a comprehensice benchmark [32] was made public, and deep convolution neural networks [33, 43] were applied in face shape regression.", + "fragments": [ + { + "start": 188, + "end": 192 + } + ] + }, + { + "text": "Images from AFLW [28], FDDB[22], 300W, 300W-Test [32] are collected and fitted with 3D facial morphable model [4] to generate 84 3D point annotations.", + "fragments": [ + { + "start": 49, + "end": 53 + } + ] + }, + { + "text": "300W-LP: 300W-LP(300W across Large Poses) [45] contains 61225 samples (1786 from IBUG, 5207 from AFW, and 16556 from LFPW) sysnthetical generated from 300-W [32].", + "fragments": [ + { + "start": 157, + "end": 161 + } + ] + }, + { + "text": "[45] sysnthetically generated a series of datasets based on 300W [32], which gradually become the benchmark of 3D shape regression methods.", + "fragments": [ + { + "start": 65, + "end": 69 + } + ] + } + ], + "isKey": true + }, + { + "id": "0a34fe39e9938ae8c813a81ae6d2d3a325600e5c", + "title": { + "text": "FacePoseNet: Making a Case for Landmark-Free Face Alignment", + "fragments": [] + }, + "slug": "FacePoseNet:-Making-a-Case-for-Landmark-Free-Face-Chang-Tran", + "venue": { + "text": "2017 IEEE International Conference on Computer Vision Workshops (ICCVW)", + "fragments": [] + }, + "year": 2017, + "authors": [ + [ + { + "name": "Feng-Ju Chang", + "ids": [ + "1752756" + ], + "slug": "Feng-Ju-Chang" + }, + { + "text": "Feng-Ju Chang", + "fragments": [] + } + ], + [ + { + "name": "Anh Tuan Tran", + "ids": [ + "46634688" + ], + "slug": "Anh-Tuan-Tran" + }, + { + "text": "Anh Tuan Tran", + "fragments": [] + } + ], + [ + { + "name": "Tal Hassner", + "ids": [ + "1756099" + ], + "slug": "Tal-Hassner" + }, + { + "text": "Tal Hassner", + "fragments": [] + } + ], + [ + { + "name": "Iacopo Masi", + "ids": [ + "11269472" + ], + "slug": "Iacopo-Masi" + }, + { + "text": "Iacopo Masi", + "fragments": [] + } + ], + [ + { + "name": "Ramakant Nevatia", + "ids": [ + "1694832" + ], + "slug": "Ramakant-Nevatia" + }, + { + "text": "Ramakant Nevatia", + "fragments": [] + } + ], + [ + { + "name": "Gérard G. Medioni", + "ids": [ + "3463966" + ], + "slug": "Gérard-G.-Medioni" + }, + { + "text": "Gérard G. Medioni", + "fragments": [] + } + ] + ], + "citationContexts": [ + { + "text": "Facial landmark detection is big business, as reflected by the numerous citation to relevant papers, the many facial landmark detection benchmarks [5, 23, 26, 40, 53], and popular international events dedicated to this problem.", + "fragments": [ + { + "start": 147, + "end": 166 + } + ] + }, + { + "text": "We evaluate performance on the 300W data set [40], the most challenging benchmark of its kind [45], using 68 landmarks.", + "fragments": [ + { + "start": 45, + "end": 49 + } + ] + }, + { + "text": "Landmarks detected in 300W [40] images by projecting an unmodified 3D face shape,", + "fragments": [ + { + "start": 27, + "end": 31 + } + ] + }, + { + "text": "Finally, (3), we test our FPN extensively and report that better landmark detection accuracy on the widely used 300W benchmark [40] does not imply better alignment and recognition on the highly challenging IJB-A [22] and IJB-B benchmarks [44].", + "fragments": [ + { + "start": 127, + "end": 131 + } + ] + } + ], + "isKey": true + }, + { + "id": "cf736f596bf881ca97ec4b29776baaa493b9d50e", + "title": { + "text": "Low Dimensional Deep Features for facial landmark alignment", + "fragments": [] + }, + "slug": "Low-Dimensional-Deep-Features-for-facial-landmark-Jalan-Mynepalli", + "venue": { + "text": "2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)", + "fragments": [] + }, + "year": 2017, + "authors": [ + [ + { + "name": "Ankit Jalan", + "ids": [ + "18090725" + ], + "slug": "Ankit-Jalan" + }, + { + "text": "Ankit Jalan", + "fragments": [] + } + ], + [ + { + "name": "Siva Chaitanya Mynepalli", + "ids": [ + "18091255" + ], + "slug": "Siva-Chaitanya-Mynepalli" + }, + { + "text": "Siva Chaitanya Mynepalli", + "fragments": [] + } + ], + [ + { + "name": "Viswanath Veera", + "ids": [ + "18178676" + ], + "slug": "Viswanath-Veera" + }, + { + "text": "Viswanath Veera", + "fragments": [] + } + ], + [ + { + "name": "Shankar M. Venkatesan", + "ids": [ + "3210146" + ], + "slug": "Shankar-M.-Venkatesan" + }, + { + "text": "Shankar M. Venkatesan", + "fragments": [] + } + ] + ], + "citationContexts": [ + { + "text": "300-W dataset [28, 31]: This dataset is created from", + "fragments": [ + { + "start": 14, + "end": 22 + } + ] + }, + { + "text": "Experimental analysis has demonstrated that LDFFA outperforms other state-of-the-art algorithms on Helen, LFPW benchmark datasets while giving comparable performance on subsets of 300-W database [26, 27, 28, 31] with 68 fiducial landmarks.", + "fragments": [ + { + "start": 195, + "end": 211 + } + ] + }, + { + "text": "We follow the method of [31] where the average L2 distance of the estimated landmark position from the ground truth is normalized by the standard definition of inter-ocular distance (douter,) to give the error (Eqn.", + "fragments": [ + { + "start": 24, + "end": 28 + } + ] + }, + { + "text": "The bounding boxes for 300-W images were provided by [28, 31] using their in-house face detector.", + "fragments": [ + { + "start": 53, + "end": 61 + } + ] + }, + { + "text": "3 (b & c) illustrate the Normalized Mean Error using definition from [31].", + "fragments": [ + { + "start": 69, + "end": 73 + } + ] + } + ], + "isKey": true + }, + { + "id": "7fcfd72ba6bc14bbb90b31fe14c2c77a8b220ab2", + "title": { + "text": "Robust FEC-CNN: A High Accuracy Facial Landmark Detection System", + "fragments": [] + }, + "slug": "Robust-FEC-CNN:-A-High-Accuracy-Facial-Landmark-He-Zhang", + "venue": { + "text": "2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)", + "fragments": [] + }, + "year": 2017, + "authors": [ + [ + { + "name": "Zhenliang He", + "ids": [ + "3469114" + ], + "slug": "Zhenliang-He" + }, + { + "text": "Zhenliang He", + "fragments": [] + } + ], + [ + { + "name": "Jie Zhang", + "ids": [ + "49050482" + ], + "slug": "Jie-Zhang" + }, + { + "text": "Jie Zhang", + "fragments": [] + } + ], + [ + { + "name": "Meina Kan", + "ids": [ + "1693589" + ], + "slug": "Meina-Kan" + }, + { + "text": "Meina Kan", + "fragments": [] + } + ], + [ + { + "name": "Shiguang Shan", + "ids": [ + "1685914" + ], + "slug": "Shiguang-Shan" + }, + { + "text": "Shiguang Shan", + "fragments": [] + } + ], + [ + { + "name": "Xilin Chen", + "ids": [ + "1710220" + ], + "slug": "Xilin-Chen" + }, + { + "text": "Xilin Chen", + "fragments": [] + } + ] + ], + "citationContexts": [ + { + "text": "Impressive works [5, 4, 6, 3, 23, 14, 20, 27, 21, 22, 9] and benchmarks [11, 29, 2, 16, 17, 10, 26] were proposed to tackle this task in the the past few decades.", + "fragments": [ + { + "start": 72, + "end": 99 + } + ] + }, + { + "text": "LFPW, HELEN, AFW, 300W Competition and Menpo 68 point subset are used as training set while IBUG is used as testing set.", + "fragments": [ + { + "start": 23, + "end": 34 + } + ] + }, + { + "text": "We employ 300W [16, 18], 300W Competition [16, 17] and Menpo dataset [26] for evaluating RFC.", + "fragments": [ + { + "start": 15, + "end": 23 + } + ] + }, + { + "text": "The 300W Competition dataset consists of indoor and outdoor subset.", + "fragments": [ + { + "start": 9, + "end": 20 + } + ] + } + ], + "isKey": true + }, + { + "id": "4ac3cd8b6c50f7a26f27eefc64855134932b39be", + "title": { + "text": "Robust Facial Landmark Detection via a Fully-Convolutional Local-Global Context Network", + "fragments": [] + }, + "slug": "Robust-Facial-Landmark-Detection-via-a-Local-Global-Merget-Rock", + "venue": { + "text": "", + "fragments": [] + }, + "authors": [ + [ + { + "name": "Daniel Merget", + "ids": [ + "3044182" + ], + "slug": "Daniel-Merget" + }, + { + "text": "Daniel Merget", + "fragments": [] + } + ], + [ + { + "name": "Matthias Rock", + "ids": [ + "28096417" + ], + "slug": "Matthias-Rock" + }, + { + "text": "Matthias Rock", + "fragments": [] + } + ], + [ + { + "name": "Gerhard Rigoll", + "ids": [ + "46343645" + ], + "slug": "Gerhard-Rigoll" + }, + { + "text": "Gerhard Rigoll", + "fragments": [] + } + ] + ], + "citationContexts": [ + { + "text": "Some qualitative results on the 300-W challenge data set [28] are presented in Figure 5.", + "fragments": [ + { + "start": 57, + "end": 61 + } + ] + }, + { + "text": "Figure 4 illustrates the results for inter-ocular distance (IOD) normalized mean absolute error (MAE) on the 300-W benchmark [28] and on a cross-data set test with Menpo [41].", + "fragments": [ + { + "start": 125, + "end": 129 + } + ] + }, + { + "text": "The overall prediction performance on 300-W [28] was worse using cross entropy.", + "fragments": [ + { + "start": 44, + "end": 48 + } + ] + }, + { + "text": "Our approach beats the state of the art on 300-W [28] and on a cross-data set test with Menpo [41].", + "fragments": [ + { + "start": 49, + "end": 53 + } + ] + }, + { + "text": "From left to right: 300-W [28], iBUG [29], LFPW [5] + HELEN [20], Menpo frontal train set [41].", + "fragments": [ + { + "start": 26, + "end": 30 + } + ] + }, + { + "text": "• We demonstrate the effectiveness of our approach by improving upon the state of the art on 300-W [28] and a cross-data set test on Menpo [41].", + "fragments": [ + { + "start": 99, + "end": 103 + } + ] + } + ], + "isKey": true + } + ], + "requestedPageSize": 10, + "pageNumber": 1, + "totalPages": 11, + "sort": "is-influential" + }, + "citedPapers": { + "citationType": "citedPapers", + "citations": [ + { + "id": "0a6d344112b5af7d1abbd712f83c0d70105211d0", + "title": { + "text": "Constrained Local Neural Fields for Robust Facial Landmark Detection in the Wild", + "fragments": [] + }, + "slug": "Constrained-Local-Neural-Fields-for-Robust-Facial-Baltrusaitis-Robinson", + "venue": { + "text": "2013 IEEE International Conference on Computer Vision Workshops", + "fragments": [] + }, + "year": 2013, + "authors": [ + [ + { + "name": "Tadas Baltrusaitis", + "ids": [ + "1756344" + ], + "slug": "Tadas-Baltrusaitis" + }, + { + "text": "Tadas Baltrusaitis", + "fragments": [] + } + ], + [ + { + "name": "Peter Robinson", + "ids": [ + "39626495" + ], + "slug": "Peter-Robinson" + }, + { + "text": "Peter Robinson", + "fragments": [] + } + ], + [ + { + "name": "Louis-Philippe Morency", + "ids": [ + "49933077" + ], + "slug": "Louis-Philippe-Morency" + }, + { + "text": "Louis-Philippe Morency", + "fragments": [] + } + ] + ], + "citationContexts": [ + { + "text": "[40] propose a probabilistic patch expert technique that learns non-linear and spatial relationships between the pixels and the probability of a landmark being aligned.", + "fragments": [ + { + "start": 0, + "end": 4 + } + ] + }, + { + "text": "[40]", + "fragments": [ + { + "start": 0, + "end": 4 + } + ] + }, + { + "text": "[40] T.", + "fragments": [ + { + "start": 0, + "end": 4 + } + ] + }, + { + "text": "[40] 0.", + "fragments": [ + { + "start": 0, + "end": 4 + } + ] + } + ], + "isKey": true + }, + { + "id": "1a8ccc23ed73db64748e31c61c69fe23c48a2bb1", + "title": { + "text": "Extensive Facial Landmark Localization with Coarse-to-Fine Convolutional Network Cascade", + "fragments": [] + }, + "slug": "Extensive-Facial-Landmark-Localization-with-Network-Zhou-Fan", + "venue": { + "text": "2013 IEEE International Conference on Computer Vision Workshops", + "fragments": [] + }, + "year": 2013, + "authors": [ + [ + { + "name": "Erjin Zhou", + "ids": [ + "1848243" + ], + "slug": "Erjin-Zhou" + }, + { + "text": "Erjin Zhou", + "fragments": [] + } + ], + [ + { + "name": "Haoqiang Fan", + "ids": [ + "1934546" + ], + "slug": "Haoqiang-Fan" + }, + { + "text": "Haoqiang Fan", + "fragments": [] + } + ], + [ + { + "name": "Zhimin Cao", + "ids": [ + "2695115" + ], + "slug": "Zhimin-Cao" + }, + { + "text": "Zhimin Cao", + "fragments": [] + } + ], + [ + { + "name": "Yuning Jiang", + "ids": [ + "1691963" + ], + "slug": "Yuning-Jiang" + }, + { + "text": "Yuning Jiang", + "fragments": [] + } + ], + [ + { + "name": "Qi Yin", + "ids": [ + "2274228" + ], + "slug": "Qi-Yin" + }, + { + "text": "Qi Yin", + "fragments": [] + } + ] + ], + "citationContexts": [ + { + "text": "[47] 0% 2.", + "fragments": [ + { + "start": 0, + "end": 4 + } + ] + }, + { + "text": "[47] 0.", + "fragments": [ + { + "start": 0, + "end": 4 + } + ] + }, + { + "text": "[47] propose a four-level convolutional network cascade, where each level is trained to locally refine the outputs of the previous network levels.", + "fragments": [ + { + "start": 0, + "end": 4 + } + ] + }, + { + "text": "[47]", + "fragments": [ + { + "start": 0, + "end": 4 + } + ] + }, + { + "text": "[47] (convolutional network framework), can continually achieve better results with continuous rise in the amount of training data.", + "fragments": [ + { + "start": 0, + "end": 4 + } + ] + }, + { + "text": "[47] E.", + "fragments": [ + { + "start": 0, + "end": 4 + } + ] + }, + { + "text": "[47] from Megvii company.", + "fragments": [ + { + "start": 0, + "end": 4 + } + ] + }, + { + "text": "[47]) and second (J.", + "fragments": [ + { + "start": 0, + "end": 4 + } + ] + } + ], + "isKey": true + }, + { + "id": "1c1a98df3d0d5e2034ea723994bdc85af45934db", + "title": { + "text": "Guided Unsupervised Learning of Mode Specific Models for Facial Point Detection in the Wild", + "fragments": [] + }, + "slug": "Guided-Unsupervised-Learning-of-Mode-Specific-for-Jaiswal-Almaev", + "venue": { + "text": "2013 IEEE International Conference on Computer Vision Workshops", + "fragments": [] + }, + "year": 2013, + "authors": [ + [ + { + "name": "Shashank Jaiswal", + "ids": [ + "2736086" + ], + "slug": "Shashank-Jaiswal" + }, + { + "text": "Shashank Jaiswal", + "fragments": [] + } + ], + [ + { + "name": "Timur R. Almaev", + "ids": [ + "2449665" + ], + "slug": "Timur-R.-Almaev" + }, + { + "text": "Timur R. Almaev", + "fragments": [] + } + ], + [ + { + "name": "Michel F. Valstar", + "ids": [ + "1795528" + ], + "slug": "Michel-F.-Valstar" + }, + { + "text": "Michel F. Valstar", + "fragments": [] + } + ] + ], + "citationContexts": [ + { + "text": "[41]", + "fragments": [ + { + "start": 0, + "end": 4 + } + ] + }, + { + "text": "[41] 0.", + "fragments": [ + { + "start": 0, + "end": 4 + } + ] + }, + { + "text": "[41] S.", + "fragments": [ + { + "start": 0, + "end": 4 + } + ] + }, + { + "text": "[41] use Local Evidence Aggregated Regression [42], in which local patches provide evidence of the location of the target facial point using Support Vector Regressors.", + "fragments": [ + { + "start": 0, + "end": 4 + } + ] + } + ], + "isKey": true + }, + { + "id": "321c8ba38db118d8b02c0ba209be709e6792a2c7", + "title": { + "text": "Learn to Combine Multiple Hypotheses for Accurate Face Alignment", + "fragments": [] + }, + "slug": "Learn-to-Combine-Multiple-Hypotheses-for-Accurate-Yan-Lei", + "venue": { + "text": "2013 IEEE International Conference on Computer Vision Workshops", + "fragments": [] + }, + "year": 2013, + "authors": [ + [ + { + "name": "Junjie Yan", + "ids": [ + "48270105" + ], + "slug": "Junjie-Yan" + }, + { + "text": "Junjie Yan", + "fragments": [] + } + ], + [ + { + "name": "Zhen Lei", + "ids": [ + "1718623" + ], + "slug": "Zhen-Lei" + }, + { + "text": "Zhen Lei", + "fragments": [] + } + ], + [ + { + "name": "Dong Yi", + "ids": [ + "1716143" + ], + "slug": "Dong-Yi" + }, + { + "text": "Dong Yi", + "fragments": [] + } + ], + [ + { + "name": "Stan Z. Li", + "ids": [ + "34679741" + ], + "slug": "Stan-Z.-Li" + }, + { + "text": "Stan Z. Li", + "fragments": [] + } + ] + ], + "citationContexts": [ + { + "text": "[46]", + "fragments": [ + { + "start": 0, + "end": 4 + } + ] + }, + { + "text": "[46], Zhou et al.", + "fragments": [ + { + "start": 0, + "end": 4 + } + ] + }, + { + "text": "[46] J.", + "fragments": [ + { + "start": 0, + "end": 4 + } + ] + }, + { + "text": "[46] from The National Laboratory of Pattern Recognition at the Institute of Automation of the Chinese Academy of Sciences, and (b) Zhou et al.", + "fragments": [ + { + "start": 0, + "end": 4 + } + ] + }, + { + "text": "[46] (cascade regression framework) and", + "fragments": [ + { + "start": 0, + "end": 4 + } + ] + }, + { + "text": "[46] 0.", + "fragments": [ + { + "start": 0, + "end": 4 + } + ] + }, + { + "text": "[46] employ a cascade regression framework, where a series of regressors are utilized to progressively refine the shape initialized by the face detector.", + "fragments": [ + { + "start": 0, + "end": 4 + } + ] + } + ], + "isKey": true + }, + { + "id": "4f77a37753c03886ca9c9349723ec3bbfe4ee967", + "title": { + "text": "Localizing Facial Keypoints with Global Descriptor Search, Neighbour Alignment and Locally Linear Models", + "fragments": [] + }, + "slug": "Localizing-Facial-Keypoints-with-Global-Descriptor-Hasan-Pal", + "venue": { + "text": "2013 IEEE International Conference on Computer Vision Workshops", + "fragments": [] + }, + "year": 2013, + "authors": [ + [ + { + "name": "Md. Kamrul Hasan", + "ids": [ + "2811524" + ], + "slug": "Md.-Kamrul-Hasan" + }, + { + "text": "Md. Kamrul Hasan", + "fragments": [] + } + ], + [ + { + "name": "Christopher Joseph Pal", + "ids": [ + "1972076" + ], + "slug": "Christopher-Joseph-Pal" + }, + { + "text": "Christopher Joseph Pal", + "fragments": [] + } + ], + [ + { + "name": "Sharon Moalem", + "ids": [ + "9422894" + ], + "slug": "Sharon-Moalem" + }, + { + "text": "Sharon Moalem", + "fragments": [] + } + ] + ], + "citationContexts": [ + { + "text": "[43] M.", + "fragments": [ + { + "start": 0, + "end": 4 + } + ] + }, + { + "text": "[43]", + "fragments": [ + { + "start": 0, + "end": 4 + } + ] + }, + { + "text": "[43] first apply a nearest neighbour search using global descriptors and, then, aim to align local neighbours by dynamically fitting a locally linear model to the global keypoint configurations of the returned neighbours.", + "fragments": [ + { + "start": 0, + "end": 4 + } + ] + }, + { + "text": "[43] 0.", + "fragments": [ + { + "start": 0, + "end": 4 + } + ] + } + ], + "isKey": true + }, + { + "id": "36e8ef2e5d52a78dddf0002e03918b101dcdb326", + "title": { + "text": "Multiview Active Shape Models with SIFT Descriptors for the 300-W Face Landmark Challenge", + "fragments": [] + }, + "slug": "Multiview-Active-Shape-Models-with-SIFT-Descriptors-Milborrow-Bishop", + "venue": { + "text": "2013 IEEE International Conference on Computer Vision Workshops", + "fragments": [] + }, + "year": 2013, + "authors": [ + [ + { + "name": "Stephen Milborrow", + "ids": [ + "2822258" + ], + "slug": "Stephen-Milborrow" + }, + { + "text": "Stephen Milborrow", + "fragments": [] + } + ], + [ + { + "name": "Tom E. Bishop", + "ids": [ + "1823550" + ], + "slug": "Tom-E.-Bishop" + }, + { + "text": "Tom E. Bishop", + "fragments": [] + } + ], + [ + { + "name": "Fred Nicolls", + "ids": [ + "2537623" + ], + "slug": "Fred-Nicolls" + }, + { + "text": "Fred Nicolls", + "fragments": [] + } + ] + ], + "citationContexts": [ + { + "text": "[44] 0.", + "fragments": [ + { + "start": 0, + "end": 4 + } + ] + }, + { + "text": "[44] S.", + "fragments": [ + { + "start": 0, + "end": 4 + } + ] + }, + { + "text": "[44] approach the problem with Active Shape Models (ASMs) that incorporate a modified version of SIFT descriptors [45].", + "fragments": [ + { + "start": 0, + "end": 4 + } + ] + }, + { + "text": "[44]", + "fragments": [ + { + "start": 0, + "end": 4 + } + ] + } + ], + "isKey": true + }, + { + "title": { + "text": "Face detection", + "fragments": [] + }, + "slug": "Face-detection-Zhu-Ramanan", + "venue": { + "text": "pose estimation, and landmark localization in the wild, in: Proceedings of IEEE International Conference on Computer Vision & Pattern Recognition (CVPR), IEEE", + "fragments": [] + }, + "year": 2012, + "authors": [ + [ + { + "name": "X. Zhu", + "ids": [], + "slug": "X.-Zhu" + }, + { + "text": "X. Zhu", + "fragments": [] + } + ], + [ + { + "name": "D. Ramanan", + "ids": [], + "slug": "D.-Ramanan" + }, + { + "text": "D. Ramanan", + "fragments": [] + } + ] + ], + "citationContexts": [ + { + "text": "The most notable databases of this category are LFPW [28], HELEN [29], AFW [17], AFLW [30] and IBUG [31] (all used for facial landmark points localization).", + "fragments": [ + { + "start": 75, + "end": 79 + } + ] + }, + { + "text": "Note that we employ DPMs [17] to estimate the initial landmarks locations for the first iteration of the above procedure.", + "fragments": [ + { + "start": 25, + "end": 29 + } + ] + }, + { + "text": "These databases can be separated in two major categories: (a) those captured under controlled conditions, e.g. Multi-PIE [24], XM2VTS [25], FRGCV2 [26], AR [27], and those captured under totally unconstrained conditions (in-the-wild), e.g. LFPW [28], HELEN [29], AFW [17], AFLW [30], IBUG [31].", + "fragments": [ + { + "start": 263, + "end": 266 + } + ] + }, + { + "text": "Furthermore, we computed the bounding boxes of all the aforementioned databases by using our in-house face detector, the one that is also employed in [16], which is a variant of [17].", + "fragments": [ + { + "start": 178, + "end": 182 + } + ] + }, + { + "text": "W), the first facial landmark localization challenge, that was\n1The annotations of XM2VTS, FRGC-V2, LFPW, HELEN, AFW and IBUG are publicly available from http://ibug.doc.ic.ac.uk/resources/facial-point-annotations/.", + "fragments": [ + { + "start": 113, + "end": 116 + } + ] + }, + { + "text": "LFPW, AFW, HELEN, XM2VTS and FRGC-V2 were provided for training, along with the corrected annotations produced with the semi-automatic annotation tool (Sec 3).", + "fragments": [ + { + "start": 6, + "end": 9 + } + ] + }, + { + "text": ", and outperform discriminative methodologies, such as CLMs [15], DPMs [17] and SDM [18].", + "fragments": [ + { + "start": 71, + "end": 75 + } + ] + }, + { + "text": "For each subject, there are available images for 15 different poses, 19 illumination conditions and 6\n5\nAC C\nEP\nTE\nD M\nAN U\nSC R\nIP T\n(a) MultiPIE/IBUG (b) XM2VTS (c) FRGC-V2 (d) AR\n(e) LFPW (f) HELEN (g) AFW (h) AFLW\nFigure 1: Landmarks configurations of existing databases.", + "fragments": [ + { + "start": 205, + "end": 208 + } + ] + }, + { + "text": "Multi-PIE [24], XM2VTS [25], FRGC-V2 [26], AR [27], LFPW [28], HELEN [29] and AFW [17].", + "fragments": [ + { + "start": 82, + "end": 86 + } + ] + }, + { + "text": "We employed the proposed tool to re-annotate all the widely used databases, i.e. Multi-PIE [24], XM2VTS [25], FRGC-V2 [26], AR [27], LFPW [28], HELEN [29] and AFW [17].", + "fragments": [ + { + "start": 159, + "end": 162 + } + ] + }, + { + "text": "The discriminative techniques can be further divided to those that use discriminative response map functions, such as Active Shape Models (ASMs) [13], Constrained Local Models (CLMs) [14, 15, 16] and Deformable Part Models (DPMs) [17], those that learn a cascade of regression functions, such as Supervised Descent Method (SDM) [18] and others [19, 20, 21], and, finally, those that employ random forests [22, 23].", + "fragments": [ + { + "start": 230, + "end": 234 + } + ] + }, + { + "text": "The accuracy of the fitting results was measured by the point-to-point RMS error between each fitted shape and the ground truth annotations, normalized by the face’s interoccular distance, as proposed in [17].", + "fragments": [ + { + "start": 204, + "end": 208 + } + ] + }, + { + "text": "HELEN, AFW, IBUG: The rest of in-the-wild databases were annotated\n11\nAC C\nEP TE\nD M\nAN U\nSC R\nIP T\n(a) Multi-PIE (b) XM2VTS\n(c) FRGC-V2 (d) LFPW\n(e) HELEN (f) AFW\nFigure 3: Examples of the annotated images.", + "fragments": [ + { + "start": 7, + "end": 10 + } + ] + }, + { + "text": "In both conducts, the training set consisted of the XM2VTS, FRGC-V2, LFPW, HELEN, AFW and IBUG databases that were annotated using the proposed semi-automatic procedure.", + "fragments": [ + { + "start": 82, + "end": 85 + } + ] + }, + { + "text": "LFPW [28], HELEN [29], AFW [17], AFLW [30], IBUG [31].", + "fragments": [ + { + "start": 27, + "end": 31 + } + ] + }, + { + "text": "AFW: The Annotated Faces in-the-wild (AFW) [17] database consists of 250 images with 468 faces, that is, more than one faces are annotated in each image.", + "fragments": [ + { + "start": 43, + "end": 47 + } + ] + }, + { + "text": "AR: The AR Face Database [27] contains over 4000 images corresponding to\n6\nAC C\nEP TE\nD M\nAN U\nSC\nR\nIP T\nDatabase conditions # faces # subjects # points pose\nMulti-PIE controlled ∼ 750000 337 68 [−45◦, 45◦] XM2VTS 2360 295 68 0◦ FRGC-V2 4950 466 5 0◦ AR ∼ 4000 126 22 0◦ LFPW\nin-the-wild\n1035\n−\n35\n[−45◦, 45◦] HELEN 2330 194 AFW 468 6 AFLW 25993 21 IBUG 135 68\nTable 1: Overview of the characteristics of existing facial databases.", + "fragments": [ + { + "start": 310, + "end": 313 + } + ] + }, + { + "text": "For example, DPMs [17] tend to return bounding boxes that only include facial texture and not any of the subject’s hair, as usually done by the Viola-Jones detector [48].", + "fragments": [ + { + "start": 18, + "end": 22 + } + ] + }, + { + "text": "The authors were encouraged, but not restricted, to use LFPW, AFW, HELEN, IBUG, FRGC-V2 and XM2VTS databases with the provided annotations.", + "fragments": [ + { + "start": 62, + "end": 65 + } + ] + } + ], + "isKey": true + }, + { + "id": "a74251efa970b92925b89eeef50a5e37d9281ad0", + "title": { + "text": "Annotated Facial Landmarks in the Wild: A large-scale, real-world database for facial landmark localization", + "fragments": [] + }, + "slug": "Annotated-Facial-Landmarks-in-the-Wild:-A-database-Köstinger-Wohlhart", + "venue": { + "text": "2011 IEEE International Conference on Computer Vision Workshops (ICCV Workshops)", + "fragments": [] + }, + "year": 2011, + "authors": [ + [ + { + "name": "Martin Köstinger", + "ids": [ + "1993853" + ], + "slug": "Martin-Köstinger" + }, + { + "text": "Martin Köstinger", + "fragments": [] + } + ], + [ + { + "name": "Paul Wohlhart", + "ids": [ + "3202367" + ], + "slug": "Paul-Wohlhart" + }, + { + "text": "Paul Wohlhart", + "fragments": [] + } + ], + [ + { + "name": "Peter M. Roth", + "ids": [ + "1791182" + ], + "slug": "Peter-M.-Roth" + }, + { + "text": "Peter M. Roth", + "fragments": [] + } + ], + [ + { + "name": "Horst Bischof", + "ids": [ + "3628150" + ], + "slug": "Horst-Bischof" + }, + { + "text": "Horst Bischof", + "fragments": [] + } + ] + ], + "citationContexts": [ + { + "text": "The most notable databases of this category are LFPW [28], HELEN [29], AFW [17], AFLW [30] and IBUG [31] (all used for facial landmark points localization).", + "fragments": [ + { + "start": 86, + "end": 90 + } + ] + }, + { + "text": "These databases can be separated in two major categories: (a) those captured under controlled conditions, e.g. Multi-PIE [24], XM2VTS [25], FRGCV2 [26], AR [27], and those captured under totally unconstrained conditions (in-the-wild), e.g. LFPW [28], HELEN [29], AFW [17], AFLW [30], IBUG [31].", + "fragments": [ + { + "start": 273, + "end": 277 + } + ] + }, + { + "text": "AFLW: The Annotated Facial Landmarks in theWild (AFLW) [30] database consists of 25993 images gathered from Flickr, exhibiting a large variety in appearance (e.g., pose, expression, ethnicity, age, gender) as well as general imaging and environmental conditions.", + "fragments": [ + { + "start": 0, + "end": 4 + } + ] + }, + { + "text": "For each subject, there are available images for 15 different poses, 19 illumination conditions and 6\n5\nAC C\nEP\nTE\nD M\nAN U\nSC R\nIP T\n(a) MultiPIE/IBUG (b) XM2VTS (c) FRGC-V2 (d) AR\n(e) LFPW (f) HELEN (g) AFW (h) AFLW\nFigure 1: Landmarks configurations of existing databases.", + "fragments": [ + { + "start": 213, + "end": 217 + } + ] + }, + { + "text": "AFLW: The Annotated Facial Landmarks in theWild (AFLW) [30] database consists of 25993 images gathered from Flickr, exhibiting a large variety in appearance (e.", + "fragments": [ + { + "start": 55, + "end": 59 + } + ] + }, + { + "text": "[30] M.", + "fragments": [ + { + "start": 0, + "end": 4 + } + ] + }, + { + "text": "LFPW [28], HELEN [29], AFW [17], AFLW [30], IBUG [31].", + "fragments": [ + { + "start": 38, + "end": 42 + } + ] + }, + { + "text": "AR: The AR Face Database [27] contains over 4000 images corresponding to\n6\nAC C\nEP TE\nD M\nAN U\nSC\nR\nIP T\nDatabase conditions # faces # subjects # points pose\nMulti-PIE controlled ∼ 750000 337 68 [−45◦, 45◦] XM2VTS 2360 295 68 0◦ FRGC-V2 4950 466 5 0◦ AR ∼ 4000 126 22 0◦ LFPW\nin-the-wild\n1035\n−\n35\n[−45◦, 45◦] HELEN 2330 194 AFW 468 6 AFLW 25993 21 IBUG 135 68\nTable 1: Overview of the characteristics of existing facial databases.", + "fragments": [ + { + "start": 320, + "end": 324 + } + ] + } + ], + "isKey": true + }, + { + "title": { + "text": "Fddb: A benchmark for face detection in unconstrained settings", + "fragments": [] + }, + "slug": "Fddb:-A-benchmark-for-face-detection-in-settings-Jain-Learned-Miller", + "venue": { + "text": "Tech. Rep. UM-CS-2010-009, University of Massachusetts, Amherst", + "fragments": [] + }, + "year": 2010, + "authors": [ + [ + { + "name": "V. Jain", + "ids": [], + "slug": "V.-Jain" + }, + { + "text": "V. Jain", + "fragments": [] + } + ], + [ + { + "name": "E. Learned-Miller", + "ids": [], + "slug": "E.-Learned-Miller" + }, + { + "text": "E. Learned-Miller", + "fragments": [] + } + ] + ], + "citationContexts": [ + { + "text": "Consequently, in order to facilitate the participants and make the competition less dependent to a face\n20\ndetector’s performance, we suggested them to use one of the face detection methods that took part in the Face Detection Data Set and Benchmark (FDDB) [49].", + "fragments": [ + { + "start": 250, + "end": 254 + } + ] + }, + { + "text": "[49] V.", + "fragments": [ + { + "start": 0, + "end": 4 + } + ] + }, + { + "text": "detector’s performance, we suggested them to use one of the face detection methods that took part in the Face Detection Data Set and Benchmark (FDDB) [49].", + "fragments": [ + { + "start": 150, + "end": 154 + } + ] + }, + { + "text": "The results presented in the Face Detection Data Set and Benchmark (FDDB) [49] show that current stateof-the-art techniques achieve very good true positive rates.", + "fragments": [ + { + "start": 74, + "end": 78 + } + ] + } + ], + "isKey": true + }, + { + "title": { + "text": "Multi-pie", + "fragments": [] + }, + "slug": "Multi-pie-Gross-Matthews", + "venue": { + "text": "Image and Vision Computing 28 (5)", + "fragments": [] + }, + "year": 2010, + "authors": [ + [ + { + "name": "R. Gross", + "ids": [], + "slug": "R.-Gross" + }, + { + "text": "R. Gross", + "fragments": [] + } + ], + [ + { + "name": "I. Matthews", + "ids": [], + "slug": "I.-Matthews" + }, + { + "text": "I. Matthews", + "fragments": [] + } + ], + [ + { + "name": "J. Cohn", + "ids": [], + "slug": "J.-Cohn" + }, + { + "text": "J. Cohn", + "fragments": [] + } + ], + [ + { + "name": "T. Kanade", + "ids": [], + "slug": "T.-Kanade" + }, + { + "text": "T. Kanade", + "fragments": [] + } + ], + [ + { + "name": "S. Baker", + "ids": [], + "slug": "S.-Baker" + }, + { + "text": "S. Baker", + "fragments": [] + } + ] + ], + "citationContexts": [ + { + "text": "The most popular such databases are Multi-PIE [24] (used for face recognition, expressions recognition, landmark points localization), FRGC-V2 [26] (used for face recognition), XM2VTS [25] and AR [27] (both used for face recognition and landmark points localization).", + "fragments": [ + { + "start": 46, + "end": 50 + } + ] + }, + { + "text": "The provided facial landmark annotations are produced by employing the annotation scheme of Multi-PIE (Fig.", + "fragments": [ + { + "start": 92, + "end": 101 + } + ] + }, + { + "text": "These databases can be separated in two major categories: (a) those captured under controlled conditions, e.g. Multi-PIE [24], XM2VTS [25], FRGCV2 [26], AR [27], and those captured under totally unconstrained conditions (in-the-wild), e.g. LFPW [28], HELEN [29], AFW [17], AFLW [30], IBUG [31].", + "fragments": [ + { + "start": 111, + "end": 120 + } + ] + }, + { + "text": "Note that in the case of Multi-PIE, even though the original and generated annotations have the same configuration, the generated ones are more accurate.\nusing a common procedure.", + "fragments": [ + { + "start": 25, + "end": 34 + } + ] + }, + { + "text": "The advantages of the generated annotations1 are twofold: (1) They all have the same landmarks configuration, i.e. the one employed in Multi-PIE (Fig.", + "fragments": [ + { + "start": 135, + "end": 144 + } + ] + }, + { + "text": "16: Fit the person-specific AOM to the image i. 17: end for 18: end for 19: end if 20: Check and manually correct, if necessary, the generated annotations of Q.\nMulti-PIE: The available Multi-PIE annotations cover only the neutral expression with pose [−45◦, 45◦] and multiple non-neutral expressions with pose 0◦.", + "fragments": [ + { + "start": 161, + "end": 170 + } + ] + }, + { + "text": "Facial databases under controlled conditions Multi-PIE: The CMU Multi Pose Illumination, and Expression (MultiPIE) Database [24] contains around 750000 images of 337 subjects captured under laboratory conditions in four different sessions.", + "fragments": [ + { + "start": 124, + "end": 128 + } + ] + }, + { + "text": "Multi-PIE [24], XM2VTS [25], FRGC-V2 [26], AR [27], LFPW [28], HELEN [29] and AFW [17].", + "fragments": [ + { + "start": 10, + "end": 14 + } + ] + }, + { + "text": "We employed the proposed tool to re-annotate all the widely used databases, i.e. Multi-PIE [24], XM2VTS [25], FRGC-V2 [26], AR [27], LFPW [28], HELEN [29] and AFW [17].", + "fragments": [ + { + "start": 81, + "end": 90 + } + ] + }, + { + "text": "For example, in Multi-PIE, the annotations for subjects with expressions “disgust” at 0◦ and “neutral” at 15◦ are provided and we want to produce the annotations for subjects with expression “disgust” at 15◦.", + "fragments": [ + { + "start": 16, + "end": 25 + } + ] + }, + { + "text": "The images of each such pose cluster were semi-automatically annotated using images from Multi-PIE with the same pose.", + "fragments": [ + { + "start": 89, + "end": 98 + } + ] + }, + { + "text": "However, the accuracy of the annotations in some cases is limited and the locations of the provided points do not correspond to ones of Multi-PIE.", + "fragments": [ + { + "start": 136, + "end": 145 + } + ] + }, + { + "text": "Multi-PIE: The CMU Multi Pose Illumination, and Expression (MultiPIE) Database [24] contains around 750000 images of 337 subjects captured under laboratory conditions in four different sessions.", + "fragments": [ + { + "start": 0, + "end": 9 + } + ] + }, + { + "text": "XM2VTS: The images of XM2VTS’s first session were semi-automatically annotated by setting V to be the subjects of Multi-PIE with neutral expression and [−15◦, 15◦] poses.", + "fragments": [ + { + "start": 113, + "end": 122 + } + ] + }, + { + "text": "HELEN, AFW, IBUG: The rest of in-the-wild databases were annotated\n11\nAC C\nEP TE\nD M\nAN U\nSC R\nIP T\n(a) Multi-PIE (b) XM2VTS\n(c) FRGC-V2 (d) LFPW\n(e) HELEN (f) AFW\nFigure 3: Examples of the annotated images.", + "fragments": [ + { + "start": 104, + "end": 113 + } + ] + }, + { + "text": "This subset was annotated by employing images from Multi-PIE with six expressions and [−15◦, 15◦] poses as V .", + "fragments": [ + { + "start": 51, + "end": 60 + } + ] + }, + { + "text": "Multi-PIE [24], XM2VTS [25], FRGCV2 [26], AR [27], and those captured under totally unconstrained conditions (in-the-wild), e.", + "fragments": [ + { + "start": 10, + "end": 14 + } + ] + }, + { + "text": "In case Q has multiple images per subject (e.g. Multi-PIE, XM2VTS, FRGC-V2, AR), the above method can be extended to further improve the generated annotations.", + "fragments": [ + { + "start": 48, + "end": 57 + } + ] + }, + { + "text": "To this end, we selected such images of N = 80 different subjects with frontal pose from the Multi-PIE database.", + "fragments": [ + { + "start": 93, + "end": 102 + } + ] + }, + { + "text": "AR: The AR Face Database [27] contains over 4000 images corresponding to\n6\nAC C\nEP TE\nD M\nAN U\nSC\nR\nIP T\nDatabase conditions # faces # subjects # points pose\nMulti-PIE controlled ∼ 750000 337 68 [−45◦, 45◦] XM2VTS 2360 295 68 0◦ FRGC-V2 4950 466 5 0◦ AR ∼ 4000 126 22 0◦ LFPW\nin-the-wild\n1035\n−\n35\n[−45◦, 45◦] HELEN 2330 194 AFW 468 6 AFLW 25993 21 IBUG 135 68\nTable 1: Overview of the characteristics of existing facial databases.", + "fragments": [ + { + "start": 158, + "end": 167 + } + ] + } + ], + "isKey": true + } + ], + "requestedPageSize": 10, + "pageNumber": 1, + "totalPages": 5, + "sort": "is-influential" + }, + "figureExtractions": { + "figures": [ + { + "name": "1", + "figureType": "figure", + "caption": "Figure 1: Landmarks configurations of existing databases. Note they all have different number of landmark points with semantically different locations.", + "uri": "https://ai2-s2-public.s3.amazonaws.com/figures/2017-08-08/e4754afaa15b1b53e70743880484b8d0736990ff/7-Figure1-1.png", + "width": 449, + "height": 449 + }, + { + "name": "1", + "figureType": "table", + "caption": "Table 1: Overview of the characteristics of existing facial databases.", + "uri": "https://ai2-s2-public.s3.amazonaws.com/figures/2017-08-08/e4754afaa15b1b53e70743880484b8d0736990ff/8-Table1-1.png", + "width": 439, + "height": 439 + }, + { + "name": "2", + "figureType": "figure", + "caption": "Figure 2: Flowchart of the proposed tool. Given a set of landmarked images V with various poses and expressions, we aim to annotate a set of non-annotated images Q (1) with the same subjects and different poses and expressions, or (2) with different subjects but similar pose and expressions.", + "uri": "https://ai2-s2-public.s3.amazonaws.com/figures/2017-08-08/e4754afaa15b1b53e70743880484b8d0736990ff/11-Figure2-1.png", + "width": 449, + "height": 449 + }, + { + "name": "2", + "figureType": "table", + "caption": "Table 2: Overview of the characteristics of the 300-W database.", + "uri": "https://ai2-s2-public.s3.amazonaws.com/figures/2017-08-08/e4754afaa15b1b53e70743880484b8d0736990ff/16-Table2-1.png", + "width": 445, + "height": 445 + }, + { + "name": "3", + "figureType": "figure", + "caption": "Figure 3: Examples of the annotated images. For each database, the image on the left has the original annotations and the one on the right shows the annotations generated by the proposed tool. Note that in the case of Multi-PIE, even though the original and generated annotations have the same configuration, the generated ones are more accurate.", + "uri": "https://ai2-s2-public.s3.amazonaws.com/figures/2017-08-08/e4754afaa15b1b53e70743880484b8d0736990ff/13-Figure3-1.png", + "width": 445, + "height": 445 + }, + { + "name": "3", + "figureType": "table", + "caption": "Table 3: Median absolute deviation of the fitting results of the first conduct of 300-W challenge in 2013, reported for both 68 and 51 points.", + "uri": "https://ai2-s2-public.s3.amazonaws.com/figures/2017-08-08/e4754afaa15b1b53e70743880484b8d0736990ff/19-Table3-1.png", + "width": 422, + "height": 422 + }, + { + "name": "4", + "figureType": "figure", + "caption": "Figure 4: The cardinality of W and V per iteration.", + "uri": "https://ai2-s2-public.s3.amazonaws.com/figures/2017-08-08/e4754afaa15b1b53e70743880484b8d0736990ff/14-Figure4-1.png", + "width": 321, + "height": 321 + }, + { + "name": "4", + "figureType": "table", + "caption": "Table 4: Overview of the characteristics of the cropped images of the 300-W database.", + "uri": "https://ai2-s2-public.s3.amazonaws.com/figures/2017-08-08/e4754afaa15b1b53e70743880484b8d0736990ff/20-Table4-1.png", + "width": 448, + "height": 448 + }, + { + "name": "5", + "figureType": "figure", + "caption": "Figure 5: Each ellipse denotes the variance of each landmark point with regards to three expert human annotators. The colours of the points rank them with respect to their standard deviation normalized by the face size.", + "uri": "https://ai2-s2-public.s3.amazonaws.com/figures/2017-08-08/e4754afaa15b1b53e70743880484b8d0736990ff/15-Figure5-1.png", + "width": 413, + "height": 413 + }, + { + "name": "5", + "figureType": "table", + "caption": "Table 5: Second conduct of the 300-W challenge. 2nd column: Number of images for which an estimation of the landmarks was returned. 3rd and 4th columns: The mean absolute deviation of the fitting results for both 68 and 51 points. 5th column: Mean computational cost per method.", + "uri": "https://ai2-s2-public.s3.amazonaws.com/figures/2017-08-08/e4754afaa15b1b53e70743880484b8d0736990ff/23-Table5-1.png", + "width": 449, + "height": 449 + }, + { + "name": "6", + "figureType": "figure", + "caption": "Figure 6: The 51-points mark-up is a subset of the 68-points one after removing the 17 points of the face’s boundary. The interoccular distance is defined between the outer points of the eyes.", + "uri": "https://ai2-s2-public.s3.amazonaws.com/figures/2017-08-08/e4754afaa15b1b53e70743880484b8d0736990ff/17-Figure6-1.png", + "width": 231, + "height": 231 + }, + { + "name": "6", + "figureType": "table", + "caption": "Table 6: Percentage of images with fitting error less than the specified values for the winners of the first (Yan et al. [46], Zhou et al. [47]) and second (J. Deng, H. Fan) 300-W challenges, and Oracle. The error is based on 68 points using both indoor and oudoor images.", + "uri": "https://ai2-s2-public.s3.amazonaws.com/figures/2017-08-08/e4754afaa15b1b53e70743880484b8d0736990ff/27-Table6-1.png", + "width": 445, + "height": 445 + }, + { + "name": "7", + "figureType": "figure", + "caption": "Figure 7: Fitting results of the first conduct of the 300-W challenge in 2013. The plots show the Cumulative Error Distribution (CED) curves with respect to the landmarks (68 and 51 points) and the condtions (indoor, outdoor or both).", + "uri": "https://ai2-s2-public.s3.amazonaws.com/figures/2017-08-08/e4754afaa15b1b53e70743880484b8d0736990ff/18-Figure7-1.png", + "width": 477, + "height": 477 + }, + { + "name": "8", + "figureType": "figure", + "caption": "Figure 8: Indicative examples of the way the images were cropped for the second conduct of the 300-W challenge.", + "uri": "https://ai2-s2-public.s3.amazonaws.com/figures/2017-08-08/e4754afaa15b1b53e70743880484b8d0736990ff/21-Figure8-1.png", + "width": 449, + "height": 449 + }, + { + "name": "9", + "figureType": "figure", + "caption": "Figure 9: Fitting results of the second conduct of the 300-W challenge in 2015. The plots show the Cumulative Error Distribution (CED) curves with respect to the landmarks (68 and 51 points) and the condtions (indoor, outdoor or both).", + "uri": "https://ai2-s2-public.s3.amazonaws.com/figures/2017-08-08/e4754afaa15b1b53e70743880484b8d0736990ff/22-Figure9-1.png", + "width": 477, + "height": 477 + }, + { + "name": "10", + "figureType": "figure", + "caption": "Figure 10: Fitting examples of the first conduct of the 300-W challenge in 2013. Each row shows the fitted landmarks for each participating method.", + "uri": "https://ai2-s2-public.s3.amazonaws.com/figures/2017-08-08/e4754afaa15b1b53e70743880484b8d0736990ff/24-Figure10-1.png", + "width": 449, + "height": 449 + }, + { + "name": "11", + "figureType": "figure", + "caption": "Figure 11: Fitting examples of the second conduct of the 300-W challenge in 2015. Each row shows the fitted landmarks for each participating method.", + "uri": "https://ai2-s2-public.s3.amazonaws.com/figures/2017-08-08/e4754afaa15b1b53e70743880484b8d0736990ff/26-Figure11-1.png", + "width": 455, + "height": 455 + } + ] + }, + "presentations": [ + { + "contentType": "PRESENTATION", + "id": "df1e22c9a92e6c506c228ef23049ab3b8c908e80", + "title": "Face detection, pose estimation, and landmark localization in the wild", + "authors": [ + "Xiangxin Zhu", + "Deva Ramanan" + ], + "published": 1412121600, + "url": "https://pdfs.semanticscholar.org/presentation/df1e/22c9a92e6c506c228ef23049ab3b8c908e80.pdf" + }, + { + "contentType": "PRESENTATION", + "id": "b4d2151e29fb12dbe5d164b430273de65103d39b", + "title": "Annotated Facial Landmarks in the Wild: A large-scale, real-world database for facial landmark localization", + "authors": [ + "Martin Köstinger", + "Paul Wohlhart", + "Peter M. Roth", + "Horst Bischof" + ], + "published": 1412121600, + "url": "https://pdfs.semanticscholar.org/presentation/b4d2/151e29fb12dbe5d164b430273de65103d39b.pdf" + }, + { + "contentType": "PRESENTATION", + "id": "d5c22cb54bd23f17289c31abb84baaf0cd439540", + "title": "Face Recognition by Humans: Nineteen Results All Computer Vision Researchers Should Know About", + "authors": [ + "Pawan Sinha", + "Benjamin J. Balas", + "Yuri Ostrovsky", + "Richard Russell" + ], + "published": 1412121600, + "url": "https://pdfs.semanticscholar.org/presentation/d5c2/2cb54bd23f17289c31abb84baaf0cd439540.pdf" + }, + { + "contentType": "PRESENTATION", + "id": "8e1f99812078e91a0d5722eaf54f20c6fc65df4c", + "title": "Challenges of the Open Source Component Marketplace in the Industry", + "authors": [ + "Claudia P. Ayala", + "Øyvind Hauge", + "Reidar Conradi", + "Xavier Franch", + "Jingyue Li", + "Ketil Sandanger Velle" + ], + "published": 1412121600, + "url": "https://pdfs.semanticscholar.org/presentation/8e1f/99812078e91a0d5722eaf54f20c6fc65df4c.pdf" + }, + { + "contentType": "PRESENTATION", + "id": "47bae254f82bdf4b66ea493afebf5a12c7291db2", + "title": "Pervasive computing: vision and challenges", + "authors": [ + "Mahadev Satyanarayanan" + ], + "published": 1412121600, + "url": "https://pdfs.semanticscholar.org/presentation/47ba/e254f82bdf4b66ea493afebf5a12c7291db2.pdf" + }, + { + "contentType": "PRESENTATION", + "id": "cf9f048a00f14ec5348a1dfe76a2e23b51e0b26a", + "title": "Challenges and opportunities in enterprise-wide optimization in the pharmaceutical industry", + "authors": [ + "José Miguel Laínez", + "E. Schaefer", + "Gintaras V. Reklaitis" + ], + "published": 1511903839.603717, + "url": "https://pdfs.semanticscholar.org/presentation/cf9f/048a00f14ec5348a1dfe76a2e23b51e0b26a.pdf" + }, + { + "contentType": "PRESENTATION", + "id": "cf0e634e608cf1d5574266d8c0110b75b3e615fd", + "title": "Industry 4.0 - Challenges in Anti-Counterfeiting", + "authors": [ + "Christian Thiel", + "Christoph Thiel" + ], + "published": 1511948094.409062, + "url": "https://pdfs.semanticscholar.org/presentation/cf0e/634e608cf1d5574266d8c0110b75b3e615fd.pdf" + }, + { + "contentType": "PRESENTATION", + "id": "6b9267cc7b4277e4abdefb52157bc69d318412c9", + "title": "Energy demand forecasting: industry practices and challenges", + "authors": [ + "Mathieu Sinn" + ], + "published": 1447968149.539, + "url": "https://pdfs.semanticscholar.org/presentation/6b92/67cc7b4277e4abdefb52157bc69d318412c9.pdf" + }, + { + "contentType": "PRESENTATION", + "id": "09e47a1b9abcce753ce124eab586d1cb03abdde4", + "title": "Hardware Acceleration Technologies in Computer Algebra : Challenges and Impact", + "authors": [ + "Sardar Anisul Haque" + ], + "published": 1427429814, + "url": "https://pdfs.semanticscholar.org/presentation/09e4/7a1b9abcce753ce124eab586d1cb03abdde4.pdf" + }, + { + "contentType": "PRESENTATION", + "id": "6442a7dc149de6763b0959d924ec5a3dbe09ec1f", + "title": "New results on the coarseness of bicolored point sets", + "authors": [ + "José Miguel Díaz-Báñez", + "Ruy Fabila Monroy", + "Pablo Pérez-Lantero", + "Inmaculada Ventura" + ], + "published": 1385693873, + "url": "https://pdfs.semanticscholar.org/presentation/6442/a7dc149de6763b0959d924ec5a3dbe09ec1f.pdf" + } + ], + "featuredContent": [] +}
\ No newline at end of file diff --git a/scraper/samples/s2-papers-api.json b/scraper/samples/s2-papers-api.json new file mode 100644 index 00000000..4db657b8 --- /dev/null +++ b/scraper/samples/s2-papers-api.json @@ -0,0 +1,2855 @@ +{ + "arxivId": "1501.05703", + "authors": [ + { + "authorId": "49410785", + "name": "Ning Zhang", + "url": "https://www.semanticscholar.org/author/49410785" + }, + { + "authorId": "2210374", + "name": "Manohar Paluri", + "url": "https://www.semanticscholar.org/author/2210374" + }, + { + "authorId": "2188620", + "name": "Yaniv Taigman", + "url": "https://www.semanticscholar.org/author/2188620" + }, + { + "authorId": "2276554", + "name": "Rob Fergus", + "url": "https://www.semanticscholar.org/author/2276554" + }, + { + "authorId": "1769383", + "name": "Lubomir D. Bourdev", + "url": "https://www.semanticscholar.org/author/1769383" + } + ], + "citationVelocity": 22, + "citations": [ + { + "arxivId": "1612.02155", + "authors": [ + { + "authorId": "2963501", + "name": "Shayan Modiri Assari", + "url": "https://www.semanticscholar.org/author/2963501" + }, + { + "authorId": "1803711", + "name": "Haroon Idrees", + "url": "https://www.semanticscholar.org/author/1803711" + }, + { + "authorId": "1745480", + "name": "Mubarak Shah", + "url": "https://www.semanticscholar.org/author/1745480" + } + ], + "doi": null, + "isInfluential": false, + "paperId": "a1e97c4043d5cc9896dc60ae7ca135782d89e5fc", + "title": "Re-identification of Humans in Crowds using Personal, Social and Environmental Constraints", + "url": "https://www.semanticscholar.org/paper/a1e97c4043d5cc9896dc60ae7ca135782d89e5fc", + "venue": "ArXiv", + "year": 2016 + }, + { + "arxivId": "1804.04779", + "authors": [ + { + "authorId": "32222907", + "name": "Qianru Sun", + "url": "https://www.semanticscholar.org/author/32222907" + }, + { + "authorId": "9102722", + "name": "Ayush Tewari", + "url": "https://www.semanticscholar.org/author/9102722" + }, + { + "authorId": "9765909", + "name": "Weipeng Xu", + "url": "https://www.semanticscholar.org/author/9765909" + }, + { + "authorId": "1739548", + "name": "Mario Fritz", + "url": "https://www.semanticscholar.org/author/1739548" + }, + { + "authorId": "1680185", + "name": "Christian Theobalt", + "url": "https://www.semanticscholar.org/author/1680185" + }, + { + "authorId": "1697100", + "name": "Bernt Schiele", + "url": "https://www.semanticscholar.org/author/1697100" + } + ], + "doi": null, + "isInfluential": false, + "paperId": "cd2c54705c455a4379f45eefdf32d8d10087e521", + "title": "A Hybrid Model for Identity Obfuscation by Face Replacement", + "url": "https://www.semanticscholar.org/paper/cd2c54705c455a4379f45eefdf32d8d10087e521", + "venue": "ArXiv", + "year": 2018 + }, + { + "arxivId": null, + "authors": [ + { + "authorId": "34949584", + "name": "Guangcai Zha", + "url": "https://www.semanticscholar.org/author/34949584" + } + ], + "doi": null, + "isInfluential": false, + "paperId": "0da75b0d341c8f945fae1da6c77b6ec345f47f2a", + "title": "The Effect of Computer-Generated Descriptions on Photo-Sharing Experiences of People With Visual Impairments", + "url": "https://www.semanticscholar.org/paper/0da75b0d341c8f945fae1da6c77b6ec345f47f2a", + "venue": "", + "year": 2017 + }, + { + "arxivId": null, + "authors": [ + { + "authorId": "2851931", + "name": "Zongji Sun", + "url": "https://www.semanticscholar.org/author/2851931" + }, + { + "authorId": "2133352", + "name": "Li Meng", + "url": "https://www.semanticscholar.org/author/2133352" + }, + { + "authorId": "1763301", + "name": "Aladdin M. Ariyaeeinia", + "url": "https://www.semanticscholar.org/author/1763301" + }, + { + "authorId": "2606013", + "name": "Xiaodong Duan", + "url": "https://www.semanticscholar.org/author/2606013" + }, + { + "authorId": "1709835", + "name": "Zheng-Hua Tan", + "url": "https://www.semanticscholar.org/author/1709835" + } + ], + "doi": "10.1109/MIPRO.2016.7522350", + "isInfluential": false, + "paperId": "8886b21f97c114a23b24dc7025bbf42885adc3a7", + "title": "Privacy protection performance of De-identified face images with and without background", + "url": "https://www.semanticscholar.org/paper/8886b21f97c114a23b24dc7025bbf42885adc3a7", + "venue": "2016 39th International Convention on Information and Communication Technology, Electronics and Microelectronics (MIPRO)", + "year": 2016 + }, + { + "arxivId": null, + "authors": [ + { + "authorId": "2963501", + "name": "Shayan Modiri Assari", + "url": "https://www.semanticscholar.org/author/2963501" + }, + { + "authorId": "1803711", + "name": "Haroon Idrees", + "url": "https://www.semanticscholar.org/author/1803711" + }, + { + "authorId": "1745480", + "name": "Mubarak Shah", + "url": "https://www.semanticscholar.org/author/1745480" + } + ], + "doi": "10.1007/978-3-319-46475-6_8", + "isInfluential": false, + "paperId": "2b339ece73e3787f445c5b92078e8f82c9b1c522", + "title": "Human Re-identification in Crowd Videos Using Personal, Social and Environmental Constraints", + "url": "https://www.semanticscholar.org/paper/2b339ece73e3787f445c5b92078e8f82c9b1c522", + "venue": "ECCV", + "year": 2016 + }, + { + "arxivId": null, + "authors": [ + { + "authorId": "1748171", + "name": "Paarijaat Aditya", + "url": "https://www.semanticscholar.org/author/1748171" + }, + { + "authorId": "50410801", + "name": "Rijurekha Sen", + "url": "https://www.semanticscholar.org/author/50410801" + }, + { + "authorId": "1736987", + "name": "Peter Druschel", + "url": "https://www.semanticscholar.org/author/1736987" + }, + { + "authorId": "2390510", + "name": "Seong Joon Oh", + "url": "https://www.semanticscholar.org/author/2390510" + }, + { + "authorId": "1798000", + "name": "Rodrigo Benenson", + "url": "https://www.semanticscholar.org/author/1798000" + }, + { + "authorId": "1739548", + "name": "Mario Fritz", + "url": "https://www.semanticscholar.org/author/1739548" + }, + { + "authorId": "1697100", + "name": "Bernt Schiele", + "url": "https://www.semanticscholar.org/author/1697100" + }, + { + "authorId": "1708336", + "name": "Bobby Bhattacharjee", + "url": "https://www.semanticscholar.org/author/1708336" + }, + { + "authorId": "40084289", + "name": "Tong Tong Wu", + "url": "https://www.semanticscholar.org/author/40084289" + } + ], + "doi": "10.1145/2938559.2938589", + "isInfluential": false, + "paperId": "10ab1b48b2a55ec9e2920a5397febd84906a7769", + "title": "Demo: I-Pic: A Platform for Privacy-Compliant Image Capture", + "url": "https://www.semanticscholar.org/paper/10ab1b48b2a55ec9e2920a5397febd84906a7769", + "venue": "MobiSys", + "year": 2016 + }, + { + "arxivId": "1607.08438", + "authors": [ + { + "authorId": "2390510", + "name": "Seong Joon Oh", + "url": "https://www.semanticscholar.org/author/2390510" + }, + { + "authorId": "1798000", + "name": "Rodrigo Benenson", + "url": "https://www.semanticscholar.org/author/1798000" + }, + { + "authorId": "1739548", + "name": "Mario Fritz", + "url": "https://www.semanticscholar.org/author/1739548" + }, + { + "authorId": "1697100", + "name": "Bernt Schiele", + "url": "https://www.semanticscholar.org/author/1697100" + } + ], + "doi": "10.1007/978-3-319-46487-9_2", + "isInfluential": true, + "paperId": "bc27434e376db89fe0e6ef2d2fabc100d2575ec6", + "title": "Faceless Person Recognition; Privacy Implications in Social Media", + "url": "https://www.semanticscholar.org/paper/bc27434e376db89fe0e6ef2d2fabc100d2575ec6", + "venue": "ECCV", + "year": 2016 + }, + { + "arxivId": "1509.03502", + "authors": [ + { + "authorId": "2390510", + "name": "Seong Joon Oh", + "url": "https://www.semanticscholar.org/author/2390510" + }, + { + "authorId": "1798000", + "name": "Rodrigo Benenson", + "url": "https://www.semanticscholar.org/author/1798000" + }, + { + "authorId": "1739548", + "name": "Mario Fritz", + "url": "https://www.semanticscholar.org/author/1739548" + }, + { + "authorId": "1697100", + "name": "Bernt Schiele", + "url": "https://www.semanticscholar.org/author/1697100" + } + ], + "doi": "10.1109/ICCV.2015.440", + "isInfluential": true, + "paperId": "0c59071ddd33849bd431165bc2d21bbe165a81e0", + "title": "Person Recognition in Personal Photo Collections", + "url": "https://www.semanticscholar.org/paper/0c59071ddd33849bd431165bc2d21bbe165a81e0", + "venue": "2015 IEEE International Conference on Computer Vision (ICCV)", + "year": 2015 + }, + { + "arxivId": "1705.07206", + "authors": [ + { + "authorId": "2757639", + "name": "Jianshu Li", + "url": "https://www.semanticscholar.org/author/2757639" + }, + { + "authorId": "46509407", + "name": "Jian Zhao", + "url": "https://www.semanticscholar.org/author/46509407" + }, + { + "authorId": "49020088", + "name": "Yunchao Wei", + "url": "https://www.semanticscholar.org/author/49020088" + }, + { + "authorId": "3350185", + "name": "Congyan Lang", + "url": "https://www.semanticscholar.org/author/3350185" + }, + { + "authorId": "2263674", + "name": "Yidong Li", + "url": "https://www.semanticscholar.org/author/2263674" + }, + { + "authorId": "1715286", + "name": "Terence Sim", + "url": "https://www.semanticscholar.org/author/1715286" + }, + { + "authorId": "1698982", + "name": "Shuicheng Yan", + "url": "https://www.semanticscholar.org/author/1698982" + }, + { + "authorId": "33221685", + "name": "Jiashi Feng", + "url": "https://www.semanticscholar.org/author/33221685" + } + ], + "doi": null, + "isInfluential": false, + "paperId": "b5968e7bb23f5f03213178c22fd2e47af3afa04c", + "title": "Multiple-Human Parsing in the Wild", + "url": "https://www.semanticscholar.org/paper/b5968e7bb23f5f03213178c22fd2e47af3afa04c", + "venue": "", + "year": 2017 + }, + { + "arxivId": "1806.03084", + "authors": [ + { + "authorId": "39360892", + "name": "Qingqiu Huang", + "url": "https://www.semanticscholar.org/author/39360892" + }, + { + "authorId": "50446092", + "name": "Yu Xiong", + "url": "https://www.semanticscholar.org/author/50446092" + }, + { + "authorId": "1807606", + "name": "Dahua Lin", + "url": "https://www.semanticscholar.org/author/1807606" + } + ], + "doi": null, + "isInfluential": true, + "paperId": "1c9efb6c895917174ac6ccc3bae191152f90c625", + "title": "Unifying Identification and Context Learning for Person Recognition", + "url": "https://www.semanticscholar.org/paper/1c9efb6c895917174ac6ccc3bae191152f90c625", + "venue": "ArXiv", + "year": 2018 + }, + { + "arxivId": "1805.05838", + "authors": [ + { + "authorId": "9517443", + "name": "Tribhuvanesh Orekondy", + "url": "https://www.semanticscholar.org/author/9517443" + }, + { + "authorId": "2390510", + "name": "Seong Joon Oh", + "url": "https://www.semanticscholar.org/author/2390510" + }, + { + "authorId": "1697100", + "name": "Bernt Schiele", + "url": "https://www.semanticscholar.org/author/1697100" + }, + { + "authorId": "1739548", + "name": "Mario Fritz", + "url": "https://www.semanticscholar.org/author/1739548" + } + ], + "doi": null, + "isInfluential": false, + "paperId": "3e0a1884448bfd7f416c6a45dfcdfc9f2e617268", + "title": "Understanding and Controlling User Linkability in Decentralized Learning", + "url": "https://www.semanticscholar.org/paper/3e0a1884448bfd7f416c6a45dfcdfc9f2e617268", + "venue": "ArXiv", + "year": 2018 + }, + { + "arxivId": null, + "authors": [ + { + "authorId": "2416019", + "name": "Lin Yuan", + "url": "https://www.semanticscholar.org/author/2416019" + }, + { + "authorId": "1681498", + "name": "Touradj Ebrahimi", + "url": "https://www.semanticscholar.org/author/1681498" + } + ], + "doi": null, + "isInfluential": true, + "paperId": "1451e7b11e66c86104f9391b80d9fb422fb11c01", + "title": "Image privacy protection with secure JPEG transmorphing", + "url": "https://www.semanticscholar.org/paper/1451e7b11e66c86104f9391b80d9fb422fb11c01", + "venue": "", + "year": 2017 + }, + { + "arxivId": null, + "authors": [ + { + "authorId": "3131569", + "name": "Haoxiang Li", + "url": "https://www.semanticscholar.org/author/3131569" + }, + { + "authorId": "1721019", + "name": "Jonathan Brandt", + "url": "https://www.semanticscholar.org/author/1721019" + }, + { + "authorId": "2012810", + "name": "Zhe L. Lin", + "url": "https://www.semanticscholar.org/author/2012810" + }, + { + "authorId": "1720987", + "name": "Xiaohui Shen", + "url": "https://www.semanticscholar.org/author/1720987" + }, + { + "authorId": "1745420", + "name": "Gang Hua", + "url": "https://www.semanticscholar.org/author/1745420" + } + ], + "doi": "10.1109/CVPR.2016.145", + "isInfluential": true, + "paperId": "1e1d7cbbef67e9e042a3a0a9a1bcefcc4a9adacf", + "title": "A Multi-level Contextual Model for Person Recognition in Photo Albums", + "url": "https://www.semanticscholar.org/paper/1e1d7cbbef67e9e042a3a0a9a1bcefcc4a9adacf", + "venue": "2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR)", + "year": 2016 + }, + { + "arxivId": null, + "authors": [ + { + "authorId": "47103251", + "name": "Yuansheng Xu", + "url": "https://www.semanticscholar.org/author/47103251" + }, + { + "authorId": "7402662", + "name": "Fangyue Peng", + "url": "https://www.semanticscholar.org/author/7402662" + }, + { + "authorId": "49521368", + "name": "Yu Yuan", + "url": "https://www.semanticscholar.org/author/49521368" + }, + { + "authorId": "36637369", + "name": "Yizhou Wang", + "url": "https://www.semanticscholar.org/author/36637369" + } + ], + "doi": "10.1109/ICASSP.2017.7952713", + "isInfluential": false, + "paperId": "d94d7ff6f46ad5cab5c20e6ac14c1de333711a0c", + "title": "Face Album: Towards automatic photo management based on person identity on mobile phones", + "url": "https://www.semanticscholar.org/paper/d94d7ff6f46ad5cab5c20e6ac14c1de333711a0c", + "venue": "2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)", + "year": 2017 + }, + { + "arxivId": null, + "authors": [ + { + "authorId": "9146911", + "name": "Shaghayegh Gharghabi", + "url": "https://www.semanticscholar.org/author/9146911" + }, + { + "authorId": "9368612", + "name": "Bita Azari", + "url": "https://www.semanticscholar.org/author/9368612" + }, + { + "authorId": "2319579", + "name": "Faraz Shamshirdar", + "url": "https://www.semanticscholar.org/author/2319579" + }, + { + "authorId": "1682051", + "name": "Reza Safabakhsh", + "url": "https://www.semanticscholar.org/author/1682051" + } + ], + "doi": null, + "isInfluential": false, + "paperId": "c8b9217ee36aebb9735e525b718490dc27c8c1cb", + "title": "Improving person recognition by weight adaptation of soft biometrics", + "url": "https://www.semanticscholar.org/paper/c8b9217ee36aebb9735e525b718490dc27c8c1cb", + "venue": "2016 6th International Conference on Computer and Knowledge Engineering (ICCKE)", + "year": 2016 + }, + { + "arxivId": "1707.06436", + "authors": [ + { + "authorId": "1730200", + "name": "Hirokatsu Kataoka", + "url": "https://www.semanticscholar.org/author/1730200" + }, + { + "authorId": "3393640", + "name": "Soma Shirakabe", + "url": "https://www.semanticscholar.org/author/3393640" + }, + { + "authorId": "1713046", + "name": "Yun He", + "url": "https://www.semanticscholar.org/author/1713046" + }, + { + "authorId": "9935341", + "name": "Shunya Ueta", + "url": "https://www.semanticscholar.org/author/9935341" + }, + { + "authorId": "5014206", + "name": "Teppei Suzuki", + "url": "https://www.semanticscholar.org/author/5014206" + }, + { + "authorId": "49897653", + "name": "Kaori Abe", + "url": "https://www.semanticscholar.org/author/49897653" + }, + { + "authorId": "2554424", + "name": "Asako Kanezaki", + "url": "https://www.semanticscholar.org/author/2554424" + }, + { + "authorId": "49133490", + "name": "Shinichiro Morita", + "url": "https://www.semanticscholar.org/author/49133490" + }, + { + "authorId": "22219521", + "name": "Toshiyuki Yabe", + "url": "https://www.semanticscholar.org/author/22219521" + }, + { + "authorId": "50544018", + "name": "Yoshihiro Kanehara", + "url": "https://www.semanticscholar.org/author/50544018" + }, + { + "authorId": "22174281", + "name": "Hiroya Yatsuyanagi", + "url": "https://www.semanticscholar.org/author/22174281" + }, + { + "authorId": "1692565", + "name": "Shinya Maruyama", + "url": "https://www.semanticscholar.org/author/1692565" + }, + { + "authorId": "10756539", + "name": "Ryousuke Takasawa", + "url": "https://www.semanticscholar.org/author/10756539" + }, + { + "authorId": "3217653", + "name": "Masataka Fuchida", + "url": "https://www.semanticscholar.org/author/3217653" + }, + { + "authorId": "2642022", + "name": "Yudai Miyashita", + "url": "https://www.semanticscholar.org/author/2642022" + }, + { + "authorId": "34935749", + "name": "Kazushige Okayasu", + "url": "https://www.semanticscholar.org/author/34935749" + }, + { + "authorId": "20505300", + "name": "Yuta Matsuzaki", + "url": "https://www.semanticscholar.org/author/20505300" + } + ], + "doi": null, + "isInfluential": false, + "paperId": "28cd46a078e8fad370b1aba34762a874374513a5", + "title": "cvpaper.challenge in 2016: Futuristic Computer Vision through 1, 600 Papers Survey", + "url": "https://www.semanticscholar.org/paper/28cd46a078e8fad370b1aba34762a874374513a5", + "venue": "ArXiv", + "year": 2017 + }, + { + "arxivId": null, + "authors": [ + { + "authorId": "2582568", + "name": "Yuhang Zhao", + "url": "https://www.semanticscholar.org/author/2582568" + }, + { + "authorId": "1968133", + "name": "Shaomei Wu", + "url": "https://www.semanticscholar.org/author/1968133" + }, + { + "authorId": "39685591", + "name": "Lindsay Reynolds", + "url": "https://www.semanticscholar.org/author/39685591" + }, + { + "authorId": "3283573", + "name": "Shiri Azenkot", + "url": "https://www.semanticscholar.org/author/3283573" + } + ], + "doi": "10.1145/3173574.3173789", + "isInfluential": false, + "paperId": "0aaf785d7f21d2b5ad582b456896495d30b0a4e2", + "title": "A Face Recognition Application for People with Visual Impairments: Understanding Use Beyond the Lab", + "url": "https://www.semanticscholar.org/paper/0aaf785d7f21d2b5ad582b456896495d30b0a4e2", + "venue": "CHI", + "year": 2018 + }, + { + "arxivId": null, + "authors": [ + { + "authorId": null, + "name": "Bouthour", + "url": null + }, + { + "authorId": null, + "name": "Khoubeib", + "url": null + }, + { + "authorId": "2488116", + "name": "Marcus S. Stefanou", + "url": "https://www.semanticscholar.org/author/2488116" + } + ], + "doi": null, + "isInfluential": true, + "paperId": "ef999ab2f7b37f46445a3457bf6c0f5fd7b5689d", + "title": "Improving face verification in photo albums by combining facial recognition and metadata with cross-matching", + "url": "https://www.semanticscholar.org/paper/ef999ab2f7b37f46445a3457bf6c0f5fd7b5689d", + "venue": "", + "year": 2016 + }, + { + "arxivId": null, + "authors": [ + { + "authorId": "11269472", + "name": "Iacopo Masi", + "url": "https://www.semanticscholar.org/author/11269472" + }, + { + "authorId": "1752756", + "name": "Feng-Ju Chang", + "url": "https://www.semanticscholar.org/author/1752756" + }, + { + "authorId": "49331459", + "name": "Jongmoo Choi", + "url": "https://www.semanticscholar.org/author/49331459" + }, + { + "authorId": "35840854", + "name": "Shai Harel", + "url": "https://www.semanticscholar.org/author/35840854" + }, + { + "authorId": "5911467", + "name": "Jungyeon Kim", + "url": "https://www.semanticscholar.org/author/5911467" + }, + { + "authorId": "2792633", + "name": "KangGeon Kim", + "url": "https://www.semanticscholar.org/author/2792633" + }, + { + "authorId": "2955822", + "name": "Jatuporn Toy Leksut", + "url": "https://www.semanticscholar.org/author/2955822" + }, + { + "authorId": "38696444", + "name": "Stephen Rawls", + "url": "https://www.semanticscholar.org/author/38696444" + }, + { + "authorId": "46220641", + "name": "Yue Wu", + "url": "https://www.semanticscholar.org/author/46220641" + }, + { + "authorId": "1756099", + "name": "Tal Hassner", + "url": "https://www.semanticscholar.org/author/1756099" + }, + { + "authorId": "17806729", + "name": "Wael AbdAlmageed", + "url": "https://www.semanticscholar.org/author/17806729" + }, + { + "authorId": "3463966", + "name": "Gérard G. Medioni", + "url": "https://www.semanticscholar.org/author/3463966" + }, + { + "authorId": "49933077", + "name": "Louis-Philippe Morency", + "url": "https://www.semanticscholar.org/author/49933077" + }, + { + "authorId": "40155735", + "name": "Prem Natarajan", + "url": "https://www.semanticscholar.org/author/40155735" + }, + { + "authorId": "51061892", + "name": "Ramkant Nevatia", + "url": "https://www.semanticscholar.org/author/51061892" + } + ], + "doi": "10.1109/TPAMI.2018.2792452", + "isInfluential": true, + "paperId": "6f22628d34a486d73c6b46eb071200a00e3abae3", + "title": "Learning Pose-Aware Models for Pose-Invariant Face Recognition in the Wild.", + "url": "https://www.semanticscholar.org/paper/6f22628d34a486d73c6b46eb071200a00e3abae3", + "venue": "IEEE transactions on pattern analysis and machine intelligence", + "year": 2018 + }, + { + "arxivId": "1807.10510", + "authors": [ + { + "authorId": "39360892", + "name": "Qingqiu Huang", + "url": "https://www.semanticscholar.org/author/39360892" + }, + { + "authorId": "40584026", + "name": "Wentao Liu", + "url": "https://www.semanticscholar.org/author/40584026" + }, + { + "authorId": "1807606", + "name": "Dahua Lin", + "url": "https://www.semanticscholar.org/author/1807606" + } + ], + "doi": null, + "isInfluential": true, + "paperId": "c97a5f2241cc6cd99ef0c4527ea507a50841f60b", + "title": "Person Search in Videos with One Portrait Through Visual and Temporal Links", + "url": "https://www.semanticscholar.org/paper/c97a5f2241cc6cd99ef0c4527ea507a50841f60b", + "venue": "ArXiv", + "year": 2018 + }, + { + "arxivId": "1705.10120", + "authors": [ + { + "authorId": "37956314", + "name": "Vijay Kumar", + "url": "https://www.semanticscholar.org/author/37956314" + }, + { + "authorId": "3185334", + "name": "Anoop M. Namboodiri", + "url": "https://www.semanticscholar.org/author/3185334" + }, + { + "authorId": "2210374", + "name": "Manohar Paluri", + "url": "https://www.semanticscholar.org/author/2210374" + }, + { + "authorId": "1694502", + "name": "C. V. Jawahar", + "url": "https://www.semanticscholar.org/author/1694502" + } + ], + "doi": "10.1109/CVPR.2017.719", + "isInfluential": false, + "paperId": "a3d8b5622c4b9af1f753aade57e4774730787a00", + "title": "Pose-Aware Person Recognition", + "url": "https://www.semanticscholar.org/paper/a3d8b5622c4b9af1f753aade57e4774730787a00", + "venue": "2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR)", + "year": 2017 + }, + { + "arxivId": "1702.06890", + "authors": [ + { + "authorId": "1715752", + "name": "Yu Liu", + "url": "https://www.semanticscholar.org/author/1715752" + }, + { + "authorId": "46382329", + "name": "Hongyang Li", + "url": "https://www.semanticscholar.org/author/46382329" + }, + { + "authorId": "31843833", + "name": "Xiaogang Wang", + "url": "https://www.semanticscholar.org/author/31843833" + } + ], + "doi": null, + "isInfluential": true, + "paperId": "6fed504da4e192fe4c2d452754d23d3db4a4e5e3", + "title": "Learning Deep Features via Congenerous Cosine Loss for Person Recognition", + "url": "https://www.semanticscholar.org/paper/6fed504da4e192fe4c2d452754d23d3db4a4e5e3", + "venue": "ArXiv", + "year": 2017 + }, + { + "arxivId": null, + "authors": [ + { + "authorId": "49410785", + "name": "Ning Zhang", + "url": "https://www.semanticscholar.org/author/49410785" + } + ], + "doi": null, + "isInfluential": false, + "paperId": "d6a9ea9b40a7377c91c705f4c7f206a669a9eea2", + "title": "Visual Representations for Fine-grained Categorization", + "url": "https://www.semanticscholar.org/paper/d6a9ea9b40a7377c91c705f4c7f206a669a9eea2", + "venue": "", + "year": 2015 + }, + { + "arxivId": "1604.02531", + "authors": [ + { + "authorId": "50853180", + "name": "Liang Zheng", + "url": "https://www.semanticscholar.org/author/50853180" + }, + { + "authorId": "1983351", + "name": "Hengheng Zhang", + "url": "https://www.semanticscholar.org/author/1983351" + }, + { + "authorId": "3141359", + "name": "Shaoyan Sun", + "url": "https://www.semanticscholar.org/author/3141359" + }, + { + "authorId": "2099305", + "name": "Manmohan Krishna Chandraker", + "url": "https://www.semanticscholar.org/author/2099305" + }, + { + "authorId": "50842217", + "name": "Qi Tian", + "url": "https://www.semanticscholar.org/author/50842217" + } + ], + "doi": "10.1109/CVPR.2017.357", + "isInfluential": false, + "paperId": "0b84f07af44f964817675ad961def8a51406dd2e", + "title": "Person Re-identification in the Wild", + "url": "https://www.semanticscholar.org/paper/0b84f07af44f964817675ad961def8a51406dd2e", + "venue": "2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR)", + "year": 2017 + }, + { + "arxivId": "1809.09189", + "authors": [ + { + "authorId": "51264689", + "name": "Sina Mokhtarzadeh Azar", + "url": "https://www.semanticscholar.org/author/51264689" + }, + { + "authorId": "51268224", + "name": "Sajjad Azami", + "url": "https://www.semanticscholar.org/author/51268224" + }, + { + "authorId": "51388019", + "name": "Mina Ghadimi Atigh", + "url": "https://www.semanticscholar.org/author/51388019" + }, + { + "authorId": "1826726", + "name": "Mohammad Javadi", + "url": "https://www.semanticscholar.org/author/1826726" + }, + { + "authorId": "1780566", + "name": "Ahmad Nickabadi", + "url": "https://www.semanticscholar.org/author/1780566" + } + ], + "doi": null, + "isInfluential": true, + "paperId": "be393cd567b338da6ed60181c8ad429627578a31", + "title": "Zoom-RNN: A Novel Method for Person Recognition Using Recurrent Neural Networks", + "url": "https://www.semanticscholar.org/paper/be393cd567b338da6ed60181c8ad429627578a31", + "venue": "", + "year": 2018 + }, + { + "arxivId": null, + "authors": [ + { + "authorId": "1956609", + "name": "Vassilios Vonikakis", + "url": "https://www.semanticscholar.org/author/1956609" + }, + { + "authorId": "48236457", + "name": "Ramanathan Subramanian", + "url": "https://www.semanticscholar.org/author/48236457" + }, + { + "authorId": "49530400", + "name": "Jonas Toft Arnfred", + "url": "https://www.semanticscholar.org/author/49530400" + }, + { + "authorId": "1702224", + "name": "Stefan Winkler", + "url": "https://www.semanticscholar.org/author/1702224" + } + ], + "doi": "10.1109/TMM.2017.2699859", + "isInfluential": false, + "paperId": "eb8a3948c4be0d23eb7326d27f2271be893b3409", + "title": "A Probabilistic Approach to People-Centric Photo Selection and Sequencing", + "url": "https://www.semanticscholar.org/paper/eb8a3948c4be0d23eb7326d27f2271be893b3409", + "venue": "IEEE Transactions on Multimedia", + "year": 2017 + }, + { + "arxivId": "1801.10442", + "authors": [ + { + "authorId": "19263506", + "name": "Arsha Nagrani", + "url": "https://www.semanticscholar.org/author/19263506" + }, + { + "authorId": "1688869", + "name": "Andrew Zisserman", + "url": "https://www.semanticscholar.org/author/1688869" + } + ], + "doi": null, + "isInfluential": false, + "paperId": "ff1f45bdad41d8b35435098041e009627e60d208", + "title": "From Benedict Cumberbatch to Sherlock Holmes: Character Identification in TV series without a Script", + "url": "https://www.semanticscholar.org/paper/ff1f45bdad41d8b35435098041e009627e60d208", + "venue": "BMVC", + "year": 2017 + }, + { + "arxivId": "1605.08247", + "authors": [ + { + "authorId": "1730200", + "name": "Hirokatsu Kataoka", + "url": "https://www.semanticscholar.org/author/1730200" + }, + { + "authorId": "2642022", + "name": "Yudai Miyashita", + "url": "https://www.semanticscholar.org/author/2642022" + }, + { + "authorId": "3108668", + "name": "Tomoaki K. Yamabe", + "url": "https://www.semanticscholar.org/author/3108668" + }, + { + "authorId": "3393640", + "name": "Soma Shirakabe", + "url": "https://www.semanticscholar.org/author/3393640" + }, + { + "authorId": "46427523", + "name": "Shin-ichi Sato", + "url": "https://www.semanticscholar.org/author/46427523" + }, + { + "authorId": "29998543", + "name": "Hironori Hoshino", + "url": "https://www.semanticscholar.org/author/29998543" + }, + { + "authorId": "2348426", + "name": "Ryo Kato", + "url": "https://www.semanticscholar.org/author/2348426" + }, + { + "authorId": "49897653", + "name": "Kaori Abe", + "url": "https://www.semanticscholar.org/author/49897653" + }, + { + "authorId": "3407486", + "name": "Takaaki Imanari", + "url": "https://www.semanticscholar.org/author/3407486" + }, + { + "authorId": "26851746", + "name": "Naomichi Kobayashi", + "url": "https://www.semanticscholar.org/author/26851746" + }, + { + "authorId": "49133490", + "name": "Shinichiro Morita", + "url": "https://www.semanticscholar.org/author/49133490" + }, + { + "authorId": "2462801", + "name": "Akio Nakamura", + "url": "https://www.semanticscholar.org/author/2462801" + } + ], + "doi": null, + "isInfluential": false, + "paperId": "2f7e9b45255c9029d2ae97bbb004d6072e70fa79", + "title": "cvpaper.challenge in 2015 - A review of CVPR2015 and DeepSurvey", + "url": "https://www.semanticscholar.org/paper/2f7e9b45255c9029d2ae97bbb004d6072e70fa79", + "venue": "ArXiv", + "year": 2016 + }, + { + "arxivId": "1807.00504", + "authors": [ + { + "authorId": "29988001", + "name": "Zhouxia Wang", + "url": "https://www.semanticscholar.org/author/29988001" + }, + { + "authorId": "1765674", + "name": "Tianshui Chen", + "url": "https://www.semanticscholar.org/author/1765674" + }, + { + "authorId": "1723599", + "name": "Jimmy S. J. Ren", + "url": "https://www.semanticscholar.org/author/1723599" + }, + { + "authorId": "12254824", + "name": "Weihao Yu", + "url": "https://www.semanticscholar.org/author/12254824" + }, + { + "authorId": "47413456", + "name": "Hui Cheng", + "url": "https://www.semanticscholar.org/author/47413456" + }, + { + "authorId": "1737218", + "name": "Liang Lin", + "url": "https://www.semanticscholar.org/author/1737218" + } + ], + "doi": "10.24963/ijcai.2018/142", + "isInfluential": true, + "paperId": "725c3605c2d26d113637097358cd4c08c19ff9e1", + "title": "Deep Reasoning with Knowledge Graph for Social Relationship Understanding", + "url": "https://www.semanticscholar.org/paper/725c3605c2d26d113637097358cd4c08c19ff9e1", + "venue": "IJCAI", + "year": 2018 + }, + { + "arxivId": "1805.01515", + "authors": [ + { + "authorId": "2582568", + "name": "Yuhang Zhao", + "url": "https://www.semanticscholar.org/author/2582568" + }, + { + "authorId": "1968133", + "name": "Shaomei Wu", + "url": "https://www.semanticscholar.org/author/1968133" + }, + { + "authorId": "39685591", + "name": "Lindsay Reynolds", + "url": "https://www.semanticscholar.org/author/39685591" + }, + { + "authorId": "3283573", + "name": "Shiri Azenkot", + "url": "https://www.semanticscholar.org/author/3283573" + } + ], + "doi": "10.1145/3134756", + "isInfluential": false, + "paperId": "2a5903bdb3fdfb4d51f70b77f16852df3b8e5f83", + "title": "The Effect of Computer-Generated Descriptions on Photo-Sharing Experiences of People with Visual Impairments", + "url": "https://www.semanticscholar.org/paper/2a5903bdb3fdfb4d51f70b77f16852df3b8e5f83", + "venue": "PACMHCI", + "year": 2017 + }, + { + "arxivId": "1704.06456", + "authors": [ + { + "authorId": "32222907", + "name": "Qianru Sun", + "url": "https://www.semanticscholar.org/author/32222907" + }, + { + "authorId": "1697100", + "name": "Bernt Schiele", + "url": "https://www.semanticscholar.org/author/1697100" + }, + { + "authorId": "1739548", + "name": "Mario Fritz", + "url": "https://www.semanticscholar.org/author/1739548" + } + ], + "doi": "10.1109/CVPR.2017.54", + "isInfluential": true, + "paperId": "23429ef60e7a9c0e2f4d81ed1b4e47cc2616522f", + "title": "A Domain Based Approach to Social Relation Recognition", + "url": "https://www.semanticscholar.org/paper/23429ef60e7a9c0e2f4d81ed1b4e47cc2616522f", + "venue": "2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR)", + "year": 2017 + }, + { + "arxivId": "1711.09001", + "authors": [ + { + "authorId": "32222907", + "name": "Qianru Sun", + "url": "https://www.semanticscholar.org/author/32222907" + }, + { + "authorId": "1847145", + "name": "Liqian Ma", + "url": "https://www.semanticscholar.org/author/1847145" + }, + { + "authorId": "2390510", + "name": "Seong Joon Oh", + "url": "https://www.semanticscholar.org/author/2390510" + }, + { + "authorId": "1681236", + "name": "Luc Van Gool", + "url": "https://www.semanticscholar.org/author/1681236" + }, + { + "authorId": "1697100", + "name": "Bernt Schiele", + "url": "https://www.semanticscholar.org/author/1697100" + }, + { + "authorId": "1739548", + "name": "Mario Fritz", + "url": "https://www.semanticscholar.org/author/1739548" + } + ], + "doi": null, + "isInfluential": false, + "paperId": "ba7b12c8e2ff3c5e4e0f70b58215b41b18ff8feb", + "title": "Natural and Effective Obfuscation by Head Inpainting", + "url": "https://www.semanticscholar.org/paper/ba7b12c8e2ff3c5e4e0f70b58215b41b18ff8feb", + "venue": "ArXiv", + "year": 2017 + }, + { + "arxivId": null, + "authors": [ + { + "authorId": "2390510", + "name": "Seong Joon Oh", + "url": "https://www.semanticscholar.org/author/2390510" + }, + { + "authorId": "1739548", + "name": "Mario Fritz", + "url": "https://www.semanticscholar.org/author/1739548" + }, + { + "authorId": "1697100", + "name": "Bernt Schiele", + "url": "https://www.semanticscholar.org/author/1697100" + }, + { + "authorId": "5514293", + "name": "J Max", + "url": "https://www.semanticscholar.org/author/5514293" + } + ], + "doi": null, + "isInfluential": true, + "paperId": "808b685d09912cbef4a009e74e10476304b4cccf", + "title": "From Understanding to Controlling Privacy against Automatic Person Recognition in Social Media", + "url": "https://www.semanticscholar.org/paper/808b685d09912cbef4a009e74e10476304b4cccf", + "venue": "", + "year": 2017 + }, + { + "arxivId": "1406.4444", + "authors": [ + { + "authorId": "7969330", + "name": "Ziming Zhang", + "url": "https://www.semanticscholar.org/author/7969330" + }, + { + "authorId": "1699322", + "name": "Venkatesh Saligrama", + "url": "https://www.semanticscholar.org/author/1699322" + } + ], + "doi": "10.1109/TCSVT.2016.2596159", + "isInfluential": false, + "paperId": "ae936628e78db4edb8e66853f59433b8cc83594f", + "title": "PRISM: Person Reidentification via Structured Matching", + "url": "https://www.semanticscholar.org/paper/ae936628e78db4edb8e66853f59433b8cc83594f", + "venue": "IEEE Transactions on Circuits and Systems for Video Technology", + "year": 2017 + }, + { + "arxivId": null, + "authors": [ + { + "authorId": "1927977", + "name": "Tousif Ahmed", + "url": "https://www.semanticscholar.org/author/1927977" + }, + { + "authorId": "1996617", + "name": "Apu Kapadia", + "url": "https://www.semanticscholar.org/author/1996617" + }, + { + "authorId": "32858206", + "name": "Venkatesh Potluri", + "url": "https://www.semanticscholar.org/author/32858206" + }, + { + "authorId": "30576065", + "name": "Manohar Swaminathan", + "url": "https://www.semanticscholar.org/author/30576065" + } + ], + "doi": "10.1145/3264899", + "isInfluential": false, + "paperId": "ca096e158912080493a898b0b8a4bd2902674fed", + "title": "Up to a Limit?: Privacy Concerns of Bystanders and Their Willingness to Share Additional Information with Visually Impaired Users of Assistive Technologies", + "url": "https://www.semanticscholar.org/paper/ca096e158912080493a898b0b8a4bd2902674fed", + "venue": "IMWUT", + "year": 2018 + }, + { + "arxivId": null, + "authors": [ + { + "authorId": null, + "name": "Micheal Cogswell", + "url": null + }, + { + "authorId": "37824829", + "name": "Michael Cogswell", + "url": "https://www.semanticscholar.org/author/37824829" + } + ], + "doi": null, + "isInfluential": false, + "paperId": "6d8eef8f8d6cd8436c55018e6ca5c5907b31ac19", + "title": "Understanding Representations and Reducing their Redundancy in Deep Networks", + "url": "https://www.semanticscholar.org/paper/6d8eef8f8d6cd8436c55018e6ca5c5907b31ac19", + "venue": "", + "year": 2016 + }, + { + "arxivId": "1703.09471", + "authors": [ + { + "authorId": "2390510", + "name": "Seong Joon Oh", + "url": "https://www.semanticscholar.org/author/2390510" + }, + { + "authorId": "1739548", + "name": "Mario Fritz", + "url": "https://www.semanticscholar.org/author/1739548" + }, + { + "authorId": "1697100", + "name": "Bernt Schiele", + "url": "https://www.semanticscholar.org/author/1697100" + } + ], + "doi": "10.1109/ICCV.2017.165", + "isInfluential": true, + "paperId": "b68150bfdec373ed8e025f448b7a3485c16e3201", + "title": "Adversarial Image Perturbation for Privacy Protection A Game Theory Perspective", + "url": "https://www.semanticscholar.org/paper/b68150bfdec373ed8e025f448b7a3485c16e3201", + "venue": "2017 IEEE International Conference on Computer Vision (ICCV)", + "year": 2017 + }, + { + "arxivId": "1609.00408", + "authors": [ + { + "authorId": "49702377", + "name": "Richard McPherson", + "url": "https://www.semanticscholar.org/author/49702377" + }, + { + "authorId": "2520493", + "name": "Reza Shokri", + "url": "https://www.semanticscholar.org/author/2520493" + }, + { + "authorId": "1723945", + "name": "Vitaly Shmatikov", + "url": "https://www.semanticscholar.org/author/1723945" + } + ], + "doi": null, + "isInfluential": false, + "paperId": "3c57e28a4eb463d532ea2b0b1ba4b426ead8d9a0", + "title": "Defeating Image Obfuscation with Deep Learning", + "url": "https://www.semanticscholar.org/paper/3c57e28a4eb463d532ea2b0b1ba4b426ead8d9a0", + "venue": "ArXiv", + "year": 2016 + }, + { + "arxivId": "1804.03287", + "authors": [ + { + "authorId": "46509484", + "name": "Jian Zhao", + "url": "https://www.semanticscholar.org/author/46509484" + }, + { + "authorId": "2757639", + "name": "Jianshu Li", + "url": "https://www.semanticscholar.org/author/2757639" + }, + { + "authorId": "47585344", + "name": "Yu Cheng", + "url": "https://www.semanticscholar.org/author/47585344" + }, + { + "authorId": "48207454", + "name": "Li Zhou", + "url": "https://www.semanticscholar.org/author/48207454" + }, + { + "authorId": "1715286", + "name": "Terence Sim", + "url": "https://www.semanticscholar.org/author/1715286" + }, + { + "authorId": "1698982", + "name": "Shuicheng Yan", + "url": "https://www.semanticscholar.org/author/1698982" + }, + { + "authorId": "33221685", + "name": "Jiashi Feng", + "url": "https://www.semanticscholar.org/author/33221685" + } + ], + "doi": "10.13140/RG.2.2.23242.67523", + "isInfluential": false, + "paperId": "5f771fed91c8e4b666489ba2384d0705bcf75030", + "title": "Understanding Humans in Crowded Scenes: Deep Nested Adversarial Learning and A New Benchmark for Multi-Human Parsing", + "url": "https://www.semanticscholar.org/paper/5f771fed91c8e4b666489ba2384d0705bcf75030", + "venue": "ArXiv", + "year": 2018 + }, + { + "arxivId": null, + "authors": [ + { + "authorId": "37956314", + "name": "Vijay Kumar", + "url": "https://www.semanticscholar.org/author/37956314" + }, + { + "authorId": "3185334", + "name": "Anoop M. Namboodiri", + "url": "https://www.semanticscholar.org/author/3185334" + }, + { + "authorId": "1694502", + "name": "C. V. Jawahar", + "url": "https://www.semanticscholar.org/author/1694502" + } + ], + "doi": "10.1007/s11760-017-1140-5", + "isInfluential": false, + "paperId": "01e27c91c7cef926389f913d12410725e7dd35ab", + "title": "Semi-supervised annotation of faces in image collection", + "url": "https://www.semanticscholar.org/paper/01e27c91c7cef926389f913d12410725e7dd35ab", + "venue": "Signal, Image and Video Processing", + "year": 2018 + }, + { + "arxivId": "1710.03224", + "authors": [ + { + "authorId": "2390510", + "name": "Seong Joon Oh", + "url": "https://www.semanticscholar.org/author/2390510" + }, + { + "authorId": "1798000", + "name": "Rodrigo Benenson", + "url": "https://www.semanticscholar.org/author/1798000" + }, + { + "authorId": "1739548", + "name": "Mario Fritz", + "url": "https://www.semanticscholar.org/author/1739548" + }, + { + "authorId": "1697100", + "name": "Bernt Schiele", + "url": "https://www.semanticscholar.org/author/1697100" + } + ], + "doi": null, + "isInfluential": true, + "paperId": "cfd4004054399f3a5f536df71f9b9987f060f434", + "title": "Person Recognition in Social Media Photos", + "url": "https://www.semanticscholar.org/paper/cfd4004054399f3a5f536df71f9b9987f060f434", + "venue": "ArXiv", + "year": 2017 + }, + { + "arxivId": null, + "authors": [ + { + "authorId": "11269472", + "name": "Iacopo Masi", + "url": "https://www.semanticscholar.org/author/11269472" + }, + { + "authorId": "38696444", + "name": "Stephen Rawls", + "url": "https://www.semanticscholar.org/author/38696444" + }, + { + "authorId": "3463966", + "name": "Gérard G. Medioni", + "url": "https://www.semanticscholar.org/author/3463966" + }, + { + "authorId": "1776379", + "name": "Premkumar Natarajan", + "url": "https://www.semanticscholar.org/author/1776379" + } + ], + "doi": "10.1109/CVPR.2016.523", + "isInfluential": true, + "paperId": "2c92839418a64728438c351a42f6dc5ad0c6e686", + "title": "Pose-Aware Face Recognition in the Wild", + "url": "https://www.semanticscholar.org/paper/2c92839418a64728438c351a42f6dc5ad0c6e686", + "venue": "2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR)", + "year": 2016 + }, + { + "arxivId": "1710.00870", + "authors": [ + { + "authorId": "1715752", + "name": "Yu Liu", + "url": "https://www.semanticscholar.org/author/1715752" + }, + { + "authorId": "46382329", + "name": "Hongyang Li", + "url": "https://www.semanticscholar.org/author/46382329" + }, + { + "authorId": "31843833", + "name": "Xiaogang Wang", + "url": "https://www.semanticscholar.org/author/31843833" + } + ], + "doi": null, + "isInfluential": true, + "paperId": "d949fadc9b6c5c8b067fa42265ad30945f9caa99", + "title": "Rethinking Feature Discrimination and Polymerization for Large-scale Recognition", + "url": "https://www.semanticscholar.org/paper/d949fadc9b6c5c8b067fa42265ad30945f9caa99", + "venue": "ArXiv", + "year": 2017 + }, + { + "arxivId": "1611.09967", + "authors": [ + { + "authorId": "48513733", + "name": "Yao Li", + "url": "https://www.semanticscholar.org/author/48513733" + }, + { + "authorId": "2604251", + "name": "Guosheng Lin", + "url": "https://www.semanticscholar.org/author/2604251" + }, + { + "authorId": "3194022", + "name": "Bohan Zhuang", + "url": "https://www.semanticscholar.org/author/3194022" + }, + { + "authorId": "2161037", + "name": "Lingqiao Liu", + "url": "https://www.semanticscholar.org/author/2161037" + }, + { + "authorId": "1780381", + "name": "Chunhua Shen", + "url": "https://www.semanticscholar.org/author/1780381" + }, + { + "authorId": "5546141", + "name": "Anton van den Hengel", + "url": "https://www.semanticscholar.org/author/5546141" + } + ], + "doi": "10.1109/CVPR.2017.600", + "isInfluential": true, + "paperId": "3d24b386d003bee176a942c26336dbe8f427aadd", + "title": "Sequential Person Recognition in Photo Albums with a Recurrent Network", + "url": "https://www.semanticscholar.org/paper/3d24b386d003bee176a942c26336dbe8f427aadd", + "venue": "2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR)", + "year": 2017 + }, + { + "arxivId": "1805.04049", + "authors": [ + { + "authorId": "2008164", + "name": "Luca Melis", + "url": "https://www.semanticscholar.org/author/2008164" + }, + { + "authorId": "3469125", + "name": "Congzheng Song", + "url": "https://www.semanticscholar.org/author/3469125" + }, + { + "authorId": "1728207", + "name": "Emiliano De Cristofaro", + "url": "https://www.semanticscholar.org/author/1728207" + }, + { + "authorId": "1723945", + "name": "Vitaly Shmatikov", + "url": "https://www.semanticscholar.org/author/1723945" + } + ], + "doi": null, + "isInfluential": true, + "paperId": "8bdf6f03bde08c424c214188b35be8b2dec7cdea", + "title": "Inference Attacks Against Collaborative Learning", + "url": "https://www.semanticscholar.org/paper/8bdf6f03bde08c424c214188b35be8b2dec7cdea", + "venue": "ArXiv", + "year": 2018 + } + ], + "doi": "10.1109/CVPR.2015.7299113", + "influentialCitationCount": 19, + "paperId": "0a85bdff552615643dd74646ac881862a7c7072d", + "references": [ + { + "arxivId": null, + "authors": [ + { + "authorId": "2064160", + "name": "Alex Krizhevsky", + "url": "https://www.semanticscholar.org/author/2064160" + }, + { + "authorId": "1701686", + "name": "Ilya Sutskever", + "url": "https://www.semanticscholar.org/author/1701686" + }, + { + "authorId": "1695689", + "name": "Geoffrey E. Hinton", + "url": "https://www.semanticscholar.org/author/1695689" + } + ], + "doi": "10.1145/3065386", + "isInfluential": false, + "paperId": "2315fc6c2c0c4abd2443e26a26e7bb86df8e24cc", + "title": "ImageNet Classification with Deep Convolutional Neural Networks", + "url": "https://www.semanticscholar.org/paper/2315fc6c2c0c4abd2443e26a26e7bb86df8e24cc", + "venue": "NIPS", + "year": 2012 + }, + { + "arxivId": null, + "authors": [ + { + "authorId": "38511654", + "name": "Douglas Gray", + "url": "https://www.semanticscholar.org/author/38511654" + }, + { + "authorId": "2491194", + "name": "Hai Tao", + "url": "https://www.semanticscholar.org/author/2491194" + } + ], + "doi": "10.1007/978-3-540-88682-2_21", + "isInfluential": false, + "paperId": "d745cf8c51032996b5fee6b19e1b5321c14797eb", + "title": "Viewpoint Invariant Pedestrian Recognition with an Ensemble of Localized Features", + "url": "https://www.semanticscholar.org/paper/d745cf8c51032996b5fee6b19e1b5321c14797eb", + "venue": "ECCV", + "year": 2008 + }, + { + "arxivId": null, + "authors": [ + { + "authorId": "48625148", + "name": "Wei Li", + "url": "https://www.semanticscholar.org/author/48625148" + }, + { + "authorId": "49832825", + "name": "Rui Zhao", + "url": "https://www.semanticscholar.org/author/49832825" + }, + { + "authorId": "31843833", + "name": "Xiaogang Wang", + "url": "https://www.semanticscholar.org/author/31843833" + } + ], + "doi": "10.1007/978-3-642-37331-2_3", + "isInfluential": false, + "paperId": "44484d2866f222bbb9b6b0870890f9eea1ffb2d0", + "title": "Human Reidentification with Transferred Metric Learning", + "url": "https://www.semanticscholar.org/paper/44484d2866f222bbb9b6b0870890f9eea1ffb2d0", + "venue": "ACCV", + "year": 2012 + }, + { + "arxivId": null, + "authors": [ + { + "authorId": "1931707", + "name": "Martin Bäuml", + "url": "https://www.semanticscholar.org/author/1931707" + }, + { + "authorId": "2103464", + "name": "Makarand Tapaswi", + "url": "https://www.semanticscholar.org/author/2103464" + }, + { + "authorId": "1742325", + "name": "Rainer Stiefelhagen", + "url": "https://www.semanticscholar.org/author/1742325" + } + ], + "doi": "10.1109/CVPR.2013.462", + "isInfluential": false, + "paperId": "2b743e9a1aa638f46f2842187136b0e32e3bc042", + "title": "Semi-supervised Learning with Constraints for Person Identification in Multimedia Data", + "url": "https://www.semanticscholar.org/paper/2b743e9a1aa638f46f2842187136b0e32e3bc042", + "venue": "2013 IEEE Conference on Computer Vision and Pattern Recognition", + "year": 2013 + }, + { + "arxivId": null, + "authors": [ + { + "authorId": "1687465", + "name": "Mor Naaman", + "url": "https://www.semanticscholar.org/author/1687465" + }, + { + "authorId": "2105696", + "name": "Ron B. Yeh", + "url": "https://www.semanticscholar.org/author/2105696" + }, + { + "authorId": "1695250", + "name": "Hector Garcia-Molina", + "url": "https://www.semanticscholar.org/author/1695250" + }, + { + "authorId": "1750481", + "name": "Andreas Paepcke", + "url": "https://www.semanticscholar.org/author/1750481" + } + ], + "doi": "10.1145/1065385.1065430", + "isInfluential": false, + "paperId": "45467c70c1857b68902acebe6086d8d455f69f1d", + "title": "Leveraging context to resolve identity in photo albums", + "url": "https://www.semanticscholar.org/paper/45467c70c1857b68902acebe6086d8d455f69f1d", + "venue": "Proceedings of the 5th ACM/IEEE-CS Joint Conference on Digital Libraries (JCDL '05)", + "year": 2005 + }, + { + "arxivId": "1311.2524", + "authors": [ + { + "authorId": "2983898", + "name": "Ross B. Girshick", + "url": "https://www.semanticscholar.org/author/2983898" + }, + { + "authorId": "7408951", + "name": "Jeff Donahue", + "url": "https://www.semanticscholar.org/author/7408951" + }, + { + "authorId": "1753210", + "name": "Trevor Darrell", + "url": "https://www.semanticscholar.org/author/1753210" + }, + { + "authorId": "1689212", + "name": "Jitendra Malik", + "url": "https://www.semanticscholar.org/author/1689212" + } + ], + "doi": "10.1109/CVPR.2014.81", + "isInfluential": false, + "paperId": "009fba8df6bbca155d9e070a9bd8d0959bc693c2", + "title": "Rich Feature Hierarchies for Accurate Object Detection and Semantic Segmentation", + "url": "https://www.semanticscholar.org/paper/009fba8df6bbca155d9e070a9bd8d0959bc693c2", + "venue": "2014 IEEE Conference on Computer Vision and Pattern Recognition", + "year": 2014 + }, + { + "arxivId": null, + "authors": [ + { + "authorId": "1769383", + "name": "Lubomir D. Bourdev", + "url": "https://www.semanticscholar.org/author/1769383" + }, + { + "authorId": "35208858", + "name": "Subhransu Maji", + "url": "https://www.semanticscholar.org/author/35208858" + }, + { + "authorId": "1689212", + "name": "Jitendra Malik", + "url": "https://www.semanticscholar.org/author/1689212" + } + ], + "doi": "10.1109/ICCV.2011.6126413", + "isInfluential": false, + "paperId": "7808937b46acad36e43c30ae4e9f3fd57462853d", + "title": "Describing people: A poselet-based approach to attribute classification", + "url": "https://www.semanticscholar.org/paper/7808937b46acad36e43c30ae4e9f3fd57462853d", + "venue": "2011 International Conference on Computer Vision", + "year": 2011 + }, + { + "arxivId": null, + "authors": [ + { + "authorId": "1807606", + "name": "Dahua Lin", + "url": "https://www.semanticscholar.org/author/1807606" + }, + { + "authorId": "2189118", + "name": "Ashish Kapoor", + "url": "https://www.semanticscholar.org/author/2189118" + }, + { + "authorId": "1745420", + "name": "Gang Hua", + "url": "https://www.semanticscholar.org/author/1745420" + }, + { + "authorId": "40039594", + "name": "Simon Baker", + "url": "https://www.semanticscholar.org/author/40039594" + } + ], + "doi": "10.1007/978-3-642-15549-9_18", + "isInfluential": false, + "paperId": "01df75d2020931ad30f827653cdde7fba06dda5f", + "title": "Joint People, Event, and Location Recognition in Personal Photo Collections Using Cross-Domain Context", + "url": "https://www.semanticscholar.org/paper/01df75d2020931ad30f827653cdde7fba06dda5f", + "venue": "ECCV", + "year": 2010 + }, + { + "arxivId": null, + "authors": [ + { + "authorId": "10338111", + "name": "Zhen Cui", + "url": "https://www.semanticscholar.org/author/10338111" + }, + { + "authorId": "50135099", + "name": "Wen Li", + "url": "https://www.semanticscholar.org/author/50135099" + }, + { + "authorId": "38188040", + "name": "Dong Xu", + "url": "https://www.semanticscholar.org/author/38188040" + }, + { + "authorId": "1685914", + "name": "Shiguang Shan", + "url": "https://www.semanticscholar.org/author/1685914" + }, + { + "authorId": "1710220", + "name": "Xilin Chen", + "url": "https://www.semanticscholar.org/author/1710220" + } + ], + "doi": "10.1109/CVPR.2013.456", + "isInfluential": false, + "paperId": "316d51aaa37891d730ffded7b9d42946abea837f", + "title": "Fusing Robust Face Region Descriptors via Multiple Metric Learning for Face Recognition in the Wild", + "url": "https://www.semanticscholar.org/paper/316d51aaa37891d730ffded7b9d42946abea837f", + "venue": "2013 IEEE Conference on Computer Vision and Pattern Recognition", + "year": 2013 + }, + { + "arxivId": null, + "authors": [ + { + "authorId": "47583369", + "name": "John Wright", + "url": "https://www.semanticscholar.org/author/47583369" + }, + { + "authorId": "2784161", + "name": "Allen Y. Yang", + "url": "https://www.semanticscholar.org/author/2784161" + }, + { + "authorId": "1701028", + "name": "Arvind Ganesh", + "url": "https://www.semanticscholar.org/author/1701028" + }, + { + "authorId": "1717598", + "name": "S. Shankar Sastry", + "url": "https://www.semanticscholar.org/author/1717598" + }, + { + "authorId": "50032052", + "name": "Yi Ma", + "url": "https://www.semanticscholar.org/author/50032052" + } + ], + "doi": "10.1109/TPAMI.2008.79", + "isInfluential": false, + "paperId": "1512b9570669a9dfe2a252900ae94b276e3aff9e", + "title": "Robust Face Recognition via Sparse Representation", + "url": "https://www.semanticscholar.org/paper/1512b9570669a9dfe2a252900ae94b276e3aff9e", + "venue": "IEEE Transactions on Pattern Analysis and Machine Intelligence", + "year": 2009 + }, + { + "arxivId": null, + "authors": [ + { + "authorId": "3219900", + "name": "Gary B. Huang", + "url": "https://www.semanticscholar.org/author/3219900" + }, + { + "authorId": "49354909", + "name": "Marwan Mattar", + "url": "https://www.semanticscholar.org/author/49354909" + }, + { + "authorId": "1685538", + "name": "Tamara L. Berg", + "url": "https://www.semanticscholar.org/author/1685538" + }, + { + "authorId": "1714536", + "name": "Erik G. Learned-Miller", + "url": "https://www.semanticscholar.org/author/1714536" + } + ], + "doi": null, + "isInfluential": true, + "paperId": "370b5757a5379b15e30d619e4d3fb9e8e13f3256", + "title": "Labeled Faces in the Wild : A Database for Studying Face Recognition in Unconstrained Environments", + "url": "https://www.semanticscholar.org/paper/370b5757a5379b15e30d619e4d3fb9e8e13f3256", + "venue": "", + "year": 2007 + }, + { + "arxivId": null, + "authors": [ + { + "authorId": "1769383", + "name": "Lubomir D. Bourdev", + "url": "https://www.semanticscholar.org/author/1769383" + }, + { + "authorId": "1689212", + "name": "Jitendra Malik", + "url": "https://www.semanticscholar.org/author/1689212" + } + ], + "doi": "10.1109/ICCV.2009.5459303", + "isInfluential": false, + "paperId": "2830fb5282de23d7784b4b4bc37065d27839a412", + "title": "Poselets: Body part detectors trained using 3D human pose annotations", + "url": "https://www.semanticscholar.org/paper/2830fb5282de23d7784b4b4bc37065d27839a412", + "venue": "2009 IEEE 12th International Conference on Computer Vision", + "year": 2009 + }, + { + "arxivId": null, + "authors": [ + { + "authorId": "3056091", + "name": "Mark Everingham", + "url": "https://www.semanticscholar.org/author/3056091" + }, + { + "authorId": "1782755", + "name": "Josef Sivic", + "url": "https://www.semanticscholar.org/author/1782755" + }, + { + "authorId": "1688869", + "name": "Andrew Zisserman", + "url": "https://www.semanticscholar.org/author/1688869" + } + ], + "doi": "10.1016/j.imavis.2008.04.018", + "isInfluential": false, + "paperId": "642a386c451e94d9c44134e03052219a7512b9de", + "title": "Taking the bite out of automated naming of characters in TV video", + "url": "https://www.semanticscholar.org/paper/642a386c451e94d9c44134e03052219a7512b9de", + "venue": "Image Vision Comput.", + "year": 2009 + }, + { + "arxivId": null, + "authors": [ + { + "authorId": "48625148", + "name": "Wei Li", + "url": "https://www.semanticscholar.org/author/48625148" + }, + { + "authorId": "49832825", + "name": "Rui Zhao", + "url": "https://www.semanticscholar.org/author/49832825" + }, + { + "authorId": "1721881", + "name": "Tong Xiao", + "url": "https://www.semanticscholar.org/author/1721881" + }, + { + "authorId": "31843833", + "name": "Xiaogang Wang", + "url": "https://www.semanticscholar.org/author/31843833" + } + ], + "doi": "10.1109/CVPR.2014.27", + "isInfluential": false, + "paperId": "6bd36e9fd0ef20a3074e1430a6cc601e6d407fc3", + "title": "DeepReID: Deep Filter Pairing Neural Network for Person Re-identification", + "url": "https://www.semanticscholar.org/paper/6bd36e9fd0ef20a3074e1430a6cc601e6d407fc3", + "venue": "2014 IEEE Conference on Computer Vision and Pattern Recognition", + "year": 2014 + }, + { + "arxivId": null, + "authors": [ + { + "authorId": "38707615", + "name": "Bryan James Prosser", + "url": "https://www.semanticscholar.org/author/38707615" + }, + { + "authorId": "3333315", + "name": "Wei-Shi Zheng", + "url": "https://www.semanticscholar.org/author/3333315" + }, + { + "authorId": "2073354", + "name": "Shaogang Gong", + "url": "https://www.semanticscholar.org/author/2073354" + }, + { + "authorId": "1700927", + "name": "Tao Xiang", + "url": "https://www.semanticscholar.org/author/1700927" + } + ], + "doi": "10.5244/C.24.21", + "isInfluential": false, + "paperId": "8aef5b3cfc80fafdcefc24c72a4796ca40f4bc8b", + "title": "Person Re-Identification by Support Vector Ranking", + "url": "https://www.semanticscholar.org/paper/8aef5b3cfc80fafdcefc24c72a4796ca40f4bc8b", + "venue": "BMVC", + "year": 2010 + }, + { + "arxivId": null, + "authors": [ + { + "authorId": "2351962", + "name": "Michela Farenzena", + "url": "https://www.semanticscholar.org/author/2351962" + }, + { + "authorId": "1809420", + "name": "Loris Bazzani", + "url": "https://www.semanticscholar.org/author/1809420" + }, + { + "authorId": "1691336", + "name": "Alessandro Perina", + "url": "https://www.semanticscholar.org/author/1691336" + }, + { + "authorId": "1727204", + "name": "Vittorio Murino", + "url": "https://www.semanticscholar.org/author/1727204" + }, + { + "authorId": "1723008", + "name": "Marco Cristani", + "url": "https://www.semanticscholar.org/author/1723008" + } + ], + "doi": "10.1109/CVPR.2010.5539926", + "isInfluential": false, + "paperId": "12fa3c73a7764cb65bb76fed0601fc5d79893bcd", + "title": "Person re-identification by symmetry-driven accumulation of local features", + "url": "https://www.semanticscholar.org/paper/12fa3c73a7764cb65bb76fed0601fc5d79893bcd", + "venue": "2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition", + "year": 2010 + }, + { + "arxivId": null, + "authors": [ + { + "authorId": "2737253", + "name": "Matthieu Guillaumin", + "url": "https://www.semanticscholar.org/author/2737253" + }, + { + "authorId": "1721683", + "name": "Jakob J. Verbeek", + "url": "https://www.semanticscholar.org/author/1721683" + }, + { + "authorId": "2462253", + "name": "Cordelia Schmid", + "url": "https://www.semanticscholar.org/author/2462253" + } + ], + "doi": "10.1109/ICCV.2009.5459197", + "isInfluential": false, + "paperId": "19296e129c70b332a8c0a67af8990f2f4d4f44d1", + "title": "Is that you? Metric learning approaches for face identification", + "url": "https://www.semanticscholar.org/paper/19296e129c70b332a8c0a67af8990f2f4d4f44d1", + "venue": "2009 IEEE 12th International Conference on Computer Vision", + "year": 2009 + }, + { + "arxivId": "1311.5591", + "authors": [ + { + "authorId": "49410785", + "name": "Ning Zhang", + "url": "https://www.semanticscholar.org/author/49410785" + }, + { + "authorId": "2210374", + "name": "Manohar Paluri", + "url": "https://www.semanticscholar.org/author/2210374" + }, + { + "authorId": "1706809", + "name": "Marc'Aurelio Ranzato", + "url": "https://www.semanticscholar.org/author/1706809" + }, + { + "authorId": "1753210", + "name": "Trevor Darrell", + "url": "https://www.semanticscholar.org/author/1753210" + }, + { + "authorId": "1769383", + "name": "Lubomir D. Bourdev", + "url": "https://www.semanticscholar.org/author/1769383" + } + ], + "doi": "10.1109/CVPR.2014.212", + "isInfluential": false, + "paperId": "42e3dac0df30d754c7c7dab9e1bb94990034a90d", + "title": "PANDA: Pose Aligned Networks for Deep Attribute Modeling", + "url": "https://www.semanticscholar.org/paper/42e3dac0df30d754c7c7dab9e1bb94990034a90d", + "venue": "2014 IEEE Conference on Computer Vision and Pattern Recognition", + "year": 2014 + }, + { + "arxivId": null, + "authors": [ + { + "authorId": "2188620", + "name": "Yaniv Taigman", + "url": "https://www.semanticscholar.org/author/2188620" + }, + { + "authorId": "2909406", + "name": "Ming Yang", + "url": "https://www.semanticscholar.org/author/2909406" + }, + { + "authorId": "1706809", + "name": "Marc'Aurelio Ranzato", + "url": "https://www.semanticscholar.org/author/1706809" + }, + { + "authorId": "1776343", + "name": "Lior Wolf", + "url": "https://www.semanticscholar.org/author/1776343" + } + ], + "doi": "10.1109/CVPR.2014.220", + "isInfluential": false, + "paperId": "14ce7635ff18318e7094417d0f92acbec6669f1c", + "title": "DeepFace: Closing the Gap to Human-Level Performance in Face Verification", + "url": "https://www.semanticscholar.org/paper/14ce7635ff18318e7094417d0f92acbec6669f1c", + "venue": "2014 IEEE Conference on Computer Vision and Pattern Recognition", + "year": 2014 + }, + { + "arxivId": null, + "authors": [ + { + "authorId": "2103464", + "name": "Makarand Tapaswi", + "url": "https://www.semanticscholar.org/author/2103464" + }, + { + "authorId": "1931707", + "name": "Martin Bäuml", + "url": "https://www.semanticscholar.org/author/1931707" + }, + { + "authorId": "1742325", + "name": "Rainer Stiefelhagen", + "url": "https://www.semanticscholar.org/author/1742325" + } + ], + "doi": "10.1109/CVPR.2012.6247986", + "isInfluential": false, + "paperId": "441132dd4ec14991644723c9642ac3a63181753e", + "title": "“Knock! Knock! Who is it?” probabilistic person identification in TV-series", + "url": "https://www.semanticscholar.org/paper/441132dd4ec14991644723c9642ac3a63181753e", + "venue": "2012 IEEE Conference on Computer Vision and Pattern Recognition", + "year": 2012 + }, + { + "arxivId": null, + "authors": [ + { + "authorId": "39460815", + "name": "Andrew C. Gallagher", + "url": "https://www.semanticscholar.org/author/39460815" + }, + { + "authorId": "1746230", + "name": "Tsuhan Chen", + "url": "https://www.semanticscholar.org/author/1746230" + } + ], + "doi": "10.1109/CVPR.2008.4587481", + "isInfluential": true, + "paperId": "22ad2c8c0f4d6aa4328b38d894b814ec22579761", + "title": "Clothing cosegmentation for recognizing people", + "url": "https://www.semanticscholar.org/paper/22ad2c8c0f4d6aa4328b38d894b814ec22579761", + "venue": "2008 IEEE Conference on Computer Vision and Pattern Recognition", + "year": 2008 + }, + { + "arxivId": null, + "authors": [ + { + "authorId": "3056091", + "name": "Mark Everingham", + "url": "https://www.semanticscholar.org/author/3056091" + }, + { + "authorId": "1782755", + "name": "Josef Sivic", + "url": "https://www.semanticscholar.org/author/1782755" + }, + { + "authorId": "1688869", + "name": "Andrew Zisserman", + "url": "https://www.semanticscholar.org/author/1688869" + } + ], + "doi": "10.5244/C.20.92", + "isInfluential": false, + "paperId": "75ebe1e0ae9d42732e31948e2e9c03d680235c39", + "title": "Hello! My name is... Buffy'' -- Automatic Naming of Characters in TV Video", + "url": "https://www.semanticscholar.org/paper/75ebe1e0ae9d42732e31948e2e9c03d680235c39", + "venue": "BMVC", + "year": 2006 + }, + { + "arxivId": null, + "authors": [ + { + "authorId": "22804340", + "name": "Gang Wang", + "url": "https://www.semanticscholar.org/author/22804340" + }, + { + "authorId": "39460815", + "name": "Andrew C. Gallagher", + "url": "https://www.semanticscholar.org/author/39460815" + }, + { + "authorId": "33642939", + "name": "Jiebo Luo", + "url": "https://www.semanticscholar.org/author/33642939" + }, + { + "authorId": "1744452", + "name": "David A. Forsyth", + "url": "https://www.semanticscholar.org/author/1744452" + } + ], + "doi": "10.1007/978-3-642-15555-0_13", + "isInfluential": false, + "paperId": "291f527598c589fb0519f890f1beb2749082ddfd", + "title": "Seeing People in Social Context: Recognizing People and Social Relationships", + "url": "https://www.semanticscholar.org/paper/291f527598c589fb0519f890f1beb2749082ddfd", + "venue": "ECCV", + "year": 2010 + }, + { + "arxivId": null, + "authors": [ + { + "authorId": "2230138", + "name": "Matthew A. Turk", + "url": "https://www.semanticscholar.org/author/2230138" + }, + { + "authorId": "1682773", + "name": "Alex Pentland", + "url": "https://www.semanticscholar.org/author/1682773" + } + ], + "doi": "10.1162/jocn.1991.3.1.71", + "isInfluential": false, + "paperId": "a6f1dfcc44277d4cfd8507284d994c9283dc3a2f", + "title": "Eigenfaces for Recognition", + "url": "https://www.semanticscholar.org/paper/a6f1dfcc44277d4cfd8507284d994c9283dc3a2f", + "venue": "Journal of Cognitive Neuroscience", + "year": 1991 + }, + { + "arxivId": null, + "authors": [ + { + "authorId": "20373504", + "name": "Mary C Dugan", + "url": "https://www.semanticscholar.org/author/20373504" + } + ], + "doi": null, + "isInfluential": false, + "paperId": "ef80015cd8de355d2cf197f050acce33d15d5ad5", + "title": "Hello, my name is ....", + "url": "https://www.semanticscholar.org/paper/ef80015cd8de355d2cf197f050acce33d15d5ad5", + "venue": "Contemporary longterm care", + "year": 1990 + }, + { + "arxivId": null, + "authors": [], + "doi": null, + "isInfluential": false, + "paperId": "bd433d471af50b571d7284afb5ee435654ace99f", + "title": "Going Deeper with Convolutional Neural Network for Intelligent Transportation", + "url": "https://www.semanticscholar.org/paper/bd433d471af50b571d7284afb5ee435654ace99f", + "venue": "", + "year": 2016 + }, + { + "arxivId": "1407.4979", + "authors": [ + { + "authorId": "1716143", + "name": "Dong Yi", + "url": "https://www.semanticscholar.org/author/1716143" + }, + { + "authorId": "1718623", + "name": "Zhen Lei", + "url": "https://www.semanticscholar.org/author/1718623" + }, + { + "authorId": "34679741", + "name": "Stan Z. Li", + "url": "https://www.semanticscholar.org/author/34679741" + } + ], + "doi": null, + "isInfluential": false, + "paperId": "afd29ac2de84c8a6d48232477be018ec57d6f564", + "title": "Deep Metric Learning for Practical Person Re-Identification", + "url": "https://www.semanticscholar.org/paper/afd29ac2de84c8a6d48232477be018ec57d6f564", + "venue": "ArXiv", + "year": 2014 + }, + { + "arxivId": null, + "authors": [ + { + "authorId": "9748713", + "name": "Rahul Garg", + "url": "https://www.semanticscholar.org/author/9748713" + }, + { + "authorId": "1679223", + "name": "Steven M. Seitz", + "url": "https://www.semanticscholar.org/author/1679223" + }, + { + "authorId": "1770537", + "name": "Deva Ramanan", + "url": "https://www.semanticscholar.org/author/1770537" + }, + { + "authorId": "1830653", + "name": "Noah Snavely", + "url": "https://www.semanticscholar.org/author/1830653" + } + ], + "doi": "10.1109/CVPR.2011.5995546", + "isInfluential": false, + "paperId": "5b2bc289b607ca1a0634555158464f28fe68a6d3", + "title": "Where's Waldo: Matching people in images of crowds", + "url": "https://www.semanticscholar.org/paper/5b2bc289b607ca1a0634555158464f28fe68a6d3", + "venue": "CVPR 2011", + "year": 2011 + }, + { + "arxivId": null, + "authors": [ + { + "authorId": "39460815", + "name": "Andrew C. Gallagher", + "url": "https://www.semanticscholar.org/author/39460815" + }, + { + "authorId": "1746230", + "name": "Tsuhan Chen", + "url": "https://www.semanticscholar.org/author/1746230" + } + ], + "doi": "10.1109/CVPRW.2009.5206828", + "isInfluential": false, + "paperId": "21d9d0deed16f0ad62a4865e9acf0686f4f15492", + "title": "Understanding images of groups of people", + "url": "https://www.semanticscholar.org/paper/21d9d0deed16f0ad62a4865e9acf0686f4f15492", + "venue": "2009 IEEE Conference on Computer Vision and Pattern Recognition", + "year": 2009 + }, + { + "arxivId": null, + "authors": [ + { + "authorId": "33612681", + "name": "Doug Gray", + "url": "https://www.semanticscholar.org/author/33612681" + }, + { + "authorId": "7207299", + "name": "Shane Brennan", + "url": "https://www.semanticscholar.org/author/7207299" + }, + { + "authorId": "2491194", + "name": "Hai Tao", + "url": "https://www.semanticscholar.org/author/2491194" + } + ], + "doi": null, + "isInfluential": false, + "paperId": "6273b3491e94ea4dd1ce42b791d77bdc96ee73a8", + "title": "Evaluating Appearance Models for Recognition , Reacquisition , and Tracking", + "url": "https://www.semanticscholar.org/paper/6273b3491e94ea4dd1ce42b791d77bdc96ee73a8", + "venue": "", + "year": 2007 + }, + { + "arxivId": null, + "authors": [ + { + "authorId": "1782755", + "name": "Josef Sivic", + "url": "https://www.semanticscholar.org/author/1782755" + }, + { + "authorId": "3056091", + "name": "Mark Everingham", + "url": "https://www.semanticscholar.org/author/3056091" + }, + { + "authorId": "1688869", + "name": "Andrew Zisserman", + "url": "https://www.semanticscholar.org/author/1688869" + } + ], + "doi": "10.1109/CVPRW.2009.5206513", + "isInfluential": false, + "paperId": "03d1d0a665e358863ff4de9ee7d78f64edd7e756", + "title": "“Who are you?” - Learning person specific classifiers from video", + "url": "https://www.semanticscholar.org/paper/03d1d0a665e358863ff4de9ee7d78f64edd7e756", + "venue": "2009 IEEE Conference on Computer Vision and Pattern Recognition", + "year": 2009 + }, + { + "arxivId": null, + "authors": [ + { + "authorId": "2405613", + "name": "Omar Oreifej", + "url": "https://www.semanticscholar.org/author/2405613" + }, + { + "authorId": "35422941", + "name": "Ramin Mehran", + "url": "https://www.semanticscholar.org/author/35422941" + }, + { + "authorId": "1745480", + "name": "Mubarak Shah", + "url": "https://www.semanticscholar.org/author/1745480" + } + ], + "doi": "10.1109/CVPR.2010.5540147", + "isInfluential": false, + "paperId": "34c256893c3e2dfcf77b7da1c5daa77981a79196", + "title": "Human identity recognition in aerial images", + "url": "https://www.semanticscholar.org/paper/34c256893c3e2dfcf77b7da1c5daa77981a79196", + "venue": "2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition", + "year": 2010 + }, + { + "arxivId": null, + "authors": [ + { + "authorId": "1838674", + "name": "Dragomir Anguelov", + "url": "https://www.semanticscholar.org/author/1838674" + }, + { + "authorId": "2457452", + "name": "Kuang-chih Lee", + "url": "https://www.semanticscholar.org/author/2457452" + }, + { + "authorId": "2437408", + "name": "Salih Burak Göktürk", + "url": "https://www.semanticscholar.org/author/2437408" + }, + { + "authorId": "1702811", + "name": "Baris Sumengen", + "url": "https://www.semanticscholar.org/author/1702811" + } + ], + "doi": "10.1109/CVPR.2007.383057", + "isInfluential": false, + "paperId": "28f232498c6e14950ef9b6a3ae0781f2348be3b7", + "title": "Contextual Identity Recognition in Personal Photo Albums", + "url": "https://www.semanticscholar.org/paper/28f232498c6e14950ef9b6a3ae0781f2348be3b7", + "venue": "2007 IEEE Conference on Computer Vision and Pattern Recognition", + "year": 2007 + }, + { + "arxivId": "1409.1556", + "authors": [ + { + "authorId": "34838386", + "name": "Karen Simonyan", + "url": "https://www.semanticscholar.org/author/34838386" + }, + { + "authorId": "1688869", + "name": "Andrew Zisserman", + "url": "https://www.semanticscholar.org/author/1688869" + } + ], + "doi": null, + "isInfluential": false, + "paperId": "061356704ec86334dbbc073985375fe13cd39088", + "title": "Very Deep Convolutional Networks for Large-Scale Image Recognition", + "url": "https://www.semanticscholar.org/paper/061356704ec86334dbbc073985375fe13cd39088", + "venue": "ArXiv", + "year": 2014 + }, + { + "arxivId": null, + "authors": [ + { + "authorId": "3264124", + "name": "Ryan Layne", + "url": "https://www.semanticscholar.org/author/3264124" + }, + { + "authorId": "1697755", + "name": "Timothy M. Hospedales", + "url": "https://www.semanticscholar.org/author/1697755" + }, + { + "authorId": "2073354", + "name": "Shaogang Gong", + "url": "https://www.semanticscholar.org/author/2073354" + } + ], + "doi": "10.5244/C.26.24", + "isInfluential": false, + "paperId": "34aa3dca30dc5cbf86c92d5035e35d264540a829", + "title": "Person Re-identification by Attributes", + "url": "https://www.semanticscholar.org/paper/34aa3dca30dc5cbf86c92d5035e35d264540a829", + "venue": "BMVC", + "year": 2012 + }, + { + "arxivId": null, + "authors": [ + { + "authorId": "8342699", + "name": "Jia Deng", + "url": "https://www.semanticscholar.org/author/8342699" + }, + { + "authorId": "49191744", + "name": "Wei Dong", + "url": "https://www.semanticscholar.org/author/49191744" + }, + { + "authorId": "2166511", + "name": "Richard Socher", + "url": "https://www.semanticscholar.org/author/2166511" + }, + { + "authorId": "33642044", + "name": "Li-Jia Li", + "url": "https://www.semanticscholar.org/author/33642044" + }, + { + "authorId": "2168945", + "name": "Kai Li", + "url": "https://www.semanticscholar.org/author/2168945" + }, + { + "authorId": "3216322", + "name": "Li Fei-Fei", + "url": "https://www.semanticscholar.org/author/3216322" + } + ], + "doi": "10.1109/CVPRW.2009.5206848", + "isInfluential": false, + "paperId": "38211dc39e41273c0007889202c69f841e02248a", + "title": "ImageNet: A large-scale hierarchical image database", + "url": "https://www.semanticscholar.org/paper/38211dc39e41273c0007889202c69f841e02248a", + "venue": "2009 IEEE Conference on Computer Vision and Pattern Recognition", + "year": 2009 + }, + { + "arxivId": "1310.1531", + "authors": [ + { + "authorId": "7408951", + "name": "Jeff Donahue", + "url": "https://www.semanticscholar.org/author/7408951" + }, + { + "authorId": "39978391", + "name": "Yangqing Jia", + "url": "https://www.semanticscholar.org/author/39978391" + }, + { + "authorId": "1689108", + "name": "Oriol Vinyals", + "url": "https://www.semanticscholar.org/author/1689108" + }, + { + "authorId": "50196944", + "name": "Judy Hoffman", + "url": "https://www.semanticscholar.org/author/50196944" + }, + { + "authorId": "49410785", + "name": "Ning Zhang", + "url": "https://www.semanticscholar.org/author/49410785" + }, + { + "authorId": "2368132", + "name": "Eric Tzeng", + "url": "https://www.semanticscholar.org/author/2368132" + }, + { + "authorId": "1753210", + "name": "Trevor Darrell", + "url": "https://www.semanticscholar.org/author/1753210" + } + ], + "doi": null, + "isInfluential": false, + "paperId": "33da83b54410af11d0cd18fd07c74e1a99f67e84", + "title": "DeCAF: A Deep Convolutional Activation Feature for Generic Visual Recognition", + "url": "https://www.semanticscholar.org/paper/33da83b54410af11d0cd18fd07c74e1a99f67e84", + "venue": "ICML", + "year": 2014 + }, + { + "arxivId": null, + "authors": [ + { + "authorId": "1688882", + "name": "Yann LeCun", + "url": "https://www.semanticscholar.org/author/1688882" + }, + { + "authorId": "2219581", + "name": "Bernhard E. Boser", + "url": "https://www.semanticscholar.org/author/2219581" + }, + { + "authorId": "1747317", + "name": "John S. Denker", + "url": "https://www.semanticscholar.org/author/1747317" + }, + { + "authorId": "37274089", + "name": "Donnie Henderson", + "url": "https://www.semanticscholar.org/author/37274089" + }, + { + "authorId": "32295804", + "name": "Richard E. Howard", + "url": "https://www.semanticscholar.org/author/32295804" + }, + { + "authorId": "34859193", + "name": "Wayne E. Hubbard", + "url": "https://www.semanticscholar.org/author/34859193" + }, + { + "authorId": "2307573", + "name": "Lawrence D. Jackel", + "url": "https://www.semanticscholar.org/author/2307573" + } + ], + "doi": "10.1162/neco.1989.1.4.541", + "isInfluential": false, + "paperId": "a8e8f3c8d4418c8d62e306538c9c1292635e9d27", + "title": "Backpropagation Applied to Handwritten Zip Code Recognition", + "url": "https://www.semanticscholar.org/paper/a8e8f3c8d4418c8d62e306538c9c1292635e9d27", + "venue": "Neural Computation", + "year": 1989 + }, + { + "arxivId": null, + "authors": [ + { + "authorId": "49832825", + "name": "Rui Zhao", + "url": "https://www.semanticscholar.org/author/49832825" + }, + { + "authorId": "3001348", + "name": "Wanli Ouyang", + "url": "https://www.semanticscholar.org/author/3001348" + }, + { + "authorId": "31843833", + "name": "Xiaogang Wang", + "url": "https://www.semanticscholar.org/author/31843833" + } + ], + "doi": "10.1109/CVPR.2013.460", + "isInfluential": false, + "paperId": "46638b810bf69023bca41db664b49bc935bcba3c", + "title": "Unsupervised Salience Learning for Person Re-identification", + "url": "https://www.semanticscholar.org/paper/46638b810bf69023bca41db664b49bc935bcba3c", + "venue": "2013 IEEE Conference on Computer Vision and Pattern Recognition", + "year": 2013 + }, + { + "arxivId": null, + "authors": [ + { + "authorId": "49832825", + "name": "Rui Zhao", + "url": "https://www.semanticscholar.org/author/49832825" + }, + { + "authorId": "3001348", + "name": "Wanli Ouyang", + "url": "https://www.semanticscholar.org/author/3001348" + }, + { + "authorId": "31843833", + "name": "Xiaogang Wang", + "url": "https://www.semanticscholar.org/author/31843833" + } + ], + "doi": "10.1109/CVPR.2014.26", + "isInfluential": false, + "paperId": "3fd7bfd90f0dfc3369bfe718e27aff30cf268c23", + "title": "Learning Mid-level Filters for Person Re-identification", + "url": "https://www.semanticscholar.org/paper/3fd7bfd90f0dfc3369bfe718e27aff30cf268c23", + "venue": "2014 IEEE Conference on Computer Vision and Pattern Recognition", + "year": 2014 + }, + { + "arxivId": null, + "authors": [ + { + "authorId": "1782755", + "name": "Josef Sivic", + "url": "https://www.semanticscholar.org/author/1782755" + }, + { + "authorId": "1699161", + "name": "C. Lawrence Zitnick", + "url": "https://www.semanticscholar.org/author/1699161" + }, + { + "authorId": "1717841", + "name": "Richard Szeliski", + "url": "https://www.semanticscholar.org/author/1717841" + } + ], + "doi": "10.5244/C.20.93", + "isInfluential": false, + "paperId": "586f106ba7aede43554c4287eb6c5b99d4725f8f", + "title": "Finding People in Repeated Shots of the Same Scene", + "url": "https://www.semanticscholar.org/paper/586f106ba7aede43554c4287eb6c5b99d4725f8f", + "venue": "BMVC", + "year": 2006 + } + ], + "title": "Beyond frontal faces: Improving Person Recognition using multiple cues", + "topics": [ + { + "topic": "Finite-state machine", + "topicId": "4280", + "url": "https://www.semanticscholar.org/topic/4280" + }, + { + "topic": "Flickr", + "topicId": "67227", + "url": "https://www.semanticscholar.org/topic/67227" + }, + { + "topic": "Image resolution", + "topicId": "881", + "url": "https://www.semanticscholar.org/topic/881" + }, + { + "topic": "Convolutional neural network", + "topicId": "29860", + "url": "https://www.semanticscholar.org/topic/29860" + } + ], + "url": "https://www.semanticscholar.org/paper/0a85bdff552615643dd74646ac881862a7c7072d", + "venue": "2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR)", + "year": 2015 +}
\ No newline at end of file diff --git a/scraper/samples/s2-search-api.json b/scraper/samples/s2-search-api.json new file mode 100644 index 00000000..207fb98c --- /dev/null +++ b/scraper/samples/s2-search-api.json @@ -0,0 +1,270 @@ +{ + "id": "0d2dd4fc016cb6a517d8fb43a7cc3ff62964832e", + "title": { + "text": "Large age-gap face verification by feature injection in deep networks", + "fragments": [ + { + "start": 0, + "end": 5 + }, + { + "start": 6, + "end": 9 + }, + { + "start": 10, + "end": 13 + }, + { + "start": 14, + "end": 18 + }, + { + "start": 19, + "end": 31 + }, + { + "start": 35, + "end": 42 + }, + { + "start": 43, + "end": 52 + }, + { + "start": 56, + "end": 60 + }, + { + "start": 61, + "end": 69 + } + ] + }, + "slug": "Large-age-gap-face-verification-by-feature-in-deep-Bianco", + "paperAbstract": { + "text": "This paper introduces a new method for face verification across large age gaps and also a dataset containing variations of age in the wild, the Large Age-Gap (LAG) dataset, with images ranging from child/young to adult/old. The proposed method exploits a deep convolutional neural network (DCNN) pre-trained for the face recognition task on a large dataset and then fine-tuned for the large age-gap face verification task. Finetuning is performed in a Siamese architecture using a contrastive loss function. A feature injection layer is introduced to boost verification accuracy, showing the ability of the DCNN to learn a similarity metric leveraging external features. Experimental results on the LAG dataset show that our method is able to outperform the face verification solutions in the state of the art considered.", + "fragments": [ + { + "start": 39, + "end": 43 + }, + { + "start": 44, + "end": 56 + }, + { + "start": 64, + "end": 69 + }, + { + "start": 70, + "end": 73 + }, + { + "start": 74, + "end": 78 + }, + { + "start": 123, + "end": 126 + }, + { + "start": 144, + "end": 149 + }, + { + "start": 150, + "end": 153 + }, + { + "start": 154, + "end": 157 + }, + { + "start": 255, + "end": 259 + }, + { + "start": 281, + "end": 288 + }, + { + "start": 316, + "end": 320 + }, + { + "start": 343, + "end": 348 + }, + { + "start": 385, + "end": 390 + }, + { + "start": 391, + "end": 394 + }, + { + "start": 395, + "end": 398 + }, + { + "start": 399, + "end": 403 + }, + { + "start": 404, + "end": 416 + }, + { + "start": 510, + "end": 517 + }, + { + "start": 518, + "end": 527 + }, + { + "start": 557, + "end": 569 + }, + { + "start": 758, + "end": 762 + }, + { + "start": 763, + "end": 775 + } + ] + }, + "authors": [ + [ + { + "name": "Simone Bianco", + "ids": [ + "2217051" + ], + "slug": "Simone-Bianco" + }, + { + "text": "Simone Bianco", + "fragments": [] + } + ] + ], + "structuredAuthors": [ + { + "firstName": "Simone", + "middleNames": [], + "lastName": "Bianco" + } + ], + "year": { + "text": "2017", + "fragments": [] + }, + "venue": { + "text": "Pattern Recognition Letters", + "fragments": [] + }, + "citationContexts": [], + "citationStats": { + "citedByBuckets": [ + { + "startKey": 2017, + "endKey": 2017, + "count": 2, + "estimate": { + "min": 2, + "value": 2.543286309486936, + "max": 6.143286309486936, + "confidence": 0.9 + } + }, + { + "startKey": 2018, + "endKey": 2018, + "count": 3, + "estimate": { + "min": 4, + "value": 5.086572618973872, + "max": 12.286572618973873, + "confidence": 0.9 + } + } + ], + "keyCitedByBuckets": [], + "numCitations": 5, + "estNumCitations": { + "min": 5, + "value": 6.3582157737173395, + "max": 15.35821577371734, + "confidence": 0.9 + }, + "numReferences": 40, + "numKeyCitations": 0, + "numKeyReferences": 8, + "numViewableReferences": 40, + "keyCitationRate": 0, + "estCitationVelocity": { + "estimate": { + "min": 0, + "value": 0, + "max": 0, + "confidence": 1 + }, + "estCitationsByRange": [] + }, + "estCitationAcceleration": { + "estimate": { + "min": 0, + "value": 0, + "max": 0, + "confidence": 1 + }, + "estCitationsByRange": [] + } + }, + "sources": [ + "DBLP", + "Grobid", + "ScienceParse", + "SPv2", + "ArXiv", + "Crawler", + "DBLP" + ], + "journal": { + "name": "Pattern Recognition Letters", + "volume": "90", + "pages": "36-42" + }, + "socialLinks": [], + "presentationUrls": [], + "doiInfo": { + "doi": "10.1016/j.patrec.2017.03.006", + "doiUrl": "http://doi.org/10.1016/j.patrec.2017.03.006" + }, + "links": [ + { + "url": "https://arxiv.org/pdf/1602.06149.pdf", + "linkType": "arxiv" + } + ], + "primaryPaperLink": { + "url": "https://arxiv.org/pdf/1602.06149.pdf", + "linkType": "arxiv" + }, + "alternatePaperLinks": [], + "entities": [], + "entityRelations": [], + "readerId": "1602.06149", + "blogs": [], + "videos": [], + "githubReferences": [], + "faqs": [], + "scorecardStats": [], + "hasPdf": true +}
\ No newline at end of file |
