{"id":4116,"date":"2024-05-01T13:42:25","date_gmt":"2024-05-01T13:42:25","guid":{"rendered":"https:\/\/www.eg.org\/wp\/?page_id=4116"},"modified":"2024-05-01T13:42:25","modified_gmt":"2024-05-01T13:42:25","slug":"young-researcher-award-2024-justus-thies","status":"publish","type":"page","link":"https:\/\/www.eg.org\/wp\/eurographics-awards-programme\/the-young-researcher-award\/young-researcher-award-2024-justus-thies\/","title":{"rendered":"Young Researcher Award 2024 &#8211; Justus Thies"},"content":{"rendered":"\n<div class=\"wp-block-image\"><figure class=\"alignleft is-resized\"><img decoding=\"async\" loading=\"lazy\" src=\"https:\/\/www.eg.org\/wp\/wp-content\/uploads\/2024\/04\/Justus-Thies.jpg\" alt=\"\" class=\"wp-image-4117\" width=\"385\" height=\"508\" srcset=\"https:\/\/www.eg.org\/wp\/wp-content\/uploads\/2024\/04\/Justus-Thies.jpg 769w, https:\/\/www.eg.org\/wp\/wp-content\/uploads\/2024\/04\/Justus-Thies-227x300.jpg 227w, https:\/\/www.eg.org\/wp\/wp-content\/uploads\/2024\/04\/Justus-Thies-768x1014.jpg 768w\" sizes=\"(max-width: 385px) 100vw, 385px\" \/><\/figure><\/div>\n\n\n\n<p>Justus Thies receives the EUROGRAPHICS Young Researcher Award 2024. Justus obtained his PhD from the University of Erlangen Nu-remberg. He is now a full Professor at the Technical University of Darmstadt where he is leading the 3D Graphics &amp; Vision group. In addition, he is an in-dependent research group leader at the Max Planck Institute for Intelligent Systems.  <\/p>\n\n\n\n<p>Justus studies the capture and syn-thesis of digital humans. His work blends elements of computer graphics, computer vision, and ma-chine learning, with the end goal of capturing and re-synthesizing reality. Justus has done profound work on important aspects of \u201cdigital humans\u201d, namely marker-less capture, neural synthesis, and multi-media forensics, among others. <\/p>\n\n\n\n<p>Justus\u2019 Face2Face algorithm was a pioneering and very successful ap-proach to real-time facial performance capture from video data which also enabled facial re-enactment. Later, with the advent of neural networks, Justus again considerably advanced the state of the art with his work on deferred neural rendering. Importantly, new methods for high-quality image and video, colloquially known as \u201cdeep fakes,\u201d can also be abused, e.g. for misinformation. Justus and col-leagues have made great strides on this equally important aspect in their work FaceForensics++, which allows to detect manipulated facial images. As the saying goes: \u201cIt takes (someone that makes) one to know one\u201d, thus it is of paramount importance that re-searchers that develop new machine learning-based image synthesis ap-proaches with all their great potential for computer graphics, additionally invest in identifying synthesized im-agery. Justus\u2019 groundbreaking work is a demonstration of this principle. <\/p>\n\n\n\n<p>The work of Justus Thies is published in the top tier conferences and jour-nals of computer graphics and com-puter vision, and has been very widely cited. His work also received several honors, such as the emerging tech-nology award at SIGGRAPH 2016.  With his considerable and impactful contributions to the field of digital hu-mans, Justus helps to shape the future of how the real and the virtual interact.<\/p>\n\n\n\n<p>EUROGRAPHICS is extremely pleased to recognize Justus Thies with the 2024 Young Researcher Award in recognition of his outstand-ing contributions to Computer Graphics in the area of marker-less motion capture and synthesis.<\/p>\n","protected":false},"excerpt":{"rendered":"<p>Justus Thies receives the EUROGRAPHICS Young Researcher Award 2024. Justus obtained his PhD from the University of Erlangen Nu-remberg. He is now a full Professor&#8230;<\/p>\n<div class=\"more-link-wrapper\"><a class=\"more-link\" href=\"https:\/\/www.eg.org\/wp\/eurographics-awards-programme\/the-young-researcher-award\/young-researcher-award-2024-justus-thies\/\">Read more<span class=\"screen-reader-text\">Young Researcher Award 2024 &#8211; Justus Thies<\/span><\/a><\/div>\n","protected":false},"author":18,"featured_media":0,"parent":137,"menu_order":0,"comment_status":"closed","ping_status":"closed","template":"templates\/full-width.php","meta":[],"_links":{"self":[{"href":"https:\/\/www.eg.org\/wp\/wp-json\/wp\/v2\/pages\/4116"}],"collection":[{"href":"https:\/\/www.eg.org\/wp\/wp-json\/wp\/v2\/pages"}],"about":[{"href":"https:\/\/www.eg.org\/wp\/wp-json\/wp\/v2\/types\/page"}],"author":[{"embeddable":true,"href":"https:\/\/www.eg.org\/wp\/wp-json\/wp\/v2\/users\/18"}],"replies":[{"embeddable":true,"href":"https:\/\/www.eg.org\/wp\/wp-json\/wp\/v2\/comments?post=4116"}],"version-history":[{"count":2,"href":"https:\/\/www.eg.org\/wp\/wp-json\/wp\/v2\/pages\/4116\/revisions"}],"predecessor-version":[{"id":4119,"href":"https:\/\/www.eg.org\/wp\/wp-json\/wp\/v2\/pages\/4116\/revisions\/4119"}],"up":[{"embeddable":true,"href":"https:\/\/www.eg.org\/wp\/wp-json\/wp\/v2\/pages\/137"}],"wp:attachment":[{"href":"https:\/\/www.eg.org\/wp\/wp-json\/wp\/v2\/media?parent=4116"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}