{"id":747889,"date":"2021-06-10T17:51:07","date_gmt":"2021-06-11T00:51:07","guid":{"rendered":"https:\/\/newed.any0.dpdns.org\/en-us\/research\/?post_type=msr-event&#038;p=747889"},"modified":"2025-08-06T11:51:19","modified_gmt":"2025-08-06T18:51:19","slug":"cvpr-2021","status":"publish","type":"msr-event","link":"https:\/\/newed.any0.dpdns.org\/en-us\/research\/event\/cvpr-2021\/","title":{"rendered":"Microsoft at CVPR 2021"},"content":{"rendered":"\n\n<p><strong>Website:<\/strong> <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" href=\"http:\/\/cvpr2021.thecvf.com\/\" target=\"_blank\" rel=\"noopener\">CVPR 2021<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><span id=\"label-external-link\" class=\"sr-only\" aria-hidden=\"true\">Opens in a new tab<\/span><\/p>\n<p>Microsoft is proud to be a sponsor of the <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" href=\"http:\/\/cvpr2021.thecvf.com\/\" target=\"_blank\" rel=\"noopener\">Computer Vision and Pattern Recognition (CVPR 2021)<span class=\"sr-only\"> (opens in new tab)<\/span><\/a> event.<\/p>\n<div style=\"height: 20px\"><\/div>\n<h3>Synthetic Data with Digital Humans<\/h3>\n<p><em>Microsoft Sponsor Session<\/em><\/p>\n<p>Join Erroll Wood and Tadas Baltrusaitis from Microsoft&#8217;s <a href=\"https:\/\/newed.any0.dpdns.org\/en-us\/research\/lab\/mixed-reality-ai-lab-cambridge\/\">Mixed Reality & AI Lab in Cambridge, UK<\/a>, for a talk on how synthetics drives work on understanding human faces and hands, including how it powers Fully Articulated Hand Tracking on HoloLens 2.<\/p>\n<div class=\"video-wrapper margin-bottom-sp1\"><iframe loading=\"lazy\" title=\"Synthetic Data with Digital Humans\" width=\"500\" height=\"281\" src=\"https:\/\/www.youtube-nocookie.com\/embed\/4rRF4UMppjY?feature=oembed&rel=0\" frameborder=\"0\" allow=\"accelerometer; autoplay; clipboard-write; encrypted-media; gyroscope; picture-in-picture; web-share\" referrerpolicy=\"strict-origin-when-cross-origin\" allowfullscreen><\/iframe><\/div>\n<div style=\"height: 30px\"><\/div>\n<h3>Longuet-Higgins Prize Winner<\/h3>\n<p><strong>Real-time human pose recognition in parts from single depth image<br \/>\n<\/strong><a href=\"https:\/\/newed.any0.dpdns.org\/en-us\/research\/people\/jamiesho\/\">Jamie Shotton<\/a>, <a href=\"https:\/\/newed.any0.dpdns.org\/en-us\/research\/people\/awf\/\">Andrew Fitzgibbon<\/a>, Mat Cook, <a href=\"https:\/\/newed.any0.dpdns.org\/en-us\/research\/people\/tsharp\/\">Toby Sharp<\/a>, Mark Finocchio, Richard Moore, Alex Kipman, <a href=\"https:\/\/newed.any0.dpdns.org\/en-us\/research\/people\/ablake\/\">Andrew Blake<\/a><\/p>\n<div style=\"height: 20px\"><\/div>\n<h3>Nominations for CVPR 2021 Best Paper Award<\/h3>\n<p><strong>Privacy-Preserving Image Features via Adversarial Affine Subspace Embeddings<\/strong><br \/>\nMihai Dusmanu, <a href=\"https:\/\/newed.any0.dpdns.org\/en-us\/research\/people\/joschonb\/\">Johannes L Sch\u00f6nberger<\/a>, <a href=\"https:\/\/newed.any0.dpdns.org\/en-us\/research\/people\/sudipsin\/\">Sudipta Sinha<\/a>, <a href=\"https:\/\/newed.any0.dpdns.org\/en-us\/research\/people\/mapoll\/\">Marc Pollefeys<\/a><\/p>\n<p><strong>CoCosNet v2: Full-Resolution Correspondence Learning for Image Translation<\/strong><br \/>\n<a href=\"https:\/\/newed.any0.dpdns.org\/en-us\/research\/people\/zhanbo\/\">Bo Zhang<\/a>, <a href=\"https:\/\/newed.any0.dpdns.org\/en-us\/research\/people\/tinzhan\/\">Ting Zhang<\/a>, Pan Zhang, <a href=\"https:\/\/newed.any0.dpdns.org\/en-us\/research\/people\/jianbao\/\">Jianmin Bao<\/a>, <a href=\"https:\/\/newed.any0.dpdns.org\/en-us\/research\/people\/doch\/\">Dong Chen<\/a>, Zhongfei Zhang, <a href=\"https:\/\/newed.any0.dpdns.org\/en-us\/research\/people\/fangwen\/\">Fang Wen<\/a><\/p>\n<p><strong>Less Is More: ClipBERT for Video-and-Language Learning via Sparse Sampling<\/strong><br \/>\nJie Lei, <a href=\"https:\/\/newed.any0.dpdns.org\/en-us\/research\/people\/linjli\/\">Linjie Li<\/a>, <a href=\"https:\/\/newed.any0.dpdns.org\/en-us\/research\/people\/luozhou\/\">Luowei Zhou<\/a>, <a href=\"https:\/\/newed.any0.dpdns.org\/en-us\/research\/people\/zhgan\/\">Zhe Gan<\/a>, Tamara Berg, Mohit Bansal, Jingjing Liu<\/p>\n<div style=\"height: 20px\"><\/div>\n<h3>Area Chairs<\/h3>\n<p><a href=\"https:\/\/newed.any0.dpdns.org\/en-us\/research\/people\/meic\/\">Mei Chen<\/a>, Mentor for the Doctoral Consortium<br \/>\n<a href=\"https:\/\/newed.any0.dpdns.org\/en-us\/research\/people\/hanhu\/\">Han Hu<\/a><br \/>\n<a href=\"https:\/\/newed.any0.dpdns.org\/en-us\/research\/people\/jingdw\/\">Jingdong Wang<\/a><br \/>\n<a href=\"https:\/\/newed.any0.dpdns.org\/en-us\/research\/people\/jiaoyan\/\">Jiaolong Yang<\/a><br \/>\n<a href=\"https:\/\/newed.any0.dpdns.org\/en-us\/research\/people\/chazhang\/\">Cha Zhang<\/a><span id=\"label-external-link\" class=\"sr-only\" aria-hidden=\"true\">Opens in a new tab<\/span><\/p>\n","protected":false},"excerpt":{"rendered":"<p>Microsoft is proud to be a sponsor of the Computer Vision and Pattern Recognition (CVPR 2021) event.<\/p>\n","protected":false},"featured_media":750469,"template":"","meta":{"msr-url-field":"","msr-podcast-episode":"","msrModifiedDate":"","msrModifiedDateEnabled":false,"ep_exclude_from_search":false,"_classifai_error":"","msr_startdate":"2021-06-19","msr_enddate":"2021-06-25","msr_location":"Virtual","msr_expirationdate":"","msr_event_recording_link":"","msr_event_link":"","msr_event_link_redirect":false,"msr_event_time":"","msr_hide_region":false,"msr_private_event":false,"msr_hide_image_in_river":0,"footnotes":""},"research-area":[13562],"msr-region":[256048],"msr-event-type":[197941],"msr-video-type":[],"msr-locale":[268875],"msr-program-audience":[],"msr-post-option":[],"msr-impact-theme":[],"class_list":["post-747889","msr-event","type-msr-event","status-publish","has-post-thumbnail","hentry","msr-research-area-computer-vision","msr-region-global","msr-event-type-conferences","msr-locale-en_us"],"msr_about":"<!-- wp:msr\/event-details {\"title\":\"Microsoft at CVPR 2021\",\"backgroundColor\":\"grey\",\"image\":{\"id\":750469,\"url\":\"https:\/\/newed.any0.dpdns.org\/en-us\/research\/wp-content\/uploads\/2021\/05\/CVPR-2021_CV_header_1920x720.jpg\",\"alt\":\"\"},\"imageType\":\"full-bleed\"} \/-->\n\n<!-- wp:msr\/content-tabs --><!-- wp:msr\/content-tab {\"title\":\"About\"} --><!-- wp:freeform --><p><strong>Website:<\/strong> <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" href=\"http:\/\/cvpr2021.thecvf.com\/\" target=\"_blank\" rel=\"noopener\">CVPR 2021<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><span id=\"label-external-link\" class=\"sr-only\" aria-hidden=\"true\">Opens in a new tab<\/span><\/p>\n<p>Microsoft is proud to be a sponsor of the <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" href=\"http:\/\/cvpr2021.thecvf.com\/\" target=\"_blank\" rel=\"noopener\">Computer Vision and Pattern Recognition (CVPR 2021)<\/a> event.<\/p>\n<div style=\"height: 20px\"><\/div>\n<h3>Synthetic Data with Digital Humans<\/h3>\n<p><em>Microsoft Sponsor Session<\/em><\/p>\n<p>Join Erroll Wood and Tadas Baltrusaitis from Microsoft&#8217;s <a href=\"https:\/\/newed.any0.dpdns.org\/en-us\/research\/lab\/mixed-reality-ai-lab-cambridge\/\">Mixed Reality &amp; AI Lab in Cambridge, UK<\/a>, for a talk on how synthetics drives work on understanding human faces and hands, including how it powers Fully Articulated Hand Tracking on HoloLens 2.<\/p>\n<div class=\"video-wrapper margin-bottom-sp1\"><iframe loading=\"lazy\" title=\"Synthetic Data with Digital Humans\" width=\"500\" height=\"281\" src=\"https:\/\/www.youtube-nocookie.com\/embed\/4rRF4UMppjY?feature=oembed&#038;rel=0\" frameborder=\"0\" allow=\"accelerometer; autoplay; clipboard-write; encrypted-media; gyroscope; picture-in-picture; web-share\" referrerpolicy=\"strict-origin-when-cross-origin\" allowfullscreen><\/iframe><\/div>\n<div style=\"height: 30px\"><\/div>\n<h3>Longuet-Higgins Prize Winner<\/h3>\n<p><strong>Real-time human pose recognition in parts from single depth image<br \/>\n<\/strong><a href=\"https:\/\/newed.any0.dpdns.org\/en-us\/research\/people\/jamiesho\/\">Jamie Shotton<\/a>, <a href=\"https:\/\/newed.any0.dpdns.org\/en-us\/research\/people\/awf\/\">Andrew Fitzgibbon<\/a>, Mat Cook, <a href=\"https:\/\/newed.any0.dpdns.org\/en-us\/research\/people\/tsharp\/\">Toby Sharp<\/a>, Mark Finocchio, Richard Moore, Alex Kipman, <a href=\"https:\/\/newed.any0.dpdns.org\/en-us\/research\/people\/ablake\/\">Andrew Blake<\/a><\/p>\n<div style=\"height: 20px\"><\/div>\n<h3>Nominations for CVPR 2021 Best Paper Award<\/h3>\n<p><strong>Privacy-Preserving Image Features via Adversarial Affine Subspace Embeddings<\/strong><br \/>\nMihai Dusmanu, <a href=\"https:\/\/newed.any0.dpdns.org\/en-us\/research\/people\/joschonb\/\">Johannes L Sch\u00f6nberger<\/a>, <a href=\"https:\/\/newed.any0.dpdns.org\/en-us\/research\/people\/sudipsin\/\">Sudipta Sinha<\/a>, <a href=\"https:\/\/newed.any0.dpdns.org\/en-us\/research\/people\/mapoll\/\">Marc Pollefeys<\/a><\/p>\n<p><strong>CoCosNet v2: Full-Resolution Correspondence Learning for Image Translation<\/strong><br \/>\n<a href=\"https:\/\/newed.any0.dpdns.org\/en-us\/research\/people\/zhanbo\/\">Bo Zhang<\/a>, <a href=\"https:\/\/newed.any0.dpdns.org\/en-us\/research\/people\/tinzhan\/\">Ting Zhang<\/a>, Pan Zhang, <a href=\"https:\/\/newed.any0.dpdns.org\/en-us\/research\/people\/jianbao\/\">Jianmin Bao<\/a>, <a href=\"https:\/\/newed.any0.dpdns.org\/en-us\/research\/people\/doch\/\">Dong Chen<\/a>, Zhongfei Zhang, <a href=\"https:\/\/newed.any0.dpdns.org\/en-us\/research\/people\/fangwen\/\">Fang Wen<\/a><\/p>\n<p><strong>Less Is More: ClipBERT for Video-and-Language Learning via Sparse Sampling<\/strong><br \/>\nJie Lei, <a href=\"https:\/\/newed.any0.dpdns.org\/en-us\/research\/people\/linjli\/\">Linjie Li<\/a>, <a href=\"https:\/\/newed.any0.dpdns.org\/en-us\/research\/people\/luozhou\/\">Luowei Zhou<\/a>, <a href=\"https:\/\/newed.any0.dpdns.org\/en-us\/research\/people\/zhgan\/\">Zhe Gan<\/a>, Tamara Berg, Mohit Bansal, Jingjing Liu<\/p>\n<div style=\"height: 20px\"><\/div>\n<h3>Area Chairs<\/h3>\n<p><a href=\"https:\/\/newed.any0.dpdns.org\/en-us\/research\/people\/meic\/\">Mei Chen<\/a>, Mentor for the Doctoral Consortium<br \/>\n<a href=\"https:\/\/newed.any0.dpdns.org\/en-us\/research\/people\/hanhu\/\">Han Hu<\/a><br \/>\n<a href=\"https:\/\/newed.any0.dpdns.org\/en-us\/research\/people\/jingdw\/\">Jingdong Wang<\/a><br \/>\n<a href=\"https:\/\/newed.any0.dpdns.org\/en-us\/research\/people\/jiaoyan\/\">Jiaolong Yang<\/a><br \/>\n<a href=\"https:\/\/newed.any0.dpdns.org\/en-us\/research\/people\/chazhang\/\">Cha Zhang<\/a><span id=\"label-external-link\" class=\"sr-only\" aria-hidden=\"true\">Opens in a new tab<\/span><\/p>\n<!-- \/wp:freeform --><!-- \/wp:msr\/content-tab --><!-- \/wp:msr\/content-tabs -->","tab-content":[{"id":0,"name":"About","content":"Microsoft is proud to be a sponsor of the <a href=\"http:\/\/cvpr2021.thecvf.com\/\" target=\"_blank\" rel=\"noopener\">Computer Vision and Pattern Recognition (CVPR 2021)<\/a> event.\r\n<div style=\"height: 20px\"><\/div>\r\n<h3>Synthetic Data with Digital Humans<\/h3>\r\n<em>Microsoft Sponsor Session<\/em>\r\n\r\nJoin Erroll Wood and Tadas Baltrusaitis from Microsoft's <a href=\"https:\/\/newed.any0.dpdns.org\/en-us\/research\/lab\/mixed-reality-ai-lab-cambridge\/\">Mixed Reality &amp; AI Lab in Cambridge, UK<\/a>, for a talk on how synthetics drives work on understanding human faces and hands, including how it powers Fully Articulated Hand Tracking on HoloLens 2.\r\n\r\nhttps:\/\/youtu.be\/4rRF4UMppjY\r\n<div style=\"height: 30px\"><\/div>\r\n<h3>Longuet-Higgins Prize Winner<\/h3>\r\n<strong>Real-time human pose recognition in parts from single depth image\r\n<\/strong><a href=\"https:\/\/newed.any0.dpdns.org\/en-us\/research\/people\/jamiesho\/\">Jamie Shotton<\/a>, <a href=\"https:\/\/newed.any0.dpdns.org\/en-us\/research\/people\/awf\/\">Andrew Fitzgibbon<\/a>, Mat Cook, <a href=\"https:\/\/newed.any0.dpdns.org\/en-us\/research\/people\/tsharp\/\">Toby Sharp<\/a>, Mark Finocchio, Richard Moore, Alex Kipman, <a href=\"https:\/\/newed.any0.dpdns.org\/en-us\/research\/people\/ablake\/\">Andrew Blake<\/a>\r\n<div style=\"height: 20px\"><\/div>\r\n<h3>Nominations for CVPR 2021 Best Paper Award<\/h3>\r\n<strong>Privacy-Preserving Image Features via Adversarial Affine Subspace Embeddings<\/strong>\r\nMihai Dusmanu, <a href=\"https:\/\/newed.any0.dpdns.org\/en-us\/research\/people\/joschonb\/\">Johannes L Sch\u00f6nberger<\/a>, <a href=\"https:\/\/newed.any0.dpdns.org\/en-us\/research\/people\/sudipsin\/\">Sudipta Sinha<\/a>, <a href=\"https:\/\/newed.any0.dpdns.org\/en-us\/research\/people\/mapoll\/\">Marc Pollefeys<\/a>\r\n\r\n<strong>CoCosNet v2: Full-Resolution Correspondence Learning for Image Translation<\/strong>\r\n<a href=\"https:\/\/newed.any0.dpdns.org\/en-us\/research\/people\/zhanbo\/\">Bo Zhang<\/a>, <a href=\"https:\/\/newed.any0.dpdns.org\/en-us\/research\/people\/tinzhan\/\">Ting Zhang<\/a>, Pan Zhang, <a href=\"https:\/\/newed.any0.dpdns.org\/en-us\/research\/people\/jianbao\/\">Jianmin Bao<\/a>, <a href=\"https:\/\/newed.any0.dpdns.org\/en-us\/research\/people\/doch\/\">Dong Chen<\/a>, Zhongfei Zhang, <a href=\"https:\/\/newed.any0.dpdns.org\/en-us\/research\/people\/fangwen\/\">Fang Wen<\/a>\r\n\r\n<strong>Less Is More: ClipBERT for Video-and-Language Learning via Sparse Sampling<\/strong>\r\nJie Lei, <a href=\"https:\/\/newed.any0.dpdns.org\/en-us\/research\/people\/linjli\/\">Linjie Li<\/a>, <a href=\"https:\/\/newed.any0.dpdns.org\/en-us\/research\/people\/luozhou\/\">Luowei Zhou<\/a>, <a href=\"https:\/\/newed.any0.dpdns.org\/en-us\/research\/people\/zhgan\/\">Zhe Gan<\/a>, Tamara Berg, Mohit Bansal, Jingjing Liu\r\n<div style=\"height: 20px\"><\/div>\r\n<h3>Area Chairs<\/h3>\r\n<a href=\"https:\/\/newed.any0.dpdns.org\/en-us\/research\/people\/meic\/\">Mei Chen<\/a>, Mentor for the Doctoral Consortium\r\n<a href=\"https:\/\/newed.any0.dpdns.org\/en-us\/research\/people\/hanhu\/\">Han Hu<\/a>\r\n<a href=\"https:\/\/newed.any0.dpdns.org\/en-us\/research\/people\/jingdw\/\">Jingdong Wang<\/a>\r\n<a href=\"https:\/\/newed.any0.dpdns.org\/en-us\/research\/people\/jiaoyan\/\">Jiaolong Yang<\/a>\r\n<a href=\"https:\/\/newed.any0.dpdns.org\/en-us\/research\/people\/chazhang\/\">Cha Zhang<\/a>"}],"msr_startdate":"2021-06-19","msr_enddate":"2021-06-25","msr_event_time":"","msr_location":"Virtual","msr_event_link":"","msr_event_recording_link":"","msr_startdate_formatted":"June 19, 2021","msr_register_text":"Watch now","msr_cta_link":"","msr_cta_text":"","msr_cta_bi_name":"","featured_image_thumbnail":"<img width=\"960\" height=\"540\" src=\"https:\/\/newed.any0.dpdns.org\/en-us\/research\/wp-content\/uploads\/2021\/05\/CVPR-2021_CV_header_1920x720-960x540.jpg\" class=\"img-object-cover\" alt=\"CVPR 2021 header: organic network on dark background\" decoding=\"async\" loading=\"lazy\" srcset=\"https:\/\/newed.any0.dpdns.org\/en-us\/research\/wp-content\/uploads\/2021\/05\/CVPR-2021_CV_header_1920x720-960x540.jpg 960w, https:\/\/newed.any0.dpdns.org\/en-us\/research\/wp-content\/uploads\/2021\/05\/CVPR-2021_CV_header_1920x720-1066x600.jpg 1066w, https:\/\/newed.any0.dpdns.org\/en-us\/research\/wp-content\/uploads\/2021\/05\/CVPR-2021_CV_header_1920x720-655x368.jpg 655w, https:\/\/newed.any0.dpdns.org\/en-us\/research\/wp-content\/uploads\/2021\/05\/CVPR-2021_CV_header_1920x720-343x193.jpg 343w, https:\/\/newed.any0.dpdns.org\/en-us\/research\/wp-content\/uploads\/2021\/05\/CVPR-2021_CV_header_1920x720-640x360.jpg 640w, https:\/\/newed.any0.dpdns.org\/en-us\/research\/wp-content\/uploads\/2021\/05\/CVPR-2021_CV_header_1920x720-1280x720.jpg 1280w\" sizes=\"auto, (max-width: 960px) 100vw, 960px\" \/>","event_excerpt":"Microsoft is proud to be a sponsor of the Computer Vision and Pattern Recognition (CVPR 2021) event.","msr_research_lab":[],"related-researchers":[],"msr_impact_theme":[],"related-academic-programs":[],"related-groups":[],"related-projects":[],"related-opportunities":[],"related-publications":[749431,749701,749716,765223,766174,785446],"related-videos":[756043,756415],"related-posts":[],"_links":{"self":[{"href":"https:\/\/newed.any0.dpdns.org\/en-us\/research\/wp-json\/wp\/v2\/msr-event\/747889","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/newed.any0.dpdns.org\/en-us\/research\/wp-json\/wp\/v2\/msr-event"}],"about":[{"href":"https:\/\/newed.any0.dpdns.org\/en-us\/research\/wp-json\/wp\/v2\/types\/msr-event"}],"version-history":[{"count":3,"href":"https:\/\/newed.any0.dpdns.org\/en-us\/research\/wp-json\/wp\/v2\/msr-event\/747889\/revisions"}],"predecessor-version":[{"id":1146868,"href":"https:\/\/newed.any0.dpdns.org\/en-us\/research\/wp-json\/wp\/v2\/msr-event\/747889\/revisions\/1146868"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/newed.any0.dpdns.org\/en-us\/research\/wp-json\/wp\/v2\/media\/750469"}],"wp:attachment":[{"href":"https:\/\/newed.any0.dpdns.org\/en-us\/research\/wp-json\/wp\/v2\/media?parent=747889"}],"wp:term":[{"taxonomy":"msr-research-area","embeddable":true,"href":"https:\/\/newed.any0.dpdns.org\/en-us\/research\/wp-json\/wp\/v2\/research-area?post=747889"},{"taxonomy":"msr-region","embeddable":true,"href":"https:\/\/newed.any0.dpdns.org\/en-us\/research\/wp-json\/wp\/v2\/msr-region?post=747889"},{"taxonomy":"msr-event-type","embeddable":true,"href":"https:\/\/newed.any0.dpdns.org\/en-us\/research\/wp-json\/wp\/v2\/msr-event-type?post=747889"},{"taxonomy":"msr-video-type","embeddable":true,"href":"https:\/\/newed.any0.dpdns.org\/en-us\/research\/wp-json\/wp\/v2\/msr-video-type?post=747889"},{"taxonomy":"msr-locale","embeddable":true,"href":"https:\/\/newed.any0.dpdns.org\/en-us\/research\/wp-json\/wp\/v2\/msr-locale?post=747889"},{"taxonomy":"msr-program-audience","embeddable":true,"href":"https:\/\/newed.any0.dpdns.org\/en-us\/research\/wp-json\/wp\/v2\/msr-program-audience?post=747889"},{"taxonomy":"msr-post-option","embeddable":true,"href":"https:\/\/newed.any0.dpdns.org\/en-us\/research\/wp-json\/wp\/v2\/msr-post-option?post=747889"},{"taxonomy":"msr-impact-theme","embeddable":true,"href":"https:\/\/newed.any0.dpdns.org\/en-us\/research\/wp-json\/wp\/v2\/msr-impact-theme?post=747889"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}