{"id":868,"date":"2022-10-23T10:29:00","date_gmt":"2022-10-23T01:29:00","guid":{"rendered":"https:\/\/www.sonycsl.co.jp\/kyoto\/?post_type=news&#038;p=868"},"modified":"2023-04-17T11:11:49","modified_gmt":"2023-04-17T02:11:49","slug":"868","status":"publish","type":"news","link":"https:\/\/www.sonycsl.co.jp\/kyoto\/news\/868\/","title":{"rendered":"\u767a\u8868\u6587\u732e\u306e\u30da\u30fc\u30b8\u3092\u66f4\u65b0\u3057\u307e\u3057\u305f"},"content":{"rendered":"\n<p>DualVoice: Speech Interaction that Discriminates between Normal and Whispered Voice Input <\/p>\n\n\n\n<p>Jun Rekimoto | ACM UIST 2022 <\/p>\n\n\n\n<p><\/p>\n\n\n\n<p>NeARportation: A Remote Real-time Neural Rendering Framework <\/p>\n\n\n\n<p>Yuichi Hiroi, Yuta Itoh, Jun Rekimoto | ACM VRST 2022 <\/p>\n\n\n\n<p><a href=\"https:\/\/www.sonycsl.co.jp\/kyoto\/publications\/\">https:\/\/www.sonycsl.co.jp\/kyoto\/publications\/<\/a><\/p>\n","protected":false},"template":"","class_list":["post-868","news","type-news","status-publish","hentry"],"aioseo_notices":[],"_links":{"self":[{"href":"https:\/\/www.sonycsl.co.jp\/kyoto\/wp-json\/wp\/v2\/news\/868","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.sonycsl.co.jp\/kyoto\/wp-json\/wp\/v2\/news"}],"about":[{"href":"https:\/\/www.sonycsl.co.jp\/kyoto\/wp-json\/wp\/v2\/types\/news"}],"version-history":[{"count":8,"href":"https:\/\/www.sonycsl.co.jp\/kyoto\/wp-json\/wp\/v2\/news\/868\/revisions"}],"predecessor-version":[{"id":898,"href":"https:\/\/www.sonycsl.co.jp\/kyoto\/wp-json\/wp\/v2\/news\/868\/revisions\/898"}],"wp:attachment":[{"href":"https:\/\/www.sonycsl.co.jp\/kyoto\/wp-json\/wp\/v2\/media?parent=868"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}