{"id":3470,"date":"2019-04-15T13:30:04","date_gmt":"2019-04-15T04:30:04","guid":{"rendered":"http:\/\/ahcweb01.naist.jp\/?p=3470"},"modified":"2019-04-15T13:36:45","modified_gmt":"2019-04-15T04:36:45","slug":"speaker-and-emotion-recognition-of-tv-series-data-using-multimodal-and-multitask-deep-learning","status":"publish","type":"post","link":"https:\/\/ahcweb02.naist.jp\/2019\/04\/15\/speaker-and-emotion-recognition-of-tv-series-data-using-multimodal-and-multitask-deep-learning\/","title":{"rendered":"Speaker and Emotion Recognition of TV-Series Data Using Multimodal and Multitask Deep Learning"},"content":{"rendered":"","protected":false},"excerpt":{"rendered":"","protected":false},"author":6,"featured_media":0,"comment_status":"closed","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":[],"categories":[13],"tags":[],"_links":{"self":[{"href":"https:\/\/ahcweb02.naist.jp\/wp-json\/wp\/v2\/posts\/3470"}],"collection":[{"href":"https:\/\/ahcweb02.naist.jp\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/ahcweb02.naist.jp\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/ahcweb02.naist.jp\/wp-json\/wp\/v2\/users\/6"}],"replies":[{"embeddable":true,"href":"https:\/\/ahcweb02.naist.jp\/wp-json\/wp\/v2\/comments?post=3470"}],"version-history":[{"count":1,"href":"https:\/\/ahcweb02.naist.jp\/wp-json\/wp\/v2\/posts\/3470\/revisions"}],"predecessor-version":[{"id":3471,"href":"https:\/\/ahcweb02.naist.jp\/wp-json\/wp\/v2\/posts\/3470\/revisions\/3471"}],"wp:attachment":[{"href":"https:\/\/ahcweb02.naist.jp\/wp-json\/wp\/v2\/media?parent=3470"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/ahcweb02.naist.jp\/wp-json\/wp\/v2\/categories?post=3470"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/ahcweb02.naist.jp\/wp-json\/wp\/v2\/tags?post=3470"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}