{"id":2905,"date":"2018-05-21T19:12:56","date_gmt":"2018-05-21T10:12:56","guid":{"rendered":"http:\/\/ahcweb01.naist.jp\/?p=2905"},"modified":"2018-05-21T19:13:19","modified_gmt":"2018-05-21T10:13:19","slug":"multi-modal-multi-task-deep-learning-for-speaker-and-emotion-recognition-of-tv-series-data","status":"publish","type":"post","link":"https:\/\/ahcweb02.naist.jp\/2018\/05\/21\/multi-modal-multi-task-deep-learning-for-speaker-and-emotion-recognition-of-tv-series-data\/","title":{"rendered":"Multi-Modal Multi-Task Deep Learning for Speaker and Emotion Recognition of TV-Series Data"},"content":{"rendered":"","protected":false},"excerpt":{"rendered":"","protected":false},"author":6,"featured_media":0,"comment_status":"closed","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":[],"categories":[13],"tags":[],"_links":{"self":[{"href":"https:\/\/ahcweb02.naist.jp\/wp-json\/wp\/v2\/posts\/2905"}],"collection":[{"href":"https:\/\/ahcweb02.naist.jp\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/ahcweb02.naist.jp\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/ahcweb02.naist.jp\/wp-json\/wp\/v2\/users\/6"}],"replies":[{"embeddable":true,"href":"https:\/\/ahcweb02.naist.jp\/wp-json\/wp\/v2\/comments?post=2905"}],"version-history":[{"count":1,"href":"https:\/\/ahcweb02.naist.jp\/wp-json\/wp\/v2\/posts\/2905\/revisions"}],"predecessor-version":[{"id":2906,"href":"https:\/\/ahcweb02.naist.jp\/wp-json\/wp\/v2\/posts\/2905\/revisions\/2906"}],"wp:attachment":[{"href":"https:\/\/ahcweb02.naist.jp\/wp-json\/wp\/v2\/media?parent=2905"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/ahcweb02.naist.jp\/wp-json\/wp\/v2\/categories?post=2905"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/ahcweb02.naist.jp\/wp-json\/wp\/v2\/tags?post=2905"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}