{"id":1464,"date":"2020-04-03T14:03:11","date_gmt":"2020-04-03T05:03:11","guid":{"rendered":"https:\/\/sr.sfc.keio.ac.jp\/www\/?page_id=1464"},"modified":"2026-03-31T12:40:51","modified_gmt":"2026-03-31T03:40:51","slug":"publications-2","status":"publish","type":"page","link":"https:\/\/sr.sfc.keio.ac.jp\/www\/en\/publications-2\/","title":{"rendered":"Publications"},"content":{"rendered":"<h2 class = \"adjust-margin-1\">2026\u301c<\/h2>\n<hr \/>\n\n\n<div class=\"su-posts su-posts-default-loop \">\n\n\t\n\t\t\t\t\t\n\t\t\t\n\t\t\t<div id=\"su-post-3157\" class=\"su-post \">\n\n\t\t\t\t\n\t\t\t\t<h2 class=\"su-post-title\"><a href=\"https:\/\/sr.sfc.keio.ac.jp\/www\/en\/publication\/3157\/\">Adaptive Control of Utterance Timing based on User Conversational Silence in Spoken Dialogue Systems, Yosuke Ujigawa\u30fbKazunori Takashio, Human-Agent Interaction Symposium 2026, Feb. 2026<\/a><\/h2>\n\n\t\t\t\t<div class=\"su-post-meta\">\n\t\t\t\t\tPosted: 2026\u5e743\u670831\u65e5\t\t\t\t<\/div>\n\n\t\t\t\t<div class=\"su-post-excerpt\">\n\t\t\t\t\t<p>\u5186\u6ed1\u3067\u5354\u529b\u7684\u306a\u5bfe\u4eba\u30b3\u30df\u30e5\u30cb\u30b1\u30fc\u30b7\u30e7\u30f3\u306b\u304a\u3044\u3066\uff0c\u8a71\u8005\u9593\u306e\u4ea4\u66ff\u6f5c\u6642\u304c\u76f8\u4e92\u306b\u5f15\u304d\u8fbc\u307e \u308c\uff0c\u540c\u8abf\u3057\u3066\u3044\u304f\u73fe\u8c61\u306f\u5e83\u304f\u77e5\u3089\u308c\u3066\u3044\u308b\uff0e\u672c\u7814\u7a76\u3067\u306f\uff0c\u975e\u8a00\u8a9e\u7684\u306a\u30bf\u30a4\u30df\u30f3\u30b0\u306e\u540c\u8abf\u304c\uff0c\u5bfe\u8a71\u30a8\u30fc \u30b8\u30a7\u30f3\u30c8\u306b\u5bfe\u3059\u308b\u793e\u4f1a\u7684\u53d7\u5bb9\u6027\u3084\u5bfe\u8a71\u306e\u8cea\u306b\u4e0e\u3048\u308b\u5f71\u97ff\u3092\u660e\u3089\u304b\u306b\u3059\u308b\u3053\u3068\u3092\u76ee\u7684\u3068\u3059\u308b\uff0e\u30e6\u30fc\u30b6\u30fc \u306e\u6f5c\u6642\u3092\u5bfe\u8a71\u306e\u4e2d\u3067\u8a08\u6e2c\u3057\uff0c\u6f5c\u6642\u3092\u52d5\u7684\u306b\u5236\u5fa1\u3059\u308b\u5bfe\u8a71\u30b7\u30b9\u30c6\u30e0\u3092\u5b9f\u88c5\u3057\u305f\uff0e\u300c\u53ce\u675f\u300d\u3068\u300c\u540c\u671f\u300d\u306e 2 \u3064\u306e\u540c\u8abf\u624b\u6cd5\u306e\u6709\u52b9\u6027\uff0c\u304a\u3088\u3073\u5bfe\u8a71\u306b\u4e0e\u3048\u308b\u5f71\u97ff\u306b\u3064\u3044\u3066\uff0c\u4e3b\u89b3\u8a55\u4fa1\u3068\u751f\u4f53\u6307\u6a19\u3092\u7528\u3044\u305f\u5ba2\u89b3\u7684\u6307 \u6a19\u306e\u4e21\u9762\u304b\u3089\u5206\u6790\u3092\u884c\u3063\u305f\uff0e<\/p>\n\t\t\t\t<\/div>\n\n\t\t\t\t\n\t\t\t<\/div>\n\n\t\t\n\t\n<\/div>\n\n<h2 class = \"adjust-margin-1\">2025\u301c<\/h2>\n<hr \/>\n\n\n<div class=\"su-posts su-posts-default-loop \">\n\n\t\n\t\t\t\t\t\n\t\t\t\n\t\t\t<div id=\"su-post-2968\" class=\"su-post \">\n\n\t\t\t\t\n\t\t\t\t<h2 class=\"su-post-title\"><a href=\"https:\/\/sr.sfc.keio.ac.jp\/www\/en\/publication\/2968\/\">\u5171\u540c\u56f3\u5f62\u914d\u7f6e\u8ab2\u984c\u3092\u5bfe\u8c61\u3068\u3057\u305f\u30e2\u30c0\u30ea\u30c6\u30a3\u3068\u793e\u4f1a\u7684\u95a2\u4fc2\u6027\u306e\u5171\u901a\u57fa\u76e4\u69cb\u7bc9\u3078\u306e\u5f71\u97ff\u5206\u6790, \u53e4\u8c37 \u512a\u6a39\u30fb\u9f4b\u85e4 \u5149\u8f1d\u30fb\u5c0f\u5009 \u529f\u88d5\u30fb\u7dd1\u5ddd \u8a60\u4ecb\u30fb\u5149\u7530 \u822a\u30fb\u6771\u4e2d \u7adc\u4e00\u90ce\u30fb\u9ad8\u6c50 \u4e00\u7d00, \u77e5\u80fd\u3068\u60c5\u5831, 2025, 37 \u5dfb, 3 \u53f7, p. 662-670, 2025\u5e748\u6708<\/a><\/h2>\n\n\t\t\t\t<div class=\"su-post-meta\">\n\t\t\t\t\tPosted: 2025\u5e749\u67082\u65e5\t\t\t\t<\/div>\n\n\t\t\t\t<div class=\"su-post-excerpt\">\n\t\t\t\t\t<p>\u30ed\u30dc\u30c3\u30c8\u3084\u30d0\u30fc\u30c1\u30e3\u30eb\u30a8\u30fc\u30b8\u30a7\u30f3\u30c8\u304c\u30e6\u30fc\u30b6\u3068\u81ea\u7136\u306b\u5bfe\u8a71\u3059\u308b\u305f\u3081\u306b\u306f\uff0c\u5171\u901a\u57fa\u76e4\u306e\u69cb\u7bc9\u304c\u4e0d\u53ef\u6b20\u3067\u3042\u308b\uff0e\u4eba\u9593\u540c\u58eb\u306e\u5bfe\u8a71\u306b\u304a\u3051\u308b\u5171\u901a\u57fa\u76e4\u306e\u69cb\u7bc9\u904e\u7a0b\u3092\u8abf\u67fb\u3057\u305f\u5148\u884c\u7814\u7a76\u306f\u5c11\u306a\u304f\uff0c\u591a\u304f\u306f\u30c6\u30ad\u30b9\u30c8\u30c1\u30e3\u30c3\u30c8\u3092\u5bfe\u8c61\u3068\u3057\u3066\u3044\u308b\uff0e\u305d\u3053\u3067\uff0c\u672c\u7814\u7a76\u3067\u306f\uff0c\u5bfe\u8a71\u306e\u30e2\u30c0\u30ea\u30c6\u30a3\uff08\u97f3\u58f0\u30fb\u6620\u50cf\uff09\u3068\u8a71\u8005\u9593\u306e\u793e\u4f1a\u7684\u95a2\u4fc2\u6027\uff08\u521d\u5bfe\u9762\u30fb\u77e5\u4eba\uff09\u304c\u5171\u901a\u57fa\u76e4\u69cb\u7bc9\u306b\u4e0e\u3048\u308b\u5f71\u97ff\u3092\u8abf\u67fb\u3057\u305f\uff0e\u5177\u4f53\u7684\u306b\u306f\uff0c\u3053\u308c\u3089\u306e\u6761\u4ef6\u4e0b\u3067\u5171\u901a\u57fa\u76e4\u304c\u69cb\u7bc9\u3055\u308c\u308b\u5bfe\u8a71\u306e\u53ce\u96c6\u3068\u5206\u6790\u3092\u884c\u3063\u305f\uff0e\u305d\u306e\u7d50\u679c\uff0c\u30e2\u30c0\u30ea\u30c6\u30a3\u306e\u62e1\u5f35\u3084\u77e5\u4eba\u95a2\u4fc2\u304c\u5171\u901a\u57fa\u76e4\u306e\u69cb\u7bc9\u3092\u52a0\u901f\u3055\u305b\u308b\u3053\u3068\u304c\u660e\u3089\u304b\u306b\u306a\u3063\u305f\uff0e\u8a00\u8a9e\u884c\u52d5\u306e\u5206\u6790\u306b\u3088\u308a\uff0c\u30e2\u30c0\u30ea\u30c6\u30a3\u306e\u62e1\u5f35\u304c\u5171\u611f\u306e\u610f\u56f3\u4f1d\u9054\u3092\u5bb9\u6613\u306b\u3057\uff0c\u77e5\u4eba\u95a2\u4fc2\u304c\u30bf\u30b9\u30af\u9032\u884c\u306b\u5fc5\u8981\u306a\u767a\u8a71\u3092\u5897\u52a0\u3055\u305b\u308b\u3053\u3068\u304c\u5224\u660e\u3057\u305f\uff0e\u3053\u308c\u306b\u3088\u308a\uff0c\u30ed\u30dc\u30c3\u30c8\u3084\u30d0\u30fc\u30c1\u30e3\u30eb\u30a8\u30fc\u30b8\u30a7\u30f3\u30c8\u306e\u8a2d\u8a08\u306b\u306f\uff0c\u5171\u611f\u3092\u4f1d\u3048\u308b\u975e\u8a00\u8a9e\u884c\u52d5\u3084\u77e5\u4eba\u540c\u58eb\u306e\u767a\u8a71\u884c\u52d5\u306e\u518d\u73fe\u304c\u91cd\u8981\u3067\u3042\u308b\u3068\u793a\u5506\u3055\u308c\u305f\uff0e<\/p>\n\t\t\t\t<\/div>\n\n\t\t\t\t\n\t\t\t<\/div>\n\n\t\t\t\t\t\n\t\t\t\n\t\t\t<div id=\"su-post-2954\" class=\"su-post \">\n\n\t\t\t\t\n\t\t\t\t<h2 class=\"su-post-title\"><a href=\"https:\/\/sr.sfc.keio.ac.jp\/www\/en\/publication\/2954\/\">Empirical Evaluation of Healthcare Communication Robot Encouraging Self-Disclosure of Chronic Pain, Airi Shimada\u30fbKazunori Takashio, The 2024 IEEE International Conference on Robot and Human Interactive Communication (RO-MAN),2025\u5e748\u6708<\/a><\/h2>\n\n\t\t\t\t<div class=\"su-post-meta\">\n\t\t\t\t\tPosted: 2025\u5e749\u67082\u65e5\t\t\t\t<\/div>\n\n\t\t\t\t<div class=\"su-post-excerpt\">\n\t\t\t\t\t<p>Self-disclosure of pain is essential to communicate pain, which is a subjective sensation, to a third party. However,many elderly people, especially those with chronic pain, are hesitant to communicate their pain. As a result, many patients do not receive appropriate treatment at the right time. It is important to detect small discomfort in daily life and not overlook pain occurrence or changes. The ultimate goal of this study is to create a robot for people with chronic pain that notices the user \u2019s discomfort through many modalities in daily interactions, and tell the recorded information to a hospital or family if necessary. In this study, we conducted fieldwork at Nichinan Hospital and based on the findings, we propose and verify the dialogue system that encourages selfdisclosure of pain. In this paper, we implemented a system that detects discomfort based on the user \u2019s utterance about pain and the action of rubbing, and asks detailed questions about the pain. We conducted a demonstration experiment with patients of Nichinan Hospital, and the content of the dialogue was evaluated by a physical therapist. The proposed method received significantly higher ratings for the naturalness of the conversation, the ease of use of the system, and the length of the conversation. The physical therapist\u2019s evaluation suggested that the ability of the dialogue system to \u201dnotice\u201d the user \u2019s discomfort or unusualness had a positive effect on facilitating pain communication and encouraging self-disclosure. The results suggest that it is possible to realize a dialogue system that facilitates self-disclosure of pain by users who suffer from chronic pain.<\/p>\n\t\t\t\t<\/div>\n\n\t\t\t\t\n\t\t\t<\/div>\n\n\t\t\t\t\t\n\t\t\t\n\t\t\t<div id=\"su-post-2940\" class=\"su-post \">\n\n\t\t\t\t\n\t\t\t\t<h2 class=\"su-post-title\"><a href=\"https:\/\/sr.sfc.keio.ac.jp\/www\/en\/publication\/2940\/\">Impact Analysis of Switching Pause Synchronization for Spoken Dialogue Systems, Yosuke Ujigawa\u30fbKazunori Takashio,  The 2025 IEEE International Conference on Robot and Human Interactive Communication (RO-MAN), 2025\u5e748\u6708<\/a><\/h2>\n\n\t\t\t\t<div class=\"su-post-meta\">\n\t\t\t\t\tPosted: 2025\u5e749\u67082\u65e5\t\t\t\t<\/div>\n\n\t\t\t\t<div class=\"su-post-excerpt\">\n\t\t\t\t\t<p>Each individual has a unique mental tempo (referred to as personal tempo), and the alignment of this tempo plays a crucial role in facilitating smooth interactions with spoken dialogue systems. This study focuses on the \u201cswitching pause,\u201d a key component of conversational tempo that is established during interaction. Using a dialogue corpus, we analyzed the impact of switching pauses on dialogue and the process of synchronization. Through the analysis of synchronization between pairs, we examined dialogues with high similarity in switching pauses to elucidate the impact of this synchronization on goal achievement and cooperativity in dialogue. Furthermore, we conducted a time-series analysis within pairs to investigate the synchronization process and proposed a method for determining switching pauses for implementation in dialogue systems. These findings contribute significantly to the investigation of individual differences in users and the identification of personal factors that enable effective dialogue with dialogue systems<\/p>\n\t\t\t\t<\/div>\n\n\t\t\t\t\n\t\t\t<\/div>\n\n\t\t\t\t\t\n\t\t\t\n\t\t\t<div id=\"su-post-2926\" class=\"su-post \">\n\n\t\t\t\t\n\t\t\t\t<h2 class=\"su-post-title\"><a href=\"https:\/\/sr.sfc.keio.ac.jp\/www\/en\/publication\/2926\/\">Exploring the Impact of Modalities on Building Common Ground Using the Collaborative Scene Reconstruction Task Yosuke Ujigawa\u30fbAsuka Shiotani\u30fbMasato Takizawa\u30fbEisuke Midorikawa\u30fbRyuichiro Higashinaka\u30fbKazunori Takashio, IWSDS2025, 2025\u5e745\u6708<\/a><\/h2>\n\n\t\t\t\t<div class=\"su-post-meta\">\n\t\t\t\t\tPosted: 2025\u5e749\u67082\u65e5\t\t\t\t<\/div>\n\n\t\t\t\t<div class=\"su-post-excerpt\">\n\t\t\t\t\t<p>To deepen our understanding of verbal and non-verbal modalities in establishing commonground, this study introduces a novel \u201ccollaborative scene reconstruction task.&#8221; In this task, pairs of participants, each provided with distinct image sets derived from the same video, work together to reconstruct the sequence of the original video. The level of agreement between the participants on the image order\u2014quantified using Kendall\u2019s rank correlation coefficient\u2014serves as a measure of common ground construction. This approach enables the analysis of how various modalities contribute to the construction of commonground. A corpus comprising 40 dialogues from 20 participants was collected and analyzed. The findings suggest that specific gestures play a significant role in fostering common ground, offering valuable insights for thedevelopment of dialogue systems that leverage multimodal information to enhance the user construction of common ground.<\/p>\n\t\t\t\t<\/div>\n\n\t\t\t\t\n\t\t\t<\/div>\n\n\t\t\t\t\t\n\t\t\t\n\t\t\t<div id=\"su-post-2912\" class=\"su-post \">\n\n\t\t\t\t\n\t\t\t\t<h2 class=\"su-post-title\"><a href=\"https:\/\/sr.sfc.keio.ac.jp\/www\/en\/publication\/2912\/\">Transparent Barriers: Natural Language Access Control Policies for XR-Enhanced Everyday Objects, Kentaro Taninaka\u30fbRahul Jain\u30fbJingyu Shi\u30fbKazunori Takashio\u30fbKarthik Ramani, CHI2025, 2025\u5e744\u6708<\/a><\/h2>\n\n\t\t\t\t<div class=\"su-post-meta\">\n\t\t\t\t\tPosted: 2025\u5e749\u67082\u65e5\t\t\t\t<\/div>\n\n\t\t\t\t<div class=\"su-post-excerpt\">\n\t\t\t\t\t<p>Extended Reality (XR)-enabled headsets that overlay digital content onto the physical world, are gradually finding their way into our daily life. This integration raises significant concerns about privacy and access control, especially in shared spaces where XR applications interact with everyday objects. Such issues remain subtle in the absence of widespread applications of XR and studies in shared spaces are required for a smooth progress. This study evaluated a prototype system facilitating natural language policy creation for flexible, context-aware access control of personal objects. We assessed its usability, focusing on balancing precision and user effort in creating access control policies. Qualitative interviews and task-based interactions provided insights into users\u2019 preferences and behaviors, informing future design directions. Findings revealed diverse user needs for controlling access to personal items in various situations, emphasizing the need for flexible, user-friendly access control in XR-enhanced shared spaces that respects boundaries and considers social contexts.<\/p>\n\t\t\t\t<\/div>\n\n\t\t\t\t\n\t\t\t<\/div>\n\n\t\t\t\t\t\n\t\t\t\n\t\t\t<div id=\"su-post-2521\" class=\"su-post \">\n\n\t\t\t\t\n\t\t\t\t<h2 class=\"su-post-title\"><a href=\"https:\/\/sr.sfc.keio.ac.jp\/www\/en\/publication\/2521\/\">\u4ea4\u66ff\u6f5c\u6642\u306e\u540c\u8abf\u304c\u5bfe\u8a71\u306b\u4e0e\u3048\u308b\u5f71\u97ff\u5206\u6790\u3068\u6c7a\u5b9a\u30e2\u30c7\u30eb\u306e\u69cb\u7bc9\uff0cYosuke Ujigawa \u30fbKazunori Takashio \uff0cHuman-Agent Interaction Symposium 2025\uff0c2025\u5e742\u6708<\/a><\/h2>\n\n\t\t\t\t<div class=\"su-post-meta\">\n\t\t\t\t\tPosted: 2025\u5e744\u67089\u65e5\t\t\t\t<\/div>\n\n\t\t\t\t<div class=\"su-post-excerpt\">\n\t\t\t\t\t<p>\u4eba\u9593\u306f\u56fa\u6709\u306e\u7cbe\u795e\u30c6\u30f3\u30dd\uff08\u30d1\u30fc\u30bd\u30ca\u30eb\u30c6\u30f3\u30dd\uff09\u3092\u6301\u3061\uff0c\u30c6\u30f3\u30dd\u306e\u4e00\u81f4\u304c\u30b7\u30b9\u30c6\u30e0\u3068\u306e\u5186\u6ed1\u306a\u5bfe\u8a71\u306b\u304a\u3044\u3066\u91cd\u8981\u306a\u5f79\u5272\u3092\u679c\u305f\u3059\uff0e\u672c\u7814\u7a76\u306f\uff0c\u5bfe\u8a71\u306b\u304a\u3051\u308b\u30c6\u30f3\u30dd\u306e\u540c\u8abf\uff0c\u7279\u306b\u4ea4\u66ff\u6f5c\u6642\u3068\u767a\u8a71\u901f\u5ea6\u306e\u95a2\u4fc2\u6027\u306b\u7740\u76ee\u3057\uff0c\u3053\u308c\u3089\u304c\u52b9\u679c\u7684\u306a\u30b3\u30df\u30e5\u30cb\u30b1\u30fc\u30b7\u30e7\u30f3\u306b\u679c\u305f\u3059\u5f79\u5272\u3092\u5206\u6790\u3057\u305f\uff0e\u52d5\u7684\u6642\u9593\u4f38\u7e2e\u6cd5\u3092\u7528\u3044\u305f\u5206\u6790\u306b\u3088\u308a\uff0c\u4ea4\u66ff\u6f5c\u6642\u306e\u540c\u8abf\u304c\u5bfe\u8a71\u306b\u4e0e\u3048\u308b\u5f71\u97ff\u3092\u660e\u3089\u304b\u306b\u3057\u305f\uff0e\u3055\u3089\u306b\uff0c\u767a\u8a71\u901f\u5ea6\u3068\u4ea4\u66ff\u6f5c\u6642\u306e\u76f8\u95a2\u95a2\u4fc2\u3092\u660e\u3089\u304b\u306b\u3057\uff0c\u97f3\u58f0\u5bfe\u8a71\u30b7\u30b9\u30c6\u30e0\u306e\u305f\u3081\u306e\u767a\u8a71\u30bf\u30a4\u30df\u30f3\u30b0\u6c7a\u5b9a\u30e2\u30c7\u30eb\u3092\u69cb\u7bc9\u3057\u305f\uff0e\u3053\u306e\u7814\u7a76\u6210\u679c\u306f\uff0c\u3088\u308a\u81ea\u7136\u3067\u9069\u5fdc\u7684\u306a\u5bfe\u8a71\u30b7\u30b9\u30c6\u30e0\u306e\u958b\u767a\u306b\u5bc4\u4e0e\u3059\u308b\u3082\u306e\u3067\u3042\u308b\uff0e<\/p>\n\t\t\t\t<\/div>\n\n\t\t\t\t\n\t\t\t<\/div>\n\n\t\t\n\t\n<\/div>\n\n<h2 class = \"adjust-margin-1\">2024\u301c<\/h2>\n<hr \/>\n\n\n<div class=\"su-posts su-posts-default-loop \">\n\n\t\n\t\t\t\t\t\n\t\t\t\n\t\t\t<div id=\"su-post-2445\" class=\"su-post \">\n\n\t\t\t\t\n\t\t\t\t<h2 class=\"su-post-title\"><a href=\"https:\/\/sr.sfc.keio.ac.jp\/www\/en\/publication\/2445\/\">Synchronization of Speech Rate to User&#8217;s Personal Tempo in Dialogue Systems and Its Effects,Yosuke Ujigawa\u30fbKazunori Takashio\uff08Keio Univ\uff09,2024 IEEE International Conference on Robot and Human Interactive Communication (RO-MAN),2024\u5e748\u6708<\/a><\/h2>\n\n\t\t\t\t<div class=\"su-post-meta\">\n\t\t\t\t\tPosted: 2024\u5e7410\u670816\u65e5\t\t\t\t<\/div>\n\n\t\t\t\t<div class=\"su-post-excerpt\">\n\t\t\t\t\t<p>Every individual lives in daily lives in own unique tempo, called Personal Tempo. Tempo is also highly important in dialogue situations, and it is thought that if the tempo can be matched with the conversational partner, it will lead to smoother communication with a higher level of comprehension. Spoken-dialogue systems have been used in many situations, and by personalizing dialogue on the basis of the user&#8217;s tempo, it is thought that dialogue will be able to make it easier to speak and make people want to speak. Previous research has focused on methods for encouraging users to change their tempo to be in tune with the tempo of their dialogue partner. However, a conversation that differs from the user&#8217;s tempo can be stressful and burdensome for the user in the process of tuning in.Therefore, we define personal tempo as speech speed, which is the number of moras divided by the duration of speech and propose a speech-speed control method for spoken-dialogue systems. We implemented our method in a spoken-dialogue system that synchronizes speech with the user. We verified the effectiveness of the proposed method by analyzing its impact on the comprehension of speech and user impressions of the spoken-dialogue system. The results indicate that significant differences were obtained with the proposed method between impression and comprehension of the speech content.<\/p>\n\t\t\t\t<\/div>\n\n\t\t\t\t\n\t\t\t<\/div>\n\n\t\t\t\t\t\n\t\t\t\n\t\t\t<div id=\"su-post-2431\" class=\"su-post \">\n\n\t\t\t\t\n\t\t\t\t<h2 class=\"su-post-title\"><a href=\"https:\/\/sr.sfc.keio.ac.jp\/www\/en\/publication\/2431\/\">(\u65e5\u672c\u8a9e) \u5171\u901a\u57fa\u76e4\u69cb\u7bc9\u904e\u7a0b\u306b\u304a\u3051\u308b\u30e2\u30c0\u30ea\u30c6\u30a3\u306e\u5f71\u97ff\u5206\u6790\u3092\u76ee\u7684\u3068\u3057\u305f\u5354\u50cd\u30bf\u30b9\u30af\u8ab2\u984c\u306e\u63d0\u6848,\u7dd1\u5ddd\u8a60\u4ecb\u30fb\u5b87\u6cbb\u5ddd\u9065\u7950\u30fb\u9ad8\u6c50\u4e00\u7d00\uff08\u6176\u5927\uff09\u30fb\u6771\u4e2d\u7adc\u4e00\u90ce\uff08NTT\uff09,\u96fb\u5b50\u60c5\u5831\u901a\u4fe1\u5b66\u4f1a\u6280\u8853\u7814\u7a76\u5831\u544a,vol. 123, no. 394,pp. 47 &#8211; 52,2024\u5e742\u6708<\/a><\/h2>\n\n\t\t\t\t<div class=\"su-post-meta\">\n\t\t\t\t\tPosted: 2024\u5e747\u67084\u65e5\t\t\t\t<\/div>\n\n\t\t\t\t<div class=\"su-post-excerpt\">\n\t\t\t\t\t<p>Sorry, this entry is only available in \u65e5\u672c\u8a9e.<\/p>\n\t\t\t\t<\/div>\n\n\t\t\t\t\n\t\t\t<\/div>\n\n\t\t\t\t\t\n\t\t\t\n\t\t\t<div id=\"su-post-2443\" class=\"su-post \">\n\n\t\t\t\t\n\t\t\t\t<h2 class=\"su-post-title\"><a href=\"https:\/\/sr.sfc.keio.ac.jp\/www\/en\/publication\/2443\/\">\u690d\u7269\u00d7AR\u30a8\u30fc\u30b8\u30a7\u30f3\u30c8\u306b\u3088\u308b\u4e00\u4eba\u66ae\u3089\u3057\u306e\u4e2d\u3067\u306e\u767a\u8a71\u4fc3\u9032,\u6238\u6ca2\u5b9f\u30fb\u9ad8\u6c50\u4e00\u7d00\uff08\u6176\u5927\uff09,\u96fb\u5b50\u60c5\u5831\u901a\u4fe1\u5b66\u4f1a\u6280\u8853\u7814\u7a76\u5831\u544a,vol. 124, no. 143,2024\u5e748\u6708<\/a><\/h2>\n\n\t\t\t\t<div class=\"su-post-meta\">\n\t\t\t\t\tPosted: 2024\u5e7410\u670816\u65e5\t\t\t\t<\/div>\n\n\t\t\t\t<div class=\"su-post-excerpt\">\n\t\t\t\t\t<p>Since the pandemic, the decrease in face-to-face communication and the increase in feelings of loneliness have become significant issues. Houseplants, which are often displayed as \u201cgreen amenities\u201d in rooms, are the most familiar plants that humans can easily form attachments . Research that considers houseplants as conversation partners aims to draw out positive emotions and improve mental health while maintaining privacy. By utilizing Mixed Reality technology, this approach promotes self-care through interaction with plants. It contributes to the reduction of loneliness by building trust and controlling negative emotions. It is expected that the agents, by encouraging anthropomorphism and self-dialogue, will have a positive impact on daily life.<\/p>\n\t\t\t\t<\/div>\n\n\t\t\t\t\n\t\t\t<\/div>\n\n\t\t\t\t\t\n\t\t\t\n\t\t\t<div id=\"su-post-2441\" class=\"su-post \">\n\n\t\t\t\t\n\t\t\t\t<h2 class=\"su-post-title\"><a href=\"https:\/\/sr.sfc.keio.ac.jp\/www\/en\/publication\/2441\/\">\u8a00\u8a9e\u30e2\u30c7\u30eb\u3092\u7528\u3044\u305f\u767a\u8a71\u5185\u5bb9\u306b\u57fa\u3065\u304fFACS\u751f\u6210\u30e2\u30c7\u30eb\u306e\u63d0\u6848,\u5c0f\u6a4b\u9f8d\u4eba\u30fb\u5b87\u6cbb\u5ddd\u9065\u7950\u30fb\u9ad8\u6c50\u4e00\u7d00\uff08\u6176\u5927\uff09,\u96fb\u5b50\u60c5\u5831\u901a\u4fe1\u5b66\u4f1a\u6280\u8853\u7814\u7a76\u5831\u544a,vol. 124, no. 143,2024\u5e748\u6708<\/a><\/h2>\n\n\t\t\t\t<div class=\"su-post-meta\">\n\t\t\t\t\tPosted: 2024\u5e7410\u670816\u65e5\t\t\t\t<\/div>\n\n\t\t\t\t<div class=\"su-post-excerpt\">\n\t\t\t\t\t<p>This study proposes a model for generating facial expressions from speech text. While previous research has focused on generating facial animation from audio, this study concentrates on directly generating expressions from text. The output utilizes Action Units (AUs) based on the Facial Action Coding System (FACS). To reduce computational complexity and enhance model scalability, the proposed architecture employs only the encoder component of the Transformer, omitting the decoder. The model is trained using a sliding window approach, enabling generation of expressions for each token in temporal sequence. The dataset for training was constructed by collecting publicly available videos from the web, performing facial expression detection, and transcribing the speech content.<\/p>\n\t\t\t\t<\/div>\n\n\t\t\t\t\n\t\t\t<\/div>\n\n\t\t\t\t\t\n\t\t\t\n\t\t\t<div id=\"su-post-2435\" class=\"su-post \">\n\n\t\t\t\t\n\t\t\t\t<h2 class=\"su-post-title\"><a href=\"https:\/\/sr.sfc.keio.ac.jp\/www\/en\/project\/2435\/\">(\u65e5\u672c\u8a9e) \u30b9\u30c8\u30ec\u30b9\u7de9\u548c\u3092\u76ee\u7684\u3068\u3057\u305f\u63e1\u308b\u30ed\u30dc\u30c3\u30c8\u306e\u691c\u8a0e\u3068\u30d7\u30ed\u30c8\u30bf\u30a4\u30d4\u30f3\u30b0,\u5869\u8c37\u660e\u65e5\u9999\u30fb\u79c0\u5cf6\u88d5\u6a39\u30fb\u9ad8\u6c50\u4e00\u7d00\uff08\u6176\u5927\uff09,\u96fb\u5b50\u60c5\u5831\u901a\u4fe1\u5b66\u4f1a\u6280\u8853\u7814\u7a76\u5831\u544a,vol. 123, no. 394,pp. 35 &#8211; 40,2024\u5e742\u6708<\/a><\/h2>\n\n\t\t\t\t<div class=\"su-post-meta\">\n\t\t\t\t\tPosted: 2024\u5e747\u67084\u65e5\t\t\t\t<\/div>\n\n\t\t\t\t<div class=\"su-post-excerpt\">\n\t\t\t\t\t<p>Sorry, this entry is only available in \u65e5\u672c\u8a9e.<\/p>\n\t\t\t\t<\/div>\n\n\t\t\t\t\n\t\t\t<\/div>\n\n\t\t\t\t\t\n\t\t\t\n\t\t\t<div id=\"su-post-2433\" class=\"su-post \">\n\n\t\t\t\t\n\t\t\t\t<h2 class=\"su-post-title\"><a href=\"https:\/\/sr.sfc.keio.ac.jp\/www\/en\/project\/2433\/\">(\u65e5\u672c\u8a9e) \u75db\u307f\u306b\u95a2\u3059\u308b\u81ea\u5df1\u958b\u793a\u3092\u4fc3\u3059\u5bfe\u8a71\u30ed\u30dc\u30c3\u30c8\u306e\u691c\u8a0e,\u5cf6\u7530\u611b\u91cc\u30fb\u5b87\u6cbb\u5ddd\u9065\u7950\u30fb\u9ad8\u6c50\u4e00\u7d00\uff08\u6176\u5927\uff09,\u96fb\u5b50\u60c5\u5831\u901a\u4fe1\u5b66\u4f1a\u6280\u8853\u7814\u7a76\u5831\u544a,vol. 123, no. 394,pp. 41 &#8211; 46,2024\u5e742\u6708<\/a><\/h2>\n\n\t\t\t\t<div class=\"su-post-meta\">\n\t\t\t\t\tPosted: 2024\u5e747\u67084\u65e5\t\t\t\t<\/div>\n\n\t\t\t\t<div class=\"su-post-excerpt\">\n\t\t\t\t\t<p>Sorry, this entry is only available in \u65e5\u672c\u8a9e.<\/p>\n\t\t\t\t<\/div>\n\n\t\t\t\t\n\t\t\t<\/div>\n\n\t\t\t\t\t\n\t\t\t\n\t\t\t<div id=\"su-post-2428\" class=\"su-post \">\n\n\t\t\t\t\n\t\t\t\t<h2 class=\"su-post-title\"><a href=\"https:\/\/sr.sfc.keio.ac.jp\/www\/en\/project\/2428\/\">(\u65e5\u672c\u8a9e) \u30d1\u30fc\u30bd\u30ca\u30eb\u30c6\u30f3\u30dd\u306b\u540c\u8abf\u3059\u308b\u767a\u8a71\u30b7\u30b9\u30c6\u30e0\u306e\u5b9f\u88c5\u3068\u8a55\u4fa1,\u5b87\u6cbb\u5ddd\u9065\u7950\u30fb\u9ad8\u6c50\u4e00\u7d00\uff08\u6176\u5927\uff09,\u96fb\u5b50\u60c5\u5831\u901a\u4fe1\u5b66\u4f1a\u6280\u8853\u7814\u7a76\u5831\u544a,vol. 123, no. 347CNR2023-30,pp. 25 &#8211; 30,2023\u5e741\u6708<\/a><\/h2>\n\n\t\t\t\t<div class=\"su-post-meta\">\n\t\t\t\t\tPosted: 2024\u5e747\u67084\u65e5\t\t\t\t<\/div>\n\n\t\t\t\t<div class=\"su-post-excerpt\">\n\t\t\t\t\t<p>Sorry, this entry is only available in \u65e5\u672c\u8a9e.<\/p>\n\t\t\t\t<\/div>\n\n\t\t\t\t\n\t\t\t<\/div>\n\n\t\t\n\t\n<\/div>\n\n<h2 class = \"adjust-margin-1\">2023\u301c<\/h2>\n<hr \/>\n\n\n<div class=\"su-posts su-posts-default-loop \">\n\n\t\n\t\t\t\t\t\n\t\t\t\n\t\t\t<div id=\"su-post-2447\" class=\"su-post \">\n\n\t\t\t\t\n\t\t\t\t<h2 class=\"su-post-title\"><a href=\"https:\/\/sr.sfc.keio.ac.jp\/www\/en\/publication\/2447\/\">Face Robot Performing Interaction with Emphasis on Eye Blink Entrainment,Iimori, Masato\u30fbFuruya, Yuki\u30fbTakashio, Kazunori\uff08Keio Univ\uff09,2023 32nd IEEE International Conference on Robot and Human Interactive Communication (RO-MAN),2023\u5e748\u6708<\/a><\/h2>\n\n\t\t\t\t<div class=\"su-post-meta\">\n\t\t\t\t\tPosted: 2024\u5e7410\u670816\u65e5\t\t\t\t<\/div>\n\n\t\t\t\t<div class=\"su-post-excerpt\">\n\t\t\t\t\t<p>Eyes play a significant role in human-human interaction, and blinking is particularly important as it can indicate a pause in the conversation and even lead to eye blink entrainment. However, most communication robots cannot reproduce eye blink movements due to cost constraints. Thus, our aim is to create a low-cost robot that can physically reproduce eye blink movements and induce eye blink entrainment. In this paper, we describe the implementation of the robot and evaluate the subjective impression of the robot&#8217;s eye blink movements. Our results suggest that the robot&#8217;s blinking behavior at pauses in the conversation facilitated the participants&#8217; understanding of the robot&#8217;s speech. Our findings also suggest that simulating eye blink entrainment movement can increase the participant&#8217;s affinity and acceptance towards the robot in certain cases, and if the blinking is not well designed, affinity may be adversely affected<\/p>\n\t\t\t\t<\/div>\n\n\t\t\t\t\n\t\t\t<\/div>\n\n\t\t\t\t\t\n\t\t\t\n\t\t\t<div id=\"su-post-2420\" class=\"su-post \">\n\n\t\t\t\t\n\t\t\t\t<h2 class=\"su-post-title\"><a href=\"https:\/\/sr.sfc.keio.ac.jp\/www\/en\/project\/2420\/\">Design and Implementation of a Dialogue Breakdown Avoidance Method Using Excuses in a Chat Dialogue System<\/a><\/h2>\n\n\t\t\t\t<div class=\"su-post-meta\">\n\t\t\t\t\tPosted: 2024\u5e747\u67084\u65e5\t\t\t\t<\/div>\n\n\t\t\t\t<div class=\"su-post-excerpt\">\n\t\t\t\t\t<p>\u5bfe\u8a71\u30b7\u30b9\u30c6\u30e0\u3068\u306e\u81ea\u7136\u306a\u5bfe\u8a71\u3092\u5b9f\u73fe\u3059\u308b\u306b\u306f,\u5bfe\u8a71\u7834\u7dbb\u3092\u89e3\u6d88\u3059\u308b\u5fc5\u8981\u304c\u3042\u308b.\u3057\u304b\u3057,\u5bfe\u8a71\u7834\u7dbb\u3092\u691c\u51fa\u3059\u308b\u7814\u7a76\u306f\u591a\u304f\u884c\u308f\u308c\u3066\u304a\u308a,\u624b\u6cd5\u3082\u78ba\u7acb\u3055\u308c\u3066\u3044\u308b\u4e00\u65b9\u3067,\u5bfe\u8a71\u7834\u7dbb\u691c\u51fa\u3092\u884c\u3063\u305f\u5f8c,\u3069\u306e\u3088\u3046\u306b\u5bfe\u8a71\u7834\u7dbb\u3092\u56de\u907f\u3059\u308b\u306e\u304b\u3068\u3044\u3046\u5bfe\u8a71\u7834\u7dbb\u56de\u907f\u3092\u30c6\u30fc\u30de\u3068\u3057\u3066\u6271\u3063\u305f\u7814\u7a76\u306f,\u76f8\u5bfe\u7684\u306b\u898b\u3066\u6570\u304c\u5c11\u306a\u304f,\u660e\u78ba\u306a\u624b\u6cd5\u3082\u78ba\u7acb\u3055\u308c\u3066\u3044\u306a\u3044.\u305d\u3053\u3067,\u672c\u7814\u7a76\u3067\u306f\u3053\u306e\u5bfe\u8a71\u7834\u7dbb\u3092,\u30b7\u30b9\u30c6\u30e0\u304c\u8a00\u3044\u8a33\u3092\u8fd4\u3059\u3053\u3068\u3067\u56de\u907f\u3059\u308b\u624b\u6cd5\u3092\u63d0\u6848\u3059\u308b.\u8a00\u3044\u8a33\u306f\u4e3b\u306b\u81ea\u5206\u306e\u5931\u6557\u30fb\u904e\u5931\u306b\u5bfe\u3059\u308b\u5f01\u660e\u30fb\u8b1d\u7f6a\u3068\u5b9a\u7fa9\u3055\u308c\u3066\u3044\u308b.\u3053\u308c\u3092\u5bfe\u8a71\u30b7\u30b9\u30c6\u30e0\u306b\u7f6e\u304d\u63db\u3048\u308b\u3068,\u5bfe\u8a71\u7834\u7dbb\u3092\u30b7\u30b9\u30c6\u30e0\u8996\u70b9\u3067\u306e\u5931\u6557\u30fb\u904e\u5931\u3068\u3057\u3066\u89e3\u91c8\u3067\u304d\u308b.\u4ee5\u4e0a\u306e\u7d4c\u7def\u304b\u3089,\u30e6\u30fc\u30b6\u30fc\u306b\u597d\u5370\u8c61\u3092\u4e0e\u3048\u308b\u8a00\u3044\u8a33\u5fdc\u7b54\u3092\u8fd4\u3059\u3053\u3068\u3067\u5bfe\u8a71\u7834\u7dbb\u56de\u907f\u3092\u884c\u3046\u30e2\u30b8\u30e5\u30fc\u30eb\u3092\u4f5c\u6210\u3057,\u5bfe\u8a71\u30b7\u30b9\u30c6\u30e0\u3078\u306e\u5b9f\u88c5\u3092\u884c\u3063\u305f.\u307e\u305f,\u30b7\u30b9\u30c6\u30e0\u306e\u5b9f\u904b\u7528\u3092\u60f3\u5b9a\u3068\u3057\u305f\u4e3b\u89b3\u8996\u70b9\u306e\u5370\u8c61\u8a55\u4fa1\u5b9f\u9a13\u3068,\u5bfe\u8a71\u5185\u5bb9\u306e\u5b9a\u6027\u7684\u306a\u8a55\u4fa1\u3092\u884c\u3046\u5ba2\u89b3\u8996\u70b9\u306e\u5370\u8c61\u8a55\u4fa1\u5b9f\u9a13\u306e2\u7a2e\u985e\u306e\u5b9f\u9a13\u3092\u5b9f\u65bd\u3057,\u63d0\u6848\u624b\u6cd5\u306e\u6709\u52b9\u6027\u3092\u793a\u3057\u305f.<\/p>\n\t\t\t\t<\/div>\n\n\t\t\t\t\n\t\t\t<\/div>\n\n\t\t\t\t\t\n\t\t\t\n\t\t\t<div id=\"su-post-2417\" class=\"su-post \">\n\n\t\t\t\t\n\t\t\t\t<h2 class=\"su-post-title\"><a href=\"https:\/\/sr.sfc.keio.ac.jp\/www\/en\/project\/2417\/\">(\u65e5\u672c\u8a9e) \u30ed\u30dc\u30c6\u30a3\u30af\u30b9\u3068 XR \u306b\u3088\u308b New Experience ,\u9ad8\u6c50 \u4e00\u7d00,NATURE INTERFACE Aug. 2023 no.88 pp.16-19<\/a><\/h2>\n\n\t\t\t\t<div class=\"su-post-meta\">\n\t\t\t\t\tPosted: 2024\u5e747\u67084\u65e5\t\t\t\t<\/div>\n\n\t\t\t\t<div class=\"su-post-excerpt\">\n\t\t\t\t\t<p>Sorry, this entry is only available in \u65e5\u672c\u8a9e.<\/p>\n\t\t\t\t<\/div>\n\n\t\t\t\t\n\t\t\t<\/div>\n\n\t\t\n\t\n<\/div>\n\n<h2 class = \"adjust-margin-1\">2022\u301c<\/h2>\n<hr \/>\n\n\n<div class=\"su-posts su-posts-default-loop \">\n\n\t\n\t\t\t\t\t\n\t\t\t\n\t\t\t<div id=\"su-post-2424\" class=\"su-post \">\n\n\t\t\t\t\n\t\t\t\t<h2 class=\"su-post-title\"><a href=\"https:\/\/sr.sfc.keio.ac.jp\/www\/en\/project\/2424\/\">A Preliminary Study for the Ethereum Blockchain-Based Smart Home Systems,C. Yiyang \u30fbK. Takashio (Keio Univ.), 2022 13th International Conference on Information and Communication Technology Convergence (ICTC),2022, pp. 71-76<\/a><\/h2>\n\n\t\t\t\t<div class=\"su-post-meta\">\n\t\t\t\t\tPosted: 2024\u5e747\u67084\u65e5\t\t\t\t<\/div>\n\n\t\t\t\t<div class=\"su-post-excerpt\">\n\t\t\t\t\t<p>The application of blockchain technology means that smart home devices will be greatly improved in terms of versatility and security, allowing people to use them more safely and conveniently in their daily home life. For smart home practitioners, blockchain technology means the arrival of the next \u201cwindow\u201d, indicating that the industry will develop from the budding stage to the popular stage. This paper proposes a smart home system based on the Ethereum blockchain as a platform. Through the use of smart contract technology, it not only builds a closed-loop management platform from energy suppliers to end users but also realizes cross-ecological communication. At the same time, this paper also examined the latency problem of \u201csolidity\u201d in the multi-contract call environment. By comparing the local pace test and the blockchain-side pace test, we found that the function invoked pace inconsistent between local-side and chain-side existed but the delay can be controlled within one second.<\/p>\n\t\t\t\t<\/div>\n\n\t\t\t\t\n\t\t\t<\/div>\n\n\t\t\t\t\t\n\t\t\t\n\t\t\t<div id=\"su-post-2062\" class=\"su-post \">\n\n\t\t\t\t\n\t\t\t\t<h2 class=\"su-post-title\"><a href=\"https:\/\/sr.sfc.keio.ac.jp\/www\/en\/project\/2062\/\">Analysis of the Effect of Modality and Social Relationships on Building Common Ground ,  Yuki Furuya, Koki Saito,Kosuke Ogura ( Keio Univ.),Koh Mitsuda,Ryuichiro Higashinaka (NTT),Kazunori Takashio ( Keio Univ.)<\/a><\/h2>\n\n\t\t\t\t<div class=\"su-post-meta\">\n\t\t\t\t\tPosted: 2022\u5e7410\u670820\u65e5\t\t\t\t<\/div>\n\n\t\t\t\t<div class=\"su-post-excerpt\">\n\t\t\t\t\t<p>Building common ground with users is essential for dialogue agent systems and robots to interact naturally with people. While a few previous studies have investigated the process of building common ground in human-human dialogue, most of them have been conducted on the basis of text chat. In this study, we constructed a dialogue corpus to investigate the process of building common ground with a particular focus on the modality of dialogue and the social relationship between the participants in the process of building common ground, which are important but have not been investigated in the previous work. The results of our analysis suggest that adding the modality or developing the relationship between workers speeds up the building of common ground. Specifically, regarding the modality, the presence of video rather than only audio may unconsciously facilitate work, and as for the relationship, it is easier to convey information about emotions and turn-taking among friends than in first meetings. These findings and the corpus should prove useful for developing a system to support remote communication.<\/p>\n\t\t\t\t<\/div>\n\n\t\t\t\t\n\t\t\t<\/div>\n\n\t\t\t\t\t\n\t\t\t\n\t\t\t<div id=\"su-post-2058\" class=\"su-post \">\n\n\t\t\t\t\n\t\t\t\t<h2 class=\"su-post-title\"><a href=\"https:\/\/sr.sfc.keio.ac.jp\/www\/en\/project\/2058\/\">Dialogue Corpus Construction Considering Modality and Social Relationships in Building Common Ground. Yuki Furuya, Koki Saito, Kosuke Ogura, Koh Mitsuda, Ryuichiro Higashinaka, and Kazunori Takashio. 2022.In Proceedings of the Thirteenth Language Resources and Evaluation Conference, pages 4088\u20134095, Marseille, France. European Language Resources Association.<\/a><\/h2>\n\n\t\t\t\t<div class=\"su-post-meta\">\n\t\t\t\t\tPosted: 2022\u5e7410\u670820\u65e5\t\t\t\t<\/div>\n\n\t\t\t\t<div class=\"su-post-excerpt\">\n\t\t\t\t\t<p>Building common ground with users is essential for dialogue agent systems and robots to interact naturally with people. While a few previous studies have investigated the process of building common ground in human-human dialogue, most of them have been conducted on the basis of text chat. In this study, we constructed a dialogue corpus to investigate the process of building common ground with a particular focus on the modality of dialogue and the social relationship between the participants in the process of building common ground, which are important but have not been investigated in the previous work. The results of our analysis suggest that adding the modality or developing the relationship between workers speeds up the building of common ground. Specifically, regarding the modality, the presence of video rather than only audio may unconsciously facilitate work, and as for the relationship, it is easier to convey information about emotions and turn-taking among friends than in first meetings. These findings and the corpus should prove useful for developing a system to support remote communication.<\/p>\n\t\t\t\t<\/div>\n\n\t\t\t\t\n\t\t\t<\/div>\n\n\t\t\t\t\t\n\t\t\t\n\t\t\t<div id=\"su-post-2042\" class=\"su-post \">\n\n\t\t\t\t\n\t\t\t\t<h2 class=\"su-post-title\"><a href=\"https:\/\/sr.sfc.keio.ac.jp\/www\/en\/publication\/2042\/\">An implementation of creating human presence of AR navigation system, Ryoma Ariyoshi, Kentaro Taninaka, Kazunori Takashio, IEICE Tech. Rep., vol. 122, no. 193, CNR2022-12, pp. 23-28, September 2022.<\/a><\/h2>\n\n\t\t\t\t<div class=\"su-post-meta\">\n\t\t\t\t\tPosted: 2022\u5e7410\u670820\u65e5\t\t\t\t<\/div>\n\n\t\t\t\t<div class=\"su-post-excerpt\">\n\t\t\t\t\t<p>In the field of AR, smartphones are often used for navigation, but since the hands are occupied and the immersive experience is compromised, it is preferable to use a stand-alone HMD for AR. However, AR navigation using HMDs is easily affected by the external environment, which inhibits immersion and realism. In addition, the only way to know the traces of a person who was there in the past is to view two-dimensional content, which does not provide a sense of presence or realism. In this research, we aim to implement an AR navigation system that enables interaction with avatars by eliminating factors that impair the sense of presence and realism. As a preliminary step, we investigated how to display avatars with a sense of presence, using HoloLens 2, and experimented with three levels of abstraction, each with motion, superimposed on the Keio University Shonan Fujisawa Campus. The results suggest that the avatar with the lowest level of abstraction and the appearance closest to that of a human being has a greater presence on certain questions. We also observed the &#8220;uncanny valley phenomenon,&#8221; in which people feel fear and discomfort toward the Avatar when its appearance is closer to reality.<\/p>\n\t\t\t\t<\/div>\n\n\t\t\t\t\n\t\t\t<\/div>\n\n\t\t\t\t\t\n\t\t\t\n\t\t\t<div id=\"su-post-2045\" class=\"su-post \">\n\n\t\t\t\t\n\t\t\t\t<h2 class=\"su-post-title\"><a href=\"https:\/\/sr.sfc.keio.ac.jp\/www\/en\/publication\/2045\/\">(\u65e5\u672c\u8a9e) \u77ac\u304d\u306e\u5f15\u304d\u8fbc\u307f\u3092\u5229\u7528\u3057\u305f\u30a4\u30f3\u30bf\u30e9\u30af\u30c6\u30a3\u30d6\u30ed\u30dc\u30c3\u30c8\u306e\u5b9f\u88c5\u3068\u8a55\u4fa1\uff0c\u98ef\u68ee\u512a\u6597\uff0c\u53e4\u8c37\u512a\u6a39\uff0c\u9ad8\u6c50\u4e00\u7d00\uff0c\u4fe1\u5b66\u6280\u5831, vol. 122, no. 193, CNR2022-10, pp. 13-17, 2022\u5e749\u6708.<\/a><\/h2>\n\n\t\t\t\t<div class=\"su-post-meta\">\n\t\t\t\t\tPosted: 2022\u5e7410\u670820\u65e5\t\t\t\t<\/div>\n\n\t\t\t\t<div class=\"su-post-excerpt\">\n\t\t\t\t\t<p>Sorry, this entry is only available in \u65e5\u672c\u8a9e.<\/p>\n\t\t\t\t<\/div>\n\n\t\t\t\t\n\t\t\t<\/div>\n\n\t\t\n\t\n<\/div>\n\n<h2 class = \"adjust-margin-1\">2021\u301c<\/h2>\n<hr \/>\n\n\n<div class=\"su-posts su-posts-default-loop \">\n\n\t\n\t\t\t\t\t\n\t\t\t\t\t\t\t\t\t\t\n\t\t\t\n\t\t\t<div id=\"su-post-1954\" class=\"su-post \">\n\n\t\t\t\t\n\t\t\t\t<h2 class=\"su-post-title\"><a href=\"https:\/\/sr.sfc.keio.ac.jp\/www\/en\/publication\/1954\/\">An Analysis of Acceptance Factors of Partner Robot -What is &#8220;Animal-Like&#8221;- , Koki Saito, Furuya Yuki, Kazunori Takashio, IEICE Tech. Rep., vol. 121, no. 267, CNR2021-9, pp. 11-15, Nov. 2021.<\/a><\/h2>\n\n\t\t\t\t<div class=\"su-post-meta\">\n\t\t\t\t\tPosted: 2022\u5e743\u67086\u65e5\t\t\t\t<\/div>\n\n\t\t\t\t<div class=\"su-post-excerpt\">\n\t\t\t\t\t<p>In this study, we analyzed what kind of animal-like qualities are important in the acceptance of partner robots in households. The questionnaire surveyed the animal-like qualities of visual, auditory and tactile elements, and analyzed the relation between the partial animal-like qualities of each element and the duration of use. We also showed what aspects of the robot&#8217;s animal-like nature the user perceived by analyzing the relationship between the animal-like nature of each element of the robot and the sense of living things. The results suggest that items such as the intelligence of the robot and the animal-like shape of the robot might be factors in acceptance. It was also suggested that the smoothness of the robot&#8217;s movement and the animal-like sound of the volume other than the robot&#8217;s voice were factors that made users feel a sense of living-likeness toward the robot.<\/p>\n\t\t\t\t<\/div>\n\n\t\t\t\t\n\t\t\t<\/div>\n\n\t\t\t\t\t\n\t\t\t\n\t\t\t<div id=\"su-post-1952\" class=\"su-post \">\n\n\t\t\t\t\n\t\t\t\t<h2 class=\"su-post-title\"><a href=\"https:\/\/sr.sfc.keio.ac.jp\/www\/en\/publication\/1952\/\">A Feasibility Study of An Intelligent Environmental Monitoring System Based On The Ethereum Blockchains, Cheng Yiyang, Kazunori Takashio,  2021 International Conference on Information and Communication Technology Convergence (ICTC 2021), pp. 435-439, doi: 10.1109\/ICTC52510.2021.9621007.<\/a><\/h2>\n\n\t\t\t\t<div class=\"su-post-meta\">\n\t\t\t\t\tPosted: 2022\u5e743\u67086\u65e5\t\t\t\t<\/div>\n\n\t\t\t\t<div class=\"su-post-excerpt\">\n\t\t\t\t\t<p>In today&#8217;s era, the security and authenticity of information have always received extensive attention, especially when people have doubts about the authenticity of data. Hence, a traceable data storage platform with the feature that theoretically cannot be tampered with becomes extremely important. Under such demands, the deployment of smart contracts can effectively trace the source and discover problems. This article introduces an environmental data detection platform based on blockchain technology and smart contracts. It aims to record the data continuously and verify the authority of each transaction.<\/p>\n\t\t\t\t<\/div>\n\n\t\t\t\t\n\t\t\t<\/div>\n\n\t\t\t\t\t\n\t\t\t\n\t\t\t<div id=\"su-post-1949\" class=\"su-post \">\n\n\t\t\t\t\n\t\t\t\t<h2 class=\"su-post-title\"><a href=\"https:\/\/sr.sfc.keio.ac.jp\/www\/en\/publication\/1949\/\">(\u65e5\u672c\u8a9e) \u30c6\u30ec\u30d7\u30ec\u30bc\u30f3\u30b9\u30ed\u30dc\u30c3\u30c8\u306b\u304a\u3051\u308b\u9060\u9694\u30e6\u30fc\u30b6\u306e\u8eab\u4f53\u6027\u3092\u8003\u616e\u3057\u305f\u8868\u793a\u624b\u6cd5\u306e\u691c\u8a0e, \u53e4\u8c37\u512a\u6a39, \u9ad8\u6c50\u4e00\u7d00, \u4fe1\u5b66\u6280\u5831, vol. 121, no. 93, CNR2021-3, pp. 8-13, 2021\u5e747\u6708.<\/a><\/h2>\n\n\t\t\t\t<div class=\"su-post-meta\">\n\t\t\t\t\tPosted: 2022\u5e743\u67086\u65e5\t\t\t\t<\/div>\n\n\t\t\t\t<div class=\"su-post-excerpt\">\n\t\t\t\t\t<p>Sorry, this entry is only available in Japanese.<\/p>\n\t\t\t\t<\/div>\n\n\t\t\t\t\n\t\t\t<\/div>\n\n\t\t\t\t\t\n\t\t\t\n\t\t\t<div id=\"su-post-1667\" class=\"su-post \">\n\n\t\t\t\t\n\t\t\t\t<h2 class=\"su-post-title\"><a href=\"https:\/\/sr.sfc.keio.ac.jp\/www\/en\/publication\/1667\/\">(\u65e5\u672c\u8a9e) HRI\u306b\u304a\u3051\u308b\u8996\u7dda\u884c\u52d5\u304c\u53c2\u4e0e\u8005\u9593\u306e\u5171\u901a\u57fa\u76e4\u69cb\u7bc9\u306b\u53ca\u307c\u3059\u5f71\u97ff, \u6851\u539f\u591a\u745b, \u5c71\u53e3\u7559\u5b9f, \u6e21\u8fba\u5de7\u767b, \u53e4\u8c37\u512a\u6a39, \u9ad8\u6c50\u4e00\u7d00, \u4fe1\u5b66\u6280\u5831, vol. 120, no. 394, CNR2020-17, pp. 22-27, 2021\u5e743\u6708.<\/a><\/h2>\n\n\t\t\t\t<div class=\"su-post-meta\">\n\t\t\t\t\tPosted: 2021\u5e745\u670829\u65e5\t\t\t\t<\/div>\n\n\t\t\t\t<div class=\"su-post-excerpt\">\n\t\t\t\t\t<p>Sorry, this entry is only available in \u65e5\u672c\u8a9e.<\/p>\n\t\t\t\t<\/div>\n\n\t\t\t\t\n\t\t\t<\/div>\n\n\t\t\n\t\n<\/div>\n\n<h2 class = \"adjust-margin-1\">2020\u301c<\/h2>\n<hr \/>\n\n\n<div class=\"su-posts su-posts-default-loop \">\n\n\t\n\t\t\t\t\t\n\t\t\t\n\t\t\t<div id=\"su-post-1583\" class=\"su-post \">\n\n\t\t\t\t\n\t\t\t\t<h2 class=\"su-post-title\"><a href=\"https:\/\/sr.sfc.keio.ac.jp\/www\/en\/publication\/1583\/\">v-IoT : Construction of Virtual IoT Infrastructure by AR and Selection Method of Appropriate Information-giving Objects Using Associative Concept, Kentaro Taninaka, Kazunori Takashio, The IEICE Transactions on Information and Systems<\/a><\/h2>\n\n\t\t\t\t<div class=\"su-post-meta\">\n\t\t\t\t\tPosted: 2020\u5e7410\u67082\u65e5\t\t\t\t<\/div>\n\n\t\t\t\t<div class=\"su-post-excerpt\">\n\t\t\t\t\t<p>The miniaturization and decreasing cost of networking computers, along with the advancement of cloud infrastructure has eased the implementation of it in commercial products. However, it is difficult to give such function to the simple products like daily necessities, thus having devices such as Smart Speakers providing wide variety of information. In this research, we propose v-IoT system, the approach which add information-giving functions to consumables such as plastic bottles using augmented reality technology. This system will determine the information and the action suggested the user to take based on the verb and function of the object. The closest associable pairs of suggested action and object will be determined with reference to the Associative Concept Dictionary, enabling v-IoT to offer a notification system that is apparent and intelligible. The result showed proposed system can select objects which is easy to associate action for users.<\/p>\n\t\t\t\t<\/div>\n\n\t\t\t\t\n\t\t\t<\/div>\n\n\t\t\t\t\t\n\t\t\t\n\t\t\t<div id=\"su-post-1633\" class=\"su-post \">\n\n\t\t\t\t\n\t\t\t\t<h2 class=\"su-post-title\"><a href=\"https:\/\/sr.sfc.keio.ac.jp\/www\/en\/publication\/1633\/\">(\u65e5\u672c\u8a9e) \u8907\u6570\u4eba\u5bfe\u8a71\u3067\u306e\u5171\u901a\u57fa\u76e4\u69cb\u7bc9\u306b\u304a\u3051\u308b\u975e\u8a00\u8a9e\u884c\u52d5\u56e0\u5b50\u304a\u3088\u3073\u793e\u4f1a\u7684\u884c\u52d5\u56e0\u5b50\u306e\u5206\u6790\uff0c\u5c71\u53e3\u7559\u5b9f\uff0c\u6e21\u8fba\u5de7\u767b\uff0c\u6851\u7530\u591a\u745b\uff0c\u53e4\u8c37\u512a\u6a39\uff0c\u7b2c37\u56de\u4eba\u9593\u60c5\u5831\u5b66\u4f1a\u8b1b\u6f14\u96c62020\u5e7412\u670810\u65e5<\/a><\/h2>\n\n\t\t\t\t<div class=\"su-post-meta\">\n\t\t\t\t\tPosted: 2021\u5e741\u670825\u65e5\t\t\t\t<\/div>\n\n\t\t\t\t<div class=\"su-post-excerpt\">\n\t\t\t\t\t<p>Sorry, this entry is only available in \u65e5\u672c\u8a9e.<\/p>\n\t\t\t\t<\/div>\n\n\t\t\t\t\n\t\t\t<\/div>\n\n\t\t\t\t\t\n\t\t\t\n\t\t\t<div id=\"su-post-1579\" class=\"su-post \">\n\n\t\t\t\t\n\t\t\t\t<h2 class=\"su-post-title\"><a href=\"https:\/\/sr.sfc.keio.ac.jp\/www\/en\/publication\/1579\/\">Telepresence Robot Blended with a Real Landscape and Its Impact on User Experiences, Yuki Furuya, Kazunori Takashio, 2020 29th IEEE International Conference on Robot and Human Interactive Communication (RO-MAN 2020), Aug to Sept, 2020<\/a><\/h2>\n\n\t\t\t\t<div class=\"su-post-meta\">\n\t\t\t\t\tPosted: 2020\u5e7410\u67082\u65e5\t\t\t\t<\/div>\n\n\t\t\t\t<div class=\"su-post-excerpt\">\n\t\t\t\t\t<p>Telepresence robots using traditional video chats do not yet fully represent the presence of a remote user. To enhance that presence, we propose a novel approach that replaces the background of the remote user with a real-time captured image of the area behind the robot. We assumed that the background of the remote user, which was visible on the telepresence robot&#8217;s display, was one reason for the reduction in presence because it might remind local viewers that the remote user is in another place. We found that our approach effectively enhanced the presence of remote user and reduced the negative impression of a delay. In this paper, we describe the new approach, its implementation, and an evaluation of the system. We evaluated the subjective impressions of communication through the proposed system. This approach also protects the privacy of the remote user and makes it more acceptable to people using telepresence robots from private places, e.g., a sickroom in a hospital.<\/p>\n\t\t\t\t<\/div>\n\n\t\t\t\t\n\t\t\t<\/div>\n\n\t\t\t\t\t\n\t\t\t\n\t\t\t<div id=\"su-post-1413\" class=\"su-post \">\n\n\t\t\t\t\n\t\t\t\t<h2 class=\"su-post-title\"><a href=\"https:\/\/sr.sfc.keio.ac.jp\/www\/en\/publication\/1413\/\">(\u65e5\u672c\u8a9e) [\u5b66\u751f\u8b1b\u6f14\u5968\u52b1\u8cde]\u8907\u6570\u4eba\u4f1a\u8a71\u306b\u304a\u3051\u308b\u753b\u9762\u5185\u30a8\u30fc\u30b8\u30a7\u30f3\u30c8\u3068\u306e\u30b3\u30df\u30e5\u30cb\u30b1\u30fc\u30b7\u30e7\u30f3\u30c1\u30e3\u30cd\u30eb\u78ba\u7acb\u624b\u6cd5\u306e\u63d0\u6848, \u6e21\u9089\u5de7\u767b, \u9ad8\u6c50\u4e00\u7d00, \u4fe1\u5b66\u6280\u5831, vol. 119, no. 446, CNR2019-53, pp. 47-51, 2020\u5e743\u6708.<\/a><\/h2>\n\n\t\t\t\t<div class=\"su-post-meta\">\n\t\t\t\t\tPosted: 2020\u5e744\u67083\u65e5\t\t\t\t<\/div>\n\n\t\t\t\t<div class=\"su-post-excerpt\">\n\t\t\t\t\t<p>Sorry, this entry is only available in \u65e5\u672c\u8a9e.<\/p>\n\t\t\t\t<\/div>\n\n\t\t\t\t\n\t\t\t<\/div>\n\n\t\t\t\t\t\n\t\t\t\n\t\t\t<div id=\"su-post-1415\" class=\"su-post \">\n\n\t\t\t\t\n\t\t\t\t<h2 class=\"su-post-title\"><a href=\"https:\/\/sr.sfc.keio.ac.jp\/www\/en\/publication\/1415\/\">(\u65e5\u672c\u8a9e) [\u5b66\u751f\u8b1b\u6f14\u5968\u52b1\u8cde]\u9060\u9694\u8005\u306e\u8eab\u4f53\u7684\u5b58\u5728\u611f\u3092\u9ad8\u3081\u308b\u30c6\u30ec\u30d7\u30ec\u30bc\u30f3\u30b9\u30ed\u30dc\u30c3\u30c8\u306e\u63d0\u6848, \u53e4\u8c37\u512a\u6a39, \u9ad8\u6c50\u4e00\u7d00, \u4fe1\u5b66\u6280\u5831, vol. 119, no. 446, CNR2019-54, pp. 53-57, 2020\u5e743\u6708.<\/a><\/h2>\n\n\t\t\t\t<div class=\"su-post-meta\">\n\t\t\t\t\tPosted: 2020\u5e744\u67083\u65e5\t\t\t\t<\/div>\n\n\t\t\t\t<div class=\"su-post-excerpt\">\n\t\t\t\t\t<p>Sorry, this entry is only available in \u65e5\u672c\u8a9e.<\/p>\n\t\t\t\t<\/div>\n\n\t\t\t\t\n\t\t\t<\/div>\n\n\t\t\t\t\t\n\t\t\t\n\t\t\t<div id=\"su-post-1411\" class=\"su-post \">\n\n\t\t\t\t\n\t\t\t\t<h2 class=\"su-post-title\"><a href=\"https:\/\/sr.sfc.keio.ac.jp\/www\/en\/publication\/1411\/\">(\u65e5\u672c\u8a9e) HRI\u306b\u304a\u3051\u308b\u5272\u8fbc\u307f\u6642\u306e\u4f1a\u8a71\u9806\u5408\u610f\u5f62\u6210\u30b3\u30fc\u30d1\u30b9\u306e\u69cb\u7bc9\u306b\u3080\u3051\u3066,\u5800\u6c5f\u62d3\u5b9f, \u5c71\u53e3\u7559\u5b9f, \u6851\u539f\u591a\u745b, \u6e21\u8fba\u5de7\u767b, \u9ad8\u6c50\u4e00\u7d00, \u4fe1\u5b66\u6280\u5831, vol. 119, no. 446, CNR2019-52, pp. 41-46, 2020\u5e743\u6708.<\/a><\/h2>\n\n\t\t\t\t<div class=\"su-post-meta\">\n\t\t\t\t\tPosted: 2020\u5e744\u67083\u65e5\t\t\t\t<\/div>\n\n\t\t\t\t<div class=\"su-post-excerpt\">\n\t\t\t\t\t<p>Sorry, this entry is only available in \u65e5\u672c\u8a9e.<\/p>\n\t\t\t\t<\/div>\n\n\t\t\t\t\n\t\t\t<\/div>\n\n\t\t\n\t\n<\/div>\n\n<h2 class = \"adjust-margin-1\">2019<\/h2>\n<hr \/>\n\n\n<div class=\"su-posts su-posts-default-loop \">\n\n\t\n\t\t\t\t\t\n\t\t\t\n\t\t\t<div id=\"su-post-1406\" class=\"su-post \">\n\n\t\t\t\t\n\t\t\t\t<h2 class=\"su-post-title\"><a href=\"https:\/\/sr.sfc.keio.ac.jp\/www\/en\/publication\/1406\/\">(\u65e5\u672c\u8a9e) [\u512a\u79c0\u767a\u8868\u8cde] \u9060\u9694\u8005\u306e\u591a\u4eba\u6570\u4f1a\u8a71\u3078\u306e\u53c2\u4e0e\u3092\u652f\u63f4\u3059\u308b\u30c6\u30ec\u30d7\u30ec\u30bc\u30f3\u30b9\u4ef2\u4ecb\u30ed\u30dc\u30c3\u30c8, \u5800\u6c5f\u62d3\u5b9f, \u5c71\u53e3\u7559\u5b9f, \u6851\u539f\u591a\u745b, \u53e4\u8c37\u512a\u6a39, \u4f0a\u8cc0\u7406\u5fc3, \u98ef\u68ee\u512a\u6597, \u52e2\u5b50\u96c4\u5927, \u9ad8\u6c50\u4e00\u7d00,  \u7b2c34\u56de \u4eba\u9593\u60c5\u5831\u5b66\u4f1a\u8b1b\u6f14\u96c6, pp. 9-10,2019\u5e7412\u6708<\/a><\/h2>\n\n\t\t\t\t<div class=\"su-post-meta\">\n\t\t\t\t\tPosted: 2020\u5e745\u67083\u65e5\t\t\t\t<\/div>\n\n\t\t\t\t<div class=\"su-post-excerpt\">\n\t\t\t\t\t<p>Sorry, this entry is only available in \u65e5\u672c\u8a9e.<\/p>\n\t\t\t\t<\/div>\n\n\t\t\t\t\n\t\t\t<\/div>\n\n\t\t\t\t\t\n\t\t\t\n\t\t\t<div id=\"su-post-1379\" class=\"su-post \">\n\n\t\t\t\t\n\t\t\t\t<h2 class=\"su-post-title\"><a href=\"https:\/\/sr.sfc.keio.ac.jp\/www\/en\/publication\/1379\/\">, Development and Evaluation of Telepresence Robot with Transparent Background, Yuki Furuya, Takumi Horie, Kazunori Takashio, IEICE Tech. Rep., vol. 119, no. 267, CNR2019-37, pp. 77-78, Nov. 2019.<\/a><\/h2>\n\n\t\t\t\t<div class=\"su-post-meta\">\n\t\t\t\t\tPosted: 2019\u5e7412\u670812\u65e5\t\t\t\t<\/div>\n\n\t\t\t\t<div class=\"su-post-excerpt\">\n\t\t\t\t\t<p>\u30c6\u30ec\u30d7\u30ec\u30bc\u30f3\u30b9\u30ed\u30dc\u30c3\u30c8\u306a\u3069\u306b\u642d\u8f09\u3055\u308c\u3066\u3044\u308b\u30d3\u30c7\u30aa\u30c1\u30e3\u30c3\u30c8\u3067\u306f\uff0c\u9060\u9694\u3067\u4f1a\u8a71\u306b\u53c2\u52a0\u3057\u3066\u3044\u308b\u76f8\u624b\u306e\u80cc\u666f\u304c\u898b\u3048\u3066\u3057\u307e\u3046\u3053\u3068\u306b\u3088\u3063\u3066\uff0c\u76f8\u624b\u304c\u5225\u306e\u7a7a\u9593\u306b\u3044\u308b\u3068\u3044\u3046\u3053\u3068\u304c\u610f\u8b58\u3055\u308c\u3066\u3057\u307e\u3046\uff0e\u305d\u3053\u3067\u6211\u3005\u306f\uff0c\u30d3\u30c7\u30aa\u30c1\u30e3\u30c3\u30c8\u306b\u304a\u3051\u308b\u76f8\u624b\u306e\u6620\u50cf\u306e\u80cc\u666f\u90e8\u5206\u306e\u307f\u3092\u30c6\u30ec\u30d7\u30ec\u30bc\u30f3\u30b9\u30ed\u30dc\u30c3\u30c8\u304c\u7f6e\u304b\u308c\u3066\u3044\u308b\u5834\u306e\u80cc\u666f\u6620\u50cf\u306b\u5165\u308c\u66ff\u3048\u308b\u3053\u3068\u3067\uff0c\u9060\u9694\u53c2\u52a0\u8005\u3068\u306e\u4e00\u4f53\u611f\u304c\u9ad8\u307e\u308a\u30b3\u30df\u30e5\u30cb\u30b1\u30fc\u30b7\u30e7\u30f3\u304c\u3088\u308a\u6d3b\u767a\u306b\u306a\u308b\uff0c\u3068\u3044\u3046\u4eee\u8aac\u3092\u7acb\u3066\u7814\u7a76\u3092\u884c\u306a\u3063\u3066\u3044\u308b\uff0e\u672c\u7814\u7a76\u3067\u306f\u30b7\u30b9\u30c6\u30e0\u306e\u30d7\u30ed\u30c8\u30bf\u30a4\u30d7\u306e\u5b9f\u88c5\u304a\u3088\u3073\u8a55\u4fa1\u5b9f\u9a13\u3092\u884c\u3044\uff0c\u80cc\u666f\u3092\u30ed\u30dc\u30c3\u30c8\u304c\u7f6e\u304b\u308c\u3066\u3044\u308b\u5834\u306e\u753b\u50cf\u306b\u7f6e\u304d\u63db\u3048\u308b\u3053\u3068\u3067\u4f1a\u8a71\u306e\u6d3b\u767a\u5ea6\u3084\u9060\u9694\u8005\u306e\u5b58\u5728\u611f\u306e\u5411\u4e0a\u306b\u5bc4\u4e0e\u3059\u308b\u3053\u3068\u304c\u660e\u3089\u304b\u306b\u306a\u3063\u305f\uff0e<\/p>\n\t\t\t\t<\/div>\n\n\t\t\t\t\n\t\t\t<\/div>\n\n\t\t\t\t\t\n\t\t\t\n\t\t\t<div id=\"su-post-1372\" class=\"su-post \">\n\n\t\t\t\t\n\t\t\t\t<h2 class=\"su-post-title\"><a href=\"https:\/\/sr.sfc.keio.ac.jp\/www\/en\/publication\/1372\/\">Attentive in-screen agent &#8212; In-screen agent that looks like a specific direction \u2014,Takuto Watanabe,IEICE Tech. Rep., vol. 119, no. 189, CNR2019-15, pp. 19-23, Aug. 2019.<\/a><\/h2>\n\n\t\t\t\t<div class=\"su-post-meta\">\n\t\t\t\t\tPosted: 2019\u5e7412\u670812\u65e5\t\t\t\t<\/div>\n\n\t\t\t\t<div class=\"su-post-excerpt\">\n\t\t\t\t\t<p>In recent years, 3DCG characters have become widespread and are being used in-screen agents as avatarsfor interactive agents. In this study, we propose a method to give a sense of coexistence and unity of space by lettingthe user follow the background and the method of projecting the gaze target to the pupil. For the experiment, anagent that can manually select the gaze target was created and evaluated.<\/p>\n\t\t\t\t<\/div>\n\n\t\t\t\t\n\t\t\t<\/div>\n\n\t\t\t\t\t\n\t\t\t\n\t\t\t<div id=\"su-post-1381\" class=\"su-post \">\n\n\t\t\t\t\n\t\t\t\t<h2 class=\"su-post-title\"><a href=\"https:\/\/sr.sfc.keio.ac.jp\/www\/en\/publication\/1381\/\">Video Gaming with Emotion-Expressive Virtual Rival Agent, Shinsuke Kiyomoto, Kazunori Takashio, IEICE Tech. Rep., vol. 119, no. 267, CNR2019-38, pp. 79-80, Nov. 2019.<\/a><\/h2>\n\n\t\t\t\t<div class=\"su-post-meta\">\n\t\t\t\t\tPosted: 2019\u5e7412\u670812\u65e5\t\t\t\t<\/div>\n\n\t\t\t\t<div class=\"su-post-excerpt\">\n\t\t\t\t\t<p>With the evolution of computer gaming hardware causing entertainment to change, on-screen PvP video games still remain to be the most popular option. Most of these games have a game mode versus CP(Computer Player). However the experience in such modes can lack satisfaction on the player side due to the lack of engagement, challenge, and involvement. This research will design and evaluate an auxiliary system that enhances the experience of playing against CP by using an on-screen virtual agent. Emotion will be synthesized through an emotion engine using the game state as an input, and the agent will display facial expressions and appropriate utterances.<\/p>\n\t\t\t\t<\/div>\n\n\t\t\t\t\n\t\t\t<\/div>\n\n\t\t\t\t\t\n\t\t\t\n\t\t\t<div id=\"su-post-1383\" class=\"su-post \">\n\n\t\t\t\t\n\t\t\t\t<h2 class=\"su-post-title\"><a href=\"https:\/\/sr.sfc.keio.ac.jp\/www\/en\/publication\/1383\/\">Virtual IoT: An IoT Platform with MR Technologies Realizing Low-cost and Flexible Notification of Life-support Information,Kentaro Taninaka, Kazunori Takashio, IoTaIS2019,Nov. 2019.<\/a><\/h2>\n\n\t\t\t\t<div class=\"su-post-meta\">\n\t\t\t\t\tPosted: 2019\u5e7412\u670812\u65e5\t\t\t\t<\/div>\n\n\t\t\t\t<div class=\"su-post-excerpt\">\n\t\t\t\t\t<p>Virtual IoT: An IoT Platform with MR Technologies Realizing Low-cost and Flexible Notification of Life-support Information Abstract : The miniaturization and decreasing cost of networking computers, along with the advancement of cloud infrastructure has eased the implementation of IoT products for manufacturers. The increase in such products has led to the rise in IoT devices with information provision function. However, it is difficult to make products display information in consideration of the design on the surface, like daily commodities, since it requires the device to have a wealth of computational resource. In this paper, we propose Virtual IoT System, the approach which adds information-giving functions to the non-computing objects using Mixed Reality technology. This system enables appropriate things to give adequate information at the appropriate timing. In this paper, we explain the detail and design of Virtual IoT System. In addition, we conduct an experiment to evaluate the usability and expressive power of Virtual IoT system.<\/p>\n\t\t\t\t<\/div>\n\n\t\t\t\t\n\t\t\t<\/div>\n\n\t\t\t\t\t\n\t\t\t\n\t\t\t<div id=\"su-post-1284\" class=\"su-post \">\n\n\t\t\t\t\n\t\t\t\t<h2 class=\"su-post-title\"><a href=\"https:\/\/sr.sfc.keio.ac.jp\/www\/en\/publication\/1284\/\">Voice Expression with Pseudo-emotion in Human-Robot Communication, Tae Kuwahara, Takumi Horie, Kazunori Takashio, IEICE Tech. Rep., vol. 119, no. 81, CNR2019-2, pp. 7-11, June 2019.<\/a><\/h2>\n\n\t\t\t\t<div class=\"su-post-meta\">\n\t\t\t\t\tPosted: 2019\u5e747\u67089\u65e5\t\t\t\t<\/div>\n\n\t\t\t\t<div class=\"su-post-excerpt\">\n\t\t\t\t\t<p>In a robot and human conversation, the robot\u2019s flat voice is a one of the factor that boring conversation. Automatically adjust emotional expressions to speech from virtual emotion in a robot. In this study, I verify that communication will be smooth when it considers the flow of emotional change from the conversation history.<\/p>\n\t\t\t\t<\/div>\n\n\t\t\t\t\n\t\t\t<\/div>\n\n\t\t\t\t\t\n\t\t\t\n\t\t\t<div id=\"su-post-1282\" class=\"su-post \">\n\n\t\t\t\t\n\t\t\t\t<h2 class=\"su-post-title\"><a href=\"https:\/\/sr.sfc.keio.ac.jp\/www\/en\/publication\/1282\/\">Design of Context Aware Conversation Task Scheduler, Takumi Horie, Kazunori Takashio, IEICE Tech. Rep., vol. 119, no. 81, CNR2019-11, pp. 55-60, June 2019.<\/a><\/h2>\n\n\t\t\t\t<div class=\"su-post-meta\">\n\t\t\t\t\tPosted: 2019\u5e747\u67089\u65e5\t\t\t\t<\/div>\n\n\t\t\t\t<div class=\"su-post-excerpt\">\n\t\t\t\t\t\t\t\t\t<\/div>\n\n\t\t\t\t\n\t\t\t<\/div>\n\n\t\t\t\t\t\n\t\t\t\n\t\t\t<div id=\"su-post-1280\" class=\"su-post \">\n\n\t\t\t\t\n\t\t\t\t<h2 class=\"su-post-title\"><a href=\"https:\/\/sr.sfc.keio.ac.jp\/www\/en\/publication\/1280\/\">Matching Things and Information Using Associative Concept Dictionary in v-IoT, Kentaro Taninaka, Kazunori Takashio, IEICE Tech. Rep., vol. 119, no. 81, CNR2019-8, pp. 37-42, June 2019.<\/a><\/h2>\n\n\t\t\t\t<div class=\"su-post-meta\">\n\t\t\t\t\tPosted: 2019\u5e747\u67089\u65e5\t\t\t\t<\/div>\n\n\t\t\t\t<div class=\"su-post-excerpt\">\n\t\t\t\t\t<p>The miniaturization and decreasing cost of networking computers, alongwith the advancement of cloud infrastructure has eased the implementation of it inproducts for manufacturers. The increase in such products has led to the rise in IoTdevices with information-provision function. However, it is difficult to give suchfunction to products which engage in an one-way association in our daily life likedaily commodities, since it requires the device to have a wealth of computationalresource. In this paper, we propose Virtual IoT system, the approach which addinformation-giving functions to the non-computing objects using Mixed Realitytechnology. This system enables appropriate things to give adequate informationat the right timing. In this paper, we explain the detail and design of Virtual IoTsystem. In addition, we conduct an experiment to evaluate impression and usabilityof Virtual IoT to disclose the descriptive power.<\/p>\n\t\t\t\t<\/div>\n\n\t\t\t\t\n\t\t\t<\/div>\n\n\t\t\t\t\t\n\t\t\t\n\t\t\t<div id=\"su-post-1278\" class=\"su-post \">\n\n\t\t\t\t\n\t\t\t\t<h2 class=\"su-post-title\"><a href=\"https:\/\/sr.sfc.keio.ac.jp\/www\/en\/publication\/1278\/\">Character Making of Personal Robot by Dressing Up Exterior Parts, Riku Takai, Rintaro Miyamoto, Kazunori Takashio, IEICE Tech. Rep., vol. 119, no. 81, CNR2019-6, pp. 27-30, June 2019.<\/a><\/h2>\n\n\t\t\t\t<div class=\"su-post-meta\">\n\t\t\t\t\tPosted: 2019\u5e747\u67089\u65e5\t\t\t\t<\/div>\n\n\t\t\t\t<div class=\"su-post-excerpt\">\n\t\t\t\t\t\t\t\t\t<\/div>\n\n\t\t\t\t\n\t\t\t<\/div>\n\n\t\t\t\t\t\n\t\t\t\n\t\t\t<div id=\"su-post-1275\" class=\"su-post \">\n\n\t\t\t\t\n\t\t\t\t<h2 class=\"su-post-title\"><a href=\"https:\/\/sr.sfc.keio.ac.jp\/www\/en\/publication\/1275\/\">Consideration of Robot Motion Library to Promote Helping Behavior, Mai Ohira, Kentaro Taninaka, Takuto Watanabe, Kazunori Takashio, IEICE Tech. Rep., vol. 119, no. 81, CNR2019-5, pp. 23-25, June 2019.<\/a><\/h2>\n\n\t\t\t\t<div class=\"su-post-meta\">\n\t\t\t\t\tPosted: 2019\u5e747\u67089\u65e5\t\t\t\t<\/div>\n\n\t\t\t\t<div class=\"su-post-excerpt\">\n\t\t\t\t\t\t\t\t\t<\/div>\n\n\t\t\t\t\n\t\t\t<\/div>\n\n\t\t\n\t\n<\/div>\n\n<h2 class = \"adjust-margin-1\">2018<\/h2>\n<hr \/>\n\n\n<div class=\"su-posts su-posts-default-loop \">\n\n\t\n\t\t\t\t\t\n\t\t\t\n\t\t\t<div id=\"su-post-546\" class=\"su-post \">\n\n\t\t\t\t\n\t\t\t\t<h2 class=\"su-post-title\"><a href=\"https:\/\/sr.sfc.keio.ac.jp\/www\/en\/publication\/546\/\">Design of Context Aware Conversation Task Scheduler and its Evaluation : Takumi Horie, Kazunori Takashio, IEICE Tech. Rep., vol. 117, no. 443, CNR2017-30, pp. 45-50, Feb. 2018.<\/a><\/h2>\n\n\t\t\t\t<div class=\"su-post-meta\">\n\t\t\t\t\tPosted: 2018\u5e743\u670820\u65e5\t\t\t\t<\/div>\n\n\t\t\t\t<div class=\"su-post-excerpt\">\n\t\t\t\t\t<p>\u3000<\/p>\n\t\t\t\t<\/div>\n\n\t\t\t\t\n\t\t\t<\/div>\n\n\t\t\t\t\t\n\t\t\t\n\t\t\t<div id=\"su-post-765\" class=\"su-post \">\n\n\t\t\t\t\t\t\t\t\t<a class=\"su-post-thumbnail\" href=\"https:\/\/sr.sfc.keio.ac.jp\/www\/en\/publication\/765\/\"><img loading=\"lazy\" decoding=\"async\" width=\"1024\" height=\"768\" src=\"https:\/\/sr.sfc.keio.ac.jp\/www\/wp-content\/uploads\/2018\/07\/\u30b9\u30af\u30ea\u30fc\u30f3\u30b7\u30e7\u30c3\u30c8-2018-08-09-23.25.42.png\" class=\"attachment-post-thumbnail size-post-thumbnail wp-post-image\" alt=\"\" srcset=\"https:\/\/sr.sfc.keio.ac.jp\/www\/wp-content\/uploads\/2018\/07\/\u30b9\u30af\u30ea\u30fc\u30f3\u30b7\u30e7\u30c3\u30c8-2018-08-09-23.25.42.png 1024w, https:\/\/sr.sfc.keio.ac.jp\/www\/wp-content\/uploads\/2018\/07\/\u30b9\u30af\u30ea\u30fc\u30f3\u30b7\u30e7\u30c3\u30c8-2018-08-09-23.25.42-300x225.png 300w, https:\/\/sr.sfc.keio.ac.jp\/www\/wp-content\/uploads\/2018\/07\/\u30b9\u30af\u30ea\u30fc\u30f3\u30b7\u30e7\u30c3\u30c8-2018-08-09-23.25.42-768x576.png 768w\" sizes=\"(max-width: 1024px) 100vw, 1024px\" \/><\/a>\n\t\t\t\t\n\t\t\t\t<h2 class=\"su-post-title\"><a href=\"https:\/\/sr.sfc.keio.ac.jp\/www\/en\/publication\/765\/\">Handling Conversation Interruption in Many-to-Many HR Interaction Considering Emotional Behaviors and Human Relationships : Takumi Horie, Kazunori Takashio, 27th IEEE International Conference on Robot and Human Interactive Communication (RO-MAN 2018), Aug, 2018.<\/a><\/h2>\n\n\t\t\t\t<div class=\"su-post-meta\">\n\t\t\t\t\tPosted: 2018\u5e747\u67084\u65e5\t\t\t\t<\/div>\n\n\t\t\t\t<div class=\"su-post-excerpt\">\n\t\t\t\t\t<p>In the future, communication robots are expected to join many-to-many human-robot interactions. Thus, robots must handle interruptions requesting a new task outside of the current conversation. In this paper, we propose a novel scheduler which decides switch timing of conversational tasks when a robot is interrupted. The model grasps the structure of the conversation and finds its breakpoints based on adjacency pairs. In order to decide whether to switch conversational tasks on each breakpoint or not, the model prioritizes conversational tasks considering an importance of its topic and a length as contexts of each conversational task. The model also uses human relationships and emotional behaviors to decide priority of conversational tasks. The result of an evaluation experiment shows that our proposed scheduler could impress subjects more favorably than that which always prioritizes an interrupter.<\/p>\n\t\t\t\t<\/div>\n\n\t\t\t\t\n\t\t\t<\/div>\n\n\t\t\t\t\t\n\t\t\t\n\t\t\t<div id=\"su-post-775\" class=\"su-post \">\n\n\t\t\t\t\n\t\t\t\t<h2 class=\"su-post-title\"><a href=\"https:\/\/sr.sfc.keio.ac.jp\/www\/en\/publication\/775\/\">C<sup>2<\/sup>AT<sup>2<\/sup> HUB: Long-term Characterization of Robots based on Human Child\u2019s Personality Development, Shintaro Kawanago, Kazunori Takashio, 27th IEEE International Conference on Robot and Human Interactive Communication (RO-MAN 2018), Aug. 2018.<\/a><\/h2>\n\n\t\t\t\t<div class=\"su-post-meta\">\n\t\t\t\t\tPosted: 2018\u5e747\u67084\u65e5\t\t\t\t<\/div>\n\n\t\t\t\t<div class=\"su-post-excerpt\">\n\t\t\t\t\t<p>Inadequate variety of personalities for communication robots may cause unnatural interac- tion with them and reduction in attachment. This paper proposes C2 AT2 HUB, where communication robots are characterized by changing tendency of affect transition based on long-term interaction with users. In C2AT2 HUB, robots\u2019 affect is defined as two types; &#8220;interpersonal affect&#8221; and &#8220;emotions&#8221;, and transition of each type of affect is adjusted by history of users\u2019 actions to robots in order to characterize robots gradually. Evaluation experiment revealed our approach characterizes robots naturally and improves impression of robots.<\/p>\n\t\t\t\t<\/div>\n\n\t\t\t\t\n\t\t\t<\/div>\n\n\t\t\t\t\t\n\t\t\t\n\t\t\t<div id=\"su-post-800\" class=\"su-post \">\n\n\t\t\t\t\n\t\t\t\t<h2 class=\"su-post-title\"><a href=\"https:\/\/sr.sfc.keio.ac.jp\/www\/en\/publication\/800\/\">Social Interaction with Cloud Network Robots : Kazunori Takashio, 14th International Conference on Intelligent Environments (IE\u201918) \/ Tutorial, June, 2018, Roma Italy.<\/a><\/h2>\n\n\t\t\t\t<div class=\"su-post-meta\">\n\t\t\t\t\tPosted: 2018\u5e747\u67084\u65e5\t\t\t\t<\/div>\n\n\t\t\t\t<div class=\"su-post-excerpt\">\n\t\t\t\t\t<p>http:\/\/ebooks.iospress.nl\/volumearticle\/49339<\/p>\n\t\t\t\t<\/div>\n\n\t\t\t\t\n\t\t\t<\/div>\n\n\t\t\t\t\t\n\t\t\t\n\t\t\t<div id=\"su-post-807\" class=\"su-post \">\n\n\t\t\t\t\n\t\t\t\t<h2 class=\"su-post-title\"><a href=\"https:\/\/sr.sfc.keio.ac.jp\/www\/en\/publication\/807\/\">Design of Physical Interaction between Robots to Build a Sense of Affinity : Lisa Kashiwagi (Sacred Heart School), Rintaro Miyamoto, Kazunori Takashio, IEICE Tech. Rep., vol. 118, no. 94, CNR2018-6, pp. 29-32, June 2018.<\/a><\/h2>\n\n\t\t\t\t<div class=\"su-post-meta\">\n\t\t\t\t\tPosted: 2018\u5e747\u67084\u65e5\t\t\t\t<\/div>\n\n\t\t\t\t<div class=\"su-post-excerpt\">\n\t\t\t\t\t<p>\u3000<\/p>\n\t\t\t\t<\/div>\n\n\t\t\t\t\n\t\t\t<\/div>\n\n\t\t\t\t\t\n\t\t\t\n\t\t\t<div id=\"su-post-809\" class=\"su-post \">\n\n\t\t\t\t\n\t\t\t\t<h2 class=\"su-post-title\"><a href=\"https:\/\/sr.sfc.keio.ac.jp\/www\/en\/publication\/809\/\">A Robot Taking Contrastive Response Strategy to Engage Natural : Shunta Nakajima, Takumi Horie, Tae Kuwahara, Kentaro Taninaka, and Kazunori Takashio, IEICE Tech. Rep., vol. 118, no. 94, CNR2018-5, pp. 23-27, June 2018.<\/a><\/h2>\n\n\t\t\t\t<div class=\"su-post-meta\">\n\t\t\t\t\tPosted: 2018\u5e747\u67084\u65e5\t\t\t\t<\/div>\n\n\t\t\t\t<div class=\"su-post-excerpt\">\n\t\t\t\t\t<p>\u3000<\/p>\n\t\t\t\t<\/div>\n\n\t\t\t\t\n\t\t\t<\/div>\n\n\t\t\t\t\t\n\t\t\t\n\t\t\t<div id=\"su-post-814\" class=\"su-post \">\n\n\t\t\t\t\n\t\t\t\t<h2 class=\"su-post-title\"><a href=\"https:\/\/sr.sfc.keio.ac.jp\/www\/en\/publication\/814\/\">The Robot Design of step by step Growth Process : Kouki Saito (Keio Fujisawa HS.), Kentaro Taninaka, Kazunori Takashio, IEICE Tech. Rep., vol. 118, no. 94, CNR2018-4, pp. 19-22, June 2018.<\/a><\/h2>\n\n\t\t\t\t<div class=\"su-post-meta\">\n\t\t\t\t\tPosted: 2018\u5e747\u67084\u65e5\t\t\t\t<\/div>\n\n\t\t\t\t<div class=\"su-post-excerpt\">\n\t\t\t\t\t<p>\u3000<\/p>\n\t\t\t\t<\/div>\n\n\t\t\t\t\n\t\t\t<\/div>\n\n\t\t\t\t\t\n\t\t\t\n\t\t\t<div id=\"su-post-1012\" class=\"su-post \">\n\n\t\t\t\t\n\t\t\t\t<h2 class=\"su-post-title\"><a href=\"https:\/\/sr.sfc.keio.ac.jp\/www\/en\/publication\/1012\/\">Effective use of AR and IoT Technology for Displaying Life-support Information &#8212; Comparison between IoT Devises versus Information-giving from Commodity &#8211;, Kentaro Taninaka, Kazunori Takashio, IEICE Tech. Rep., vol. 118, no. 184, CNR2018-12, pp. 7-11, Aug. 2018.<\/a><\/h2>\n\n\t\t\t\t<div class=\"su-post-meta\">\n\t\t\t\t\tPosted: 2019\u5e742\u670828\u65e5\t\t\t\t<\/div>\n\n\t\t\t\t<div class=\"su-post-excerpt\">\n\t\t\t\t\t<p>IoT devices which display life-support information to users is increasing. However, it is difficult to add information-giving function to the product which close to user\u2019s life, i.e. commodity. In fact, this information is given from the device which have a wealth computation source. In this research, we propose the approach which add information-giving function to those IoT introduction difficult things, using augmented reality technology. This system enables right things to give right information in right timing.<\/p>\n\t\t\t\t<\/div>\n\n\t\t\t\t\n\t\t\t<\/div>\n\n\t\t\t\t\t\n\t\t\t\n\t\t\t<div id=\"su-post-1014\" class=\"su-post \">\n\n\t\t\t\t\n\t\t\t\t<h2 class=\"su-post-title\"><a href=\"https:\/\/sr.sfc.keio.ac.jp\/www\/en\/publication\/1014\/\">Effectiveness of Manga-Effects To Telepresence Communication System, Risa Takemura, Takumi Horie, Kazunori Takashio, IEICE Tech. Rep., vol. 118, no. 306, CNR2018-33, pp. 87-92, Nov. 2018.<\/a><\/h2>\n\n\t\t\t\t<div class=\"su-post-meta\">\n\t\t\t\t\tPosted: 2019\u5e742\u670828\u65e5\t\t\t\t<\/div>\n\n\t\t\t\t<div class=\"su-post-excerpt\">\n\t\t\t\t\t<p>Today\u2019s telecommunication systems have been greatly evolving, while the video calls used on telepresence robots still seems bland. We hypothesize that the augmentation of video calls with Manga-Effects will help users attain a sense of familiarity and find entertaining components within their conversations. In specific, adding various Manga-Effects over the video call, respectively selected upon the users\u2019 facial expression, will help visually emphasize certain emotions. This experiment analyzes the level of users\u2019 involvement in the conversation, and verifies our hypothesis.<\/p>\n\t\t\t\t<\/div>\n\n\t\t\t\t\n\t\t\t<\/div>\n\n\t\t\n\t\n<\/div>\n\n<h2 class = \"adjust-margin-1\">\u301c2018<\/h2>\n<hr \/>\n\n\n<div class=\"su-posts su-posts-default-loop \">\n\n\t\n\t\t\t\t\t\n\t\t\t\n\t\t\t<div id=\"su-post-471\" class=\"su-post \">\n\n\t\t\t\t\n\t\t\t\t<h2 class=\"su-post-title\"><a href=\"https:\/\/sr.sfc.keio.ac.jp\/www\/en\/2017\/471\/\">Ex-Amp Robot: Expressive Robotic Avatar with Multimodal Emotion Detection to Enhance Communication of Users with Motor Disabilities : Ai Kashii,  Kazunori Takashio and Hideyuki Tokuda, 26th IEEE International Symposium on Robot and Human Interactive Communication (RO-MAN 2017), Aug to Sept, 2017.<\/a><\/h2>\n\n\t\t\t\t<div class=\"su-post-meta\">\n\t\t\t\t\tPosted: 2018\u5e743\u670819\u65e5\t\t\t\t<\/div>\n\n\t\t\t\t<div class=\"su-post-excerpt\">\n\t\t\t\t\t<p>In current society, there are numerous robots made for various purposes, including manufacturing, cleaning, therapy, and customer service. Other robots are used for enhancing H2H communication. In this research, we proposed a robotic system which detects the user\u2019s emotions and enacts them on a humanoid robot. By using this robotic avatar, users with motor disabilities are able to extend their methods of communication, as a physical form of expression will be added to the conversation.<\/p>\n\t\t\t\t<\/div>\n\n\t\t\t\t\n\t\t\t<\/div>\n\n\t\t\t\t\t\n\t\t\t\n\t\t\t<div id=\"su-post-540\" class=\"su-post \">\n\n\t\t\t\t\n\t\t\t\t<h2 class=\"su-post-title\"><a href=\"https:\/\/sr.sfc.keio.ac.jp\/www\/en\/2017\/540\/\">Atomicity Detection in Dialogue Flow Considering Sequences of Adjacency Pairs : Takumi Horie, Kazunori Takashio, IEICE Tech. Rep., vol. 117, no. 95, CNR2017-5, pp. 23-28, June 2017.<\/a><\/h2>\n\n\t\t\t\t<div class=\"su-post-meta\">\n\t\t\t\t\tPosted: 2018\u5e743\u670820\u65e5\t\t\t\t<\/div>\n\n\t\t\t\t<div class=\"su-post-excerpt\">\n\t\t\t\t\t<p>\u3000<\/p>\n\t\t\t\t<\/div>\n\n\t\t\t\t\n\t\t\t<\/div>\n\n\t\t\t\t\t\n\t\t\t\n\t\t\t<div id=\"su-post-542\" class=\"su-post \">\n\n\t\t\t\t\n\t\t\t\t<h2 class=\"su-post-title\"><a href=\"https:\/\/sr.sfc.keio.ac.jp\/www\/en\/2017\/542\/\">MoDe: a Real-time Deformed Motion Generation Tool for Humanoid Robots : Ryota Hayashi (MUHS), Takumi Horie, Taiki Majima, Shintaro Kawanago, Rintaro Miyamoto, Kazunori Takashio, IEICE Tech. Rep., vol. 117, no. 95, CNR2017-8, pp. 41-44, June 2017.<\/a><\/h2>\n\n\t\t\t\t<div class=\"su-post-meta\">\n\t\t\t\t\tPosted: 2018\u5e743\u670820\u65e5\t\t\t\t<\/div>\n\n\t\t\t\t<div class=\"su-post-excerpt\">\n\t\t\t\t\t<p>\u3000<\/p>\n\t\t\t\t<\/div>\n\n\t\t\t\t\n\t\t\t<\/div>\n\n\t\t\t\t\t\n\t\t\t\n\t\t\t<div id=\"su-post-544\" class=\"su-post \">\n\n\t\t\t\t\n\t\t\t\t<h2 class=\"su-post-title\"><a href=\"https:\/\/sr.sfc.keio.ac.jp\/www\/en\/2017\/544\/\">Social Pods: a Robot Kit for Design-Oriented STEM Education using Digital Fabrication Environment : Taiki Majima, Kazunori Takashio, IEICE Tech. Rep., vol. 117, no. 198, CNR2017-14, pp. 31-36, Sept. 2017.<\/a><\/h2>\n\n\t\t\t\t<div class=\"su-post-meta\">\n\t\t\t\t\tPosted: 2018\u5e743\u670820\u65e5\t\t\t\t<\/div>\n\n\t\t\t\t<div class=\"su-post-excerpt\">\n\t\t\t\t\t<p>\u3000<\/p>\n\t\t\t\t<\/div>\n\n\t\t\t\t\n\t\t\t<\/div>\n\n\t\t\t\t\t\n\t\t\t\n\t\t\t<div id=\"su-post-548\" class=\"su-post \">\n\n\t\t\t\t\n\t\t\t\t<h2 class=\"su-post-title\"><a href=\"https:\/\/sr.sfc.keio.ac.jp\/www\/en\/2017\/548\/\">C<sup>2<\/sup>AT<sup>2<\/sup> HUB: Long-term Characterization of Robots based on Human Child\u2019s Personality Development, Shintaro Kawanago, Kazunori Takashio, IEICE Tech. Rep., vol. 117, no. 443, CNR2017-45, pp. 127-132, Feb. 2018.<\/a><\/h2>\n\n\t\t\t\t<div class=\"su-post-meta\">\n\t\t\t\t\tPosted: 2018\u5e743\u670820\u65e5\t\t\t\t<\/div>\n\n\t\t\t\t<div class=\"su-post-excerpt\">\n\t\t\t\t\t<p>\u3000<\/p>\n\t\t\t\t<\/div>\n\n\t\t\t\t\n\t\t\t<\/div>\n\n\t\t\n\t\n<\/div>\n\n\n\n<div class=\"su-posts su-posts-default-loop \">\n\n\t\n\t\t\t\t\t\n\t\t\t\n\t\t\t<div id=\"su-post-483\" class=\"su-post \">\n\n\t\t\t\t\n\t\t\t\t<h2 class=\"su-post-title\"><a href=\"https:\/\/sr.sfc.keio.ac.jp\/www\/en\/2016\/483\/\">Ex-Amp Robot: Physical Avatar for Enhancing Human to Human Communication : Ai Kashii,  Kazunori Takashio,  Hideyuki Tokuda, International Conference on Human-Agent Interaction 2016 (HAI 2016), Dec, 2016.<\/a><\/h2>\n\n\t\t\t\t<div class=\"su-post-meta\">\n\t\t\t\t\tPosted: 2018\u5e743\u670819\u65e5\t\t\t\t<\/div>\n\n\t\t\t\t<div class=\"su-post-excerpt\">\n\t\t\t\t\t<p>Our research focuses on creating a robotic system that aids human-to-human communication. The robot acts as a personal companion that understands the user\u2019s emotions and helps express them alongside the user. First, the user\u2019s facial expression is detected through a connected camera device and relays the retrieved information to a humanoid robot. The humanoid robot then performs physical gestures that match the detected emotion. By using this system, those who are unable to freely move their own bodies can add a physical component to their communication method. In this paper, we have determined the efficacy of translating detected facial expressions into robot movements. Through experiments and surveys, we determined whether our proposed \u2018Ex-Amp Robot\u2019 helped enhance the communication of a hypothetically tetraplegic user.<\/p>\n\t\t\t\t<\/div>\n\n\t\t\t\t\n\t\t\t<\/div>\n\n\t\t\t\t\t\n\t\t\t\n\t\t\t<div id=\"su-post-485\" class=\"su-post \">\n\n\t\t\t\t\n\t\t\t\t<h2 class=\"su-post-title\"><a href=\"https:\/\/sr.sfc.keio.ac.jp\/www\/en\/2016\/485\/\">Robot Avatar for Enhancing the Communication of Physically Disabled Humans : Ai Kashii,  Kazunori Takashio and Hideyuki Tokuda 25th IEEE International Symposium on Robot and Human Interactive Communication (RO-MAN 2016), Aug, 2016<\/a><\/h2>\n\n\t\t\t\t<div class=\"su-post-meta\">\n\t\t\t\t\tPosted: 2018\u5e743\u670819\u65e5\t\t\t\t<\/div>\n\n\t\t\t\t<div class=\"su-post-excerpt\">\n\t\t\t\t\t<p>The robot avatar proposed in this paper is a personal companion that captures users\u2019 facial expressions and translates retrieved emotion data into gestures best suited to express the user\u2019s emotion in real-time. This robot allows users to enjoy physical aspects of communication, as well as put an impact on parts of conversations to raise conversation quality. In this research, we have conducted experiments to validate the efficacy of translating facial expressions into robot movements.<\/p>\n\t\t\t\t<\/div>\n\n\t\t\t\t\n\t\t\t<\/div>\n\n\t\t\t\t\t\n\t\t\t\n\t\t\t<div id=\"su-post-538\" class=\"su-post \">\n\n\t\t\t\t\n\t\t\t\t<h2 class=\"su-post-title\"><a href=\"https:\/\/sr.sfc.keio.ac.jp\/www\/en\/2016\/538\/\">Expression Amplifying Robot &#8212; A Personal Expression Translator &#8212; : Ai Kashii, Kazunori Takashio, Hideyuki Tokuda IEICE Tech. Rep., vol. 116, no. 106, CNR2016-3, pp. 11-16, June 2016.<\/a><\/h2>\n\n\t\t\t\t<div class=\"su-post-meta\">\n\t\t\t\t\tPosted: 2018\u5e743\u670820\u65e5\t\t\t\t<\/div>\n\n\t\t\t\t<div class=\"su-post-excerpt\">\n\t\t\t\t\t<p>Currently, various different robots aid and enhance our daily lives. They help us in the form of communication, manufacturing, and physical labor. Even within the category of enhancing communication, robots may be used in a myriad of ways, including use as an avatar (a representation of an identity) or as a conversation companion on its own. In this paper, we have focused on creating a robotic system that would aid human-to-human communication. This robot acts as a personal companion that understands the user\u2019s emotions and helps express them alongside the user. It first detects the user\u2019s facial expression through a connected camera device, which then relays the retrieved information to a humanoid robot. The humanoid robot then performs physical gestures, according to the detected emotion. By using this personal robot system, a person who is unable to freely move their own body would be able to add a physical component to their method of communication, as the robot will act as the user\u2019s body in the communication. This time around, we have conducted experiments to validate the efficacy of translating detected facial expressions into robot movements. Through experiments and surveys, we determined whether our proposed \u2018Ex-Amp Robot\u2019 helped enhance communication between a hypothetically tetraplegic user and another person.<\/p>\n\t\t\t\t<\/div>\n\n\t\t\t\t\n\t\t\t<\/div>\n\n\t\t\n\t\n<\/div>\n\n<p><!--\n\n\n<h3 class=\"adjust-margin-1\">Other Publications of Prof. Takashio<\/h3>\n\n\n\n\n<p>SFC Faculty Page\n<a href=\"https:\/\/vu.sfc.keio.ac.jp\/faculty_profile\/cgi\/f_profile_e.cgi?id=a464a3ec55ed9a77\">https:\/\/vu.sfc.keio.ac.jp\/faculty_profile\/cgi\/f_profile_e.cgi?id=a464a3ec55ed9a77<\/a><\/p>\n\n\n--><\/p>\n","protected":false},"excerpt":{"rendered":"<p>2026\u301c 2025\u301c 2024\u301c 2023\u301c 2022\u301c 2021\u301c 2020\u301c 2019 2018 \u301c2018<\/p>\n","protected":false},"author":9,"featured_media":0,"parent":0,"menu_order":0,"comment_status":"closed","ping_status":"closed","template":"","meta":{"footnotes":""},"class_list":["post-1464","page","type-page","status-publish","hentry","post"],"_links":{"self":[{"href":"https:\/\/sr.sfc.keio.ac.jp\/www\/en\/wp-json\/wp\/v2\/pages\/1464"}],"collection":[{"href":"https:\/\/sr.sfc.keio.ac.jp\/www\/en\/wp-json\/wp\/v2\/pages"}],"about":[{"href":"https:\/\/sr.sfc.keio.ac.jp\/www\/en\/wp-json\/wp\/v2\/types\/page"}],"author":[{"embeddable":true,"href":"https:\/\/sr.sfc.keio.ac.jp\/www\/en\/wp-json\/wp\/v2\/users\/9"}],"replies":[{"embeddable":true,"href":"https:\/\/sr.sfc.keio.ac.jp\/www\/en\/wp-json\/wp\/v2\/comments?post=1464"}],"version-history":[{"count":39,"href":"https:\/\/sr.sfc.keio.ac.jp\/www\/en\/wp-json\/wp\/v2\/pages\/1464\/revisions"}],"predecessor-version":[{"id":3171,"href":"https:\/\/sr.sfc.keio.ac.jp\/www\/en\/wp-json\/wp\/v2\/pages\/1464\/revisions\/3171"}],"wp:attachment":[{"href":"https:\/\/sr.sfc.keio.ac.jp\/www\/en\/wp-json\/wp\/v2\/media?parent=1464"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}