東京大学大学院 情報学環・学際情報学府 The University of Tokyo III / GSII

イベント Event

April 27, 2015

HCI研究者の講演会 (CHI researcher talk)

ソウルで開催されたHCI (Human-Computer Interaction)のトップカンファレンスである、ACM SIGCHIの後、2名の研究者が東大を訪問して、講演会を4月27日に以下要領で開催することになりました。どなたでも参加可能ですのでぜひお越しください。
(主催:情報学環暦本研究室、共催:ユビキタス情報社会基盤研究センター)
After SIGCHI conference in Seoul, two HCI researchers will visit Tokyo University and give talks.  It is open for everybody so please join us.
Date: April 27 (mon) 13:30-15:00 (door open at 13:15)
Venue: Ishibashi Hall, 3F, Daiwa Ubiquitous Computing Research Building
TALK 1
Alex Olwal
Ph.D, Google, USA
<img src=”http://olwal.com/images/alex/alex_olwal_400_400_bw.jpg” alt=”” width=”200″ height=”200″ />
TITLE: Augmented Realism through Relevant Minimalism
ABSTRACT
Augmented Reality (AR) merges virtual information with a real environment for intuitive and direct user interfaces. I will give an overview of our research projects that emphasize a seamless blend of the digital and physical. We leverage exotic displays, sensing and context, to register digital content with the environment and to support rich interactions. Our goal is to render less, with every pixel being relevant, dynamic and grounded in the space. This is one of our key philosophies for enhancing the human senses and capabilities through minimal, yet highly relevant augmentations, that embrace the realism of our physical world. Our projects include novel interaction techniques, see-through displays, sensing technologies, immaterial user interfaces, and dynamic shape displays.
BIO
Alex Olwal (Ph.D., M.Sc.) is an Interaction Researcher at Google, Affiliate Faculty at KTH, and Research Affiliate at the MIT Media Lab.
Alex designs and develops interactions and technologies that embrace digital and physical experiences. He is interested in tools, techniques and devices that enable new interaction concepts for the augmentation and empowerment of the human senses.
Alex’s research (olwal.com) includes augmented reality, spatially aware mobile devices, medical user interfaces, ubiquitous computing, touch-screens, as well as novel interaction devices and displays.
He has previously worked with the development of new technologies for Human-Computer Interaction at MIT – Massachusetts Institute of Technology (Cambridge, MA), KTH – Royal Institute of Technology (Stockholm), Columbia University (NY), University of California (Santa Barbara, CA) and Microsoft Research (Redmond, WA).
At Google, Alex conducts applied research and development of novel and exotic input and output mechanisms for wearables, and explore associated
interaction techniques. Alex’s goal is to expand expressiveness while avoiding interference with user’s experience of the physical reality.
TALK 2:
Pedro Lopes
Ph.D Candidate, Computer Interaction Lab, Hasso Platner Institute, Germany
TITLE: Affordance++ and Proprioceptive Interaction
ABSTRACT
We propose extending the affordance of objects by allowing them to communicate dynamic use, such as (1) motion (e.g., spray can shakes when touched), (2) multi-step processes (e.g., spray can sprays only after shaking), and (3) behaviors that change over time (e.g., empty spray can does not allow spraying anymore). Rather than enhancing objects directly, however, we implement this concept by enhancing the user. We call this affordance++. By stimulating the user’s arms using electrical muscle stimulation, our prototype allows objects not only to make the user actuate them, but also perform required movements while merely approaching the object, such as not to touch objects that do not “want” to be touched. In our user study, affordance++ helped participants to successfully operate devices of poor natural affordance, such as a multi-functional slicer tool or a magnetic nail sweeper, and to stay away from cups filled with hot liquids. We call this concept of creating object behavior by controlling user behavior affordance++. Conceptually there are many ways of implementing affordance++, generally by applying sensors and actuators to the user’s body, such as the arm. In this paper, we actuate users by controlling their arm poses using electrical muscle stimulation, i.e., users wear a device on their arm that talks to the user’s muscles by means of electrodes attached to the user’s arm (we describe the device in detail in section “Prototype”). This allows for a particularly compact form factor and is arguably even more “direct” than the indirection through a mechanical system. However, the concept of affordance++ needs not to be tied to a particular means of actuating the user, but to the concept of doing so instead of actuating the objects that the user interacts with.
BIO