Investigating Accuracy of Tilting Operation on Wrist-worn Devices with
Touchscreens
Late-Breaking Works: Novel Interactions
/
Shima, Keigo
/
Onishi, Kazusa
/
Takada, Ryosuke
/
Adachi, Takuya
/
Shizuki, Buntarou
/
Tanaka, Jiro
Extended Abstracts of the ACM CHI'16 Conference on Human Factors in
Computing Systems
2016-05-07
v.2
p.2705-2711
© Copyright 2016 ACM
Summary: The tilting operation has been shown to be useful in expanding the input
vocabulary of a small device; therefore, this operation could also be useful in
expanding the input vocabulary of wrist-worn devices. In this study, in order
to explore new design guidelines of GUI and operation methods for smartwatches,
we have investigated the accuracy of the tilting operation for smartwatches. We
conducted an experiment using a cursor whose position is determined according
to the direction and angle of tilt. The result shows that the deviation in the
x-axis (i.e., the inclination of the arm) direction ranges from -0.10 degrees
to +0.07 degrees and from -0.10 degrees to +0.07 degrees under Seating and
Standing condition, respectively; the deviation in the y-axis (i.e., the
rotation of the wrist) direction from -0.12 degrees to +0.10 degrees and from
-0.12 degrees to +0.09 degrees.
Virtual Liver Surgical Simulator by Using Z-Buffer for Object Deformation
Universal Access to Health Applications and Services
/
Onishi, Katsuhiko
/
Noborio, Hiroshi
/
Koeda, Masanao
/
Watanabe, Kaoru
/
Mizushino, Kiminori
/
Kunii, Takahiro
/
Kaibori, Masaki
/
Matsui, Kosuke
/
Kon, Masanori
UAHCI 2015: 9th International Conference on Universal Access in
Human-Computer Interaction, Part III: Access to Learning, Health and Well-Being
2015-08-02
v.3
p.345-351
Keywords: Z-buffer; Liver surgical simulator; GPU; Object deformation
© Copyright 2015 Springer International Publishing Switzerland
Summary: Virtual surgical simulator which is using computer graphics is much popular
system than before. It is generally used in the medical areas, such as medical
hospital or medical university. The simulator uses virtual organ models like
liver, brain and so on. These models are usually based on the scanning data
from patients and are used as volume models. Fortunately, the volume model is
familiar with its cutting or deforming operation in a surgical system. For this
reason, there are many kinds of surgical simulation or navigation systems using
the volume model. However, visual reality of the volume model is not sufficient
for human being including doctors. This means that the doctors cannot identify
shape or location of a target organ from volume objects. In order to overcome
this, we should use the translating method, such as marching cubes method and
so on, for getting precisely polygon models which is included normal vectors of
volume object. However, the method is quite time consuming and consequently the
doctors cannot operate the virtual model in real-time.
On these observations, we propose the virtual surgical simulator for
operating the human liver in a virtual environment, which is based on the
cooperation of polygon models and Z-buffer in GPU. By using parallel processing
of GPU, the simulator allows uses to cut or deform a virtual liver model by
using several kinds of medical tools like a scalpel in this system. In
addition, visual reality of polygon model is wonderful for a doctor to identify
its shape or location because this model maintains their precise normal vector.
Investigation of Transferring Touch Events for Controlling a Mobile Device
with a Large Touchscreen
Touch-Based and Haptic Interaction
/
Onishi, Kazusa
/
Shizuki, Buntarou
/
Tanaka, Jiro
HCI International 2015: 17th International Conference on HCI, Part II:
Interaction Technologies
2015-08-02
v.2
p.250-261
Keywords: Large mobile device; Touch gesture; Single-handed control
© Copyright 2015 Springer International Publishing Switzerland
Summary: When users hold large mobile devices equipped with a large touchscreen in
one hand, the region distant from the thumb is too distant for users to
control. This forces users to change their hand posture so that their thumb can
reach to the top half. To address this problem, we explore a technique that
transfers touch events on the bottom half of a touchscreen to its top half.
This technique may allow users to control all regions of a large touchscreen by
using only the bottom half. Thus, users can control a mobile device without
changing hand posture. We conducted a user study to investigate the feasibility
of our technique. From the results, our technique is marginally faster than
direct touch and thus, might be feasible.
Study of Tile Menu Selection Technique Using the Relative Position of Joints
for Gesture Operation
Gesture, Gaze and Motion Detection, Modelling and Recognition
/
Gomi, Yamato
/
Onishi, Katsuhiko
HCI International 2015: 17th International Conference on HCI: Posters'
Extended Abstracts, Part I
2015-08-02
v.4
p.481-484
Keywords: 3D pointing; Selection method; Gesture; Two-hand manipulation
© Copyright 2015 Springer International Publishing Switzerland
Summary: In this paper, we describe about our selected method by using the 3D tile
menu by hand gesture motion. In the discussion of our approach, it is a mainly
subject to realize the efficient selection gesture by user's arm for any
position on the screen. Therefore, we designed the selection method by using
relative position of the user's arm joints. The method uses user's hands,
elbows and shoulder position at each arm. It recognized user's selected points
by the relative position of these joints. We make the prototype system which
has been implemented our method. And we examined the basic evaluation of our
selection method by comparing with the conventional method. As a result of this
evaluation, it is confirmed that our method allows users to perform smooth
selecting operation regardless of the position.
Depth Camera Calibration and Knife Tip Position Estimation for Liver Surgery
Support System
Gesture, Gaze and Motion Detection, Modelling and Recognition
/
Koeda, Masanao
/
Tsukushi, Akio
/
Noborio, Hiroshi
/
Onishi, Katsuhiko
/
Mizushino, Kiminori
/
Kunii, Takahiro
/
Watanabe, Kaoru
/
Kaibori, Masaki
/
Matsui, Kosuke
/
Kwon, Masanori
HCI International 2015: 17th International Conference on HCI: Posters'
Extended Abstracts, Part I
2015-08-02
v.4
p.496-502
© Copyright 2015 Springer International Publishing Switzerland
Summary: We have developed a liver surgery support system that uses two depth cameras
and measures positional relationships between a surgical knife and a liver in
real time. In this report, the overview of our system, the method for depth
camera calibration, the estimation for knife tip positioning, and some
experimental results are described.
Haptic AR Dental Simulator Using Z-buffer for Object Deformation
Health and Rehabilitation Applications
/
Onishi, Katsuhiko
/
Mizushino, Kiminori
/
Noborio, Hiroshi
/
Koeda, Masanao
UAHCI 2014: 8th International Conference on Universal Access in
Human-Computer Interaction, Part III: Aging and Assistive Environments
2014-06-22
v.3
p.342-348
Keywords: Collision detection; dental surgical simulator; augmented reality; GPU
© Copyright 2014 Springer International Publishing
Summary: Dental surgical simulator could be one of the efficient tools to learning
and practicing dental surgical skills. To these simulators, the visual and
tactile feedback is desirable to be processed in real time. And, in the dental
operation, the hand position during operations is one of the skills to learn
and practice. Therefore, we develop the dental surgical simulator which use
virtual tooth surface model for processing real time rendering. And we develop
a display system which allow users to training dental operation by a right hand
position.. The tooth model is deformed by cutting and drilling operation using
haptic device. And the display is set close to user's hand position and shows
combined image with virtual tooth model as a surgical target and a real tooth
model as other parts of the patient dental model. The system uses a collision
detection and deformation method by using Z-buffer for virtual objects. This
method enables users to view the complex shape of virtual tooth model by the
surgical operation tasks and practicing dental surgical tasks. We developed
prototype system and confirmed about the capability of our system.
HandyScope: A Remote Control Technique Using Circular Widget on Tabletops
Gesture-Based Interaction
/
Kuribara, Takuro
/
Mita, Yusaku
/
Onishi, Kazusa
/
Shizuki, Buntarou
/
Tanaka, Jiro
HCI International 2014: 16th International Conference on HCI, Part II:
Advanced Interaction Modalities and Techniques
2014-06-22
v.2
p.69-80
Keywords: bimanual interaction; multi-touch; gesture; dynamic control-display gain;
pointing; target acquisition; pull-out
© Copyright 2014 Springer International Publishing
Summary: A large multi-touch tabletop has remote areas that the users might not touch
by their hands. This forces users to move around the tabletop. In this paper,
we present a novel remote control technique which we call HandyScope. This
technique allows users to manipulate those remote areas. Moreover, users can
move an object between the nearby area and the remote areas using a widget. In
addition, users can precisely point a remote area quickly because this system
includes our proposed control-display ratio changing system. To evaluate the
performance of HandyScope, we compared HandyScope with direct touch
manipulation. The results show that HandyScope is significantly faster in
selection.
AR Dental Surgical Simulator Using Haptic Feedback
Health and Medicine
/
Onishi, Katsuhiko
/
Mizushino, Kiminori
/
Ikemoto, Hiroki
/
Noborio, Hiroshi
HCI International 2013: 15th International Conference on HCI: Posters'
Extended Abstracts Part II
2013-07-21
v.7
p.202-205
Keywords: Dental surgical simulator; augmented reality; direct manipulation
© Copyright 2013 Springer-Verlag
Summary: We describe about our dental surgical simulator which enable users to
simulate dental surgical operation. Our simulator which enables the user to
learn dental surgical methods through actual hand and body postures. The
proposed system uses a display showing a virtual tooth model and real teeth and
gums that are positioned close to the hands of the user, which allows the user
to directly manipulate objects with haptic feedback. As a preliminary
evaluation, in display system, we measured the deviation between real object
image and virtual object image at user's view positions. And we confirmed the
capability and the limitation of our system.
Development of AR Display System for Dental Surgical Simulator
Part II / Virtual and Augmented Environments
/
Onishi, Katsuhiko
/
Ito, Shota
/
Kawamura, Yusuke
/
Noborio, Hiroshi
HCI International 2011: 14th International Conference on HCI - Posters'
Extended Abstracts, Part II
2011-07-09
v.6
p.130-133
Keywords: dental surgical simulator; direct manipulation; augmented reality
Copyright © 2011 Springer-Verlag
Summary: In this paper, we describe about a display system for dental surgical
simulator. The design concept of our system is that the system allows the user
to learning the dental surgical methods with the real hand and body posture.
Therefore our system set a display, which shows virtual teeth model and real
teeth and gums, close to user's hand position and allows the user to
manipulating object directly.
Designing space for socially meaningful creativity enhanced by new
technologies
Demonstrations
/
Miyata, Yoshiro
/
Ueda, Nobuyuki
/
Onishi, Keiko
/
Sowa, Tomoyuki
/
Harada, Yasushi
/
Mogi, Kazuzi
/
Tetsuka, Chihiro
/
Inoue, Masaki
Proceedings of the 2009 Conference on Creativity and Cognition
2009-10-26
p.431-432
Keywords: collaborative creativity, lifelong creativity, scratch/cricket,
trans-cultural learning, workshops
© Copyright 2009 ACM
Summary: We will demonstrate some workshop space that we have designed for creating
and sharing socially meaningful activities that transcend differences in
culture and age. We argue for the importance of: 1. Sharing at multiple levels:
bodily/sensory, emotional, intellectual, and social levels. 2. Connecting these
levels by introducing technologies in response to socially shared meanings. In
our demonstration, we will show some activities incorporating Scratch/Cricket
technologies to enhance sharing and connecting these levels.
Design of a learning place for collaborative creation by the mode of Byo-bu
(Japanese folding screen)
Workshops
/
Tetsuka, Chihiro
/
Mogi, Kazuzi
/
Miyata, Yoshiro
/
Ueda, Nobuyuki
/
Harada, Yasushi
/
Sowa, Tomoyuki
/
Onishi, Keiko
/
Inoue, Masaki
Proceedings of the 2009 Conference on Creativity and Cognition
2009-10-26
p.489-490
Keywords: collaborative creation, collaborative learning, Japanese art, Japanese
traditional culture
© Copyright 2009 ACM
Summary: This workshop aims increase understanding through experience the style of a
traditional Japanese art. The idea is to develop growing "images" from the
impressions gotten from the paintings on the "Byo-bu (Folding screen)", and
then to re-create with the others a "New story" coming from these "images".
Thus, we would like to think about a collaborative creation based on individual
or common ideas and the process of creative thinking that were raised by
Byo-bu.
AirFlip-Undo: Quick Undo using a Double Crossing In-Air Gesture in Hover
Zone
Posters
/
Shima, Keigo
/
Takada, Ryosuke
/
Onishi, Kazusa
/
Adachi, Takuya
/
Shizuki, Buntarou
/
Tanaka, Jiro
Adjunct Proceedings of the 2015 ACM Symposium on User Interface Software and
Technology
2005-11-08
v.2
p.97-98
© Copyright 2005 ACM
Summary: In this work, we use AirFlip to undo text input on mobile touchscreen
devices. AirFlip involves a quick double crossing in-air gesture in the
boundary surfaces of hover zone of devices that have hover sensing capability.
To evaluate the effectiveness of undoing text input with AirFlip, we
implemented two QWERTY soft keyboards (AirFlip keyboard and Typical keyboard).
With these keyboards, we conducted a user study to investigate the users'
workload and to collect subjective opinions. The results show that there is no
significant difference in workload between keyboards.
Interactive modeling of trees by using growth simulation
Interaction techniques
/
Onishi, Katsuhiko
/
Hasuike, Shoichi
/
Kitamura, Yoshifumi
/
Kishino, Fumio
Proceedings of the 2003 ACM Symposium on Virtual Reality Software and
Technology
2003-10-01
p.66-72
Keywords: L-system, data structure, intuitive interaction, multimodal interface, tree
models, virtual reality
© Copyright 2003 ACM
Summary: We propose a real-time interactive system that enables users to generate,
manipulate and edit the shape model of a tree based on growth simulation by
directly indicating its global and spatial information. For this purpose,
three-dimensional (3D) spatial information is introduced to the well-known
L-system as an attribute of the growth simulation. Moreover, we propose an
efficient data structure of L-strings in order to speed up the process.