2.literature Review: 2.1 Bare Hand Computer Interaction
2.literature Review: 2.1 Bare Hand Computer Interaction
LITERATURE REVIEW
2.1 Bare Hand Computer Interaction
classification were discussed. They applied their work for control of an on-
brainstorming tool that would allow different users to arrange text across the
1
Hand segmentation techniques such as stereo image segmentation,
detecting the fingers and hands. They describe a simple and reliable
algorithm based on finding fingertips and from which fingers and eventually
the whole hand can be identified. The algorithm is based on a simple model
the fingertips, a model of the fingers and eventually hand can be generated
2
Figure 2-2. Finger Model used by Hardenburg and Berard
The end system that was developed had a real time capacity of around
20-25 Hz. Data from their evaluation shows about 6 frames out of 25 are
0.5 and 1.9 pixels. They have concluded in their paper that the system
conditions.
Menus to Develop Command Set for Computer Vision” [6] published in the
for the present interaction tools such as remote control, mouse, etc.
Perspective and Multimodal User Interfaces are the two main scenarios
3
discussed for gestural interfaces. First the Perspective User Interface aims
gestural command sets. The first dimension was the Cognitive aspect, this
aspects being the second dimension tackle on how gestures are easy to
perform or how tiring it will be for the user. The last dimension was on the
Technical aspects. This refers to the command sets must be state of the art
commands can be then easily recognize. Pie and Marking menus are the two
types of menu structures that the authors discussed and explained. Pie
menus are pop-up menus with included alternatives that are arranged in a
implementing sub-menus.
4
As a test, a prototype for hand gesture interaction was performed.
Lenman, Bretzer and Thrusson have chosen a hierarchic menu system for
computer vision system was the representation of the hand. The system will
multiscale color detector and particle filtering. Hand poses are then
has three hierarchical levels and four choices. Menus then are shown on a
As for their future work, they are attempting to increase the speed and
Imaging and Image processing. It contains or talks about different areas like
5
gesture recognition, segmentation, etc. that are needed to complete the
research and the techniques that will be used for it. Already there has been a
sensing the environment with sense of sight and hearing must be imitated by
the computer.
on gesture recognition and finger tracking used as interface in the PC. There
structure will project a display on a Place Holder Object (PHO) and by the
use of your own hand the system can create controls and situations for the
display, movements and gestures of the hand are detected by the Head
There are two main areas of problem and the presentation of their
solutions is the main bulk of the paper. The first was segmentation; the use
of segmentation was to detect the PHO and hands in 2D images that are
forms of the hand as it move and the varying of its size from different
6
illumination which is dependent on the intensity changes and color changes.
but implementing this method creates several issues one of which is that
of hand pixels from the image next task is to recognize the gesture, they are
hands and second is for the point and click gesture. For gesture recognition
transformation around the center of the hand and count the number of
circles. Second area of concern is detection of point and click gestures. The
algorithm in the gesture recognition is used and when it detects only one
actual position. The center of the finger is found for each radius and the
values are fitted into a straight line, this line is searched until the final point
is reached.
7
The paper is a research step for gesture recognition. It is implemented
has proven qualitatively that it can be a useful alternate interface for use in
augmented reality. Also it was proven that it is robust enough for the
LCD projector with motorized focus and zoom and a computer controlled
they called ED-lite which functions same as the high-end version and differs
only in the devices used. In the low-end version the group used a portable
8
Several group of professionals were researching and working for a
new method of improving the present HCI. The most common method they
make use of for HCI is by use of mouse, keyboard and touch-screens. But
eliminate the use of such external device that would link the communication
of human and computers. The most popular method under research was
the aid of computer vision, the system was able to steer the projected
9
The particular application used by IBM for demonstration is a slide
screen like function using devices which was mentioned earlier. The ED
computer is used to control the ED unit and performs all other functions
location in the projected image which the user could use to navigate the
slide or to move the content of the projected display from one surface area
to another. The user controls the slide by touching the buttons superimposed
in the specified projected surface area. With this technology the user
interacts with the computer using bare hand and without using such input
Projector designs are now shrinking and are now just in the threshold
of being compact for handheld use. That is why Beardsley and his
10
which can make every surface into displays, a vision to make the world as
its desktop.
The prototype has buttons that serves as the I/O of the device. It also
has a built-in camera to detect the input of the user. Here, it discusses three
display surface for the projected display. Another class creates a projection
first stage is object recognition and the next is to project an overlay that
gives some information about the object. The last class is to project physical
processing. This is similar to a mouse that creates a box to select the region-
There are two main issues when using a handheld device to create
projection and next is the correct aspect ratio. Keystoning occurs when the
11
instead of a square. Keystone correction is used to fix this kind of problem.
Second is the removal of the effects of hand motions. Here it describes the
the markers and to infer the target area in camera image coordinates and
projection data is mapped into these coordinates giving the right placement
of projection.
Examples of applications for each main class given above are also
display goes to the projector and the input is taken form the buttons of the
Region of Interest (ROI) just like in a desktop but without the use of a
mouse.
12
Interest Group on Graphics (SIGGRAPH): Sketches, proposes an interactive
with the projected image. They call this technology the Everywhere Display
help the customers find and give them information about a certain product
and it also tells where the product is located. The ED projector is installed
on the ceiling and it can project images on boards that are hung on every
aisle of the store. At the entrance of the store, there is a table where a larger
on the table and the user can move the wooden red slider to find a product.
The camera detects this motion and the list scrolls up and down copying the
13
2.7 Real-Time Fingertip Tracking and Gesture
Recognition
video projector the users can operate projected desktop application with
system. The application uses the proposed tracking and gesture recognition
methods which assigns different roles to each hand. The gesture recognition
lets users draw objects of different shapes and directly manipulate those
objects using right hand and fingers. Figure 7 shows the set-up used by the
14
group which includes infrared camera, color camera, LCD projector and
Plasma display.
extracting hand regions, finding fingertips and finding palm’s center. In the
defined rather than arm extraction since the searching process in this
15
Figure 2.8. Fingertip Detection
compare it to the previous. Finding the best combination among these two
16
In the evaluation of the tracking method, the group used Linux based
PC with Intel Pentium III 500-MHz and Hitatchi IP5000 image processing
board, and a Nikon Laird-S270 infrared camera. The testing involves seven
image frames. The method reliably tracks multiple fingertips and could
recognition works well with the tracking method and able the user to
such as rotate, move, or resize and other control mode parameters if direct
needed is because the infrared camera didn’t work well on cold hands. A
17
is also planning to extend the system to 3D tracking since the current system
prevent unwanted effects and occlusion detection can be used for hand and
predicted images to the projected scene. This is needed because Hilario and
predicted and observed camera images. Their system is used with a single
18
constant lightning conditions and negligible intra-projector color calibration
camera frames of reference. It will center the projected image and aligned
registration the paper adopted the same approach based on the work of
Second step in the calibration process is the offline color calibration. Due to
colors match exactly those from the source of image. For us to determine
average RGB color over corresponding patch pixels measured over multiple
19