Contour-based Hand Pose Recognition for Sign Language Recognition

Similar documents
TWO HANDED SIGN LANGUAGE RECOGNITION SYSTEM USING IMAGE PROCESSING

Analyzing Well-Formedness of Syllables in Japanese Sign Language

doi: / _59(

Recognition of sign language gestures using neural networks

1. INTRODUCTION. Vision based Multi-feature HGR Algorithms for HCI using ISL Page 1

Cancer Cells Detection using OTSU Threshold Algorithm

Analysis of Recognition System of Japanese Sign Language using 3D Image Sensor

Figure 1: The relation between xyz and HSV. skin color in HSV color space from the extracted skin regions. At each frame, our system tracks the face,

Facial expression recognition with spatiotemporal local descriptors

3. MANUAL ALPHABET RECOGNITION STSTM

Design of Palm Acupuncture Points Indicator

Characterization of 3D Gestural Data on Sign Language by Extraction of Joint Kinematics

International Journal of Software and Web Sciences (IJSWS)

IDENTIFICATION OF REAL TIME HAND GESTURE USING SCALE INVARIANT FEATURE TRANSFORM

This is the accepted version of this article. To be published as : This is the author version published as:

Finger spelling recognition using distinctive features of hand shape

TURKISH SIGN LANGUAGE RECOGNITION USING HIDDEN MARKOV MODEL

Building an Application for Learning the Finger Alphabet of Swiss German Sign Language through Use of the Kinect

Noise-Robust Speech Recognition Technologies in Mobile Environments

Speaking System For Mute

A Lip Reading Application on MS Kinect Camera

Agitation sensor based on Facial Grimacing for improved sedation management in critical care

Implementation of image processing approach to translation of ASL finger-spelling to digital text

Analysis of Emotion Recognition using Facial Expressions, Speech and Multimodal Information

Learning Utility for Behavior Acquisition and Intention Inference of Other Agent

Real Time Sign Language Processing System

A Survey on Hand Gesture Recognition for Indian Sign Language

Sign Language in the Intelligent Sensory Environment

Accuracy and validity of Kinetisense joint measures for cardinal movements, compared to current experimental and clinical gold standards.

Intelligent Frozen Shoulder Self-Home Rehabilitation Monitoring System

An approach for Brazilian Sign Language (BSL) recognition based on facial expression and k-nn classifier

Using Deep Convolutional Networks for Gesture Recognition in American Sign Language

Intelligent Sensor Systems for Healthcare: A Case Study of Pressure Ulcer TITOLO. Prevention and Treatment TESI. Rui (April) Dai

International Journal of Advance Engineering and Research Development. Gesture Glove for American Sign Language Representation

Development of an interactive digital signage based on F-formation system

COMPARATIVE STUDY ON FEATURE EXTRACTION METHOD FOR BREAST CANCER CLASSIFICATION

HAND GESTURE RECOGNITION FOR HUMAN COMPUTER INTERACTION

A Kinematic Assessment of Knee Prosthesis from Fluoroscopy Images

AUTOMATIC DIABETIC RETINOPATHY DETECTION USING GABOR FILTER WITH LOCAL ENTROPY THRESHOLDING

Image processing applications are growing rapidly. Most

Skin color detection for face localization in humanmachine

Estimation of the Upper Limb Lifting Movement Under Varying Weight and Movement Speed

Extraction of Blood Vessels and Recognition of Bifurcation Points in Retinal Fundus Image

A Sleeping Monitor for Snoring Detection

ABSTRACT I. INTRODUCTION

Effect of Sensor Fusion for Recognition of Emotional States Using Voice, Face Image and Thermal Image of Face

PROJECT PLAN. Kinecticons. Version 2.2 May 07, 2013 TEXAS CHRISTIAN UNIVERSITY Computer Science Department, Texas Christian University

Overview Detection epileptic seizures

Finger Braille teaching system-teaching of prosody of finger Braille

AN ALGORITHM FOR EARLY BREAST CANCER DETECTION IN MAMMOGRAMS

MRI Image Processing Operations for Brain Tumor Detection

Facial Expression Biometrics Using Tracker Displacement Features

Research Article Development of a Gastric Cancer Diagnostic Support System with a Pattern Recognition Method Using a Hyperspectral Camera

PROCEEDINGS OF SPIE. Automated microaneurysm detection method based on double ring filter in retinal fundus images

Recognition of Tamil Sign Language Alphabet using Image Processing to aid Deaf-Dumb People

Online Speaker Adaptation of an Acoustic Model using Face Recognition

Sign Language to English (Slate8)

Implementation of Automatic Retina Exudates Segmentation Algorithm for Early Detection with Low Computational Time

EARLY STAGE DIAGNOSIS OF LUNG CANCER USING CT-SCAN IMAGES BASED ON CELLULAR LEARNING AUTOMATE

HCS 7367 Speech Perception

Gesture Recognition using Marathi/Hindi Alphabet

Experimental evaluation of the accuracy of the second generation of Microsoft Kinect system, for using in stroke rehabilitation applications

Posture Monitor. User Manual. Includes setup, guidelines and troubleshooting information for your Posture Monitor App

Koji Sakai. Kyoto Koka Women s University, Ukyo-ku Kyoto, Japan

Face Analysis : Identity vs. Expressions

Segmentation of Cell Membrane and Nucleus by Improving Pix2pix

PAPER REVIEW: HAND GESTURE RECOGNITION METHODS

Artificial Intelligence for Robot-Assisted Treatment of Autism

Performance Analysis of different Classifiers for Chinese Sign Language Recognition

An Approach to Hand Gesture Recognition for Devanagari Sign Language using Image Processing Tool Box

Using Perceptual Grouping for Object Group Selection

Set Your World in Motion. Skeleton measurement

1 Introduction. Fig. 1 Examples of microaneurysms

Automated Blood Vessel Extraction Based on High-Order Local Autocorrelation Features on Retinal Images

International Journal of Engineering Trends and Applications (IJETA) Volume 4 Issue 2, Mar-Apr 2017

Automated Assessment of Diabetic Retinal Image Quality Based on Blood Vessel Detection

Sound Texture Classification Using Statistics from an Auditory Model

CSE 118/218 Final Presentation. Team 2 Dreams and Aspirations

FACIAL EXPRESSION RECOGNITION FROM IMAGE SEQUENCES USING SELF-ORGANIZING MAPS

Detection of suspicious lesion based on Multiresolution Analysis using windowing and adaptive thresholding method.

Human Activities: Handling Uncertainties Using Fuzzy Time Intervals

Improved Intelligent Classification Technique Based On Support Vector Machines

Design and Implementation study of Remote Home Rehabilitation Training Operating System based on Internet

Valence-arousal evaluation using physiological signals in an emotion recall paradigm. CHANEL, Guillaume, ANSARI ASL, Karim, PUN, Thierry.

FEATURE EXTRACTION USING GAZE OF PARTICIPANTS FOR CLASSIFYING GENDER OF PEDESTRIANS IN IMAGES

Automatic Hemorrhage Classification System Based On Svm Classifier

Principals of Object Perception

A Study on the Effect of Inspection Time on Defect Detection in Visual Inspection

Learning Spatiotemporal Gaps between Where We Look and What We Focus on

Introduction to Computational Neuroscience

Automated Estimation of mts Score in Hand Joint X-Ray Image Using Machine Learning

Quality Assessment of Human Hand Posture Recognition System Er. ManjinderKaur M.Tech Scholar GIMET Amritsar, Department of CSE

Quick detection of QRS complexes and R-waves using a wavelet transform and K-means clustering

Feasibility Evaluation of a Novel Ultrasonic Method for Prosthetic Control ECE-492/3 Senior Design Project Fall 2011

Filipino Sign Language Recognition using Manifold Learning

An Avatar-Based Weather Forecast Sign Language System for the Hearing-Impaired

Sign Language Recognition with the Kinect Sensor Based on Conditional Random Fields

Earlier Detection of Cervical Cancer from PAP Smear Images

Biologically-Inspired Human Motion Detection

An Evaluation of RGB-D Skeleton Tracking for Use in Large Vocabulary Complex Gesture Recognition

Transcription:

Contour-based Hand Pose Recognition for Sign Language Recognition Mika Hatano, Shinji Sako, Tadashi Kitamura Graduate School of Engineering, Nagoya Institute of Technology {pia, sako, kitamura}@mmsp.nitech.ac.jp Abstract We are developing a real-time Japanese sign language recognition system that employs abstract hand motions based on three elements familiar to sign language: hand motion, position, and pose. This study considers the method of hand pose recognition using depth images obtained from the Kinect v2 sensor. We apply the contour-based method proposed by Keogh to hand pose recognition. This method s a contour by means of discriminators generated from contours. We conducted experiments on recognizing 23 hand poses from 0 Japanese sign language words. Index Terms: hand pose, contour, sign language recognition, real-time, Kinect 1. Introduction In Japan, Japanese sign language is usually used among hearing impaired people to communicate. In addition, these people often communicate with others through a third person who understands both oral and sign language. The alternative is to use a computer that acts as an interpreter. However, no practical sign language recognition system exists, even one that s isolated words. The difficulties lie in the nature of visual language and its complex structure. Compared with speech recognition, sign language recognition incorporates various visual components, such as hand motions, hand poses and facial expressions. In addition, no established study exists on representing the structure of Japanese sign language in a similar manner to that of spoken language. Therefore, few attempts sign language by units such as hand motions and hand poses [1, 2]. Our study develops with real-time recognition of sign language words. In Japanese sign language, a sentence consists of several words and non-manual signals such as facial expressions. To words is a first step and essential to sentences. The number of Japanese sign language words is said to be 3,000 or more. Recognition by discriminators that are independent of every word has proven ineffective. To produce a practical system, analysis and reconstruction of sign language words are critical. We want to emphasize that database of sign language words is required when we analyze such words. However, no established database currently exists for sign language recognition. Therefore, we employ a database from a computerized sign language word dictionary instead. Our system is based on three elements of sign language: hand motion, position, and pose. This study considers the method of hand pose recognition for our system. Speeding up hand pose recognition is difficult, because of the number and variety of hand poses caused by rotations, altering the angle from the sensor, and diversities in bone structures. This study considers a hand pose recognition using depth images obtained from a single depth sensor. We apply the contour-based method proposed by Keogh [3] to hand pose recognition. This method s a contour by means of discriminators learned from contours. We conducted experiments to 23 hand poses from 0 Japanese sign language words. 2. System overview Figure 1 shows the flowchart of the entire system. We use Kinect v2 sensor [4] to obtain data from sign motions produced by an actual person. First, data obtained from the sensor is segmented into sign language words. Second, the three aforementioned elements are d individually. Finally, the recognition result is determined by the weighted sum of each score. The recognition process of the hand pose and other two components employs depth data of the hand region and coordinates of joints, respectively. This study partially considers the method of hand pose recognition and does not discuss other processes on the flowchart. To utilize the structure in sign language recognition requires an expert knowledge of sign language. We apply a database from the computerized sign language word dictionary produced by Kimura [5] to sign language recognition. Our hand pose recognition is based on the classification of hand types employed in this dictionary. Table 1 shows a portion of the database in the dictionary. This database includes approximately 2,600 Japanese sign language words. Each word is represented by specific sign language types in Table 2 and other elements are indicated in Figure 2. For example, the word red which belongs to the type 1 in Table 2 is expressed by the dominant hand and the other hand is not used. 3. Method of hand pose recognition Some methods of hand pose estimation classify depth pixels into parts to obtain joint coordinates [6, 7]. However, these methods present difficulties when the palm does not the process by coordinates of joints motion Kinect v2 sensor segment sign motion position unify 3 results cut out hand region hand shape process by depth image Figure 1: Flowchart of the entire system.

Table 1: Portion of the database in the dictionary. SL Hand Palm Word Type type direction Position Motion love 3 B down NS circle between 4 B side NS down blue 1 B back lower back red 1 1 back lower right baby 4 B up NS up-down autumn 4 B back whole front-back open 4 B front NS right morning 1 S side upper down shallow 2 B side NS up tomorrow 1 1 front whole front play 4 1 side upper front-back rain 4 5 back NS up-down walk 1 U back NS front relief 4 B back body down say 1 1 side lower front. Table 2: Sign Language (SL) types. 1 2 3 4 5 use both hands hand type is same through two hands non-dominant hand moves Hand type 1 1-b 3 5 5-b 7(S) A B B4-f C F I L L-f R S U V W Y Palm direction Position NS (neutral space) Motion put whole front back side down up upper lower right up-down right-left front-back ear nose body arm circle Figure 2: Elements in sign language dictionary. Depth Image (Original Image) Distance Transform Contour and Center Point camera and some fingers are invisible. We use the contourbased method proposed by Keogh [3]. Contour-based methods work efficiently when recognition objects have distinct shapes. This method treats a contour that encircles an area as a recognition object and uses discriminators called wedges generated from contours. This method is described below. 3.1. Feature extraction Shapes can be converted to distance vectors to form onedimensional series. Figure 3 shows the procedure for extracting a distance vector from a hand image. First, the center point of the hand region is determined by distance transform. Distance transform labels each pixel whose value is 1 with the distance to the nearest pixel whose value is 0 in a binary image. The center point is a pixel that has a maximal value after distance transform. Next, each distance from the center point to every pixel on the contour is calculated. The distance vector represents a series of these distances. 3.2. Calculation of distance A distance D between two distance vectors P = {p 0, p 1,..., p n } and Q = {q 0, q 1,..., q n } is calculated according to the followings. D(P, Q) = n (p i q i ) 2 (1) i=1 If the length of two distance vectors is different, dynamic time warping (DTW) should be used to adjust for size variations. However, we do not use DTW to avoid excessive computation time. Instead, we unify their length by resizing them in advance. We can compare contours by calculating their distances or using discriminators generated from contours. These discriminators are called wedges. Wedges have maximal and minimal values at each point. If a contour is located inside a Distance from center point Figure 3: Feature extraction from an image of a hand region. wedge, the distance is 0. The distance D between a wedge W (U = {u 0, u 1,..., u n } is its top L = {l 0, l 1,..., l n } is its bottom) and a contour P = {p 0, p 1,..., p n } is calculated based on the following equation. For example, the sum of broken lines in Figure 4 is a distance between W and P. n D(W, P ) = i=1 3.3. Producing wedges (p i u i ) 2 (p i > u i ) (p i l i ) 2 (p i < l i ) 0 (otherwise) Wedges are produced according to the following procedures. 1. Extract features from hand images. 2. Calculate distances of all contours. 3. Unify two contours in ascending order of distances. The maximal and minimal values of merged contours become a wedge. 4. Repeat process 3. until the number of wedges decreases to a definite number. P L W Figure 4: Distance between a wedge and contour. U (2)

5 wedges 3 wedges 2 wedges 1 wedge Figure 5: Producing wedges from five contours. When Figure 5 shows an example of producing wedges. A wide wedge produced by contours that are diverse does not function as a discriminator. We prepare various wedges for recognizing each hand type in order to consider the details of contours. 3.4. Speeding up calculation When we consider a rotation invariant matching of two distance vectors, the calculation must be repeated many times with shifting one of the distance vectors. We can speed up this computation by aborting when the current sum of squared differences exceeds a threshold. In addition, although existing research does not attempt this, we try to speed up the calculation by means of the followings. The length of the distance vectors is unified and shortened, and the accuracy does not diminish. When the number of wedges per hand type is greater than one, recognition that uses one-by-one wedge is performed prior to help targeting candidates. 4.1. Datasets 4. Experiments We conducted experiments on recognizing 23 hand poses in 0 Japanese sign language words in the national sign language test grade 5. To these 0 words requires to distinguish 23 hand poses in Table 3 defined by hand types and palm directions. Some words have the same hand poses but different position and motion. Our system distinguish each word after recognizing 3 components and unifying recognition results. Because hand shapes transform with motions, each hand type remains independent even if the palm direction is different. However, some exceptions exist to distinguish sign language words that have the same motion, position, and hand type, but have a different palm direction. For example, Groups 3 and 4 in Table 3 should be distinguished even though the hand type is the same. To simplify the collection of data in our experiments, we used depth images of stationary hands instead of those obtained during natural sign motions. Table 4 shows the experimental conditions. We conducted four experiments examining the robustness of the recognition method about the variety of hand shapes and the computation time. The objectives of the experiments are described as follows. Experiment 1 Recognize 100 hand images by wedges produced from the same 100 images per hand type, palm direction, and tester (close-dataset, close-tester). Experiment 2 Recognize 50 hand images by wedges produced from the other 50 hand images per hand type, palm direc- Table 3: List of 23 hand pose groups. ID Hand type Palm direction 0 1 front-back, right-left 1 1-b right-left 2 3 front-back 3 5 front-back 4 5 up-down 5 5-b front-back, right-left, up-down 6 7(S) front-back 7 A front-back, right-left 8 B front-back 9 B right-left 10 B up-down 11 B4-f right-left 12 C right-left 13 F front-back 14 I front-back 15 L front-back 16 L-f right-left 17 R right-left 18 S front-back, right-left, up-down 19 U front-back 20 V front-back 21 W front-back 22 Y front-back Table 4: Experimental condition. Hand type 20 types in Figure 2 Palm direction 3 patterns (front-back, right-left, up-down) Hand pose group 23 groups* *determined by hand types and palm directions Tester s profile A (female, hand size* 16 cm) B (female, hand size* 18 cm) C (male, hand size* 19 cm) D (male, hand size* 21 cm) *measured from the wrist to the tip of the middle finger Depth image 100 100 pixel 100 images of the hand region per tester, hand type and palm direction Length of distance vector 30 or 180 PC specs OS Windows 8.1 64 bit RAM: 4 GB CPU Intel Core i5-4570 (3.20 GHz, 4-core) tion, and tester (open-dataset, close-tester). Experiments were repeated with different data. Experiment 3 Recognize 100 hand images of a person by wedges produced from 300 hand images of the other three persons per hand type, and palm direction (opendataset, open-tester). Experiments ware repeated with different data. Experiment 4 Examine the relationship between the computation time required to a hand image and the average recognition rate from Experiment 2. We attempted to speed up the calculation by the methods in Section 3.4. The threshold value when the calculation was aborted was determined by the preliminary experiment. The length of distance vectors was 30 in this experiment. Each recognition was aided to target candidates as many as five hand pose groups by the recognition that uses one-by-one wedge performed prior. 4.2. Results 4.2.1. Experiment 1, 2 Figure 6 shows the average recognition rates for Experiment 1 and 2. The accuracy can be improved by increasing the number of wedges. This can be accomplished because of the variety of hand shapes caused by posing of hand and by altering the angle from the camera.

100 50 Recognition rate (%) 80 60 length of distance vector: 180 Experiment 1{ length of distance vector: 30 length of distance vector: 180 Experiment 2{ length of distance vector: 30 1 5 10 Number of wedges Recognition rate (%) 30 20 10 Test data A B C D 1 10 20 30 Number of wedges Figure 6: Experiment 1, 2: Recognition rate and number of wedges per hand type and palm direction (person is known). Experiment 1 was conducted with close-dataset. This experiment is just for sanity check and its condition is impossible in real-life. The result was sufficient for our system. Erroneous recognition in Experiment 1 was primarily caused by misidentifying hand pose Groups 4 and 5. These two groups have a common point that includes a hand pose whose palm direction is down. When we obtain data from a single depth camera, capturing the characteristics of hand shapes when the palm does not the camera is difficult. Group 6 had the lowest recognition rate among the hand pose group (when the length of the distance vector is 180, the number of wedges was 10 per hand type and palm direction, and the group s recognition rate was 80%). This is because the group was misd as Group 0. These two groups have similar shapes. In addition, the recognition rates of Group 2, 13, 15, 20, and 22 were high under all conditions because other groups do not possess similar shapes. Experiment 2 was conducted with open-dataset and closetester. The result showed a similar trend to that of Experiment 1 concerning the causes of erroneous recognition. Because no hand shapes from the learning data are included in the evaluation data, the recognition rate was lower than that of Experiment 1. However, no significant difference in recognition rate of Experiment 1 and 2 appeared when the number of wedges is one per hand type and palm direction. Therefore, if the wedges are generated from samples of a certain number, applying unknown data from the same person is possible. The recognition rate from Experiment 2 is expected to approach that of Experiment 1 by increasing the amount of learning data. Experiments were conducted after changing the length of distance vectors. Although shortening the distance vectors reduces the calculations, the accuracy is expected to fall because of the loss of detailed features. However, no significant differences between the experiments appeared when the length of the distance vectors is 30 and 180. Therefore, if small sized hand images are used or the contours are rough because of noises, a robust recognition can be accomplished. The maximal number of wedges was between 20 and 25 in Experiment 1 and between 8 and 13 in Experiment 2. The number fluctuated with the complexity of the hand types. 4.2.2. Experiment 3 Experiment 3 was tester-independent setup. Figure 7 shows the results of Experiment 3. The recognition rates shown are the results when the length of distance vectors is 30. If we change the length to 180, recognition rates do not change significantly. We specified causes of erroneous recognition when the number of wedges is 30 per hand type and palm direction. The results Figure 7: Experiment 3: Recognition rate and number of wedges per hand type and palm direction (person is unknown). Tester A B C D Hand types (palm direction) 1(back) F(front) 5(down) 5-b(left) B(left) Figure 8: Variety of hand shapes among people. show the same tendency as in Experiments 1 and 2, that is, 13 % of all data were misd as Groups 4 and 5. The detailed findings for each hand pose group reveal the following: 41 % of Group 6 were misd as Group 0, 53 % of Group 19 were misd as Group 0, 45 % of Group 12 were misd as Group 5. The low recognition rate is due to individual differences in hand shapes caused by differences in bone structure and posing of hand shown in Figure 8. Wedges produced from the hand images of various people include other hand types. This caused misrecognitions. Per person details show that the recognition rate was lowest when the system attempted to hand poses of tester A, whose hand size was the smallest. When the number of wedges increases, the recognition rate of tester B, whose hand size is between that of A and C is higher than that of other testers. Although we normalized the scale of distance vectors according to each hand size, hand pose recognition by contours possesses other difficulties when the bone structures are considered. The accuracy diminishes when the system s hand images of a person whose bone structure is dissimilar to any learning data. When we want to hand poses of an unknown person, wedges generated from people who have similar bone structure should be used. Therefore, additional hand images that reveal various characteristics in bone structures should be collected. 4.2.3. Experiment 4 Experiment 4 was for checking the computation time. Figure 9 shows the relationship between the computation time required to a hand image and the average recognition rate in Experiment 2. The speed-up process did not affect the recognition rate.

Recognition rate (%) 70 60 50 0 20 60 80 100 Computation time (ms) Figure 9: Experiment 4: Average computation time and the recognition rate required to a hand image. When the person is known, 88 ms (corresponding to 11 fps) was required to a hand image with 70 % accuracy. Recognizing all hand images obtained from the sensor with a frame rate of 30 fps is impossible. However, the number of frames required to specify a hand pose is limited because the hand pose does not change at every frame. We can in real-time selected hand images by means of comparison method employing a small calculation such as image moment [8]. This experiment has been implemented in a single-thread. The processing speed can be improved by utilizing a high-speed technique such as multi-threading. [2] Arata Sato and Koichi Shinoda, Large Vocabrary Sign Language Recognition Based on Cheremes, in IEICE Technical Report PRMU2011-222, SP2011-137, 2012, pp. 155 160. [3] Eamonn Keogh, Li Wei, Xiaopeng Xi, Sang-Hee Lee and Michail Vlachos, LB Keogh Supports Exact Indexing of Shapes under Rotation Invariance with Arbitrary Representations and Distance Measures, in 32nd International Conference on Very Large Data Bases (VLDB2006), 2006, pp. 882 893. [4] Kinect for Windows, http://kinectforwindows.org. [5] Tsutomu Kimura, Daisuke Hara, Kazuyuki Kanda and Kazunari Morimoto, Expansion of the System of JSL-Japanese Electronic Dictionary: An Evaluation for the Compound Research System, in Proceedings of the 2nd International Conference on Human Centered Design, ser. HCD 11, 2011, pp. 7 416. [6] Hui Liang, Junsong Yuan and Daniel Thalmann, Parsing the Hand in Depth Images, Multimedia, IEEE Transactions on, vol. 16, no. 5, pp. 1241 1253, Aug 2014. [7] Danhang Tang, Tsz-Ho Yu and Tae-Kyun Kim, Real-Time Articulated Hand Pose Estimation Using Semi-supervised Transductive Regression Forests, in Proceedings of the 2013 IEEE International Conference on Computer Vision, ser. ICCV 13, 2013, pp. 3224 3231. [8] Ming-Kuei Hu, Visual Pattern Recognition by Moment Invariants, Information Theory, IRE Transactions on, vol. 8, no. 2, pp. 179 187, February 1962. 5. Conclusion We are developing a real-time Japanese sign language recognition system based on three elements of sign language: motion, position, and pose. This study examined hand pose recognition by means of contour-based method proposed by Keogh using depth images obtained from a single depth sensor. We conducted experiments on recognizing 23 hand poses from 0 Japanese sign language words. Under the condition of close-tester, the recognition rate was approximately 90 % for close-dataset, 70 % for open-dataset. In addition, we conducted an experiment to the hand poses of an unknown person by means of discriminators learned from hand poses of other people. The recognition rate dropped considerably because diversities in bone structure of each person s hand generated loose discriminators that are unable to consider the details of contours. We also evaluated the computation time. Regarding close-tester and open-dataset, 88 ms (corresponding to 11 fps) was required to a hand image with 70 % accuracy. When we the hand poses of an unknown person, discriminators generated from people who have similar bone structure should be used. Future research in this area requires that hand images of various people be collected and applied for the purpose of recognizing unknown persons. 6. Acknowledgement This research was supported in part by Japan Society for the Promotion of Science KAKENHI (No. 25350666), and Toukai Foundation for Technology. 7. References [1] Rung-Huei Liang and Ming Ouhyoung, A Real-time Continuous Gesture Recognition System for Sign Language, in Automatic Face and Gesture Recognition, 1998. Proceedings. Third IEEE International Conference on, Apr 1998, pp. 558 567.