New Project Report
New Project Report
CHAPTER 1
INTRODUCTION
Page 1
Application of Deep Learning for sign Language Gesture Recognition with Efficient Hand Gesture Representation
Speech impaired people use hand signs and gestures to communicate. Normal people face
difficulty in understanding their language. Hence there is a need of a system which recognizes
the different signs, gestures and conveys the information to the normal people. It bridges the
gap between physically challenged people and normal people.
Page 2
Application of Deep Learning for sign Language Gesture Recognition with Efficient Hand Gesture Representation
image into a corresponding digital image and the extraction of significant information from the
digital image by applying various algorithms.
any system with these capabilities thus far. during this project our aim is to develop a system
which may classify signing accurately.
American Sign Language (ASL) is a complete, natural language that has the same linguistic
properties as spoken languages, with grammar that differs from English. ASL is expressed by
movements of the hands and face. It is the primary language of many North Americans who are
deaf and hard of hearing, and is used by many hearing people as well.
Page 4
Application of Deep Learning for sign Language Gesture Recognition with Efficient Hand Gesture Representation
1.4 Motivation
The 2011 Indian census cites roughly 1.3 million people with “hearingimpairment”. In contrast
to that numbers from India’s National Association of the Deaf estimates that 18 million people
–roughly 1 per cent of Indian population are deaf. These statistics formed the motivation for
our project. As these speech impairment and deaf people need a proper channel to
communicate with normal people there is a need for a system . Not all normal people can
understand sign language of impaired people. Our project hence is aimed at converting the sign
language gestures into text that is readable for normal people.
Page 5
Application of Deep Learning for sign Language Gesture Recognition with Efficient Hand Gesture Representation
CHAPTER 2
LITERATURE SURVEY
Page 6
Application of Deep Learning for sign Language Gesture Recognition with Efficient Hand Gesture Representation
In Literature survey we have gone through other similar works that are implemented in the
domain of sign language recognition.The summaries of each of the project works are mentioned
below
Sign Language Recognition (SLR) system, which is required to recognize sign
languages, has been widely studied for years.The studies are based on various input sensors,
gesture segmentation, extraction of features and classifcation methods.This paper aims to
analyze and compare the methods employed in the SLR systems, classi cations methods that
have been used, and suggests the most promising method for future research. Due to recent
advancement in classifcationmethods, many of the recent proposed works mainly contribute on
the classifcation methods, such as hybrid method and Deep Learning. This paper focuses on the
classifcation methodsused in prior Sign Language Recognition system. Based on our
review, HMM- based approaches have been explored extensively in prior research, including
its modifcations.
This study is based on various input sensors, gesture segmentation, extraction of
features and classification methods. This paper aims to analyze and compare the methods
employed in the SLR systems, classifications methods that have been used, and suggests the
most reliable method for future research. Due to recent advancement in classification methods,
many of the recently proposed works mainly contribute to the classification methods, such as
hybrid method and Deep Learning. Based on our review, HMM-based approaches have been
explored extensively in prior research, including its modifications.Hybrid CNN-HMM and
fully Deep Learning approaches have shown promising results and offer opportunities for
further exploration.
Chat applications have become a powerful mediathat assist people to communicate in different
languages witheach other. There are lots of chat applications that are useddifferent people in
different languages but there are not such achat application that has facilitate to communicate
with signlanguages. The developed system isbased on Sinhala Sign language. The system has
included fourmain components as text messages are converted to sign messages, voice
messages are converted to sign messages, signmessages are converted to text messages and
sign messages areconverted to voice messages. Google voice recognition API hasused to
develop speech character recognition for voice messages.The system has been trained for the
speech and text patterns by usingsome text parameters and signs of Sinhala Sign language
isdisplayed by emoji. Those emoji and signs that are included inthis system will bring the
normal people more close to the disabled people. This is a 2 way communication system but it
uses pattern of gesture recognition which is not very realiable in getting appropriate output.
In this paper we proposed some methods,through which the recognition of the signs becomes
easy forpeoples while communication. And the result of thosesymbols signs will be converted
Page 7
Application of Deep Learning for sign Language Gesture Recognition with Efficient Hand Gesture Representation
into the text. In this project,we are capturing hand gestures through webcam andconvert this
image into gray scale image. The segmentationof gray scale image of a hand gesture is
performed usingOtsu thresholdingalgorithm.. Total image level is dividedinto two classes one
is hand and other is background. Theoptimal threshold value is determined by computing
theratio between class variance and total class variance. Tofind the boundary of hand gesture in
image Canny edgedetection technique is used.In Canny edge detection we used edge based
segmentation and threshold based segmentation.Then Otsu’s algorithm is used because of its
simple calculation and stability.This algorithm fails, when the global distribution of the target
and background vary widely.
Computer recognition of sign language is an important research problem for enabling
communication with hearing impaired people. This project introduces an efficient and fast
algorithm for identification of the number of fingers opened in a gesture representing an
alphabet of the Binary Sign Language. The system does not require the hand to be perfectly
aligned to the camera. The project uses image processing system to identify, especially English
alphabetic sign language used by the deaf people to communicate. The basic objective of this
project is to develop a computer based intelligent system that will enable dumb people
significantly to communicate with all other people using their natural hand gestures. The idea
consisted of designing and building up an intelligent system using image processing, machine
learning and artificial intelligence concepts to take visual inputs of sign language’s hand
gestures and generate easily recognizable form of outputs. Hence the objective of this project is
to develop an intelligent system which can act as a translator between the sign language and
the spoken language dynamically and can make the communication between people with
hearing impairment and normal people both effective and efficient. The system is we are
implementing for Binary sign language but it can detect any sign language with prior image
processing
One of the major drawback of our society is the barrier that is created between disabled
or handicapped persons and the normal person. Communication is the only medium by which
we can share our thoughts or convey the message but for a person with disability (deaf and
dumb) faces difficulty in communication with normal person. For many deaf and dumb
people , sign language is the basic means of communication. Sign language recognition (SLR)
aims to interpret sign languages automatically by a computer in order to help the deaf
communicate with hearing society conveniently. Our aim is to design a system to help the
person who trained the hearing impaired to communicate with the rest of the world using sign
language or hand gesture recognition techniques. In this system, feature detection and feature
extraction of handgesture is done with the help of SURF algorithm using image processing.
All this work is done using MATLAB software. With the help of this algorithm, a person can
easily trained a deaf and dumb.
Page 8
Application of Deep Learning for sign Language Gesture Recognition with Efficient Hand Gesture Representation
Speech impairment is a disability which affects one’s ability to speak and hear. Such
individuals use sign language to communicate with other people. Although it is an effective
form of communication, there remains a challenge for people who do not understand sign
language to communicate with speech impaired people. The aim of this paper is to develop an
application which will translate sign language to English in the form of text and audio, thus
aiding communication with sign language. The application acquires image data using the
webcam of the computer, then it is preprocessed using a combinational algorithm and
recognition is done using template matching. The translation in the form of text is then
converted to audio. The database used for this system includes 6000 images of English
alphabets. We used 4800 images for training and 1200 images for testing. The system produces
88% accuracy.
This research work presents a prototype system that helps to recognize hand gesture to
normal people in order to communicate more effectively with the special people. Aforesaid
research work focuses on the problem of gesture recognition in real time that sign language
used by the community of deaf people. The problem addressed is based on Digital Image
Processing using Color Segmentation, Skin Detection, Image Segmentation, Image Filtering,
and Template Matching techniques. This system recognizes gestures of ASL (American Sign
Language) including the alphabet and a subset of its words.
2.1 Libraries
2.1.1 TensorFlow:
TensorFlow is a free and open-source software library for dataflow and differentiable
programming across a range of tasks. It is a symbolic math library, and is also used for
machine learning applications such as neural networks. It is used for both research and
production at Google.
Features: TensorFlow provides stable Python (for version 3.7 across all platforms) and C APIs;
and without API backwards compatibility guarantee: C++, Go, Java, JavaScript and Swift
(early release). Third-party packages are available for C#, Haskell Julia, MATLAB,R, Scala,
Rust, OCaml, and Crystal."New language support should be built on top of the C API.
However, not all functionality is available in C yet." Some more functionality is provided by
the Python API. Application: Among the applications for which TensorFlow is the foundation,
are automated image-captioning software, suchas DeepDream.
2.1.2 Opencv:
OpenCV (Open Source Computer Vision Library) is a library of programming
Page 9
Application of Deep Learning for sign Language Gesture Recognition with Efficient Hand Gesture Representation
functions mainly aimed at real-time computer vision.[1] Originally developed by Intel, it was
later supported by Willow Garage then Itseez (which was later acquired by Intel[2]). The
library is cross-platform and free for use under the open-source BSD license. OpenCV's
application areas include:
2D and 3D feature toolkits
Egomotion estimation
Facial recognition system
Gesture recognition
Human–computer interaction (HCI)
Mobile robotics
Motion understanding
Object identification
Segmentation and recognition
Stereopsis stereo vision: depth perception from 2 cameras
Structure from motion (SFM).
Motion tracking
Augmented reality
To support some of the above areas, OpenCV includes a statistical machine learning library
that contains:
Boosting
Decision tree learning
Gradient boosting trees
Expectation-maximization algorithm
k-nearest neighbor algorithm
Naive Bayes classifier
Artificial neural networks
Random forest
Support vector machine (SVM)
Deep neural networks (DNN)
AForge.NET, a computer vision library for the Common Language Runtime (.NET
Framework and Mono).
ROS (Robot Operating System). OpenCV is used as the primary vision package in ROS.
CVIPtools, a complete GUI-based computer-vision and image-processing software
environment, with C function libraries, a COM-based DLL, along with two utility programs for
algorithm development and batch processing.
Page 10
Application of Deep Learning for sign Language Gesture Recognition with Efficient Hand Gesture Representation
2.1.3 Keras:
Keras is an open-source neural-network library written in Python. It is capable of
running on top of TensorFlow, Microsoft Cognitive Toolkit, R, Theano, or PlaidML. Designed
to enable fast experimentation with deep neural networks, it focuses on being user-friendly,
modular, and extensible. It was developed as part of the research effort of project ONEIROS
(Open-ended Neuro-Electronic Intelligent Robot Operating System), and its primary author
and maintainer is François Chollet, a Google engineer. Chollet also is the author of the
XCeption deep neural network model. Features: Keras contains numerous implementations of
commonly used neural- network building blocks such as layers, objectives, activation
functions, optimizers, anda host of tools to make working with image and text data easier to
simplify the coding necessary for writing deep neural network code. The code is hosted on
GitHub, and community support forums include the GitHub issues page, and a Slack channel.
In addition to standard neural networks, Keras has support for convolutional and recurrent
neural networks. It supports other common utility layers like dropout, batch normalization, and
pooling.
Keras allows users to productize deep models on smartphones (iOS and Android), on the web,
or on the Java Virtual Machine. It also allows use of distributed training of deep-learning
models on clusters of Graphics processing units (GPU) and tensor processing units (TPU)
principally in conjunction with CUDA.
Keras applications module is used to provide pre-trained model for deep neural networks.
Keras models are used for prediction, feature extraction and fine tuning.
Pre-trained models
Trained model consists of two parts model Architecture and model Weights. Model
weights are large file so we have to download and extract the feature from ImageNet database.
Some of the popular pre-trained models are listed below,
ResNet
VGG16
MobileNet
InceptionResNetV2
InceptionV3
2.1.4 Numpy:
NumPy (pronounced /ˈnʌmpaɪ/ (NUM-py) or sometimes /ˈnʌmpi/ (NUM-pee)) is a
Page 11
Application of Deep Learning for sign Language Gesture Recognition with Efficient Hand Gesture Representation
library for the Python programming language, adding support for large, multi- dimensional
arrays and matrices, along with a large collection of high-level mathematical functions to
operate on these arrays. The ancestor of NumPy, Numeric, was originally created by Jim
Hugunin with contributions from several other developers. In 2005, Travis Oliphant created
NumPy by incorporating features of the competing Numarray into Numeric, with extensive
modifications. NumPy is open- source software and has many contributors. Features: NumPy
targets the CPython reference implementation of Python, which is a non-optimizing bytecode
interpreter. Mathematical algorithms written for this version of Python often run much slower
than compiled equivalents. NumPy addresses the slowness problem partly by providing
multidimensional arrays and functions and operators that operate efficiently on arrays,
requiring rewriting some code, mostly inner loops using NumPy. Using NumPy in Python
gives functionality comparable to MATLAB since they are both interpreted,and they both
allow the user to write fast programs as long as most operations work on arrays or matrices
instead of scalars. In comparison, MATLAB boasts a large number of additional toolboxes,
notably Simulink, whereas NumPy is intrinsically integrated with Python, a more modern and
complete programming language. Moreover, complementary Python packages are available;
SciPy is a library that adds more MATLAB-like functionality and Matplotlib is aplotting
package that providesMATLAB-like plotting functionality. Internally, both MATLAB and
NumPy rely on BLAS and LAPACK for efficient linear algebra computations. Python bindings
of the widely used computer vision library OpenCV utilize NumPy arrays to store and operate
on data. Since images with multiple channels are simply represented as three-dimensional
arrays, indexing, slicing or masking with other arrays are very efficient ways to access specific
pixels of an image. The NumPy array as universal data structure in OpenCV for images,
extracted feature points, filter kernels and many more vastly simplifies the programming
workflow and debugging. Limitations: Inserting or appending entries to an array is not as
trivially possible as it is with Python's lists. The np.pad(...) routine to extend arrays actually
creates new arrays of the desired shape and padding values, copies the given array into the new
one and returns it. NumPy'snp.concatenate([a1,a2]) operation does not actually link the two
arrays but returns a new one, filled with the entries from both given arrays in sequence.
Reshaping the dimensionality of an array with np.reshape(...) is only possible as long as the
number of elements in the array does not change. These circumstances originate from the fact
that NumPy's arrays must be views on contiguous memory buffers. A replacement package
called Blaze attempts to overcome this limitation.
Algorithms that are not expressible as a vectorized operation will typically run slowly because
they must be implemented in "pure Python", while vectorization may increase memory
complexity of some operations from constant to linear, because temporary arrays must be
created that are as large as the inputs. Runtime compilation of numerical code has been
Page 12
Application of Deep Learning for sign Language Gesture Recognition with Efficient Hand Gesture Representation
implemented by several groups to avoid these problems; open source solutions that
interoperate with NumPy include scipy.weave, numexpr and Numba. Cython and Pythran are
static-compiling alternatives to these.
Page 13
Application of Deep Learning for sign Language Gesture Recognition with Efficient Hand Gesture Representation
Above all, these neural nets are capable of discovering latent structures within unlabeled,
unstructured data, which is the vast majority of data in the world. Another word for
unstructured data is raw media; i.e. pictures, texts, video and audio recordings. Therefore, one
of the problems deep learning solves best is in processing and clustering the world’s raw,
unlabeled media, discerning similarities and anomalies in data that no human has organized in
a relational database or ever put a name to.
For example, deep learning can take a million images, and cluster them according to their
similarities: cats in one corner, ice breakers in another, and in a third all the photos of your
grandmother. This is the basis of so-called smart photo albums.
Deep-learning networks perform automatic feature extraction without human intervention,
unlike most traditional machine-learning algorithms. Given that feature extraction is a task that
can take teams of data scientists years to accomplish, deep learning is a way to circumvent the
chokepoint of limited experts. It augments the powers of small data science teams, which by
their nature do not scale. When training on unlabeled data, each node layer in a deep network
learns features automatically by repeatedly trying to reconstruct the input from which it draws
its samples, attempting to minimize the difference between the network’s guesses and the
probability distribution of the input data itself. Restricted Boltzmann machines, for examples,
create so-called reconstructions in this manner. In the process, these neural networks learn to
recognize correlations between certain relevant features and optimal results – they draw
connections between feature signals and what those features represent, whether it be a full
reconstruction, or with labeled data. A deep-learning network trained on labeled data can then
be applied to unstructured data, giving it access to much more input than machine-learning
nets.
Page 14
Application of Deep Learning for sign Language Gesture Recognition with Efficient Hand Gesture Representation
computer, these characteristics are boundaries or curvatures. And then through the groups of
convolutional layers the computer constructs more abstract concepts.In more detail: the image
is passed through a series of convolutional, nonlinear, pooling layers and fully connected
layers, and then generates the output.
Comparing all the collected data with already existing data in the database to match a face
with a name. A similar process is followed for scene labeling as well. Analyzing Documents.
Convolutional neural networks can also be used for document analysis. This is not just useful
for handwriting analysis, but also has a major stake in recognizers. For a machine to be able to
scan an individual's writing, and then compare that to the wide database it has, it must execute
almost a million commands a minute. It is said with the use of CNNs and newer models and
algorithms, the error rate has been brought down to a minimum of 0.4% at a character level,
though it's complete testing is yet to be widely seen.
Page 15
Application of Deep Learning for sign Language Gesture Recognition with Efficient Hand Gesture Representation
CHAPTER 3
METHODOLOGY
Page 16
Application of Deep Learning for sign Language Gesture Recognition with Efficient Hand Gesture Representation
Page 17
Application of Deep Learning for sign Language Gesture Recognition with Efficient Hand Gesture Representation
Page 18
Application of Deep Learning for sign Language Gesture Recognition with Efficient Hand Gesture Representation
4. model evaluation
Model construction: It depends on machine learning algorithms. In this projectscase, it was
neural networks.Such an agorithm looks like:
1. begin with its object: model = Sequential()
2. then consist of layers with their types: model.add(type_of_layer())
3. after adding a sufficient number of layers the model is compiled. At this moment Keras
communicates with TensorFlow for construction of the model. During model compilation it is
important to write a loss function and an optimizer algorithm. Before model training it is
important to scale data for their further use.
Model training:
After model construction it is time for model training. In this phase, the model is trained using training
data and expected output for this data. It’s look this way: model.fit(training_data, expected_output).
Progress is visible on the console when the script runs. At the end it will report the final accuracy of the
model.
Model Testing:
During this phase a second set of data is loaded. This data set has never been seen by the model and
therefore it’s true accuracy will be verified. After the model training is complete, and it is understood that
the model shows the right result, it can be saved by: model.save(“name_of_file.h5”). Finally, the saved
model can be used in the real world. The name of this phase is model evaluation. This means that the
model can be used to evaluate new data.
3.2.2 Preprocessing:
Uniform aspect ratio
Understanding aspect ratios:
An aspect ratio is a proportional relationship between an image's width and height. Essentially,
it describes an image's shape.Aspect ratios are written as a formula of width to height, like
this: For example, a square image has an aspect ratio of 1:1, since the height and width are the
same. The image could be 500px × 500px, or 1500px × 1500px, and the aspect ratio would still
be 1:1.As another example, a portrait-style image might have a ratio of 2:3. With this aspect
ratio, the height is 1.5 times longer than the width. So the image could be 500px × 750px,
1500px × 2250px, etc.
Page 19
Application of Deep Learning for sign Language Gesture Recognition with Efficient Hand Gesture Representation
Page 20
Application of Deep Learning for sign Language Gesture Recognition with Efficient Hand Gesture Representation
Page 21
Application of Deep Learning for sign Language Gesture Recognition with Efficient Hand Gesture Representation
3.3.2 Optimizer(Adam):
Adam can be looked at as a combination ofRMSprop and Stochastic Gradient Descent with
momentum. It uses the squared gradients to scale the learning rate like RMSprop and it takes
advantage of momentum by using moving average of the gradient instead of gradient itself
like SGD with momentum. Adam is an adaptive learning rate method, which means, it
computes individual learning rates for different parameters. Its name is derived from adaptive
moment estimation, and the reason it’s called that is because Adam uses estimations of first
and second moments of gradient to adapt the learning rate for each weight of the neural
network. Now, what is moment ? N-th moment of a random variable is defined as the expected
value of that variable to the power of n. More formally
3.4 Segmentation
Image segmentation is the process of partitioning a digital image into multiple
segments(sets of pixels, also known as image objects). The goal of segmentation is to simplify
and/or change the representation of an image into something that is more meaningful and easier
to analyze. Modern image segmentation techniques are powered by deep learning technology.
Here are several deep learning architectures used for segmentation: If we take an example of
Autonomous Vehicles, they need sensory input devices like cameras, radar, and lasers to allow
the car to perceive the world around it, creating a digital map. Autonomous driving is not even
possible without object detection which itself involves image classification/segmentation.
How Image Segmentation works
Image Segmentation involves converting an image into a collection of regions of pixels that
are represented by a mask or a labeled image. By dividing an image into segments, you can
process only the important segments of the image instead of processing the entire image. A
common technique is to look for abrupt discontinuities in pixel values, which typically indicate
edges that define a region.Another common approach is to detect similarities in the regions of
an image. Some techniques that follow this approach are region growing, clustering, and
Page 22
Application of Deep Learning for sign Language Gesture Recognition with Efficient Hand Gesture Representation
Page 23
Application of Deep Learning for sign Language Gesture Recognition with Efficient Hand Gesture Representation
Equation…..(3.1) …….
Here are the three elements that enter into the convolution operation:
• Input image
• Feature detector
• Feature map
Steps to apply convolution layer:
• You place it over the input image beginning from the top-left corner within the
borders you see demarcated above, and then you count the number of cells in which the
feature detector matches the input image.
• The number of matching cells is then inserted in the top-left cell of the feature map
• You then move the feature detector one cell to the right and do the same thing. This
movement is called a and since we are moving the feature detector one cell at time, that would
be called a stride of one pixel.
• What you will find in this example is that the feature detector's middle-left cell with
the number 1 inside it matches the cell that it is standing over inside the input image. That's the
only matching cell, and so you write “1” in the next cell in the feature map, and so on and so
forth.
• After you have gone through the whole first row, you can then move it over to the
next row and go through the same process.
There are several uses that we gain from deriving a feature map. These are the most important
of them: Reducing the size of the input image, and you should know that the larger your strides
(the movements across pixels), the smaller your feature map.
Page 24
Application of Deep Learning for sign Language Gesture Recognition with Efficient Hand Gesture Representation
remove all the black elements from it, keeping only those carrying a positive value (the grey
and white colors).The essential difference between the non-rectified version of the image and
the rectified one is the progression of colors. After we rectify the image, you will find the
colors changing more abruptly. The gradual change is no longer there. That indicates that the
linearity has been disposed of.
Page 26
Application of Deep Learning for sign Language Gesture Recognition with Efficient Hand Gesture Representation
CHAPTER 4
DESIGN
Page 27
Application of Deep Learning for sign Language Gesture Recognition with Efficient Hand Gesture Representation
Page 28
Application of Deep Learning for sign Language Gesture Recognition with Efficient Hand Gesture Representation
synchronous message, it must wait until the message is done, such as invoking a subroutine. If
a caller sends an asynchronous message, it can continue processing and doesn’t have to wait
for a response. Asynchronous calls are present in multithreaded applications, event-driven
applications and in message-oriented middleware. Activation boxes, or method-call boxes, are
opaque rectangles drawn on top of lifelines to represent that processes are being performed in
response to the message (ExecutionSpecifications in UML).
Objects calling methods on themselves use messages and add new activation boxes on top of
any others to indicate a further level of processing. If an object is destroyed (removed from
memory), an X is drawn on bottom of the lifeline, and the dashed line ceases to be drawn
below it. It should be the result of a message, either from the object itself, or another. A
message sent from outside the diagram can be represented by a message originating from a
filled-in circle (found message in UML) or from a border of the sequence diagram (gate in
UML).
UML has introduced significant improvements to the capabilities of sequence diagrams. Most
of these improvements are based on the idea of interaction fragmentswhich represent smaller
pieces of an enclosing interaction. Multiple interaction fragments are combined to create a
variety of combined fragments, which are then used to model interactions that include
parallelism, conditional branches, optional interactions
Page 30
Application of Deep Learning for sign Language Gesture Recognition with Efficient Hand Gesture Representation
Page 31
Application of Deep Learning for sign Language Gesture Recognition with Efficient Hand Gesture Representation
CHAPTER 5
IMPLEMENTATION AND RESULTS
Page 32
Application of Deep Learning for sign Language Gesture Recognition with Efficient Hand Gesture Representation
5.1 Screenshot
Page 33
Application of Deep Learning for sign Language Gesture Recognition with Efficient Hand Gesture Representation
Page 34
Application of Deep Learning for sign Language Gesture Recognition with Efficient Hand Gesture Representation
Page 35
Application of Deep Learning for sign Language Gesture Recognition with Efficient Hand Gesture Representation
Page 36
Application of Deep Learning for sign Language Gesture Recognition with Efficient Hand Gesture Representation
Page 37
Application of Deep Learning for sign Language Gesture Recognition with Efficient Hand Gesture Representation
Page 38
Application of Deep Learning for sign Language Gesture Recognition with Efficient Hand Gesture Representation
Page 39
Application of Deep Learning for sign Language Gesture Recognition with Efficient Hand Gesture Representation
Page 40
Application of Deep Learning for sign Language Gesture Recognition with Efficient Hand Gesture Representation
Page 41
Application of Deep Learning for sign Language Gesture Recognition with Efficient Hand Gesture Representation
CHAPTER 6
COMPARISON BETWEEN PREVIOUS SYSTEM AND PROPOSED
SYSTEM
Page 42
Application of Deep Learning for sign Language Gesture Recognition with Efficient Hand Gesture Representation
Point assembling 59 81
Data retrieving 56 83
Trained
65 91
parameters
Model creation 55 85
Page 43
Application of Deep Learning for sign Language Gesture Recognition with Efficient Hand Gesture Representation
160
140
120
100
80
60
40 Proposed (Accuracy %)
Previous(Accuracy %)
20
0
n
lin
g
in
g rs on
itio b e v ete ati
gn m r i m e
co se et ra cr
Re as tar pa del
t o
in Da ed M
Po r ain
T
CHAPTER 7
CONCLUSION AND FUTURE SCOPE
Page 44
Application of Deep Learning for sign Language Gesture Recognition with Efficient Hand Gesture Representation
7.1 Conclusion
Nowadays, applications need several kinds of images as sources of information for elucidation
and analysis. Several features are to be extracted so as to perform various applications. When
an image is transformed from one form to another such as digitizing, scanning, and
communicating, storing, etc. degradation occurs. Therefore the output image has to
undertake a process called image enhancement, which contains of a group of methods that
seek to develop the visual presence of an image. Image enhancement is fundamentally
enlightening the interpretability or awareness of information in images for human listeners and
providing better input for other automatic image processing systems. Image then undergoes
feature extraction using various methods to make the image more readable by the
computer.Sign language recognition system is a powerful tool to preparae an expert
knowledge, edge detect and the combination of inaccurate information from different
sources. the intend of convolution neural network is to get the appropriate classification.
Page 45
Application of Deep Learning for sign Language Gesture Recognition with Efficient Hand Gesture Representation
REFERENCES
[1] Ahmed, Mohamed Aktham, et al. ”A review on systems-based sensory gloves for sign
language recognition state of the art between 2007 and 2017.” Sensors 18.7 (2018).
[2] Han, Rui, et al. ”A Data Glove-based KEM Dynamic Gesture Recognition Algorithm.”
International Journal of Performability Engineering 14.11 (2018).
[3] Ronchetti, Franco, Facundo Quiroga, César Armando Estrebou, and Laura Cristina Lanzarini.
"Handshape recognition for argentinian sign language using probsom." Journal of Computer
Science & Technology 16 (2016).
[4] Abhishek, Kalpattu S., Lee Chun Fai Qubeley, and Derek Ho. ”Glovebased hand gesture
recognition sign language translator using capacitive touch sensor.” Electron Devices and
Solid-State Circuits (EDSSC), 2016 IEEE International Conference on. IEEE, 2016.
[5] Ronchetti, Franco, Facundo Quiroga, César Armando Estrebou, Laura Cristina Lanzarini, and
Alejandro Rosete. "LSA64: An Argentinian Sign Language Dataset." In XXII Congreso
Argentino de Ciencias de la Computación (CACIC 2016). 2016.
[6] Das, Abhinandan, et al. ”Smart glove for Sign Language communications.” Accessibility to
Digital World (ICADW), 2016 International Conference on. IEEE, 2016.
[7] Abadi, Martín, Ashish Agarwal, Paul Barham, Eugene Brevdo, Zhifeng Chen, Craig Citro,
Greg S. Corrado et al. "Tensorflow: Largescale machine learning on heterogeneous
distributed systems." arXiv preprint arXiv:1603.04467 (2016).
[8] Lokhande, Priyanka, Riya Prajapati, and Sandeep Pansare. ”Data gloves for sign language
recognition system.” International Journal of Computer Applications (2015): 11-14.
[9] Singha, Joyeeta, and Karen Das. "Automatic Indian Sign Language Recognition for
Page 46
Application of Deep Learning for sign Language Gesture Recognition with Efficient Hand Gesture Representation
Page 47
Application of Deep Learning for sign Language Gesture Recognition with Efficient Hand Gesture Representation
Page 48