CPP Project
CPP Project
REPORT ON
SIGN LANGUAGE ACCESSIBILITY
ACADEMIC YEAR
2023-24
GROUP MEMBERS
SR.NO NAME OF STUDENT ENROLLMENT
GUIDED BY
Mrs. Madhura Mahindrakar
MAHARASHTRA STATE BOARD OF TECHNICAL
EDUCATION
CERTIFICATE
This is to certify that the project titled "SIGN LANGUAGE ACCESSIBILITY" is a bonafied work
carried out by student of Diploma in Computer Engineering as a part of curriculum as prescribed by
MSBTE. I hereby declare that the project work has not formed the basis for the award previously of
any Diploma, Associate ship. Fellowship or any other similar title according to my knowledge.
Signature of student
1.
2.
3.
4.
ACKNOWLEDGEMENT
We extend our special thanks to all teaching and non-teaching staff. Success is nourished under the
combination of perfect guidance, care and blessing. Acknowledgement is the best way to convey Last few
years spend in estimated institution has moulded us into confident and aspiring engineers. We express our
sense of gratitude towards our project guide MRS. MADHURA MAHINDRAKAR It is because of his
valuable guidance, analytical approach encouragement that we could learn, work and complete the project.
We will always cherish great experience work under the enthusiastic guidance.
We are also grateful to our principal and our vice principal who not only supported us in our project but
also encouraged for every creative activity. We also sincerely give thanks to our head of department
MRS. MADHURA MAHINDRAKAR of computer and its sector, friends and well-wishers to directly or
indirectly contribute for the success of our maiden mission.
ABSTRACT
The aims to bridge the communication gap for individuals with hearing or speech disabilities by
translating American Sign Language (ASL) into both text and speech in real-time. Using computer vision
and machine learning techniques, the app captures hand gestures through the device's camera and
processes them into meaningful words and sentences, allowing disabled individuals to communicate
effectively with those who do not understand sign language.
The app integrates powerful technologies like MediaPipe for hand tracking and a deep learning model
built with TensorFlow for gesture recognition. By focusing on hand movements, it ensures precise
detection of each ASL sign and dynamically converts these gestures into readable text or spoken words
using text-to-speech functionalities. Additionally, the app includes a feature where users can input text,
and a 3D model created using Blender performs the corresponding ASL signs. This feature enables non-
signers to effectively communicate with individuals who rely on sign language, further promoting
inclusivity.
The app's user-friendly interface, built using Flutter, offers a seamless experience for users of all ages.
Designed to be accessible, promotes inclusivity by providing a tool for disabled individuals to interact
more easily in both social and professional settings. Whether used for everyday conversations or more
formal situations, the app enhances communication, making it a valuable asset for the disabled community
and beyond.
INDEX
1. Introduction 1
2 Problem Statement 3
3. Review of Literature 5
5. Use Case 10
6. Plan Of Work 12
8. Future Scope 17
9. Conclusion 19
10. Reference 21
CHAPTER 1
INTRODUCTION
1
INTRODUCTION
Sign language accessibility is essential for fostering inclusivity and equal communication opportunities
for people who are Deaf or hard of hearing. Sign language serves as a primary mode of communication
for millions globally, bridging the gap between spoken languages and ensuring that people with hearing
impairments can access information, education, and social interaction. Inaccessible communication
channels can lead to isolation and disadvantage, making it crucial to integrate sign language into various
domains such as media, technology, and public services.
Advances in technology have opened new doors for improving sign language accessibility. Digital tools,
real-time sign language interpretation, and assistive devices empower individuals and support effective
communication across diverse environments. As awareness grows about the importance of sign language
inclusion, efforts to enhance accessibility aim to provide a more equitable and connected world, where
language barriers are minimized and full participation is made possible for everyone.
2
CHAPTER 2
PROBLEM STATEMENT
3
PROBLEM STATEMENT
Sign language accessibility is a groundbreaking solution aimed at addressing a long-standing challenge for
individuals with hearing or speech disabilities. By translating sign language into both text and speech in
real-time, this app empowers individuals to communicate effectively with those unfamiliar with sign
language, breaking down communication barriers and fostering inclusivity. Additionally, the app includes
a unique 3D model featurethat visually performs ASL gestures, enabling non-signers to follow along and
better understand the signs being communicated.
4
3.Who will benefit from Sign language accessibility?
The app is designed to benefit individuals with hearing or speech impairments, enabling them to
communicate more freely and independently. Beyond this, the app’s functionality also supports those
around them family members, friends, colleagues, healthcare professionals by allowing smoother, more
accessible interactions. The 3D model benefits non-signers, providing a visual aid to understand and
follow ASL gestures, fostering better communication between disabled individuals and the general public.
By removing communication barriers, Sign language accessibility has the potential to impact society at
large, promoting inclusivity and better understanding.
5
CHAPTER 3
REVIEW OF LITERATURE
6
REVIEW OF LITERATURE
7
CHAPTER 4
DESIGN AND PLANNING
8
DESIGN AND PLANNING
Modules:
This component is responsible for capturing American Sign Language (ASL) gestures using the
device's camera. It processes the visual data and prepares it for further translation.
● Text Input:
Users can input text that they want to be translated into sign language. This feature allows
communicating users (those without hearing or speech disabilities) to interact effectively with the
disabled user.
● Settings:
A configurable area where users can customize accessibility options, user preferences, and other
application settings to enhance their experience and usability.
● Output Display:
This component presents the translated output to the user, either in the form of text or visual
animations. It ensures that the user receives clear and understandable communication results.
● Gesture Processing:
A backend process that analyzes the captured gestures from the "Gesture Capture" component. It
prepares the data for translation into ASL and further processing.
● Text Processing:
Similar to gesture processing, this component handles the text entered by communicating users. It
prepares the input for translation into ASL gestures.
● ASL Translation:
The core functionality of the application, where both the gestures and text are translated into sign
language. This component integrates the results of gesture and text processing to generate
appropriate responses.
● 3D Model Animation:
This component generates a 3D visual representation of the ASL signs based on the translated
output. It helps non-signers visually understand the signs being communicated.
● Text-to-Speech:
Converts the translated text into spoken language, allowing users to express themselves vocally
and bridging the gap between signers and non-signers. It is particularly useful in social, educational,
and professional contexts.
10
CHAPTER 5
USE CASE DIIAGRAM
11
USE CASE
A use-case diagram is usually referred to as behaviour diagrams used to describe a set of actions (use cases)
that some system or systems (subject) should or can perform in collaboration with one or more external
users of the system (actors). A use case diagram at its simplest is a representation of a user's interaction
with the system that shows the relationship between the user and the different use cases in which the user
is involved. A use case diagram can identify the different types of users of a system and the different use
cases and will often be accompanied by other types of diagrams as well.
12
CHAPTER 6
PLAN OF WORK
13
PLAN OF WORK
1. Requirement Analysis :
Software Development Life Cycle begins with Requirement Analysis phase, where the
stakeholders discuss the requirements of the software that needs to be developed to achieve a goal.
The aim of the requirement analysis phase is to capture the detail of each requirement and to make
sure everyone understands the scope of the work and how each requirement is going to be fulfilled.
2. Design:
The next stage of Software Development Life Cycle is the Design phase. During the design phase,
developers and technical architects start the high-level design of the software and system to be able
to deliver each requirement.
3. Implementation or coding:
After the requirements and design activity is completed, the next phase of the Software
Development Life Cycle is the implementation or development of the software. In this phase,
developers start coding according to the requirements.
4. Testing:
Testing is the last phase of the Software Development Life Cycle before the software is delivered
to customers. During testing, experienced testers start to test the system against the requirements.
The testers aim to find defects within the system as well as verifying whether the application
behaves as expected and according to what was documented in the requirements analysis phase.
14
Figure 6.1. Phases of software development life cycle
Agile Methodology:
Agile is a flexible and iterative approach to software development that is well-suited to projects
with evolving requirements and a need for continuous adaptation. Agile, work is typically
divided into small, manageable units known as sprints.
Each sprint delivers a potentially shippable product increment. This approach allows
you to prioritize and develop modules incrementally, making it easier to adjust to changing
needs and add new modules as required.
As shown in the figure, every module has been planned, designed, coded, tested and
then launched again the same process is repeated for other module.
15
CHAPTER 7
PROS & CONS
16
Pros of Sign language accessibility:
1. Real-Time Translation: Converts sign language gestures into both text and speech instantly, making
communication faster and more efficient.
2. Inclusivity: Helps bridge the communication gap between those who use sign language and those who
don’t, fostering a more inclusive society.
3. User-Friendly Interface: The intuitive design ensures ease of use for individuals of all ages and tech
experience.
4. 3D Model Feature: Non-signers can learn and understand ASL visually through the animated 3D model
feature, enhancing comprehension.
5. Cross-Platform Support: Built using Flutter, the app works on both Android and iOS, increasing
accessibility.
17
CHAPTER 8
FUTURE SCOPE
18
FUTURE SCOPE
Community Engagement:
Build a community for peer support and learning through forums, challenges, and live classes.
Accessibility Focus:
Enhance features to accommodate users with various disabilities, ensuring inclusivity.
Global Outreach:
Conduct awareness campaigns to educate the public about the app and the importance of
communication accessibility.
19
CHAPTER 9
CONCLUSION
20
CONCLUSION
Sign language accessibility is more than just an application; it is a transformative tool that bridges
communication gaps for individuals with hearing or speech disabilities. By facilitating real-time
translation of sign language into text and speech, the app empowers users to engage more fully in
everyday interactions, education, and professional environments. Its diverse applications promote
inclusivity, foster understanding, and enhance the quality of life for individuals with disabilities. As Sign
language accessibilitycontinues to evolve and expand, it holds the potential to make a significant impact
on society, paving the way for a world where everyone can communicate effortlessly, regardless of their
abilities. Through this innovative approach,Sign language accessibilityaims to create a more connected
and empathetic community for all.
21
CHAPTER 10
REFERENCE
21
REFERENCE
1. https://www.handtalk.me/en/app/
2. https://www.google.co.in/
3. https://www.youtube.com/
4. https://www.mixamo.com/#/
5. https://www.blender.org/
22