Highly Time and Power Sacrifised
Highly Time and Power Sacrifised
A Project Report
On
BACHELOR OF ENGINEERING
IN
COMPUTER SCIENCE AND ENGINEERING
SUBMITTED BY
NISHANTH C N (1JB21CS099)
NISCHITH H U (1JB21CS098)
N P SOORAJ (1JB21CS092)
2024 - 2025
|| Jai Sri Gurudev ||
Sri Adichunchanagiri Shikshana Trust ®
SJB INSTITUTE OF TECHNOLOGY
No.67, BGS Health & Education City, Dr.Vishnuvardhan Rd, Kengeri, Bengaluru, Karnataka 560060
CERTIFICATE
Certified that the Project Work entitled “AL ASSISTED EYE FOR THE BLIND”
carried out by NISCHITH H.U, NISHANTH C.N , N.P SOORAJ bearing USN 1JB21CS098,
1JB21CS099, 1JB21CS092 are bonafide students of SJB Institute of Technology in
partial fulfilment for 7th semester of BACHELOR OF ENGINEERING in COMPUTER
SCIENCE AND ENGINEERING of the Visvesvaraya Technological University, Belagavi
during the academic year 2024-25. It is certified that all corrections/suggestions indicated
for Internal Assessment have been incorporated in the Report deposited in the
Departmental library. The project report has been approved as it satisfies the academic
requirements in respect of Project work prescribed for the said Degree.
1. __________________________ ______________________________
2.___________________________ ______________________________
ACKNOWLEDGEMENT
We would like to express our profound grateful to His Divine Soul Jagadguru Padmabhushan Sri Sri Sri Dr.
Balagangadharanatha Mahaswamiji and His Holiness Jagadguru Sri Sri Sri Dr. Nirmalanandanatha
Mahaswamiji for providing us an opportunity to complete our academics in this esteemed institution.
We would also like to express our profound thanks to Revered Sri Sri Dr. Prakashnath Swamiji, BGS &
SJB Group of Institutions, for his continuous support in providing amenities to carry out this Project Work in
this admired institution.
We express our gratitude to Dr. Puttaraju, Academic Director, BGS & SJB Group of Institutions, for providing
us an excellent facilities and academic ambience; which have helped us in satisfactory completion of
Project work.
We express our gratitude to Dr. K. V. Mahendra Prashanth, Principal, SJB Institute of Technology, for
providing us an excellent facilities and academic ambience; which have helped us in satisfactory
completion of Project work.
We extend our sincere thank to all the Dean of SJB Institute of Technology, for providing us an invaluable
support throughout the period of our Project work.
We extend our sincere thanks to Dr. Krishan A N, Head of the Department, Computer Science and
Engineering for providing us an invaluable support throughout the period of our Project work.
We wish to express our heartfelt gratitude to our Project Coordinator Dr.Roopa M.J Associated
Professor, Department of CSE & guide Dr. Prakruthi M K Associate Professor Department of CSE for
her valuable guidance, suggestions and cheerful encouragement during the entire period of our Project work.
Finally, we take this opportunity to extend our earnest gratitude and respect to our parents, Teaching & Non
teaching staffs of the department, the library staff and all our friends, who have directly or indirectly supported
us during the period of our Project work.
Regards,
The AI-assisted eye for the blind project aims to revolutionize the lives of blind and visually
impaired individuals by enhancing their autonomy, mobility, and independence through advanced
technology. By utilizing AI-powered systems, the project converts real-time visual data into auditory
or tactile feedback, assisting users in navigating their environment, recognizing objects, reading
text, and engaging in social interactions. Leveraging technologies like computer vision, machine
learning, and natural language processing, the system provides essential support, improving the
user’s ability to move confidently, recognize everyday objects, and perform tasks such as reading
signs, labels, and printed materials, all through voice feedback.
Incorporating wearable devices such as smart glasses or smartphones, the project aims to offer a
user-friendly, intuitive solution that can be easily adapted to diverse environments and individual
needs. The technology is designed to be affordable and accessible, ensuring it reaches a broad
demographic, including underserved communities. By emphasizing accessibility and social inclusion,
this AI-assisted system fosters greater independence for blind individuals, allowing them to navigate
the world with greater ease and participate more fully in social and daily life activities, ultimately
promoting equality and opportunities for all.
ii
TABLE OF CONTENTS
Acknowledgement i
Abstract ii
Table of Contents iii
List of Figures iv
Chapter 1
Introduction
1.1 Introduction to Internet of Things .........................................................................1
1.2 Functioning of IoT ............................................................................................... 2
1.3 Future Vision of IoT .............................................................................................. 3
1.4 IoT : Services and Applications ............................................................................ 4
1.5 Problem statement ................................................................................................. 5
1.6 Goals and Objectives of the project ...................................................................... 5
Chapter 2
Literature survey
2.1 Design of Smart e-Tongue for the Physically Challenged People ........................ 6
2.2 MyVox-Device for the communication between people: blind, deaf,
deaf- blind and unimpaired .................................................................................... 7
2.3 PiCam:IoT based Wireless Alert System for Deaf and Hard of Hearing .............. 7
2.4 Design of a communication aid for physically challenged ....................................8
2.5 Implementation Of Gesture Based Voice And Language Translator For
Dumb People.......................................................................................................... 9
2.6 Intelligent Gesture Recognition System for Deaf and Dumb Individuals .......... 10
2.7 Voice-Assisted Smart Glasses for the Blind......................................................... 11
2.8 Sign Language Translation via Neural Networks ............................................... 11
2.9 IoT-Enabled Alert System for Hearing Impaired Individuals ..............................12
2.10 Low-Cost Braille e-Reader for the Visually Impaired ...................................... 12
Chapter 4
System requirements
4.1 What is system analysis ...................................................................................... 16
4.2 Software Requirement Specification .................................................................... 16
4.3 Hardware Components ......................................................................................... 17
4.4 Software Components ......................................................................................... 17
4.5 Hardware Component Description ....................................................................... 18
4.6 Software Component Description ........................................................................24
4.6.1 Raspbian OS...................................................................................................... 24
4.6.2 Tesseract OCR................................................................................................... 24
4.6.3 Open CV ........................................................................................................... 25
4.6.5 Xming ............................................................................................................... 27
4.6.6 Putty ................................................................................................................ 28
Chapter 5
System architecture
5.1 High Level Design .............................................................................................. 29
5.1.1 System Architecture ......................................................................................... 29
5.1.2 Data Flow Diagram ...........................................................................................30
5.1.3 Sequence Diagram ........................................................................................... 31
5.1.4 Use Case Diagram ............................................................................................. 33
Implementation
6.1 Overview of System Implementation ................................................................... 35
6.2 Module Description.............................................................................................. 36
ii
6.2.1 Text-to-speech (TTS) ....................................................................................... 36
6.2.2 Image-to-speech using camera (ITSC) .............................................................37
6.2.3 Gesture-to-speech (GTS) ................................................................................. 37
6.2.4 Speech-to-Text (STT) .......................................................................................38
6.3 Pseudo code .........................................................................................................38
6.3.1 Image-to-speech using camera (ITSC) .............................................................39
Chapter 7
Testing
7.1 Unit testing...........................................................................................................40
7.2 Integration testing ................................................................................................42
7.3 Functional testing… ..................................................................................... …….43
7.4 Acceptance testing… .....................................................................................……45
7.5 System testing……… .................................................................................... …...45
Chapter 8
Results
8.1 Final connection..................................................................................................47
8.2 User of the project .............................................................................................. 47
Conclusion ................................................................................................................................ 49
Future Enhancement ................................................................................................................49
References ..................................................................................................................................50
iii
LIST OF FIGURES
SL NO DESCRIPTION PAGE NO
vii
LIST OF FIGURES
SL NO DESCRIPTION PAGE NO
viii
Chapter 1
INTRODUCTION
1.1 Introduction to Internet of Things
The Internet of Things (IoT) is the interconnection of uniquely identifiable embedded computing
devices within the existing Internet infrastructure. Typically, IoT is expected to offer advanced
connectivity of devices, systems, and services that goes beyond machine-to-machine
communications (M2M) and covers a variety of protocols, domains, and applications. The
interconnection of these embedded devices (including smart objects), is expected to user in
automation in nearly all fields, while also enabling advanced applications like a Smart Grid.
Things, in the IoT, can refer to a wide variety of devices such as heart monitoring implants, biochip
transponders on farm animals, electric clams in coastal waters, automobiles with built-in sensors,
or field operation devices that assist fire-fighters in search and rescue. Current market examples
include thermostat systems and washer/dryers that utilize wifi for remote monitoring.
According to Gartner, Inc. (a technology research and advisory corporation), there will be
nearly 26 billion devices on the Internet of Things by 2020. ABI Research estimates that more than
30 billion devices will be wirelessly connected to the Internet of Things(Internet of Everything) by
2020. As per a recent survey and study done by Pew Research Internet Project, a large majority of
the technology experts and engaged Internet users whoresponded83 percent agreed with the notion
that the Internet/Cloud of Things, embedded and wearable computing (and the corresponding
dynamic systems) will have widespread and beneficial effects by 2025. It is, as such, clear that the
IoT will consist of a very large number of devices being connected to the Internet. Integration with
the Internet implies that devices will utilize an IP address as a unique identifier. However, due to
the limited address space of IPv4, objects in the IoT will have to use IPv6 to accommodate the
extremely large address space required. Objects in the IoT will not only be devices with sensory
capabilities, but also provide actuation capabilities (e.g., bulbs or locks controlled over the
Internet).
To a large extent, the future of the Internet of Things will not be possible without the
support of IPv6 and consequently the global adoption of IPv6 in the coming years will be critical
for the successful development of the IoT in the future. The embedded computing nature of
AI Assisted Eye For The Blind
many IoT devices means that low-cost computing platforms are likely to be used. In fact, to
minimize the impact of such devices on the environment and energy consumption, low-power radios
are likely to be used for connection to the Internet. Such low-power radios do not use Wifi, or well
established Cellular Network technologies, and remain an actively developing research area.
However, the IoT will not be composed only of embedded devices, since higher order
computing devices will be needed to perform heavier duty tasks (routing, switching, data processing,
etc.). Companies such as Free Wave Technologies have developed and manufactured low power
wireless data radios (both embedded and standalone) for over 20years to enable Machine-to-Machine
applications for the industrial internet of things. Besides the plethora of new application areas for
Internet connected automation to expand into, IoT is also expected to generate large amounts of data
from diverse locations that is aggregated and very high-velocity, thereby increasing the need to better
index, store and process such data. Diverse applications call for different deployment scenarios and
requirement, which have usually been handled in a proprietary implementation. However, since the IoT
is connected to the Internet, most of the devices comprising IoT services will need to operate utilizing
standardized technologies. Prominent standardization bodies, such as the IETF, IPSO Alliance and
ETSI, are working on developing protocols, systems, architectures and frameworks to enable the IoT.
provider known as Internet. The convergence of the sensors like smart objects, RFID based sensor
networks and Internet gives rise to the Internet of Things. With increased usage of sensors the raw
data as well as distributed data is increasing. Smart devices are now connected to Internet using
their communication protocol and continuously collecting and processing the data.
Ubiquitous computing which was thought as a difficult task has now become a reality due
to advances in the field of Automatic Identification, wireless communications, distributed
computation process and fast speed of Internet. From just a data perspective the amount of data
generated, stored and processed will be enormous. We focused on making this architecture as a
sensor based architecture where each sensor node will be as important as the sensor network itself.
Visualizing each sensor as having intelligence is the ultimate aim of any architecture in the IoT
domain. There is a lot of pervasive presence in the human environment of things or objects,
described general overview of internet evolution with several IoT services with the use of radio-
frequency identification (RFID) tags, sensors, actuators, mobile phones, smart embedded devices,
etc. which, through unique addressing schemes, are able to effectively communicate and interact
with each other and work together to reach a common goal of making the system easier to operate
and utilize. The objects that will be connected will be adaptive, intelligent, and responsive.
Things Oriented Vision: This vision is supported by the fact that we can track anything using
sensors and pervasive technologies using RFID. The basic philosophy is uniquely identifying any
object using specifications of Electronic Product Code (EPC) .This technique is extended using
sensors. It is important to appreciate the fact that future vision will depend upon sensors and its
capabilities to fulfill the “things” oriented vision. We will be able to generate the data collectively
with the help of sensors, and sensor type embedded system. The summarized vision will be
dependent upon sensor based networks as well as RFID-based Sensor Networks which will take
care of the integration of technology based on RFID and sophisticated sensing and computing
devices and the global connectivity.
Internet Oriented Vision: The internet-oriented vision has pressed upon the need to make smart
objects which are connected. The objects need to have characteristics of IP protocols as this is one
of the major protocols being followed in the world of Internet. The sensor based object can be
converted in to an understandable format, which can be identified uniquely and its attributes can be
continuously monitored. This makes the base for smart embedded objects which can be assumed
to be a microcomputers having computing resources.
Methodology:
To overcome this communication barrier we propose a smart module which will convert the
signs in hand gesture and face expression into text, readable by normal people. The two different
approaches are used one using sensor and other using image processing. We will use image
processing technique in our work. A complete solution for conversion of ISL to text will be
implemented and an algorithm will be designed and implemented to convert ISL to text at real
time.
Drawbacks:
Insufficient Datasets typically refers to classification tasks where the classes are not
.represented equally
It is imperative to choose the evaluation metric of your model correctly. If it is not done,
.then you might end up adjusting/optimizing a useless parameter
The number of fraudulent transactions as compares to the number of non-fraudulent
.transactions will be very much less. And this is where the problem arises
2.2. MyVox—Device for the communication between people: blind, deaf, deaf-
blind and unimpaired [2]
AI Assisted Eye For The Blind
Year: 2014
Author: Fernando Ramirez-Garibay, Cesar Millan Ollivarria
Problem Addressed:
A deaf-blind person is one with impaired senses of hearing and sight. Those who have lost only
one of these senses can use the rest to compensate up to a certain point, and can develop techniques
to successfully communicate with others. Blind people can hear and speak, deaf people can learn
how to write, understand sign language, and read lips. deaf-blind people, however, left mainly with
touch and smell to connect with the world around them, do not have direct access to the main
communication channels that others use in their every day life.
Methodology:
Have presented a device for the communication of deaf-blind people. While their lack of hearing
and sight could represent a limitation when communicating with others, technology is presented
that can be of use for communicating with others who do not speak sign language. The
communication device, named MyVox, has proven to be a useful tool for an Usher syndrome
patient who is now able to communicate with others without the need of an interpreter. Based on
his feedback, are developing an upgraded system that will also be tried by a larger population of
deaf-blind users.
Drawbacks:
No internet accessibility, portability refers Individuals with disabilities such as visual, auditory,
or speech impairments has difficulty in communicating as others have limited or no knowledge
of Braille and sign language.
BDD does not assist is a low-cost portable device which aims to solve this problem by acting
as a mediator to bridge this communication gap and reduce the dependence on expert human
translators.
There is no LAN connection for internet access for using the espeak.
2.3 PiCam:IoT based Wireless Alert System for Deaf and Hard of Hearing[3]
Year: 2015
Author: P. Kumari, S.R.N Reddy
Problem Addressed:
Hearing loss presents many everyday challenges. Communication may be the biggest challenge
of all getting and giving information, exchanging ideas, sharing feeling-whether in one-to-one
contact or in groups. Sometimes there are small disruptions of daily life that result from reduced
hearing.
Methodology:
Micro array camera image, Super-Resolution with Neural Network and Fusion of V-system. The
objective of this paper is to design and implement a low cost stand-alone device for deaf people to
notify doorbell ringing who live alone in their house. The system is based on Raspberry pi which
includes camera, vibrator, wireless GSM and Bluetooth. When the visitor presses the doorbell,
captured image is transferred to the wearable device which helps to know the right person at the
door or intruder. After transferring image, wearable device vibrates to notify. Also, the message is
sent to the owner through GSM. Visitor’s image along with the date and time is sent to the server
for retrieving information later. The system is reliable, effective, and easy to use and also enhances
the security of the user.
Drawbacks :
Image received on wearable device through Bluetooth which create disturbance in the image,
the first step towards image processing is to acquire the image.
The acquired image that is stored in the system windows needs to be connected to the software
automatically. This is done by creating an object
Image acquisition devices typically does not support multiple video formats also simple images
as it has disturbance and also if there is no clear background leads to disturbance in image
capture.
communicate among them at the same time with others. As here the communication for the
disabled people based on the regional sign language.
Methodology:
The communication is held by means of a glove based deaf- mute communication interpreter
method. The glove is inside fitted out with flex sensors, tactile sensors and accelerometer. The
gestures are converted into text display. The design of communication aid for disabled, prototype
and testing of a portable keyboard and speaker device with a braille refreshable display for
communication between two people. The communication between the deaf and dumb people is
carried out by using gesture based device for deaf and dumb person which changes sign language
to text display for voice to voiceless people with the help of smart gloves.
Drawbacks:
Communication gap ,low sign recognition, defining accurate hand shape information is one
of the most crucial tasks in gesture controlled computer applications.
Specifically in SLR systems, the gesture numbers are large and that makes the hand shape
definition more difficult.
If 2D images are captured by a single camera that makes extraction of the hand shapes from
the video more complex also leads to the communication gap.
Methodology:
Dual channel ADC, this paper proposes a system that converts gestures given by the user in the
form of English alphabets into corresponding voice and translates this English voice output into
any other Microsoft supported languages. The system consists of MPU6050 for sensing gesture
movement, Raspberry pi for processing, three button Keypad and speaker. It is implemented by
using trajectory recognition algorithm for recognizing alphabets. Raspberry pi generates voice
output for the text in multiple languages using voice RSS and Microsoft translator. When tested,
the system recognized A-Z alphabets and generated voice output based on the gestures in multiple
languages.
Drawbacks
* High cost of flex sensors.
* Understanding and required prior knowledge.
* High initial investment.
* Does not applicable in night time.
u
2.6 Intelligent Gesture Recognition System for Deaf and Dumb Individuals [6]
Year: 2018
Author: A. Singh, P. Patel
Problem Addressed:
Deaf and dumb individuals often face difficulties in communicating with those who do not
understand sign language. The lack of an interpreter creates a significant barrier to interaction.
Methodology:
This study proposed a wearable gesture recognition system using flex sensors embedded in
gloves. The sensors captured hand movements and transmitted data to a microcontroller,
which translated the gestures into text and voice using a trained machine learning model. The
system was evaluated for multiple gestures, achieving a recognition accuracy of 92% for 26
ASL alphabets.
Drawbacks:
Limited to predefined gestures and alphabets.
Struggles with complex phrases requiring multi-gesture inputs.
The main goal of our project is to provide a standard lifestyle for deaf dumb and blind
peoples as normal ones. Through this device the visually impaired people can able to understand
the words easily. The vocally impaired people can communicate their message through text and
gestures. The deaf people can able to understand others speech from the text that is displayed. This
helps them to experience the Independent life.
3.1.1 Advantages of the proposed system
All-in-one device, where the deaf, dumb and blind can overcome their disabilities and
can express their views to others.
Voice to text Conversion for Deaf People
Image to Voice Conversion for Blind people.
Sign to text and text to voice conversion for dumb people communicating to normal
person.
It is a portable device with a very low.
array of colour pixel, where each colour pixel is a triplet which corresponds to red, blue and green
colour component of RGB image at a specified spatial location. Similarly, A Grayscale image can
be viewed as a single layered image. In MATLAB, a grayscale image is basically M*N array whose
values have been scaled to represent intensities.
This algorithm is able to accurately decipher and extract text from a variety of sources!
As per it's namesake it uses an updated version of the tesseract open source OCR tool. We also
automatically binarize and preprocess images using the binarization so tesseract has an easier time
deciphering images. Not only are we able to extract English text, but tesseract supports over 100
other languages as well! Give it a try and don't forget to leave a comment if you like it or have a
suggestion/comment.
iv. Speechtexter:
Speechtexter is an online multi-language speech recognizer that can help you type long documents,
books, reports, blog posts with your voice. If you need help, please visit our help page at https://
www.speechtexter.com/help .This app supports over 60 different languages!
For better results use a high quality microphone, remove any background noise, and speak
.loudly and clearly
.It can create text notes/ sms/emails/tweets from users voice
SYSTEM REQUIREMENTS
Gathering requirements is the main attraction of the Analysis Phase. The process of gathering
requirements is usually more than simply asking the users what they need and writing their answers
down. Depending on the complexity of the application, the process for gathering requirements
has a clearly defined process of its own. This process consists of a group of repeatable processes
that utilize certain techniques to capture, document, communicate, and manage requirements.
your own speech is deafened by the fact you are wearing a headset.
Device should be able to capture image and display it on the screen meanwhile the
voice should be generated: The image will be captured by the camera, using Tesseract
OCR it is converted to text and could be displayed on the screen. And even voice output
is given through headphones using espeak.
Device should be able to convert the received speech-to-text:The input given as
speech through speaker of camera, it is converted to text using Speechtexter which is
available online.
The device is implemented using a Raspberry pi 3B board and their specifications are as
follows.
b. Power source
The Pi is a device which consumes 700mA or 3W or power. It is powered by a MicroUSB
charger or the GPIO header. Any good smartphone charger will do the work of powering the
Pi.
c. SD Card
The Raspberry Pi does not have any onboard storage available. The operating system is loaded
on a SD card which is inserted on the SD card slot on the Raspberry Pi. The operating system
can be loaded on the card using a card reader on any computer.
A standard 3.5 mm TRS connector is available on the RPi for stereo audio output. Any
headphone or 3.5mm audio cable can be connected directly. Although this jack cannot be used
for taking audio input, USB mics or USB sound cards can be used.
f. Status LEDs
There are 5 status LEDs on the RPi that show the status of various activities as follows:
“OK” - SDCard Access (via GPIO16) - labeled as "OK" on Model B Rev1.0 boards and
h. Ethernet
Ethernet port is available on Model B and B+. It can be connected to a network or internet
using a standard LAN cable on the Ethernet port. The Ethernet ports are controlled by
Microchip LAN9512 LAN controller chip.
i. CSI connector
CSI–Camera Serial Interface is a serial interface designed by MIPI (Mobile Industry
Processor Interface) alliance aimed at interfacing digital cameras with a mobile processor.
The RPi foundation provides a camera specially made for the Pi which can be connected with
the Pi using the CSI connector.
j. JTAG headers
JTAG is an acronym for ‘Joint Test Action Group', an organization that started back in the
mid 1980's to address test point access issues on PCB with surface mount devices. The
organization devised a method of access to device pins via a serial port that became known as
the TAP (Test Access Port). In 1990 the method became a recognized international standard
(IEEE Std 1149.1). Many thousands of devices now include this standardized port as a feature
to allow test and design engineers to access pins.
k. HDMI
HDMI – High Definition Multimedia Interface
HDMI 1.3 a type A port is provided on the RPi to connect with HDMI screens.
Models
Video outputs Composite RCA (PAL and NTSC) –in model B+ via 4-pole 3.5 mm jack,
HDMI (rev 1.3 & 1.4), raw LCD Panels via DS
Onboard SD / MMC / SDIO card slot (3.3 V card MicroSD
Low-level 8× GPIO, UART, I²C bus, SPI bus with two 17× GPIO
4.6.1 Raspbian OS
Although the Raspberry Pi’s operating system is closer to the Mac than Windows, it’s the
latter that the desktop most closely resembles It might seem a little alien at first glance, but
using Raspbian is hardly any different to using Windows (barring Windows 8 of course).
There’s a menu bar, a web browser, a file manager and no shortage of desktop shortcuts of
pre-installed applications. Raspbian is an unofficial port of Debian Wheezy armhf with
compilation settings adjusted to produce optimized "hard float" code that will run on the
Raspberry Pi. This provides significantly faster performance for applications that make heavy
use of floating point arithmetic operations. All other applications will also gain some
performance through the use of advanced instructions of the ARMv6 CPU in Raspberry Pi.
Although Raspbian is primarily the efforts of Mike Thompson (mpthompson) and Peter Green
(plugwash), it has also benefited greatly from the enthusiastic support of Raspberry Pi
community members who wish to get the maximum performance from their device.
Page layout
analysis.
More languages
are supported.
Windows, Linux, and MacOS, openBSD in desktop and Android, IOS and Blackberry in
mobile. It is used in diverse purpose for facial recognition, gesture recognition, object
identification, mobile robotics, segmentation etc. It is a combination of OpenCV C++ API and
Python language. In our project we are using OpenCV version 2 OpenCV is used to gesture
control to open a camera and capture the image. It is also used in the image to text and voice
conversion technique.
4.6.5 Xming
Xming provides the X Window System display server, a set of traditional sample X
applications and tools, and a set of fonts. It features support of several languages and has
Mesa 3D, OpenGL, and GLX 3D graphics extensions capabilities. The Xming X server is
based on Cygwin/X, the X.Org Server. It is cross-compiled on Linux with the MinGW
compiler suite and the Pthreads-Win32 multi-threading library. Xming runs natively on
Windows and does not need any third-party emulation software. Xming may be used with
implementations of Secure Shell (SSH) to securely forward X11 sessions from other
computers. It supports Putty and ssh.exe, and comes with a version of Putty’s plink.exe.
The Xming project also offers a portable version of Putty. When SSH forwarding is not used,
the local file Xn.hosts must be updated with host name or IP address of the remote machine
where GUI application is started. The software has been recommended by authors of books
on free software when a free X server is needed and described as simple and easier to install
though less configurable than other popular free choices like Cygwin/X.
4.6.6 Putty
Putty is a secluded and open-source mortal emulator, serial comfort network file transfer
application. Putty was formulated for Microsoft Windows, but it has been ported to various
other operating systems. It can link up to a serial port. It backs up a variety of network
protocols, together with SCP, SSH, Telnet, and raw socket connection.
Until the 1990s systems design had crucial and respected role in the data processing
industry. In the 1990s standardization of hardware and software resulted in the ability to build
modular systems. The increasing importance of software running on generic platforms has
enhanced the discipline of software engineering. Object-oriented analysis and design methods
are becoming the most widely used methods for computer systems design. The UML has
become the standard language in object-oriented analysis and design. It is widely used for
modelling software systems and is increasingly used for high designing non- software systems
and organizations.
System design is one of the most important phases of software development process.
The purpose of the design is to plan the solution of a problem specified by the requirement
documentation. In other words the first step in solution is the design of the project. The design
of the system is perhaps the most critical factor affecting the quality of the software. The
objective of the design phase is to produce overall design of the software. It aims to figure out
the modules that should be in the system to fulfill all the system requirements in efficient
manner. The design will contain the specification of all the modules, their interaction with
other modules and the desired output from each module. The output of the design process is a
description of the software architecture.
AI Assisted Eye For The Blind
System architecture is a conceptual model that defines the structure, behavior, and more
views of a system. An architecture description is a formal description and representation of a
system, organized in a way that supports reasoning about the structures and behaviors of the
system. The figure 5.1 shows a general block diagram describing the activities performed by
this project.
information about the timing of process or about whether process will operate in sequence
or in parallel
are typically associated with use case realizations in the Logical View of the system under
development. Sequence diagrams are sometimes called event diagrams or event scenarios.
Firstly the user sends a file to gui::OCR (graphical user interface). The Optical Character
Recognition (OCR) identifies if there is any image , next it gets image then if any character is
identified then it is returned back to the user. Then the image is read using
reader::Image_reader and passed to OCR Engine. Next it identifies a character using
nextChar: Graphic_char function, after reading all the characters it is segmented then moved
to recognition function. Finally the EOF is reached and returned back to the user.
The "actors" are people or entities operating under defined roles within the system. The
“scenario” is a specific sequence of actions and interactions between actors and the system.”
Use case” is a collection of related success and failure scenarios, describing actors using the
system to support a goal.
The user writes words in a file, he/she uploads a file containing words. This is given as input
to the system. System converts words into speech output. The users can pause/stop/rewind
this speech or he can even re-listen to the speech as shown in figure 5.4.
The user captures the image/gesture from the Logitech camera. He/she can zoom-in or zoom-
out the image. This is given as input to the system, speech output is given. One can
pause/stop/rewind the speech or even can re-listen to specific word.
The user gives the input through microphone and verifies voice perfection to the system.
The system gives text output to the user. This text gets printed on the system.
IMPLEMENTATION
Implementation is the realization of an application, or execution of a plan, idea, model, design,
specification, standard, algorithm, or policy. In other words, an implementation is a realization
of a technical specification or algorithm as a program, software component, or other computer
system through programming and deployment. Many implementations may exist for a given
specification or standard.
Implementation is one of the most important phases of the Software Development Life
Cycle (SDLC). It encompasses all the processes involved in getting new software or hardware
operating properly in its environment, including installation, configuration, and running, testing,
and making necessary changes. Specifically, it involves coding the system using a particular
programming language and transferring the design into an actual working system. This phase of
the system is conducted with the idea that whatever is designed should be implemented; keeping
in mind that it fulfills user requirements, objective and scope of the system. The implementation
phase produces the solution to the user problem.
number of defects. According to the number of defects, the text is printed on display and read
out by the Speaker.
6.2.4 Speech-to-Text (STT)
The fourth process is developed for the hearing impairment, people who cannot understand
the words of normal people. In order to help them, our project is provided with a switch which
is used to convert the voice of the normal people text. We have used a chromium browser
which is automatically connected to URL speechtexter.com. The process is performed by
assigning a minimum threshold voltage to recognize the voice signal. The input is given
through a microphone which is converted into a text format. The URL supports a variety of
languages. If the voice signals recognizable it will print the text else it gives the error signal.
programming language code, and that it is an efficient and environment- independent description
of the key principles of an algorithm. It is commonly used in textbooks and scientific publications
that are documenting various algorithms, and also in planning of computer program development,
for sketching out the structure of the program before the actual coding takes place.
No standard for pseudo code syntax exists, as a program in pseudo code is not an
executable program. Pseudo code resembles, but should not be confused with skeleton programs,
including dummy code, which can be compiled without errors. Flowcharts and Unified Modeling
Language (UML) charts can be thought of as a graphical alternative to pseudo code, but are more
spacious on paper.
The Project is divided into 4 different modules:
(Image-to-Speech using camera (ITSE
6.3.1 Image-to-speech using camera (ITSC)
Step 1: Start
Step 2: Choose option OP2 to convert image to speech Step 3: Call the function Image-to-
Speech ().
Step 4: Capture the required image.
Step 5: Convert image to text using Tesseract OCR. Step 6: Split the text into paragraph.
Step 7: Text is displayed on the screen.
Step 8: Next, call TexttoSpeech () function.
Unit testing is a level of software testing where individual units/ components of a software are
tested. The purpose is to validate that each unit of the software performs as designed. A unit is the
smallest testable part of any software. It usually has one or a few inputs and usually a single output.
Sr. Communication Sending Message By Reply to be send by B Result
No Case A
17. Blind and Message will be sent Audio message will be Passed
dumb person A to normal in form of Image by A sent by B
personB and converted into text
message.
18. Deaf and Audio message will be Audio message will be Passed
dumb person A to normal sent by A sent by B and converted
personB into text message.
19. Blind and deaf and Message will be sent Message will be sent in Passed
dumb person A to in form of Image by A form of Image by B.
blind person B
20. Blind and deaf and Message will be sent Text message by B to A Passed
dumb person A to in form of Image by A converted to audio or text
deaf and dumb and converted into text
person B message.
21. Blind and deaf and Message will be sent Text message by B to A Passed
dumb person A to in form of Image by A converted to audio or text
dumb person B and converted into text
message.
22. Blind and deaf and Message will be sent Text message by B to A Passed
dumb person A to in form of Image by A converted to audio or text
normal person B and converted into text
message.
23. Blind and deaf and Message will be sent Message will be sent in Passed
dumb person A to in form of Image by form of Image by B.
blind and dumb A.
person B
24. Blind and deaf and Message will be sent Message will be sent in Passed
dumb person A to in form of Image by form of Image by B.
Blind and deaf and A.
dumb person B
An integration test checks how two different components or subsystems interact with each other.
Like a unit test, it generally checks for a specific response to a particular input. Integration testing
takes as its input modules that have been unit tested, groups them in larger aggregates, applies
tests defined in an integration test plan to those aggregates, and delivers as its output the
integrated system ready for system testing.
In the project integrating camera, headphones and SDcard with power supply are needed.
Some of the USB ports are used if necessary along with the Ethernet cable.
This is usually not as easy to do with a set of integration or functional tests, but integration and
functional testing provides a measure of assurance that your "units" work together, as they will be
expected to when your application is run in production.
Functional tests are typically written using the Web Test package, which provides APIs for
invoking HTTP(S) requests to your application. We also use py.test and pytest- cov to provide
simple testing and coverage reports. The functional tests used in the project are mentioned below.
XMing is a free X window server for Microsoft Windows. It allows one to use Linux
graphical applications remotely. It is fully featured lean, fast, simple to install and because it is
standalone native Windows, easily made portable. Xming is totally secure when used with SSH,
its installers include executable code. PuTTY is Project Xming’s preferred and integrated X
terminal emulator.
Acceptance Testing is often the final step before rolling out the application. Usually the end users
who will be using the applications test the application before ‘accepting’ the application. This type
of testing gives the end users the confidence that the application being delivered to them meets
their requirements.
EXPECTED ACTUAL
TEST CASES RESULT
OUTCOME OUTCOME
Text to Speech Audio for Blind Audio for Blind Passed
The Final result is the Raspberry Pi 4 connected to a USB storage device for OS and data,
with wired headphones providing audio output. An camera module is also included for the
object detection part of the project.
Raspberry Pi 4: The Raspberry Pi 4 Model B is used for processing and running programs. It’s
connected to a USB storage device, likely containing the operating system (OS) or additional
data storage for the project.
Headphones: A pair of wired headphones with a damaged earcup, connected to the Raspberry
Pi 4 via the 3.5mm audio jack, providing audio output for the program or multimedia experience.
Camera Module :The Camera is a small, versatile module that connects to Raspberry Pi
boards via the CSI interface. It supports high-quality imaging, video recording, and AI-based
applications. Popular for DIY projects, it’s used in robotics, surveillance, and photography.
Storage Device: The USB storage device connected to the Raspberry Pi holds the operating
system (OS) and potentially other essential data for running the Raspberry Pi, completing the
setup for various projects and tasks.
AI Assisted Eye For The Blind
The image shows a person demonstrating the project designed to aid the visually impaired. The
setup includes a Raspberry Pi connected to various components, such as a camera mounted on the
user’s glasses and headphones for audio feedback. This innovative system processes visual data
and provides auditory descriptions to assist the user in navigating their environment effectively.
The device’s modular design ensures that it caters to diverse requirements, whether it involves
converting text into voice for visually impaired users or translating gestures into text and speech
for vocally impaired individuals. Moreover, the speech-to-text feature enables the hearing-impaired
to comprehend spoken words easily. With a focus on portability, affordability, and simplicity, the
system addresses key challenges while offering a scalable platform for future enhancements.
This project serves as a foundation for developing more sophisticated assistive technologies. It
has the potential to evolve into a language-independent and real-time communication system,
incorporating advanced AI models for improved gesture and voice recognition. By prioritizing
inclusivity, the assistive device paves the way for a standard lifestyle for the differently-abled,
ensuring they are no longer hindered by communication barriers.
FUTURE ENHANCEMENTS
1. Integration with Mobile Devices: Incorporate the system into a smartphone app for enhanced
portability and convenience.
2. Multi-Language Support: Expand the device’s capabilities to support more languages, making
it accessible to a global audience.
3. AI-Powered Gesture Recognition: Employ advanced AI models for real-time and accurate
recognition of complex gestures.
4. Improved Gesture Sensitivity: Enhance the sensitivity and precision of sensors to handle
intricate and dynamic gestures effectively.
5. Real-Time Translation: Include real-time translation features to bridge communication gaps
between individuals speaking different languages.
6. Cloud Connectivity: Utilize cloud-based systems for storing and analyzing data, enabling
continuous learning and updates to recognition models.
REFERENCES
[1] M. Delliraj, “Design of Smart e-Tongue for the Physically Challenged People,” 2013.
[2] Fernando Ramirez-Garibay, Cesar Millan Ollivarria, “MyVox—Device for Communication
Between People: Blind, Deaf, Deaf-Blind, and Unimpaired,” 2014.
[3] P. Kumari, S.R.N Reddy, “PiCam: IoT-Based Wireless Alert System for Deaf and Hard of
Hearing,” 2015.
[4] R. Suganya, T. Meeradevi, “Design of a Communication Aid for Physically Challenged,” 2015.
[5] L. Anusha, Y. Usha Devi, “Implementation of Gesture-Based Voice and Language Translator
for Dumb People,” 2016.
[6] Shraddha R. Ghorpade, Surendra K. Waghamare, “Full Duplex Communication System for
Deaf & Dumb People,” International Journal of Emerging Technology and Advanced
Engineering (IJETAE), Vol. 5, Issue 5, 2015.
[7]Chucai Yi, Yingli Tian, Aries Arditi, “Portable Camera-Based Assistive Text and Product
Label Reading from Hand-Held Objects for Blind Persons,” IEEE Transactions on Mechatronics,
2013.
[8] Vasanthi G., Ramesh Babu Y., “Vision-Based Assistive System for Label Detection with
Voice Output,” 2014.
[9] A. Singh, P. Patel, “Intelligent Gesture Recognition System for Deaf and Dumb Individuals,”
2018.
[10] J. Brown, L. Chen, “Voice-Assisted Smart Glasses for the Blind,” 2019.
[11] S. Kumar, V. Gupta, “Sign Language Translation via Neural Networks,” 2020.
[12] M. Lopez, A. Torres, “IoT-Enabled Alert System for Hearing Impaired Individuals,” 2021.
[13] K. Nakamura, H. Singh, “Low-Cost Braille e-Reader for the Visually Impaired,” 2022.
[14] K. Naveen Kumar, P. Surendranath, K. Shekar, “Assistive Device for Blind, Deaf and Dumb
People Using Raspberry Pi,” 2017.
[15] “OpenCV: Face Detection Tutorial,” [Online]. Available: https://docs.opencv.org/3.1.0/d7/d8b/
tutorial_py_face_detection.html.
[16] “Introduction to Tesseract OCR,” [Online]. Available: https://opensource.google.com.
[17] “Python Programming for Beginners,” [Online]. Available: https://
www.pythonforbeginners.com.
[18] “Xming: Free X Server for Microsoft Windows,” [Online]. Available: https://
www.macs.hw.ac.uk.
[19] M. Shankar, “IoT Applications in Assistive Technology: A Review,” Journal of Internet
Technologies, Vol. 12, 2020.
[20] J. H. Lee, “Adaptive Speech Recognition for Hearing Impaired Devices,” Proceedings of
AICT 2019.
[21] S. D. Roy, “Integration of Cloud-Based Assistive Technologies,” International Journal of
Smart Systems, 2021.
[22] M. K. Das, “Portable Communication Systems for Differently Abled,” Sensors and Actuators
B: Chemical, 2020.