Chat GPT DOCUMENT
Chat GPT DOCUMENT
i
ABSTRACT
ii
TABLE OF CONTENTS
ABSTRACT iv
LIST OF TABLES ix
LIST OF FIGURES x
ABBREVATION xi
1. INTRODUCTION 1
2. EMBEDDED SYSTEMS 2
2.1 Characteristics 3
2.2 User interface 3
2.3 Processors in embedded systems 3
2.4 Peripherals 4
2.5 Embedded software architectures 5
2.5.1.Simple control loop 5
2.5.2.Interrupt-controlled system 5
2.5.3.Cooperative multitasking 5
2.5.4 Preemptive multitasking or 6
Multi-threading
2.6 Applications 6
3. LITERATURE REVIEW 8
3.1 Introduction 8
3.2 Existing Work 8
3.3 Proposed Work 13
iii
4. HARD WARE DESCRIPTION 14
4.1 Block Diagram of Online Doctor Consultant 14
4.1.1 Hardware Requirements 14
4.1.2 Software Requirements 16
4.1.3 Operation 16
4.2 PIC Microcontroller 18
4.2.1 Introduction 18
4.2.2 Basic Features of PIC Microcontroller 19
4.2.3 Pin Descriptions 20
4.2.4 Applications of PIC Microcontroller 25
4.3 Peripheral Features 31
4.3.1 Advantages 32
4.4.2 Disadvantages 32
iv
4.5.4 Timing diagram 35
4.5.5 Theory of Operation 35
4.5.6 Circuit Diagram of Oxygen Sensor 37
4.5.7 Advantages of Oxygen Sensor 41
4.5.8 Disadvantages of Heartbeat Sensor 41
4.5.9 Applications 42
5 PROJECT OVERVIEW 53
6 PROGRAM 54
v
7 RESULT AND DISCUSSION 59
8 APPLICATIONS 61
9 FUTURE SCOPE 62
10 CONCLUSION 63
11 REFERENCES 64
vi
LIST OF TABLES
vii
LIST OF FIGURES
IO - Input Output
US sensor - Ultrasonic sensor
SRAM - Static Random Access Memory
EEPROM - Electrically Erasable Programmable Read Only
Memory
FTDI - Future Technology Devices International
RX - Receiver
TX - Transmitter
USB - Universal Serial Bus
PWM - Pulse Width Modulation
SPI - Serial Peripheral Interface
SS - Stack Segment register
MOSI - Master Out Slave In
MISO - Master In Slave Out
SCK - Serial Clock
SDA - Serial Data
SCL - Serial Clock
PWI - Precision Winding
I2C - Inter Integrated Circuit
AREF - Analog Reference
OS - Operating System
RISC - Reduced Instruction Set
Computer
AVR - Advanced Virtual RISC
DTR - Data Terminal Ready
ix
ADC - Analog to Digital Converter
DAC - Digital to Analog Converter
UART - Universal Asynchronous Receiver /
Transmitter USART - Universal Synchronous / Asynchronous
Receiver /
Transmitter
MIPS - Million Instructions Per Second
PDIP - Payload Data Interface Panel
TQFP - Thin Quad Flat Pack
RTC - Real – Time Clock
BLDC - Brushless DC
SNR - Signal – to – Noise Ratio
BPS - Bits Per Second
MCU - Microcontroller Unit
x
1. INTRODUCTION
In today’s world, technology plays an essential role in making life easier for people
across different sections of society. However, despite all advancements, there
remains a significant digital divide for people with disabilities, particularly those
with visual impairments. For the visually impaired community, accessing digital
content, interacting with smart devices, navigating new environments, and
performing daily tasks independently remain considerable challenges. Traditional
aids like walking canes, Braille literacy systems, guide dogs, and audio books have
helped to some extent, but they come with limitations in adaptability, affordability,
and independence.
The advent of Artificial Intelligence (AI), conversational agents, and the Internet of
Things (IoT) has opened up immense possibilities to bridge this accessibility gap. AI-
powered models, especially language models like ChatGPT, demonstrate the ability
to understand and generate human-like conversations, providing opportunities for
natural and intuitive communication interfaces. When combined with voice
assistance and connected IoT devices, such technologies can create a supportive
ecosystem where visually impaired individuals can receive real-time information,
navigate physical spaces, and interact seamlessly with their environment.
The motivation behind this project lies in creating a cost-effective, intelligent, and
portable hardware system that leverages ChatGPT, voice recognition, IoT, and smart
sensing to offer real-time guidance, interaction, and support. A device capable of
understanding and responding naturally can empower visually impaired individuals
with unprecedented autonomy, enhancing their overall quality of life, fostering
inclusion, and enabling them to participate actively in the digital world.
While various assistive technologies exist, most current systems are either too
expensive, too bulky, require complex interactions, or offer only limited
functionality. Many voice assistants available today, such as Siri, Alexa, or Google
Assistant, are highly dependent on cloud services and cannot function effectively
1
offline or in low-connectivity environments. Furthermore, they are often designed for
general users, lacking the specific adaptations needed for the visually impaired.
Visually impaired individuals not only need voice-based interaction but also require
real-time awareness of their surroundings, reliable navigation aids, and control over
their IoT-enabled environments. Existing solutions often fail to combine all these
elements into a single, integrated device that can provide continuous, natural, and
context-aware assistance. The specific problem addressed by this project is the lack
of an affordable, standalone, conversational AI-powered device that supports voice
interaction, environment sensing, IoT control, and real-world situational assistance
for visually impaired users. Developing such a system requires a careful integration
of natural language processing, real-time hardware interfacing, low-power design,
and user-centric interaction models.
2. Problem Statement
Despite the existence of various assistive devices, most current solutions are limited
by high costs, limited functionality, complexity of use, or lack of real-time
2
adaptability. Devices often require tactile input or are restricted to predefined tasks,
thus limiting user flexibility. Visually impaired users face difficulties in navigating
unknown environments, accessing real-time information, and controlling smart
devices independently.
There is a critical need for a comprehensive, voice-driven, AI-enabled hardware
solution that can bridge the gap between the digital world and the real needs of the
visually impaired community. Such a system must be portable, affordable, and
capable of understanding natural human language commands while providing smart
feedback and control over IoT devices.
3
IoT Connectivity: Enabling communication between the device and other
smart devices or sensors over Wi-Fi, Bluetooth, or cellular networks.
Voice Recognition and Synthesis: For converting spoken commands into
machine instructions and reading out responses.
Obstacle Detection and Navigation Assistance: Using sensors like
ultrasonic, infrared, or LIDAR to enhance mobility.
Hardware Design: Building a compact, low-power hardware platform
suitable for wearable or portable use cases.
AI-Based Adaptation: Learning user preferences and frequently used
commands to personalize the experience over time.
6. Technology Overview
ChatGPT Integration: ChatGPT serves as the core brain behind the system’s
understanding and conversation generation. Fine-tuned models can be
employed for better handling of common queries and specific user needs.
4
Hardware Components: Microcontrollers like ESP32, Raspberry Pi Zero, or
custom boards are utilized along with microphones, speakers, camera modules,
and sensors.
Voice Processing: Open-source speech-to-text and text-to-speech engines are
deployed for real-time interaction.
IoT Framework: MQTT, HTTP, or CoAP protocols are employed for data
communication across IoT networks.
8. Methodology
The methodology adopted for this project includes:
Requirement Analysis: Studying the needs of visually impaired users.
System Architecture Design: Designing both the software and hardware
frameworks.
Prototyping: Developing initial models for testing different functions like
voice command processing, obstacle detection, and IoT control.
Testing and Validation: Field-testing the prototypes in real-world
environments.
Iteration and Enhancement: Incorporating feedback to improve system
functionality and reliability.
5
The project report is organized as follows:
Chapter 1: Introduction and background
Chapter 2: Literature review on related works
Chapter 3: System design and architecture
Chapter 4: Hardware and software implementation
Chapter 5: Testing, results, and evaluation
Chapter 6: Conclusion and future enhancements
This project holds immense significance for social, technological, and humanitarian
reasons. For the visually impaired community, it promises to deliver a new degree of
independence by allowing users to control their environment, receive situational
assistance, and interact with information systems naturally and intuitively. By
removing barriers to access, the solution aims to enhance the autonomy, dignity, and
overall quality of life of visually impaired individuals.
Technologically, the project pushes the boundaries of how AI models like ChatGPT
can be utilized beyond traditional chatbots or virtual assistants, embedding them into
hardware for physical-world applications. It also demonstrates how AI and IoT
convergence can deliver smart, adaptive systems capable of real-time decision-
making and communication.
From a humanitarian perspective, ensuring accessibility to modern technology is
crucial for creating an inclusive society where disabilities are not viewed as
limitations. The project's cost-effectiveness makes it scalable and replicable, opening
possibilities for deployment across different geographies, particularly in
underdeveloped or resource-constrained regions where commercial assistive
technologies are often out of reach.
6
2. EMBEDDED SYSTEMS
7
2.1 Characteristics
8
Ordinary microprocessors (μP): Use separate integrated
circuits for memory and peripherals.
Microcontrollers (μC): Have on-chip peripherals, thus
reducing power consumption, size and cost.
2.4 Peripherals
Embedded systems talk with the outside world via peripherals, such as:
9
2.5 Embedded software architectures
10
2.5.4 Preemptive multitasking or multi-threading
2.6 Applications
11
user. Computer networking uses dedicated routers and network
bridges to route data.
13
preparing it for AI-based interpretation.
ChatGPT Integration: Handling the communication pipeline between the
embedded device and the ChatGPT language model (either via cloud API or a
localized lightweight inference engine), interpreting the user's commands, and
generating appropriate responses.
Speech Synthesis: Converting the AI-generated text response back into
speech output that can be clearly understood by the user.
Obstacle Detection and Alerting: Reading distance data from ultrasonic or
infrared sensors, identifying potential hazards, and immediately notifying the
user via voice or sound alarms.
IoT Device Control: Sending commands to IoT-enabled devices (smart lights,
doors, alarms, etc.) via Wi-Fi or Bluetooth communication protocols, allowing
users to manage their surroundings effortlessly.
Emergency Handling: Triggering SOS alerts by sending messages through
GSM modules or Wi-Fi networks when emergencies are detected or manually
initiated by the user.
14
2.11 Power Management in Embedded Systems
Since the device is intended for portable use, low power consumption is a priority.
Embedded systems are inherently energy-efficient and are optimized further by
implementing techniques like:
Using sleep and deep sleep modes when the device is idle.
Activating sensors only when necessary.
Reducing the transmission frequency for non-critical data.
Selecting low-power wireless protocols (e.g., BLE over classic Bluetooth).
Battery-operated embedded platforms are carefully chosen and configured to provide
several hours of uninterrupted operation, ensuring the device remains practical for
everyday use.
2.12 Advantages of Embedded Systems in This Project
The decision to use embedded systems architecture brings multiple benefits to the
project:
Portability: The small form factor enables the entire system to be integrated
into wearable or handheld devices.
Energy Efficiency: Ensures long operational life even on limited battery
resources.
Cost-Effectiveness: Off-the-shelf microcontrollers and modules significantly
reduce the overall project cost.
Customization: Embedded programming allows fine-tuning features
according to user feedback and evolving requirements.
Reliability: Rugged and stable operation suitable for continuous use in diverse
environments.
15
identification using vision modules, dynamic real-time location tracking, and even
predictive analytics for user health monitoring. By leveraging advances in embedded
AI, Edge computing, and low-power communications, next-generation assistive
devices could become even smarter, more intuitive, and more personalized for each
individual user's needs.
16
3. LITERATURE REVIEW
3. Design of Smart Assistive Systems for Blind People Using Machine Learning
(2021)
The researchers developed a lightweight machine learning model capable of running
on microcontrollers (e.g., STM32, ESP32) to classify nearby objects like cars,
people, and trees. Using datasets of obstacle images, they trained a compact
convolutional neural network (CNN) and implemented it on a wearable device. This
advancement significantly enhanced environmental awareness. The study also
pointed out the importance of continuous learning and updating models to adapt to
changing urban environments, suggesting future devices should incorporate OTA
17
(Over-the-Air) learning updates.
18
7. A Study on Speech Synthesis Techniques for Assistive Devices (2018)
This study discussed different methods for generating synthetic speech in assistive
devices. Techniques like concatenative synthesis (joining pre-recorded speech
samples) and parametric synthesis (using mathematical models) were compared.
Researchers concluded that parametric methods, particularly deep neural network-
based speech synthesis, offer superior adaptability and naturalness for embedded
assistive systems. In the context of this project, high-quality speech output is crucial
for delivering intelligible and emotionally neutral communication to the visually
impaired.
19
10. Real-Time Object Detection for Visually Impaired Assistance Using Embedded
Vision (2022)
A system using low-power CNNs like MobileNetV2 was proposed to identify objects
such as stairs, vehicles, and pedestrians in real time. Implemented on devices like
Raspberry Pi and NVIDIA Jetson Nano, the object detection systems provided audio
feedback on nearby hazards. However, the study noted that energy consumption was
a serious constraint, suggesting further exploration into event-based cameras and
neuromorphic processors for ultra-low power real-time vision.
20
summaries, and conducting verbal quizzes. The chatbot was fine-tuned to accept
voice inputs and deliver audio responses with a human-like tone. Integration with
cloud platforms like Dialogflow and GPT-3 improved the richness of interaction.
Such findings prove the potential of conversational AI in the broader landscape of
assistive devices beyond basic navigation.
15. Emergency Alert Systems for Visually Impaired Using GSM and IoT (2021)
A GSM and IoT-based emergency alert system was developed where visually
impaired users could trigger an alert by pressing a hidden button or speaking a
predefined word. An SMS or IoT notification would immediately be sent to
caregivers, along with location data if available. The study stressed that having multi-
channel emergency communication (voice, GSM, IoT cloud) greatly increases
reliability and ensures that users can seek help even if one network fails.
21
3.1 Existing Work
Assistive technologies for the visually impaired have witnessed significant growth
over the past two decades, with major advancements in voice-based systems, obstacle
detection, and IoT-based navigation aids. Existing solutions largely focus on
providing support through smart canes, smartphone applications, wearable sensors,
and voice-enabled devices. Each of these innovations has aimed to bridge the gap
between the visually impaired and their environment by enhancing mobility, safety,
and access to information.
One of the earliest and most widely adopted technologies was the smart walking
stick, embedded with ultrasonic sensors to detect nearby obstacles. When an object is
detected within a certain range, the stick vibrates or emits a sound, warning the user
of the obstruction ahead. While effective in static environments, these systems often
struggle in dynamic conditions like crowded urban areas where real-time decision-
making and adaptive navigation are crucial.
Another major stream of development has been voice-assisted smartphones and
screen readers such as TalkBack (Android) and VoiceOver (iOS). These tools
convert text to speech and allow visually impaired users to navigate their devices
independently. While powerful, their dependence on a stable internet connection for
full functionality often limits their use in rural or low-connectivity regions.
Moreover, these solutions are often not deeply integrated with physical navigation
support or obstacle detection in real-world environments.
IoT-based solutions have further enhanced accessibility by introducing real-time
monitoring and assistance. Some devices send location data to caregivers or
emergency services when needed. However, these systems often lack robust AI
capabilities for predictive assistance or complex environment interpretation.
Moreover, most existing IoT devices operate on basic data transmission models
without intelligent contextual processing.
Voice-activated home automation systems have also emerged, enabling visually
impaired individuals to control household appliances through simple voice
22
commands. Integration with smart hubs like Amazon Alexa and Google Assistant has
made life easier indoors. Nevertheless, these systems are not specifically designed
with visually impaired users in mind and may sometimes require visual
confirmations or interactions that are not accessible.
Recent works involving AI-based object recognition have begun to demonstrate the
potential of machine learning in assistive devices. Systems using lightweight CNN
models can now identify objects like traffic signals, stairs, and moving vehicles.
However, running such computationally intensive models on low-power embedded
hardware remains a major challenge. While platforms like Raspberry Pi or Jetson
Nano can support these models, they often lack the battery efficiency and
compactness required for wearables.
Chatbot integration and voice-based AI assistance have shown promise for
improving communication and access to information. However, most chatbot
systems are cloud-dependent, requiring stable internet access and considerable
processing power, which limits their portability and real-time responsiveness in
mobile or wearable devices. Additionally, current systems are limited in their
contextual awareness and cannot fully understand or predict the complex needs of
visually impaired users while navigating real-world environments.
Despite these advancements, current systems exhibit several limitations:
Limited integration of AI-driven real-time decision-making.
Dependency on internet connectivity for core functionality.
Lack of highly optimized embedded AI models suitable for wearable devices.
Minimal proactive support based on environmental context.
No single compact system that combines voice assistance, AI-based object
recognition, IoT data transmission, and emergency response.
These gaps highlight the need for a comprehensive, hardware-based embedded
system that combines ChatGPT-level natural language processing, real-time
obstacle detection, voice-based assistance, and IoT connectivity in a lightweight,
portable, and energy-efficient device specifically tailored for visually impaired users.
23
The existing research serves as a critical foundation, but there remains a vast scope
for innovation by integrating AI, Embedded Systems, Speech Recognition, and IoT
into a single, intelligent wearable platform—which this project aims to achieve.
25
4. HARD WARE DESCRIPTION
26
TEXT TO SPEECH
28
Output is played through a speaker or headphone connected to the hardware.
Summary
In this project, Text-to-Speech technology is not just an add-on but a core
communication bridge between the hardware AI system and the visually impaired
user. By effectively delivering audible responses, TTS empowers users to interact
naturally with their environment, making technology more inclusive, accessible,
29
and human-centered.
SPEECH TO TEXT
30
Purpose of STT in the Project
Voice Command Input:
The user can control connected IoT devices (e.g., turn on/off lights, fans) simply by
speaking.
Interaction with ChatGPT AI:
Users can ask questions, give instructions, or seek assistance, and their voice is
converted into text that the AI can understand and process.
Obstacle and Navigation Queries:
Users can ask about nearby obstacles, navigation help, or environmental conditions
(e.g., "Is there an object ahead?").
Emergency Commands:
Quick emergency voice commands (e.g., "Call for help") can trigger predefined
actions like sending alerts to caregivers.
31
Working Flow:
User speaks a command or question.
Microphone captures the audio input.
STT engine processes the speech and converts it into text.
The text is either interpreted as a command or sent to the ChatGPT AI.
A corresponding action or response is triggered.
32
Summary
In this project, Speech-to-Text technology transforms the user’s spoken language
into actionable digital commands, making interaction with AI and IoT devices
seamless, efficient, and accessible for visually impaired individuals. Combined
with TTS, it creates a full voice-based two-way communication system that
eliminates the need for screens, keyboards, or buttons, enhancing independence and
quality of life.
ESP32 Microcontroller
The ESP32 microcontroller plays a central role in the development of the proposed
hardware ChatGPT with Voice Assistance and AI–IoT system for visually
impaired users. The ESP32 is a powerful, low-cost, dual-core microcontroller with
built-in Wi-Fi and Bluetooth connectivity, making it ideal for real-time data
communication, IoT integration, and wireless voice command processing. In this
project, the ESP32 is responsible for managing all sensor inputs, including
ultrasonic sensors for obstacle detection, GPS modules for location tracking, and
environmental sensors for enhanced situational awareness. Its multi-tasking
capability allows it to handle simultaneous operations such as running lightweight
AI models for obstacle recognition, processing voice commands, and
communicating with cloud platforms like ThingSpeak or AWS IoT.
Moreover, the ESP32’s high processing speed (up to 240 MHz) and large SRAM
(520 KB) enable it to support embedded AI functionalities such as keyword
spotting for offline voice recognition, and basic natural language understanding to
respond intelligently to user queries. The microcontroller’s GPIO flexibility
facilitates the integration of various output components like vibration motors,
buzzers, and voice output modules, ensuring timely and appropriate feedback to the
user.
33
The device's low-power modes are particularly beneficial in designing a wearable
solution, allowing the system to conserve battery life during idle periods and
extend operational time. Additionally, the dual Wi-Fi and Bluetooth functionality
ensures that the device can both connect to IoT cloud servers for real-time
monitoring and communicate locally with smartphones or emergency systems
without any extra hardware. In case of network unavailability, the ESP32 can
operate in offline mode using its onboard processing capabilities, maintaining
essential services like obstacle warnings and voice interactions.
Overall, the ESP32 microcontroller offers the perfect balance of processing power,
connectivity, low energy consumption, and versatility, making it an indispensable
component for creating an intelligent, reliable, and efficient assistive device for
visually impaired individuals in this project.
Power Source
The system is powered using a regulated power supply, typically 5V or 3.3V
depending on the sensor and microcontroller requirements. The power source may
include a DC adapter, battery, or rechargeable power bank, depending on
deployment conditions. Ensuring an uninterrupted power supply is essential for
continuous monitoring and communication in industrial automation systems.
Backup power options like batteries may be used for wearable components or GSM
alerts during power failures.
5.PROJECT OVERVIEW
The "Design of Hardware ChatGPT with Voice Assistance and AI–IoT for
Visually Impaired" project seeks to develop a wearable assistive device that
combines cutting-edge technologies like AI-based natural language processing
(NLP), IoT integration, voice assistance, and real-time environmental sensing to
help visually impaired individuals navigate and interact with their surroundings
independently. This device will provide a seamless and intelligent system to aid
34
users in daily activities, offering both safety and autonomy through the
combination of hardware and software solutions.
The project integrates a hardware platform built around the ESP32 microcontroller,
known for its low energy consumption, dual-core processing power, and built-in
Wi-Fi and Bluetooth capabilities. This allows the device to communicate in real-
time with other devices, such as smartphones or cloud-based servers, enabling
remote monitoring and emergency alerting. The ESP32 is complemented by a
series of environmental sensors including ultrasonic sensors for obstacle detection,
GPS for location tracking, and voice feedback systems for direct interaction with
the user.
The system aims to empower visually impaired users with AI-driven voice
assistance, enabling them to interact with the device using natural language. By
integrating ChatGPT-level language models, the device can answer questions,
provide contextual navigation support, and offer real-time information about the
environment, such as identifying objects, detecting nearby people, and providing
directions. This voice-driven interface will eliminate the need for manual controls,
giving the user hands-free access to critical information.
In addition to real-time navigation and environmental assistance, the system
supports IoT-based monitoring and emergency alerting. The device is capable of
sending distress signals and user location details to caregivers or emergency
contacts in case of sudden incidents such as falls or distress. Moreover, the device’s
location-based features allow for dynamic adjustments to guidance depending on
the user’s surroundings, such as providing specific alerts when approaching street
crossings or obstacles.
The device also features offline functionality, ensuring that it continues to operate
effectively in areas with poor or no internet connectivity. The system utilizes local
processing capabilities of the ESP32 and lightweight AI models to run essential
features such as obstacle detection and voice interaction without needing constant
cloud access.
35
Ultimately, the Project Overview aims to create a compact, energy-efficient, multi-
sensor wearable device that enhances the mobility, safety, and independence of
visually impaired users. It combines AI-powered assistance, real-time
environmental awareness, and IoT communication in one integrated solution,
bridging the gap between users and their surroundings. This system seeks to
empower the visually impaired by providing them with tools that foster confidence,
autonomy, and access to information in their daily lives.
6. PROGRAM
// Necessary Libraries
#include "Audio.h"
#include "CloudSpeechClient.h"
int i = 0;
void setup()
{
pinMode(button, INPUT);
pinMode(led_1, OUTPUT);
pinMode(led_2, OUTPUT);
pinMode(led_3, OUTPUT);
36
Serial.begin(115200);
Serial2.begin(115200, SERIAL_8N1, RXp2, TXp2);
Serial2.println("Intialising");
// Serial.println(My_Data);
}
void loop()
{
digitalWrite(led_1, 0);
digitalWrite(led_2, 0);
digitalWrite(led_3, 0);
if (i == 0) {
Serial.println("Press button");
i = 1;
}
// if(i==1){delay(1);}
delay(500);
if (digitalRead(button) == 0)
{
Serial2.println("\r\nPlease Ask!\r\n");
// Green LED ON
digitalWrite(led_1, 1);
digitalWrite(led_2, 0);
digitalWrite(led_3, 0);
delay(2100);
Serial.println("\r\nRecord start!\r\n");
Audio* audio = new Audio(ADMP441);
audio->Record();
Serial.println("Processing your Audio File");
// Blue LED ON
digitalWrite(led_1, 0);
digitalWrite(led_2, 1);
digitalWrite(led_3, 0);
/*
This is the code for Text to Speech conversion of
String respose coming from other ESP32 board.
If you still facing any issues, kindly watch out our video about this project on
our YouTube channel
YouTube channel - https://www.youtube.com/techiesms
*/
#include "Arduino.h"
#include "WiFi.h"
#include "Audio.h" // Download this Library ->
https://github.com/schreibfaul1/ESP32-audioI2S
#define uart_en 15
#define RXp2 16
#define TXp2 17
#define I2S_DOUT 25
#define I2S_BCLK 27
#define I2S_LRC 26
38
Audio audio;
void setup()
{
Serial.begin(115200);
Serial2.begin(115200,SERIAL_8N1, RXp2, TXp2);
WiFi.disconnect();
WiFi.mode(WIFI_STA);
WiFi.begin( ssid, password);
void loop()
if (Serial2.available()) {
String Answer = Serial2.readString();
//-----
// Split the answer into chunks and send each chunk to connecttospeech
size_t chunkSize = 80; // Define chunk size (adjust if necessary)
for (size_t i = 0; i < Answer.length(); i += chunkSize) {
while(audio.isRunning()){
audio.loop();
}
//--------
}
39
}
audio.loop();
SOFTWARE DESCRIPTION
The Software System for the "Design of Hardware ChatGPT with Voice Assistance
and AI–IoT for Visually Impaired" integrates multiple layers of technologies to
ensure seamless functionality, from sensor data processing to voice interaction and
cloud communication. The software system is divided into several components,
each responsible for specific functions such as sensor data acquisition, AI
processing, voice recognition, and cloud connectivity. Below, we provide a
detailed description of the software components and how they work together to
deliver a fully integrated system for visually impaired users.
41
Vibration feedback through motors to alert the user of nearby obstacles in their
immediate path.
Emergency notification: If a critical obstacle is detected or a fall is detected
(through additional sensors), the system sends an SMS alert or push notification to
caregivers or emergency contacts via SMS API or IoT cloud platforms.
42
models can leverage services like Google Speech API or Amazon Transcribe. The
speech-to-text engine provides the text that is then processed by the AI system.
b. Natural Language Understanding (NLU)
Once the speech is converted to text, the next step is natural language
understanding. This component identifies the intent behind the command or query,
whether it’s a request for navigation, obstacle detection, location updates, or
general information. The ChatGPT API or a customized NLP model processes the
input to generate an appropriate response.
For example, if the user says, “What’s around me?”, the AI will respond with
information gathered from the sensors, such as “There is a wall 1 meter ahead,” or
“You are near a bus stop.”
c. Text-to-Speech (TTS)
Once the AI system processes the query or command, the result is then converted
back to speech using Text-to-Speech (TTS) technology. The response is spoken
aloud to the user through an onboard speaker or earphones. The system uses TTS
services such as eSpeak or Google Text-to-Speech (for cloud-based options) to
provide clear and natural responses.
43
4. System Integration
All software components work in concert to deliver a seamless experience for the
user:
Real-time data processing occurs on the ESP32, ensuring that obstacle detection,
voice interaction, and IoT communication are handled locally or via cloud-based
services as needed.
The cloud integration ensures that data is logged and shared in real-time with
emergency contacts or caregivers, enabling remote monitoring.
AI-powered voice assistance provides users with interactive feedback, navigation
aids, and real-time object identification, ensuring that the device offers dynamic,
personalized support.
8. APPLICATIONS
44
powered voice assistance.
6. Healthcare Monitoring
With additional health sensors (like heart rate, temperature, or fall detection),
the system can monitor the user’s vital signs and alert medical personnel if
45
necessary.
Reduces dependency on hospital visits by providing real-time health
monitoring at home.
7. Educational Support
Can assist visually impaired students by reading out study material,
recognizing text, and guiding them in academic environments like
classrooms, libraries, or campuses.
Supports voice-based learning modules and information access.
Thus, the applications of this system span across various fields, making life easier,
46
safer, and more independent for visually impaired users. It leverages the power of
AI, IoT, and embedded technology to create a truly smart assistive device that
brings practical benefits to both individuals and society as a whole.
47
9.FUTURE SCOPE
The future of IoT in industrial automation and security holds immense potential as it
continues to evolve and transform industries globally. With advancements in IoT
devices, sensor technologies, machine learning, artificial intelligence, and cloud
computing, the applications of IoT will further extend into more advanced and
intelligent systems that enhance operational efficiency, safety, and sustainability.
Below are several key areas where IoT will continue to have a significant impact in
the future:
49
Additionally, 5G will facilitate the implementation of smart factories, where
machines, robots, sensors, and workers are all interconnected in a real-time
ecosystem, allowing for more efficient and flexible production lines. Enhanced
communication networks will support the massive number of IoT devices expected to
be deployed, enabling the seamless integration of different systems and technologies
across industries.
50
contractual agreements. This will enhance trust between parties, reduce fraud, and
streamline operations in industries that rely on secure data exchanges.
51
promising. As technologies like AI, 5G, blockchain, and advanced sensors continue
to evolve, IoT systems will become more autonomous, intelligent, and efficient. The
integration of IoT in industrial settings will lead to smarter factories, safer working
environments, and more sustainable industrial practices. As industries embrace these
innovations, IoT will remain a cornerstone in the future of industrial automation,
transforming the way businesses operate and interact with their environments.
52
10.CONCLUSION
The project titled “Design of Hardware ChatGPT with Voice Assistance and
AI–IoT for Visually Impaired” presents a significant step forward in the
development of intelligent, assistive technologies for individuals with visual
impairments. Through the integration of AI, IoT, embedded systems, and
voice-based communication, this system provides real-time, interactive support
that enhances user mobility, independence, and confidence.
By using the ESP32 microcontroller, voice recognition, and ChatGPT-based
conversational AI, the system is capable of offering natural, human-like
interactions. It allows users to receive auditory guidance, detect obstacles,
request information, and control IoT devices seamlessly. The use of sensors and
real-time data processing ensures that users are kept safe from immediate
environmental hazards, while the IoT capabilities enable communication with
caregivers or emergency services if needed.
The project's hardware design is compact, portable, and adaptable, making it
suitable for daily wear in different environments, from homes to public spaces.
The integration of smart features like obstacle detection, voice feedback, and
remote monitoring provides a complete solution addressing both safety and
convenience.
Furthermore, the future scope of the system is immense, with opportunities to
integrate advanced AI vision, GPS navigation, health monitoring, machine
learning personalization, and smart home control. These enhancements would
elevate the system from an assistive tool to a comprehensive lifestyle support
device.
In conclusion, the Hardware ChatGPT with Voice Assistance and AI–IoT for
Visually Impaired project not only addresses a crucial societal need but also
demonstrates how modern technology can bridge the accessibility gap. It
empowers visually impaired individuals to lead more independent, safer, and
53
fuller lives, showcasing the transformative potential of AI and IoT in creating
an inclusive world.
54
11.REFERENCES
56
3(3), 301-314. doi:10.1142/S2424906418400137.
21.Zhang, X., & Lin, C. (2020). IoT-Enabled Smart Factory: Challenges,
Applications, and Future Directions. Journal of Manufacturing Science and
Engineering, 142(2), 021014. doi:10.1115/1.4045880.
22.Samaras, S., & Dooley, L. (2018). Challenges and Opportunities in IoT for
Smart Cities and Industries. International Journal of Internet of Things and
Cyber-Assurance, 1(1), 1-10.
23.Mahmoud, M. M., & Elmogy, M. (2017). Security Challenges in Industrial
IoT Applications. The 13th International Symposium on Autonomous
Decentralized Systems (ISADS). doi:10.1109/ISADS.2017.52.
24.Islam, S. R., & Kwak, D. (2015). The Internet of Things for Smart Cities:
Applications, Opportunities, and Challenges. IEEE Access, 3, 1886-1899.
doi:10.1109/ACCESS.2015.2440935.
25.Das, S., & Banerjee, M. (2021). Real-Time Industrial Automation System
Based on IoT. Proceedings of the 2021 IEEE International Conference on
Smart Devices, Circuits, and Systems.
57