Autonomous Surveillance Robot (ASaR)
Autonomous Surveillance Robot (ASaR)
https://doi.org/10.22214/ijraset.2020.30298
International Journal for Research in Applied Science & Engineering Technology (IJRASET)
ISSN: 2321-9653; IC Value: 45.98; SJ Impact Factor: 7.429
Volume 8 Issue VII July 2020- Available at www.ijraset.com
Abstract: India has a very large and complex land border. Thus, managing our borders with effective security monitoring
against intrusions, infiltrations and other potential threats becomes a challenging task. This project aims to implement a smart
border management system which will overcome these shortcomings and will help in securing our frontiers and safeguarding
our nation from the risks involved across the border. The overcoming of this problem will enable us to provide unmanned
surveillance at the borders as well as sensitive unattended critical sites, where the site can be remotely monitored such as
airports, seaports and industries. The method involves implementation of smart fences which consists of several switches, which
gets activated when any force is applied over the fence and alerts the control room. The robot runs over the rail structure which
is attached to the fence and is controlled by Raspberry Pi 4 development board. The robot consists of a PTZ camera, a fixed
camera, a LiDAR module, a DC motor which are interfaced with Raspberry Pi 4. Finally the target is detected by the smart
fence, LiDAR or motion detection using the method of image processing. The robot simultaneously alerts the control room and
continuously tracks the target until it moves away from the fence.
Keywords: Raspberry Pi 4, smart fence, motion detection, image processing, PTZ camera, fixed camera, LiDAR module.
I. INTRODUCTION
India has a large, complex land border and shares this border with seven other countries covering an area of approximately 15,106.7
km. In recent decades, a rising trend in smuggling, trafficking, crime, terrorism and illegal migration across the borders has been
noticed. As we know border management is a difficult task due to topology and challenging relations with some countries.
Surveillance systems provide the capability of collecting authentic and purposeful information and forming appropriate decisions to
enhance safety [1]. Video surveillance has a large market as the number of installed cameras around us can show. This work
introduces a new classification for smart video surveillance systems depending on their commercial applications [2]. Surveillance
robots can provide many benefits like greater efficiency and enhanced safety of security personnel. Thus, the current requirement is
a smart border management system which provides unmanned surveillance along with intrusion recognition, detection and tracking.
The method is applied for Red, Green and Blue (RGB) colour space. Metrically trimmed mean, Mean Absolute Deviation (MAD)
and chromaticity, brightness difference acts as estimators for background subtraction and shadow removal respectively. Deshmukh
et al. [9] proposes a system which consists of three layers of lasers, cameras and a central monitoring system. The system heavily
depends on extensive use of lasers and consists of three layer of lasers, running parallel behind each other. Background Subtraction,
frame differencing and
thresholding are used for intrusion detection. Finally, a Global System for Mobile communication (GSM) module is used for
sending notification to the end user. Ghanem et al. [10] presents a Raspbian operating system based spy robot platform with remote
monitoring and control algorithm through Internet of Things (IoT). This IoT application is developed by python, Hyper Text
Markup Language (HTML) and Javascript. The spy robot system comprises of Raspberry Pi, night vision pi camera and sensors.
The information regarding the detection of living objects by PIR sensor is sent to the users through the web server and pi camera
captures the moving object which is posted inside the webpage simultaneously. The user in the control room is able to access the
robot with wheel drive control buttons on the webpage. The movement of the robot is also controlled automatically through obstacle
detecting sensors to avoid collisions.
B. Pulley
The pulleys which will slide on the dual-rail structure were designed in SolidWorks and were manufactured on the lathe machine.
The material used for fabrication was Nylon. We used Nylon pulleys instead of Metal pulleys since we required pulleys which were
large in size so that we could slide the robot on the rail structure and since Nylon pulleys have less weight compared to the ordinary
Metal pulleys which in turn reduces the robot weight thereby increasing the speed with which the robot can move.
C. Smart Fence
The smart fence is equipped with switch which acts similar to a Single Pole Double Throw switch. SPDT switch has a single input
and can connect to switch between 2 outputs. Any force applied on the fence will cause the switch to oscillate. During oscillation,
the switch will come in contact with the ground plates and produce current signal which will be sent to the Raspberry Pi.
D. Internal Circuitry
Various components like L298N motor driver, 12V DC motor, fixed camera, PTZ camera, Lidar module are interfaced with
Raspberry Pi 4 for intrusion recognition, detection and tracking.
V. IMPLEMENTATION
A 3D model of the robot body was designed using Solidworks, for the purpose of balancing the robot on the rail, visualising the
correct shape of the body for placement of components and dimensions of construction. The body is implemented such that it can
balance on the rail holding necessary components.
B. Working
The robot is powered by a battery, which travels along a rail, which is attached to the smart fence. The Wi-Fi module ensures
constant connection between robot and control room. The fixed camera and the PTZ Camera are equipped with IR Illuminator
which is used for threat verification. Also the images and video captured by the cameras undergoes image processing for clear and
better images and videos. LiDAR is used for detection of the target, and to monitor constant changes across the fence.
C. Flowchart
Fig 6: Flowchart
As soon as the robot is powered using the battery, the robot starts sliding on the dual-rail structure, scanning the fence and two
operations starts simultaneously i.e.
1) Continuous scanning using LiDAR: While scanning the fence the robot checks if any target is detected or not using LiDAR. If
any target is detected the robot alerts the control room and keeps tracing the target else it keeps scanning the fence.
2) Smart fence is activated: As the smart fence is activated, any force applied on the fence will give an alert of intrusion to the
control room. As soon as force is applied on the fence the intrusion location is traced and the robot moves towards the intrusion
site.
For both above actions further decision will be taken by control room.
VI. RESULT
A. Motor
After interfacing the motor with Raspberry Pi 4, we could rotate the pulley connected to the motor’s shaft. This in turn resulted in
the movement of the robot body. The RPM of the motor was good enough to move the robot body.
B. LiDAR
After interfacing the Lidar with Raspberry Pi 4, we could calculate the distance of the object from the robot body, based on the time
the laser took to return from the object to the LiDAR module.
C. Smart Fence
After designing the switch of the smart fence, we could send data to the control room whenever force was applied on the fence. As
soon as force was applied the switch touched the ground plates and current signal was generated which was transferred to the
Raspberry Pi.
D. Camera
After interfacing the cameras with Raspberry Pi 4, we could capture live images and also live streaming was possible.
E. Image Processing
We successfully implemented the method of image processing using OpenCV library to the captured images and the recorded
videos in real time which helped us to detect the intruder using motion detection.
1) RGB Components: We successfully implemented the RGB colour model. An illustration which shows the captured image along
with the red, green, blue components on the screen is shown in Fig 7.
2) Continuous Motion Detection: We successfully implemented the method of continuous motion detection. An illustration of the
captured image and the processed image is shown in Fig 8 and Fig 9.
3) Current Motion Detection: We successfully implemented the method of current motion detection. An illustration of the
captured image and the processed image is shown in Fig 10 and Fig 11.
F. Prototype
After designing the 3-Dimensional model of the prototype, it was successfully implemented. The top section of the body has a
camera, a motor pulley attached to shaft of the motor. The centre section has shafts, pulleys, driving pulley connected to the motor
pulley with the help of a pulley belt. The bottom section has an inner compartment for holding the battery.
On one of the centre shaft three pulleys are placed. The third pulley and the driving pulley are connected with a help of a belt. As
soon as the motor is powered the pulley on the motor starts rotating which in turn will rotate the driving pulley on the centre shaft
and help the robot to slide on the rail.
VII. CONCLUSION
A wireless automatic outdoor intruder security system was developed which runs on the dual-rail structure and responds to the
intrusion alerts. The LiDAR module used for detection constantly monitors the changes across the fence and gives us the accurate
distance between the intruder and the robot. The smart fence detects the intrusion whenever any force is applied on the fence. The
intruder could be also detected by motion detection using image processing. The 12V DC motor allows us to control the speed of the
robot. The robot designed is applicable for security purpose at borders, sensitive unattended critical sites such as airports, seaports,
industries, farms and it can be also used in societies. Thus, a smart border management system which provides unmanned
surveillance along with intrusion recognition, detection and tracking was successfully designed and implemented.
IX. ACKNOWLEDGEMENT
It is our privilege to express our sincerest regards to our project guide, Prof. Devendra Sutar, Assistant Professor, Goa College of
Engineering for his valuable inputs, able guidance, encouragement, whole-hearted cooperation and constructive criticism throughout
the duration of project. We deeply extend our sincere thanks to Dr. Krupashankara M. S., Principal, Goa College of Engineering,
Dr. H.G. Virani, Head of Department, E & TC Dept. and Dr. Akshay V. Nigalye, Workshop Superintendent, Mechanical Dept. for
encouraging and allowing us to present and conduct the project at their premises for partial fulfilment of the requirements for the
project. We take this opportunity to thank all our teachers, laboratory assistants and non teaching staff who have directly or
indirectly helped our project. We pay our respects and love to our parents and all other family members and friends for their love
and encouragement throughout our career.
REFERENCES
[1] Tomi D Räty, “Survey on contemporary remote surveillance systems for public safety”, IEEE Transactions on Systems, Man, and Cybernatics, Part C
(Applications and Reviews), 40(5), pp. 493-515, IEEE, 2010.
[2] Mohamed H Sedky, Mansour Moniri, Claude C Chibelushi, “Classification of smart video surveillance systems for commercial applications”, IEEE
Conference on Advanced Video and Signal Based Surveillance, pp. 638-643, IEEE, 2005.
[3] Arjun D, Indukala P K, K A Unnikrishna Menon, “Border surveillance and intruder detection using wireless sensor networks: A brief survey”, International
Conference on Communication and Signal Processing (ICCSP), pp.1125-1130, IEEE, 2017.
[4] Palagati Harish, R.Subhashini, K.Priya, “Intruder Detection by Extracting Semantic Content from Surveillance Videos”, International Conference on Green
Computing Communication and Electrical Engineering (ICGCCEE), pp. 1-5, IEEE, 2014.
[5] Jisha R C, Maneesha V. Ramesh, Lekshmi G S, “Intruder tracking using wireless sensor network”, IEEE International Conference on Computational
Intelligence and Computing Research (ICCIC), pp. 1-5, IEEE, 2010.
[6] Sagar R N, Sharmila S P, Suma B V, “Smart Home Intruder Detection System”, International Journal of Advanced Research in Computer Engineering and
Technology (IJARCET), 6(4), pp. 439-443, 2017.
[7] Dushyant Kumar Singh, Dharmender Singh, “Automatic Intruder Combat System: A way to Smart Border Surveillance”, Defence Science Journal, 67(1), pp.
50-58, 2016.
[8] Jianhua Ye, Tao Gao, Jun Zhang, “Moving object detection with background subtraction and shadow removal”, 2012 9th International Conference on Fuzzy
Systems and Knowledge Discovery (FSKD 2012), pp. 1859-1863, IEEE, 2012.
[9] R. A. Deshmukh, Shubham Kamdi, Mohak Pingle, Shivani Rajebhosale, Asawari Bhosale, “Intelligent surveillance system using energy efficient intrusion
detection and tracking techniques”, 2nd International Conference on Electronics, Communication and Aerospace Technology (ICECA), pp. 1214-1218, IEEE,
2018.
[10] Ghanem Osman Elhaj Abdalla, T. Veeramanikandasamy, “Implementation of Spy Robot for A Surveillance System using Internet Protocol of Raspberry Pi”,
2nd International Conference on Recent Trends in Electronics, Information and Communication Technology (RTEICT), pp. 86-89, IEEE, 2017.