Local cover image
Local cover image
Image from OpenLibrary
Custom cover image
Custom cover image

Wireless blind-spot monitoring system with real-time motion detection and risk estimation using yolov11 and optical flow on nvidia jetson orin nano for toyota hilux 2.4 fx 4x2 mt (w/rear a/c)/ Sarah Bint Mubarak F. Al-hajri, Elizhea C. Camonias, Jason D. Del Rosario, Alexandre Jr. Dela Cruz, and Ivan James Gabriel R. Yray.--

By: Contributor(s): Material type: TextTextPublication details: Manila: Technological University of the Philippines, 2025.Description: xvii, 177pages: 29cmContent type:
Media type:
Carrier type:
Subject(s): LOC classification:
  • BTH TK 870 A45 2025
Dissertation note: College of Engineering.-- Bachelor of science in electronics engineering: Technological University of the Philippines, 2025. Summary: This study presents the development and evaluation of a Wireless Blind-Spot Monitoring System with Real-Time Motion Detection and Risk Estimation using YOLOv11 and Optical Flow on the NVIDIA Jetson Orin Nano, designed specifically for utility vehicles such as the Toyota Hilux 2.4 FX 4x2 M/T. The system aims to enhance driver awareness and safety by providing real-time object detection and motion tracking in the vehicle’s blind spots through a non-intrusive Heads-Up Display (HUD). It integrates four ESP32-CAM modules for video capture, wirelessly transmitting data to the Jetson Orin Nano for processing. YOLOv11 is utilized for object detection while a CUDA- accelerated optical flow algorithm, in conjunction with OBD-based vehicle speed input, is employed for estimating the relative motion of surrounding objects. Experimental results demonstrate strong object detection performance with a mean Average Precision ([email protected]) of 0.78 and consistent tracking of moving objects. Real- time processing was achieved with an average inference time of 0.06 seconds per frame, though memory constraints on the Jetson Orin Nano presented limitations. Speed estimation tests yielded mean absolute percentage errors ranging from 4.4% to 69.1%, depending on object distance and motion complexity. A bird’s-eye view (BEV) map was integrated into the HUD for intuitive spatial awareness. The system was deployed and evaluated across 30 road trips in urban environments, with user surveys indicating improved driving confidence and reduced distraction. This study concludes that the proposed system offers a scalable, affordable, and real-time solution to blind-spot monitoring, contributing to safer road conditions and intelligent vehicle systems in developing transportation infrastructures.
Tags from this library: No tags from this library for this title. Log in to add tags.
Star ratings
    Average rating: 0.0 (0 votes)
Holdings
Item type Current library Shelving location Call number Copy number Status Date due Barcode
Bachelor's Thesis COE Bachelor's Thesis COE TUP Manila Library Thesis Section-2nd floor BTH TK 870 A45 2025 (Browse shelf(Opens below)) c.1. Not for loan BTH0006402

Bachelor's thesis

College of Engineering.-- Bachelor of science in electronics engineering: Technological University of the Philippines, 2025.

Includes bibliographic references and index.

This study presents the development and evaluation of a Wireless Blind-Spot
Monitoring System with Real-Time Motion Detection and Risk Estimation using
YOLOv11 and Optical Flow on the NVIDIA Jetson Orin Nano, designed specifically for
utility vehicles such as the Toyota Hilux 2.4 FX 4x2 M/T. The system aims to enhance
driver awareness and safety by providing real-time object detection and motion tracking in
the vehicle’s blind spots through a non-intrusive Heads-Up Display (HUD). It integrates
four ESP32-CAM modules for video capture, wirelessly transmitting data to the Jetson

Orin Nano for processing. YOLOv11 is utilized for object detection while a CUDA-
accelerated optical flow algorithm, in conjunction with OBD-based vehicle speed input, is

employed for estimating the relative motion of surrounding objects.
Experimental results demonstrate strong object detection performance with a mean

Average Precision ([email protected]) of 0.78 and consistent tracking of moving objects. Real-
time processing was achieved with an average inference time of 0.06 seconds per frame,

though memory constraints on the Jetson Orin Nano presented limitations. Speed
estimation tests yielded mean absolute percentage errors ranging from 4.4% to 69.1%,
depending on object distance and motion complexity. A bird’s-eye view (BEV) map was
integrated into the HUD for intuitive spatial awareness.
The system was deployed and evaluated across 30 road trips in urban environments,
with user surveys indicating improved driving confidence and reduced distraction. This
study concludes that the proposed system offers a scalable, affordable, and real-time
solution to blind-spot monitoring, contributing to safer road conditions and intelligent
vehicle systems in developing transportation infrastructures.

There are no comments on this title.

to post a comment.

Click on an image to view it in the image viewer

Local cover image



© 2025 Technological University of the Philippines.
All Rights Reserved.

Powered by Koha