๐ Efficient Real-Time Vision for Edge AI Optimization of Lightweight Deep Learning Models for Resource-Constrained Systems ๐ 1. Introduction
Recent advances in deep learning have significantly improved the performance of computer vision systems. However, state-of-the-art models often require high computational resources, making them unsuitable for deployment on edge devices with limited processing power, memory, and energy constraints.
This project focuses on designing and optimizing lightweight deep learning models capable of performing real-time object detection while maintaining a balance between accuracy, speed, and model size.
๐ฏ 2. Problem Statement
Many modern vision models achieve high accuracy at the cost of increased computational complexity and latency. Such models cannot be efficiently deployed in real-time systems operating on low-power hardware.
Applications affected by this limitation include:
Autonomous robots
Smart surveillance systems
Traffic monitoring
Assistive technologies
Agricultural automation
Low-cost embedded AI solutions
The key objective of this project is:
To develop an optimized vision model that achieves real-time performance on resource-constrained devices without significant loss in detection accuracy.
๐ 3. Related Work
This work is inspired by research on efficient deep learning architectures and model optimization techniques:
MobileNet โ Introduced depthwise separable convolutions for lightweight CNN design
EfficientNet โ Proposed compound scaling for improved performance efficiency
YOLO-Tiny / YOLO-Nano โ Designed for real-time object detection
Model Pruning Techniques โ Reduce redundant parameters to compress networks
Quantization Methods โ Convert floating-point weights to lower precision for faster inference
These approaches demonstrate that careful optimization can significantly improve deployment feasibility on edge systems.
๐งช 4. Methodology 4.1 Baseline Model
Model: YOLOv8 Nano (lightweight object detection model)
Framework: PyTorch
Dataset: COCO dataset (subset for experimentation)
Hardware: CPU-based execution environment (no dedicated GPU required)
4.2 Optimization Techniques
To improve deployment efficiency, the following techniques are explored:
Structured pruning to remove redundant parameters
Post-training quantization to reduce model precision
Hyperparameter tuning for performance improvement
Evaluation of trade-offs between speed and accuracy
4.3 Evaluation Metrics
Model performance is evaluated using:
Mean Average Precision (mAP)
Frames Per Second (FPS)
Inference latency
Model size (MB)
Parameter count
๐ 5. Experimental Results
Initial experiments indicate that lightweight models can achieve real-time inference on CPU while maintaining acceptable accuracy for practical applications.
๐น Performance Comparison Model Version mAP (%) FPS Model Size (MB) Inference Time (ms) Baseline 37.2 18 6.2 55 Pruned 35.8 22 4.9 45 Quantized 34.6 27 3.8 37
(Results obtained on CPU-based testing environment)
๐ฅ 6. Demo
The model performs real-time object detection on standard video input using CPU-only inference.
๐น Demo video: (to be added)
๐ธ Sample outputs available in the /demo/ directory
๐ป Real-time inference tested without GPU acceleration
๐ง 7. Key Findings
Quantization significantly reduces model size with minimal accuracy loss
Pruning improves inference speed by removing redundant weights
Lightweight models can achieve practical real-time performance on CPU
Optimization techniques enable deployment on low-power systems
๐ 8. Applications
The optimized system can be deployed in:
Edge robotics and autonomous systems
Smart city surveillance solutions
Intelligent transportation systems
Assistive technologies for accessibility
Precision agriculture monitoring
Low-cost AI deployments in developing regions
๐ฎ 9. Future Work
Planned extensions of this research include:
Deployment on embedded hardware platforms (Raspberry Pi / Jetson Nano)
Energy consumption analysis
Integration with real-time tracking systems
Exploration of Neural Architecture Search (NAS) for automated optimization
Expansion to multi-object tracking in video streams
โ๏ธ 10. Installation & Usage git clone https://github.com/yourusername/efficient-real-time-vision-edge-ai cd efficient-real-time-vision-edge-ai pip install -r requirements.txt python run_inference.py ๐ Repository Structure โโโ training/ โโโ optimization/ โโโ evaluation/ โโโ results/ โ โโโ graphs/ โโโ demo/ โโโ report/ ๐ 11. Technical Report
A detailed technical report documenting methodology, experiments, and analysis will be included in the /report/ directory.
๐ 12. Author
Varshakaleeswaran B.Tech โ Artificial Intelligence & Data Science
Research Interests: Artificial Intelligence โข Machine Learning โข Computer Vision โข Efficient Deep Learning โข Edge AI