Autonomous Obstacle Avoidance with Monocular Percept
Autonomous navigtion of automobiles is a long standing problem. Navigation task requires heavy and expensive sensors like lidar, and kinect to percieve the environment. These sensors drain a lot of power. However, single cameras are cheap and found on most quadcopters today
2025-06-28 16:30:32 - Adil Khan
Autonomous Obstacle Avoidance with Monocular Percept
Project Area of Specialization Artificial IntelligenceProject SummaryAutonomous navigtion of automobiles is a long standing problem. Navigation task requires heavy and expensive sensors like lidar, and kinect to percieve the environment. These sensors drain a lot of power. However, single cameras are cheap and found on most quadcopters today. Repurposing these cameras for depth estimation of a scene is a useful feature to have. Our method uses deep learning to train a camera to percieve distances from obstacles from a single image, exploiting Stucture from Motion techniques. This estimated depth map of the scene is used to determine steering commands for a quadcopter to avoid obstacle collision for safe driving.
Project ObjectivesThe objective of my project is to train an obstacle avoiding deep network that is fast and accurate. It should be fast enoght for real time response on a quadcopter, and accurate enough for it to be practical.
Project Implementation MethodThe project will be implemented with a layer of deep convolutional neural network for depth estimation with supervised learning in a simulation environment. The simulation environment of choice is AirSim by Microsoft. This depth map is further passed into another network that estimates the navigation command for the quadcopter. This will be trained in AirSim with reinforcement learning.
Benefits of the ProjectMonocular camera based navigation is a low cost, power efficient, and lightweight alternative to conventional lidar sensor and kinect camera percept used for obstacle avoidance. This improvement will make quadcopters cheaper and more accessible and promote autonomy and safety in navigation.
Technical Details of Final DeliverableThe network will be trained on a gpu in a simulated environment, but it is very generalizable and the learning can be transferred to real environment. Deploying it on a quadcopter with real time response would require gpu embedding on the quadcopter for fast response,or alterbnatively remote processing can employed. The quadcopter will transmit the state of environment captured through the camera to a remote server over network. This image is processsed on a gpu enabled system and a steering command is returned in real time.
Final Deliverable of the Project Software SystemType of Industry Transportation Technologies Artificial Intelligence(AI), RoboticsSustainable Development Goals Industry, Innovation and InfrastructureRequired Resources| Item Name | Type | No. of Units | Per Unit Cost (in Rs) | Total (in Rs) |
|---|---|---|---|---|
| Total in (Rs) | 75000 | |||
| NVIDIA GTX 1070 | Equipment | 1 | 60000 | 60000 |
| SSD 120gb | Equipment | 1 | 10000 | 10000 |
| IEEE & CIS membership | Miscellaneous | 1 | 5000 | 5000 |