
Gesture-Controlled Bluetooth Speaker
Project Overview
The project uses an ESP32-CAM and lightweight vision models to detect gestures and control playback via Bluetooth AVRCP commands.
Tech Stack
Team
Mentors
- Vinay
- Mahadev
- Shreesha
- Siri
Mentees
- Aditya Khankar
- Devendranath Reddy
- Partha Sarathi
- Vrushank
Problem Statement
Touch-based audio interfaces are inconvenient in hygiene-sensitive and accessibility-focused use cases.
Objectives
- - Build fully gesture-controlled portable speaker
- - Run vision gesture recognition on ESP32-CAM
- - Deploy lightweight on-device ML
- - Use AVRCP for playback controls
Methodology
The project follows a structured implementation approach that includes Collect and label gesture dataset, Train CNN with TFLite/Edge Impulse, Deploy inference to ESP32-CAM, and Integrate hardware enclosure and audio module. These steps are executed iteratively to validate assumptions, improve performance, and ensure reliable delivery of the final solution.
Expected Outcome
By the end of this project, the team is expected to deliver Gesture-based playback control prototype, and On-device inference with high practical accuracy. Together, these outcomes reflect both technical feasibility and practical value for demos, evaluation, and future scaling.
Future Scope
- - Voice plus gesture multimodal control
- - Smart-home control integration
- - Expanded datasets for robustness
Components and Budget
ESP32-CAM modules (x2): Rs. 1,600
Bluetooth speaker module: Rs. 1,500