What if we could make music when we mimic instruments? Using ML and paying attention to certain hand gestures maybe we can!
This is the first stage of such an app, starting off with drums and might expand to other instruments in the future.
- Import MediaPipe
- Getting vision library via cdn @mediapipe/tasks-vision
- Console logging finger values
- Checking if left and right hand are present
- Calculate pinch
- Update pinch ref
- Calculate velocity
- Import Tone or other music library.
- Tie pinch to trigger drum sample.
- Velocity pass a certain threshold triggers music.
- New notes for each hand!
- Build UI to let people know if theyre hands are triggering.
- Let people use notes for each hand.
- Upload to netlify.
- Debug online production drawing thing? (made my own drawing functions)
- Add a loading screen takes longer on mobile.
- Make sure its working on mobile.
-
Make a config.json to store threshold values for app.
-
Refactor refs to be in one state to make it simpler to mutate in child functions
-
Expand app to include piano and guitar guestures.
Audio wont work on IOS if phone is in silent mode!