DEV Community

Cover image for An App where you can Train your Own Hand Pose Model for your Project! 🤌

An App where you can Train your Own Hand Pose Model for your Project! 🤌

FrancisTRᴅᴇᴠ (っ◔◡◔)っ on February 28, 2026

This is a submission for the DEV Weekend Challenge: Community The Community This targets the AI/ML community in general, especially tho...
Collapse
 
itsugo profile image
Aryan Choudhary

Wow, great project Francis! and even though I'm still trying to wrap my head around it. From what I've learned, it appears to be a game-changer for those who want to train hand pose models without having to sign up or pay for API keys. I'm really curious to see how this project will shape up in the future - it has the potential to make a significant impact on the community.

Collapse
 
francistrdev profile image
FrancisTRᴅᴇᴠ (っ◔◡◔)っ

Hey Aryan! Thanks for reading! I appreciate the comment. I think I could have word it better. My apologies if it didn't make sense from reading, but glad you understood it!

Collapse
 
bytethecarrot profile image
viv

Hey Francis, really cool project!

I've actually been wanting to try making one of these hand gesture projects myself, especially after seeing a similar one that used Naruto's shadow clone gesture. And I couldn't agree more about the struggle of not having many free resources for creating your own models, and that whole 'sign up here, pay here' dance, truly hate it. So, it's awesome that you're providing this kind of freedom.

I'm definitely curious about trying it out and maybe even contributing!

Collapse
 
francistrdev profile image
FrancisTRᴅᴇᴠ (っ◔◡◔)っ

Thanks Viv! Glad you liked it! Note that this is an MVP (since it was coded with Gemini), so there is a lot to work with and I might need to add a sample code in that website in order for the model to work in your project (since it is not simply a drag and drop to your project of choice). Again, thanks for reading! Let me know if anything! :D

Collapse
 
trinhcuong-ast profile image
Kai Alder

This is really cool — the fact that you can download the trained model files and use them offline is huge. Most tools lock you into their ecosystem and that's exactly why a lot of hobby projects never get past the prototype stage.

Quick question: how does the model handle variations in lighting and hand distance from the camera? I've worked with MediaPipe hand tracking before and the landmark detection can get pretty noisy when lighting conditions change. Did you find ml5.js handles that well enough, or did you have to do any preprocessing/normalization on the landmark data before training?

Also love the Domain Expansion reference lol. Now I want to build a gesture-controlled Jujutsu Kaisen soundboard.

Collapse
 
francistrdev profile image
FrancisTRᴅᴇᴠ (っ◔◡◔)っ

Hey Kai! Hope you are well! I appreciate you enjoyed it!

For your question, it handles lighting pretty well. If it was image classification, then it won't since it captures the whole image. ml5.js is built on top of Tensorflow.js and it allows to use machine learning on the web quite easier since I believe it handles most of the conditions you mentioned. I did test it out when all the lights are off and it seems to detect it just fine. I did have to normalize the landmark data, which is expected since you don't really want to mess with raw data.

I link you the documentation for ml5.js and mediapipe for reference just in case, even though you have experience already. Just in case :)

docs.ml5js.org/#/reference/handpose
github.com/google-ai-edge/mediapip...

Hope this helps and thanks for reading!!

Collapse
 
maxxmini profile image
MaxxMini

The "download and own your model files" approach is what makes this stand out. I've built a bunch of browser-based tools and games, and the pattern I keep running into is exactly what you described — most ML services want you locked into their API, which is a non-starter for client-side projects where you can't hide keys.

ml5.js + p5.js is a smart stack choice for this. Vanilla HTML/CSS/JS means zero build tooling friction, which matters a lot for an open-source tool people are supposed to fork and extend.

Curious about one thing: when you expand to image classification, are you planning to keep the same train-in-browser approach? The hand pose model is relatively lightweight since it's working with landmark coordinates, but image classification with raw pixel data might hit browser memory limits pretty quickly depending on training set size. Would be interesting to see how far you can push client-side training before needing a different architecture.

Cool challenge submission 🤌

Collapse
 
francistrdev profile image
FrancisTRᴅᴇᴠ (っ◔◡◔)っ

Hey MaxxMini! Thanks for reading! I appreciate your comment!

I would hope to keep the train-in-browser approach. I know Teachable Machine's did this capability and it work out. The only problem is the more classes and data you have, the laggier the site will become, which is expected to happen. I might limit how many classes you can have, but I don't think you will have 500 classes that does image classification realistically (as far as I know).

Again, thanks!

Collapse
 
luftietheanonymous profile image
Luftie The Anonymous

Hi dude, I've got actually a question. I'm not coming from AI field, so could you tell me what for have u built such project ? I mean are you collecting the data on finger points and movements, so that you could build a robotic hardware hand, or what is it for ?

Collapse
 
francistrdev profile image
FrancisTRᴅᴇᴠ (っ◔◡◔)っ

Hey Luftie! Thanks for reading.

The MVP I have is essentially collecting the whole hand (as you mentioned) and can use that ML model for other projects. From the example, I was doing ASL (As I demonstrated doing the ABC since I named 3 of my classes A, B, and C). ml5.js is more towards the creative coding side, but can be used for other applications.

It can be use for robotics like you mentioned, but I don't know the exact use case where you will be needing a camera that captures the hand and the robotic hand moves along with it (Unless you are building a remote robotic suite, sure). You might need to do more since I am not an expert in the hardware side of development. Here is the documentation I followed:

github.com/google-ai-edge/mediapip...
docs.ml5js.org/#/reference/handpose

Overall, it can be use for anything as long as you are using the Hand pose model. I hope this makes sense. Let me know if it doesn't or if there is any other questions. Thanks for reading man!

Collapse
 
deshawn_wilson_a5e386afff profile image
Deshawn Wilson • Edited

“HandTracker” is not just another demo — it actually solves a real problem that many beginners in AI/ML face. When people start learning machine learning, they often run into platforms that require sign-ups, API keys, limited free tiers, or paid subscriptions. That can be discouraging, especially when you just want to experiment and learn.

Here you learn more Cheats and Codm Injectors

The fact that you built a completely free, client-side tool where users can train their own hand pose model and download it without depending on external APIs is incredibly valuable for the community.

I especially love that you built this using and with vanilla HTML, CSS, and JavaScript. That makes it approachable for web developers who are curious about AI but might feel intimidated by Python-heavy ML ecosystems. It really lowers the barrier to entry.

Some standout features that I found particularly powerful:

The ability to define your own custom classes

Controlling how many samples are collected per class

Adjustable training parameters

Automatic pause when the hand is not visible (great for dataset quality!)

Downloadable model files, model.weights.bin, model

That last point is huge. Giving users full ownership of their trained model instead of locking them into a hosted service is empowering. It also makes this tool perfect for projects like Chrome extensions or fully client-side apps where hiding API keys can be difficult.

I also appreciate how creative this can get. It’s not limited to something formal like ASL — users can create any gesture they want, from simple “Hello” signs to fun pop-culture-inspired poses. That mix of creativity and machine learning makes learning more engaging and memorable.

Another thing worth highlighting is your vision for expansion — turning this into a broader ML playground for image classification, text generation, and more. If you continue building in that direction, this could evolve into a powerful open, community-driven ML training platform for beginners.

Overall, this feels bigger than just a weekend challenge submission. It’s a thoughtful, practical tool that makes AI more accessible and hands-on. Huge respect for building something that removes barriers instead of creating them.

Collapse
 
francistrdev profile image
FrancisTRᴅᴇᴠ (っ◔◡◔)っ

Hey Deshawn! Hope you are well. Thanks for taking the time to read and reflect! I appreciated it! :D

Collapse
 
m_c2 profile image
meherab hossain

Wow, it's great!!! It really cool..😍😍

Collapse
 
francistrdev profile image
FrancisTRᴅᴇᴠ (っ◔◡◔)っ

Thanks meherab! Glad you liked it! :D

Collapse
 
matsync profile image
Matsync

Hey, cool project!

Collapse
 
francistrdev profile image
FrancisTRᴅᴇᴠ (っ◔◡◔)っ

Thanks Matsync! Glad you enjoyed it!