Hand-Detection | Android application which uses Mediapipe | Camera library
kandi X-RAY | Hand-Detection Summary
kandi X-RAY | Hand-Detection Summary
Android application which uses Mediapipe to detect hands in a live stream from a phone camera
Support
Quality
Security
License
Reuse
Top functions reviewed by kandi - BETA
- Create the options menu
- Start camera
- Initializes the preview
- Setup the preview display view
- Closes the converter
- This method is called when a request has been granted
Hand-Detection Key Features
Hand-Detection Examples and Code Snippets
Community Discussions
Trending Discussions on Hand-Detection
QUESTION
Pretty much brand new to ML here. I'm trying to create a hand-detection CoreML model using turicreate.
The dataset I'm using is from https://github.com/aurooj/Hand-Segmentation-in-the-Wild , which provides images of hands from an egocentric perspective, along with masks for the images. I'm following the steps in turicreate's "Data Preparation" (https://github.com/apple/turicreate/blob/master/userguide/object_detection/data-preparation.md) step-by-step to create the SFrame. Checking the contents of the variables throughout this process, there doesn't appear to be anything wrong.
Following data preparation, I follow the steps in the "Introductory Example" section of https://github.com/apple/turicreate/tree/master/userguide/object_detection
I get the hint of an error when turicreate is performing iterations to create the model. There doesn't appear to be any Loss at all, which doesn't seem right.
After the model is created, I try to test it with a test_data portion of the SFrame. The results of these predictions are just empty arrays though, which is obviously not right.
After exporting the model as a CoreML .mlmodel and trying it out in an app, it is unable to recognize anything (not surprisingly).
Me being completely new to model creation, I can't figure out what might be wrong. The dataset seems quite accurate to me. The only changes I made to the dataset were that some of the masks didn't have explicit file extensions (they are PNGs), so I added the .png extension. I also renamed the images to follow turicreate's tutorial formats (i.e. vid4frame025.image.png
and vid4frame025.mask.0.png
. Again, the SFrame creation process using this data seems correct at each step. I was able to follow the process with turicreate's tutorial dataset (bikes and cars) successfully. Any ideas on what might be going wrong?
ANSWER
Answered 2020-Mar-06 at 20:49I found the problem, and it basically stemmed from my unfamiliarity with Python.
In one part of the Data Preparation section, after creating bounding boxes out of the mask images, each annotation is assigned a 'label'
indicating the type of object the annotation is meant to be. My data had a different name format than the tutorial's data, so rather than each annotation having 'label': 'bike'
, my annotations had 'label': 'vid4frame25`, 'label': 'vid4frame26', etc
.
Correcting this such that each annotation has 'label': 'hand'
seems to have corrected this (or at least it's creating a legitimate-seeming model so far).
Community Discussions, Code Snippets contain sources that include Stack Exchange Network
Vulnerabilities
No vulnerabilities reported
Install Hand-Detection
You can use Hand-Detection like any standard Java library. Please include the the jar files in your classpath. You can also use any IDE and you can run and debug the Hand-Detection component as you would do with any other Java program. Best practice is to use a build tool that supports dependency management such as Maven or Gradle. For Maven installation, please refer maven.apache.org. For Gradle installation, please refer gradle.org .
Support
Reuse Trending Solutions
Find, review, and download reusable Libraries, Code Snippets, Cloud APIs from over 650 million Knowledge Items
Find more librariesStay Updated
Subscribe to our newsletter for trending solutions and developer bootcamps
Share this Page