MobileNet-Yolo | MobileNetV2-YoloV3-Nano: 0.5BFlops 3MB HUAWEI P40: 6ms/img, YoloFace-500k:0.1Bflops 420KB:fire::fire | Computer Vision library
kandi X-RAY | MobileNet-Yolo Summary
kandi X-RAY | MobileNet-Yolo Summary
For the close-range face detection model in a specific scene, the recommended detection distance is 1.5m.
Support
Quality
Security
License
Reuse
Top functions reviewed by kandi - BETA
Currently covering the most popular Java, JavaScript and Python libraries. See a Sample of MobileNet-Yolo
MobileNet-Yolo Key Features
MobileNet-Yolo Examples and Code Snippets
Community Discussions
Trending Discussions on MobileNet-Yolo
QUESTION
There are some questions when I read ssd-caffe code and I really need your help.
Native caffe only supports classification, data reading layer is commonly used to read LMDB database and read image for training
In order to support the input of multiple labels and input annotation boxes, I decide to use ssd-caffe, which adds an AnnotatedDataLayer layer to the native caffe. This newly added layer can support multiple labels and annotation boxes, but it has limitations. The reason is that the type of data it reads is still lmdb;
We now need to read the data of the data set randomly, but according to the query results, lmdb is a B+ tree structure, which can only be read sequentially through the iterator, so we want to change lmdb to read the images directly. However, the direct reading pictures of native caffe do not support multi-labels and annotation boxes. How can I modify the image_data_layers of caffe to support the input of annotation boxes (Can I follow AnnotatedDataLayer's approach to solve the problem)?
Note:
Modified ssd-caffe source code: https://github.com/eric612/MobileNet-YOLO
The file path of the newly added annotation box: /MobileNet-YOLO/src/caffe/layers/annotated_data_layer.cpp
Native caffe file path for reading pictures directly: /MobileNet-YOLO/src/caffe/layers/image_data_layer.cpp
ANSWER
Answered 2020-Jun-12 at 10:06Data layer offers the possibility of reading random data from the hard disk asynchronously (it uses 2 threads: in one it reads and in the other it delivers the data to the neural network). Your top blob is made up of the data and the label. Unfortunately, the label is 1-dimensional. To solve this problem it is possible to organize our lmdb database in a special order. Then when we read the data, before delivering it to the neural network, we transform it to adapt it to our problem. Below I show this in an example: First I will write an LMDB database with 10 different images (it is the same image, but we will assume they are different), 10 random bounding boxes and 10 random labels of dimension 3 each one.
NOTE: to reproduce the following codes you must have caffe installed. If you only have the caffe folder compiled, then create the folder in root_caffe/examples/new_folder, put the code in there and then compile doing make.
Community Discussions, Code Snippets contain sources that include Stack Exchange Network
Vulnerabilities
No vulnerabilities reported
Install MobileNet-Yolo
Support
Reuse Trending Solutions
Find, review, and download reusable Libraries, Code Snippets, Cloud APIs from over 650 million Knowledge Items
Find more librariesStay Updated
Subscribe to our newsletter for trending solutions and developer bootcamps
Share this Page