-
Notifications
You must be signed in to change notification settings - Fork 16
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Machine Learning / Deep Learning #47
Comments
After a couple of weeks working in the project I would love to share some ideas. Multi-View CNN - Something to definitely have a look atSee for example: http://vis-www.cs.umass.edu/mvcnn/ What does it do? It takes into account an ordered sequence of pictures. It is the 360deg view we have been talking about a couple of times. How to turn this into a process? What are the alternatives for capturing images?
As a start we could comb through the already existing data and see if a fitting dataset could be derived from that. I believe this can be generalized to use also pointclouds instead of pictures. One option for showing the point: Artificial data.You can get some inspiration from the following, excellent Apple-article: Bottom line: Usually you need way too many images to solve problems that are similar to ours. A solution would be to use rendered data. Apple found out that this works. And they had to add a refining step that makes the artificial data more realistic. I did some quick research and experiments. There is already some research done that is in our domain: All you need to to is to create a pipeline that automatically generates images/pointclouds from some parametrized models. I tried Blender for a couple of minutes. It can be called from the command-line, this making it perfect for automatic rendering of 3D data. Note that I did not do an extensive research. I just used Blender as an example. Here are some simple images rendered from a simple model: Definitely have a look at Apple's ARKit2Look first: See this picture animated in this great article: ARKit2 is currently beta. It will be available in autumn. Amongst other things it makes the World-Map available to developers. This is a data-structure that maps reality into a point-cloud of some sorts based on data from the motion sensors and 2d-camera data. Questions:
|
The Learning/Prediction Pipeline
Objective:
3 million children are dying of malnutrition every year. We need a game-changer to identify malnutrition of children, to replace the manual measures of weight and height, which are costly, slow and often inaccurate. Our mobile app scans collects 3D point clouds and video data from children to extract anthropometric measurements like
Also, for rapid assessment especially in offline regions, and for cheap mass-market smartphones, classification of severely or moderate acute malnourished, normal or overweight children from video would be valuable.
Special attention is given to the fact that uncooperative children and imperfect lighting and internet connectivity need to be addressed for a useful general approach.
Our goal is to do online learning, so we can gradually improve the quality of our measurements.
Available data
Our current Dataset consist of
which amounts to 10 GB of zip compressed point cloud data
Throughout the next 6 month we will collect data of
more than 250 GB of data
Concepts to explore
Absolute millimeters from Point Clouds
SAM/MAM/Normal/Overweight Classification
The text was updated successfully, but these errors were encountered: