100-YR Health Management Plan where's the app?
To my entreprenuerial advantage, no one has thought to deploy a health app UI based on daily (or more frequent) total body skin (hair and nail) imaging collected simulatenously with whole-body surface scanning, preferably naked. Such imaging and scanning equipment is not affordable for the average Walmart shopper yet but it will be in a decade or sooner if I have any success. The software engineering to automatically bring the scans and images of the UI alive (realtime interaction with your digital double) with machine learning of near continous tracking and monitoring of whole-body and part position and posture for a revolutionary UX is the grand challenge. Processing power, programming languages, storage devices, and networking (MBAN, PAN, LAN) are not worries but design parameters for the mobile platforms today to render graphics and acoustics of human biotelemetry and health data archives for goal setting, healthcare planning and health analyses. I believe the solution to the health apps' high attrition rates will be an international standard for the whole-body, true-color surface with hair and nails binded to a skeleton that can derive joint position and motion from a panoly of sensors and qualitative behaviors (e.g., operation of man-made objects). Once this hurdle is cleared, all the health literacy/education content, that's not quantified, can smartly be delivered to users with precision and timeliness.