{"id":7964,"date":"2018-12-23T17:09:57","date_gmt":"2018-12-24T01:09:57","guid":{"rendered":"http:\/\/jnack.com\/blog\/?p=7964"},"modified":"2018-12-23T17:09:57","modified_gmt":"2018-12-24T01:09:57","slug":"ar-ai-help-blind-users-navigate-space-perceive-emotions","status":"publish","type":"post","link":"http:\/\/jnack.com\/blog\/2018\/12\/23\/ar-ai-help-blind-users-navigate-space-perceive-emotions\/","title":{"rendered":"AR &#038; AI help blind users navigate space &#038; perceive emotions"},"content":{"rendered":"<p>I love assistive superpowers like this work from Caltech:<\/p>\n<p><iframe loading=\"lazy\" width=\"604\" height=\"340\" src=\"https:\/\/www.youtube.com\/embed\/vAGybwLb3kg?feature=oembed\" frameborder=\"0\" allow=\"accelerometer; autoplay; encrypted-media; gyroscope; picture-in-picture\" allowfullscreen><\/iframe><\/p>\n<p><a href=\"https:\/\/vrscout.com\/news\/caltech-ar-helps-blind-navigate\/?utm_source=VRScout+Scouting+Report&amp;utm_campaign=c5722af3e5-VRScoutReport_121418&amp;utm_medium=email&amp;utm_term=0_f3642cd298-c5722af3e5-162007301\">VR Scout<\/a> shares numerous details:<\/p>\n<blockquote>\n<p>[T]he team used the Microsoft HoloLens\u2019s capability to create a digital mesh over a \u201cscene\u201d of the real-world. Using unique software called Cognitive Augmented Reality Assistant (CARA), they were able to convert information into audio messages, giving each object a \u201cvoice\u201d that you would hear while wearing the headset. [&#8230;]<\/p>\n<p>If the object is at the left, the voice will come from the left side of the AR headset, while any object on the right will speak out to you from the right side of the headset. The pitch of the voice will change depending on how far you are from the object.<\/p>\n<\/blockquote>\n<p><img decoding=\"async\" loading=\"lazy\" title=\"NewImage.png\" src=\"http:\/\/jnack.com\/blog\/wp-content\/uploads\/2018\/12\/NewImage-20.png\" alt=\"NewImage\" width=\"592\" height=\"320\" border=\"0\" \/><\/p>\n<p>Meanwhile Huawei is using AI to help visually impaired users <a href=\"https:\/\/venturebeat.com\/2018\/12\/18\/huaweis-facing-emotions-app-using-ai-to-help-visually-impaired-users-hear-facial-expressions\/\">\u201chear\u201d facial expressions<\/a>:<\/p>\n<blockquote>\n<p><a href=\"https:\/\/play.google.com\/store\/apps\/details?id=facing.emotions\">Facing Emotions<\/a> taps the Mate 20 Pro\u2019s back cameras to scan the faces of conversation partners, identifying facial features like eyes, nose, brows, and mouth, and their positions in relation to each other. An offline, on-device machine learning algorithm interprets the detected emotions as sounds, which the app plays on the handset\u2019s loudspeaker.<\/p>\n<\/blockquote>\n<p><img decoding=\"async\" loading=\"lazy\" title=\"NewImage.png\" src=\"http:\/\/jnack.com\/blog\/wp-content\/uploads\/2018\/12\/NewImage-21.png\" alt=\"NewImage\" width=\"600\" height=\"339\" border=\"0\" \/><\/p>\n<p>[<a href=\"https:\/\/youtu.be\/vAGybwLb3kg\">YouTube<\/a>] [Via Helen Papagiannis]<\/p>\n","protected":false},"excerpt":{"rendered":"<p>I love assistive superpowers like this work from Caltech: VR Scout shares numerous details: [T]he team used the Microsoft HoloLens\u2019s capability to create a digital mesh over a \u201cscene\u201d of the real-world. Using unique software called Cognitive Augmented Reality Assistant (CARA), they were able to convert information into audio messages, giving each object a \u201cvoice\u201d [&hellip;]<\/p>\n","protected":false},"author":1,"featured_media":0,"comment_status":"open","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":[],"categories":[13],"tags":[],"_links":{"self":[{"href":"http:\/\/jnack.com\/blog\/wp-json\/wp\/v2\/posts\/7964"}],"collection":[{"href":"http:\/\/jnack.com\/blog\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"http:\/\/jnack.com\/blog\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"http:\/\/jnack.com\/blog\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"http:\/\/jnack.com\/blog\/wp-json\/wp\/v2\/comments?post=7964"}],"version-history":[{"count":1,"href":"http:\/\/jnack.com\/blog\/wp-json\/wp\/v2\/posts\/7964\/revisions"}],"predecessor-version":[{"id":7965,"href":"http:\/\/jnack.com\/blog\/wp-json\/wp\/v2\/posts\/7964\/revisions\/7965"}],"wp:attachment":[{"href":"http:\/\/jnack.com\/blog\/wp-json\/wp\/v2\/media?parent=7964"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"http:\/\/jnack.com\/blog\/wp-json\/wp\/v2\/categories?post=7964"},{"taxonomy":"post_tag","embeddable":true,"href":"http:\/\/jnack.com\/blog\/wp-json\/wp\/v2\/tags?post=7964"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}