{"id":2262,"date":"2015-02-10T23:50:20","date_gmt":"2015-02-11T04:50:20","guid":{"rendered":"http:\/\/www.jhclaura.com\/?p=2262"},"modified":"2015-11-20T17:10:16","modified_gmt":"2015-11-20T22:10:16","slug":"experiment_01","status":"publish","type":"post","link":"http:\/\/www.jhclaura.com\/experiment_01\/","title":{"rendered":"Experiment_01"},"content":{"rendered":"
For my first experiment, in order to create strong nonsense moment<\/em>, I want to connect virtual world with\u00a0the reality, so as\u00a0to play with perception and make\u00a0conflict, and one way to do this is importing\u00a0the camera feed from the mobile phone. This way, the user can not only be able to see the real world<\/strong>, but also experience illogical, contradictory events in\u00a0the virtual world<\/strong>, which is triggered by the real world.<\/p>\n <\/a>–> Contradict to reality<\/a>\u00a0–> Think everyone is monkey<\/a>\u00a0–> Encourage to say Hi<\/a>–> Focus enhancement<\/p>\n <\/p>\n <\/a><\/p>\n After talking with professor Shawn Van Every, I decided to use browser instead of App as platform first, and test the limitation of browser and Javascript.\u00a0With flexible\u00a0HTML5 and Chrome browser, I can get camera feed from mobile just like from webcam of laptop, and\u00a0getUserMedia<\/strong> &\u00a0MediaStreamTrack<\/strong> function of WebRTC<\/strong> API allow users to choose camera and set up constraints as they want. Below are the gists of it:<\/p>\n <\/p>\n <\/a><\/p>\n Nonsense is built on sense, so in order to make nonsense in virtual world, the V\u00a0World\u00a0needs to know what’s going on in the real world, and my first attempt is using computer vision to analyze the image captured from camera. Below are the computer vision JS libraries I found:<\/p>\n Issue #1<\/strong> –\u00a0Currently I used jsfeat to grayscale the footage first, and then found the bright area pixels by pixels. It’s obviously slow. Next step will be trying the combination of OpenCV and Node.js (Thanks to Pedro), see if “perform CPU-intense image processing routines in the cloud, let Node.js server handle client requests and call C++ back-end” will optimize the performance or not.<\/p>\n Issue #2<\/strong> – Have to figure out how to translate the pixel location from canvas to 3D world, since let eyeScreen rotating with camera (head) making everything complicated.<\/p>\n","protected":false},"excerpt":{"rendered":" Camera Feed For my first experiment, in order to create strong nonsense moment, I want to connect virtual world with\u00a0the reality, so as\u00a0to play with perception and make\u00a0conflict, and one… Read The Rest →<\/a><\/p>\n","protected":false},"author":1,"featured_media":0,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":[],"categories":[244],"tags":[249,250],"_links":{"self":[{"href":"http:\/\/www.jhclaura.com\/wp-json\/wp\/v2\/posts\/2262"}],"collection":[{"href":"http:\/\/www.jhclaura.com\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"http:\/\/www.jhclaura.com\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"http:\/\/www.jhclaura.com\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"http:\/\/www.jhclaura.com\/wp-json\/wp\/v2\/comments?post=2262"}],"version-history":[{"count":1,"href":"http:\/\/www.jhclaura.com\/wp-json\/wp\/v2\/posts\/2262\/revisions"}],"predecessor-version":[{"id":2263,"href":"http:\/\/www.jhclaura.com\/wp-json\/wp\/v2\/posts\/2262\/revisions\/2263"}],"wp:attachment":[{"href":"http:\/\/www.jhclaura.com\/wp-json\/wp\/v2\/media?parent=2262"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"http:\/\/www.jhclaura.com\/wp-json\/wp\/v2\/categories?post=2262"},{"taxonomy":"post_tag","embeddable":true,"href":"http:\/\/www.jhclaura.com\/wp-json\/wp\/v2\/tags?post=2262"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}Possible scenario<\/h3>\n
Real + Virtual Mashup<\/h3>\n
\n
Computer Vision on Phone<\/h3>\n
\n