{"id":548,"date":"2006-11-10T15:19:43","date_gmt":"2006-11-10T20:19:43","guid":{"rendered":"http:\/\/blogs.law.harvard.edu\/hoanga\/2006\/11\/10\/photosynth-a-new-way-of-seeing-relatio"},"modified":"2006-11-10T15:19:43","modified_gmt":"2006-11-10T20:19:43","slug":"photosynth-a-new-way-of-seeing-relations-in-your-photo-collection","status":"publish","type":"post","link":"https:\/\/archive.blogs.harvard.edu\/hoanga\/2006\/11\/10\/photosynth-a-new-way-of-seeing-relations-in-your-photo-collection\/","title":{"rendered":"Photosynth, a new way of seeing relations in your photo collection"},"content":{"rendered":"<p><a name='a676'><\/a><\/p>\n<p><a href=\"http:\/\/labs.live.com\/photosynth\/default.html\"><\/p>\n<p><img decoding=\"async\" src=\"http:\/\/labs.live.com\/photosynth\/images\/whatisgrapic_new.jpg\"><\/p>\n<p><\/a><\/p>\n<p>I just took a glance at <\/p>\n<p><a href=\"http:\/\/labs.live.com\/photosynth\/default.html\">photosynch<\/a> and<\/p>\n<p>am really impressed with the applications of computer vision and image<\/p>\n<p>processing techniques to create a really unique application.<\/p>\n<p>The basic<\/p>\n<p>idea is to take a pile of photos that are related to each other somehow<\/p>\n<p>(imagine taking zillions of pictures of the Taj Mahal from tons of<\/p>\n<p>different places) find similar features in all the images and try to<\/p>\n<p>reconstruct a mock 3d space that shows the spacial relation between<\/p>\n<p>all of your photos.  This is really cool as you might be able to create<\/p>\n<p>a very interesting photo tour from your photo collection in a 3d<\/p>\n<p>navigatable space.<\/p>\n<p>Oddly, I was trying to come up with a similar idea to link videos stills<\/p>\n<p>in QuickTimeVR movies and try to use the linkable features in QuicktimeVR<\/p>\n<p>to provide clickable hotspots that would take you to another photo that<\/p>\n<p>was a picture of the same scene however this is far slicker and if it works<\/p>\n<p>with very little intervention from the user besides pointing to a pile<\/p>\n<p>of photos and letting it do its job that would be great.<\/p>\n<p>However, there are still caveats.  The whole process takes hours or days<\/p>\n<p>to currently do and the current technology preview is only for a pre-rendered<\/p>\n<p>project.   The true acid test will be in my opinion the ability to<\/p>\n<p>just point to a folder of pictures and have it do its job with as little<\/p>\n<p>possible human intervention as possible.  That is a not a trivial problem<\/p>\n<p>but I&#8217;m sure we&#8217;ll see something interesting especially since it has<\/p>\n<p><a href=\"http:\/\/www.cs.washington.edu\/homes\/seitz\/\">two<\/a> (very well)<\/p>\n<p><a href=\"http:\/\/research.microsoft.com\/~szeliski\/\">known<\/a> researchers<\/p>\n<p>in the computer vision field. I&#8217;m really looking forward to the results<\/p>\n<p>of their labor.  My last question is how many technologies behind this<\/p>\n<p>are patented already.  It&#8217;d be great if it an OSS implementation inspired<\/p>\n<p>from this project could be made however patents are a sticky problem.<\/p>\n<p><a href=\"http:\/\/labs.live.com\/photosynth\/SystemCheck.htm\">Try it yourself<\/a><\/p>\n<p>(<img decoding=\"async\" src=\"http:\/\/www.microsoft.com\/library\/media\/1033\/windows\/ie\/images\/ie7\/home\/ie7_thumb_logo.gif\"> 6 or 7 REQUIRED)<\/p>\n","protected":false},"excerpt":{"rendered":"<p>I just took a glance at photosynch and am really impressed with the applications of computer vision and image processing techniques to create a really unique application. The basic idea is to take a pile of photos that are related to each other somehow (imagine taking zillions of pictures of the Taj Mahal from tons [&hellip;]<\/p>\n","protected":false},"author":704,"featured_media":0,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[1212],"tags":[],"class_list":["post-548","post","type-post","status-publish","format-standard","hentry","category-tagme"],"jetpack_featured_media_url":"","_links":{"self":[{"href":"https:\/\/archive.blogs.harvard.edu\/hoanga\/wp-json\/wp\/v2\/posts\/548","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/archive.blogs.harvard.edu\/hoanga\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/archive.blogs.harvard.edu\/hoanga\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/archive.blogs.harvard.edu\/hoanga\/wp-json\/wp\/v2\/users\/704"}],"replies":[{"embeddable":true,"href":"https:\/\/archive.blogs.harvard.edu\/hoanga\/wp-json\/wp\/v2\/comments?post=548"}],"version-history":[{"count":0,"href":"https:\/\/archive.blogs.harvard.edu\/hoanga\/wp-json\/wp\/v2\/posts\/548\/revisions"}],"wp:attachment":[{"href":"https:\/\/archive.blogs.harvard.edu\/hoanga\/wp-json\/wp\/v2\/media?parent=548"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/archive.blogs.harvard.edu\/hoanga\/wp-json\/wp\/v2\/categories?post=548"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/archive.blogs.harvard.edu\/hoanga\/wp-json\/wp\/v2\/tags?post=548"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}