{"id":397,"date":"2010-05-28T11:50:54","date_gmt":"2010-05-28T15:50:54","guid":{"rendered":"http:\/\/blogs.law.harvard.edu\/niftyc\/?p=397"},"modified":"2011-07-29T22:59:09","modified_gmt":"2011-07-30T02:59:09","slug":"confessions-of-a-spy-car-driver","status":"publish","type":"post","link":"https:\/\/archive.blogs.harvard.edu\/niftyc\/archives\/397","title":{"rendered":"Confessions of a Spy Car Driver"},"content":{"rendered":"<p>(or: <strong>Inadvertently Illegal Programming, A Primer<\/strong>)<\/p>\n<p>Earlier this month, Google&#8217;s official engineering blog confessed that the company&#8217;s Street View cars and bikes have <strong>&#8220;inadvertently&#8221; gathered personal data in transit on unencrypted Wi-Fi networks<\/strong> for the past three years (see the post:\u00a0<a href=\"http:\/\/googleblog.blogspot.com\/2010\/05\/wifi-data-collection-update.html\">Wi-Fi Data Collection<\/a>). \u00a0As chronicled in major news stories in the past three weeks, Google&#8217;s actions are under scrutiny by government regulators everywhere (see links to news stories at the end of this post).<\/p>\n<p style=\"text-align: center\"><a href=\"http:\/\/blogs.law.harvard.edu\/niftyc\/files\/2010\/05\/google_street_view_car.jpg\"><img loading=\"lazy\" decoding=\"async\" class=\"aligncenter size-medium wp-image-398\" src=\"http:\/\/blogs.law.harvard.edu\/niftyc\/files\/2010\/05\/google_street_view_car-300x225.jpg\" alt=\"\" width=\"300\" height=\"225\" srcset=\"https:\/\/archive.blogs.harvard.edu\/niftyc\/files\/2010\/05\/google_street_view_car-300x225.jpg 300w, https:\/\/archive.blogs.harvard.edu\/niftyc\/files\/2010\/05\/google_street_view_car.jpg 1024w\" sizes=\"auto, (max-width: 300px) 100vw, 300px\" \/><\/a><em> <\/em><\/p>\n<p style=\"text-align: center\"><em>[One of Google&#8217;s Ominous-Looking Spy Cars<br \/>\nphoto by <a href=\"http:\/\/www.flickr.com\/photos\/byrion\/2666901841\/\">byrion<\/a> &#8212; click to enlarge] <\/em><\/p>\n<p>This is a topic close to my heart because <strong>my research group has been conducting similar surveys of wireless signals<\/strong> for the past five years as part of a project funded by the US National Science Foundation. \u00a0Here&#8217;s a picture of our own slightly less obtrusive <strong>Wi-Fi sampling car<\/strong> in South Central Los Angeles in 2005. \u00a0(On second thought, we shouldn&#8217;t have chosen a black SUV. \u00a0Too scary.)<\/p>\n<p><!--more--><\/p>\n<p style=\"text-align: center\"><a href=\"http:\/\/blogs.law.harvard.edu\/niftyc\/files\/2010\/05\/South_Central_LA_011.sized_.jpg\"><img loading=\"lazy\" decoding=\"async\" class=\"aligncenter size-medium wp-image-399\" src=\"http:\/\/blogs.law.harvard.edu\/niftyc\/files\/2010\/05\/South_Central_LA_011.sized_-225x300.jpg\" alt=\"\" width=\"225\" height=\"300\" srcset=\"https:\/\/archive.blogs.harvard.edu\/niftyc\/files\/2010\/05\/South_Central_LA_011.sized_-225x300.jpg 225w, https:\/\/archive.blogs.harvard.edu\/niftyc\/files\/2010\/05\/South_Central_LA_011.sized_.jpg 480w\" sizes=\"auto, (max-width: 225px) 100vw, 225px\" \/><\/a><\/p>\n<p style=\"text-align: center\"><em>[At least our antenna isn&#8217;t as scary as Google&#8217;s.<br \/>\nClick to enlarge &#8212; photo Credit: Hope Hall]<\/em><\/p>\n<p style=\"text-align: left\">Our project was research and not commerce, so thanks to something called the <a href=\"http:\/\/ohsr.od.nih.gov\/guidelines\/belmont.html\">National Research Act of 1974<\/a>, in order to begin our research project we needed ethics approval from a university panel of researchers and civilians. \u00a0We had to <strong>investigate, explain and justify the privacy implications<\/strong> of our study to a group called the university&#8217;s &#8220;institutional review board&#8221; before we started doing anything.<\/p>\n<p style=\"text-align: left\">My argument to the board went like this: \u00a0We want to count the presence and absence of Wi-Fi networks, and we want to uniquely identify them so that we can tell where Wi-Fi exists and where it doesn&#8217;t. \u00a0(This is the same thing Google wants to do. \u00a0<strong>Other companies do it too<\/strong>, like Microsoft and <a href=\"http:\/\/www.skyhookwireless.com\/\">Skyhook Wireless<\/a>.) \u00a0A main commercial motivation for this kind of project is to improve GPS accuracy (try it: <a href=\"http:\/\/loki.com\/\">http:\/\/loki.com\/<\/a>). \u00a0Our research motivation was to understand the evolution and diffusion of computer networks.<\/p>\n<p style=\"text-align: left\">This is akin to <strong>doing a survey of telephone adoption by counting telephone poles<\/strong>. \u00a0We can do this research from public streets and sidewalks. \u00a0We are looking at unencrypted information that is broadcast in the clear to everyone anyway (called the &#8220;management frame&#8221; &#8212; this information is what creates the list of available Wi-Fi access points that is on the upper right (Mac) or lower right (PC) of your laptop). \u00a0We don&#8217;t look at the content of the transmissions.<\/p>\n<p style=\"text-align: left\">Although our equipment <em>looks <\/em>different from your laptop (and works faster and on more channels), our code does essentially the same thing that your laptop does when you open it in a new place. \u00a0It listens to see if there is any Wi-Fi around. \u00a0That&#8217;s it. \u00a0To me, it <strong>didn&#8217;t seem like a difficult ethical case to make<\/strong>. \u00a0Indeed we easily passed ethics review and our research was declared exempt from further review.<\/p>\n<p style=\"text-align: left\">To give you an example of what we see, here is a screenshot from a popular open source wireless sniffer, <a href=\"http:\/\/www.kismetwireless.net\/\">kismet<\/a>. \u00a0(We use a slightly modified version.)<\/p>\n<p style=\"text-align: center\"><a href=\"http:\/\/blogs.law.harvard.edu\/niftyc\/files\/2010\/05\/kismet-screenshot.jpg\"><img loading=\"lazy\" decoding=\"async\" class=\"aligncenter size-medium wp-image-409\" src=\"http:\/\/blogs.law.harvard.edu\/niftyc\/files\/2010\/05\/kismet-screenshot-300x180.jpg\" alt=\"\" width=\"300\" height=\"180\" srcset=\"https:\/\/archive.blogs.harvard.edu\/niftyc\/files\/2010\/05\/kismet-screenshot-300x180.jpg 300w, https:\/\/archive.blogs.harvard.edu\/niftyc\/files\/2010\/05\/kismet-screenshot.jpg 800w\" sizes=\"auto, (max-width: 300px) 100vw, 300px\" \/><\/a><\/p>\n<p style=\"text-align: center\"><em>[Kismet screenshot &#8212; click to enlarge.]<\/em><\/p>\n<p style=\"text-align: left\">Google was trying to do the same thing that my wireless research group was doing &#8212; again, no ethical problems there. \u00a0However, they claim to have &#8220;inadvertently&#8221; also <strong>listened to the content of communications<\/strong>. \u00a0(This is called &#8220;payload&#8221; data.) \u00a0<strong>Here&#8217;s the problem with the story<\/strong> we&#8217;re getting from Google: the word &#8220;<strong>inadvertently<\/strong>.&#8221;<\/p>\n<p style=\"text-align: left\">I see no way that this could be inadvertent. \u00a0Continuing my earlier metaphor: If your plan is to count telephone poles <strong>how would you &#8220;inadvertently&#8221; tap telephone lines and transcribe everything that you hear<\/strong>? \u00a0The two actions are quite different. \u00a0Of course I don&#8217;t know how Google wrote its software for these capture platforms. \u00a0With our team we use slightly modified versions of open source wireless tools. \u00a0It is possible to use tools like these to save the &#8220;payload&#8221; data from wireless systems. \u00a0There are legitimate engineering reasons for doing so if you are trying to improve the performance of your network. \u00a0That&#8217;s why these tools exist.<\/p>\n<p style=\"text-align: left\">However, I don&#8217;t understand how we could ever have &#8220;inadvertently&#8221; done that. \u00a0It isn&#8217;t like stumbling over a banana peel or forgetting to leave a light switch (or variable) turned on. \u00a0Even sampling <em>some <\/em>of the payload data would start producing about 10x as much data as listening only to the management frame. \u00a0<strong>Do you ever go to the store for a can of soda and &#8220;inadvertently&#8221; fill your cart with ten cans? <\/strong>I didn&#8217;t think so.<\/p>\n<p style=\"text-align: left\">If you <strong>inadvertently started buying 10x as many groceries as you wanted<\/strong>, I bet you&#8217;d notice. \u00a0I bet it would take you less than three years to notice, too.<\/p>\n<p style=\"text-align: left\">The only interpretation I can think of is that the word &#8220;inadvertently&#8221; is being applied by the legal department. \u00a0The real chain of events is probably that the coders at <strong>Google intentionally designed these systems to act in this (illegal) way<\/strong>, but they didn&#8217;t understand the legal and PR implications. \u00a0Programmers may have set it up on their own initiative and not briefed anyone else who could have seen this disaster looming, but that isn&#8217;t &#8220;inadvertent.&#8221; \u00a0And <strong>it isn&#8217;t a &#8220;programming error&#8221; <\/strong>&#8212; another phrase that is being used in the press.<\/p>\n<p style=\"text-align: left\"><strong>From here Google looks pretty guilty<\/strong>. \u00a0Now in the most recent news reports it looks like they are trying to destroy the data as quickly as possible as a way out of the scandal. \u00a0But looking at the data would make it even clearer that its collection wasn&#8217;t accidental. \u00a0So \u00a0in this case destroying the private data may <em>not <\/em>be a way to protect the privacy of those they snooped on &#8212; instead it seems like a way to protect Google&#8217;s nontraditional use of this word: &#8220;inadvertently.&#8221;<\/p>\n<p style=\"text-align: left\">P.S.<br \/>\n<strong>See visualizations<\/strong> from our (legal and ethical) Wi-Fi research: \u00a0<a href=\"http:\/\/pact.uiuc.edu\/red\">The RED Project<\/a>. \u00a0From: Sandvig, C. (2007). The RED Project: Rendering Electromagnetic Distributions. <em>Vectors: Journal of Culture and Technology in a Dynamic Vernacular<\/em>. <em>3<\/em>(1).<\/p>\n<p style=\"text-align: left\">P.P.S.<br \/>\nThanks to <a href=\"http:\/\/www.rajivshah.com\/\">Rajiv Shah<\/a> for suggesting this post.<\/p>\n<p style=\"text-align: left\">________<\/p>\n<p>Related news coverage:<\/p>\n<p>&#8220;<a href=\"http:\/\/www.nytimes.com\/2010\/05\/15\/business\/15google.html\">Google Says it Collected Private Data by Mistake<\/a>&#8221; (NYT), &#8220;<a href=\"http:\/\/www.nytimes.com\/2010\/05\/28\/technology\/28google.html?adxnnl=1&amp;adxnnlx=1275059019-kn10\/nUyMLynUNaOVq+mBg\">Google Balks at Turning Over Private Internet Data to Regulators<\/a>&#8221; (NYT), &#8220;<a href=\"http:\/\/www.washingtonpost.com\/wp-dyn\/content\/article\/2010\/05\/27\/AR2010052705595.html\">FTC Asks Google to retain WiFi data<\/a>&#8221; (Washington Post).<\/p>\n<p>________<\/p>\n<p><strong><span style=\"color: #ff0000\">UPDATED <\/span><\/strong>on 5\/29: \u00a0Added <strong>kismet screenshot<\/strong>, fixed typo.<\/p>\n","protected":false},"excerpt":{"rendered":"<p>(or: Inadvertently Illegal Programming, A Primer) Earlier this month, Google&#8217;s official engineering blog confessed that the company&#8217;s Street View cars and bikes have &#8220;inadvertently&#8221; gathered personal data in transit on unencrypted Wi-Fi networks for the past three years (see the post:\u00a0Wi-Fi Data Collection). \u00a0As chronicled in major news stories in the past three weeks, Google&#8217;s [&hellip;]<\/p>\n","protected":false},"author":2132,"featured_media":0,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"_jetpack_newsletter_access":"","_jetpack_dont_email_post_to_subs":false,"_jetpack_newsletter_tier_id":0,"_jetpack_memberships_contains_paywalled_content":false,"_jetpack_memberships_contains_paid_content":false,"footnotes":"","jetpack_publicize_message":"","jetpack_publicize_feature_enabled":true,"jetpack_social_post_already_shared":false,"jetpack_social_options":{"image_generator_settings":{"template":"highway","default_image_id":0,"font":"","enabled":false},"version":2},"jetpack_post_was_ever_published":false},"categories":[48405,1321],"tags":[],"class_list":["post-397","post","type-post","status-publish","format-standard","hentry","category-geeking","category-research"],"jetpack_publicize_connections":[],"jetpack_featured_media_url":"","jetpack_sharing_enabled":true,"jetpack_shortlink":"https:\/\/wp.me\/p4M7Bm-6p","_links":{"self":[{"href":"https:\/\/archive.blogs.harvard.edu\/niftyc\/wp-json\/wp\/v2\/posts\/397","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/archive.blogs.harvard.edu\/niftyc\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/archive.blogs.harvard.edu\/niftyc\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/archive.blogs.harvard.edu\/niftyc\/wp-json\/wp\/v2\/users\/2132"}],"replies":[{"embeddable":true,"href":"https:\/\/archive.blogs.harvard.edu\/niftyc\/wp-json\/wp\/v2\/comments?post=397"}],"version-history":[{"count":10,"href":"https:\/\/archive.blogs.harvard.edu\/niftyc\/wp-json\/wp\/v2\/posts\/397\/revisions"}],"predecessor-version":[{"id":403,"href":"https:\/\/archive.blogs.harvard.edu\/niftyc\/wp-json\/wp\/v2\/posts\/397\/revisions\/403"}],"wp:attachment":[{"href":"https:\/\/archive.blogs.harvard.edu\/niftyc\/wp-json\/wp\/v2\/media?parent=397"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/archive.blogs.harvard.edu\/niftyc\/wp-json\/wp\/v2\/categories?post=397"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/archive.blogs.harvard.edu\/niftyc\/wp-json\/wp\/v2\/tags?post=397"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}