<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0"
	xmlns:content="http://purl.org/rss/1.0/modules/content/"
	xmlns:wfw="http://wellformedweb.org/CommentAPI/"
	xmlns:dc="http://purl.org/dc/elements/1.1/"
	xmlns:atom="http://www.w3.org/2005/Atom"
	xmlns:sy="http://purl.org/rss/1.0/modules/syndication/"
	xmlns:slash="http://purl.org/rss/1.0/modules/slash/"
	>

<channel>
	<title>Google Data &#187; Research Admin</title>
	<atom:link href="/author/research-admin/feed/" rel="self" type="application/rss+xml" />
	<link>https://googledata.org</link>
	<description>Everything Google: News, Products, Services, Content, Culture</description>
	<lastBuildDate>Thu, 19 Mar 2015 22:49:02 +0000</lastBuildDate>
	<language>en-US</language>
		<sy:updatePeriod>hourly</sy:updatePeriod>
		<sy:updateFrequency>1</sy:updateFrequency>
	<generator>http://wordpress.org/?v=3.7.5</generator>
	<item>
		<title>Academic Successes in Cluster Computing</title>
		<link>https://googledata.org/google-research/academic-successes-in-cluster-computing/?utm_source=rss&#038;utm_medium=rss&#038;utm_campaign=academic-successes-in-cluster-computing</link>
		<comments>https://googledata.org/google-research/academic-successes-in-cluster-computing/#comments</comments>
		<pubDate>Thu, 22 Dec 2011 23:00:00 +0000</pubDate>
		<dc:creator><![CDATA[Research Admin]]></dc:creator>
				<category><![CDATA[Google Research]]></category>

		<guid isPermaLink="false">https://googledata.org/?guid=beba9c09955e14772c7e23d5747f278a</guid>
		<description><![CDATA[Posted by Alfred Spector, VP of ResearchAccess to massive computing resources is foundational to Research and Development. Fifteen awardees of the National Science Foundation (NSF) Cluster Exploratory Service (CLuE) program have been applying large sca...]]></description>
				<content:encoded><![CDATA[<span class="byline-author">Posted by Alfred Spector, VP of Research</span><br /><br />Access to massive computing resources is foundational to Research and Development. Fifteen awardees of the National Science Foundation (NSF) <a href="http://www.nsf.gov/awardsearch/progSearch.do?SearchType=progSearch&page=2&QueryText=&ProgOrganization=&ProgOfficer=&ProgEleCode=7782&BooleanElement=false&ProgRefCode=&BooleanRef=false&ProgProgram=&ProgFoaCode=&RestrictActive=on&Search=Search#results">Cluster Exploratory Service</a> (CLuE) program have been applying large scale computational resources <a href="http://googleblog.blogspot.com/2008/02/supporting-cluster-computing-in.html">donated by Google and IBM</a>. <br /><br />Overall, 1,328 researchers have used the cluster to perform over 120 million computing tasks on the cluster and in the process, have published 49 scientific publications, educated thousands of students on parallel computing and supported numerous post-doctoral candidates in their academic careers. Researchers have used the program for such diverse fields as astronomy, oceanography and linguistics. Besides validating <a href="http://research.google.com/archive/mapreduce.html">MapReduce</a> as a useful tool in academic research, the program has also generated significant scientific knowledge. <br /><br />Three years later, there are many viable, affordable alternatives to the Academic Cloud Computing Initiative, so we have decided to bring our part of the program to a close. It has been a great opportunity to collaborate with IBM, the NSF and the many universities on this program. It was state-of-the-art four years ago when it was started; now, Academic Cloud Computing is a worldwide phenomena and there are many low-cost cloud computing options that provide viable alternatives to the Academic Cloud Computing Initiative.<div class="blogger-post-footer"><img width='1' height='1' src='https://blogger.googleusercontent.com/tracker/21224994-1809234908579026219?l=googleresearch.blogspot.com' alt='' /></div>]]></content:encoded>
			<wfw:commentRss>https://googledata.org/google-research/academic-successes-in-cluster-computing/feed/</wfw:commentRss>
		<slash:comments>0</slash:comments>
		</item>
		<item>
		<title>ACM Fellows for 2011</title>
		<link>https://googledata.org/google-research/acm-fellows-for-2011/?utm_source=rss&#038;utm_medium=rss&#038;utm_campaign=acm-fellows-for-2011</link>
		<comments>https://googledata.org/google-research/acm-fellows-for-2011/#comments</comments>
		<pubDate>Thu, 08 Dec 2011 15:30:00 +0000</pubDate>
		<dc:creator><![CDATA[Research Admin]]></dc:creator>
				<category><![CDATA[Google Research]]></category>

		<guid isPermaLink="false">https://googledata.org/?guid=cd59abf3d4d49800881aefdc8b651349</guid>
		<description><![CDATA[Posted by Alfred Spector, Google ResearchCross-posted with the Official Google BlogCongratulations to three Googlers elected ACM FellowsIt gives me great pleasure to share that the Association for Computing Machinery (ACM) has announced that three Goog...]]></description>
				<content:encoded><![CDATA[<span class="byline-author">Posted by Alfred Spector, Google Research</span><br /><br /><i>Cross-posted with the <a href="http://googleblog.blogspot.com/2011/12/congratulations-to-three-googlers.html">Official Google Blog</a></i><br /><br /><b>Congratulations to three Googlers elected ACM Fellows</b><br /><br />It gives me great pleasure to share that the <a href="http://www.acm.org/">Association for Computing Machinery</a> (ACM) has <a href="http://www.acm.org/press-room/news-releases/2011/fellows-2011/">announced</a> that three Googlers have been elected ACM Fellows in 2011. The ACM is the world’s largest educational and scientific computing society, and the Fellows Program celebrates the exceptional contributions of leaders in the computing field. This year the society has selected <a href="https://plus.google.com/115744399689614835150/about">Amit Singhal</a>, <a href="https://plus.google.com/110401818717224273095/posts">Peter S. Magnusson</a> and <a href="http://cseweb.ucsd.edu/~vahdat/">Amin Vahdat</a> for their outstanding work, which has provided fundamental knowledge to the field.<br /><br />The recently-named Fellows join 14 <a href="http://googleblog.blogspot.com/2010/12/four-googlers-elected-acm-fellows-this.html">prior Googler ACM Fellows</a> and other professional society honorees in exemplifying our extraordinarily talented people. On behalf of Google, I congratulate our colleagues. They embody Google’s commitment to innovation with impact, and I hope that they’ll serve as inspiration to students as well as the broader community of computer scientists.<br /><br />You can read more detailed summaries of their achievements below, including the official citations from the ACM. <br /><br /><b>Dr. Amit Singhal, Google Fellow</b><br /><br /><i>For contributions to search and information retrieval</i><br /><br />Since 2000, Dr. Amit Singhal has been pioneering search as the technical lead for Google's core search algorithms. He is credited with most of the information retrieval design decisions in Google Search – a massive system that has responded to hundreds of billions of queries. More than anyone, Amit has a deep understanding of Google’s entire algorithmic system. He is responsible for prioritization and has overseen the development of numerous algorithmic signals and their progression over time. He is the clear thought and managerial leader who has led critically important initiatives at the company. Among many other things, Amit catalyzed Universal Search, which returns multi-modal results from all available corpora; he was the force behind Realtime Search, which returns results from dynamic corpora with low latency; and he championed Google Instant, which returns search results as the user types. <br /><br />Prior to joining Google, Amit boasted a prolific publication record averaging 5 publications/year from 1996-9 while at AT&amp;T Labs. Since that time, you could say Google Search has been one long, sustained publication demonstrating a constant advancement in the state of the art of information retrieval. <br /><br /><br /><b>Peter S. Magnusson, Engineering Director</b><br /><br /><i>For contributions to full-system simulation</i><br /><br />Peter has made a tremendous impact by driving full-system simulation. His approach was so advanced, it can be used in real world production of commercial CPUs and prototyping of system software. Starting in 1991, Peter began to challenge the notion that simulators could not be made fast enough to run large workloads, nor accurate enough to run commercial operating systems. His innovations in simulator design culminated in Simics, the first academic simulator that could boot and run commercial multiprocessor workloads. Simics saw huge academic success and has been used to run simulations for research presented in several hundred subsequent publications. <br /><br />Peter founded Virtutech in 1998 to commercially develop Simics, and he ultimately forged and became the leader in a new market segment for software tools. With Peter at the helm, Virtutech pushed Simics beyond several performance barriers to make it the first simulator to exceed 1 billion instructions per second and the first simulator to model over 1,000 processors. Peter joined Google in 2010 to work with cloud computing.<br /><br /><br /><b>Dr. Amin Vahdat, Principal Engineer</b><br /><br /><i>For contributions to data center scalability and management</i><br /><br />Amin’s work made an impact at Google long before he arrived here. Amin is known for conducting research through bold, visionary projects that combine creativity with careful consideration of the engineering constraints needed to make them applicable in real world applications. Amin’s infrastructure ideas have underpinned the shift in the computing field from the pure client-server paradigm to a landscape in which major web services are hosted “in the cloud” across multiple data centers. In addition to pioneering “third-party cloud computing” through his work on WebOS and Rent-A-Server in the mid-90s, Amin has made important advancements in managing wide-area consistency between data centers, scalable modeling of data center applications, and building scalable data center networks. <br /><br />Amin’s innovations have penetrated and broadly influenced the networking community within academia and industry, including Google, and his research has been recapitulated and expanded upon in a number of publications. Conferences that formerly did not even cover data centers now have multiple sessions covering variants of what Amin and his team have proposed. At Google, Amin continues to drive next-generation data center infrastructure focusing on Software Defined Networking and new opportunities from optical technologies. This is emblematic of Amin’s ability to build real systems, and perhaps more significantly, convince people of their value.<div class="blogger-post-footer"><img width='1' height='1' src='https://blogger.googleusercontent.com/tracker/21224994-1347738637523703608?l=googleresearch.blogspot.com' alt='' /></div>]]></content:encoded>
			<wfw:commentRss>https://googledata.org/google-research/acm-fellows-for-2011/feed/</wfw:commentRss>
		<slash:comments>0</slash:comments>
		</item>
		<item>
		<title>Our second round of Google Research Awards for 2011</title>
		<link>https://googledata.org/google-research/our-second-round-of-google-research-awards-for-2011/?utm_source=rss&#038;utm_medium=rss&#038;utm_campaign=our-second-round-of-google-research-awards-for-2011</link>
		<comments>https://googledata.org/google-research/our-second-round-of-google-research-awards-for-2011/#comments</comments>
		<pubDate>Tue, 06 Dec 2011 16:00:00 +0000</pubDate>
		<dc:creator><![CDATA[Research Admin]]></dc:creator>
				<category><![CDATA[Google Research]]></category>

		<guid isPermaLink="false">https://googledata.org/?guid=f081be95537828b6ca714f9cbe113723</guid>
		<description><![CDATA[Posted by Maggie Johnson, Director of Education &#38; University RelationsWe’ve just finished the review process for the latest round of the Google Research Awards, which provide funding to full-time faculty working on research in areas of mutual int...]]></description>
				<content:encoded><![CDATA[<span class="byline-author">Posted by Maggie Johnson, Director of Education &amp; University Relations</span><br /><br />We’ve just finished the review process for the latest round of the <a href="http://research.google.com/university/relations/research_awards.html">Google Research Awards</a>, which provide funding to full-time faculty working on research in areas of mutual interest with Google. We are delighted to be funding 119 awards across 21 different focus areas for a total of $6 million. The subject areas that received the highest level of support this time were systems and infrastructure, human-computer interaction, social and mobile. In addition, 24% of the funding was awarded to universities outside the U.S.<br /><br />One way in which we measure the impact of the research award program is through surveys of Principal Investigators (PIs) and their Google sponsors (a Googler with whom grantees can discuss research directions, provide progress updates, engage in knowledge transfer, etc.). Here are some highlights from our most recent survey, covering projects funded over the last two years:<br /><br /><ul><li>433 papers were published as a result of a Google research award</li><li>126 projects made data sets or software publicly available</li><li>63 research talks were given by sponsored PIs at Google offices</li></ul><br />An important aspect of the program is that it often gives early career academics a head start on their research agenda.  Many new PIs commented on how a Google research award allowed them to explore their initial ideas and build a foundation for obtaining more significant funding from other sources. This type of seed funding is especially hard to get in the current economic environment.  <br /><br />The goal of the research award program is to initiate and sustain strong collaborations with our academic colleagues. The collaborations take many forms, from working on a project together, to co-writing a paper, to coming to Google to give a research talk. Whatever the form, the most important aspect is building strong relationships that last. Case in point, many of our <a href="http://research.google.com/university/relations/focused_research_awards.html">focused awards</a> (multi-year, unrestricted grants that include access to Google’s tools, technology and expertise) started as Google research awards.<br /><br />Congratulations to the <a href="http://services.google.com/fh/files/blogs/2011R2%20recipients%20for%20blog.pdf">well-deserving recipients of this round’s awards</a>, and if you are interested in applying for the next round (deadline is April 15), please visit <a href="http://research.google.com/university/relations/research_awards.html">our website</a> for more information.<div class="blogger-post-footer"><img width='1' height='1' src='https://blogger.googleusercontent.com/tracker/21224994-2153916703336370099?l=googleresearch.blogspot.com' alt='' /></div>]]></content:encoded>
			<wfw:commentRss>https://googledata.org/google-research/our-second-round-of-google-research-awards-for-2011/feed/</wfw:commentRss>
		<slash:comments>0</slash:comments>
		</item>
		<item>
		<title>2011 Google China Faculty Summit in Hangzhou</title>
		<link>https://googledata.org/google-research/2011-google-china-faculty-summit-in-hangzhou/?utm_source=rss&#038;utm_medium=rss&#038;utm_campaign=2011-google-china-faculty-summit-in-hangzhou</link>
		<comments>https://googledata.org/google-research/2011-google-china-faculty-summit-in-hangzhou/#comments</comments>
		<pubDate>Fri, 02 Dec 2011 19:30:00 +0000</pubDate>
		<dc:creator><![CDATA[Research Admin]]></dc:creator>
				<category><![CDATA[Google Research]]></category>
		<category><![CDATA[education]]></category>

		<guid isPermaLink="false">https://googledata.org/?guid=b5d970122ed0d03ca0679b8b4363af47</guid>
		<description><![CDATA[Posted by Aimin Zhu, University Relationship Manager, Google ChinaWe just wrapped up a highly successful 2011 Google China Faculty Summit in Hangzhou, China. On November 17 and 18, Googlers from China and the U.S. gathered with more than 80 faculty mem...]]></description>
				<content:encoded><![CDATA[<span class="byline-author">Posted by Aimin Zhu, University Relationship Manager, Google China</span><br /><br />We just wrapped up a highly successful 2011 Google China Faculty Summit in <a href="http://maps.google.com/maps?q=hangzhou,china&amp;hl=en&amp;ll=30.221102,120.146484&amp;spn=45.110857,79.013672&amp;sll=34.01609,-117.856509&amp;sspn=0.086083,0.154324&amp;vpsrc=6&amp;hnear=Hangzhou,+Zhejiang,+China&amp;t=m&amp;z=4">Hangzhou, China</a>. On November 17 and 18, Googlers from China and the U.S. gathered with more than 80 faculty members representing more than 45 universities and institutes, including Tsinghua University, Peking University and The Chinese Academy of Sciences. The two-day event revolved around the theme of “Communication, Exploration and Expansion,” with day one covering research and day two focusing on academic development. <br /><br />The summit provided a unique setting for both sides to share the results of their research and exchange ideas. Speakers included: <br /><br /><ul><li>Maggie Johnson, director of education and university relations at Google, presenting on innovation in Google research and global university relations programs,</li><li>Dr. Boon-Lock Yeo, head of engineering and research for Google China, providing an overview of  innovation in China engineering and corporate social responsibility efforts and accomplishments, and</li><li>Prof. Edward Chang, director of research for Google China, delivering a keynote on mobile information management and retrieval.</li></ul><br />The discussions on November 17 focused on two tracks, mobile computing and natural language processing, while discussions on November 18 focused on curriculum development with a special focus on Android app development. The attendees also spent time discussing joint research and development between universities and industry.<br /><br />This summit is part of a continuing to effort to collaborate with Chinese universities in order to support education in China. Click <a href="http://www.google.com/intl/zh-CN/corporate/university/en/index.html">here</a> for a list of the variety of education programs we have launched there in recent years. We look forward to expanding partnership opportunities in the future.<div class="blogger-post-footer"><img width='1' height='1' src='https://blogger.googleusercontent.com/tracker/21224994-621494373520213050?l=googleresearch.blogspot.com' alt='' /></div>]]></content:encoded>
			<wfw:commentRss>https://googledata.org/google-research/2011-google-china-faculty-summit-in-hangzhou/feed/</wfw:commentRss>
		<slash:comments>0</slash:comments>
		</item>
		<item>
		<title>Discovering Talented Musicians with Acoustic Analysis</title>
		<link>https://googledata.org/youtube/discovering-talented-musicians-with-acoustic-analysis/?utm_source=rss&#038;utm_medium=rss&#038;utm_campaign=discovering-talented-musicians-with-acoustic-analysis</link>
		<comments>https://googledata.org/youtube/discovering-talented-musicians-with-acoustic-analysis/#comments</comments>
		<pubDate>Wed, 02 Nov 2011 15:00:00 +0000</pubDate>
		<dc:creator><![CDATA[Research Admin]]></dc:creator>
				<category><![CDATA[Google Research]]></category>
		<category><![CDATA[Youtube]]></category>

		<guid isPermaLink="false">https://googledata.org/?guid=0af35c1e77e438c9f7db760b7e3b5a0c</guid>
		<description><![CDATA[Posted by Charles DuHadway, YouTube Slam Team, Google Research In an earlier post we talked about the technology behind Instant Mix for Music Beta by Google. Instant Mix uses machine hearing to characterize music attributes such as its timbre, mood and...]]></description>
				<content:encoded><![CDATA[<span class="byline-author">Posted by Charles DuHadway, YouTube Slam Team, Google Research </span><br /><br />In an <a href="http://googleresearch.blogspot.com/2011/06/instant-mix-for-music-beta-by-google.html">earlier post</a> we talked about the technology behind Instant Mix for <a href="http://music.google.com/">Music Beta by Google</a>. Instant Mix uses machine hearing to characterize music attributes such as its timbre, mood and tempo. Today we would like to talk about acoustic and visual analysis -- this time on YouTube. A fundamental part of YouTube's mission is to allow anyone anywhere to showcase their talents -- occasionally leading to <a href="http://www.youtube.com/watch?v=eQOFRZ1wNLw">life-changing success</a> -- but many talented performers are never discovered. Part of the problem is the sheer volume of videos: forty eight hours of video are uploaded to YouTube every minute (that’s eight years of content every day). We wondered if we could use acoustic analysis and machine learning to pore over these videos and automatically identify talented musicians.<br /><br />First we analyzed audio and visual features of videos being uploaded. We wanted to find “singing at home” videos -- often correlated with features such as ambient indoor lighting, head-and-shoulders view of a person singing in front of a fixed camera, few instruments and often a single dominant voice. Here’s a sample set of videos we found.<br /><br /><a href="http://2.bp.blogspot.com/-uIV4jRrylSI/TrB14eH590I/AAAAAAAAASc/xrHHAIj50X0/s1600/cover_wall_short.png" imageanchor="1" style="margin-left: 1em; margin-right: 1em;"><img border="0" src="http://2.bp.blogspot.com/-uIV4jRrylSI/TrB14eH590I/AAAAAAAAASc/xrHHAIj50X0/s1600/cover_wall_short.png" /></a><br /><br />Then we estimated the quality of singing in each video. Our approach is based on acoustic analysis similar to that used by Instant Mix, coupled with a small set of singing quality annotations from human raters. Given these data we used machine learning to build a ranker that predicts if an average listener would like a performance. <br /><br />While machines are useful for weeding through thousands of not-so-great videos to find potential stars, we know they alone can't pick the next great star. So we turn to YouTube users to help us identify the real hidden gems by playing a voting game called <a href="http://www.youtube.com/slam">YouTube Slam</a>. We're putting an equal amount of effort into the game itself -- how do people vote? What makes it fun? How do we know when we have a true hit? We're looking forward to your feedback to help us refine this process: <a href="http://www.youtube.com/slam/music/vote">give it a try</a>*.  You can also check out singer and voter <a href="http://www.youtube.com/slam/music">leaderboards</a>. Toggle “All time” to “Last week” to find emerging talent in fresh videos or all-time favorites. <br /><br />Our “Music Slam” has only been running for a few weeks and we have already found some very talented musicians. Many of the videos have less than 100 views when we find them.<br /><br /><div><table><tr><td><object class="BLOGGER-youtube-video" classid="clsid:D27CDB6E-AE6D-11cf-96B8-444553540000" codebase="http://download.macromedia.com/pub/shockwave/cabs/flash/swflash.cab#version=6,0,40,0" data-thumbnail-src="http://2.gvt0.com/vi/ZMZr83rwdNI/0.jpg" height="266" width="320"><param name="movie" value="http://www.youtube.com/v/ZMZr83rwdNI&fs=1&source=uds" /><param name="bgcolor" value="#FFFFFF" /><embed width="320" height="266"  src="http://www.youtube.com/v/ZMZr83rwdNI&fs=1&source=uds" type="application/x-shockwave-flash"></embed></object></td><td><object class="BLOGGER-youtube-video" classid="clsid:D27CDB6E-AE6D-11cf-96B8-444553540000" codebase="http://download.macromedia.com/pub/shockwave/cabs/flash/swflash.cab#version=6,0,40,0" data-thumbnail-src="http://3.gvt0.com/vi/OY2vWMtSsIM/0.jpg" height="266" width="320"><param name="movie" value="http://www.youtube.com/v/OY2vWMtSsIM&fs=1&source=uds" /><param name="bgcolor" value="#FFFFFF" /><embed width="320" height="266"  src="http://www.youtube.com/v/OY2vWMtSsIM&fs=1&source=uds" type="application/x-shockwave-flash"></embed></object></td></tr><tr><td><object class="BLOGGER-youtube-video" classid="clsid:D27CDB6E-AE6D-11cf-96B8-444553540000" codebase="http://download.macromedia.com/pub/shockwave/cabs/flash/swflash.cab#version=6,0,40,0" data-thumbnail-src="http://0.gvt0.com/vi/wRCPCNtViGA/0.jpg" height="266" width="320"><param name="movie" value="http://www.youtube.com/v/wRCPCNtViGA&fs=1&source=uds" /><param name="bgcolor" value="#FFFFFF" /><embed width="320" height="266"  src="http://www.youtube.com/v/wRCPCNtViGA&fs=1&source=uds" type="application/x-shockwave-flash"></embed></object></td><td><object class="BLOGGER-youtube-video" classid="clsid:D27CDB6E-AE6D-11cf-96B8-444553540000" codebase="http://download.macromedia.com/pub/shockwave/cabs/flash/swflash.cab#version=6,0,40,0" data-thumbnail-src="http://2.gvt0.com/vi/gBfynvifkOY/0.jpg" height="266" width="320"><param name="movie" value="http://www.youtube.com/v/gBfynvifkOY&fs=1&source=uds" /><param name="bgcolor" value="#FFFFFF" /><embed width="320" height="266"  src="http://www.youtube.com/v/gBfynvifkOY&fs=1&source=uds" type="application/x-shockwave-flash"></embed></object></td></tr><tr><td><object class="BLOGGER-youtube-video" classid="clsid:D27CDB6E-AE6D-11cf-96B8-444553540000" codebase="http://download.macromedia.com/pub/shockwave/cabs/flash/swflash.cab#version=6,0,40,0" data-thumbnail-src="http://2.gvt0.com/vi/LVFe6P-C7iY/0.jpg" height="266" width="320"><param name="movie" value="http://www.youtube.com/v/LVFe6P-C7iY&fs=1&source=uds" /><param name="bgcolor" value="#FFFFFF" /><embed width="320" height="266"  src="http://www.youtube.com/v/LVFe6P-C7iY&fs=1&source=uds" type="application/x-shockwave-flash"></embed></object></td><td><object class="BLOGGER-youtube-video" classid="clsid:D27CDB6E-AE6D-11cf-96B8-444553540000" codebase="http://download.macromedia.com/pub/shockwave/cabs/flash/swflash.cab#version=6,0,40,0" data-thumbnail-src="http://3.gvt0.com/vi/Qh-4qF07V1s/0.jpg" height="266" width="320"><param name="movie" value="http://www.youtube.com/v/Qh-4qF07V1s&fs=1&source=uds" /><param name="bgcolor" value="#FFFFFF" /><embed width="320" height="266"  src="http://www.youtube.com/v/Qh-4qF07V1s&fs=1&source=uds" type="application/x-shockwave-flash"></embed></object></td></tr></table></div><br /><br />And while we're excited about what we've done with music, there's as much undiscovered potential in almost any subject you can think of. Try our other slams: <a href="http://www.youtube.com/slam/cute">cute</a>, <a href="http://www.youtube.com/slam/bizarre">bizarre</a>, <a href="http://www.youtube.com/slam/comedy">comedy</a>, and <a href="http://www.youtube.com/slam/dance">dance</a>*. Enjoy!<br /><br />Related work by Google Researchers:<br />“<a href="http://research.google.com/pubs/pub35638.html">Video2Text: Learning to Annotate Video Content</a>”, <a href="http://research.google.com/pubs/author37818.html">Hrishikesh Aradhye</a>, <a href="http://research.google.com/pubs/author38233.html">George Toderici</a>, <a href="http://research.google.com/pubs/author36197.html">Jay Yagnik</a>, ICDM Workshop on Internet Multimedia Mining, 2009.<br /><br />* Music and dance slams are currently available only in the US.<div class="blogger-post-footer"><img width='1' height='1' src='https://blogger.googleusercontent.com/tracker/21224994-5952642640319180625?l=googleresearch.blogspot.com' alt='' /></div>]]></content:encoded>
			<wfw:commentRss>https://googledata.org/youtube/discovering-talented-musicians-with-acoustic-analysis/feed/</wfw:commentRss>
		<slash:comments>0</slash:comments>
<enclosure url="" length="" type="" />
		</item>
		<item>
		<title>Fresh Perspectives about People and the Web from Think Quarterly</title>
		<link>https://googledata.org/uncategorized/fresh-perspectives-about-people-and-the-web-from-think-quarterly/?utm_source=rss&#038;utm_medium=rss&#038;utm_campaign=fresh-perspectives-about-people-and-the-web-from-think-quarterly</link>
		<comments>https://googledata.org/uncategorized/fresh-perspectives-about-people-and-the-web-from-think-quarterly/#comments</comments>
		<pubDate>Wed, 28 Sep 2011 19:00:00 +0000</pubDate>
		<dc:creator><![CDATA[Research Admin]]></dc:creator>
				<category><![CDATA[Google Research]]></category>
		<category><![CDATA[Uncategorized]]></category>

		<guid isPermaLink="false">https://googledata.org/?guid=7e3109ebd8c899f9ed8aaa02afe32177</guid>
		<description><![CDATA[Posted by Allison Mooney, Christina Park, and Caroline McCarthy, The Think Quarterly TeamThere’s a lot of research, analysis and insights—from inside and outside Google—that we use in building our products and making decisions. To share what we]]></description>
				<content:encoded><![CDATA[<span class="byline-author">Posted by Allison Mooney, Christina Park, and Caroline McCarthy, The Think Quarterly Team</span><br /><br />There’s a lot of research, analysis and insights—from inside and outside Google—that we use in building our products and making decisions. To share what we’ve learned with our partners, we created <a href="http://www.google.com/url?q=http://www.thinkwithgoogle.com/quarterly%23utm_medium=Blogs&utm_campaign=Google+Research+Blog&utm_source=Google">Think Quarterly</a>. It’s intended to be a snapshot of what Google and other industry leaders are talking about and inspired by right now.<br /><br />Today we’re launching our second edition, the <a href="http://www.google.com/url?q=http://www.thinkwithgoogle.com/quarterly%23utm_medium=Blogs&utm_campaign=Google+Research+Blog&utm_source=Google">“People” issue</a>, exploring the latest technologies connecting us and the big ideas driving society forward. It also includes some of the research and analysis that helps us shape our strategies. <br /><br />For those who love data as much as we do, here are a few articles worth reading:<br /><br /><ul><li>“Following Generation Z,” in which Google research scientist Ed Chi details what he’s learned from monitoring the course of digital innovation and mapping patterns of digital technology use in the future</li></ul><ul><li>“Predicting the Present,” by chief economist Hal Varian, about how publicly available search tools can help anyone gain valuable insights into the behavior of web users and predict what they might do next</li></ul><ul><li>“Power to the People,” by Meg Pickard, anthropologist turned head of digital engagement at Guardian News and Media, about tracking the influence and power of online communities</li></ul><ul><li>“From Cash to Contentment,” about the use of happiness as a measurable metric of success, with insights coming from Nobel Prize winner Joseph Stiglitz</li></ul><br /><a href="http://www.google.com/url?q=http://www.thinkwithgoogle.com/quarterly%23utm_medium=Blogs&utm_campaign=Google+Research+Blog&utm_source=Google">Click here</a> to read all the articles, and if you have a suggestion for our next issue please tell us <a href="https://services.google.com/fb/forms/contactthinkquarterlyus/#utm_medium=Blogs&amp;utm_campaign=Google+Research+Blog&amp;utm_source=Google">here</a>. We hope you enjoy (and +1) it!<div class="blogger-post-footer"><img width='1' height='1' src='https://blogger.googleusercontent.com/tracker/21224994-751723561172313008?l=googleresearch.blogspot.com' alt='' /></div>]]></content:encoded>
			<wfw:commentRss>https://googledata.org/uncategorized/fresh-perspectives-about-people-and-the-web-from-think-quarterly/feed/</wfw:commentRss>
		<slash:comments>0</slash:comments>
		</item>
		<item>
		<title>Sorting Petabytes with MapReduce &#8211; The Next Episode</title>
		<link>https://googledata.org/uncategorized/sorting-petabytes-with-mapreduce-the-next-episode/?utm_source=rss&#038;utm_medium=rss&#038;utm_campaign=sorting-petabytes-with-mapreduce-the-next-episode</link>
		<comments>https://googledata.org/uncategorized/sorting-petabytes-with-mapreduce-the-next-episode/#comments</comments>
		<pubDate>Wed, 07 Sep 2011 23:50:00 +0000</pubDate>
		<dc:creator><![CDATA[Research Admin]]></dc:creator>
				<category><![CDATA[Google Research]]></category>
		<category><![CDATA[Uncategorized]]></category>

		<guid isPermaLink="false">https://googledata.org/?guid=a054571029d32be4fd79898888d24151</guid>
		<description><![CDATA[Posted by Grzegorz Czajkowski, Marián Dvorský, Jerry Zhao, and Michael Conley, Systems InfrastructureAlmost three years ago we announced results of the first ever "petasort" (sorting a petabyte-worth of 100-byte records, following the Sort Benchmark ...]]></description>
				<content:encoded><![CDATA[<span itemscope itemtype="http://schema.org/Article"><meta itemprop="name" content="Sorting Petabytes with MapReduce - The Next Episode"><meta itemprop="description" content="Sorting a ten petabyte input set took 6 hours and 27 minutes to complete on 8000 computers. We are not aware of any other sorting experiment successfully completed at this scale."></span><span class="byline-author">Posted by Grzegorz Czajkowski, Marián Dvorský, Jerry Zhao, and Michael Conley, Systems Infrastructure</span><br /><br />Almost three years ago we announced <a href="http://googleblog.blogspot.com/2008/11/sorting-1pb-with-mapreduce.html">results of the first ever "petasort"</a> (sorting a petabyte-worth of 100-byte records, following the <a href="http://sortbenchmark.org/">Sort Benchmark</a> rules).  It completed in just over six hours on 4000 computers.  Recently we repeated the experiment using 8000 computers.  The execution time was 33 minutes, an order of magnitude improvement.<br /><br />Our sorting code is based on <a href="http://labs.google.com/papers/mapreduce.html">MapReduce</a>, which is a key framework for running multiple processes simultaneously at Google. Thousands of applications, supporting most services offered by Google, have been expressed in MapReduce.  While not many MapReduce applications operate at a petabyte scale, some do.  Their scale is likely to continue growing quickly.  The need to help such applications scale motivated us to experiment with data sets larger than one petabyte.  In particular, sorting a ten petabyte input set took 6 hours and 27 minutes to complete on 8000 computers. We are not aware of any other sorting experiment successfully completed at this scale.<br /><br />We are excited by these results.  While internal improvements to the MapReduce framework contributed significantly, a large part of the credit goes to numerous advances in Google's hardware, cluster management system, and storage stack.  <br /><br />What would it take to scale MapReduce by further orders of magnitude and make processing of such large data sets efficient and easy? One way to find out is to join Google’s systems infrastructure team.  If you have a passion for distributed computing, are an expert or plan to become one, and feel excited about the challenges of exascale then definitely consider applying for a <a href="http://www.google.com/intl/en/jobs/swe/#src=storageresearchblog">software engineering position</a> with Google.<div class="blogger-post-footer"><img width='1' height='1' src='https://blogger.googleusercontent.com/tracker/21224994-1105872079913092982?l=googleresearch.blogspot.com' alt='' /></div>]]></content:encoded>
			<wfw:commentRss>https://googledata.org/uncategorized/sorting-petabytes-with-mapreduce-the-next-episode/feed/</wfw:commentRss>
		<slash:comments>0</slash:comments>
		</item>
		<item>
		<title>Google at the Joint Statistical Meetings in Miami</title>
		<link>https://googledata.org/google-research/google-at-the-joint-statistical-meetings-in-miami/?utm_source=rss&#038;utm_medium=rss&#038;utm_campaign=google-at-the-joint-statistical-meetings-in-miami</link>
		<comments>https://googledata.org/google-research/google-at-the-joint-statistical-meetings-in-miami/#comments</comments>
		<pubDate>Mon, 22 Aug 2011 22:06:00 +0000</pubDate>
		<dc:creator><![CDATA[Research Admin]]></dc:creator>
				<category><![CDATA[Google Research]]></category>

		<guid isPermaLink="false">https://googledata.org/?guid=e494fc42223cfef89976ce9bda0a7872</guid>
		<description><![CDATA[Posted by Marianna Dizik, StatisticianThe Joint Statistical Meetings (JSM) were held in Miami, Florida, this year. Nearly 5,000 participants from academia and industry came to present and discuss the latest in statistical research, methodology, and app...]]></description>
				<content:encoded><![CDATA[<span class="byline-author">Posted by Marianna Dizik, Statistician</span><br /><br />The Joint Statistical Meetings (JSM) were held in Miami, Florida, this year. Nearly 5,000 participants from academia and industry came to present and discuss the latest in statistical research, methodology, and applications.  Similar to previous years, several Googlers shared expertise in large-scale experimental design and implementation, statistical inference with massive datasets and forecasting, data mining, parallel computing, and much more.<br /><br />Our session "Statistics: The Secret Weapon of Successful Web Giants" attracted over one hundred people; surprising for an 8:30 AM session!  Revolution Analytics reviewed this in their official blog post <a href="http://blog.revolutionanalytics.com/2011/08/google-r-effective-ads.html">"How Google uses R to make online advertising more effective"</a><br /><br />The following talks were given by Googlers at JSM 2011.  Please check the upcoming Proceedings of the JSM 2011 for the full papers.<br /><br /><ul><li><a href="http://www.amstat.org/meetings/jsm/2011/onlineprogram/AbstractDetails.cfm?abstractid=300146">Statistical Plumbing: Effective use of classical statistical methods for large scale applications</a></li><ul><li>Author(s): Ni Wang, Yong Li, Daryl Pregibon, and Rachel Schutt</li></ul><li><a href="http://www.amstat.org/meetings/jsm/2011/onlineprogram/AbstractDetails.cfm?abstractid=300570">Parallel Computations in R, with Applications for Statistical Forecasting</a></li><ul><li>Author(s): Murray Stokely and Farzan Rohani and Eric Tassone</li></ul><li><a href="http://www.amstat.org/meetings/jsm/2011/onlineprogram/AbstractDetails.cfm?abstractid=300008">Conditional Regression Models</a></li><ul><li>Author(s): William D. Heavlin</li></ul><li><a href="http://www.amstat.org/meetings/jsm/2011/onlineprogram/AbstractDetails.cfm?abstractid=300051">The Effectiveness of Display Ads</a></li><ul><li>Author(s): Tim Hesterberg and Diane Lambert and David X. Chan and Or Gershony and Rong Ge</li></ul><li><a href="http://www.amstat.org/meetings/jsm/2011/onlineprogram/AbstractDetails.cfm?abstractid=300257">Measuring Ad Effectiveness Using Continuous Geo Experiments</a></li><ul><li>Author(s): Jon Vaver and Deepak Kumar and Jim Koehler</li></ul><li><a href="http://www.amstat.org/meetings/jsm/2011/onlineprogram/AbstractDetails.cfm?abstractid=300200">Post-Stratification and Network Sampling</a></li><ul><li>Author(s): Rachel Schutt and Andrew Gelman  and Tyler McCormick</li></ul></ul>Google has participated at JSM each year since 2004.  We have been increasing our involvement significantly by providing sponsorship, organizing and giving talks at sessions and roundtables, teaching courses and workshops, hosting a booth with new Google products demo, submitting posters, and more.  This year Googlers participated in sessions sponsored by ASA sections for Statistical Learning and Data Mining, Statistics and Marketing, Statistical Computing, Bayesian Statistical Science , Health Policy Statistics, Statistical Graphics, Quality and Productivity, Physical and Engineering Sciences, and Statistical Education.<br /><br />We also hosted the Google faculty reception, which was well-attended by faculty and their promising students. Google hires a growing number of statisticians and we were happy to participate in JSM again this year. People had a chance to talk to Googlers, ask about working here, encounter elements of Google culture (good food! T-shirts! 3D puzzles!), meet old and make new friends, and just have fun!<br /><br />Thanks to everyone that presented, attended, or otherwise engaged with the statistical community at JSM this year.  We’re looking forward to seeing you in San Diego next year.<div class="blogger-post-footer"><img width='1' height='1' src='https://blogger.googleusercontent.com/tracker/21224994-9079268293716491067?l=googleresearch.blogspot.com' alt='' /></div>]]></content:encoded>
			<wfw:commentRss>https://googledata.org/google-research/google-at-the-joint-statistical-meetings-in-miami/feed/</wfw:commentRss>
		<slash:comments>0</slash:comments>
		</item>
		<item>
		<title>A new MIT center for mobile learning, with support from Google</title>
		<link>https://googledata.org/google-research/a-new-mit-center-for-mobile-learning-with-support-from-google/?utm_source=rss&#038;utm_medium=rss&#038;utm_campaign=a-new-mit-center-for-mobile-learning-with-support-from-google</link>
		<comments>https://googledata.org/google-research/a-new-mit-center-for-mobile-learning-with-support-from-google/#comments</comments>
		<pubDate>Tue, 16 Aug 2011 16:00:00 +0000</pubDate>
		<dc:creator><![CDATA[Research Admin]]></dc:creator>
				<category><![CDATA[Google Research]]></category>

		<guid isPermaLink="false">https://googledata.org/?guid=0278ecd55c03d334292467da6949e699</guid>
		<description><![CDATA[Posted by Hal Abelson, Professor of Computer Science and Engineering, MITMIT and Google have a long-standing relationship based on mutual interests in education and technology. Today, we took another step forward in our shared goals with the establishm...]]></description>
				<content:encoded><![CDATA[<span class="byline-author">Posted by Hal Abelson, Professor of Computer Science and Engineering, MIT</span><br /><br />MIT and Google have a <a href="http://googleblog.blogspot.com/2011/05/celebrating-150-years-of-mit.html">long-standing relationship</a> based on mutual interests in education and technology. Today, we took another step forward in our shared goals with the establishment of the MIT Center for Mobile Learning, which will strive to transform learning and education through innovation in mobile computing. The new center will be actively engaged in studying and extending <a href="http://appinventor.googlelabs.com/about/">App Inventor for Android</a>, which Google recently announced it will be open sourcing.<br /><br />The new center, housed at MIT’s Media Lab, will focus on designing and studying new mobile technologies that enable people to learn anywhere, anytime, with anyone. The center was made possible in part by support from <a href="http://research.google.com/university/">Google University Relations</a> and will be run by myself and two distinguished MIT colleagues: Professors Eric Klopfer (science education) and Mitchel Resnick (media arts and sciences).<br /><br />App Inventor for Android—a programming system that makes it easy for learners to create mobile apps for Android smartphones—currently supports a community of about 100,000 educators, students and hobbyists. Through the new initiatives at the MIT Center for Mobile Learning, App Inventor will be connected to MIT’s premier research in educational technology and MIT’s long track record of creating and supporting open software.<br /><br />Google first launched App Inventor internally in order to move it forward with speed and focus, and then developed it to a point where it started to gain critical mass. Now, its impact can be amplified by collaboration with a top academic institution. At MIT, App Inventor will adopt an enriched research agenda with increased opportunities to influence the educational community. In a way, App Inventor has now come full circle, as I actually initiated App Inventor at Google by proposing it as a project during my sabbatical with the company in 2008. The core code for App Inventor came from Eric Klopfer’s <a href="http://education.mit.edu/projects">lab</a>, and the inspiration came from Mitch Resnick’s <a href="http://scratched.media.mit.edu/">Scratch project</a>. The new center is a perfect example of how industry and academia can collaborate effectively to create change enabled by technology, and we look forward to seeing what we can do next, together.<div class="blogger-post-footer"><img width='1' height='1' src='https://blogger.googleusercontent.com/tracker/21224994-5622208624483313484?l=googleresearch.blogspot.com' alt='' /></div>]]></content:encoded>
			<wfw:commentRss>https://googledata.org/google-research/a-new-mit-center-for-mobile-learning-with-support-from-google/feed/</wfw:commentRss>
		<slash:comments>0</slash:comments>
		</item>
		<item>
		<title>Our Faculty Institute brings faculty back to the drawing board</title>
		<link>https://googledata.org/uncategorized/our-faculty-institute-brings-faculty-back-to-the-drawing-board/?utm_source=rss&#038;utm_medium=rss&#038;utm_campaign=our-faculty-institute-brings-faculty-back-to-the-drawing-board</link>
		<comments>https://googledata.org/uncategorized/our-faculty-institute-brings-faculty-back-to-the-drawing-board/#comments</comments>
		<pubDate>Fri, 12 Aug 2011 18:00:00 +0000</pubDate>
		<dc:creator><![CDATA[Research Admin]]></dc:creator>
				<category><![CDATA[Google Research]]></category>
		<category><![CDATA[Uncategorized]]></category>
		<category><![CDATA[education]]></category>

		<guid isPermaLink="false">https://googledata.org/?guid=74cbabfc14f5de662985fff936fc7689</guid>
		<description><![CDATA[Posted by Nina Kim Schultz, Google Education ResearchCross-posted with the Official Google BlogSchool may still be out for summer, but teachers remain hard at work. This week, we hosted Google’s inaugural Faculty Institute at our Mountain View, Calif...]]></description>
				<content:encoded><![CDATA[Posted by Nina Kim Schultz, Google Education Research<br /><br /><i>Cross-posted with the <a href="http://googleblog.blogspot.com/2011/08/faculty-institute-brings-faculty-back.html">Official Google Blog</a></i><br /><br />School may still be out for summer, but teachers remain hard at work. This week, we hosted Google’s inaugural Faculty Institute at our Mountain View, Calif. headquarters. The three-day event was created for esteemed faculty from schools of education and math and science to explore teaching paradigms that leverage technology in K-12 classrooms. Selected via a rigorous nomination and application process, the 39 faculty members hail from 19 California State Universities (CSUs), as well as Stanford and UC Berkeley, and teach high school STEM (Science, Technology, Engineering and Math) teachers currently getting their teaching credentials. CSU programs credential 60 percent of California’s teachers—or 10 percent of all U.S. K-12 teachers—and one CSU campus alone can credential around 1,000 new teachers in a year. The purpose of gathering together at the Institute was to ensure our teachers’ teachers have the support they need to help educators adjust to a changing landscape.<br /><br />There is so much technology available to educators today, but unless they learn how to use it effectively, it does little to change what is happening in our classrooms. Without the right training and inspiration, interactive displays become merely expensive projection screens, and laptops simply replace paper rather than shifting the way teachers teach and students learn. Although the possibilities for technology use in schools are endless, teacher preparation for the 21st century classroom also has many constraints. For example: beyond the expense involved, there’s the time it costs educators to match a technological innovation to the improvement of pedagogy and curriculum; there’s a distinct shift in thinking that needs to take place to change classrooms; and there’s an essential challenge to help teachers develop the dispositions and confidence to be lifelong evaluators, learners and teachers of technology, instead of continuing to rely on traditional skill sets that will soon be outdated.<br /><br />The Institute featured keynote addresses from respected professors from Stanford and Berkeley, case studies from distinguished high school teachers from across California, hands-on technology workshops with a variety of Google and non-Google tools, and panels with professionals in the tech-education industry. Notable guests included representatives from <a href="http://www.teachforamerica.org/">Teach for America</a>, <a href="http://tntp.org/">The New Teacher Project</a>, the <a href="http://www.ed.gov/">Department of Education</a> and <a href="http://www.edutopia.org/">Edutopia</a>. Topics covered the ability to distinguish learning paths, how to use technology to transform classrooms into project-based, collaborative spaces and how to utilize a more interactive teaching style rather than the traditional lecture model.<br /><br />On the last day of the Institute, faculty members were invited to submit grant proposals to scale best practices outside of the meeting. Deans of the participating universities will convene at the end of the month to further brainstorm ways to scale new ideas in teacher preparation programs. Congratulations to all of the faculty members who were accepted into the inaugural Institute, and thank you for all that you do to help bring technology and new ways of thinking into the classroom. <br /><br /><style="clear: both;="" center;"="" text-align:="">
<a href="http://1.bp.blogspot.com/-_JrercX4bfo/TkVnu72IZwI/AAAAAAAAIYs/6ear903_-80/s1600/Faculty+Institute.jpeg" imageanchor="1" style="margin-left: 1em; margin-right: 1em;"><img border="0" src="http://1.bp.blogspot.com/-_JrercX4bfo/TkVnu72IZwI/AAAAAAAAIYs/6ear903_-80/Faculty+Institute.jpeg" width="500" /></a>
This program is a part of Google’s continued commitment to supporting STEM education.  Details on our other programs can be found on <a href="http://www.google.com/education">www.google.com/education</a>.</style="clear:><div class="blogger-post-footer"><img width='1' height='1' src='https://blogger.googleusercontent.com/tracker/21224994-8929157003305149385?l=googleresearch.blogspot.com' alt='' /></div>]]></content:encoded>
			<wfw:commentRss>https://googledata.org/uncategorized/our-faculty-institute-brings-faculty-back-to-the-drawing-board/feed/</wfw:commentRss>
		<slash:comments>0</slash:comments>
<enclosure url="" length="" type="" />
		</item>
		<item>
		<title>Culturomics, Ngrams and new power tools for Science</title>
		<link>https://googledata.org/uncategorized/culturomics-ngrams-and-new-power-tools-for-science/?utm_source=rss&#038;utm_medium=rss&#038;utm_campaign=culturomics-ngrams-and-new-power-tools-for-science</link>
		<comments>https://googledata.org/uncategorized/culturomics-ngrams-and-new-power-tools-for-science/#comments</comments>
		<pubDate>Wed, 10 Aug 2011 22:51:00 +0000</pubDate>
		<dc:creator><![CDATA[Research Admin]]></dc:creator>
				<category><![CDATA[Google Books]]></category>
		<category><![CDATA[Google Research]]></category>
		<category><![CDATA[Uncategorized]]></category>

		<guid isPermaLink="false">https://googledata.org/?guid=f1d882a81aa84e8ad4675e2db8464d45</guid>
		<description><![CDATA[Posted by Erez Lieberman Aiden and Jean-Baptiste Michel, Visiting Faculty at GoogleFour years ago, we set out to create a research engine that would help people explore our cultural history by statistically analyzing the world’s books. In January 201...]]></description>
				<content:encoded><![CDATA[<span class="byline-author">Posted by Erez Lieberman Aiden and Jean-Baptiste Michel, Visiting Faculty at Google</span><br /><br />Four years ago, we set out to create a research engine that would help people explore our cultural history by statistically analyzing the world’s books. In January 2011, the resulting method, <a href="http://www.culturomics.org/">culturomics</a>, was featured on the cover of the journal <i><a href="http://www.sciencemag.org/content/331/6014/176">Science</a></i>. More importantly, Google implemented and launched a web-based version of our prototype research engine, the Google Books Ngram Viewer.<br /><br />Now scientists, scholars, and web surfers around the world can take advantage of the Ngram Viewer to study a vast array of phenomena. And that's exactly what they've done. Here are a few of our favorite examples.<br /><br /><b>Poverty</b><br />Martin Ravallion, head of the Development Research Group at the World Bank, has been using the ngrams to study the history of poverty. In a <a href="http://www.psocommons.org/ppp/vol3/iss2/art2/">paper</a> published in the journal Poverty and Public Policy, he argues for the existence of two ‘poverty enlightenments’ marked by increased awareness of the problem: one towards the end of the 18th century, and another in the 1970s and 80s. But he makes the point that only the second of these enlightenments brought with it a truly enlightened idea: that poverty can be and should be completely <a href="http://ngrams.googlelabs.com/graph?content=eradicate+poverty&amp;year_start=1800&amp;year_end=2000&amp;corpus=0&amp;smoothing=3">eradicated</a>.<br /><br /><a href="http://ngrams.googlelabs.com/chart?content=eradicate%20poverty&amp;corpus=0&amp;smoothing=3&amp;year_start=1800&amp;year_end=2000"><img alt="" border="0" src="http://ngrams.googlelabs.com/chart?content=eradicate%20poverty&amp;corpus=0&amp;smoothing=3&amp;year_start=1800&amp;year_end=2000" style="cursor: hand; cursor: pointer; float: center; height: 247px; margin: 0 10px 10px 0; width: 550px;" /></a><br /><br /><b>The Science Hall of Fame</b><br />Adrian Veres and John Bohannon wondered who the most famous scientists of the past two centuries were. But there was no hall of fame for scientists, or a committee that determines who deserves to get into such a hall. So they used the ngrams data to define a metric for celebrity – the milliDarwin – and algorithmically created a <a href="http://www.sciencemag.org/site/feature/misc/webfeat/gonzoscientist/episode14/index.xhtml">Science Hall of Fame</a> listing the most famous scientists born since 1800. They found that things like a popular book or a major controversy did more to increase discussion of a scientist than, for instance, winning a Nobel Prize.<br /><br />(Other users have been exploring the history of particular sciences with the Ngram Viewer, covering everything from <a href="http://egosumdaniel.blogspot.com/2011/02/brief-history-of-neuroscience-in-google.html">neuroscience</a> to the <a href="http://www.theatlantic.com/technology/archive/2011/03/the-nuclear-century-in-google-ngrams/72461/">nuclear</a> age.)<br /><br /><br /><b style="font-weight: bold;">The History of Typography</b><br />When we introduced the Ngram Viewer, we pointed out some potential pitfalls with the data. For instance, the ‘medial s’ ( ſ ), an older form of the letter s that looked like an integral sign and appeared in the beginning or middle of words, tends to be classified as an instance of the letter ‘f’ by the OCR algorithm used to create our version of the data. Andrew West, blogging at <a href="http://babelstone.blogspot.com/2006/06/rules-for-long-s.html">Babelstone</a>, found a clever way to exploit this error: using queries like ‘husband’ and ‘hufband’ to study the history of medial s typography, he pinned down the precise moment when the medial s disappeared from English (around 1800), French (1780), and Spanish (1760).<br /><br />People are clearly having a good time with the Ngram Viewer, and they have been learning a few things about science and history in the process. Indeed, the tool has proven so popular and so useful that Google recently announced that its imminent graduation from Google Labs to become a permanent part of Google Books.<br /><br />Similar ‘big data’ approaches can also be applied to a wide variety of other problems. From books to maps to the structure of the web itself, 'the world's information' is one amazing dataset. <br /><i><br />Erez Lieberman Aiden is Visiting Faculty at Google and a Fellow of the Harvard Society of Fellows. Jean-Baptiste Michel is Visiting Faculty at Google and a Postdoctoral Fellow in Harvard's Department of Psychology.</i><div class="blogger-post-footer"><img width='1' height='1' src='https://blogger.googleusercontent.com/tracker/21224994-6009201590116295031?l=googleresearch.blogspot.com' alt='' /></div>]]></content:encoded>
			<wfw:commentRss>https://googledata.org/uncategorized/culturomics-ngrams-and-new-power-tools-for-science/feed/</wfw:commentRss>
		<slash:comments>0</slash:comments>
		</item>
		<item>
		<title>What You Capture Is What You Get: A New Way for Task Migration Across Devices</title>
		<link>https://googledata.org/uncategorized/what-you-capture-is-what-you-get-a-new-way-for-task-migration-across-devices/?utm_source=rss&#038;utm_medium=rss&#038;utm_campaign=what-you-capture-is-what-you-get-a-new-way-for-task-migration-across-devices</link>
		<comments>https://googledata.org/uncategorized/what-you-capture-is-what-you-get-a-new-way-for-task-migration-across-devices/#comments</comments>
		<pubDate>Tue, 12 Jul 2011 21:45:00 +0000</pubDate>
		<dc:creator><![CDATA[Research Admin]]></dc:creator>
				<category><![CDATA[Google Research]]></category>
		<category><![CDATA[Uncategorized]]></category>
		<category><![CDATA[android]]></category>

		<guid isPermaLink="false">https://googledata.org/?guid=3f9959847a30fd61e99f8cdf6849cd7a</guid>
		<description><![CDATA[Posted by Yang Li, Research Scientist  We constantly move from one device to another while carrying out everyday tasks. For example, we might find an interesting article on a desktop computer at work, then bring the article with us on a mobile phone du...]]></description>
				<content:encoded><![CDATA[<span class="byline-author">Posted by Yang Li, Research Scientist</span>  <br /><br />We constantly move from one device to another while carrying out everyday tasks. For example, we might find an interesting article on a desktop computer at work, then bring the article with us on a mobile phone during the commute and keep reading it on a laptop or a TV when we get home. Cloud computing and web applications have made it possible to access the same data and applications on different devices and platforms. However, there are not many ways to easily move tasks across devices that are as intuitive as drag-and-drop in a graphical user interface.<br /><br />Since last year, our research team started developing new technologies for users to easily migrate their tasks across devices. In a project named Deep Shot, we demonstrated how a user can easily move web pages and applications, such as Google Maps directions, between a laptop and an Android phone by using the phone camera. With Deep Shot, a user can simply take a picture of their monitor with a phone camera, and the captured content automatically shows up and becomes instantly interactive on the mobile phone.<br /><br />This project was inspired by our observations that many people tend to take a picture of map directions on the monitor using their mobile phone camera, rather than using other approaches such as email. Taking pictures feels more direct and convenient, and fits well our everyday activity that is often more opportunistic. Instead of just capturing raw pixels, Deep Shot recovers the actual contents and applications on the mobile phone based on these pixels. You can find out how Deep Shot keeps user interaction simple and what happens behind the scenes <a href="http://research.google.com/pubs/archive/37153.pdf">here</a>. Similar to WYSIWYG—What You See Is What You Get—for graphical user interfaces, Deep Shot demonstrates WYCIWYG—What You Capture Is What You Get—for cross-device interaction. We are exploring this interaction style for various task migration situations in our everyday life.<br /><br /><object width="320" height="266" class="BLOGGER-youtube-video" classid="clsid:D27CDB6E-AE6D-11cf-96B8-444553540000" codebase="http://download.macromedia.com/pub/shockwave/cabs/flash/swflash.cab#version=6,0,40,0" data-thumbnail-src="http://1.gvt0.com/vi/iGTM6xs2sck/0.jpg"><param name="movie" value="http://www.youtube.com/v/iGTM6xs2sck&fs=1&source=uds" /><param name="bgcolor" value="#FFFFFF" /><embed width="320" height="266"  src="http://www.youtube.com/v/iGTM6xs2sck&fs=1&source=uds" type="application/x-shockwave-flash"></embed></object><br /><br />Deep Shot remains a research project at Google. With increasing capabilities of mobile phones and fast growing web applications, we hope to explore more exciting ways to help users carry out their everyday activities.<div class="blogger-post-footer"><img width='1' height='1' src='https://blogger.googleusercontent.com/tracker/21224994-3074129224865597838?l=googleresearch.blogspot.com' alt='' /></div>]]></content:encoded>
			<wfw:commentRss>https://googledata.org/uncategorized/what-you-capture-is-what-you-get-a-new-way-for-task-migration-across-devices/feed/</wfw:commentRss>
		<slash:comments>0</slash:comments>
		</item>
		<item>
		<title>Languages of the World (Wide Web)</title>
		<link>https://googledata.org/google-research/languages-of-the-world-wide-web/?utm_source=rss&#038;utm_medium=rss&#038;utm_campaign=languages-of-the-world-wide-web</link>
		<comments>https://googledata.org/google-research/languages-of-the-world-wide-web/#comments</comments>
		<pubDate>Fri, 08 Jul 2011 00:15:00 +0000</pubDate>
		<dc:creator><![CDATA[Research Admin]]></dc:creator>
				<category><![CDATA[Google Research]]></category>

		<guid isPermaLink="false">https://googledata.org/?guid=064334517a0622437d5c0ec4068baa70</guid>
		<description><![CDATA[Posted by Daniel Ford and Josh BatsonThe web is vast and infinite. Its pages link together in a complex network, containing remarkable structures and patterns. Some of the clearest patterns relate to language.Most web pages link to other pages on the s...]]></description>
				<content:encoded><![CDATA[<span class="byline-author">Posted by Daniel Ford and Josh Batson</span><br /><br />The web is vast and infinite. Its pages link together in a complex network, containing remarkable structures and patterns. Some of the clearest patterns relate to language.<br /><br />Most web pages link to other pages on the same web site, and the few off-site links they have are almost always to other pages in the same language.&nbsp;It's as if each language has its own web which is loosely linked to the webs of other languages.&nbsp;However, there are a small but significant number of off-site links between languages. These give tantalizing hints of the world beyond the virtual.<br /><br />To see the connections between languages, start by taking the several billion most important pages on the web in 2008, including all pages in smaller languages, and look at the off-site links between these pages. The particular choice of pages in our corpus here reflects decisions about what is `important'. For example, in a language with few pages every page is considered important, while for languages with more pages some selection method is required, based on pagerank for example.<br /><br />We can use our corpus to draw a very simple graph of the web, with a node for each language and an edge between two languages if more than one percent of the&nbsp;offsite links in the first language land on pages in the second. To make things a little clearer, we only show the languages which have at least a hundred thousand pages and have a strong link with another language, meaning at least 1% of off-site links go to that language. We also leave out English, which we'll discuss more in a moment. (Figure 1)<br /><br />Looking at the language web in 2008, we see a surprisingly clear map of Europe and Asia. <br />The language linkages invite explanations around geopolitics, linguistics, and historical associations.<br /><br /><div><a href="http://1.bp.blogspot.com/-icTwf3P6f58/ThZM092RU2I/AAAAAAAAAdw/5nrMSXNz-5o/s1600/PR_mapped.gif" imageanchor="1" style="margin-left: 1em; margin-right: 1em;"><img border="0" height="259" src="http://1.bp.blogspot.com/-icTwf3P6f58/ThZM092RU2I/AAAAAAAAAdw/5nrMSXNz-5o/s400/PR_mapped.gif" width="400" /></a></div><br /><div style="text-align: center;"><i>Figure 1: Language links on the web.</i>&nbsp;</div><div style="text-align: center;"><br /></div>The outlines of the Iberian and Scandinavian Peninsulas are clearly visible, which suggest geographic rather than purely linguistic associations.<br /><br />Examining links between other languages, it seems that many are explained by people and communities which speak both languages.<br /><br />The language webs of many former Soviet republics link back to the Russian web, with the strongest link from Ukrainian.&nbsp;While Russia is the major importer of Ukrainian products, the bilingual nature of Ukraine is a more plausible explanation. Most Ukrainians speak both languages, and Russian is even the dominant language in large parts of the country.<br /><br />The link from Arabic to French speaks to the long connection between France and its former colonies. In many of these countries Arabic and French are now commonly spoken together, and there has been significant emigration from these countries to France.<br /><br />Another strong link is between the Malay/Malaysian and Indonesian webs. Malaysia and Indonesia share a border, but more importantly the languages are nearly eighty percent cognate, meaning speakers of one can easily understand the other.<br /><br />What about the sizes of each language web?&nbsp;Both the number of sites in each language and the number of urls seen by Google's crawler follow an exponential distribution, although the ordering for each is slightly different (Figure 2).&nbsp;The exact number of pages in each language in 2008 is unknown, since multiple urls may point to the same page and some pages may not have been seen at all.&nbsp;However, the language of an un-crawled url can be guessed by the dominant language of its site.&nbsp;In fact, calendar pages and other infinite spaces mean that there really are an unlimited number of pages on the web, though some are more useful than others.<br /><br /><div class="separator" style="clear: both; text-align: center;"><a href="http://1.bp.blogspot.com/-DtvO5pZqfgM/ThZCN1pto-I/AAAAAAAAAdI/IaPE0e0xINg/s1600/language_sizes.gif" imageanchor="1" style="margin-left: 1em; margin-right: 1em;"><img border="0" height="229" src="http://1.bp.blogspot.com/-DtvO5pZqfgM/ThZCN1pto-I/AAAAAAAAAdI/IaPE0e0xINg/s400/language_sizes.gif" width="400" /></a></div><div style="text-align: center;"><i>Figure 2: The number of sites and seen urls per language are roughly exponentially distributed.</i>&nbsp;</div><br />The largest language on the web, in terms of size and centrality, has always been English, but where is it on our map? <br /><br />Every language on the web has strong links to English, usually with around twenty percent of offsite links and occasionally&nbsp;over forty five percent, such as from Tagalog/Filipino, spoken in the Philippines, and Urdu, principally spoken in Pakistan (Figure 3).&nbsp;Both the Philippines and Pakistan are former British colonies where English is one of the two official languages.<br /><br /><div class="separator" style="clear: both; text-align: center;"><a href="http://1.bp.blogspot.com/-JjCwlKx32vY/ThZD6XszS9I/AAAAAAAAAdQ/yc1SlARd23s/s1600/EnglishLanguageGraph_.gif" imageanchor="1" style="margin-left: 1em; margin-right: 1em;"><img border="0" height="400" src="http://1.bp.blogspot.com/-JjCwlKx32vY/ThZD6XszS9I/AAAAAAAAAdQ/yc1SlARd23s/s400/EnglishLanguageGraph_.gif" width="379" /></a></div><div style="text-align: center;"><i>Figure 3: Language links to and from English</i>&nbsp;</div><br />You might wonder whether off-site links landing on English pages can be explained simply by the number of English pages available to be linked to.  The webs of other languages in our corpus typically have sixty to eighty percent of their out-language links to English pages. However, only 38 percent of the pages and 42 percent of sites in our set are English, while it attracts 79 percent of all out-language links from other languages.<br /><br />Chinese and Japanese also seem unusual because there are relatively few links from pages in these languages to pages in English. This is despite the fact that Japanese&nbsp;and Chinese&nbsp;sites are the most popular non-English sites for English sites to link to.&nbsp;However, the number of sites in a language is a strong predictor of its `introversion', or fraction of off-site links to pages in the same language. Taking this into account shows that Chinese and Japanese webs are not unusually introverted given their size.&nbsp;In general, language webs with more sites are more introverted, perhaps due to better availability of content.&nbsp;(Figure 4)<br /><br /><div class="separator" style="clear: both; text-align: center;"><a href="http://4.bp.blogspot.com/-OCZPh30IEwI/ThZEe9v0BDI/AAAAAAAAAdY/aFzEWL0aQmM/s1600/introversion.gif" imageanchor="1" style="margin-left: 1em; margin-right: 1em;"><img border="0" height="400" src="http://4.bp.blogspot.com/-OCZPh30IEwI/ThZEe9v0BDI/AAAAAAAAAdY/aFzEWL0aQmM/s400/introversion.gif" width="400" /></a></div><div style="text-align: center;"><i>Figure 4: Language size vs introversion.</i>&nbsp;</div><br />There is a roughly linear relationship between the (log) number of sites in a language and the fraction of off-site links which point to pages in the same language, with a correlation of 0.9 if English is removed.&nbsp;However, only 45 percent of off-site links from English pages are to other English pages, making English the most extroverted web language given its size. Other notable outliers are the Hindi web, which is unusually introverted, and the Tagalog and Malay webs which are unusually extroverted.<br /><br />We can generate another map by connecting languages if the number of links from one to the other is 50 times greater than expected given the number of out-of-language links and the size of the language linked to&nbsp;(Figure 5).&nbsp;This time, the native languages of India show up clearly.&nbsp;Surprising links include those from Hindi to Ukrainian, Kurdish to Swedish, Swahili to Tagalog and Bengali, and Esperanto to Polish.<br /><br /><div class="separator" style="clear: both; text-align: center;"><a href="http://3.bp.blogspot.com/-PYI7CG2COS4/ThZEm3MvTzI/AAAAAAAAAdg/ozF6XJeYAgo/s1600/GNP_mapped.gif" imageanchor="1" style="margin-left: 1em; margin-right: 1em;"><img border="0" height="259" src="http://3.bp.blogspot.com/-PYI7CG2COS4/ThZEm3MvTzI/AAAAAAAAAdg/ozF6XJeYAgo/s400/GNP_mapped.gif" width="400" /></a></div><div style="text-align: center;"><i>Figure 5: Unexpected connections, given the size of each language.</i>&nbsp;</div><br />What's happened since 2008? The languages of the web have become more densely connected. There is now significant content in even more languages, and these languages are more closely linked. We hope that tools like Google page translation, voice translation, and other services will accelerate this process and bring more people in the world closer together, whichever languages they speak.<br /><br /><br /><div style="margin-bottom: 0px; margin-left: 0px; margin-right: 0px; margin-top: 0px;"><b>UPDATE&nbsp;</b><b>9 July 2011:</b><i><b>&nbsp;</b>As has been pointed out in the comments, in</i><i>&nbsp;both the Philippines and Pakistan, English is one of the two&nbsp;</i><i>official languages; however, the Philippines was not a British colony.</i></div><div class="blogger-post-footer"><img width='1' height='1' src='https://blogger.googleusercontent.com/tracker/21224994-8351332047248996003?l=googleresearch.blogspot.com' alt='' /></div>]]></content:encoded>
			<wfw:commentRss>https://googledata.org/google-research/languages-of-the-world-wide-web/feed/</wfw:commentRss>
		<slash:comments>0</slash:comments>
<enclosure url="" length="" type="" />
		</item>
		<item>
		<title>Google Translate welcomes you to the Indic web</title>
		<link>https://googledata.org/google-research/google-translate-welcomes-you-to-the-indic-web-2/?utm_source=rss&#038;utm_medium=rss&#038;utm_campaign=google-translate-welcomes-you-to-the-indic-web-2</link>
		<comments>https://googledata.org/google-research/google-translate-welcomes-you-to-the-indic-web-2/#comments</comments>
		<pubDate>Tue, 21 Jun 2011 16:30:00 +0000</pubDate>
		<dc:creator><![CDATA[Research Admin]]></dc:creator>
				<category><![CDATA[Google Research]]></category>

		<guid isPermaLink="false">https://googledata.org/?guid=983c96aa57e09d3acb923dfbcd6d690a</guid>
		<description><![CDATA[Posted by Ashish Venugopal, Research Scientist   (Cross-posted on the Translate Blog and the Official Google Blog)&#160;Beginning today, you can explore the linguistic diversity of the Indian sub-continent with Google Translate, which now supports five...]]></description>
				<content:encoded><![CDATA[<span class="byline-author">Posted by Ashish Venugopal, Research Scientist </span>  <br /><br /><i>(Cross-posted on the <a href="http://googletranslate.blogspot.com/2011/06/google-translate-welcomes-you-to-indic.html">Translate Blog</a> and the <a href="http://googleblog.blogspot.com/2011/06/google-translate-welcomes-you-to-indic.html">Official Google Blog</a>)</i><br /><br /><a href="http://translate.google.com/?sl=ta&amp;tl=en&amp;q=%E0%AE%A8%E0%AE%B2%E0%AF%8D%E0%AE%B5%E0%AE%B0%E0%AE%B5%E0%AF%81" imageanchor="1" style="margin-left: 1em; margin-right: 1em;"><img border="0" height="50" id=":current_picnik_image" src="http://1.bp.blogspot.com/-ipbCokO5Zn8/TgC-igicMJI/AAAAAAAAIK8/6xRSNefw6bE/translate1.jpg" /></a><a href="http://translate.google.com/?sl=te&amp;tl=en&amp;q=%E0%B0%B8%E0%B1%81%E0%B0%B8%E0%B1%8D%E0%B0%B5%E0%B0%BE%E0%B0%97%E0%B0%A4%E0%B0%82" imageanchor="1" style="margin-left: 1em; margin-right: 1em;"><img border="0" height="50" src="http://4.bp.blogspot.com/--LBOgNAXEIM/TgC-h7dSlUI/AAAAAAAAIK4/g8AjAbD_Vd4/translate2.jpg" /></a><a href="http://translate.google.com/?sl=kn&amp;tl=en&amp;q=%E0%B2%B8%E0%B3%81%E0%B2%B8%E0%B3%8D%E0%B2%B5%E0%B2%BE%E0%B2%97%E0%B2%A4" imageanchor="1" style="margin-left: 1em; margin-right: 1em;"><img border="0" height="50" src="http://1.bp.blogspot.com/-oi-gX0KyC2Q/TgC-hHnwT3I/AAAAAAAAIK0/x_Xb-1OmUPA/translate3.jpg" /></a><br /><br /><div style="text-align: center;"><a href="http://translate.google.com/?sl=bn&amp;tl=en&amp;q=%E0%A6%B8%E0%A7%8D%E0%A6%AC%E0%A6%BE%E0%A6%97%E0%A6%A4%E0%A6%AE" imageanchor="1" style="margin-left: 1em; margin-right: 1em;"><img border="0" height="50" src="http://4.bp.blogspot.com/-25bN2mnY02c/TgC-gLgqtmI/AAAAAAAAIKw/SsbTnYFiJuk/translate4.jpg" style="cursor: move;" /></a>&nbsp;<a href="http://translate.google.com/?sl=gu&amp;tl=en&amp;q=%E0%AA%B8%E0%AB%8D%E0%AA%B5%E0%AA%BE%E0%AA%97%E0%AA%A4" imageanchor="1"><img border="0" height="50" src="http://4.bp.blogspot.com/-f3Lk44n8UNs/TgC-fD657yI/AAAAAAAAIKs/SfwJGtwsBSk/translate5.jpg" /></a></div><br />Beginning today, you can explore the linguistic diversity of the Indian sub-continent with <a href="http://translate.google.com/">Google Translate</a>, which now supports five new experimental alpha languages: Bengali, Gujarati, Kannada, Tamil and Telugu. In India and Bangladesh alone, more than 500 million people speak these five languages. Since 2009, we’ve launched a total of 11 alpha languages, bringing the current number of languages supported by Google Translate to 63.<br /><br /><a href="http://en.wikipedia.org/wiki/Languages_of_South_Asia">Indic languages</a> differ from English in many ways, presenting several exciting challenges when developing their respective translation systems. Indian languages often use the <a href="http://en.wikipedia.org/wiki/Subject_Object_Verb">Subject Object Verb (SOV) ordering</a>&nbsp;to form sentences, unlike English, which uses <a href="http://en.wikipedia.org/wiki/Subject_Verb_Object">Subject Verb Object (SVO) ordering</a>. This difference in sentence structure makes it harder to produce fluent translations; the more words that need to be reordered, the more chance there is to make mistakes when moving them. Tamil, Telugu and Kannada are also highly <a href="http://en.wikipedia.org/wiki/Agglutinative_language">agglutinative</a>, meaning a single word often includes affixes that represent additional meaning, like tense or number. Fortunately, our research to improve Japanese (an SOV language) translation helped us with the word order challenge, while our work translating languages like German, Turkish and Russian provided insight into the agglutination problem.<br /><br />You can expect translations for these new alpha languages to be less fluent and include many more untranslated words than some of our more mature languages—like Spanish or Chinese—which have much more of the web content that powers our <a href="http://www.youtube.com/watch?v=_GdSC1Z1Kzs&amp;feature=player_embedded">statistical machine translation approach</a>. Despite these challenges, we release alpha languages when we believe that they help people better access the multilingual web. If you notice incorrect or missing translations for any of our languages, please <a href="http://googletranslate.blogspot.com/2010/12/when-one-translation-just-isnt-enough.html">correct us</a>; we enjoy learning from our mistakes and your feedback helps us graduate new languages from alpha status. If you’re a translator, you’ll also be able to take advantage of our machine translated output when using the <a href="http://translate.google.com/support/toolkit/bin/answer.py?hl=en&amp;answer=147809">Google Translator Toolkit</a>.<br /><br />Since these languages each have their own unique scripts, we’ve enabled a transliterated input method for those of you without Indian language keyboards. For example, if you type in the word “nandri,” it will generate the Tamil word  நன்றி (<a href="http://translate.google.com/?sl=ta&amp;tl=en&amp;q=%E0%AE%A8%E0%AE%A9%E0%AF%8D%E0%AE%B1%E0%AE%BF">see what it means</a>). To see all these beautiful scripts in action, you’ll need to install fonts* for each language.<br /><br />We hope that the launch of these new alpha languages will help you better understand the Indic web and encourage the publication of new content in Indic languages, taking us five alpha steps closer to a web without language barriers.<br /><br />*Download the fonts for each language: <a href="http://salrc.uchicago.edu/resources/fonts/available/tamil/">Tamil</a>, <a href="http://salrc.uchicago.edu/resources/fonts/available/telugu/">Telugu</a>, <a href="http://salrc.uchicago.edu/resources/fonts/available/bengali/">Bengali</a>, <a href="http://salrc.uchicago.edu/resources/fonts/available/gujarati/">Gujarati</a> and <a href="http://salrc.uchicago.edu/resources/fonts/available/kannada/">Kannada</a>.<div class="blogger-post-footer"><img width='1' height='1' src='https://blogger.googleusercontent.com/tracker/21224994-6387100852149730726?l=googleresearch.blogspot.com' alt='' /></div>]]></content:encoded>
			<wfw:commentRss>https://googledata.org/google-research/google-translate-welcomes-you-to-the-indic-web-2/feed/</wfw:commentRss>
		<slash:comments>0</slash:comments>
<enclosure url="" length="" type="" />
		</item>
		<item>
		<title>Auto-Directed Video Stabilization with Robust L1 Optimal Camera Paths</title>
		<link>https://googledata.org/youtube/auto-directed-video-stabilization-with-robust-l1-optimal-camera-paths/?utm_source=rss&#038;utm_medium=rss&#038;utm_campaign=auto-directed-video-stabilization-with-robust-l1-optimal-camera-paths</link>
		<comments>https://googledata.org/youtube/auto-directed-video-stabilization-with-robust-l1-optimal-camera-paths/#comments</comments>
		<pubDate>Mon, 20 Jun 2011 15:00:00 +0000</pubDate>
		<dc:creator><![CDATA[Research Admin]]></dc:creator>
				<category><![CDATA[Google Research]]></category>
		<category><![CDATA[Youtube]]></category>

		<guid isPermaLink="false">https://googledata.org/?guid=b65a0e624b39ca8ea56f0a48f6867686</guid>
		<description><![CDATA[Posted by Matthias Grundmann, Vivek Kwatra, and Irfan Essa, Research TeamEarlier this year, we announced the launch of new features on the YouTube Video Editor, including stabilization for shaky videos, with the ability to preview them in real-time. Th...]]></description>
				<content:encoded><![CDATA[<span class="byline-author">Posted by <a href="http://research.google.com/pubs/author38919.html">Matthias Grundmann</a>, <a href="http://research.google.com/pubs/author38000.html">Vivek Kwatra</a>, and <a href="http://www.irfanessa.com/Work/Welcome.html">Irfan Essa</a>, Research Team</span><br /><br />Earlier this year, we <a href="http://youtube-global.blogspot.com/2011/03/lights-camera-edit-new-features-for.html">announced</a> the launch of new features on the <a href="http://www.youtube.com/editor">YouTube Video Editor</a>, including stabilization for shaky videos, with the ability to preview them in real-time. The core technology behind this feature is detailed in <a href="http://research.google.com/pubs/pub37041.html">this paper</a>, which will be presented at the IEEE International Conference on Computer Vision and Pattern Recognition (<a href="http://www.cvpr2011.org/">CVPR 2011</a>).<br /><br />Casually shot videos captured by handheld or mobile cameras suffer from significant amount of shake. Existing in-camera stabilization methods dampen high-frequency jitter but do not suppress low-frequency movements and bounces, such as those observed in videos captured by a walking person. On the other hand, most professionally shot videos usually consist of carefully designed camera configurations, using specialized equipment such as tripods or camera dollies, and employ ease-in and ease-out for transitions. Our goal was to devise a completely automatic method for converting casual shaky footage into more pleasant and professional looking videos.<br /><br /><iframe width="540" height="330" src="http://www.youtube.com/embed/0MiY-PNy-GU" frameborder="0" allowfullscreen></iframe><br /><br />Our technique mimics the cinematographic principles outlined above by automatically determining the best camera path using a robust optimization technique. The original, shaky camera path is divided into a set of segments, each approximated by either a constant, linear or parabolic motion. Our optimization finds the best of all possible partitions using a computationally efficient and stable algorithm.<br /><br />To achieve real-time performance on the web, we distribute the computation across multiple machines in the cloud. This enables us to provide users with a real-time preview and interactive control of the stabilized result. Above we provide a video demonstration of how to use this feature on the YouTube Editor. We will also demo this live at <a href="http://googleresearch.blogspot.com/2011/06/google-at-cvpr-2011.html">Google’s exhibition booth</a> in CVPR 2011.<br /><br />For further details, please read our <a href="http://research.google.com/pubs/archive/37041.pdf">paper</a>.<div class="blogger-post-footer"><img width='1' height='1' src='https://blogger.googleusercontent.com/tracker/21224994-7399904819779185895?l=googleresearch.blogspot.com' alt='' /></div>]]></content:encoded>
			<wfw:commentRss>https://googledata.org/youtube/auto-directed-video-stabilization-with-robust-l1-optimal-camera-paths/feed/</wfw:commentRss>
		<slash:comments>0</slash:comments>
<enclosure url="" length="" type="" />
		</item>
		<item>
		<title>Google at CVPR 2011</title>
		<link>https://googledata.org/uncategorized/google-at-cvpr-2011/?utm_source=rss&#038;utm_medium=rss&#038;utm_campaign=google-at-cvpr-2011</link>
		<comments>https://googledata.org/uncategorized/google-at-cvpr-2011/#comments</comments>
		<pubDate>Thu, 16 Jun 2011 21:00:00 +0000</pubDate>
		<dc:creator><![CDATA[Research Admin]]></dc:creator>
				<category><![CDATA[Google Research]]></category>
		<category><![CDATA[Uncategorized]]></category>

		<guid isPermaLink="false">https://googledata.org/?guid=27a512479c1b0989d1f19495c9b6cebb</guid>
		<description><![CDATA[Posted by Mei Han and Sergey Ioffe, Research TeamThe computer vision community will get together in Colorado Springs the week of June 20th for the IEEE International Conference on Computer Vision and Pattern Recognition (CVPR 2011). This year will see ...]]></description>
				<content:encoded><![CDATA[<span class="byline-author">Posted by Mei Han and Sergey Ioffe, Research Team</span><br /><br />The computer vision community will get together in Colorado Springs the week of June 20th for the <a href="http://www.cvpr2011.org/">IEEE International Conference on Computer Vision and Pattern Recognition</a> (CVPR 2011). This year will see a record number of people attending the  conference and 27 co-located workshops and tutorials. The registration was closed at 1500 attendees even before the conference started.<br /><br />Computer Vision is at the core of many Google products, such as <a href="http://images.google.com/">Image Search</a>, <a href="http://www.youtube.com/">YouTube</a>, <a href="http://maps.google.com/help/maps/streetview/">Street View</a>, <a href="https://picasaweb.google.com/">Picasa</a>, and <a href="http://www.google.com/mobile/goggles/#text">Goggles</a>, and as always, Google is involved in several ways with CVPR. <a href="http://research.google.com/pubs/author37792.html">Andrew Senior</a> is serving as an area chair of CVPR 2011 and many Googlers are reviewers. Googlers also co-authored these papers:<br /><br /><ul><li><a href="http://www.cs.washington.edu/homes/rahul/data/cvpr2011.pdf">Where's Waldo: Matching People in Images of Crowds</a> by Rahul Garg, Deva Ramanan, Steve Seitz, Noah Snavely</li><li><a href="http://static.googleusercontent.com/external_content/untrusted_dlcp/research.google.com/en/us/pubs/archive/37065.pdf">Visual and Semantic Similarity in ImageNet</a> by Thomas Deselaers, Vittorio Ferrari</li><li><a href="http://grail.cs.washington.edu/projects/mcba/pba.pdf">Multicore Bundle Adjustment</a> by Changchang Wu, Sameer Agarwal, Brian Curless, Steve Seitz</li><li><a href="http://static.googleusercontent.com/external_content/untrusted_dlcp/research.google.com/en/us/pubs/archive/37125.pdf">A Hierarchical Conditional Random Field Model for Labeling and Segmenting Images of Street Scenes</a> by Qixing Huang, Mei Han, Bo Wu, Sergey Ioffe</li><li><a href="http://static.googleusercontent.com/external_content/untrusted_dlcp/research.google.com/en/us/pubs/archive/36985.pdf">Kernelized Structural SVM Learning for Supervised Object Segmentation</a> by Luca Bertelli, Tianli Yu, Diem Vu, Salih Gokturk</li><li><a href="http://static.googleusercontent.com/external_content/untrusted_dlcp/research.google.com/en/us/pubs/archive/36931.pdf">Discriminative Tag Learning on YouTube Videos with Latent Sub-tags</a> by Weilong Yang, George Toderici</li><li><a href="http://static.googleusercontent.com/external_content/untrusted_dlcp/research.google.com/en/us/pubs/archive/37041.pdf">Auto-Directed Video Stabilization with Robust L1 Optimal Camera Paths</a> by Matthias Grundmann, Vivek Kwatra, Irfan Essa</li><li><a href="https://docs.google.com/viewer?a=v&amp;pid=explorer&amp;chrome=true&amp;srcid=0B_szji3RgxrBYjUzMTYyOGUtY2M1ZC00MmY1LTgxMjAtNmJjMTVlM2FlNjI5&amp;hl=en_US">Image Saliency: From Local to Global Context</a> by Meng Wang, Janusz Konrad, Prakash Ishwar, Yushi Jing, Henry Rowley</li></ul><br />If you are attending the conference, stop by Google’s exhibition booth. In addition to talking with Google researchers, you will get to see examples of exciting computer vision research that has made it into Google products including, among others, the following:<br /><br /><ul><li><i>Google Earth Facade Shadow Removal</i> by Mei Han, Vivek Kwatra, and Shengyang Dai<br />We will demonstrate our technique for removing shadows and other lighting/texture artifacts from building facades in Google Earth. We obtain cleaner, clearer, and more uniform textures which provide users with an improved visual experience.</li><li><i>Video Stabilization on YouTube Editor</i> by Matthias Grundmann, Vivek Kwatra, and Irfan Essa<br />Casually shot videos captured by handheld or mobile cameras suffer from significant amount of shake. In contrast, professionally shot video usually employs stabilization equipment such as tripods or camera dollies, and employ ease-in and ease-out for transitions. Our technique mimics these cinematographic principles, by optimally dividing the original, shaky camera path into a set of segments and approximating each with either constant, linear or parabolic motion using a computationally efficient and stable algorithm. We will showcase a live version of our algorithm, featuring real-time performance and interactive control, which is publicly available at youtube.com/editor.</li><li><i>Tag Suggest for YouTube</i> by George Toderici and Mehmet Emre Sargin<br />YouTube offers millions of users the opportunity to upload videos and share them with their friends. Many users would love to have their videos discoverable but don't annotate them properly. One new feature on YouTube that seeks to address this problem is tag prediction based on video content and independently based on text metadata.</li></ul><br /><i><b>6/17/2011 UPDATE:  "Posted by" was changed to include Sergey Ioffe.</b></i><div class="blogger-post-footer"><img width='1' height='1' src='https://blogger.googleusercontent.com/tracker/21224994-8277213501716138712?l=googleresearch.blogspot.com' alt='' /></div>]]></content:encoded>
			<wfw:commentRss>https://googledata.org/uncategorized/google-at-cvpr-2011/feed/</wfw:commentRss>
		<slash:comments>0</slash:comments>
		</item>
		<item>
		<title>Our first round of Google Research Awards for 2011</title>
		<link>https://googledata.org/uncategorized/our-first-round-of-google-research-awards-for-2011/?utm_source=rss&#038;utm_medium=rss&#038;utm_campaign=our-first-round-of-google-research-awards-for-2011</link>
		<comments>https://googledata.org/uncategorized/our-first-round-of-google-research-awards-for-2011/#comments</comments>
		<pubDate>Thu, 09 Jun 2011 13:00:00 +0000</pubDate>
		<dc:creator><![CDATA[Research Admin]]></dc:creator>
				<category><![CDATA[Google Research]]></category>
		<category><![CDATA[Uncategorized]]></category>

		<guid isPermaLink="false">https://googledata.org/?guid=984804d96ec245f44a33774fcd8c0cce</guid>
		<description><![CDATA[Posted by Maggie Johnson, Director of Education &#38; University Relations  We’ve just finished awarding the latest round of Google Research Awards, which provide funding to full-time faculty working on research in areas of mutual interest with Googl...]]></description>
				<content:encoded><![CDATA[<span class="byline-author">Posted by Maggie Johnson, Director of Education &amp; University Relations</span>  <br /><br />We’ve just finished awarding the latest round of <a href="http://research.google.com/university/relations/research_awards.html">Google Research Awards</a>, which provide funding to full-time faculty working on research in areas of mutual interest with Google. A record number of submissions came in this round, and we are delighted to be funding 112 awards across 21 different focus areas for a total of more than $6.75 million. The subject areas that received the highest level of support were systems and infrastructure, human computer interaction, Geo/maps and machine learning. Thanks to strong international collaborations, 23% of the funding in this round was awarded to universities outside the U.S. <br /><br />In prior years, we’ve used this blog post to highlight some of our top-ranked projects, but this year, we’d like to give you an inside look into how we determine the award recipients. <br /><br />Designating the awards involves a careful and detailed review process. First, we have a set of internal research leads, each a well-known expert in their field, review all the proposals in their area. They assess the proposals on merit, innovation, connection to Google’s products and services and fit with our overall research agenda. The research leads then assign several volunteer reviewers—culled from experts on their team or other Google engineers holding PhDs—to weigh each proposal.<br /><br />All these reviews are recorded in an internal grant administration system, and the research leads make their funding recommendations. These recommendations are aggregated and a series of committee meetings are run, one for each research area. The research lead attends, along with members of the university relations team and executives in research. This committee reviews each proposal that the research lead has recommended for funding, using the same criteria mentioned above. This additional review process may change the proposal rankings and sometimes brings back other proposals for reconsideration.<br /><br />Once the committee meetings are complete, we make the final funding decisions, which are based on the available budget and balancing the funding across research areas and geographic regions. The final decisions are reviewed one last time by research management, and then we distribute the awards to the selected faculty.<br /><br />As the number of submissions for these research awards continues to grow, we remain committed to a merit-based review process with effective checks and balances. Congratulations to the well-deserving <a href="http://services.google.com/fh/files/blogs/google_2011researchawards.pdf">recipients</a> of this round’s awards, and if you are interested in applying for the next round (deadline is August 1), please visit our <a href="http://research.google.com/university/relations/research_awards.html">website</a> for more information.<div class="blogger-post-footer"><img width='1' height='1' src='https://blogger.googleusercontent.com/tracker/21224994-6801816548280445029?l=googleresearch.blogspot.com' alt='' /></div>]]></content:encoded>
			<wfw:commentRss>https://googledata.org/uncategorized/our-first-round-of-google-research-awards-for-2011/feed/</wfw:commentRss>
		<slash:comments>0</slash:comments>
		</item>
		<item>
		<title>Instant Mix for Music Beta by Google</title>
		<link>https://googledata.org/google-research/instant-mix-for-music-beta-by-google/?utm_source=rss&#038;utm_medium=rss&#038;utm_campaign=instant-mix-for-music-beta-by-google</link>
		<comments>https://googledata.org/google-research/instant-mix-for-music-beta-by-google/#comments</comments>
		<pubDate>Wed, 08 Jun 2011 21:16:00 +0000</pubDate>
		<dc:creator><![CDATA[Research Admin]]></dc:creator>
				<category><![CDATA[Google Research]]></category>
		<category><![CDATA[android]]></category>

		<guid isPermaLink="false">https://googledata.org/?guid=8613934c70ab80f7f3834c275c1cb209</guid>
		<description><![CDATA[Posted by Douglas Eck, Research ScientistMusic Beta by Google was announced at the Day One Keynote of  Google I/O 2011.  This service allows users to stream their music collections from the cloud to any supported device, including a web browser.  It’...]]></description>
				<content:encoded><![CDATA[<span class="byline-author">Posted by Douglas Eck, Research Scientist</span><br /><br /><a href="http://music.google.com/">Music Beta</a> by Google was announced at the <a href="http://www.google.com/events/io/2011/sessions/android-momentum-mobile-and-more-at-google-i-o.html">Day One Keynote</a> of  <a href="http://www.google.com/events/io/2011/index-live.html">Google I/O 2011</a>.  This service allows users to stream their music collections from the cloud to any supported device, including a web browser.  It’s a first step in creating a platform that gives users a range of compelling music experiences. One key component of the product, <span style="font-weight: bold;">Instant Mix</span>, is a playlist generator developed by Google Research.  Instant Mix uses <span style="font-weight: bold;">machine hearing</span> to extract attributes from audio which can be used to answer questions such as  “Is there a Hammond B-3 organ?” (instrumentation / timbre), “Is it angry?” (mood), “Can I jog to it?” (tempo / meter) and so on.   <span style="font-weight: bold;">Machine learning</span> algorithms relate these audio features to what we know about music on the web, such as the fact that Jimmy Smith is a jazz organist or that Arcade Fire and Wolf Parade are similar artists.   From this we can predict similar tracks for a seed track and, with some additional sequencing logic, generate Instant Mix playlists from songs in a user’s locker.<br /><br />Because we combine audio analysis with information about which artists and albums go well together, we can use both dimensions of similarity to compare songs. If you pick a mellow track from an album, we will make a mellower playlist than if you pick a high energy track from the same album.   For example, here we compare short Instant Mixes made from two very different tracks by U2. The first Instant Mix comes from "Mysterious Ways," an upbeat, danceable track from <i>Achtung</i> <i>Baby</i> with electric guitar and heavy percussion.<br /><br /><object width="320" height="266" class="BLOG_video_class" id="BLOG_video-89e85fbd3185b23d" classid="clsid:D27CDB6E-AE6D-11cf-96B8-444553540000" codebase="http://download.macromedia.com/pub/shockwave/cabs/flash/swflash.cab#version=6,0,40,0"><param name="movie" value="http://www.youtube.com/get_player"><param name="bgcolor" value="#FFFFFF"><param name="allowfullscreen" value="true"><param name="flashvars" value="flvurl=http://v20.nonxt2.googlevideo.com/videoplayback?id%3D89e85fbd3185b23d%26itag%3D5%26app%3Dblogger%26ip%3D0.0.0.0%26ipbits%3D0%26expire%3D1319375092%26sparams%3Did,itag,ip,ipbits,expire%26signature%3DEF6FDBF69144446DE4A37FEA8DB5164692D4053.52BB21804AF4BEE55A31CBA2252DC20E2F856F94%26key%3Dck1&amp;iurl=http://video.google.com/ThumbnailServer2?app%3Dblogger%26contentid%3D89e85fbd3185b23d%26offsetms%3D5000%26itag%3Dw160%26sigh%3Dta2a3bWnTua08tbrESY9mNlTzzA&amp;autoplay=0&amp;ps=blogger"><embed src="http://www.youtube.com/get_player" type="application/x-shockwave-flash"width="320" height="266" bgcolor="#FFFFFF"flashvars="flvurl=http://v20.nonxt2.googlevideo.com/videoplayback?id%3D89e85fbd3185b23d%26itag%3D5%26app%3Dblogger%26ip%3D0.0.0.0%26ipbits%3D0%26expire%3D1319375092%26sparams%3Did,itag,ip,ipbits,expire%26signature%3DEF6FDBF69144446DE4A37FEA8DB5164692D4053.52BB21804AF4BEE55A31CBA2252DC20E2F856F94%26key%3Dck1&iurl=http://video.google.com/ThumbnailServer2?app%3Dblogger%26contentid%3D89e85fbd3185b23d%26offsetms%3D5000%26itag%3Dw160%26sigh%3Dta2a3bWnTua08tbrESY9mNlTzzA&autoplay=0&ps=blogger"allowFullScreen="true" /></object><br /><div><div><div><div><ol><li>U2 "Mysterious Ways"</li><li>David Bowie "Fame"</li><li>Oingo Boingo "Gratitude"</li><li><span class="Apple-style-span" style="-webkit-border-horizontal-spacing: 2px; -webkit-border-vertical-spacing: 2px; font-size: small;">Infectious Grooves “Spreck”</span></li><li><span class="Apple-style-span" style="-webkit-border-horizontal-spacing: 2px; -webkit-border-vertical-spacing: 2px; font-size: small;">Red Hot Chili Peppers “Special Secret Song Inside”</span></li></ol><div class="nobr">Compare this to a short Instant Mix made from a much more laid back U2 cut, "MLK" from the album<i> Unforgettable Fire.  </i>This track has delicate vocals on top of a sparse synthesizer background and no percussion.</div><br /><object width="320" height="266" class="BLOG_video_class" id="BLOG_video-17b7d4070d48cde3" classid="clsid:D27CDB6E-AE6D-11cf-96B8-444553540000" codebase="http://download.macromedia.com/pub/shockwave/cabs/flash/swflash.cab#version=6,0,40,0"><param name="movie" value="http://www.youtube.com/get_player"><param name="bgcolor" value="#FFFFFF"><param name="allowfullscreen" value="true"><param name="flashvars" value="flvurl=http://v17.nonxt3.googlevideo.com/videoplayback?id%3D17b7d4070d48cde3%26itag%3D5%26app%3Dblogger%26ip%3D0.0.0.0%26ipbits%3D0%26expire%3D1319375092%26sparams%3Did,itag,ip,ipbits,expire%26signature%3D23EB208FEBE8F6F60416FC5BF958254C12011FFD.1CB44192D8C2FCF0BF7C3016EF4272870B39EDBA%26key%3Dck1&amp;iurl=http://video.google.com/ThumbnailServer2?app%3Dblogger%26contentid%3D17b7d4070d48cde3%26offsetms%3D5000%26itag%3Dw160%26sigh%3DKxXq1nfHw0NeL0HiMLIn29Za93k&amp;autoplay=0&amp;ps=blogger"><embed src="http://www.youtube.com/get_player" type="application/x-shockwave-flash"width="320" height="266" bgcolor="#FFFFFF"flashvars="flvurl=http://v17.nonxt3.googlevideo.com/videoplayback?id%3D17b7d4070d48cde3%26itag%3D5%26app%3Dblogger%26ip%3D0.0.0.0%26ipbits%3D0%26expire%3D1319375092%26sparams%3Did,itag,ip,ipbits,expire%26signature%3D23EB208FEBE8F6F60416FC5BF958254C12011FFD.1CB44192D8C2FCF0BF7C3016EF4272870B39EDBA%26key%3Dck1&iurl=http://video.google.com/ThumbnailServer2?app%3Dblogger%26contentid%3D17b7d4070d48cde3%26offsetms%3D5000%26itag%3Dw160%26sigh%3DKxXq1nfHw0NeL0HiMLIn29Za93k&autoplay=0&ps=blogger"allowFullScreen="true" /></object><br /><div class="nobr"><ol><li><span class="Apple-style-span" style="-webkit-border-horizontal-spacing: 2px; -webkit-border-vertical-spacing: 2px; font-size: small;">U2 "MLK"</span></li><li style="text-align: -webkit-auto;"><span class="Apple-style-span" style="-webkit-border-horizontal-spacing: 2px; -webkit-border-vertical-spacing: 2px; font-size: small;">Jewel “Don’t”</span><span class="Apple-style-span" style="-webkit-border-horizontal-spacing: 2px; -webkit-border-vertical-spacing: 2px; font-size: small;"><span class="Apple-style-span" style="-webkit-border-horizontal-spacing: 0px; -webkit-border-vertical-spacing: 0px; font-size: 16px;"></span></span></li><li style="text-align: -webkit-auto;"><span class="Apple-style-span" style="-webkit-border-horizontal-spacing: 2px; -webkit-border-vertical-spacing: 2px; font-size: small;">Antony and the Johnsons “What Can I Do?”</span></li><li style="text-align: -webkit-auto;"><span class="Apple-style-span" style="-webkit-border-horizontal-spacing: 2px; -webkit-border-vertical-spacing: 2px; font-size: small;">The Beatles “And I Love Her”</span></li><li style="text-align: -webkit-auto;"><span class="Apple-style-span" style="-webkit-border-horizontal-spacing: 2px; -webkit-border-vertical-spacing: 2px; font-size: small;">Van Morrison “Crazy Love”</span></li></ol></div>As you can hear, the “Mysterious Ways” Instant Mix is funky, with strong percussion and high-energy vocals while the “MLK” mix carries on with that track's laid-back lullaby feeling.</div><div><br /></div><div>Our approach also allows us to create mixes from music in the long tail.  Are you the lead singer in an unknown Dylan cover band?  Even if your group is new or otherwise unknown,  Instant Mix can still use audio similarity to match your tracks to real Dylan tracks (provided, of course, that you sing like Bob and your band sounds like The Band).<br /><br />Our goal with Instant Mix is to build awesome playlists from your music collection.  We achieve this by using machine learning to blend a wide range of information sources, including features derived from the music audio itself. Though we’re still in beta, and still have a lot of work to do, we believe Instant Mix is a great tool for music discovery that stands out from the crowd. Give it a try!<br /><br />Further reading by Google Researchers:<br /><a href="http://research.google.com/pubs/archive/36608.pdf">Machine Hearing: An Emerging Field</a><br />Richard F. Lyon.<br /><br /><a href="http://research.google.com/pubs/archive/35269.pdf">Sound Ranking Using Auditory Sparse-Code Representations</a><br />Martin Rehn, Richard F. Lyon, Samy Bengio, Thomas C. Walters, Gal Chechik.<br /><br /><a href="http://arxiv.org/abs/1105.5196">Large-Scale Music Annotation and Retrieval: Learning to Rank in Joint  Semantic Spaces</a><br />Jason Weston, Samy Bengio, Philippe Hamel.</div></div></div></div><div class="blogger-post-footer"><img width='1' height='1' src='https://blogger.googleusercontent.com/tracker/21224994-1907577272290907174?l=googleresearch.blogspot.com' alt='' /></div>]]></content:encoded>
			<wfw:commentRss>https://googledata.org/google-research/instant-mix-for-music-beta-by-google/feed/</wfw:commentRss>
		<slash:comments>0</slash:comments>
		</item>
		<item>
		<title>Google Scribe: Now with automatic text for links and faster formatting options</title>
		<link>https://googledata.org/uncategorized/google-scribe-now-with-automatic-text-for-links-and-faster-formatting-options/?utm_source=rss&#038;utm_medium=rss&#038;utm_campaign=google-scribe-now-with-automatic-text-for-links-and-faster-formatting-options</link>
		<comments>https://googledata.org/uncategorized/google-scribe-now-with-automatic-text-for-links-and-faster-formatting-options/#comments</comments>
		<pubDate>Thu, 26 May 2011 18:00:00 +0000</pubDate>
		<dc:creator><![CDATA[Research Admin]]></dc:creator>
				<category><![CDATA[Google Research]]></category>
		<category><![CDATA[Uncategorized]]></category>

		<guid isPermaLink="false">https://googledata.org/?guid=113454a65f5cad165652bf98efa375fb</guid>
		<description><![CDATA[Posted by Kartik Singh and Kuntal Loya, Google Scribe teamSince Google Scribe's first release on Google Labs last year, we have been poring over your feedback and busy adding the top features you asked for. Today, we're excited to announce a new versio...]]></description>
				<content:encoded><![CDATA[<span class="byline-author">Posted by Kartik Singh and Kuntal Loya, Google Scribe team</span><br /><br />Since <a href="http://scribe.googlelabs.com/">Google Scribe's</a> first release on <a href="http://www.googlelabs.com/">Google Labs</a> last year, we have been poring over your feedback and busy adding the top features you asked for. Today, we're excited to announce a new version of Google Scribe that brings more features to word processing.<br /><br />Besides formatting, Google Scribe provides features that help you author high quality documents quickly:<br /><ol><li><b>Automatic text for links<br /></b>Adding a hyperlink to your document has been a two-step process of choosing the link and the text to display for it. Google Scribe now makes it easier. Just paste or type any link into your document and Google Scribe will set an appropriate link text.<br /><br /></li><li><b>Smart toolbar<br /></b>Do you repeatedly spend time reaching out to the toolbar to format your document ? To speed-up formatting, Google Scribe now displays an abridged toolbar close-by when you select a portion of the document.<br /><br /><a href="http://3.bp.blogspot.com/-zCvHxd3QlEI/Td6UvKEa7MI/AAAAAAAAAZY/xa1SusYQ9pk/s1600/floating+toolbar.png" onblur="try {parent.deselectBloggerImageGracefully();} catch(e) {}"><img style="display:block; margin:0px auto 10px; text-align:center;cursor:pointer; cursor:hand;width: 400px; height: 75px;" src="http://3.bp.blogspot.com/-zCvHxd3QlEI/Td6UvKEa7MI/AAAAAAAAAZY/xa1SusYQ9pk/s400/floating+toolbar.png" border="0" alt="" id="BLOGGER_PHOTO_ID_5611085723627875522" /></a><br /><br /></li><li><b>Text completion in 12 languages<br /></b>Google Scribe auto-completes text as you type. In addition to saving keystrokes, the suggestions indicate correct or popular phrases to use. Google Scribe now auto-detects  document language, so you no longer need to choose a language.<br /><br /><a href="http://3.bp.blogspot.com/-2Yv-FlQxOSA/Td6VyXJ1aCI/AAAAAAAAAZg/o7uCpWiczcE/s1600/autocomplete.png" onblur="try {parent.deselectBloggerImageGracefully();} catch(e) {}"><img style="display:block; margin:0px auto 10px; text-align:center;cursor:pointer; cursor:hand;width: 400px; height: 121px;" src="http://3.bp.blogspot.com/-2Yv-FlQxOSA/Td6VyXJ1aCI/AAAAAAAAAZg/o7uCpWiczcE/s400/autocomplete.png" border="0" alt="" id="BLOGGER_PHOTO_ID_5611086878191478818" /></a><br />You can view other applicable suggestions by clicking on the options button next to the Google Scribe icon and choosing “Show Multiple Suggestions”.<br /><br /><a href="http://3.bp.blogspot.com/-67jgZUnGZr4/Td6WhyEDdhI/AAAAAAAAAZo/1_C8hBki3Rk/s1600/multiple+suggestions.png" onblur="try {parent.deselectBloggerImageGracefully();} catch(e) {}"><img style="display:block; margin:0px auto 10px; text-align:center;cursor:pointer; cursor:hand;width: 278px; height: 60px;" src="http://3.bp.blogspot.com/-67jgZUnGZr4/Td6WhyEDdhI/AAAAAAAAAZo/1_C8hBki3Rk/s400/multiple+suggestions.png" border="0" alt="" id="BLOGGER_PHOTO_ID_5611087692868843026" /></a><br />We have extended auto-complete support to Arabic, Dutch, French, German, Hungarian, Italian, Polish, Portuguese, Russian, Spanish and Swedish in addition to English that we already supported.<br /><br /></li><li><b>Correct your document as you type<br /></b>Google Scribe now has basic support for checking spelling, punctuation and phrases in your document. Google Scribe underlines incorrect usage and clicking on underlined words or phrases will display a menu of suggested corrections to choose from.<br /><br /><a href="http://3.bp.blogspot.com/-SZRdxXiQ7vY/Td6Xjp-0dAI/AAAAAAAAAZw/L13bTx21seM/s1600/proofreading.png" onblur="try {parent.deselectBloggerImageGracefully();} catch(e) {}"><img style="display:block; margin:0px auto 10px; text-align:center;cursor:pointer; cursor:hand;width: 400px; height: 139px;" src="http://3.bp.blogspot.com/-SZRdxXiQ7vY/Td6Xjp-0dAI/AAAAAAAAAZw/L13bTx21seM/s400/proofreading.png" border="0" alt="" id="BLOGGER_PHOTO_ID_5611088824570770434" /></a><br />We are continuously working on expanding the list of proofreading features. Stay tuned.</li></ol>Try out the new Google Scribe at <a href="http://scribe.googlelabs.com">scribe.googlelabs.com</a> and <a href="http://www.google.com/url?sa=t&amp;source=web&amp;cd=2&amp;ved=0CCMQFjAB&amp;url=http://www.googlelabs.com/show_details?app_key=agtnbGFiczIwLXd3d3IVCxIMTGFic0FwcE1vZGVsGIvu5QEM&amp;ei=gJjeTZftD476sAPBpty0Bw&amp;usg=AFQjCNFfnW060OQwJVXCAjZH4KmFYvvuow&amp;sig2=8a3HEdT8CDwUJy3rfVSPlw">let us know what you think</a>.<div class="blogger-post-footer"><img width='1' height='1' src='https://blogger.googleusercontent.com/tracker/21224994-7406103754441681786?l=googleresearch.blogspot.com' alt='' /></div>]]></content:encoded>
			<wfw:commentRss>https://googledata.org/uncategorized/google-scribe-now-with-automatic-text-for-links-and-faster-formatting-options/feed/</wfw:commentRss>
		<slash:comments>0</slash:comments>
<enclosure url="" length="" type="" />
		</item>
		<item>
		<title>Google at ACL 2011</title>
		<link>https://googledata.org/google-research/google-at-acl-2011/?utm_source=rss&#038;utm_medium=rss&#038;utm_campaign=google-at-acl-2011</link>
		<comments>https://googledata.org/google-research/google-at-acl-2011/#comments</comments>
		<pubDate>Thu, 19 May 2011 00:00:00 +0000</pubDate>
		<dc:creator><![CDATA[Research Admin]]></dc:creator>
				<category><![CDATA[Google Research]]></category>

		<guid isPermaLink="false">https://googledata.org/?guid=884140cc2e20ab6e228f3778ed2c2672</guid>
		<description><![CDATA[Posted by Ryan McDonald and Fernando Pereira, Research TeamThe Annual Meeting of the Association for Computational Linguistics is one of the premier conferences for language and text technologies. Many employees at Google have strong roots in the commu...]]></description>
				<content:encoded><![CDATA[<span class="byline-author">Posted by Ryan McDonald and Fernando Pereira, Research Team</span><br /><br />The Annual Meeting of the <a href="http://www.aclweb.org/">Association for Computational Linguistics</a> is one of the premier conferences for language and text technologies. Many employees at Google have strong roots in the community of researchers that attend this meeting, including many of our researchers working on <a href="http://translate.google.com/">machine translation</a> and <a href="http://www.google.com/mobile/voice-search/">speech</a>.<br /><br />At <a href="http://www.acl2011.org/">this years conference</a>, Google is particularly well represented. The General Chair is <a href="http://research.google.com/pubs/author108.html">Dekang Lin</a> and a few Googlers are serving as technical <a href="http://www.acl2011.org/call.shtml">Area Chairs</a> (in addition to the plethora of Googlers that reviewed papers for the conference). Google is also a <a href="http://www.acl2011.org/sponsors.shtml">Platinum Sponsor</a> of ACL this year. <br /><br />Research advances at Google can be seen throughout the conference’s technical content. Below is a complete list of Googler-authored or co-authored papers in the main conference. We want to give special emphasis to this year’s best paper award, given to  “<a href="http://static.googleusercontent.com/external_content/untrusted_dlcp/research.google.com/en/us/pubs/archive/37071.pdf">Unsupervised Part-of-Speech Tagging with Bilingual Graph-Based Projections</a>” by CMU graduate student and Google intern <a href="http://www.cs.cmu.edu/~dipanjan/Home.html">Dipanjan Das</a> and his internship advisor <a href="http://research.google.com/pubs/author38945.html">Slav Petrov</a>. ACL is an extremely selective conference and this award speaks volumes to the importance of <a href="http://googleresearch.blogspot.com/2011/03/building-resources-to-syntactically.html">syntactic analysis</a> and using bilingual corpora to project syntactic resources from resource rich languages (like English) to other languages. Congratulations Dipanjan and Slav!<br /><br />Googlers are also involved in two of this year’s tutorials. <a href="http://research.google.com/pubs/author107.html">Marius Pasca</a> will present “<a href="http://www.acl2011.org/tutorials_10pasca.shtml">Web Search Queries as a Corpus</a>” and <a href="http://www.seas.upenn.edu/~kuzman/">Kuzman Ganchev</a> and his colleagues will  teach about “<a href="http://www.acl2011.org/tutorials_11druck.shtml">Rich Prior Knowledge in Learning for Natural Language Processing</a>”. Finally, <a href="http://research.google.com/pubs/author39008.html">Katja Filippova</a> and her colleagues are running a workshop on “<a href="https://sites.google.com/site/texttotext2011/">Monolingual Text-to-Text Generation</a>”.<br /><br />ACL will take place this year in Portland from June 19th to June 24th.<br /><br />Papers by Googlers (a * indicates a paper that will be linked to after the conference):<br /><br />Ranking Class Labels Using Query Sessions*<br />Marius Pasca<br /><br />Fine-Grained Class Label Markup of Search Queries*<br />Joseph Reisinger and Marius Pasca<br /><br /><a href="http://www.petrovi.de/data/acl11.pdf">Unsupervised Part-of-Speech Tagging with Bilingual Graph-Based Projections</a><br />Dipanjan Das and Slav Petrov<br /><br /><a href="http://1458142612872649252-a-1802744773732722657-s-sites.googlegroups.com/site/amarsubramanya/home/Pubs/clustorie-acl.pdf?attachauth=ANoY7cqSwbXkxnmC9UA62cXItr_zzOZLbemzaMAKbVxiChjOfEOViCt6MELyu2ezRn-8mDBfaRdmBsPnW75r9VZz1_C0rZ2Yq0HkdvnEMIkMN9vs1LiTIQ5jnGKX8vJyNISCE0O1S2RnwlEL3Kd1PLDym8Pc6mOTO-z3QBtBLLjLmZqyH4FJ_DSCIywZn0djcJl1ATp5Ft5Rr7YV6W-NDDT-gJmtUTJBkCZUvD21E1lXFrb3hpNbBE4=&attredirects=1">Large-Scale Cross-Document Coreference Using Distributed Inference and Hierarchical Models</a><br />Sameer Singh, Amarnag Subramanya, Fernando Pereira and Andrew McCallum<br /><br /><a href="http://www.ims.uni-stuttgart.de/~schuetze/piggyback11/piggyback11.pdf">Piggyback: Using Search Engines for Robust Cross-Domain Named Entity Recognition</a><br />Stefan Rüd, Massimiliano Ciaramita, Jens Müller and Hinrich Schütze<br /><br /><a href="http://www.csee.ogi.edu/~bodensta/2011-beam-predict.pdf">Beam-Width Prediction for Efficient Context-Free Parsing</a><br />Nathan Bodenstab, Aaron Dunlop, Keith Hall and Brian Roark<br /><br /><a href="http://research.mtv.corp.google.com:4444/pubversion?fp=7092811193925625484">Language-independent compound splitting with morphological operations</a><br />Klaus Macherey, Andrew Dai, David Talbot, Ashok Popat and Franz Och<br /><br /><a href="http://www.denero.org/content/pubs/acl11_denero_dual.pdf">Model-Based Aligner Combination Using Dual Decomposition</a><br />John DeNero and Klaus Macherey<br /><br /><a href="http://static.googleusercontent.com/external_content/untrusted_dlcp/research.google.com/en/us/pubs/archive/37011.pdf">Binarized Forest to String Translation</a><br />Hao Zhang, Licheng Fang, Peng Xu and Xiaoyun Wu<br /><br /><a href="http://www.ryanmcd.com/papers/ssl-sentiment-acl2011.pdf">Semi-supervised Latent Variable Models for Fine-grained Sentiment Analysis</a><br />Oscar Tackstrom and Ryan McDonald<div class="blogger-post-footer"><img width='1' height='1' src='https://blogger.googleusercontent.com/tracker/21224994-6474354340858041793?l=googleresearch.blogspot.com' alt='' /></div>]]></content:encoded>
			<wfw:commentRss>https://googledata.org/google-research/google-at-acl-2011/feed/</wfw:commentRss>
		<slash:comments>0</slash:comments>
		</item>
		<item>
		<title>Make beautiful interactive maps even faster with new additions to the Fusion Tables API</title>
		<link>https://googledata.org/google-research/make-beautiful-interactive-maps-even-faster-with-new-additions-to-the-fusion-tables-api/?utm_source=rss&#038;utm_medium=rss&#038;utm_campaign=make-beautiful-interactive-maps-even-faster-with-new-additions-to-the-fusion-tables-api</link>
		<comments>https://googledata.org/google-research/make-beautiful-interactive-maps-even-faster-with-new-additions-to-the-fusion-tables-api/#comments</comments>
		<pubDate>Tue, 10 May 2011 21:00:00 +0000</pubDate>
		<dc:creator><![CDATA[Research Admin]]></dc:creator>
				<category><![CDATA[Google Research]]></category>

		<guid isPermaLink="false">https://googledata.org/?guid=a7efd92e26cc073a025cff3a8825cb50</guid>
		<description><![CDATA[Posted by Rebecca Shapley, Jayant Madhavan, Rod McChesney, and Kathryn Hurley, Fusion Tables teamGoogle Fusion Tables is a modern data management and publishing web application that makes it easy to host, manage, collaborate on, visualize, and publish ...]]></description>
				<content:encoded><![CDATA[<span class="byline-author">Posted by Rebecca Shapley, Jayant Madhavan, Rod McChesney, and Kathryn Hurley, Fusion Tables team</span><br /><br /><a href="http://www.google.com/fusiontables/public/tour/index.html">Google Fusion Tables</a> is a modern data management and publishing web application that makes it easy to host, manage, collaborate on, visualize, and publish data tables online. Since we first <a href="http://googleresearch.blogspot.com/2009/06/google-fusion-tables.html">launched</a> Fusion Tables almost two years ago, we've seen tremendous interest and usage from dozens of areas, from journalists to scientists to open-data entrepreneurs, and have been excited to see the <a href="http://googleresearch.blogspot.com/2010/06/google-fusion-tables-celebrates-one.html">innovative applications</a> that our users have been able to rapidly build and publish.<br /><br />We've been working hard to enrich what Fusion Tables offers for customization and control of visual presentation. This past fall we added the ability to style the colors and icons of mapped data with a few clicks in the <a href="http://www.google.com/fusiontables">Fusion Tables web app</a>. This spring we made it easy to use HTML and customize what users see in the info window that appears after a click on the map. We’ve enjoyed seeing the impressive visualizations you have created.  Some, like the <a href="http://www.guardian.co.uk/news/datablog/2011/mar/31/deprivation-map-indices-multiple">Guardian’s map of deprivation in the UK</a>, were created strictly within the web app, while apps like the <a href="http://www.baycitizen.org/data/bike-accidents/">Bay Citizen’s Bike Accident tracker</a> and the <a href="http://www.texastribune.org/library/data/census-2010/">Texas Tribune’s Census 2010 interactive map</a> take advantage of the <a href="http://code.google.com/apis/fusiontables/">Fusion Tables SQL API</a> to do even more customization.  <br /><br /><a onblur="try {parent.deselectBloggerImageGracefully();} catch(e) {}" href="http://www.baycitizen.org/data/bike-accidents/"><img style="display:block; margin:0px auto 10px; text-align:center;cursor:pointer; cursor:hand;width: 400px; height: 324px;" src="http://1.bp.blogspot.com/-iEUmqJmK0yw/TciJ2yRvhhI/AAAAAAAAAYM/YBTzhH9XyZ0/s400/Screen+shot+2011-05-08+at+6.23.00+PM.jpg" border="0" alt=""id="BLOGGER_PHOTO_ID_5604881310564714002" /></a><br />Of course, it’s not always convenient to do everything through a web interface, and today we’re delighted to invite trusted testers to try out the new Fusion Tables Styling and Info Window API. Now developers will be able to set a table’s map colors and info windows with code. <br /><br />Even better, this new Styling and Info Window API will be part of the <a href="http://googlecode.blogspot.com/2010/11/introducing-google-apis-console-and-our.html">Google APIs Console</a>.  The Google APIs Console helps you manage projects and teams, provision access quotas, and view analytics and metrics on your API usage. It also offers sample code that supports the <a href="http://googlecode.blogspot.com/2011/03/making-auth-easier-oauth-20-for-google.html">OAuth 2.0</a> client key management flow you need to build secure apps for your users. <br /><br />So if you've been looking for a way to programmatically create highly-customizable map visualizations from data tables, check out our new APIs and let us know what you think! To become a trusted tester, please apply to join the <a href="https://groups.google.com/group/fusion-tables-api-trusted-testers">Google Group</a> and tell us a little bit about how you use the Fusion Tables API.<div class="blogger-post-footer"><img width='1' height='1' src='https://blogger.googleusercontent.com/tracker/21224994-8761725649509826373?l=googleresearch.blogspot.com' alt='' /></div>]]></content:encoded>
			<wfw:commentRss>https://googledata.org/google-research/make-beautiful-interactive-maps-even-faster-with-new-additions-to-the-fusion-tables-api/feed/</wfw:commentRss>
		<slash:comments>0</slash:comments>
<enclosure url="" length="" type="" />
		</item>
		<item>
		<title>Google at CHI 2011</title>
		<link>https://googledata.org/google-research/google-at-chi-2011/?utm_source=rss&#038;utm_medium=rss&#038;utm_campaign=google-at-chi-2011</link>
		<comments>https://googledata.org/google-research/google-at-chi-2011/#comments</comments>
		<pubDate>Thu, 05 May 2011 14:30:00 +0000</pubDate>
		<dc:creator><![CDATA[Research Admin]]></dc:creator>
				<category><![CDATA[Google Research]]></category>

		<guid isPermaLink="false">https://googledata.org/?guid=6e2ede7d46e66569f6e3852067c3f80d</guid>
		<description><![CDATA[Posted by Yang Li, Research ScientistCross-posted with the Technical Programs and Events Blog   Google has an increasing presence at ACM CHI: Conference on Human Factors in Computing Systems, which is the premiere conference for Human Computer Interact...]]></description>
				<content:encoded><![CDATA[<span class="byline-author">Posted by Yang Li, Research Scientist</span><br /><br /><span style="font-style:italic;">Cross-posted with the <a href="http://googletechprograms.blogspot.com/2011/05/google-at-chi-2011.html">Technical Programs and Events Blog</a></span><br /><br /><div class="section">   Google has an increasing presence at <a href="http://chi2011.org/">ACM CHI: Conference on Human Factors in Computing Systems</a>, which is the premiere conference for Human Computer Interaction research. Eight Google papers will appear at the conference. These papers not only touch on our core areas such as Search, <a href="http://www.google.com/chrome/">Chrome</a> and <a href="http://www.android.com/">Android</a> but also demonstrate our growing effort in new areas where HCI is essential, such as new search user interfaces, <a href="http://gesturesearch.googlelabs.com/">gesture-based interfaces</a> and cross-device interaction. They showcase our efforts to address user experiences in diverse situations. Googlers are playing active roles in the conference in many other ways too: participating in conference committees, hosting panels, organizing workshops and teaching courses, as well as running demos and 1:1 sessions at Google's booth.<br />  </div><br /><div class="section">   This year's CHI takes place in Vancouver, BC, from May 7th - 12th.</div><br />  <div class="section_title"><b>PAPERS</b></div>   <div><a href="http://yangl.org/pdf/gestureavatar-chi2011.pdf">Gesture Avatar: A Technique for Operating Mobile User Interfaces Using Gestures</a>, by Hao Lü, Yang Li*</div><div>   <a href="http://yangl.org/pdf/motiongestures-chi2011.pdf"><br />  User-Defined Motion Gestures for Mobile Interaction</a> by Jaime Ruiz, Yang Li*, Edward Lank<br />  </div><br /><div>   <a href="http://yangl.org/pdf/gesturestudy-chi2011.pdf">Experimental Analysis of Touch-Screen Gesture Designs in Mobile Environments</a> by Andrew Bragdon, Eugene Nelson, Yang Li*, Ken Hinckley</div>   <div><br />  <a href="http://chi2011.org/program/program.html#paper2150">Many Bills: Engaging Citizens through Visualizations of Congressional Legislation</a>  by Yannick Assogba, Irene Ros, Joan DiMicco, Matt McKeon*</div>   <div><br />  <a href="http://social.cs.uiuc.edu/papers/pdfs/youPivot-CHI2011.pdf">YouPivot: Improving Recall with Contextual Search</a> by Joshua Hailpern, Nicholas Jitkoff*, Andrew Warr*, Karrie Karahalios, Robert Sesek, Nik Shkrob<br />  </div><br /><div>   <a href="http://www.cs.brown.edu/~sk/Publications/Papers/Published/eok-mit-rep-ac-err-fb/paper.pdf">Oops, I Did It Again: Mitigating Repeated Access Control Errors on Facebook</a> by Serge Egelman, Andrew Oates*, Shriram Krishnamurthi<br />  </div><br /><div>   <a href="http://yangl.org/pdf/deepshot-chi2011.pdf">Deep Shot: A Framework for Migrating Tasks Across Devices Using Mobile Phone Cameras</a> by Tsung-Hsiang Chang, Yang Li*<br />  </div><br /><div>   <a href="http://yangl.org/pdf/doubleflip-chi2011.pdf">DoubleFlip: A Motion Gesture Delimiter for Mobile Interaction</a> by Jaime Ruiz, Yang Li*<br />  </div><br /><div class="section_title"><b>   WORKSHOPS</b><br />  </div> <a href="http://crowdresearch.org/chi2011-workshop/">Crowdsourcing and Human Computation: Systems, Studies and Platforms</a> by Michael Bernstein, Ed H. Chi*, Lydia B. Chilton, Björn Hartmann, Aniket Kittur, Robert C. Miller<br /><br /><div class="section_title"><b>   PANELS</b></div><div>   <a href="http://chi2011.org/program/program.html#S1078">Designing for User Experience: Academia &amp; Industry</a> by Joseph 'Jofish' Kaye, Elizabeth Buie, Jettie Hoonhout, Kristina Höök, Virpi Roto, Scott Jenson*, Peter Wright<br />  </div><br /><a href="http://chi2011.org/program/program.html#S1114">Festschrift Panel in Honor of Stuart K. Card</a> by Ed H. Chi*, Peter Pirolli, Bonnie John, Judith S Olson, Dan Russell*, Tom Moran<br /><br /><a href="http://replichi.org/wp-content/uploads/2011/01/RepliCHI-panel_CR.pdf">CHI Should be Replicating and Validating Results More: Discuss</a> by Max L. Wilson, Wendy Mackay, Ed H. Chi*, Michael Bernstein, Dan Russell*, Harold Thimbleby<br /><br /><div>   <a href="http://chi2011.org/program/program.html#S1198">Transferability of Research Findings: Context-Dependent or Model-Driven</a> by Ed H. Chi*, Mary Czerwinski, David Millen, Dave Randall, Gunnar Stevens, Volker Wulf, John Zimmerman<br />  </div><br /><div>   <a href="http://chi2011.org/program/program.html#S1214">The Future of Child-Computer Interaction</a> by Allison Druin, Gary Knell, Elliot Soloway, Dan Russell*, Elizabeth Mynatt, Yvonne Rogers</div>   <div class="section_title"><br /><b>   CASE</b> <b>STUDIES</b></div><div>   <a href="http://www.google.com/research/pubs/pub37064.html">From Basecamp to Summit: Scaling Field Research Across 9 Locations</a> by Jens Riegelsberger*, Audrey Yang*, Konstantin Samoylov*, Elizabeth Nunge*, Molly Stevens*, Patrick Larvie*<br />  </div><br /><div class="section_title"><b>   COURSES</b></div><div>   <a href="http://research.microsoft.com/en-us/um/people/sdumais/Logs-talk-HCIC-2010.pdf">Design and Analysis of Large Scale Log Studies</a> by Susan Dumais, Robin Jeffries*, Dan Russell*, Diane Tang*, Jaime Teevan<br />  </div><br /><div class="section_title"><b>   SIG MEETING</b><br />  </div> <a href="http://www.cc.gatech.edu/~yardi/pubs/Yardi_CHI11_SIG.pdf">Participatory Culture in the Age of Social Media</a> by Dana Rotman, Sarah Vieweg, Sarita Yardi, Ed H. Chi*, Jenny Preece, Ben Shneiderman, Peter Pirolli, Tom Glaisyer<br /><br /><i>Note: * denotes a Googler</i><div class="blogger-post-footer"><img width='1' height='1' src='https://blogger.googleusercontent.com/tracker/21224994-5820293800081400183?l=googleresearch.blogspot.com' alt='' /></div>]]></content:encoded>
			<wfw:commentRss>https://googledata.org/google-research/google-at-chi-2011/feed/</wfw:commentRss>
		<slash:comments>0</slash:comments>
		</item>
		<item>
		<title>Partnering with Tsinghua University to support education in Western China</title>
		<link>https://googledata.org/google-research/partnering-with-tsinghua-university-to-support-education-in-western-china/?utm_source=rss&#038;utm_medium=rss&#038;utm_campaign=partnering-with-tsinghua-university-to-support-education-in-western-china</link>
		<comments>https://googledata.org/google-research/partnering-with-tsinghua-university-to-support-education-in-western-china/#comments</comments>
		<pubDate>Thu, 14 Apr 2011 15:30:00 +0000</pubDate>
		<dc:creator><![CDATA[Research Admin]]></dc:creator>
				<category><![CDATA[Google Research]]></category>

		<guid isPermaLink="false">https://googledata.org/?guid=63a5058c0fa3d4350419dd31b4e87d9e</guid>
		<description><![CDATA[Posted by Aimin Zhu, China University RelationsWe’re excited to announce that we’ve teamed up with Tsinghua University to provide educational support to five major universities in Western China: Qinghai, Xinjiang, Guizhou, Ningxia, and Yunnan. Toge...]]></description>
				<content:encoded><![CDATA[<font class="byline-author">Posted by Aimin Zhu, China University Relations</font><br /><br />We’re excited to announce that we’ve teamed up with <a href="http://www.tsinghua.edu.cn/publish/th/index.html">Tsinghua University</a> to provide educational support to five major universities in Western China: <a href="http://www.qhu.edu.cn/">Qinghai</a>, <a href="http://www.xju.edu.cn/">Xinjiang</a>, <a href="http://www.gzu.edu.cn/">Guizhou</a>, <a href="http://www.nxu.edu.cn/english/index.htm">Ningxia</a>, and <a href="http://www.ynu.edu.cn/">Yunnan</a>. Together, we aim to:<br /><ul><li>Support faculty development by recognizing outstanding teachers, sponsoring published papers, and funding academic exchange and cooperation with other universities</li><li>Establish specialized curricula by creating new courses focused on advanced industrial and web technologies</li><li>Cultivate student talent by inspiring scientific and technological innovation through local activities and programs.</li></ul>A ceremony held at Tsinghua today kicked off what we expect to be a long and beneficial partnership to advance educational opportunities in the region.<div class="blogger-post-footer"><img width='1' height='1' src='https://blogger.googleusercontent.com/tracker/21224994-7249161153913544860?l=googleresearch.blogspot.com' alt='' /></div>]]></content:encoded>
			<wfw:commentRss>https://googledata.org/google-research/partnering-with-tsinghua-university-to-support-education-in-western-china/feed/</wfw:commentRss>
		<slash:comments>0</slash:comments>
		</item>
		<item>
		<title>1 billion core-hours of computational capacity for researchers</title>
		<link>https://googledata.org/google-research/1-billion-core-hours-of-computational-capacity-for-researchers/?utm_source=rss&#038;utm_medium=rss&#038;utm_campaign=1-billion-core-hours-of-computational-capacity-for-researchers</link>
		<comments>https://googledata.org/google-research/1-billion-core-hours-of-computational-capacity-for-researchers/#comments</comments>
		<pubDate>Thu, 07 Apr 2011 22:00:00 +0000</pubDate>
		<dc:creator><![CDATA[Research Admin]]></dc:creator>
				<category><![CDATA[Google Research]]></category>

		<guid isPermaLink="false">https://googledata.org/?guid=d2fd738ae3d58c801e526bdf39285518</guid>
		<description><![CDATA[Posted by Dan Belov, Principal Engineer and David Konerding, Software EngineerWe’re pleased to announce a new academic research grant program: Google Exacycle for Visiting Faculty. Through this program, we’ll award up to 10 qualified researchers wi...]]></description>
				<content:encoded><![CDATA[<span class="byline-author">Posted by Dan Belov, Principal Engineer and David Konerding, Software Engineer</span><br /><br />We’re pleased to announce a new academic research grant program: <a href="http://research.google.com/university/exacycle_program.html">Google Exacycle for Visiting Faculty</a>. Through this program, we’ll award up to 10 qualified researchers with at least 100 million core-hours each, for a total of 1 billion core-hours. The program is focused on large-scale, CPU-bound batch computations in research areas such as biomedicine, energy, finance, entertainment, and agriculture, amongst others. For example, projects developing large-scale genomic search and alignment, massively scaled Monte Carlo simulations, and sky survey image analysis could be an ideal fit.<br /><br />Exacycle for Visiting Faculty expands upon our current efforts through <a href="http://research.google.com/university/index.html">University Relations</a> to stimulate advances in science and engineering research, and awardees will participate through the <a href="http://research.google.com/university/relations/visiting_faculty.html">Visiting Faculty Program</a>.  We invite full-time faculty members from universities worldwide to apply. All grantees, including those outside of the U.S., will work on-site at specific Google offices in the U.S. or abroad. The exact Google office location will be determined at the time of project selection.<br /><br />We are excited to accept proposals starting today. The application deadline is 11:59 p.m. PST May 31, 2011.  Applicants are encouraged to send in their proposals early as awards will be granted starting in June.<br /><br />More information and details on how to apply for a Google Exacycle for Visiting Faculty grant can be found on the <a href="http://research.google.com/university/exacycle_program.html">Google Exacycle for Visiting Faculty website</a>.<div class="blogger-post-footer"><img width='1' height='1' src='https://blogger.googleusercontent.com/tracker/21224994-9183793493357473422?l=googleresearch.blogspot.com' alt='' /></div>]]></content:encoded>
			<wfw:commentRss>https://googledata.org/google-research/1-billion-core-hours-of-computational-capacity-for-researchers/feed/</wfw:commentRss>
		<slash:comments>0</slash:comments>
		</item>
		<item>
		<title>Overlapping Experiment Infrastructure: More, Better, Faster Experimentation</title>
		<link>https://googledata.org/google-research/overlapping-experiment-infrastructure-more-better-faster-experimentation/?utm_source=rss&#038;utm_medium=rss&#038;utm_campaign=overlapping-experiment-infrastructure-more-better-faster-experimentation</link>
		<comments>https://googledata.org/google-research/overlapping-experiment-infrastructure-more-better-faster-experimentation/#comments</comments>
		<pubDate>Tue, 05 Apr 2011 00:00:00 +0000</pubDate>
		<dc:creator><![CDATA[Research Admin]]></dc:creator>
				<category><![CDATA[Google Research]]></category>

		<guid isPermaLink="false">https://googledata.org/?guid=a7d404405ae9f3ca7f389861362dce28</guid>
		<description><![CDATA[Posted by Deirdre O'Brien and Diane Tang, Adwords TeamAt Google, experimentation is practically a mantra; we evaluate almost every change that potentially affects what our users experience. Such changes include not only obvious user-visible changes suc...]]></description>
				<content:encoded><![CDATA[<span class="byline-author">Posted by Deirdre O'Brien and Diane Tang, Adwords Team</span><br /><br />At Google, experimentation is practically a mantra; we evaluate almost every change that potentially affects what our users experience. Such changes include not only obvious user-visible changes such as modiﬁcations to a user interface, but also more subtle changes such as different machine learning algorithms that might affect ranking or content selection. Our insatiable appetite for experimentation has led us to tackle the problems of how to run more experiments, how to run experiments that produce better decisions, and how to run them faster.<br /><br />Google's infrastructure supports this vast experimentation by using orthogonal diversion criteria for experiments in different "layers" so that each event (e.g. a web search) can be assigned to multiple experiments. The treatment and population sample are easily specified in data files allowing for fast and accurate experiment set up. We have also developed analytical tools to do experiment sizing and a metrics dashboard which provides summarized data within hours of experiment set up. Decision making is improved by the consistency and accuracy in metrics assured by these tools. We believe that Google's experimental system and processes described in <a href="http://www.google.com/research/pubs/pub36500.html">this paper</a> can be generalized and applied by any entity interested in using experimentation to improve search engines and other web applications.<div class="blogger-post-footer"><img width='1' height='1' src='https://blogger.googleusercontent.com/tracker/21224994-1349033063210880743?l=googleresearch.blogspot.com' alt='' /></div>]]></content:encoded>
			<wfw:commentRss>https://googledata.org/google-research/overlapping-experiment-infrastructure-more-better-faster-experimentation/feed/</wfw:commentRss>
		<slash:comments>0</slash:comments>
		</item>
		<item>
		<title>Ig-pay Atin-lay Oice-vay Earch-say</title>
		<link>https://googledata.org/google-research/ig-pay-atin-lay-oice-vay-earch-say/?utm_source=rss&#038;utm_medium=rss&#038;utm_campaign=ig-pay-atin-lay-oice-vay-earch-say</link>
		<comments>https://googledata.org/google-research/ig-pay-atin-lay-oice-vay-earch-say/#comments</comments>
		<pubDate>Fri, 01 Apr 2011 14:15:00 +0000</pubDate>
		<dc:creator><![CDATA[Research Admin]]></dc:creator>
				<category><![CDATA[Google Research]]></category>

		<guid isPermaLink="false">https://googledata.org/?guid=3dfb85e7a61a2068d131dcf8dab93749</guid>
		<description><![CDATA[Posted by Martin Jansche and Alex Salcianu, Google Speech TeamAs you might know, Google Voice Search is available in more than two dozen languages and dialects, making it easy to perform Google searches just by speaking into your phone.Today it is our ...]]></description>
				<content:encoded><![CDATA[<span class="byline-author">Posted by Martin Jansche and Alex Salcianu, Google Speech Team</span><br /><br />As you might know, <a href="http://www.google.com/mobile/voice-search/">Google Voice Search</a> is available in more than <a href="http://googleresearch.blogspot.com/2011/03/word-of-mouth-introducing-voice-search.html">two dozen languages and dialects</a>, making it easy to perform Google searches just by speaking into your phone.<br /><br />Today it is our pleasure to announce the launch of Pig Latin Voice Search!<br /><br />What is Pig Latin you may ask? <a href="http://en.wikipedia.org/wiki/Pig_Latin">Wikipedia describes it</a> as a language game where, for each English word, the first consonant (or consonant cluster) is moved to the end of the word and an “ay” is affixed (for example, “pig” yields “ig-pay” and “search” yields “earch-say”).<br /><br />Our Pig Latin Voice Search is even more fun than our other languages, because when you speak in Pig Latin, our speech recognizer not only recognizes your piggy speech but also translates it automagically to normal English and does a Google search.<br /><br /><iframe allowfullscreen="" frameborder="0" height="330" src="http://www.youtube.com/embed/Zg7Wayo1rLU" title="YouTube video player" width="540"></iframe><br /><br />To configure Pig Latin Voice Search in your Android phone just go to Settings, select “Voice input &amp; output settings”, and then “Voice recognizer settings”. In the list of languages you’ll see Pig Latin. Just select it and you are ready to roll in the mud!<br /><br />It also works on iPhone with the <a href="http://www.google.com/mobile/iphone/">Google Search app</a>. In the app, tap the Settings icon, then "Voice Search" and select Pig Latin.<br /><a href="http://market.android.com/search?q=pname:com.google.android.voicesearch" onblur="try {parent.deselectBloggerImageGracefully();} catch(e) {}"><img alt="" border="0" id="BLOGGER_PHOTO_ID_5590444973392737538" src="http://2.bp.blogspot.com/-_rpA8uUyhic/TZVAFXdxNQI/AAAAAAAAAWc/gxWgJcHKHiA/s400/qr-code-voice-search.png" style="cursor: hand; cursor: pointer; float: right; height: 230px; margin: 0 0 10px 10px; width: 230px;" /></a><br />Ave-hay un-fay ith-way Ig-pay Atin-lay.<br /><br /><br />Pig Latin Voice Search works on Android 2.2 (Froyo) and later Android versions. If you don't already have Google Voice Search on your Android phone, scan or tap this QR code to download it.<br /><br />The list of languages and dialects now supported by Google Voice Search includes:<br /><ul><li>US English, UK English, Australian English, Indian English, South African English</li><li>Spanish from Spain, Mexico, Argentina, and Latin America</li><li>French (France), Italian, and Portuguese (Brazil)</li><li>German (Germany) and Dutch</li><li>Russian, Polish, and Czech</li><li>Turkish</li><li>Japanese, Korean, Mandarin (Mainland China and Taiwan), and Cantonese</li><li>Bahasa Indonesia and Bahasa Malaysia</li><li>Afrikaans and isiZulu</li><li>Latin</li><li>Pig Latin</li></ul><div class="blogger-post-footer"><img width='1' height='1' src='https://blogger.googleusercontent.com/tracker/21224994-7481575886778068116?l=googleresearch.blogspot.com' alt='' /></div>]]></content:encoded>
			<wfw:commentRss>https://googledata.org/google-research/ig-pay-atin-lay-oice-vay-earch-say/feed/</wfw:commentRss>
		<slash:comments>0</slash:comments>
<enclosure url="" length="" type="" />
		</item>
		<item>
		<title>Word of Mouth: Introducing Voice Search for Indonesian, Malaysian and Latin American Spanish</title>
		<link>https://googledata.org/google-research/word-of-mouth-introducing-voice-search-for-indonesian-malaysian-and-latin-american-spanish/?utm_source=rss&#038;utm_medium=rss&#038;utm_campaign=word-of-mouth-introducing-voice-search-for-indonesian-malaysian-and-latin-american-spanish</link>
		<comments>https://googledata.org/google-research/word-of-mouth-introducing-voice-search-for-indonesian-malaysian-and-latin-american-spanish/#comments</comments>
		<pubDate>Wed, 30 Mar 2011 17:00:00 +0000</pubDate>
		<dc:creator><![CDATA[Research Admin]]></dc:creator>
				<category><![CDATA[Google Research]]></category>

		<guid isPermaLink="false">https://googledata.org/?guid=5c884f007b3cace8d38f50aa10a1a9ef</guid>
		<description><![CDATA[Posted by Linne Ha, International Program ManagerRead more about the launch of Voice Search in Latin American Spanish on the Google América Latina blog.Today we are excited to announce the launch of Voice Search in Indonesian, Malaysian, and Latin Ame...]]></description>
				<content:encoded><![CDATA[Posted by Linne Ha, International Program Manager<br /><br />Read more about the launch of Voice Search in Latin American Spanish on the Google América Latina blog.<br /><br />Today we are excited to announce the launch of Voice Search in Indonesian, Malaysian, and Latin American Spanish, making Voice Search available in over two dozen languages and accents since our first launch in November 2008. This accomplishment could not have been possible without the help of local users in the region - really, we couldn’t have done it without them. Let me explain:<br /><br />In 2010 we launched Voice Search in Dutch, the first language where we used the “word of mouth” project, a crowd-sourcing effort to collect the most accurate voice data possible.The traditional method of acquiring voice samples is to license the data from companies who specialize in the distribution of speech and text databases. However, from day one we knew that to build the most accurate Voice Search acoustic models possible, the best data would come from the people who would use Voice Search once it launched - our users.<br /><br />Since then, in each country, we found small groups of people who were avid fans of Google products and were part of a large social network, either in local communities or on online. We gave them phones and asked them to get voice samples from their friends and family. Everyone was required to sign a consent form and all voice samples were anonymized. When possible, they also helped to test early versions of Voice Search as the product got closer to launch.<br /><br />Building a speech recognizer is not just limited to localizing the user interface. We require thousands of hours of raw data to capture regional accents and idiomatic speech in all sorts of recording environments to mimic daily life use cases. For instance, when developing Voice Search for Latin American Spanish, we paid particular attention to Mexican and Argentinean Spanish. These two accents are more different from one another than any other pair of widely-used accents in all of South and Central America. Samples collected in these countries were very important bookends for building a version of Voice Search that would work across the whole of Latin America. We also chose key countries such as Peru, Chile, Costa Rica, Panama and Colombia to bridge the divergent accent varieties.<br /><br />As an International Program Manager at Google, I have been fortunate enough to travel around the world and meet many of our local Google users. They often have great suggestions for the products that they love, and word of mouth was created with the vision that our users could participate in developing the product. These Voice Search launches would not have been possible without the help of our users, and we’re excited to be able to work together on the product development with the people who will ultimately use our products.<div class="blogger-post-footer"><img width='1' height='1' src='https://blogger.googleusercontent.com/tracker/21224994-8095909017213332862?l=googleresearch.blogspot.com' alt='' /></div>]]></content:encoded>
			<wfw:commentRss>https://googledata.org/google-research/word-of-mouth-introducing-voice-search-for-indonesian-malaysian-and-latin-american-spanish/feed/</wfw:commentRss>
		<slash:comments>0</slash:comments>
		</item>
		<item>
		<title>Reading tea leaves in the tourism industry: A Case Study in the Gulf Oil Spill</title>
		<link>https://googledata.org/google-research/reading-tea-leaves-in-the-tourism-industry-a-case-study-in-the-gulf-oil-spill/?utm_source=rss&#038;utm_medium=rss&#038;utm_campaign=reading-tea-leaves-in-the-tourism-industry-a-case-study-in-the-gulf-oil-spill</link>
		<comments>https://googledata.org/google-research/reading-tea-leaves-in-the-tourism-industry-a-case-study-in-the-gulf-oil-spill/#comments</comments>
		<pubDate>Thu, 24 Mar 2011 18:00:00 +0000</pubDate>
		<dc:creator><![CDATA[Research Admin]]></dc:creator>
				<category><![CDATA[Google Research]]></category>

		<guid isPermaLink="false">https://googledata.org/?guid=d662a557beebde3f3b96d3548377af74</guid>
		<description><![CDATA[Posted by Hyunyoung Choi and Paul Liu, Senior EconomistsA few years ago, our in-house economists, Hal Varian and Hyunyoung Choi, demonstrated how to “predict the present” with monthly visitor arrivals to Hong Kong.  We took this idea further to see...]]></description>
				<content:encoded><![CDATA[<span class="byline-author">Posted by Hyunyoung Choi and Paul Liu, Senior Economists</span><br /><br />A few years ago, our in-house economists, Hal Varian and Hyunyoung Choi, demonstrated how to “<a href="http://googleresearch.blogspot.com/2009/04/predicting-present-with-google-trends.html">predict the present</a>” with monthly visitor arrivals to Hong Kong.  We took this idea further to see if search queries could predict the future. If users start to research their travel plans some weeks or months in advance, then intuitively shouldn’t we be able to extend "predicting the present" into "predicting the future?"  We decided to test it out by focusing on a region whose tourism was recently severely impacted: Florida’s gulf coast.<br /><br />With the travel industry still in the midst of recovering from a deep recession, the Gulf Oil spill had the potential to do significant economic damage. Our case study on the Gulf Oil spill helped find useful insight into people’s future travel plans to Florida; in fact, we found that travel search queries actually were good predictors for trips to Florida, and destinations within Florida, about 4 weeks later.<br /><br />The results we saw surprised us. <a href="http://www.google.com/insights/search/">Google Insights for Search</a> suggested that at least with respect to hotel bookings (using data from Smith Travel Research, Inc.), the aggregate effect of the oil spill was modest on Florida travel, since travelers tended to shift their destinations from the affected regions on the west coast to the east coast or central regions of Florida.  In particular, hotel bookings for affected areas along the Gulf coast were 4.25% less than predicted, and unaffected areas along the Atlantic coast were 4.89% greater than predicted.<br /><br />You can read the full case study <a href="http://www.google.com/url?q=http://www.google.com/googleblogs/pdfs/google_gulf_tourism_march2011.pdf">here</a> or <a href="http://www.google.com/insights/search/">try your own hand</a> at predicting the future!<div class="blogger-post-footer"><img width='1' height='1' src='https://blogger.googleusercontent.com/tracker/21224994-4098509991295395138?l=googleresearch.blogspot.com' alt='' /></div>]]></content:encoded>
			<wfw:commentRss>https://googledata.org/google-research/reading-tea-leaves-in-the-tourism-industry-a-case-study-in-the-gulf-oil-spill/feed/</wfw:commentRss>
		<slash:comments>0</slash:comments>
		</item>
		<item>
		<title>Games, auctions and beyond</title>
		<link>https://googledata.org/google-research/games-auctions-and-beyond/?utm_source=rss&#038;utm_medium=rss&#038;utm_campaign=games-auctions-and-beyond</link>
		<comments>https://googledata.org/google-research/games-auctions-and-beyond/#comments</comments>
		<pubDate>Wed, 16 Mar 2011 21:45:00 +0000</pubDate>
		<dc:creator><![CDATA[Research Admin]]></dc:creator>
				<category><![CDATA[Google Research]]></category>

		<guid isPermaLink="false">https://googledata.org/?guid=47352effd9161c6ecb9391a9ceb6d119</guid>
		<description><![CDATA[Posted by Yossi Matias, Senior Director, Head of Israel R&#38;D CenterIn an effort to advance the understanding of market algorithms and Internet economics, Google has launched an academic research initiative focused on the underlying aspects of online...]]></description>
				<content:encoded><![CDATA[<span class="byline-author">Posted by Yossi Matias, Senior Director, Head of Israel R&amp;D Center</span><br /><br />In an effort to advance the understanding of market algorithms and Internet economics, Google has launched an academic research initiative focused on the underlying aspects of online auctions, pricing, game-theoretic strategies, and information exchange. Twenty professors from three leading Israeli academic institutions - the <a href="http://www.huji.ac.il/huji/eng/">Hebrew University</a>, <a href="http://www.tau.ac.il/index-eng.html">Tel Aviv University</a> and the <a href="http://www1.technion.ac.il/en">Technion</a> - will receive a Google grant to conduct research for three years.<br /><br />In the past two decades, we have seen the Internet grow from a scientific network to an economic force that positively affects the global economy. E-commerce, online advertising, social networks and other new online business models present fascinating research questions and topics of study that can have a profound impact on society.<br /><br />Consider online advertising, which is based on principles from algorithmic game theory and online auctions.  The Internet has enabled advertising that is more segmented and measurable, making it more efficient than traditional advertising channels, such as newspaper classifieds, radio spots, and television commercials. These measurements have led to better pricing models, which are based on online real-time auctions. The original Internet auctions were designed by the industry, based on basic economic principles which have been known and appreciated for forty years.<br /><br />As the Internet grows, online advertising is becoming more sophisticated, with developments such as ad-exchanges, advertising agencies which specialize in online markets, and new analytic tools. Optimizing value for advertisers and publishers in this new environment may benefit from a better understanding of the strategies and dynamics behind online auctions, the main driving tool of Internet advertising.<br /><br />These grants will foster collaboration and interdisciplinary research by bringing together world renowned computer scientists, engineers, economists and game theorists to analyze complex online auctions and markets. Together, they will help bring this area of study into mainstream academic scientific research, ultimately advancing the field to the benefit of the industry at large.<br /><br />The professors who received research grants include:<br /><ul><li>Hebrew University: Danny Dolev, Jeffrey S. Rosenschein, Noam Nisan (<i>Computer Science and Engineering</i>); Liad Blumrosen, Alex Gershkov, Eyal Winter (<i>Economics</i>); Michal Feldman and Ilan Kremer (<i>Business</i>). The last six are also members of the <i>Center for the Study of Rationality</i>.</li><li>Tel Aviv University: Yossi Azar, Amos Fiat,  Haim Kaplan, and Yishay Mansour (<i>Computer Science</i>);  Zvika Neeman (<i>Economics</i>);  Ehud Lehrer and Eilon Solan (<i>Mathematics</i>); and Gal Oestreicher (<i>Business</i>). </li><li>Technion: Seffi Naor (<i>Computer Science</i>); Ron Lavi (<i>Industrial Engineering</i>); Shie Mannor and Ariel Orda (<i>Electrical Engineering</i>).</li></ul>In addition to providing the funds, Google will offer support by inviting the researchers to seminars, workshops, faculty summits and brainstorming events. The results of this research will be published for the benefit of the Internet industry as a whole, and will contribute to the evolving discipline of market algorithms.<div class="blogger-post-footer"><img width='1' height='1' src='https://blogger.googleusercontent.com/tracker/21224994-301139919925075951?l=googleresearch.blogspot.com' alt='' /></div>]]></content:encoded>
			<wfw:commentRss>https://googledata.org/google-research/games-auctions-and-beyond/feed/</wfw:commentRss>
		<slash:comments>0</slash:comments>
		</item>
		<item>
		<title>Large Scale Image Annotation: Learning to Rank with Joint Word-Image Embeddings</title>
		<link>https://googledata.org/google-research/large-scale-image-annotation-learning-to-rank-with-joint-word-image-embeddings/?utm_source=rss&#038;utm_medium=rss&#038;utm_campaign=large-scale-image-annotation-learning-to-rank-with-joint-word-image-embeddings</link>
		<comments>https://googledata.org/google-research/large-scale-image-annotation-learning-to-rank-with-joint-word-image-embeddings/#comments</comments>
		<pubDate>Thu, 10 Mar 2011 16:00:00 +0000</pubDate>
		<dc:creator><![CDATA[Research Admin]]></dc:creator>
				<category><![CDATA[Google Research]]></category>

		<guid isPermaLink="false">https://googledata.org/?guid=db08f11486f6f7a5efb68f4cc3044867</guid>
		<description><![CDATA[Posted by Jason Weston and Samy Bengio, Research TeamIn our paper, we introduce a generic framework to find a joint representation of images and their labels, which can then be used for various tasks, including image ranking and image annotation.We foc...]]></description>
				<content:encoded><![CDATA[<span class="byline-author">Posted by Jason Weston and Samy Bengio, Research Team</span><br /><br />In our <a href="http://www.google.com/research/pubs/pub35780.html">paper</a>, we introduce a generic framework to find a joint representation of images and their labels, which can then be used for various tasks, including image ranking and image annotation.<br /><br />We focus on the task of automatic assignment of annotations (text labels) to images given only the pixel representation of the image (i.e., with no known metadata). This is achieved by a learning algorithm, that is, where the computer learns to predict annotations for new images given annotated training images. Such training datasets are becoming larger and larger, with tens of millions of images and tens of thousands of possible annotations. In this paper, we propose a strongly performing method that scales to such datasets by simultaneously learning to optimize precision at the top of the ranked list of annotations for a given image and learning a low-dimensional joint embedding vector space for both images and annotations. Our system learns an interpretable model, where annotations with alternate wordings ("president obama" or "barack"), different languages ("tour eiﬀel" or "eiffel tower"), or similar concepts (such as "toad" or "frog") are close in the embedding space. Hence, even when our model does not predict the exact annotation given by a human labeler, it often predicts similar annotations.<br /><br />Our system is trained on ~10 million images with ~100,000 possible annotation types and it annotates a single new image in ~0.17 seconds (not including feature processing) and consumes only 82MB of memory. Our method both outperforms all the methods we tested against and in comparison to them is faster and consumes less memory, making it possible to house such a system on a laptop or mobile device.<div class="blogger-post-footer"><img width='1' height='1' src='https://blogger.googleusercontent.com/tracker/21224994-7878514956378010370?l=googleresearch.blogspot.com' alt='' /></div>]]></content:encoded>
			<wfw:commentRss>https://googledata.org/google-research/large-scale-image-annotation-learning-to-rank-with-joint-word-image-embeddings/feed/</wfw:commentRss>
		<slash:comments>0</slash:comments>
		</item>
		<item>
		<title>Julia meets HTML 5</title>
		<link>https://googledata.org/google-research/julia-meets-html-5/?utm_source=rss&#038;utm_medium=rss&#038;utm_campaign=julia-meets-html-5</link>
		<comments>https://googledata.org/google-research/julia-meets-html-5/#comments</comments>
		<pubDate>Mon, 31 Jan 2011 20:00:00 +0000</pubDate>
		<dc:creator><![CDATA[Research Admin]]></dc:creator>
				<category><![CDATA[Google Research]]></category>

		<guid isPermaLink="false"></guid>
		<description><![CDATA[Posted by Daniel Wolf, Software EngineerToday, we launched Julia Map on Google Labs, a fractal renderer in HTML 5. Julia sets are fractals that were studied by the French mathematician Gaston Julia in the early 1920s. Fifty years later, Benoît Mandelb...]]></description>
				<content:encoded><![CDATA[<span class="byline-author">Posted by Daniel Wolf, Software Engineer</span><br /><br />Today, we launched <a href="http://juliamap.googlelabs.com/">Julia Map</a> on Google Labs, a fractal renderer in HTML 5. <a href="http://en.wikipedia.org/wiki/Julia_set">Julia sets</a> are fractals that were studied by the French mathematician Gaston Julia in the early 1920s. Fifty years later, Benoît Mandelbrot studied the set z2 − c and popularized it by generating the first computer visualisation. Generating these images requires heavy computation resources. Modern browsers have optimized JavaScript execution up to the point where it is now possible to render in a browser fractals like Julia sets almost instantly.<br /><br />Julia Map uses the <a href="http://code.google.com/apis/maps/documentation/javascript/">Google Maps API</a> to zoom and pan into the fractals. The images are computed with <a href="http://www.whatwg.org/specs/web-apps/current-work/multipage/the-canvas-element.html">HTML 5 canvas</a>. Each image generally requires millions of floating point operations. <a href="http://www.whatwg.org/specs/web-workers/current-work/">Web workers</a> spread the heavy calculations on all cores of the machine.<br /><br />We hope you will enjoy exploring the different Julia sets, and share the URLs of the most artistic images you discovered. See what others have posted on Twitter under hashtag <a href="http://twitter.com/search?q=%23juliamap">#juliamap</a>. Click on the images below to dive in to infinity!<br /><br /><a onblur="try {parent.deselectBloggerImageGracefully();} catch(e) {}" href="http://www.google.com/url?q=http://juliamap.googlelabs.com/%23ll=64.168107,-0.703125&z=0&p=ffffff,ffff00,ff,ff0000,ffff00&f=newton"><img style="display:block; margin:0px auto 10px; text-align:center;cursor:pointer; cursor:hand;width: 320px; height: 179px;" src="http://4.bp.blogspot.com/_mol72XYv6Jk/TT4ZcLBTPNI/AAAAAAAAAUQ/3ylyUdSjhvk/s400/Julia4-scaled.png" border="0" alt="" id="BLOGGER_PHOTO_ID_5565914161261788370" /></a><br /><br /><a onblur="try {parent.deselectBloggerImageGracefully();} catch(e) {}" href="http://www.google.com/url?q=http://juliamap.googlelabs.com/%23ll=-34.813803,72.202148&z=4&p=ffffff,ffffff,ffffff,ffffff,ff0000,ffff00,ffff00,ff00,ff&f=julia3"><img style="display:block; margin:0px auto 10px; text-align:center;cursor:pointer; cursor:hand;width: 320px; height: 181px;" src="http://3.bp.blogspot.com/_mol72XYv6Jk/TT4Zb5BLoYI/AAAAAAAAAUI/SZJLyUIbCVY/s400/Julia3-scaled.png" border="0" alt="" id="BLOGGER_PHOTO_ID_5565914156429451650" /></a><br /><br /><a onblur="try {parent.deselectBloggerImageGracefully();} catch(e) {}" href="http://juliamap.googlelabs.com/#ll=-3.640121,97.776783&amp;z=11&amp;p=fffc00,cecc00,6c7035,d619b9,ff21dd&amp;f=mandelbrot"><img style="display:block; margin:0px auto 10px; text-align:center;cursor:pointer; cursor:hand;width: 320px; height: 229px;" src="http://2.bp.blogspot.com/_mol72XYv6Jk/TUcidyA9n8I/AAAAAAAAAUk/CsqLYHLZIQ4/s400/Elephant.png" border="0" alt="" id="BLOGGER_PHOTO_ID_5568457359304269762" /></a><br /><br /><a onblur="try {parent.deselectBloggerImageGracefully();} catch(e) {}" href="http://www.google.com/url?q=http://juliamap.googlelabs.com/%23ll=-62.296666,42.608414&z=11&p=7f7fe5,ffffff,ff00,663300&f=mandel321"><img style="display:block; margin:0px auto 10px; text-align:center;cursor:pointer; cursor:hand;width: 320px; height: 179px;" src="http://3.bp.blogspot.com/_mol72XYv6Jk/TT4ZbYd50qI/AAAAAAAAAT4/GIwuSyrQ7dM/s400/Julia1-scaled.png" border="0" alt="" id="BLOGGER_PHOTO_ID_5565914147691549346" /></a><div class="blogger-post-footer"><img width='1' height='1' src='https://blogger.googleusercontent.com/tracker/21224994-1613009550809054253?l=googleresearch.blogspot.com' alt='' /></div>]]></content:encoded>
			<wfw:commentRss>https://googledata.org/google-research/julia-meets-html-5/feed/</wfw:commentRss>
		<slash:comments>0</slash:comments>
<enclosure url="" length="" type="" />
		</item>
		<item>
		<title>Google at NIPS 2010</title>
		<link>https://googledata.org/google-research/google-at-nips-2010/?utm_source=rss&#038;utm_medium=rss&#038;utm_campaign=google-at-nips-2010</link>
		<comments>https://googledata.org/google-research/google-at-nips-2010/#comments</comments>
		<pubDate>Thu, 27 Jan 2011 15:30:00 +0000</pubDate>
		<dc:creator><![CDATA[Research Admin]]></dc:creator>
				<category><![CDATA[Google Research]]></category>

		<guid isPermaLink="false"></guid>
		<description><![CDATA[Posted by Slav Petrov, Doug Aberdeen, and Lisa McCracken, Google ResearchThe machine learning community met in Vancouver in December for the 24th Neural Information Processing Systems Conference (NIPS). As always, the single-track program of the main c...]]></description>
				<content:encoded><![CDATA[<span class="byline-author">Posted by Slav Petrov, Doug Aberdeen, and Lisa McCracken, Google Research</span><br /><br />The machine learning community met in Vancouver in December for the 24th <a href="http://nips.cc/">Neural Information Processing Systems Conference (NIPS)</a>. As always, the single-track program of the main conference featured a number of outstanding talks, followed by interesting late night poster sessions. A record number of workshops covered a wide variety of topics, while allocating sufficient time for skiing in Whistler - after all, many of the most interesting research conversations happen while riding the lift in-between ski runs. This year’s conference also featured a symposium dedicated to <a href="http://samroweis1972-2010.blogspot.com/">Sam Roweis</a>, providing a retrospective on Sam’s life and work. Sam, a fellow Googler and professor at NYU, was at the heart of the NIPS community and is terribly missed.<br /><br />As always, Google was involved in various ways with NIPS. Here at Google, we take a data-driven approach when solving problems. Therefore, Machine Learning is in one way or another at the core of most of the things that we do. It is therefore unsurprising that many Googlers helped shape the program of the conference or were in the audience. This year, three Googlers served as area chairs and even more were reviewers. Googlers also co-authored the following papers:<br /><ul><li><a href="http://static.googleusercontent.com/external_content/untrusted_dlcp/research.google.com/en/us/pubs/archive/36578.pdf">Label Embedding Trees for Large Multi-Class Tasks</a> by Samy Bengio and Jason Weston</li><li><a href="http://www.cs.nyu.edu/~mohri/pub/importance.pdf">Learning Bounds for Importance Weighting</a> by Corinna Cortes, Yishay Mansour, and Mehryar Mohri</li><li><a href="http://books.nips.cc/papers/files/nips23/NIPS2010_1015.pdf">Online Learning in the Manifold of Low-Rank Matrices</a> by Uri Shalit, Daphna Weinshall, and Gal Chechik</li><li><a href="https://docs.google.com/a/google.com/viewer?url=http://books.nips.cc/papers/files/nips23/NIPS2010_0800.pdf">Deterministic Single–Pass Algorithm for LDA</a> by Issei Sato, Kenichi Kurihara, and Hiroshi Nakagawa</li><li><a href="https://docs.google.com/a/google.com/viewer?url=http://books.nips.cc/papers/files/nips23/NIPS2010_0423_slide.pdf">Distributed Dual Averaging In Networks</a> by John Duchi, Alekh Agarwal, and Martin Wainwright</li></ul><br />Additionally, Googlers co-organized three well attended workshops:<br /><ul><li><a href="http://learning.cis.upenn.edu/coarse2fine/">Coarse–to–Fine Learning and Inference</a> by Ben Taskar, David Weiss, Benjamin Sapp, and Slav Petrov</li><li><a href="http://nips.cc/Conferences/2010/Program/event.php?ID=1982">Low–rank Methods for Large–scale Machine Learning</a> by Arthur Gretton, Michael Mahoney, Mehryar Mohri, and Ameet Talwalkar</li><li><a href="http://lccc.eecs.berkeley.edu/">Learning on Cores, Clusters, and Clouds</a> by John Duchi, Ofer Dekel, John Langford, Lawrence Cayton, and Alekh Agarwal</li></ul><br />Finally, Yoram Singer gave a great talk on <a href="http://nips.cc/Conferences/2010/Program/event.php?ID=1966">Learning Structural Sparsity</a> at the Sam Roweis symposium and Googlers presented the following talks during the workshops:<br /><ul><li><a href="http://books.nips.cc/papers/files/nips23/NIPS2010_1015.pdf">Online Learning in the Manifold of Low–Rank Matrices</a> by Uri Shalit, Daphna Weinshall, and Gal Chechik</li><li><a href="http://lccc.eecs.berkeley.edu/Papers/singh-nipsws10-lccc.pdf">Distributed MAP Inference for Undirected Graphical Models</a> by Sameer Singh, Amar Subramanya, Fernando Pereira, and Andrew McCallum </li><li><a href="http://static.googleusercontent.com/external_content/untrusted_dlcp/research.google.com/en/us/pubs/archive/36948.pdf">MapReduce/Bigtable for Distributed Optimization</a> by Keith Hall, Scott Gilpin and Gideon Mann</li><li><a href="http://wimlworkshop.org/">Self-Pruning Prediction Trees</a> by Sally Goldman</li><li><a href="http://sites.google.com/site/mlngcvc/program">Web Scale Image Annotation: Learning to Rank with Joint Word-Image Embeddings</a> by Jason Weston, Samy Bengio, and Nicolas Usunier </li><li><a href="http://learning.cis.upenn.edu/coarse2fine/Main/Petrov">Coarse–to–fine Decoding for Parsing and Machine Translation</a> by Slav Petrov</li></ul><br />Overall, it was a very successful conference and it was good to be back in Vancouver one last time. This coming year <a href="http://nips.cc/Conferences/2011/">NIPS 2011</a> will be in Granada, Spain.  Hasta luego!<div class="blogger-post-footer"><img width='1' height='1' src='https://blogger.googleusercontent.com/tracker/21224994-8051866178146583706?l=googleresearch.blogspot.com' alt='' /></div>]]></content:encoded>
			<wfw:commentRss>https://googledata.org/google-research/google-at-nips-2010/feed/</wfw:commentRss>
		<slash:comments>0</slash:comments>
		</item>
		<item>
		<title>More Google Contributions to the Broader Scientific Community</title>
		<link>https://googledata.org/google-research/more-google-contributions-to-the-broader-scientific-community/?utm_source=rss&#038;utm_medium=rss&#038;utm_campaign=more-google-contributions-to-the-broader-scientific-community</link>
		<comments>https://googledata.org/google-research/more-google-contributions-to-the-broader-scientific-community/#comments</comments>
		<pubDate>Wed, 26 Jan 2011 03:00:00 +0000</pubDate>
		<dc:creator><![CDATA[Research Admin]]></dc:creator>
				<category><![CDATA[Google Research]]></category>

		<guid isPermaLink="false"></guid>
		<description><![CDATA[Posted by Corinna Cortes and Alfred Spector, Google ResearchGooglers actively engage with the scientific community by publishing technical papers, contributing open-source packages, working on standards, introducing new APIs and tools, giving talks and...]]></description>
				<content:encoded><![CDATA[<span class="byline-author">Posted by Corinna Cortes and Alfred Spector, Google Research</span><br /><br />Googlers actively engage with the scientific community by publishing technical papers, contributing open-source packages, working on standards, introducing new APIs and tools, giving talks and presentations, participating in ongoing technical debates, and much more. Our publications offer technical and algorithmic advances, demonstrate things we learn as we develop novel products and services, and shed light on some of the technical challenges we face at Google.<br /><br />In an effort to highlight some of our work, we periodically select a number of publications to be featured on this site.  We first posted a <a href="http://googleresearch.blogspot.com/2010/07/google-publications.html">set of papers</a> on this blog in mid-2010 and subsequently discussed them in more detail in the following blog postings.  This blog posting highlights a few new noteworthy papers authored or co-authored by Googlers from the later half of 2010. In the coming weeks we will be offering a more in-depth look at these publications, but here are some summaries:<br /><br /><span style="font-weight:bold;">Algorithms and Electronic Commerce</span><div><span></span><b><br /></b><a href="http://www.stanford.edu/~qiqiyan/papers/ec10b.pdf">Robust Mechanisms for Risk-Averse Sellers</a><br /><i>ACM Conference on Electronic Commerce (EC)</i><div><div>Mukund Sundararajan and  Qiqi Yan, Stanford University</div><div><br /></div><div>In his seminal Nobel prize-winning work, Roger Myerson identified the revenue-maximizing auction for a risk-neutral auctioneer. In contrast, this work identifies good mechanisms for risk-averse auctioneers. These mechanisms trade a little revenue for better certainty, in the best possible way. We expect this work will help guide reserve-price selection in auctions where auctioneers/sellers want better control over their revenue.<br /><br /><a href="http://www.google.com/search?lr=&amp;ie=UTF-8&amp;oe=UTF-8&amp;q=Monitoring+Algorithms+for+Negative+Feedback+Systems+Sandler+Muthukrishnan">Monitoring Algorithms for Negative Feedback Systems</a><br /><i>World Wide Web Conference (WWW)</i></div><div><div>Mark Sandler and S. Muthukrishnan</div><div><br />In  negative feedback systems, users report abusive content at a site to its owner for consideration or removal, but the users might not be honest.  For the site owners, this represents a trade-off between vetting such user reports by humans vs. accepting them without vetting. This paper presents a  mathematical framework for design and analysis of such systems and presents algorithms with provably good trade-offs against malicious users.<br /><br /><span style="font-weight:bold;">HCI</span></div><div><span></span><b><br /></b><a href="http://research.google.com/pubs/pub36300.html">Children's Roles Using Keyword Search Interfaces in the Home</a><br /><i>Computer Human Interactions (CHI)</i></div><div><div>Allison Druin, University of Maryland, Elizabeth Foss, University of Maryland, Hilary Hutchinson,  Evan Golub, University of Maryland, and Leshell Hatley, University of Maryland</div><div><br />In this paper, we describe seven search roles children display as information seekers using Internet keyword interfaces, based on a home study of 83 children ages 7, 9, and 11.<br /><br /><span style="font-weight:bold;">Machine Learning</span></div><div><span></span><b><br /></b><a href="http://www.google.com/research/pubs/pub35780.html">Large Scale Image Annotation: Learning to Rank with Joint Word-Image Embeddings</a><div><a href="http://www.google.com/research/pubs/pub35780.html"></a><i>European Conference on Machine Learning (ECML) Best Paper</i></div><div>Jason Weston, Samy Bengio, and Nicolas Usunier, Universite Paris 6 - LIP6</div><div><br />In this paper, we introduce a generic framework to find a joint representation of images and their labels, which can then be used for various tasks, including image ranking and image annotation. We simultaneously propose an efficient training algorithm that scales to tens of millions of images and hundreds of thousands of labels, while focusing training on making good predictions at the top of the ranked list. The models are both fast at prediction time and have low memory usage making it possible to house such systems on a laptop or mobile device.<br /><br /><a href="http://www.google.com/research/pubs/pub36500.html">Overlapping Experiment Infrastructure: More, Better</a></div><div><a href="http://www.google.com/research/pubs/pub36500.html"></a><i>Faster Experimentation, Knowledge Discovery and Datamining (KDD)</i></div><div>Diane Tang, Ashish Agarwal, Deirdre O'Brien, and Mike Meyer</div><div><br /></div><div><i></i>Google's data driven culture requires running a large number of live traffic experiments. This paper describes Google's overlapping experimental infrastructure where a single event (e.g. a web search) can be assigned to multiple simultaneous large experiments. The infrastructure and supporting tools provide a framework that enables running experiments from design to decision making and launch, and can be generalized to many other web applications.<br /><br /><span style="font-weight:bold;">NLP</span></div><div><span></span><b><br /></b><a href="http://www.petrovi.de/data/naacl10.pdf">Products of Random Latent Variable Grammars</a></div><div><i>North American Chapter of the Association for Computational Linguistics (NAACL)</i></div><div>Slav Petrov</div><div><br />It is well known that the Expectation Maximization algorithm can converge to widely varying local maxima. This paper shows that this can be advantageous when learning latent variable grammars for syntactic parsing. By combining multiple state-of-the-art individual grammars into an unweighted product model, parsing accuracy can be improved from 90.2% to 91.8% for English, and from 80.3% to 84.5% for German.<br /><br /><span style="font-weight:bold;">Software Engineering</span></div><div><span></span><b><br /></b><a href="http://www.cs.virginia.edu/~jom5x/index.html%3Fp=641.html">Contention Aware Execution: Online Contention Detection and Response</a></div><div><i>International Symposium on Code Generation and Optimization (CGO)</i></div><div>Jason Mars, University of Virginia, Neil Vachharajani, Robert Hundt, Mary Lou Soffa, University of Virginia</div><div><br />This paper makes a big step forward in addressing an important and pressing problem in the field of Computer Science today. This work presents a lightweight runtime solution that significantly improves the utilization of datacenter servers by up to 58% on average. This work also received the CGO 2010 Best Presentation Award.<br /><br /><span style="font-weight:bold;">Speech</span></div><div><span></span><b><br /></b><a href="http://www.google.com/research/pubs/pub36577.html">Say What? Why users choose to speak their web queries</a></div><div><i>Interspeech</i></div><div>Maryam Kamvar and Doug Beeferman</div><div><br />Say What? Have you been speaking your search queries into your mobile device rather than typing them? Spoken search is available on Android, iPhone and Blackberry devices and we see an increasing numbers of searches coming in by voice on these phones. In our paper “Say What: Why users choose to speak their web queries” we investigate, on an aggregate level, what factors are most predictive of spoken queries.  Understanding context in which a speech-driven search is used (or conversely  not used) can be used to improve recognition engines and spoken interface design. So, save keystrokes and say your query!<br /><br /><a href="http://research.google.com/pubs/pub36732.html">Query Language Modeling for Voice Search</a></div><div><i>IEEE Workshop on Spoken Language Technology</i></div><div>Ciprian Chelba, Johan Schalkwyk, Thorsten Brants, Vida Ha, Boulos Harb, Will Neveitt, Carolina Parada*, Johns Hopkins University, and Peng Xu</div><div><br />The paper describes language modeling for google.com query data, and its application to speech recognition for Google Voice Search.<br />Our empirical findings include:<br /><ul><li>10% relative gains in WER from large scale modeling,</li><li>a less known yet potentially quite detrimental interaction between Kneser-Ney smoothing and entropy pruning (approx. 10% relative increase in WER)</li><li>evidence that hints at non-stationarity of the query stream, and </li><li>surprisingly strong dependence across three English locales---USA, Britain and Australia.</li></ul><br /><span style="font-weight:bold;">Structured Data</span></div><div><span></span><b><br /></b><a href="http://research.google.com/pubs/pub36632.html">Dremel: Interactive Analysis of Web-Scale Datasets</a></div><div><i>Very Large Data Bases (VLDB)</i></div><div>Sergey Melnik, Andrey Gubarev, Jing Jing Long, Geoffrey Romer, Shiva Shivakumar, Matt Tolton, and Theo Vassilakis, Google Inc.</div><div><br />Dremel is a scalable, interactive ad-hoc query system. By combining multi-level execution trees and columnar data layout, it is capable of running aggregation queries over trillion-row tables in seconds. The system is widely used at Google and serves as the foundational technology behind BigQuery, a product launched in limited preview mode.<br /><br /><span style="font-weight:bold;">Systems and Infrastructure</span></div><div><span></span><b><br /></b><a href="http://www.google.com/research/pubs/pub36726.html">Large-scale Incremental Processing Using Distributed Transactions and Notifications</a></div><div><i>USENIX Symposium on Operating Systems Design and Implementation (OSDI)</i></div><div>Daniel Peng and Frank Dabek</div><div><br />In the past, Google accumulated a whole day’s worth of changes to the web and ran a series of enormous MapReduces to apply this batch of changes to our index of the web. This system led to a delay of several days between crawling a document and presenting it to users in search results. To meet our goal of reducing the indexing delay to minutes, we needed to update the index as each individual document was crawled, rather than in daily batches. No existing infrastructure supported this kind of incremental transformation at web scale, so we built Percolator: a framework for transforming a large repository using small ACID transactions.<br /><br /><a href="http://research.google.com/pubs/pub36737.html">Availability in Globally Distributed Storage Systems</a></div><div><i>USENIX Symposium on Operating Systems Design and Implementation (OSDI)</i></div><div>Daniel Ford, Francois Labelle, Florentina Popovici, Murray Stokely, Van-Anh Truong*, Columbia University,  Luiz Barroso, Carrie Grimes, and Sean Quinlan</div><div><br />In our paper, we characterize the availability of cloud storage systems, based on extensive monitoring of Google's main storage infrastructure, and the sources of failure which affect availability. We also present statistical models for reasoning about the impact of design choices such as data placement, recovery speed, and replication strategies, including replication across multiple data centers.<br /><br /><span style="font-weight:bold;">Vision</span></div><div><span></span><b><br /></b><a href="http://www.google.com/research/pubs/pub36928.html">Improved Consistent Sampling, Weighted Minhash and L1 Sketching</a></div><div><i>IEEE International Conference on Data Mining (ICDM)</i></div><div>Sergey Ioffe</div><div><br />With the huge amounts of very high-dimensional data, such as images and videos, we frequently need to "sketch" the data -- that is, represent it in a much more compact form, while still allowing us to accurately determine how different any two images or videos are. In this paper, we describe a sketching method for L1, one of the most common distance measures. It works by first hashing the data with a new algorithm, and then compressing each hash to a small number of bits, which is learned from data. This method is fast and allows the distances to be estimated accurately, while reducing the storage requirements by a factor of 100.<br /><br /><span class="Apple-style-span">*) work carried out while at Google</span></div></div></div></div></div></div><div class="blogger-post-footer"><img width='1' height='1' src='https://blogger.googleusercontent.com/tracker/21224994-8381512424098004200?l=googleresearch.blogspot.com' alt='' /></div>]]></content:encoded>
			<wfw:commentRss>https://googledata.org/google-research/more-google-contributions-to-the-broader-scientific-community/feed/</wfw:commentRss>
		<slash:comments>0</slash:comments>
		</item>
		<item>
		<title>Robot hackathon connects with Android, browsers and the cloud</title>
		<link>https://googledata.org/google-research/robot-hackathon-connects-with-android-browsers-and-the-cloud/?utm_source=rss&#038;utm_medium=rss&#038;utm_campaign=robot-hackathon-connects-with-android-browsers-and-the-cloud</link>
		<comments>https://googledata.org/google-research/robot-hackathon-connects-with-android-browsers-and-the-cloud/#comments</comments>
		<pubDate>Fri, 17 Dec 2010 17:00:00 +0000</pubDate>
		<dc:creator><![CDATA[Research Admin]]></dc:creator>
				<category><![CDATA[Google Research]]></category>

		<guid isPermaLink="false"></guid>
		<description><![CDATA[Posted by Ryan Hickman and Mamie Rheingold, 20% Robotics Task ForceWith a beer fridge stocked and music blasting, engineers from across Google—and the world—spent the month of October soldering and hacking in their 20% time to connect hobbyist and ...]]></description>
				<content:encoded><![CDATA[<span class="byline-author">Posted by Ryan Hickman and Mamie Rheingold, 20% Robotics Task Force</span><br /><br /><a onblur="try {parent.deselectBloggerImageGracefully();} catch(e) {}" href="http://3.bp.blogspot.com/_mol72XYv6Jk/TQrpP1YvK1I/AAAAAAAAASg/ez8T1yDfKGE/s1600/IMG_2058.JPG"><img style="float:right; margin:0 0 10px 10px;cursor:pointer; cursor:hand;width: 320px; height: 214px;" src="http://3.bp.blogspot.com/_mol72XYv6Jk/TQrpP1YvK1I/AAAAAAAAASg/ez8T1yDfKGE/s320/IMG_2058.JPG" border="0" alt="" id="BLOGGER_PHOTO_ID_5551505948925700946" /></a>With a beer fridge stocked and music blasting, engineers from across Google—and the world—spent the month of October soldering and hacking in their 20% time to connect hobbyist and educational robots with Android phones.  Just two months later we’re psyched to announce three ways you can play with your iRobot Create(R), LEGO(R) MINDSTORMS(R) or VEX Pro(R) through the cloud:<br /><ul><li><a href="http://appinventor.googlelabs.com/about/">App Inventor for Android</a></li><li><a href="http://www.cellbots.com/software/java-app/">Cellbots for Android</a></li><li><a href="http://www.cellbots.com/software/python-library/">Python library</a> for the <a href="http://code.google.com/p/android-scripting/">Scripting Layer 4 Android</a></li></ul><br /><a onblur="try {parent.deselectBloggerImageGracefully();} catch(e) {}" href="http://2.bp.blogspot.com/_mol72XYv6Jk/TQrppfRMNFI/AAAAAAAAASo/T3JC8sZNE1g/s1600/IMG_2167.JPG"><img style="float:left; margin:0 10px 10px 0;cursor:pointer; cursor:hand;width: 320px; height: 214px;" src="http://2.bp.blogspot.com/_mol72XYv6Jk/TQrppfRMNFI/AAAAAAAAASo/T3JC8sZNE1g/s320/IMG_2167.JPG" border="0" alt="" id="BLOGGER_PHOTO_ID_5551506389665068114" /></a>For the month of October, we invited any Googler who wanted to contribute to connect robots to Google’s services in the cloud to pool their 20% time and participate in as much of the process as they could, from design to hard-core coding.<br /><br />Thanks to our hardware partners (<a href="http://www.irobot.com/create/">iRobot</a>, <a href="http://mindstorms.lego.com/">LEGO Group</a>, and <a href="http://www.vexrobotics.com/">VEX Robotics</a>), we never suffered a shortage of supplies. Designers flew in from London, and prototypes were passed between engineers in Tel-Aviv, Hyderabad, Zurich, Munich and California. In Mountain View, we gathered around every Thursday night, rigging up a projector against the wall to share our week’s worth of demos while chowing on pizza. And here is what we produced (so far!):<br /><ul><li><b>App Inventor:</b> Low level <a href="http://appinventor.googlelabs.com/learn/reference/components/notready.html#BluetoothClient">Bluetooth support</a> for connecting with many serial-enabled robots, and of course <a href="http://appinventor.googlelabs.com/learn/reference/components/legomindstorms.html">tight integration with LEGO MINDSTORMS</a>.</li><li><b>Cellbots for Android:</b> Brand new <a href="http://www.cellbots.com/software/java-app/">Java app</a> from Cellbots.com, which is <a href="http://code.google.com/p/cellbots/source/browse/#svn/trunk/android/java">open source</a> and available for free in the Android Market.</li><li><b>Python library:</b> Modularized version of <a href="http://www.cellbots.com/software/python-library/">the popular Cellbots project</a>, which is all <a href="http://code.google.com/p/cellbots/source/browse/#svn/trunk/android/python">open source code</a>.</li></ul><br /><a onblur="try {parent.deselectBloggerImageGracefully();} catch(e) {}" href="http://4.bp.blogspot.com/_mol72XYv6Jk/TQrsuiwZPvI/AAAAAAAAATQ/IFTr7Rdio-Q/s1600/Profile+List.png"><img style="cursor:pointer; cursor:hand;width: 112px; height: 200px;" src="http://4.bp.blogspot.com/_mol72XYv6Jk/TQrsuiwZPvI/AAAAAAAAATQ/IFTr7Rdio-Q/s200/Profile+List.png" border="0" alt=""id="BLOGGER_PHOTO_ID_5551509775035481842" /></a>   <a onblur="try {parent.deselectBloggerImageGracefully();} catch(e) {}" href="http://1.bp.blogspot.com/_mol72XYv6Jk/TQrsuOdUqfI/AAAAAAAAATI/E3IInRznN6g/s1600/Python+remote+menu.png"><img style="cursor:pointer; cursor:hand;width: 120px; height: 200px;" src="http://1.bp.blogspot.com/_mol72XYv6Jk/TQrsuOdUqfI/AAAAAAAAATI/E3IInRznN6g/s200/Python+remote+menu.png" border="0" alt=""id="BLOGGER_PHOTO_ID_5551509769586780658" /></a>   <a onblur="try {parent.deselectBloggerImageGracefully();} catch(e) {}" href="http://3.bp.blogspot.com/_mol72XYv6Jk/TQrst_4DCnI/AAAAAAAAATA/D4gJS13OElM/s1600/AppInventorBluetooth.png"><img style="cursor:pointer; cursor:hand;width: 186px; height: 200px;" src="http://3.bp.blogspot.com/_mol72XYv6Jk/TQrst_4DCnI/AAAAAAAAATA/D4gJS13OElM/s200/AppInventorBluetooth.png" border="0" alt=""id="BLOGGER_PHOTO_ID_5551509765672340082" /></a><br /><br />We hope these applications provide some fun and inspire you to build upon this lightweight connectivity between robots, Android, the cloud and your browser.<br /><br /><a onblur="try {parent.deselectBloggerImageGracefully();} catch(e) {}" href="http://2.bp.blogspot.com/_mol72XYv6Jk/TQrtQKbVyrI/AAAAAAAAATY/_9Rmqw1otGE/s1600/LEGO.JPG"><img style="display:block; margin:0px auto 10px; text-align:center;cursor:pointer; cursor:hand;width: 307px; height: 400px;" src="http://2.bp.blogspot.com/_mol72XYv6Jk/TQrtQKbVyrI/AAAAAAAAATY/_9Rmqw1otGE/s400/LEGO.JPG" border="0" alt=""id="BLOGGER_PHOTO_ID_5551510352620276402" /></a><div class="blogger-post-footer"><img width='1' height='1' src='https://blogger.googleusercontent.com/tracker/21224994-5480635977004607229?l=googleresearch.blogspot.com' alt='' /></div>]]></content:encoded>
			<wfw:commentRss>https://googledata.org/google-research/robot-hackathon-connects-with-android-browsers-and-the-cloud/feed/</wfw:commentRss>
		<slash:comments>0</slash:comments>
<enclosure url="" length="" type="" />
		</item>
		<item>
		<title>Letting everyone do great things with App Inventor</title>
		<link>https://googledata.org/google-research/letting-everyone-do-great-things-with-app-inventor/?utm_source=rss&#038;utm_medium=rss&#038;utm_campaign=letting-everyone-do-great-things-with-app-inventor</link>
		<comments>https://googledata.org/google-research/letting-everyone-do-great-things-with-app-inventor/#comments</comments>
		<pubDate>Wed, 15 Dec 2010 18:45:00 +0000</pubDate>
		<dc:creator><![CDATA[Research Admin]]></dc:creator>
				<category><![CDATA[Google Research]]></category>

		<guid isPermaLink="false"></guid>
		<description><![CDATA[Posted by Karen Parker, App Inventor Program ManagerIn July, we announced App Inventor for Android, a Google Labs experiment that makes it easier for people to access the capabilities of their Android phone and create apps for their personal use.  We w...]]></description>
				<content:encoded><![CDATA[<font class="byline-author">Posted by Karen Parker, App Inventor Program Manager</font><br /><br /><a onblur="try {parent.deselectBloggerImageGracefully();} catch(e) {}" href="http://3.bp.blogspot.com/_mol72XYv6Jk/TQjxk4g1ANI/AAAAAAAAASQ/Q2rnUDzq93Q/s1600/android_app_inventor-src+%25281%2529+%25281%2529.jpg"><img style="float:right; margin:0 0 10px 10px;cursor:pointer; cursor:hand;width: 84px; height: 84px;" src="http://3.bp.blogspot.com/_mol72XYv6Jk/TQjxk4g1ANI/AAAAAAAAASQ/Q2rnUDzq93Q/s400/android_app_inventor-src+%25281%2529+%25281%2529.jpg" border="0" alt="" id="BLOGGER_PHOTO_ID_5550952156681797842"></a>In July, we <a href="http://googleblog.blogspot.com/2010/07/app-inventor-for-android.html">announced</a> App Inventor for Android, a <a href="http://www.googlelabs.com/faq">Google Labs</a> experiment that makes it easier for people to access the capabilities of their Android phone and create apps for their personal use.  We were delighted (and honestly a bit overwhelmed!) by the interest that our announcement generated.  We were even more delighted to hear the stories of what you were doing with App Inventor.  All sorts of people (teachers and students, parents and kids, programming hobbyists and programming newbies) were building Android apps that perfectly fit their needs.<br /><br />For example, we’ve heard of people building <a href="http://dadhoc.com/2010/11/simple-sight-words-android-app/">vocabulary apps</a> for their children, SMS broadcasting apps for their community events, apps that <a href="https://sites.google.com/site/edricspage/final-project">track their favorite public transportation routes</a> and—our favorite—a <a href="http://www.jonq.com/jq/proposal/">marriage proposal app</a>.<br /><br />We are so impressed with the great things people have done with App Inventor,  we want to allow more people the opportunity to do great things. So we’re excited to announce that App Inventor (beta) is now available in <a href="http://www.googlelabs.com/faq">Labs</a> to anyone with a <a href="https://www.google.com/accounts/NewAccount">Google account</a>.<br /><br />Visit the <a href="http://appinventor.googlelabs.com/">App Inventor home page</a> to get set up and start building your first app.  And be sure to share your App Inventor story on the A<a href="http://appinventor.googlelabs.com/forum/">pp Inventor user forum</a>.  Maybe this holiday season you can make a new kind of homemade gift—an app perfectly designed for the recipient’s needs!<br /><br /><object width="640" height="385"><param name="movie" value="http://www.youtube.com/v/8ADwPLSFeY8?fs=1&amp;hl=en_US&amp;rel=0"><param name="allowFullScreen" value="true"><param name="allowscriptaccess" value="always"><embed src="http://www.youtube.com/v/8ADwPLSFeY8?fs=1&amp;hl=en_US&amp;rel=0" type="application/x-shockwave-flash" allowscriptaccess="always" allowfullscreen="true" width="640" height="385"></object><div class="blogger-post-footer"><img width='1' height='1' src='https://blogger.googleusercontent.com/tracker/21224994-6319278394435437570?l=googleresearch.blogspot.com' alt='' /></div>]]></content:encoded>
			<wfw:commentRss>https://googledata.org/google-research/letting-everyone-do-great-things-with-app-inventor/feed/</wfw:commentRss>
		<slash:comments>0</slash:comments>
<enclosure url="" length="" type="" />
		</item>
		<item>
		<title>$6 million to faculty in Q4 Research Awards</title>
		<link>https://googledata.org/google-research/6-million-to-faculty-in-q4-research-awards/?utm_source=rss&#038;utm_medium=rss&#038;utm_campaign=6-million-to-faculty-in-q4-research-awards</link>
		<comments>https://googledata.org/google-research/6-million-to-faculty-in-q4-research-awards/#comments</comments>
		<pubDate>Wed, 08 Dec 2010 22:30:00 +0000</pubDate>
		<dc:creator><![CDATA[Research Admin]]></dc:creator>
				<category><![CDATA[Google Research]]></category>

		<guid isPermaLink="false"></guid>
		<description><![CDATA[Posted by Maggie Johnson, Director of Education and University RelationsWe've just completed the latest round of Google Research Awards, our program which identifies and supports faculty pursuing research in areas of mutual interest. We had a record nu...]]></description>
				<content:encoded><![CDATA[<font class="byline-author">Posted by Maggie Johnson, Director of Education and University Relations</font><br /><br />We've just completed the latest round of Google Research Awards, our program which identifies and supports faculty pursuing research in areas of mutual interest. We had a record number of submissions this round, and are funding 112 awards across 20 different areas—for a total of more than $6 million. We’re also providing more than 150 Android devices for research and curriculum development to faculty whose projects rely heavily on Android hardware.<br /><br />The areas that received the highest level of funding, due to the large number of proposals in these areas, were systems and infrastructure, human computer interaction, security and multimedia. We also continue to support international research; in this round, 29 percent of the funding was awarded to universities outside the U.S.<br /><br />Some examples from this round of awards:<br /><ul><li><b>Injong Rhee, North Carolina State University</b>. Experimental Evaluation of Increasing TCP Initial Congestion Window (Systems)</li><li><b>James Jones, University of California, Irvine</b>. Bug Comprehension Techniques to Assist Software Debugging (Software Engineering)</li><li><b>Yonina Eldar, Technion, Israel</b>.  Semi-Supervised Regression with Auxiliary Knowledge (Machine Learning)</li><li><b>Victor Lavrenko, University of Edinburgh, United Kingdom</b>. Interactive Relevance Feedback for Mobile Search (Information Retrieval)</li><li><b>James Glass, MIT</b>. Crowdsourcing to Acquire Semantically Labelled Text and Speech Data for Speech Understanding (Speech)</li><li><b>Chi Keung Tang, The Hong Kong University of Science and Technology</b>. Quasi-Dense 3D Reconstruction from 2D Uncalibrated Photos (Geo/Maps)</li><li><b>Phil Blunsom, Oxford, United Kingdom</b>. Unsupervised Induction of Multi-Nonterminal Grammars for Statistical Machine Translation (Machine Translation)</li><li><b>Oren Etzioni, University of Washington</b>. Accessing the Web utilizing Android Phones, Dialogue, and Open Information Extraction (Mobile)</li><li><b>Matthew Salganik, Princeton</b>. Developments in Bottom-Up Social Data Collection (Social)</li></ul><br />The full list of this round’s award recipients can be found in this <a href="http://www.google.com/url?q=http://www.google.com/googleblogs/pdfs/google_research_awards_dec2010.pdf">PDF</a>. For more information on our research award program, visit our <a href="http://www.google.com/url?q=http://research.google.com/university/relations/research_awards.html">website</a>. And if you’re a faculty member, we welcome you to apply for one of next year’s two rounds. The deadline for the first round is February 1.<div class="blogger-post-footer"><img width='1' height='1' src='https://blogger.googleusercontent.com/tracker/21224994-5603455061012429213?l=googleresearch.blogspot.com' alt='' /></div>]]></content:encoded>
			<wfw:commentRss>https://googledata.org/google-research/6-million-to-faculty-in-q4-research-awards/feed/</wfw:commentRss>
		<slash:comments>0</slash:comments>
		</item>
		<item>
		<title>Four Googlers elected ACM Fellows this year</title>
		<link>https://googledata.org/google-research/four-googlers-elected-acm-fellows-this-year/?utm_source=rss&#038;utm_medium=rss&#038;utm_campaign=four-googlers-elected-acm-fellows-this-year</link>
		<comments>https://googledata.org/google-research/four-googlers-elected-acm-fellows-this-year/#comments</comments>
		<pubDate>Wed, 08 Dec 2010 00:00:00 +0000</pubDate>
		<dc:creator><![CDATA[Research Admin]]></dc:creator>
				<category><![CDATA[Google Research]]></category>

		<guid isPermaLink="false"></guid>
		<description><![CDATA[Posted by Alfred Spector, VP of ResearchI am delighted to share with you that, like last year, the Association for Computing Machinery (ACM) has announced that four Googlers have been elected ACM Fellows in 2010, the most this year from any single corp...]]></description>
				<content:encoded><![CDATA[<span class="byline-author">Posted by Alfred Spector, VP of Research</span><br /><br />I am delighted to share with you that, <a href="http://googleresearch.blogspot.com/2009/12/four-googlers-elected-acm-fellows.html">like last year</a>, the Association for Computing Machinery (ACM) has <a href="http://www.acm.org/press-room/news-releases/2010/fellows-2010">announced</a> that four Googlers have been elected ACM Fellows in 2010, the most this year from any single corporation or institution.<br /><br /><a href="http://research.google.com/pubs/LuizBarroso.html">Luiz Barroso</a>, <a href="http://research.google.com/pubs/author35932.html">Dick Lyon</a>, <a href="http://research.google.com/pubs/author2318.html">Muthu Muthukrishnan</a> and <a href="http://research.google.com/pubs/author1092.html">Fernando Pereira</a> were chosen for their contributions to computing and computer science that have provided fundamental knowledge to the field and have generated multiple innovations.<br /><br />On behalf of Google, I congratulate our colleagues, who join the 10 other ACM Fellows and other professional society awardees at Google in exemplifying our extraordinarily talented people. I’ve been struck by the breadth and depth of their contributions, and I hope that they will serve as inspiration for students and computer scientists around the world.<br /><br />You can read more detailed summaries of their achievements below, including the official citations from ACM—although it’s really hard to capture everything they’ve accomplished in one paragraph!<br /><br /><b>Dr. Luiz Barroso: Distinguished Engineer</b><br /><i>For contributions to multi-core computing, warehouse scale data-center architectures, and energy proportional computing</i><br /><blockquote>Over the past decade, Luiz has played a leading role in the definition and implementation of Google’s cluster architecture which has become a blueprint for the computing systems behind the world’s leading Internet services. As the first manager of Google’s Platforms Engineering team, he helped deliver multiple generations of cluster systems, including the world’s first container-based data center.  His theoretical and engineering insights into the requirements of this class of machinery have influenced the processor industry roadmap towards more effective products for server-class computing. His book "The Datacenter as a Computer" (co-authored with Urs Hoelzle) was the first authoritative publication describing these so-called warehouse-scale computers for computer systems professionals and researchers.  Luiz was among the first computer scientists to recognize and articulate the importance of energy-related costs for large data centers, and identify energy proportionality as a key property of energy efficient data centers.  Prior to Google, at Digital Equipment's Western Research Lab, he worked on Piranha, a pioneering chip-multiprocessing architecture that inspired today’s popular multi-core products.  As one of the lead architects and designers of Piranha, his papers, ideas and numerous presentations stimulated much of the research that led to products decades later.</blockquote><b>Richard Lyon: Research Scientist</b><br /><i>For contributions to machine perception and for the invention of the optical mouse</i><br /><blockquote>In the last four years at Google, Dick led the team developing new camera systems and improved photographic image processing for Street View, while leading another team developing technologies for machine hearing and their application to sound retrieval and ranking.  He is now writing a book with Cambridge University Press, and will teach a Stanford course this fall on "Human and Machine Hearing," returning to a line of work that he carried out at Xerox, Schlumberger, and Apple while also doing the optical mouse, bit-serial VLSI computing machines, and handwriting recognition.  The optical mouse (1980) is especially called out in the citation, because it exemplifies the field of "semi-digital" techniques that he developed, which also led to his work on the first single-chip Ethernet device.  And more recently, as chief scientist at Foveon, Dick invented and developed several new techniques for color image sensing and processing, and delivered acclaimed cameras and end-user software.  A hallmark of Dick’s work during his distinguished career has been a practical interplay between theory, including biological theory, and practical computing.</blockquote><b>Dr. S. Muthukrishnan: Research Scientist</b><br /><i>For contributions to efficient algorithms for string matching, data streams, and Internet ad auctions</i><br /><blockquote>Muthu has made significant contributions to the theory and practice of Internet ad systems during his more than four years at Google.  Muthu's breakthrough WWW’09 paper presented a general stable matching framework that produces a (desirable) truthful mechanism capturing all of the common variations and more, in contradiction to prevailing wisdom. In display ads, where image, video and other types of ads are shown as users browse, Muthu led Ad Exchange at Google, to automate placement of display ads that were previously negotiated offline by sales teams.  Prior to Google, Muthu was well known for his pioneering work in the area of data stream algorithmics (including a definitive book on the subject), which led to theoretical and practical advances still in use today to monitor the health and smooth operation of the Internet.  Muthu has a talent for bringing new perspectives to longstanding open problems as exemplified in the work he did on string processing.  Muthu has made influential contributions to many other areas and problems including IP networks, data compression, scheduling, computational biology, distributed algorithms and database technology.   As an educator, Muthu’s avant garde teaching style won him the Award for Excellence in Graduate Teaching at Rutgers CS, where is on the faculty.  As a student remarked in his blog: "there is a magic in his class which kinda spellbinds you and it doesn't feel like a class. It’s more like a family sitting down for dinner to discuss some real world problems. It was always like that even when we were 40 people jammed in for cs-513." </blockquote><b>Dr. Fernando Pereira: Research Director</b><br /><i>For contributions to machine-learning models of natural language and biological sequences</i><br /><blockquote>For the past three years, Fernando has been leading some of Google’s most advanced natural language understanding efforts and some of the most important applications of machine learning technology. He has just the right mix of forward thinking ideas and the ability to put ideas into practice.  With this balance, Fernando has has helped his team of research scientists apply their ideas at the scale needed for Google.  From when he wrote the first Prolog compiler (for the PDP-10 with David Warren) to his days as Chair at University of Pennsylvania,  Fernando has demonstrated a unique understanding of the challenges and opportunities that faced companies like Google with their unprecedented access to massive data sets and its application to the world of speech recognition, natural language processing and machine translation.  At SRI, he pioneered probabilistic language models at a time when logic-based models were more popular.  At AT&amp;T, his work on a toolkit for finite-state models became an industry standard, both as a useful piece of software and in setting the direction for building ever larger language models.  And his year at WhizBang had an influence on other leaders of the field, such as Andrew McCallum at University of Massachusetts and John Lafferty and Tom Mitchell at Carnegie Mellon University, with whom Fernando developed the Conditional Random Field model for sequence processing that has become one of the leading tools of the trade.</blockquote><br />Finally, we also congratulate <a href="http://www.cs.cmu.edu/~christos/">Professor Christos Faloutsos</a> of Carnegie Mellon, who is on sabbatical and a <a href="http://www.google.com/research/university/relations/visiting_faculty.html">Visiting Faculty Member</a> at Google this academic year.  Professor Faloutsos is cited for contributions to data mining, indexing, fractals and power laws.<br /><br /><i><b>Update</b> 12/8</i>: Updated Dick Lyon's title and added information about Professor Faloutsos.<div class="blogger-post-footer"><img width='1' height='1' src='https://blogger.googleusercontent.com/tracker/21224994-4780925651881355452?l=googleresearch.blogspot.com' alt='' /></div>]]></content:encoded>
			<wfw:commentRss>https://googledata.org/google-research/four-googlers-elected-acm-fellows-this-year/feed/</wfw:commentRss>
		<slash:comments>0</slash:comments>
		</item>
		<item>
		<title>Google Launches Cantonese Voice Search in Hong Kong</title>
		<link>https://googledata.org/google-research/google-launches-cantonese-voice-search-in-hong-kong/?utm_source=rss&#038;utm_medium=rss&#038;utm_campaign=google-launches-cantonese-voice-search-in-hong-kong</link>
		<comments>https://googledata.org/google-research/google-launches-cantonese-voice-search-in-hong-kong/#comments</comments>
		<pubDate>Thu, 02 Dec 2010 18:46:00 +0000</pubDate>
		<dc:creator><![CDATA[Research Admin]]></dc:creator>
				<category><![CDATA[Google Research]]></category>

		<guid isPermaLink="false"></guid>
		<description><![CDATA[Posted by Posted by Yun-hsuan Sung (宋雲軒) and Martin Jansche, Google ResearchOn November 30th 2010, Google launched Cantonese Voice Search in Hong Kong. Google Search by Voice has been available in a growing number of languages since we launched o...]]></description>
				<content:encoded><![CDATA[<span class="byline-author">Posted by Posted by Yun-hsuan Sung (宋雲軒) and Martin Jansche, Google Research</span><br /><br />On November 30th 2010, Google launched Cantonese Voice Search in Hong Kong. Google Search by Voice has been available in a growing number of languages since we launched our first US English system in 2008. In addition to US English, we already support Mandarin for Mainland China, Mandarin for Taiwan, Japanese, Korean, French, Italian, German, Spanish, Turkish, Russian, Czech, Polish, Brazilian Portuguese, Dutch, Afrikaans, and Zulu, along with special recognizers for English spoken with British, Indian, Australian, and South African accents.<br /><br />Cantonese is widely spoken in Hong Kong, where it is written using traditional Chinese characters, similar to those used in Taiwan. Chinese script is much harder to type than the Latin alphabet, especially on mobile devices with small or virtual keyboards. People in Hong Kong typically use either “<a href="http://en.wikipedia.org/wiki/Cangjie_input_method">Cangjie</a>” (<a href="http://zh.wikipedia.org/zh/%E5%80%89%E9%A0%A1%E8%BC%B8%E5%85%A5%E6%B3%95">倉頡</a>) or “Handwriting” (手寫輸入) input methods. Cangjie (倉頡) has a steep learning curve and requires users to break the Chinese characters down into sequences of graphical components. The Handwriting (手寫輸入) method is easier to learn, but slow to use. Neither is an ideal input method for people in Hong Kong trying to use Google Search on their mobile phones.<br /><br />Speaking is generally much faster and more natural than typing.  Moreover, some Chinese characters – like “滘” in “<a href="http://zh.wikipedia.org/zh-hk/%E6%BB%98%E8%A5%BF%E6%B4%B2">滘西州</a>” (<a href="http://en.wikipedia.org/wiki/Kau_Sai_Chau">Kau Sai Chau</a>) and “砵” in “<a href="http://zh.wikipedia.org/zh-hk/%E7%A0%B5%E5%85%B8%E4%B9%8D%E8%A1%97">砵典乍街</a>” (<a href="http://en.wikipedia.org/wiki/Pottinger_Street">Pottinger Street</a>) – are so rarely used that people often know only the pronunciation, and not how to write them. Our Cantonese Voice Search begins to address these situations by allowing Hong Kong users to speak queries instead of entering Chinese characters on mobile devices. We believe our development of Cantonese Voice Search is a step towards solving the text input challenge for devices with small or virtual keyboards for users in Hong Kong.<br /><br />There were several challenges in developing Cantonese Voice Search, some unique to Cantonese, some typical of Asian languages and some universal to all languages. Here are some examples of problems that stood out:<br /><ul><li><b>Data Collection</b>: In contrast to English, there are few existing Cantonese datasets that can be used to train a recognition system.  Building a recognition system requires both audio and text data so it can recognize both the sounds and the words. For audio data, our efficient <a href="http://www.google.com/research/pubs/pub36801.html">DataHound</a> collection technique uses smartphones to record and upload large numbers of audio samples from local Cantonese-speaking volunteers. For text data, we sample from anonymized search query logs from http://www.google.com.hk to obtain the large amounts of data needed to train language models.</li><li><b>Chinese Word Boundaries</b>: Chinese writing doesn’t use spaces to indicate word boundaries. To limit the size of the vocabulary for our speech recognizer and to simplify lexicon development, we use characters, rather than words, as the basic units in our system and allow multiple pronunciations for each character.</li><li><b>Mixing of Chinese Characters and English Words</b>: We found that Hong Kong users mix more English into their queries than users in Mainland China and Taiwan.  To build a lexicon for both Chinese characters and English words, we map English words to a sequence of Cantonese pronunciation units.</li><li><b>Tone Issues</b>: Linguists disagree on the best count of the number of tones in Cantonese – some say 6, some say 7, or 9, or 10.  In any case, it’s a lot. We decided to model tone-plus-vowel combinations as single units.  In order to limit the complexity of the resulting model, some rarely-used tone-vowel combinations are merged into single models.</li><li><b>Transliteration</b>: We found that some users use English words while others use the Cantonese transliteration (e.g.,: “<a href="http://en.wikipedia.org/wiki/Jordan,_Hong_Kong">Jordan</a>” vs. “<a href="http://zh-yue.wikipedia.org/wiki/%E4%BD%90%E6%95%A6">佐敦</a>­”). This makes it challenging to develop and evaluate the system, since it’s often impossible for the recognizer to distinguish between an English word and its Cantonese transliteration. During development we use a metric that simply checks whether the correct search results are returned.</li><li><b>Different Accents and Noisy Environment</b>: People speak in different styles with different accents. They use our systems in a variety of environments, including offices, subways, and shopping malls. To make our system work in all these different conditions, we train it using data collected from many different volunteers in many different environments.</li></ul>Cantonese is Google’s third spoken language for Voice Search in the Chinese linguistic family, after Mandarin for Mainland China and Mandarin for Taiwan.  We plan to continue to use our data collection and language modeling technologies to help speakers of Chinese languages easily input text and look up information.<div class="blogger-post-footer"><img width='1' height='1' src='https://blogger.googleusercontent.com/tracker/21224994-980416476665192090?l=googleresearch.blogspot.com' alt='' /></div>]]></content:encoded>
			<wfw:commentRss>https://googledata.org/google-research/google-launches-cantonese-voice-search-in-hong-kong/feed/</wfw:commentRss>
		<slash:comments>0</slash:comments>
		</item>
		<item>
		<title>Voice Search in Underrepresented Languages</title>
		<link>https://googledata.org/google-research/voice-search-in-underrepresented-languages/?utm_source=rss&#038;utm_medium=rss&#038;utm_campaign=voice-search-in-underrepresented-languages</link>
		<comments>https://googledata.org/google-research/voice-search-in-underrepresented-languages/#comments</comments>
		<pubDate>Tue, 09 Nov 2010 22:21:00 +0000</pubDate>
		<dc:creator><![CDATA[Research Admin]]></dc:creator>
				<category><![CDATA[Google Research]]></category>

		<guid isPermaLink="false"></guid>
		<description><![CDATA[Posted by Pedro J. Moreno, Staff Research Scientist and Johan Schalkwyk, Senior Staff EngineerWelkom*!Today we’re introducing Voice Search support for Zulu and Afrikaans, as well as South African-accented English. The addition of Zulu in particular r...]]></description>
				<content:encoded><![CDATA[<span class="byline-author">Posted by Pedro J. Moreno, Staff Research Scientist and Johan Schalkwyk, Senior Staff Engineer</span><br /><br />Welkom*!<br /><br />Today we’re introducing Voice Search support for Zulu and Afrikaans, as well as South African-accented English. The addition of Zulu in particular represents our first effort in building Voice Search for <i>underrepresented languages</i>.<br /><br />We define underrepresented languages as those which, while spoken by millions, have little presence in electronic and physical media, e.g., webpages, newspapers and magazines.  Underrepresented languages have also often received little attention from the speech research community.  Their phonetics, grammar, acoustics, etc., haven’t been extensively studied, making the development of ASR (automatic speech recognition) voice search systems challenging.<br /><br />We believe that the speech research community needs to start working on many of these underrepresented languages to advance progress and build speech recognition, translation and other Natural Language Processing (NLP) technologies. The development of NLP technologies in these languages is critical for enabling <a href="http://www.google.com/corporate/facts.html">information access for everybody</a>. Indeed, these technologies have the potential to break language barriers.<br /><br />We also think it’s important that researchers in these countries take a leading role in advancing the state of the art in their own languages. To this end, we’ve collaborated with the Multilingual Speech Technology group at <a href="http://www.nwu.ac.za/nwu/index.html">South Africa’s North-West University</a> led by Prof. Ettiene Barnard (also of the <a href="http://www.meraka.org.za/">Meraka Research Institute</a>), an authority in speech technology for South African languages.  Our development effort was spearheaded by Charl van Heerden, a South African intern and a student of Prof. Barnard. With the help of Prof. Barnard’s team, we collected acoustic data in the three languages, developed lexicons and grammars, and Charl and others used those to develop the three Voice Search systems. A team of language specialists traveled to several cities collecting audio samples from hundreds of speakers in multiple acoustic conditions such as street noise, background speech, etc. Speakers were asked to read typical search queries into an <a href="http://research.google.com/pubs/pub36801.html">Android app</a> specifically designed for audio data collection.<br /><br />For Zulu, we faced the additional challenge of few text sources on the web. We often analyze the search queries from local versions of Google to build our lexicons and language models. However, for Zulu there weren’t enough queries to build a useful language model. Furthermore, since it has few online data sources, native speakers have learned to use a mix of Zulu and English when searching for information on the web. So for our Zulu Voice Search product, we had to build a truly hybrid recognizer, allowing free mixture of both languages. Our phonetic inventory covers both English and Zulu and our grammars allow natural switching from Zulu to English, emulating speaker behavior.<br /><br />This is our first release of Voice Search in a native African language, and we hope that it won’t be the last. We’ll continue to work on technology for languages that have until now received little attention from the speech recognition community.<br /><br />Salani kahle!**<br /><br />* “Welcome” in Afrikaans<br />** “Stay well” in Zulu<div class="blogger-post-footer"><img width='1' height='1' src='https://blogger.googleusercontent.com/tracker/21224994-3288065901288467706?l=googleresearch.blogspot.com' alt='' /></div>]]></content:encoded>
			<wfw:commentRss>https://googledata.org/google-research/voice-search-in-underrepresented-languages/feed/</wfw:commentRss>
		<slash:comments>0</slash:comments>
		</item>
		<item>
		<title>Suggesting a Better Remote Control</title>
		<link>https://googledata.org/google-research/suggesting-a-better-remote-control/?utm_source=rss&#038;utm_medium=rss&#038;utm_campaign=suggesting-a-better-remote-control</link>
		<comments>https://googledata.org/google-research/suggesting-a-better-remote-control/#comments</comments>
		<pubDate>Thu, 04 Nov 2010 20:12:00 +0000</pubDate>
		<dc:creator><![CDATA[Research Admin]]></dc:creator>
				<category><![CDATA[Google Research]]></category>

		<guid isPermaLink="false"></guid>
		<description><![CDATA[Posted by Ullas Gargi and Rich Gossweiler, Research TeamIt seems clear that the TV is a growing source of online audio-video content that you select by searching. Entering characters of a search string one by one using a traditional remote control and ...]]></description>
				<content:encoded><![CDATA[<span class="byline-author">Posted by Ullas Gargi and Rich Gossweiler, Research Team</span><br /><br />It seems clear that the TV is a growing source of online audio-video content that you select by searching. Entering characters of a search string one by one using a traditional remote control and onscreen keyboard is extremely tiresome. People have been working on building better ways to search on the TV, ranging from small keyboards to voice input to interesting gestures you might make to let the TV know what you want. But currently the traditional left-right-up-down clicker dominates as the family room input device. To enter the letters of a show, you click over and over until you get to the desired letter on the on-screen keyboard and then you hit enter to select it. You repeat this mind-numbingly slow process until you type in your query string or at least enough letters that the system can put up a list of suggested completions. Can we instead use a Google AutoComplete style recommendation model and novel interface to make character entry less painful?<br /><br />We have developed an interaction model that reduces the distance to the predicted next letter without scrambling or moving letters on the underlying keyboard (which is annoying and increases the time it takes to find the next letter). We reuse the highlight ring around the currently selected letter and fill it with 4 possible characters that might be next, but we do not change the underlying keyboard layout. With 4 slots to suggest the next letter and a good prediction model trained on the target corpus, the next letter is often right where you are looking and just a click away.<br /><br />To learn more about this combination of User Experience and Machine Learning to address a growing problem with searching on TVs, check out our WWW 2010 publication,<a href="http://research.google.com/pubs/pub36239.html">QuickSuggest.</a><div class="blogger-post-footer"><img width='1' height='1' src='https://blogger.googleusercontent.com/tracker/21224994-7297099746362425648?l=googleresearch.blogspot.com' alt='' /></div>]]></content:encoded>
			<wfw:commentRss>https://googledata.org/google-research/suggesting-a-better-remote-control/feed/</wfw:commentRss>
		<slash:comments>0</slash:comments>
		</item>
		<item>
		<title>Exploring Computational Thinking</title>
		<link>https://googledata.org/google-research/exploring-computational-thinking/?utm_source=rss&#038;utm_medium=rss&#038;utm_campaign=exploring-computational-thinking</link>
		<comments>https://googledata.org/google-research/exploring-computational-thinking/#comments</comments>
		<pubDate>Mon, 25 Oct 2010 17:00:00 +0000</pubDate>
		<dc:creator><![CDATA[Research Admin]]></dc:creator>
				<category><![CDATA[Google Research]]></category>

		<guid isPermaLink="false"></guid>
		<description><![CDATA[Posted by Elaine Kao, Education Program ManagerOver the past year, a group of California-credentialed teachers along with our own Google engineers came together to discuss and explore ideas about how to incorporate computational thinking into the K-12 ...]]></description>
				<content:encoded><![CDATA[<span class="byline-author">Posted by Elaine Kao, Education Program Manager</span><br /><br />Over the past year, a group of California-credentialed teachers along with our own Google engineers came together to discuss and explore ideas about how to incorporate computational thinking into the K-12 curriculum to enhance student learning and build this critical 21st century skill in everyone.<br /><br />What exactly is computational thinking? Well, that would depend on who you ask as there are several existing <a href="http://www.google.com/edu/computational-thinking/resources.html#ct">resources</a> on the web that may define this term slightly differently. We define <a href="http://www.google.com/edu/computational-thinking/what-is-ct.html">computational thinking</a> (CT) as a set of skills that software engineers use to write the programs that underlay all of the computer applications you use every day. Specific CT techniques include:<br /><ul><li>Problem decomposition: the ability to break down a problem into sub-problems</li><li>Pattern recognition: the ability to notice similarities, differences, properties, or trends in data</li><li>Pattern generalization: the ability to extract out unnecessary details and generalize those that are necessary in order to define a concept or idea in general terms</li><li>Algorithm design: the ability to build a repeatable, step-by-step process to solve a particular problem</li></ul>Given the increasing prevalence of technology in our day-to-day lives and in most careers outside of computer science, we believe that it is important to raise this base level of understanding in everyone.<br /><br />To this end, we’d like to introduce you to a new resource: <a href="http://www.google.com/edu/computational-thinking/index.html">Exploring Computational Thinking</a>. Similar to some of our other initiatives in education, including <a href="http://cs4hs.com/">CS4HS</a> and <a href="http://code.google.com/edu/">Google Code University</a>, this program is committed to providing educators with access to our curriculum models, resources, and communities to help them learn more about CT, discuss it as a strategy for teaching and understanding core curriculum, as well as easily incorporate CT into their own curriculum, whether it be in math, science, language, history or beyond. The materials developed by the team reflect both the teachers’ expertise in pedagogy and K-12 curriculum as well as our engineers’ problem-solving techniques that are critical to our industry.<br /><br />Prior to launching this program, we reached out to several educators and classrooms and had them try our materials. Here’s some of the feedback we received:<br /><ul><li>CT as a strategy for teaching and student learning works well with many subjects, and can easily be incorporated to support the existing K-12 curriculum</li><li>Our models help to call out the specific CT techniques and provide more structure around the topics taught by educators, many of who were already unknowingly applying CT in their classrooms</li><li>Including programming exercises in the classroom can significantly enrich a lesson by both challenging the advanced students and motivating the students who have fallen behind</li><li>Our examples provide educators with a means of re-teaching topics that students have struggled with in the past, without simply going through the same lesson that frustrated them before</li></ul>To learn more about our program or access CT curriculum materials and other resources, visit us at <a href="http://www.google.com/edu/computational-thinking/index.html">www.google.com/edu/ect</a>.<div class="blogger-post-footer"><img width='1' height='1' src='https://blogger.googleusercontent.com/tracker/21224994-4692792540142565485?l=googleresearch.blogspot.com' alt='' /></div>]]></content:encoded>
			<wfw:commentRss>https://googledata.org/google-research/exploring-computational-thinking/feed/</wfw:commentRss>
		<slash:comments>0</slash:comments>
		</item>
		<item>
		<title>Google at the Conference on Empirical Methods in Natural Language Processing (EMNLP &#8217;10)</title>
		<link>https://googledata.org/google-research/google-at-the-conference-on-empirical-methods-in-natural-language-processing-emnlp-10/?utm_source=rss&#038;utm_medium=rss&#038;utm_campaign=google-at-the-conference-on-empirical-methods-in-natural-language-processing-emnlp-10</link>
		<comments>https://googledata.org/google-research/google-at-the-conference-on-empirical-methods-in-natural-language-processing-emnlp-10/#comments</comments>
		<pubDate>Mon, 18 Oct 2010 21:30:00 +0000</pubDate>
		<dc:creator><![CDATA[Research Admin]]></dc:creator>
				<category><![CDATA[Google Research]]></category>

		<guid isPermaLink="false"></guid>
		<description><![CDATA[Posted by Slav Petrov, Research ScientistThe Conference on Empirical Methods in Natural Language Processing (EMNLP '10) was recently held at the MIT Stata Center in Massachusetts.  Natural Language Processing is at the core of many of the things that w...]]></description>
				<content:encoded><![CDATA[<span class="byline-author">Posted by Slav Petrov, Research Scientist</span><br /><br /><a href="http://www.lsi.upc.edu/events/emnlp2010/">The Conference on Empirical Methods in Natural Language Processing</a> (EMNLP '10) was recently held at the MIT Stata Center in Massachusetts.  Natural Language Processing is at the core of many of the things that we do here at Google. Googlers have therefore been traditionally part of this research community, participating as program committee members, paper authors and attendees.<br /><br />At this year's EMNLP conference Google Fellow, <a href="http://singhal.info/">Amit Singhal</a> gave an invited keynote talk on "<a href="http://www.lsi.upc.edu/events/emnlp2010/speakers.html">Challenges in running a commercial search engine</a>"  where he highlighted some of the exciting opportunities, as well as challenges, that Google is currently facing. Furthermore, Terry Koo (who recently joined Google), David Sontag (former Google PhD Fellowship recipient) and their collaborators from MIT received the Fred Jelinek Best Paper Award for their innovative work on syntactic parsing with the title "Dual Decomposition for Parsing with Non-Projective Head Automata".<br /><br />Here is a complete list of the papers presented by Googlers at the conference:<br /><ul><li><a href="http://www.aclweb.org/anthology/D/D10/D10-1125.pdf">Dual Decomposition for Parsing with Non-Projective Head Automata</a> (Fred Jelinek Best Paper Award) by Terry Koo, Alexander M. Rush, Michael Collins, Tommi Jaakkola, and David Sontag</li><li><a href="http://www.aclweb.org/anthology/D/D10/D10-1016.pdf">"Poetic" Statistical Machine Translation: Rhyme and Meter</a> (see also <a href="http://googleresearch.blogspot.com/2010/10/poetic-machine-translation.html">here</a>) by <a href="http://research.google.com/pubs/author14721.html">Dmitriy Genzel</a>, <a href="http://research.google.com/pubs/author37567.html">Jakob Uszkoreit</a>, and <a href="http://research.google.com/pubs/och.html">Franz Och</a> </li><li><a href="http://www.aclweb.org/anthology/D/D10/D10-1017.pdf">Efficient Graph-Based Semi-Supervised Learning of Structured Tagging Models</a> by <a href="http://research.google.com/pubs/author38552.html">Amarnag Subramanya</a>, <a href="http://research.google.com/pubs/author38945.html">Slav Petrov</a>, and <a href="http://research.google.com/pubs/author1092.html">Fernando Pereira</a></li><li><a href="http://www.aclweb.org/anthology/D/D10/D10-1069.pdf">Uptraining for Accurate Deterministic Question Parsing</a> by <a href="http://research.google.com/pubs/author38945.html">Slav Petrov</a>, Pi-Chuan Chang, <a href="http://research.google.com/pubs/author38358.html">Michael Ringgaard</a>, and <a href="http://research.google.com/pubs/author88.html">Hiyan Alshawi</a> </li><li><a href="http://www.aclweb.org/anthology/D/D10/D10-1002.pdf">Self-training with Products of Latent Variable Grammars</a> by Zhongqiang Huang, Mary Harper, and <a href="http://research.google.com/pubs/author38945.html">Slav Petrov</a></li></ul><div class="blogger-post-footer"><img width='1' height='1' src='https://blogger.googleusercontent.com/tracker/21224994-2809016617578644981?l=googleresearch.blogspot.com' alt='' /></div>]]></content:encoded>
			<wfw:commentRss>https://googledata.org/google-research/google-at-the-conference-on-empirical-methods-in-natural-language-processing-emnlp-10/feed/</wfw:commentRss>
		<slash:comments>0</slash:comments>
		</item>
		<item>
		<title>Kuzman Ganchev Receives Presidential Award from the Republic of Bulgaria</title>
		<link>https://googledata.org/google-research/kuzman-ganchev-receives-presidential-award-from-the-republic-of-bulgaria/?utm_source=rss&#038;utm_medium=rss&#038;utm_campaign=kuzman-ganchev-receives-presidential-award-from-the-republic-of-bulgaria</link>
		<comments>https://googledata.org/google-research/kuzman-ganchev-receives-presidential-award-from-the-republic-of-bulgaria/#comments</comments>
		<pubDate>Fri, 15 Oct 2010 13:45:00 +0000</pubDate>
		<dc:creator><![CDATA[Research Admin]]></dc:creator>
				<category><![CDATA[Google Research]]></category>

		<guid isPermaLink="false"></guid>
		<description><![CDATA[Posted by Slav Petrov, Research ScientistWe would like to congratulate Kuzman Ganchev for being the runner-up for the John Atanasoff award from the President of the Republic of Bulgaria. Kuzman recently joined our New York office as a research scientis...]]></description>
				<content:encoded><![CDATA[<span class="byline-author">Posted by Slav Petrov, Research Scientist</span><br /><br /><a onblur="try {parent.deselectBloggerImageGracefully();} catch(e) {}" href="http://3.bp.blogspot.com/_mol72XYv6Jk/TLeBGz5GlvI/AAAAAAAAAP8/zkTyOtfI3iY/s1600/kuzman+pres1.jpg"><img style="float:left; margin:0 10px 10px 0;cursor:pointer; cursor:hand;width: 400px; height: 290px;" src="http://3.bp.blogspot.com/_mol72XYv6Jk/TLeBGz5GlvI/AAAAAAAAAP8/zkTyOtfI3iY/s400/kuzman+pres1.jpg" border="0" alt=""id="BLOGGER_PHOTO_ID_5528029021629486834" /></a>We would like to congratulate <a href="http://www.seas.upenn.edu/~kuzman/">Kuzman Ganchev</a> for being the runner-up for the <a href="http://www.president.bg/news.php?id=4017&amp;st=0">John Atanasoff award</a> from the President of the Republic of Bulgaria. Kuzman recently joined our New York office as a research scientist, after completing his doctoral studies at the University of Pennsylvania.<br /><br />The John Atanasoff award was established in 2003 and is given annually to a Bulgarian scientist under 35 for scientific or practical contributions to the development of computer and information technology worldwide and significant economic or social importance for Bulgaria. Kuzman received the award for his contributions to computational linguistics and machine learning. Kuzman is the co-author of more than 20 publications that have appeared in international conferences and journals.<div class="blogger-post-footer"><img width='1' height='1' src='https://blogger.googleusercontent.com/tracker/21224994-4231255786923511142?l=googleresearch.blogspot.com' alt='' /></div>]]></content:encoded>
			<wfw:commentRss>https://googledata.org/google-research/kuzman-ganchev-receives-presidential-award-from-the-republic-of-bulgaria/feed/</wfw:commentRss>
		<slash:comments>0</slash:comments>
<enclosure url="" length="" type="" />
		</item>
		<item>
		<title>Korean Voice Input &#8212; Have you Dictated your E-Mails in Korean lately?</title>
		<link>https://googledata.org/google-research/korean-voice-input-have-you-dictated-your-e-mails-in-korean-lately/?utm_source=rss&#038;utm_medium=rss&#038;utm_campaign=korean-voice-input-have-you-dictated-your-e-mails-in-korean-lately</link>
		<comments>https://googledata.org/google-research/korean-voice-input-have-you-dictated-your-e-mails-in-korean-lately/#comments</comments>
		<pubDate>Thu, 14 Oct 2010 16:00:00 +0000</pubDate>
		<dc:creator><![CDATA[Research Admin]]></dc:creator>
				<category><![CDATA[Google Research]]></category>

		<guid isPermaLink="false"></guid>
		<description><![CDATA[Posted by Mike Schuster &#38; Kaisuke Nakajima, Google ResearchGoogle Voice Search has been available in various flavors of English since 2008, in Mandarin and Japanese since 2009, in French, Italian, German and Spanish since June 2010 (see also in thi...]]></description>
				<content:encoded><![CDATA[<span class="byline-author">Posted by Mike Schuster &amp; Kaisuke Nakajima, Google Research</span><br /><br />Google Voice Search has been available in various flavors of <a href="http://googlemobile.blogspot.com/2008/11/google-mobile-app-for-iphone-now-with.html">English</a> since 2008, in <a href="http://googleresearch.blogspot.com/2009/11/google-search-by-voice-learns-mandarin.html">Mandarin</a> and <a href="http://googleresearch.blogspot.com/2009/12/teaching-computer-to-understand.html">Japanese</a> since 2009, in <a href="http://googlemobile.blogspot.com/2010/06/salut-willkommen-benvenuto-bienvenido.html">French, Italian, German and Spanish</a> since June 2010 (see also in this <a href="http://googleresearch.blogspot.com/2010/06/google-search-by-voice-now-available-in.html">blog post</a>), and shortly after that in Taiwanese. On June 16th 2010, we took the next step by launching our <a href="http://googlemobile.blogspot.com/2010/06/annyeong-haseyo-to-google-search-by.html">Korean Voice Search system</a>.<br /><br />Korean Voice Search, by focusing on finding the correct web page for a spoken query, has been quite successful since launch. We have improved the acoustic models several times which resulted in significantly higher accuracy and reduced latency, and we are committed to improving it even more over time.<br /><br />While voice search significantly simplifies input for search, especially for longer queries, there are numerous applications on any smartphone that could also benefit from general voice input, such as dictating an email or an SMS. Our experience with US English has taught us that voice input is as important as voice search, as the time savings from speaking rather than typing a message are substantial. Korean is the first non-English language where we are launching general voice input. This launch extends voice input to emails, SMS messages, and more on Korean Android phones. Now every text field on the phone will accept Korean speech input.<br /><br />Creating a general voice input service had different requirements and technical challenges compared to voice search. While voice search was optimized to give the user the correct web page, voice input was optimized to minimize (Hangul) character error rate. Voice inputs are usually longer than searches (short full sentences or parts of sentences), and the system had to be trained differently for this type of data. The current system’s language model was trained on millions of Korean sentences that are similar to those we expect to be spoken. In addition to the queries we used for training voice search, we also used parts of web pages, selected blogs, news articles and more. Because the system expects spoken data similar to what it was trained on, it will generally work well on normal spoken sentences, but may yet have difficulty on random or rare word sequences -- we will work to keep improving on those.<br /><br />Korean voice input is part of Google’s long-term goal to make speech input an acceptable and useful form of input on any mobile device. As with voice search, our cloud computing infrastructure will help us to improve quality quickly, as we work to better support all noise conditions, all Korean dialects, and all Korean users.<div class="blogger-post-footer"><img width='1' height='1' src='https://blogger.googleusercontent.com/tracker/21224994-6584976533166381481?l=googleresearch.blogspot.com' alt='' /></div>]]></content:encoded>
			<wfw:commentRss>https://googledata.org/google-research/korean-voice-input-have-you-dictated-your-e-mails-in-korean-lately/feed/</wfw:commentRss>
		<slash:comments>0</slash:comments>
		</item>
		<item>
		<title>Clustering Related Queries Based on User Intent</title>
		<link>https://googledata.org/google-research/clustering-related-queries-based-on-user-intent/?utm_source=rss&#038;utm_medium=rss&#038;utm_campaign=clustering-related-queries-based-on-user-intent</link>
		<comments>https://googledata.org/google-research/clustering-related-queries-based-on-user-intent/#comments</comments>
		<pubDate>Wed, 13 Oct 2010 21:10:00 +0000</pubDate>
		<dc:creator><![CDATA[Research Admin]]></dc:creator>
				<category><![CDATA[Google Research]]></category>

		<guid isPermaLink="false"></guid>
		<description><![CDATA[Posted by Jayant Madhavan and Alon HalevyPeople today use search engines for all their information needs, but when they pose a particular search query, they typically have a specific underlying intent. However, when looking at any query in isolation, i...]]></description>
				<content:encoded><![CDATA[<span class="byline-author">Posted by Jayant Madhavan and Alon Halevy</span><br /><br />People today use search engines for all their information needs, but when they pose a particular search query, they typically have a specific underlying intent. However, when looking at any query in isolation, it might not entirely be clear what the underlying intent is. For example, when querying for <a href="http://www.google.com/search?q=mars">mars</a>, a user might be looking for more information about <a href="http://www.google.com/search?q=planet+mars">the planet Mars</a>, or the <a href="http://www.google.com/search?q=mercury+venus+earth+mars">planets in the solar system</a> in general, or the <a href="http://www.google.com/search?q=mars+candy">Mars candy bar</a>, or <a href="http://www.google.com/search?q=mars+god+of+war">Mars the Roman god of war</a>. The ambiguity in intent is most pronounced for queries that are inherently ambiguous and for queries about prominent entities about which there are various different types of information on the Internet. Given such ambiguity, modern search engines try to complement their search results with lists of related queries that can be used to further explore a particular intent.<br /><br />In a recent <a href="http://www.stanford.edu/~esadikov/www2010_paper.pdf">paper</a>, we explored the problem of clustering the related queries as a means of understanding the different intents underlying a given user query. We propose an approach that combines an analysis of anonymized document-click logs (what results do users click on) and query-session logs (what sequences of queries do users pose in a search session). We model typical user search behavior as a traversal of a graph whose nodes are related queries and clicked documents. We propose that  the nodes in the graph, when grouped based on the probability of a typical user visiting them within a single search session, yield clusters that correspond to distinct user intents.<br /><br />Our results show that underlying intents (clusters of related queries) almost always correspond to well-understood, high-level concepts. For example, for mars, in addition to re-constructing each of the intents listed earlier, we also find distinct clusters grouping queries about <a href="http://www.google.com/search?q=mars+phoenix">NASA’s missions to the planet</a>, about specific interest in <a href="http://www.google.com/search?q=life+on+mars">life on Mars</a>, as well as a <a href="http://www.google.com/search?q=mars+manga">Japanese comic series</a>, and a <a href="http://www.google.com/search?q=mars+supermarket">grocery chain</a> named Mars. We found that our clustering approach yields better results than earlier approaches that either only used document-click or only query-session information. More details about our proposed approach and an analysis of the resulting clusters can be found in our paper that was presented at the <a href="http://www2010.org/www/">International World Wide Web conference</a> earlier this year.<div class="blogger-post-footer"><img width='1' height='1' src='https://blogger.googleusercontent.com/tracker/21224994-8065684740198901350?l=googleresearch.blogspot.com' alt='' /></div>]]></content:encoded>
			<wfw:commentRss>https://googledata.org/google-research/clustering-related-queries-based-on-user-intent/feed/</wfw:commentRss>
		<slash:comments>0</slash:comments>
		</item>
		<item>
		<title>Google at USENIX Symposium on Operating Systems Design and Implementation (OSDI ‘10)</title>
		<link>https://googledata.org/google-research/google-at-usenix-symposium-on-operating-systems-design-and-implementation-osdi-%e2%80%9810/?utm_source=rss&#038;utm_medium=rss&#038;utm_campaign=google-at-usenix-symposium-on-operating-systems-design-and-implementation-osdi-%25e2%2580%259810</link>
		<comments>https://googledata.org/google-research/google-at-usenix-symposium-on-operating-systems-design-and-implementation-osdi-%e2%80%9810/#comments</comments>
		<pubDate>Tue, 12 Oct 2010 18:01:00 +0000</pubDate>
		<dc:creator><![CDATA[Research Admin]]></dc:creator>
				<category><![CDATA[Google Research]]></category>

		<guid isPermaLink="false"></guid>
		<description><![CDATA[Posted by Murray Stokely, Software EngineerThe 9th USENIX Symposium on Operating Systems Design and Implementation (OSDI ‘10) was recently held in Vancouver, B.C.  This biennial conference is one of the premiere forums for presenting innovative resea...]]></description>
				<content:encoded><![CDATA[<span class="byline-author">Posted by Murray Stokely, Software Engineer</span><br /><p>The <a href="http://usenix.com/events/osdi10/">9th USENIX Symposium on Operating Systems Design and Implementation</a> (OSDI ‘10) was recently held in Vancouver, B.C.  This biennial conference is one of the premiere forums for presenting innovative research in distributed systems from both academia and industry, and we were glad to be a part of it.</p><p>In addition to sponsoring this conference since 2002, Googlers contributed to the exchange of scientific ideas through authoring or co-authoring 3 published papers, organizing workshops, and serving on the program committee.  A short summary of the contributions:</p><ul><li><a href="http://research.google.com/pubs/pub36726.html">Large-scale Incremental Processing Using Distributed Transactions and Notifications</a>.<br />Google replaced its batch-oriented indexing system with an incremental system, Percolator. Rather than running a series of high-latency map-reduces over large batches of documents, we now <a href="http://googleblog.blogspot.com/2010/06/our-new-search-index-caffeine.html">index individual documents at very low latency</a>. The result is a 50% reduction in search result age; our paper discusses this project and the implications of the result.</li><li><a href="http://research.google.com/pubs/pub36737.html">Availability in Globally Distributed Storage Systems</a>.<br />Reliable and efficient storage systems are a key component of cloud-based services. In this paper we characterize the availability properties of cloud storage systems based on extensive monitoring of Google's main storage infrastructure and present statistical models that enable further insight into the impact of multiple design choices, such as data placement and replication strategies. We demonstrate the utility of these models by computing data availability under a variety of replication schemes given the real patterns of failures observed in our fleet.<br /></li><li><a href="http://www.icsi.berkeley.edu/cgi-bin/pubs/publication.pl?ID=002971">Onix: A Distributed Control Platform for Large-scale Production Networks</a>.<br />There has been recent interest in a new networking paradigm called Software-Defined Networking (SDN). The crucial enabler for SDN is distributed control platform that shields developers from the details of the underlying physical infrastructure and allows them to write sophisticated control logic against a high-level API. Onix provides such a control platform for large-scale production networks.<br /></li></ul><p>In addition to the papers presented by current Googlers, we were also happy to see that the recipient of the <a href="http://googleblog.blogspot.com/2009/05/best-and-brightest.html">2009 Google Ph.D. Fellowship in Cloud Computing</a>, <a href="http://www.cs.washington.edu/homes/roxana/">Roxana Geambasu</a>, presented her work on <a href="http://www.cs.washington.edu/homes/roxana/acads/projects/vanish/osdi2010comet.pdf">Comet: An active distributed key-value store</a>.</p><p>Videos of all of the talks from OSDI are available on the <a href="http://www.usenix.org/events/osdi10/tech/">conference website</a> for attendees and current USENIX members.  There is also a <a href="http://www.youtube.com/user/USENIXAssociation">USENIX YouTube channel</a> with a growing subset of the conference videos open to everyone.</p><p>Google is making substantial progress on many of the grand challenge problems in computer science and artificial intelligence as part of its mission to organize the worlds information and make it useful.  Given the continuing increase in the scale of our distributed systems it’s fair to say we’ll have some other exciting new work to share at the next OSDI.  Hope to see you in 2012.</p><div class="blogger-post-footer"><img width='1' height='1' src='https://blogger.googleusercontent.com/tracker/21224994-1735983275886289638?l=googleresearch.blogspot.com' alt='' /></div>]]></content:encoded>
			<wfw:commentRss>https://googledata.org/google-research/google-at-usenix-symposium-on-operating-systems-design-and-implementation-osdi-%e2%80%9810/feed/</wfw:commentRss>
		<slash:comments>0</slash:comments>
		</item>
		<item>
		<title>Making an Impact on a Thriving Speech Research Community</title>
		<link>https://googledata.org/google-research/making-an-impact-on-a-thriving-speech-research-community/?utm_source=rss&#038;utm_medium=rss&#038;utm_campaign=making-an-impact-on-a-thriving-speech-research-community</link>
		<comments>https://googledata.org/google-research/making-an-impact-on-a-thriving-speech-research-community/#comments</comments>
		<pubDate>Mon, 11 Oct 2010 23:04:00 +0000</pubDate>
		<dc:creator><![CDATA[Research Admin]]></dc:creator>
				<category><![CDATA[Google Research]]></category>

		<guid isPermaLink="false"></guid>
		<description><![CDATA[Posted by Vincent Vanhoucke, Google ResearchWhile we continue to launch exciting new speech products--most recently Voice Actions and Google Search by Voice in Russian, Czech and Polish--we also strive to contribute to the academic research community b...]]></description>
				<content:encoded><![CDATA[<span class="byline-author">Posted by Vincent Vanhoucke, Google Research</span><br /><br /><a onblur="try {parent.deselectBloggerImageGracefully();} catch(e) {}" href="http://4.bp.blogspot.com/_mol72XYv6Jk/TLObmixIKlI/AAAAAAAAAP0/uZoeiLx8lgw/s1600/thad.JPG"><img style="float:right; margin:0 0 10px 10px;cursor:pointer; cursor:hand;width: 400px; height: 400px;" src="http://4.bp.blogspot.com/_mol72XYv6Jk/TLObmixIKlI/AAAAAAAAAP0/uZoeiLx8lgw/s400/thad.JPG" border="0" alt="" id="BLOGGER_PHOTO_ID_5526932254183926354" /></a>While we continue to launch exciting new speech products--most recently <a href="http://www.google.com/mobile/voice-actions/">Voice Actions</a> and <a href="http://www.google.com/mobile/">Google Search by Voice</a> in Russian, Czech and Polish--we also strive to contribute to the academic research community by sharing both innovative techniques and experiences with large-scale systems.<br /><br />This year’s gathering of the world’s experts in speech technology research, <a href="http://www.interspeech2010.org/">Interspeech 2010</a> in Makuhari, Japan, which Google co-sponsored, was a fantastic demonstration of the momentum of this community, driven by new challenges such as mobile voice communication, voice search, and the increasing international reach of speech technologies.<br /><br />Googlers published papers that showcased the breadth and depth of our speech recognition research.  Our work addresses both fundamental problems in acoustic and language modeling,  as well as the practical issues of building scalable speech interfaces that real people use everyday to make their lives easier.<br /><br />Here is a list of the papers presented by Googlers at the conference:<br /><ul><li><a href="http://research.google.com/pubs/pub36834.html">Direct Construction of Compact Context-Dependency Transducers From Data</a>, David Rybach and <a href="http://research.google.com/pubs/author125.html">Michael Riley</a> (Computer Speech &amp; Language Best Paper Award).</li><li><a href="http://research.google.com/pubs/pub36833.html">Voice Search for Development</a>, Etienne Barnard, Johan Schalkwyk, Charl van Heerden and <a href="http://research.google.com/pubs/author5289.html">Pedro J. Moreno</a>.</li><li><a href="http://research.google.com/pubs/pub36487.html">Unsupervised Discovery and Training of Maximally Dissimilar Cluster Models</a>, <a href="http://research.google.com/pubs/author21120.html">Françoise Beaufays</a>, <a href="http://research.google.com/pubs/author37534.html">Vincent Vanhoucke</a> and Brian Strope.</li><li><a href="http://research.google.com/pubs/pub36463.html">Search by Voice in Mandarin Chinese</a>, Jiulong Shan, Genqing Wu, Zhihong Hu, Xiliu Tang, <a href="http://research.google.com/pubs/author35845.html">Martin Jansche</a> and <a href="http://research.google.com/pubs/author5289.html">Pedro J. Moreno</a>.</li><li><a href="http://research.google.com/pubs/pub36756.html">On-Demand Language Model Interpolation for Mobile Speech Input</a>, Brandon Ballinger, <a href="http://research.google.com/pubs/author130.html">Cyril Allauzen</a>, <a href="http://research.google.com/pubs/author39407.html">Alexander Gruenstein</a>, and Johan Schalkwyk.</li><li><a href="http://research.google.com/pubs/pub36801.html">Building Transcribed Speech Corpora Quickly and Cheaply for Many Languages</a>, <a href="http://research.google.com/pubs/author37811.html">Thad Hughes</a>, Kaisuke Nakajima, Linne Ha, Atul Vasu, Pedro J. Moreno and Mike LeBeau.</li><li><a href="http://research.google.com/pubs/pub36577.html">Say What? Why Users Choose to Speak their Web Queries</a>, <a href="http://research.google.com/pubs/author34.html">Maryam Kamvar</a> and Doug Beeferman.</li><li><a href="http://research.google.com/pubs/pub36472.html">Study on Interaction between Entropy Pruning and Kneser-Ney Smoothing</a>, <a href="http://research.google.com/pubs/author6342.html">Ciprian Chelba</a>, Thorsten Brants, Will Neveitt and Peng Xu.</li><li><a href="http://research.google.com/pubs/pub36828.html">Decision Tree State Clustering with Word and Syllable Features</a>, <a href="http://research.google.com/pubs/author37501.html">Hank Liao</a>, Chris Alberti, Michiel Bacchiani and Olivier Siohan.</li></ul><div class="blogger-post-footer"><img width='1' height='1' src='https://blogger.googleusercontent.com/tracker/21224994-3666648602928390231?l=googleresearch.blogspot.com' alt='' /></div>]]></content:encoded>
			<wfw:commentRss>https://googledata.org/google-research/making-an-impact-on-a-thriving-speech-research-community/feed/</wfw:commentRss>
		<slash:comments>0</slash:comments>
<enclosure url="" length="" type="" />
		</item>
		<item>
		<title>Bowls and Learning</title>
		<link>https://googledata.org/google-research/bowls-and-learning/?utm_source=rss&#038;utm_medium=rss&#038;utm_campaign=bowls-and-learning</link>
		<comments>https://googledata.org/google-research/bowls-and-learning/#comments</comments>
		<pubDate>Thu, 07 Oct 2010 22:00:00 +0000</pubDate>
		<dc:creator><![CDATA[Research Admin]]></dc:creator>
				<category><![CDATA[Google Research]]></category>

		<guid isPermaLink="false"></guid>
		<description><![CDATA[Posted by Phil Long, Research TeamIt is easy to find the bottom of a bowl no matter where you start -- if you toss a marble anywhere into the bowl, it will roll downhill and find its way to the bottom.What does this have to do with Machine Learning?  A...]]></description>
				<content:encoded><![CDATA[<span class="byline-author">Posted by Phil Long, Research Team</span><br /><br />It is easy to find the bottom of a bowl no matter where you start -- if you toss a marble anywhere into the bowl, it will roll downhill and find its way to the bottom.<br /><br />What does this have to do with Machine Learning?  A natural way to try to construct an accurate classifier is to minimize the number of prediction errors the classifier makes on training data. The trouble is, even for moderate-sized data sets, minimizing the number of training errors is a computationally intractable problem.  A popular way around this is to assign different training errors different costs and to minimize the total cost.  If the costs are assigned in a certain way (according to a “<a href="http://www.stat.rutgers.edu/home/tzhang/papers/aos04_consistency.pdf">convex loss function</a>”), the total cost can be efficiently minimized the way a marble rolls to the bottom of a bowl.  <br /><br />In a recent <a href="http://www.phillong.info/publications/LS10_potential.pdf">paper</a>, Rocco Servedio and I show that no algorithm that works this way can achieve a simple and natural theoretical noise-tolerance guarantee that can be achieved by other kinds of algorithms.  A result like this is interesting for two reasons: first, it's important to understand what you cannot do with convex optimization in order to get a fuller understanding of what you can do with it. Second, this result may spur more research into noise-tolerant training algorithms using alternative approaches.<div class="blogger-post-footer"><img width='1' height='1' src='https://blogger.googleusercontent.com/tracker/21224994-807934366799730370?l=googleresearch.blogspot.com' alt='' /></div>]]></content:encoded>
			<wfw:commentRss>https://googledata.org/google-research/bowls-and-learning/feed/</wfw:commentRss>
		<slash:comments>0</slash:comments>
		</item>
		<item>
		<title>Remembering Fred Jelinek</title>
		<link>https://googledata.org/google-research/remembering-fred-jelinek/?utm_source=rss&#038;utm_medium=rss&#038;utm_campaign=remembering-fred-jelinek</link>
		<comments>https://googledata.org/google-research/remembering-fred-jelinek/#comments</comments>
		<pubDate>Sat, 18 Sep 2010 01:01:00 +0000</pubDate>
		<dc:creator><![CDATA[Research Admin]]></dc:creator>
				<category><![CDATA[Google Research]]></category>

		<guid isPermaLink="false"></guid>
		<description><![CDATA[Posted by Ciprian Chelba, Research TeamIt is with great sadness that we note the passing of Fred Jelinek, teacher and colleague to many of us here at Google. His seminal contributions to statistical modeling of speech and language influenced not only u...]]></description>
				<content:encoded><![CDATA[<span class="byline-author">Posted by Ciprian Chelba, Research Team</span><br /><br />It is with great sadness that we note the passing of Fred Jelinek, teacher and colleague to many of us here at Google. His seminal contributions to statistical modeling of speech and language influenced not only us, but many more members of the research community.<br /><br />Several of us at Google remember Fred:<br /><br /><span style="font-style:italic;">Ciprian Chelba</span>:<br />Fred was my thesis advisor at CLSP. My ten years of work in the field after graduation led me to increasingly appreciate the values that Fred instilled by personal example: work on the hard problem because it simply cannot be avoided, bring fundamental and original contributions that steer clear of incrementalism, exercise your creativity despite the risks entailed, and pursue your ideas with determination.<br /><br />I recently heard a comment from a colleague,  “A natural born leader is someone you follow even if only out of curiosity.” I immediately thought of Fred. Working with him marked a turning point in my life, and his influential role will be remembered.<br /><br /><span style="font-style:italic;">Bob Moore</span>:<br />I first met Fred Jelinek in 1984 at an IBM-sponsored workshop on natural-language processing. Fred's talk was my first exposure to the application of statistical ideas to language, and about the only thing I understood was the basic idea of N-gram language modeling: estimate the probability of the next word in a sequence based on a small fixed number of immediately preceding words. At the time, I was so steeped in the tradition of linguistically-based formal grammars that I was sure Fred's approach could not possibly be useful.<br /><br />Starting about five years later, however, I began to interact with Fred often at speech and language technology meetings organized by DARPA, as well as events affiliated with the Association for Computational Linguistics. Gradually, I (along with much of the computational linguistics community) began to understand and appreciate the statistical approach to language technology that Fred and his colleagues were developing, to the point that it now dominates the field of computational linguistics, including my own research. The importance of Fred's technical contributions and visionary leadership in bringing about this revolution in language technology cannot be overstated. The field is greatly diminished by his passing.<br /><br /><span style="font-style:italic;">Fernando Pereira</span>:<br />I met Fred first at a DARPA-organized workshop where one of the main topics was how to put natural language processing research on a more empirical, data-driven path. Fred was leading the charge for the move, drawing from his successes in speech recognition. Although I had already started exploring those ideas, I was not fully convinced by Fred’s vision. Nevertheless, Fred’s program raised many interesting research questions, and I could not resist some of them. Working on search for speech recognition at AT&amp;T, I was part of the  small team that invented the finite-state transducer representation of recognition models. I gave what I think was the first public talk on the approach at a workshop session that Fred chaired. It was Fred’s turn to be skeptical, and we had a spirited exchange in the discussion period. At the time, I was disappointed that I had failed to interest Fred in the work, but later I was delighted when Fred became a strong supporter of our work after a JHU Summer workshop where Michael Riley led the use of our software tools in successful experiments with a team of JHU researchers and students. Indeed, in hindsight, Fred was right to be skeptical before we had empirical validation for the approach, and his strong support when the results started coming in was thus much more meaningful and gratifying. Through these experiences and much more, I came to respect immensely Fred’s pioneer spirit, vision, and sharp mind. Many of my most successful projects benefited directly or indirectly from his ideas, his criticism, and his building of thriving institutions, from CLSP to links with the research team at Charles University in Prague. I saw Fred last at ACL in Uppsala. He was in great form, and we had a good discussion on funding for the summer workshops. I am very sad that he will not be with us to continue these conversations.<br /><br /><span style="font-style:italic;">Shankar Kumar</span>:<br />Fred was my academic advisor at CLSP/JHU and I interacted with him throughout my Ph.D. program. I had the privilege of having him on my thesis committee. My very first exposure to research in speech and NLP was through an independent study that I did under him. A few years later, I was his teaching assistant for the speech recognition class. Fred's energy and passion for research made a strong impression on me back then and continues to influence my work to this day. I remember Fred carefully writing up his ideas and sending them out as a starting point to our discussions. While I found this curiously amusing at the time, I now think this was his unique approach to ensure clarity of thought and to steer the discussion without distractions. Fred's enthusiasm for learning new concepts was infectious! I attended several classes and guest lectures with him - graphical models, NLP, and many more. His insightful questions and his active participation in each one of these classes made them memorable for me. He epitomized what a life-long learner should be. I will always recall Fred's advice on sharing credit generously. In his own words, "The contribution of a research paper does not get divided by the number of authors". By his passing, we have lost a role model who dedicated his life to research and whose contributions will continue to impact and shape the field for years to come.<br /><br /><span style="font-style:italic;">Michael Riley</span>:<br />I got to know Fred pretty well having attended two of the CLSP six-week summer workshops, working on a few joint grants, and visiting CLSP in between. If there is a ‘father of speech recognition’, its got to be Fred Jelinek - he led the IBM team that invented and popularized many of the key methods used today. His intellect, wide knowledge, and force of will served him well later as the leader of the JHU Center for Language and Speech Processing - a sort of academic hearth where countless speech/NLP researchers and students interacted over the years in seminars and workshops. I was impressed that at an age when many retired and after which most of his IBM colleagues had gone into (very lucrative) financial engineering, he remained a vigorous, leading academic. Fernando mentioned the initial skepticism he had for our work on weighted FSTs for ASR. Some years later though I heard that he praised the work to my lab director, Larry Rabiner, on a plane ride that likely helped my promotion shortly thereafter. And no discussion of Fred would be complete without a mention of his inimitable humor, delivered in that loud Czech-accented voice:<br /><blockquote><i>Riley</i> [at workshop planning meeting]: “Could they hold the summer workshop in some nicer place than Baltimore to help attract people?”<br /><i>Fred</i>: “Riley, we’ll hold it in Rome next year and get better people than you!”<br /><br /><i>Seminar presenter</i>: [fumbling with Windows configuration for minutes].<br /><i>Fred</i> [very loud]: “How long do we have to endure this high-tech torture?”</blockquote><br />The website of The Johns Hopkins University’s Center for Language and Speech Processing links to Fred’s own descriptions of his <a href="http://www.clsp.jhu.edu/people/jelinek/promoce.html">life</a> and <a href="http://www.mitpressjournals.org/doi/pdf/10.1162/coli.2009.35.4.35401">technical achievements</a>.<div class="blogger-post-footer"><img width='1' height='1' src='https://blogger.googleusercontent.com/tracker/21224994-5186426778493489804?l=googleresearch.blogspot.com' alt='' /></div>]]></content:encoded>
			<wfw:commentRss>https://googledata.org/google-research/remembering-fred-jelinek/feed/</wfw:commentRss>
		<slash:comments>0</slash:comments>
		</item>
		<item>
		<title>Frowns, Sighs, and Advanced Queries &#8212; How does search behavior change as search becomes more difficult?</title>
		<link>https://googledata.org/google-research/frowns-sighs-and-advanced-queries-how-does-search-behavior-change-as-search-becomes-more-difficult/?utm_source=rss&#038;utm_medium=rss&#038;utm_campaign=frowns-sighs-and-advanced-queries-how-does-search-behavior-change-as-search-becomes-more-difficult</link>
		<comments>https://googledata.org/google-research/frowns-sighs-and-advanced-queries-how-does-search-behavior-change-as-search-becomes-more-difficult/#comments</comments>
		<pubDate>Fri, 17 Sep 2010 15:18:00 +0000</pubDate>
		<dc:creator><![CDATA[Research Admin]]></dc:creator>
				<category><![CDATA[Google Research]]></category>

		<guid isPermaLink="false"></guid>
		<description><![CDATA[Posted by Anne Aula, Rehan Khan, and Zhiwei Guan, User Experience TeamHow does search behavior change as search becomes more difficult?At Google, we strive to make finding information easy, efficient, and even fun. However, we know that once in a while...]]></description>
				<content:encoded><![CDATA[<span class="byline-author">Posted by Anne Aula, Rehan Khan, and Zhiwei Guan, User Experience Team</span><br /><br /><span style="font-weight:bold;"><a href="http://portal.acm.org/citation.cfm?id=1753326.1753333">How does search behavior change as search becomes more difficult?</a></span><br /><br />At Google, we strive to make finding information easy, efficient, and even fun. However, we know that once in a while, finding a specific piece of information turns out to be tricky. Based on dozens of user studies over the years, we know that it’s relatively easy for an observer to notice that the user is having problems finding the information, by watching changes in language, body language, and facial expressions:<br /><br /><a onblur="try {parent.deselectBloggerImageGracefully();} catch(e) {}" href="http://2.bp.blogspot.com/_mol72XYv6Jk/TJOIEbs_YYI/AAAAAAAAAPk/TPfDrzCCpeg/s1600/Picture+52.png"><img style="display:block; margin:0px auto 10px; text-align:center;cursor:pointer; cursor:hand;width: 400px; height: 297px;" src="http://2.bp.blogspot.com/_mol72XYv6Jk/TJOIEbs_YYI/AAAAAAAAAPk/TPfDrzCCpeg/s400/Picture+52.png" border="0" alt=""id="BLOGGER_PHOTO_ID_5517903578196304258" /></a><br /><br />Computers, however, don’t have the luxury of observing a user the way another person would.  But would it be possible for a computer to somehow tell that the user is struggling to find information? <br /><br />We decided to find out. We first ran a study in the usability lab where we gave users search tasks, some of which we knew to be difficult. The first couple of searches always looked pretty much the same independent of task difficulty: users formulated a query, quickly scanned the results and either clicked on a result or refined the query. However, after a couple of unsuccessful searches, we started noticing interesting changes in behavior. In addition to many of them sighing or starting to bite their nails, users sometimes started to type their searches as natural language questions, they sometimes spent a very long time simply staring at the results page, and they sometimes completely changed their approach to the task. <br /><br />We were fascinated by these findings as they seemed to be signals that the computer could potentially detect while the user is searching. We formulated the initial findings from the usability lab study as hypotheses which we then tested in a larger web-based user study. <br /><br />The overall findings were promising: we found five signals that seemed to indicate that users were struggling in the search task. Those signals were: use of question queries, use of advanced operators, spending more time on the search results page, formulating the longest query in the middle of the session, and spending a larger proportion of the time on the search results page. None of these signals alone are strong enough predictors of users having problems in search tasks. However, when used together, we believe we can use them to build a model that will one day make it possible for computers to detect frustration in real time.<br /><br />You can read the full text of the paper <a href="http://dub.washington.edu/pubs/215">here</a>.<div class="blogger-post-footer"><img width='1' height='1' src='https://blogger.googleusercontent.com/tracker/21224994-5775364719966833024?l=googleresearch.blogspot.com' alt='' /></div>]]></content:encoded>
			<wfw:commentRss>https://googledata.org/google-research/frowns-sighs-and-advanced-queries-how-does-search-behavior-change-as-search-becomes-more-difficult/feed/</wfw:commentRss>
		<slash:comments>0</slash:comments>
<enclosure url="" length="" type="" />
		</item>
		<item>
		<title>Focusing on Our Users: The Google Health Redesign</title>
		<link>https://googledata.org/google-research/focusing-on-our-users-the-google-health-redesign/?utm_source=rss&#038;utm_medium=rss&#038;utm_campaign=focusing-on-our-users-the-google-health-redesign</link>
		<comments>https://googledata.org/google-research/focusing-on-our-users-the-google-health-redesign/#comments</comments>
		<pubDate>Wed, 15 Sep 2010 13:00:00 +0000</pubDate>
		<dc:creator><![CDATA[Research Admin]]></dc:creator>
				<category><![CDATA[Google Research]]></category>

		<guid isPermaLink="false"></guid>
		<description><![CDATA[Posted by Hendrik Mueller, User Experience ResearcherWhen I relocated to New York City a few years ago, some of the most important health information for me to have on hand was my immunization history. At the time, though, my health records were scatte...]]></description>
				<content:encoded><![CDATA[<span class="byline-author">Posted by Hendrik Mueller, User Experience Researcher</span><br /><br />When I relocated to New York City a few years ago, some of the most important health information for me to have on hand was my immunization history. At the time, though, my health records were scattered, and it felt like a daunting task to organize them -- a not-uncommon problem that many people face. For me, the solution came when Google Health became available in May of 2008, and I started using it to organize my health information and keep it more manageable. I also saw the potential to do much more within Google Health, such as tracking my overall fitness goals. When I joined the Google Health team as the lead user experience researcher, I was curious about the potential for Google Health to impact people’s lives beyond things like immunization tracking and how we could make the product a lot easier to use. So I set out to explore how to expand and improve Google Health.<br /><br />Here at Google, we focus on the user throughout the entire product development process. So before Google Health was first launched, we interviewed many people about how they managed their medical records and other health information to better understand their needs. We then iteratively created and tested multiple concepts and designs. After our initial launch, we followed up with actual Google Health users through surveys, interviews, and usability studies to understand how well we were meeting their needs.<br /><br /><a onblur="try {parent.deselectBloggerImageGracefully();} catch(e) {}" href="http://2.bp.blogspot.com/_mol72XYv6Jk/TI_f_GNEFTI/AAAAAAAAAOo/ja65Ya1lhIU/s1600/HealthSurvey_Cutout_Final.png"><img style="display:block; margin:0px auto 10px; text-align:center;cursor:pointer; cursor:hand;width: 400px; height: 331px;" src="http://2.bp.blogspot.com/_mol72XYv6Jk/TI_f_GNEFTI/AAAAAAAAAOo/ja65Ya1lhIU/s400/HealthSurvey_Cutout_Final.png" border="0" alt="" id="BLOGGER_PHOTO_ID_5516874343642305842" /></a><br /><br />From this user research, we learned what was working in the product and what needed to be improved. Here are some of the things our users found especially useful:<br /><ul><li>Organizing and tracking health-related information in a single place that is accessible from anywhere at any time</li><li>Sharing medical records easily with loved ones and health care providers, either by allowing online access or by printing out health summaries</li><li>Referencing rich information about <a href="https://health.google.com/health/ref/index.html">health topics</a>, aggregated from trusted sources and Google search results</li></ul><br />Our users also described to us the benefits they saw from using Google Health:<br /><blockquote><span style="font-style:italic;">“Google Health gives me many tools to research my prescriptions and symptoms, and to track all of the many tests I keep having. Google Health made several necessary and cumbersome tasks easy and worry free.”<br /><br />“For years now, I've tried to remember my son’s allergies and medications, but the list has grown so long, that I kept forgetting one or two when a doctor asked me about them. That can't happen again because I now have a single place to keep up with them. And I love the fact that I can print off information for situations when I really need it.”<br /><br />“I really like that I can share my profile with others. I want my mom to know my medical information, just in case anything ever happens to me.”</span></blockquote><br />While we learned that our users were clearly getting positive results from using Google Health, our research also taught us that more was needed. We learned that we needed to make fundamental changes to fully meet the needs of all of our current and prospective users, such as those that are chronically ill, those who care for family members, and especially those users looking to track and improve their wellness and fitness.<br /><br />On this last point, our user surveys already pointed out that there was more we could do to help our users track and manage their wellness, not just their sickness, so we conducted further research about how people collect, monitor, track, and analyze their wellness data. We interviewed several people in their homes and invited others into our usability labs. As a result, we identified several areas where we could improve Google Health to make it a more useful wellness tool, including:<br /><ul><li>Dedicated wellness tracking including pre-built and custom trackers</li><li>Efficient manual data entry as well as automatic data collection through devices</li><li>A customizable summary dashboard of wellness and other health topics</li><li>Goal setting and progress tracking using interactive charts</li><li>Personalized pages for each topic with rich charts, journaling, and related information</li></ul><br />These insights led us to a whole new set of design proposals. We gathered feedback on the resulting sketches, wire-frames, and screenshots from active and new Google Health users. The results throughout this process were eye-opening. While we were on the right track for some parts of the design, other parts had to be corrected or even redesigned. We went through several iterations until we had a design that tested well and we felt met the user needs our research had uncovered. Finally, we conducted several usability studies with a functioning prototype throughout the product development process to continuously improve usability and function.<br /><br /><a onblur="try {parent.deselectBloggerImageGracefully();} catch(e) {}" href="http://1.bp.blogspot.com/_mol72XYv6Jk/TI_j8wnAR0I/AAAAAAAAAO4/UDjqUYCHCK4/s1600/LabPicture_left_Final.JPG"><img style="display:block; margin:0px auto 10px; text-align:center;cursor:pointer; cursor:hand;width: 400px; height: 256px;" src="http://1.bp.blogspot.com/_mol72XYv6Jk/TI_j8wnAR0I/AAAAAAAAAO4/UDjqUYCHCK4/s400/LabPicture_left_Final.JPG" border="0" alt="" id="BLOGGER_PHOTO_ID_5516878701532301122" /></a><br /><br />At the end, the collaboration between the user experience, engineering, and product management teams resulted in an entirely new user experience for Google Health combined with a set of new functionality that is now available for you to try out at <a href="http://www.google.com/health">www.google.com/health</a>. See for yourself how the old and new versions compare. Here is a screenshot of a health profile in the new version:<br /><br /><a onblur="try {parent.deselectBloggerImageGracefully();} catch(e) {}" href="http://2.bp.blogspot.com/_mol72XYv6Jk/TI_khVVfxbI/AAAAAAAAAPA/1wRoHHAvA9I/s1600/Dashboard_newUI_Final.png"><img style="display:block; margin:0px auto 10px; text-align:center;cursor:pointer; cursor:hand;width: 400px; height: 260px;" src="http://2.bp.blogspot.com/_mol72XYv6Jk/TI_khVVfxbI/AAAAAAAAAPA/1wRoHHAvA9I/s400/Dashboard_newUI_Final.png" border="0" alt="" id="BLOGGER_PHOTO_ID_5516879329866270130" /></a><br /><br />And this is how the same account and profile looked in the old user interface:<br /><br /><a onblur="try {parent.deselectBloggerImageGracefully();} catch(e) {}" href="http://1.bp.blogspot.com/_mol72XYv6Jk/TI_k8uNlUDI/AAAAAAAAAPI/ns7BhVvJozo/s1600/Dashboard_oldUI_Final.png"><img style="display:block; margin:0px auto 10px; text-align:center;cursor:pointer; cursor:hand;width: 400px; height: 260px;" src="http://1.bp.blogspot.com/_mol72XYv6Jk/TI_k8uNlUDI/AAAAAAAAAPI/ns7BhVvJozo/s400/Dashboard_oldUI_Final.png" border="0" alt="" id="BLOGGER_PHOTO_ID_5516879800400433202" /></a><br /><br />As a Google Health user, I am excited to take advantage of the new design and have already started using it for my own exercise and weight tracking. And on behalf of the user experience team and the entire Google Health team, we’re excited about being able to bring you a new design and more powerful tool that we think will meet more of your health and wellness needs.<br /><br />We look forward to continuing to explore how we can make Google Health even more useful and easier to use for people like you. As you use Google Health, you may see a link to a feedback survey at the top of the application. If you do, please take the time to fill it out - we will be listening to your input!<div class="blogger-post-footer"><img width='1' height='1' src='https://blogger.googleusercontent.com/tracker/21224994-3067348915733480180?l=googleresearch.blogspot.com' alt='' /></div>]]></content:encoded>
			<wfw:commentRss>https://googledata.org/google-research/focusing-on-our-users-the-google-health-redesign/feed/</wfw:commentRss>
		<slash:comments>0</slash:comments>
<enclosure url="" length="" type="" />
		</item>
		<item>
		<title>Discontinuous Seam Carving for Video Retargeting</title>
		<link>https://googledata.org/google-research/discontinuous-seam-carving-for-video-retargeting/?utm_source=rss&#038;utm_medium=rss&#038;utm_campaign=discontinuous-seam-carving-for-video-retargeting</link>
		<comments>https://googledata.org/google-research/discontinuous-seam-carving-for-video-retargeting/#comments</comments>
		<pubDate>Mon, 13 Sep 2010 19:24:00 +0000</pubDate>
		<dc:creator><![CDATA[Research Admin]]></dc:creator>
				<category><![CDATA[Google Research]]></category>

		<guid isPermaLink="false"></guid>
		<description><![CDATA[Posted by Matthias Grundmann and Vivek Kwatra, Google ResearchVideos come in different sizes, resolutions and aspect ratios, but the device used for playback, may it be your TV, mobile phone, or laptop, only has a fixed resolution and form factor. As a...]]></description>
				<content:encoded><![CDATA[<span class="byline-author">Posted by Matthias Grundmann and Vivek Kwatra, Google Research</span><br /><br />Videos come in different sizes, resolutions and aspect ratios, but the device used for playback, may it be your TV, mobile phone, or laptop, only has a fixed resolution and form factor. As a result, you cannot watch your favorite old show that came in 4:3 on your new 16:9 HDTV without having black bars on the side, referred to as letterboxing. Likewise, widescreen movies and user-videos uploaded on YouTube are shot using various cameras with wide-ranging formats, so they do not fit completely on the screen. As an alternative to letterboxing, several devices try to <span style="font-style:italic;">upscale</span> the content uniformly, which either changes the aspect ratio, making everything look stretched out, or simply crop the frame, thereby discarding any content that cannot fit the screen after scaling.<br /><br />At Google Research, together with collaborators from Georgia Tech, we have developed an algorithm that resizes (or <span style="font-style:italic;">retargets</span>) videos to fit the form factor of a given device without cropping, stretching or letterboxing. Our approach uses all of the screen’s precious pixels, while striving to deliver as much video-content of the original as possible. The result is a video that adapts to your needs, so you don’t have to adapt to the video.<br /><br /><a onblur="try {parent.deselectBloggerImageGracefully();} catch(e) {}" href="http://2.bp.blogspot.com/_mol72XYv6Jk/TI6ybOhVHqI/AAAAAAAAAOg/H29dlCANEy8/s1600/Blogpostandhighlightsentences.jpg"><img style="display:block; margin:0px auto 10px; text-align:center;cursor:pointer; cursor:hand;width: 400px; height: 246px;" src="http://2.bp.blogspot.com/_mol72XYv6Jk/TI6ybOhVHqI/AAAAAAAAAOg/H29dlCANEy8/s400/Blogpostandhighlightsentences.jpg" border="0" alt="" id="BLOGGER_PHOTO_ID_5516542774399540898" /></a><br /><span class="Apple-style-span" style="font-size: small;">Six frames from the result of our retargeting algorithm applied to a sub-clip of “Apologize”, © 2006 One Republic. Original frame is shown on the left, our resized result on the right. The original content is fit to a new aspect ratio.</span><br /><br />The key insight is that we can separate the video into <i>salient</i> and <i>non-salient</i> content, which are then treated differently. Think of salient content as actors, faces, or structured objects, where the viewer anticipates specific, important details to perceive it as being correct and unaltered. We cannot change this content beyond uniform scaling without it being noticeable. On the other hand, non-salient content, such as sky, water or a blurry out-of-focus background can be squished or stretched without changing the overall appearance or the viewer noticing a dramatic change.<br /><br />Our technique, which we call <i>discontinuous seam carving</i> -- named so because it modifies the video by adding or removing disconnected seams (or chains) of pixels -- allows greater freedom in the resizing process than previous approaches. By optimizing for the retargeted video to be consistent with the original, we carefully preserve the <i>shape and motion</i> of the salient content while being less restrictive with non-salient content. The key innovations of our research include: (a) a solution that maintains temporal continuity of the video in addition to preserving its spatial structure, (b) space-time smoothing for automatic as well as interactive (user-guided) salient content selection, and (c) sequential frame-by-frame processing conducive for arbitrary length and streaming video. The outcome is a scalable system capable of retargeting videos featuring complex motions of actors and cameras, highly dynamic content and camera shake. For more details, please refer to our <a href="http://static.googleusercontent.com/external_content/untrusted_dlcp/research.google.com/en/us/pubs/archive/36246.pdf">paper</a> or visit the <a href="http://cpl.cc.gatech.edu/projects/videoretargeting/">project web-site</a>.<div class="blogger-post-footer"><img width='1' height='1' src='https://blogger.googleusercontent.com/tracker/21224994-8324635245137123258?l=googleresearch.blogspot.com' alt='' /></div>]]></content:encoded>
			<wfw:commentRss>https://googledata.org/google-research/discontinuous-seam-carving-for-video-retargeting/feed/</wfw:commentRss>
		<slash:comments>0</slash:comments>
<enclosure url="" length="" type="" />
		</item>
		<item>
		<title>Google Search by Voice:  A Case Study</title>
		<link>https://googledata.org/google-research/google-search-by-voice-a-case-study/?utm_source=rss&#038;utm_medium=rss&#038;utm_campaign=google-search-by-voice-a-case-study</link>
		<comments>https://googledata.org/google-research/google-search-by-voice-a-case-study/#comments</comments>
		<pubDate>Thu, 09 Sep 2010 23:59:00 +0000</pubDate>
		<dc:creator><![CDATA[Research Admin]]></dc:creator>
				<category><![CDATA[Google Research]]></category>

		<guid isPermaLink="false"></guid>
		<description><![CDATA[Posted by Johan Schalkwyk, Google ResearchWind the clock back two years with your smart phone in hand. Try to recall doing a search for a restaurant or the latest scores of your favorite sports team. If you’re like me you probably won’t even bother...]]></description>
				<content:encoded><![CDATA[<span class="byline-author">Posted by Johan Schalkwyk, Google Research</span><br /><br />Wind the clock back two years with your smart phone in hand. Try to recall doing a search for a restaurant or the latest scores of your favorite sports team. If you’re like me you probably won’t even bother, or you’ll suffer with tiny keys or fat fingers on a touch screen. With Google Search by Voice all that has changed. Now you just tap the microphone, speak, and within seconds you see the result. No more fat fingers.<br /><br />Google Search by Voice is a result of many years of investment in speech at Google. We started by building our own recognizer (aka GReco ) from the ground up. Our first foray in search by voice was doing local searches with <a href="http://www.google.com/goog411/index.html">GOOG-411</a>. Then, in November 2008, we launched <a href="http://googleblog.blogspot.com/2008/11/now-you-can-speak-to-google-mobile-app.html">Google Search by Voice</a>. Now you can search the entire Web using your voice.<br /><br />What makes search by voice really interesting is that it requires much more than a just good speech recognizer. You also need a good user interface and a good phone like an Android in the hands of millions of people. Besides the excellent computational platform and data availability, the project succeeded due to Google’s culture built around teams that wholeheartedly tackle such challenges with the conviction that they will set a new bar. <br /><br />In our book chapter, “Google Search by Voice: A Case Study”, we describe the basic technology, the supporting technologies, and the user interface design behind Google Search by Voice. We describe how we built it and what lessons we have learned. As the product required many helping hands to build, this chapter required many helping hands to write. We believe it provides a valuable contribution to the academic community.<br /><br />The book, <a href="http://www.springer.com/engineering/signals/book/978-1-4419-5950-8">Advances in Speech Recognition</a>, is available for purchase from Springer.<div class="blogger-post-footer"><img width='1' height='1' src='https://blogger.googleusercontent.com/tracker/21224994-2178210901062200452?l=googleresearch.blogspot.com' alt='' /></div>]]></content:encoded>
			<wfw:commentRss>https://googledata.org/google-research/google-search-by-voice-a-case-study/feed/</wfw:commentRss>
		<slash:comments>0</slash:comments>
		</item>
		<item>
		<title>Towards Energy-Proportional Datacenters</title>
		<link>https://googledata.org/google-research/towards-energy-proportional-datacenters/?utm_source=rss&#038;utm_medium=rss&#038;utm_campaign=towards-energy-proportional-datacenters</link>
		<comments>https://googledata.org/google-research/towards-energy-proportional-datacenters/#comments</comments>
		<pubDate>Wed, 01 Sep 2010 19:18:00 +0000</pubDate>
		<dc:creator><![CDATA[Research Admin]]></dc:creator>
				<category><![CDATA[Google Research]]></category>

		<guid isPermaLink="false"></guid>
		<description><![CDATA[Posted by Dennis Abts, Michael R. Marty, Philip M. Wells, Peter Klausler, and Hong LiuThis is part of the series highlighting some notable publications by Googlers.  At Google, we operate large datacenters containing clusters of servers, networking swi...]]></description>
				<content:encoded><![CDATA[<span class="byline-author">Posted by Dennis Abts, Michael R. Marty, Philip M. Wells, Peter Klausler, and Hong Liu</span><br /><br /><span style="font-style:italic;">This is part of the <a href="http://googleresearch.blogspot.com/2010/07/google-publications.html">series</a> highlighting some notable publications by Googlers.  </span><br /><br />At Google, we operate large datacenters containing clusters of servers, networking switches, and more.   While this gear costs a lot of money, an increasingly important cost -- both in terms of dollars and environmental impact -- is the electricity that drives the computing clusters and  the cooling infrastructure.  Since our clusters often do not run at full utilization, Google recently put forth a call to industry and researchers to develop energy proportional computer systems.  With such systems, the power consumed by our clusters would be directly proportional to utilization. Servers consume the most electricity, and therefore researchers have responded to Google’s call by focusing their attention towards servers.  As the servers become increasingly energy proportional, however, the “always on” network fabric that connects servers together will consume an increasing fraction of datacenter power unless it too becomes energy proportional.<br /><br />In a <a href="http://static.googleusercontent.com/external_content/untrusted_dlcp/www.google.com/en/us/research/pubs/archive/36462.pdf">paper</a> recently published at the International Symposium on Computer Architecture (ISCA), we push further towards the goal of energy-proportional computing by focusing on the energy usage of high-bandwidth, highly-scalable cluster networking fabrics.  This research considers a broad set of architectural and technological solutions to optimize energy usage without sacrificing performance. First, we show how the Flattened Butterfly network topology uses less power since it uses less switching chips and fewer links than a comparable-performance network built using the more conventional Fat Tree topology.  Second, our approach takes advantage of the observation that when network demand is low, we can reduce the speed at which links transmit data.  We show via simulation, that by tuning the speeds of the links very rapidly, we can reduce power consumption with little impact on performance. Finally, our research is a further call to action for the academic and industry research communities to make energy efficiency, and energy proportionality in particular, a first-class citizen in networking research.  Put together, our proposed techniques can reduce energy cost for typical Google workloads seen in our production datacenters by millions of dollars!<div class="blogger-post-footer"><img width='1' height='1' src='https://blogger.googleusercontent.com/tracker/21224994-2295552561482451008?l=googleresearch.blogspot.com' alt='' /></div>]]></content:encoded>
			<wfw:commentRss>https://googledata.org/google-research/towards-energy-proportional-datacenters/feed/</wfw:commentRss>
		<slash:comments>0</slash:comments>
		</item>
		<item>
		<title>Google Publications</title>
		<link>https://googledata.org/google-research/google-publications/?utm_source=rss&#038;utm_medium=rss&#038;utm_campaign=google-publications</link>
		<comments>https://googledata.org/google-research/google-publications/#comments</comments>
		<pubDate>Fri, 30 Jul 2010 18:41:00 +0000</pubDate>
		<dc:creator><![CDATA[Research Admin]]></dc:creator>
				<category><![CDATA[Google Research]]></category>

		<guid isPermaLink="false"></guid>
		<description><![CDATA[Posted by Corinna Cortes and Alfred Spector, Google ResearchWe often get asked if Google scientists and engineers publish technical papers, and the answer is, “Most certainly, yes.”    Indeed, we have a formidable research capability, and we encour...]]></description>
				<content:encoded><![CDATA[<span class="byline-author">Posted by Corinna Cortes and Alfred Spector, Google Research</span><div><span class="byline-author"></span><br />We often get asked if Google scientists and engineers publish technical papers, and the answer is, “Most certainly, yes.”    Indeed, we have a formidable research capability, and we encourage publications as well as other forms of technical dissemination--including our contributions to open source and standards and the introduction of new APIs and tools, which have proven to sometimes be foundational.<br /><br />Needless to say, with our great commitment to technical excellence in computer science and related disciplines, we find it natural and rewarding to contribute to the scientific community and to ongoing technical debates.  And we know that it is important for Google to help create the fundamental building blocks upon which continuing advances can occur.<br /><br />To be specific, Googlers publish hundreds of technical papers that appear in journals, books, and conference and workshop proceedings every year. These deal with specific applications and engineering questions, algorithmic and data structure problems, and important theoretical problems in computer science, mathematics, and other areas, that can guide our algorithmic choices.  While the publications are interesting in their own right, they also offer a glance at some of the key problems we face when dealing with very large data sets and demonstrate other questions that arise in our engineering design at Google.<br /><br />We’d like to highlight a few of the more noteworthy papers from the first trimester of this year.  The papers reflect the breadth and depth of the problems on which we work.  We find that virtually all aspects of computer science, from systems and programming languages, to algorithms and theory, to security, data mining, and machine learning are relevant to our research landscape.  A more complete list of our publications can be found <a href="http://research.google.com/pubs/papers.html">here</a>.<br /><br />In the coming weeks we will be offering a more in-depth look at these publications, but here are some summaries:<br /><br /><b>Speech Recognition</b><br /><br /><span class="Apple-style-span" style="font-size: small;">"Google Search by Voice: A Case Study," by Johan Schalkwyk, Doug Beeferman, Francoise Beaufays, Bill Byrne, Ciprian Chelba, Mike Cohen, Maryam Garrett, Brian Strope, to appear in </span><i><span class="Apple-style-span" style="font-size: small;">Advances in Speech Recognition: Mobile Environments, Call Centers, and Clinics</span></i><span class="Apple-style-span" style="font-size: small;">, Amy Neustein (Ed.), Springer-Verlag 2010.</span><br /><br />Google Search by Voice is a result of many years of investment in speech at Google. In our book chapter, “Google Search by Voice: A Case Study,” we describe the basic technology, the supporting technologies, and the user interface design behind Google Search by Voice. We describe how we built it and what lessons we have learned. Google search by voice is growing rapidly and being built in many languages. Along the way we constantly encounter new research problems providing the perfect atmosphere for doing research on real world problems.<br /><br /><span style="font-weight:bold;">Computer Architecture &amp; Networks &amp; Distributed Systems</span><br /><br /><span class="Apple-style-span" style="font-size: small;">"Energy-proportional Datacenter Networks," by Dennis Abts, Mike Marty, Philip Wells, Peter Klausler, Hong Liu, </span><span style="font-style:italic;"><span class="Apple-style-span" style="font-size: small;">International Symposium on Computer Architecture</span></span><span class="Apple-style-span" style="font-size: small;">, ISCA, June 201</span>0.<br /><br />Google researchers have called on industry and academia to develop energy-proportional computing systems, where the energy consumed is directly proportional to the utilization of the system. In this work, we focus on the energy usage of high-bandwidth, highly scalable cluster networks. Through a combination of an energy-efficient topology and dynamic fine-grained control of link speeds, our proposed techniques show the potential to significantly reduce both electricity and environmental costs.<br /><br /><span style="font-weight:bold;">Economics &amp; Market Algorithms</span><br /><br /><span class="Apple-style-span" style="font-size: small;">"Quasi-Proportional Mechanisms: Prior-free Revenue Maximization," by  Vahab S. Mirrokni, S. Muthukrishnan, Uri Nadav, </span><span style="font-style:italic;"><span class="Apple-style-span" style="font-size: small;">Latin American Theoretical Informatics Symposium</span></span><span class="Apple-style-span" style="font-size: small;">, LATIN, April 2010.</span><br /><br />Say a seller wishes to sell an item, but the buyers value it vastly differently. What is a suitable auction to sell the item, in terms of efficiency as well as revenue? First and second price auctions will be efficient but will only extract the lower value in equilibrium; if one knows the distributions from which values are drawn, then setting a reserve price will get optimal revenue but will not be efficient. This paper views this problem as prior-free auction and proposes a quasi-proportional allocation in which the probability that an item is allocated to a bidder depends (quasi-proportionally) on their bids. The paper also proves existence of an equilibrium for quasi-proportional auctions and shows how to compute them efficiently. Finally, the paper shows that these auctions have high efficiency and revenue.<br /><br /><span class="Apple-style-span" style="font-size: small;">"Auctions with Intermediaries," Jon Feldman, Vahab Mirrokni, S. Muthukrishnan, Mallesh Pai, </span><span style="font-style:italic;"><span class="Apple-style-span" style="font-size: small;">ACM Conference on Electronic Commerce</span></span><span class="Apple-style-span" style="font-size: small;">, EC, June 2010. </span><br /><br />We study an auction where the bidders are middlemen, looking in turn to auction off the item if they win it.  This setting arises naturally in online advertisement exchange systems, where the participants in the exchange are ad networks looking to sell ad impressions to their own advertisers. We present optimal strategies for both the bidders and the auctioneer in this setting. In particular, we show that  the optimal strategy for bidders is to choose a  randomized reserve price, and the optimal reserve price of the centeral auctioneer may depend on the number of bidders (unlike the case when there are no middlemen).<br /><br /><span style="font-weight:bold;">Computer Vision</span><br /><br /><span class="Apple-style-span" style="font-size: small;">"Discontinuous Seam-Carving for Video Retargeting," Matthias Grundmann, Vivek Kwatra, Mei Han, Irfan Essa, </span><span style="font-style:italic;"><span class="Apple-style-span" style="font-size: small;">Computer Vision and Pattern Recognition</span></span><span class="Apple-style-span" style="font-size: small;">, CVPR, June 2010.</span><br /><br />Playing a video on devices with different form factors requires resizing (or <span style="font-style:italic;">retargeting</span>) the video to fit the resolution of the given device. We have developed a content-aware technique for video retargeting based on <span style="font-style:italic;">discontinuous seam-carving</span>, which unlike standard methods like uniform scaling and cropping, strives to retain salient content (such as actors, faces and structured objects) while discarding relatively unimportant pixels (such as the sky or a blurry background). The key innovations of our research include: (a) a solution that maintains temporal continuity of the video in addition to preserving its spatial structure, (b) space-time smoothing for automatic as well as interactive (user-guided) salient content selection, and (c) sequential frame-by-frame processing conducive for arbitrary length and streaming video.<br /><br /><span style="font-weight:bold;">Machine Learning</span><br /><br /><span class="Apple-style-span" style="font-size: small;">"Random classification noise defeats all convex potential boosters," Philip M. Long, Rocco A. Servedio, Machine Learning, vol. 78 (2010), pp. 287-304.</span><br /><br />A popular approach that has been used to tackle many machine learning problems recently is to formulate them as optimization problems in which the goal is to minimize some “convex loss function.”  This is an appealing formulation because these optimization problems can be solved in much the same way that a marble rolls to the bottom of a bowl.  However, it turns out that there are drawbacks to this formulation.  In "Random Classification Noise Defeats All Convex Potential Boosters," we show that any learning algorithm that works in this way can fail badly if there are noisy examples in the training data.  This research motivates further study of other approaches to machine learning, for which there are algorithms that are provably more robust in the presence of noise.<br /><br /><span style="font-weight:bold;">IR</span><br /><br /><span class="Apple-style-span" style="font-size: small;">"Clustering Query Refinements by User Intent," Eldar Sadikov, Jayant Madhavan, Lu Wang, Alon Halevy, </span><span style="font-style:italic;"><span class="Apple-style-span" style="font-size: small;">Proceedings of the International World Wide Web Conference</span></span><span class="Apple-style-span" style="font-size: small;">, WWW, April 2010.</span><br /><br />When users pose a search query, they usually have an underlying intent or information need, and the sequence of queries he or she poses in single search sessions is usually determined by the user's underlying intent. Our research demonstrates that there typically are only a small number of prominent underlying intents for a given user query. Further, these intents can be identified very accurately by an analysis of anonymized search query logs. Our results show that underlying intents almost always correspond to well-understood high-level concepts.<br /><br /><span style="font-weight:bold;">HCI</span><br /><br /><span class="Apple-style-span" style="font-size: small;">"How does search behavior change as search becomes more difficult?", Anne Aula, Rehan Khan, Zhiwei Guan, </span><span style="font-style:italic;"><span class="Apple-style-span" style="font-size: small;">Proceedings of the ACM Conference on Human Factors in Computing Systems</span></span><span class="Apple-style-span" style="font-size: small;">, CHI , April 2010.</span><br /><br />Seeing that someone is getting frustrated with a difficult search task is easy for another person--just look for the frowns, and listen for the sighs.  But could a computer tell that you're getting frustrated from just the limited behavior a search engine can observe?  Our study suggests that it can: when getting frustrated, our data shows that users start to formulate question queries, they start to use advanced operators, and they spend a larger proportion of the time on the search results page. Used together, these signals can be used to build a model that can potentially detect user frustration.<br /></div><div class="blogger-post-footer"><img width='1' height='1' src='https://blogger.googleusercontent.com/tracker/21224994-5992154403033942385?l=googleresearch.blogspot.com' alt='' /></div>]]></content:encoded>
			<wfw:commentRss>https://googledata.org/google-research/google-publications/feed/</wfw:commentRss>
		<slash:comments>0</slash:comments>
		</item>
		<item>
		<title>Googlers receive multiple awards at the 2010 International Conference on Machine Learning</title>
		<link>https://googledata.org/google-research/googlers-receive-multiple-awards-at-the-2010-international-conference-on-machine-learning/?utm_source=rss&#038;utm_medium=rss&#038;utm_campaign=googlers-receive-multiple-awards-at-the-2010-international-conference-on-machine-learning</link>
		<comments>https://googledata.org/google-research/googlers-receive-multiple-awards-at-the-2010-international-conference-on-machine-learning/#comments</comments>
		<pubDate>Tue, 27 Jul 2010 23:13:00 +0000</pubDate>
		<dc:creator><![CDATA[Research Admin]]></dc:creator>
				<category><![CDATA[Google Research]]></category>

		<guid isPermaLink="false"></guid>
		<description><![CDATA[Posted by Fernando Pereira, Research DirectorGooglers were recognized in three of the four paper awards at ICML 2010: Sajid Siddiqi was co-recipient of the best paper award for Hilbert Space Embeddings of Hidden Markov Models with Le Song, Byron Boots,...]]></description>
				<content:encoded><![CDATA[<span class="byline-author">Posted by Fernando Pereira, Research Director</span><br /><br />Googlers were recognized in three of the four <a href="http://www.icml2010.org/awards.html">paper awards</a> at ICML 2010: <div><ul><li><a href="http://research.google.com/pubs/siddiqi.html">Sajid Siddiqi</a> was co-recipient of the best paper award for <span style="font-style:italic;"><a href="http://www.icml2010.org/abstracts.html#495">Hilbert Space Embeddings of Hidden Markov Models</a></span> with Le Song, Byron Boots, Geoff Gordon, and Alex Smola </li><li><a href="http://www.cs.berkeley.edu/~jduchi/">John Duchi</a>, who is also a graduate student at UC Berkeley, was co-recipient of the best student paper award for <span style="font-style:italic;"><a href="http://www.icml2010.org/abstracts.html#421">On the Consistency of Ranking Algorithms</a> <span class="Apple-style-span" style="font-style: normal;">with Lester Mackey and Michael Jordan</span></span></li><li><span>And last but not the least, </span><a href="http://research.google.com/pubs/author28.html">Yoram Singer</a><span> was co-recipient of the best 10-year paper award for the most influential paper of ICML 2000, </span><a href="http://research.google.com/pubs/author28.html">Reducing Multiclass to Binary: A Unifying Approach for Margin Classifiers</a><span> (pdf) with Erin Allwein and Robert Schapire, which has currently </span><a href="http://scholar.google.com/scholar?hl=en&amp;as_sdt=2000&amp;q=reducing+Multiclass+to+Binary:+A+Unifying+Approach+for+Margin+Classifiers">852 citations</a><span> in Google Scholar. </span></li></ul>I feel a particular connection to this last paper as Rob and Yoram were members of technical staff and Erin a student intern at the department I headed at AT&amp;T Labs when this work was done.</div><div><br /></div>Congratulations to all!<div class="blogger-post-footer"><img width='1' height='1' src='https://blogger.googleusercontent.com/tracker/21224994-6044955422802897296?l=googleresearch.blogspot.com' alt='' /></div>]]></content:encoded>
			<wfw:commentRss>https://googledata.org/google-research/googlers-receive-multiple-awards-at-the-2010-international-conference-on-machine-learning/feed/</wfw:commentRss>
		<slash:comments>0</slash:comments>
		</item>
		<item>
		<title>Announcing our Q2 Research Awards</title>
		<link>https://googledata.org/google-research/announcing-our-q2-research-awards/?utm_source=rss&#038;utm_medium=rss&#038;utm_campaign=announcing-our-q2-research-awards</link>
		<comments>https://googledata.org/google-research/announcing-our-q2-research-awards/#comments</comments>
		<pubDate>Thu, 22 Jul 2010 16:23:00 +0000</pubDate>
		<dc:creator><![CDATA[Research Admin]]></dc:creator>
				<category><![CDATA[Google Research]]></category>

		<guid isPermaLink="false"></guid>
		<description><![CDATA[Posted by Maggie Johnson, Director of Education &#38; University RelationsWe’re excited to announce the latest round of Google Research Awards, our program which identifies and supports full-time faculty pursuing research in areas of mutual interest....]]></description>
				<content:encoded><![CDATA[<span class="byline-author">Posted by Maggie Johnson, Director of Education &amp; University Relations</span><br /><br />We’re excited to announce the latest round of Google Research Awards, our program which identifies and supports full-time faculty pursuing research in areas of mutual interest. From a record number of submissions, we are funding 75 awards across 18 different areas—a total of more than $4 million.<br /><br />The areas that received the highest level of funding for this round were systems and infrastructure, human computer interaction, multimedia and security. We also continue to develop more collaborations internationally. In this round, 26 percent of the funding was awarded to universities outside the U.S.<br /><br />Here are some examples from this round of awards:<br /><ul><li><b>Jeremy Cooperstock, McGill University</b>. <i>A Spatialized Audio Map System for Mobile Blind Users</i> (Geo/maps): A mobile audio system that provides location-based information, primarily for use by the blind and visually impaired communities. </li><li><b>Alexander Pretschner, Karlsruhe Institute of Technology, Germany</b>. <i>Towards Operational Privacy </i>(Security and privacy): Provide a framework for precise semantic definitions in policies for domain-specific applications to give users a way to define the exact behaviour they expect from a system in application-specific contexts.</li><li><b>Erik Brynjolfsson, Massachusetts Institute of Technology</b>.  <i>The Future of Prediction - How Google Searches Foreshadow Housing Prices and Quantities </i>(Economics and market algortihms): How data from search engines like Google provide a highly accurate but simple way to predict future business activities. </li><li><b>Stephen Pulman, Oxford University Computing Laboratory</b>. <i>Automatic Generation of Natural Language Descriptions of Visual Scenes</i> (Natural language processing): Develop a system that automatically generates a description of a visual scene.</li><li><b>Jennifer Rexford, Princeton</b>. <i>Rethinking Wide-Area Traffic Management </i>(Software and hardware systems infrastructure): Drawing on mature techniques from optimization theory, design new traffic-management solutions where the hosts, routers, and management system cooperate in a more effective way. </li><li><b>John Quinn, Makerere University, Uganda</b>.  <i>Mobile Crop Surveillance in the Developing World </i>(Multimedia search and audio/video processing): A computer vision system using camera-enabled mobile devices to monitor the spread of viral disease among staple crops. </li><li><b>Allison Druin, University of Maryland</b>. <i>Understanding how Children Change as Searchers </i>(Human-computer interaction): Do children change as searchers as they age?  How do searchers typically shift between roles over time? If children change, how many of them become Power Searchers? If children don’t change, what roles do they typically demonstrate?</li><li><b>Ronojoy Adhikari, The Institute of Mathematical Sciences, India</b>. <i>Machine Learning of Syntax in Undeciphered Scripts </i>(Machine learning): Devise algorithms that would learn to search for evidence of semantics in datasets such as the Indus script.</li></ul><br />You can find the full list of this round’s award recipients <a href="http://static.googleusercontent.com/external_content/untrusted_dlcp/www.google.com/en/us/googleblogs/pdfs/google_research_awards_july2010.pdf">here</a> (pdf). More information on our research award program can be found on our <a href="http://research.google.com/university/relations/research_awards.html">website</a>.<div class="blogger-post-footer"><img width='1' height='1' src='https://blogger.googleusercontent.com/tracker/21224994-7811078160479700959?l=googleresearch.blogspot.com' alt='' /></div>]]></content:encoded>
			<wfw:commentRss>https://googledata.org/google-research/announcing-our-q2-research-awards/feed/</wfw:commentRss>
		<slash:comments>0</slash:comments>
		</item>
		<item>
		<title>Google launches Korean Voice Search</title>
		<link>https://googledata.org/google-research/google-launches-korean-voice-search/?utm_source=rss&#038;utm_medium=rss&#038;utm_campaign=google-launches-korean-voice-search</link>
		<comments>https://googledata.org/google-research/google-launches-korean-voice-search/#comments</comments>
		<pubDate>Wed, 30 Jun 2010 21:45:00 +0000</pubDate>
		<dc:creator><![CDATA[Research Admin]]></dc:creator>
				<category><![CDATA[Google Research]]></category>

		<guid isPermaLink="false"></guid>
		<description><![CDATA[Posted by Mike Schuster &#38; Martin Jansche, Google ResearchOn June 16th, we launched our Korean voice search system. Google Search by Voice has been available in various flavors of English since 2008, in Mandarin and Japanese since 2009, and in Frenc...]]></description>
				<content:encoded><![CDATA[<span class="byline-author">Posted by Mike Schuster &amp; Martin Jansche, Google Research</span><br /><br />On June 16th, we launched our <a href="http://googlemobile.blogspot.com/2010/06/annyeong-haseyo-to-google-search-by.html">Korean voice search system</a>. Google Search by Voice has been available in various flavors of <a href="http://googlemobile.blogspot.com/2008/11/google-mobile-app-for-iphone-now-with.html">English</a> since 2008, in <a href="http://googleresearch.blogspot.com/2009/11/google-search-by-voice-learns-mandarin.html">Mandarin</a> and <a href="http://googleresearch.blogspot.com/2009/12/teaching-computer-to-understand.html">Japanese</a> since 2009, and in <a href="http://googlemobile.blogspot.com/2010/06/salut-willkommen-benvenuto-bienvenido.html">French, Italian, German and Spanish</a> just a few weeks ago (some more details in a recent blog <a href="http://googleresearch.blogspot.com/2010/06/google-search-by-voice-now-available-in.html">post</a>).<br /><br />Korean speech recognition has received less attention than English, which has been studied extensively around the world by teams in both English and non-English speaking countries.  Fundamentally, our methodology for developing a Korean speech recognition system is similar to the process we have used for other languages. We created a set of statistical models: an acoustic model for the basic sounds of the language, a language model for the words and phrases of the language, and a dictionary mapping the words to their pronunciations. We trained our acoustic model using a large quantity of recorded and transcribed Korean speech. The language model was trained using anonymized Korean web search queries. Once these models were trained, given an audio input, we can compute and display the most likely spoken phrase, along with its search result.<br /><br />There were several challenges in developing a Korean speech recognition system, some unique to Korean, some typical of Asian languages and some universal to all languages. Here are some examples of problems that stood out:<br /><br /><ul><li>Developing a Korean dictionary:   Unlike English, where there are many publicly-available dictionaries for mapping words to their pronunciations, there are very few available for Korean. Since our Korean recognizer knows several hundred thousand words, we needed to create these mappings ourselves. Luckily, Korean has one of the most elegant and simple writing systems in the world (created in the 15th century!) and this makes mapping Korean words to pronunciations relatively straightforward. However, we found that Koreans also use quite a few English words in their queries, which complicates the mapping process. To predict these pronunciations, we built a statistical model using data from an existing (smaller) Korean dictionary.</li><li>Korean word boundaries:   Although Korean orthography uses spaces to indicate word boundaries (unlike Japanese or Mandarin), we found that people use word boundaries inconsistently for search queries. To limit the size of the vocabulary generated from the search queries, we used statistical techniques to cut rare long words into smaller sub-words (similarly to the system we developed for Japanese).<br /></li><li>Pronunciation exceptions:  Korean (like all other languages) has many exceptions for pronunciations that are not immediately obvious. For example, numbers are often written as digit sequences but not necessarily spoken this way (2010 = 이천십). The same is true for many common alphanumeric sequences like “mp3”, “kbs2” or mixed queries like “삼성 tv”, which often contain spelled letters and possibly English spoken digits as opposed to Korean ones.</li><li>Encoding issues:  Korean script (Hangul) is written in syllabic blocks, with each block containing at least two of the 24 modern Hangul letters (Jamo), at least one consonant and one vowel.  Including the normal ASCII characters this brings the total number of possible basic characters to over 10000, not including Hanja (used mostly in the formal spelling of names). So, despite its simple writing system, Korean still presents the same challenge of handling a large alphabet that is typical of Asian languages.</li><li>Script ambiguity:  We found that some users like to use English native words and others the Korean transliteration (example: “ncis season 6” vs. “ncis 시즌6”). This makes it challenging to train and evaluate the system. We use a metric that estimates whether our transcription will give the correct web page result on the user’s smart phone screen, and such script variations make this tricky.</li><li>Recognizing rare words:  The recognizer is good at recognizing things users often type into the search engine, such as cities, shops, addresses, common abbreviations, common product model numbers and well-known names like “김연아”. However, rare words (like many personal names) are often harder for us to recognize. We continue to work on improving those.</li><li>Every speaker sounds different:  People speak in different styles, slow or fast, with an accent or without, have lower or higher pitched voices, etc. To make our system work for all these different conditions, we trained our system using data from many different sources to capture as many conditions as possible.<br /></li></ul><br />When speech recognizers make errors, the reason is usually that the models are not good enough, and that often means they haven’t been trained on enough data. For Korean (and all other languages) our cloud computing infrastructure allows us to retrain our models frequently and using an ever growing amount of data to continually improve performance. Over time, we are committed to improve the system regularly to make speech a user-friendly input method on mobile devices.<div class="blogger-post-footer"><img width='1' height='1' src='https://blogger.googleusercontent.com/tracker/21224994-5838927280305250053?l=googleresearch.blogspot.com' alt='' /></div>]]></content:encoded>
			<wfw:commentRss>https://googledata.org/google-research/google-launches-korean-voice-search/feed/</wfw:commentRss>
		<slash:comments>0</slash:comments>
		</item>
		<item>
		<title>Google Search by Voice now available in France, Italy, Germany and Spain</title>
		<link>https://googledata.org/google-research/google-search-by-voice-now-available-in-france-italy-germany-and-spain/?utm_source=rss&#038;utm_medium=rss&#038;utm_campaign=google-search-by-voice-now-available-in-france-italy-germany-and-spain</link>
		<comments>https://googledata.org/google-research/google-search-by-voice-now-available-in-france-italy-germany-and-spain/#comments</comments>
		<pubDate>Mon, 14 Jun 2010 23:00:00 +0000</pubDate>
		<dc:creator><![CDATA[Research Admin]]></dc:creator>
				<category><![CDATA[Google Research]]></category>

		<guid isPermaLink="false"></guid>
		<description><![CDATA[Posted by Thad Hughes, Martin Jansche, and Pedro Moreno, Google ResearchGoogle’s speech team is composed of people from many different cultural backgrounds. Indeed, if we count the languages spoken by our teammates, the number comes to well over a do...]]></description>
				<content:encoded><![CDATA[<span class="byline-author">Posted by Thad Hughes, Martin Jansche, and Pedro Moreno, Google Research</span><br /><br />Google’s speech team is composed of people from many different cultural backgrounds. Indeed, if we count the languages spoken by our teammates, the number comes to well over a dozen. Given our own backgrounds and interests, we are naturally excited to extend our software to work with many different languages and dialects.  After testing the waters with <a href="http://googlemobile.blogspot.com/2008/11/google-mobile-app-for-iphone-now-with.html">English</a>, <a href="http://googleresearch.blogspot.com/2009/11/google-search-by-voice-learns-mandarin.html">Mandarin Chinese</a>, and <a href="http://googleresearch.blogspot.com/2009/12/teaching-computer-to-understand.html">Japanese</a>, we decided to tackle <a href="http://googlemobile.blogspot.com/2010/06/salut-willkommen-benvenuto-bienvenido.html">four main European languages</a> which are often referred to as FIGS - French, Italian, German and Spanish.<br /><br />Developing Voice Search systems in each of these languages presented its own challenges.  French and Spanish required special work to deal with diacritic and accent marks (e.g. ç in French, ñ in Spanish).  When we develop a new language we tweak our dictionaries based on user generated content.  To our surprise we found that a lot of this content in French and Spanish often uses non-standard orthography.  For example a French speaker might type “francoise” into a search engine and still expect it to return results for “Françoise”.  Likewise in Spanish a user might type “espana” and expect results for the term “España”.  Of course a lot of this has to do with the fact that, until recently, domain names (like www.elpais.es) did not allow diacritics, and that entering special characters is often painful but omitting diacrictics is usually not an obstacle to communication.  However, non-standard spellings distort the intended pronunciations.  For example, if “francoise” were a real French word, one would expect it to be pronounced “franquoise”.  In order to capture the intended pronunciation of the non-standard spellings, we fixed the orthography in our dictionaries for Spanish and French automatically.  While this is not perfect, it deals with many of the offending cases.<br /><br />Since our Voice search systems typically understand more than a million different words in each language, developing pronunciation dictionaries is one of the most critical tasks.  We need the dictionary to match what the user said with the written form.  Not surprisingly we found that dictionary development for some languages like Spanish and Italian to be extremely easy, as they have very regular orthographies.  In fact the core of our Spanish pronunciation module consists of less than 100 lines of source code.  Other languages like German and French have more complex orthographies.  For example in French “au”, “eaux” and “hauts” are all pronounced “o”.<br /><br />A notable aspect of German (especially “Internet German”) is that a lot of English words are in common usage.  We do our best to recognize thousands of English words, even though English contains some sounds that don’t exist in German, like “th” in “the”.  One of the trickiest examples we came across was when one of our volunteers read “nba playoffs 2009”, saying “nba playoffs” in English followed by “zwei tausend neun” in German.  So go ahead and search for “Germany’s Next Topmodel” or “Postbank Online”, see if it works for you.<br /><br />German is also notorious for having long, complex words.  Our favorite examples include:<br /><ul><li><a href="http://de.wikipedia.org/wiki/BKrFQG#Berufskraftfahrer-Qualifikations-Gesetz">Berufskraftfahrerqualifikationsgesetz </a>(or shorter: BKrFQG)</li><li><a href="http://en.wiktionary.org/wiki/Eierschalensollbruchstellenverursacher">Eierschalensollbruchstellenverursacher</a></li><li><a href="http://de.wikipedia.org/wiki/Verkehrsinfrastrukturfinanzierungsgesellschaft">Verkehrsinfrastrukturfinanzierungsgesellschaft</a></li><li><a href="http://de.wikipedia.org/wiki/Stichpimpulibockforcelorum">Stichpimpulibockforcelorum</a></li><li>Hypothalamus-Hypophysen-Nebennierenrinde-Achse</li></ul><br />Just for fun, compare how long it takes you to say these to Voice Search vs. typing them.<br /><br />Even though a vocabulary size of one million words sounds like a large number, each of these languages has even more words, so we need a procedure to select which ones to model. We obviously do not do this manually and instead use statistical procedures to identify the list of words we will allow. We do this by looking at many sources of data and looking at the frequency of words. It is therefore surprising to find sometimes really weird terms selected by our algorithms. For example in Spanish we found these unusual words:<br /><ul><li><a href="http://en.wikipedia.org/wiki/Supercalifragilisticexpialidocious">supercalifragilisticespialidoso</a></li><li><a href="http://es.wikipedia.org/wiki/Los_Chiripitifl%C3%A1uticos">chiripitiflautico</a></li><li><a href="http://es.wikipedia.org/wiki/M%C3%BAsculo_esternocleidomastoideo">esternocleidomastoideo</a></li></ul> <br />So, in the unlikely event that  you ever try a Spanish voice search query like this “imágenes del músculo supercalifragilisticoespialidoso chiripitiflautico esternocleidomastoideo” you may be surprised to see that it works.<br /><br />French, Italian, German, and Spanish are spoken in many parts of the world.  In this first release of Google Search by Voice in these languages, we initially only support the varieties spoken in France, Italy, Germany, and Spain, respectively.  The reason is that almost all aspects of a Voice Search system are affected by regional variation: French speakers from different regions have slightly different accents, use a number of different words, and will want to search for different things.  Eventually, we plan to support other regions as well, and we will work hard to make sure our systems work well for all of you.<br /><br />So, we hope you find these new voice search system useful and fun to use. We definitely had a “supercalifragilisticoespialidoso chiripitiflautico” time developing them.<div class="blogger-post-footer"><img width='1' height='1' src='https://blogger.googleusercontent.com/tracker/21224994-4472073388895647452?l=googleresearch.blogspot.com' alt='' /></div>]]></content:encoded>
			<wfw:commentRss>https://googledata.org/google-research/google-search-by-voice-now-available-in-france-italy-germany-and-spain/feed/</wfw:commentRss>
		<slash:comments>0</slash:comments>
		</item>
		<item>
		<title>Google Fusion Tables celebrates one year of data management</title>
		<link>https://googledata.org/google-research/google-fusion-tables-celebrates-one-year-of-data-management/?utm_source=rss&#038;utm_medium=rss&#038;utm_campaign=google-fusion-tables-celebrates-one-year-of-data-management</link>
		<comments>https://googledata.org/google-research/google-fusion-tables-celebrates-one-year-of-data-management/#comments</comments>
		<pubDate>Wed, 09 Jun 2010 20:00:00 +0000</pubDate>
		<dc:creator><![CDATA[Research Admin]]></dc:creator>
				<category><![CDATA[Google Research]]></category>

		<guid isPermaLink="false"></guid>
		<description><![CDATA[Posted by Alon Halevy, Google Research and Rebecca Shapley, User ExperienceA year ago we launched Google Fusion Tables, an easy way to integrate, visualize and collaborate on data tables in the Google cloud. You used it and saw the potential, and told ...]]></description>
				<content:encoded><![CDATA[<span class="byline-author">Posted by Alon Halevy, Google Research and Rebecca Shapley, User Experience</span><br /><br />A year ago we <a href="http://googleresearch.blogspot.com/2009/06/google-fusion-tables.html">launched Google Fusion Tables</a>, an easy way to integrate, visualize and collaborate on data tables in the Google cloud. You used it and saw the potential, and told us what else you wanted. Since then, we’ve responded by offering programmatic access through the <a href="http://googlecode.blogspot.com/2009/12/google-fusion-tables-api.html">Fusion Tables API</a>, math across data columns owned by multiple people, and search on the collection of tables that have been made public. We published about Fusion Tables in SIGMOD 2010 and in the First Symposium on Cloud Computing.  And since the map visualizations were such a hit, we made them even better by supporting <a href="http://google-latlong.blogspot.com/2010/02/mapping-your-data-with-google-fusion.html">large numbers of points, lines and polygons</a>, <a href="http://earth.google.com/outreach/tutorial_fusion_sample.html#custompopup">custom HTML in map pop-up balloons</a> complete with <a href="http://earth.google.com/outreach/tutorials.html#tab1">tutorials</a> and <a href="http://googlegeodevelopers.blogspot.com/2010/05/map-your-data-with-maps-api-and-fusion.html">integration with the Google Maps API</a>. We’ve made all this capability available on Google’s cloud and are excited to see examples every day of how our cloud approach to data tables is changing the game and making structured data management, collaboration, and publishing fast, easy, and open. <br /><br />But more exciting than all the features we’ve been releasing is the things that people have been *doing* with Fusion Tables. News agencies have been taking advantage of Fusion Tables to map data that governments make public, and tell a more complete story (see the <a href="http://latimesblogs.latimes.com/lanow/2010/05/the-interactive-map-above-lists-the-439-los-angeles-medical-marijuana-dispensaries-that-must-shut-down-by-june-7-when-the.html">L.A. Times</a>, <a href="http://www.knoxnews.com/data/meth-lab-busts/">Knoxville News</a>, and <a href="http://blog.apps.chicagotribune.com/2010/03/04/quickly-visualize-and-map-a-data-set-using-google-fusion-tables/">Chicago Tribune</a>).  Just this month the State of California kicked off an <a href="http://www.ca.gov/appsforcalifornians/">application development contest</a>, hosting data sets like <a href="http://www.programmableweb.com/api/california-race-and-ethnic-population-projections">this one</a> in Fusion Tables for easy API access for developers.  And the US Department of Health and Human Services held the Community Health Data Forum, where attendees presented data applications such as the <a href="http://googlepublicpolicy.blogspot.com/2010/06/making-us-community-health-data.html">heart-friendly and people-friendly hospital-finder</a>, built with Google Fusion Tables. <br /><br />It continues to astound us how quickly our users are able to pull together these kinds of compelling data applications with Fusion Tables, again showing the power of a cloud approach to data. Fusion Tables were the multimedia extension to Joseph Rossano’s art exhibit on Butterflies and DNA barcodes, an easy way to <a href="http://www.thecalifornian.com/article/99999999/NEWS01/100406036/Mapping+Monterey+County+real+estate+sales&template=theme&theme=">map real-estate in Monterey county</a> or <a href="http://otrobache.com/">potholes in Spain</a>, provided the geo-catalog for <a href="http://windpoleventures.com/wpv/">wind power data</a> and <a href="http://www.ethanolretailer.com/find-an-e85-station">ethanol-selling stations</a>, and even the data backend for an <a href="http://www.gwopa.org/grubs/">geo portal to organize water data for Africa</a>, among many, many other uses. <br /><br />As we head into our second year, we’re looking forward to delivering more tools that make data management easier and more powerful on the web. What’s next for <a href="http://tables.googlelabs.com/">Fusion Tables</a>? Request your favorite features on our <a href="http://productideas.appspot.com/#25/e=1260b5">Feature Request</a> (a special implementation of Google Moderator), and follow the latest progress of Fusion Tables on our <a href="https://groups.google.com/group/fusion-tables-users-group?pli=1">User Group</a>, <a href="http://www.facebook.com/pages/Google-Fusion-Tables/127016660645629">Facebook</a>, and <a href="http://twitter.com/GoogleFT">Twitter</a>.  We love to hear from you!<div class="blogger-post-footer"><img width='1' height='1' src='https://blogger.googleusercontent.com/tracker/21224994-5301267421941643386?l=googleresearch.blogspot.com' alt='' /></div>]]></content:encoded>
			<wfw:commentRss>https://googledata.org/google-research/google-fusion-tables-celebrates-one-year-of-data-management/feed/</wfw:commentRss>
		<slash:comments>0</slash:comments>
		</item>
	</channel>
</rss>
