Real Time Search Event

Video Statistics and Information

Video
Captions Word Cloud
Reddit Comments
Captions
>> MAYER: Good morning. Thank you for joining us here at the Computer History Museum. We're so excited to be here within so many and surrounded by so many wonderful pieces of computing history. And it's an especially perfect backdrop for today's search event, which focuses on the future of search and the innovation that's happening in search. At Google, when we think about the future of search, we think about four key elements: modes, media, language, and personalization. What do I mean about each of these? Modes really refers to modalities. How do people search? Today, the predominant mode is that people type keywords into desktop computers. When we look at the future of search, we think they'll be many more modalities. What happens with mobile phones? What happens if you can talk to the search engine? What happens if you could give concepts or pictures that could cue off of--the search results could cue off of? We're really excited about the fact that in the future there'll be many different ways of searching. And when you look at how search will grow in the future and change, we think these many different modalities are really what will drive search forward and really grow with overall adoptions to even greater numbers than we see today. On the flipside, there is media. Media is what is appearing within the search results. And the Web, of course, has gotten very, very rich; books, movies, news, local, maps, all on the Web. And so, our search results have to mirror that same richness that we feel on the Web. So we're constantly looking at how can we make Google more comprehensive and more relevant with regard to media. So it's not just about 10 blue links; it's about the best answers. Media also leads the next piece, which is language. We really are working at Google to try and branch across languages, break down the language barrier because this focus on language in translation is what unlocks the Web. Today, we are able to translate from 51 different languages into all known pairs of those 51 languages. And we have 173 local domains, because we really foresee a world in the future where you can search and find your answer wherever they exist in the world, whatever language it's written in. And the final component that we see in the future of search is personalization. We don't know a lot about what the search engine will look like 30 years from now, 50 years from now, and it's because search changes so quickly that we--it's hard to actually pinpoint what that future looks like. But one thing we do know is that the results would be more personalized. They'll know what you are an expert in. They'll know who your friends with and where you're located. And, ultimately, our results will become more rich and more relevant to our users because of that personalization. And there is a fifth component to the future of search, which is the rate of progress. There has to be a consistent rate of innovation pushing us towards this future. And we've always been a company that likes to launch early and often and iterate. And we've done lots and lots of search features over the years. And in October, we released a new blog series called "This We Can Search." And "This We Can Search" chronicles each week all of the new user visible features that appear in search. So, our users get to see the latest and greatest, and they also get a sense of this pace that's driving us towards the future of search. And since October 2nd, in the past 67 days, we've actually launched 33 different search innovations. So that's one innovation every two days. And if you look at the innovations, they include things like the new homepage, Flu Shot Finder, personalized search on mobile, and they all fit neatly into these four categories of media, of personalization, modes, and languages. And we're really excited about this overall focus that we have on search, this rate of innovation. And today's event really drives us forward even further increasing that rate of innovation and new features that we're launching. Today's event will focus mostly on modes and media. And so, without further ado, I want to introduce our master of modalities who's driving our Mobile Innovation, vice president of engineering, Vic Gundotra. >> GUNDOTRA: Well, thank you very much, Marissa. You know, I'm very excited to have a chance today to speak to you about innovations in mobile search. But before I dive into those, I'd like you to join me and just take a step back for a moment. Think about the technological innovation that was so profound that it changed mankind. What comes to your mind? Maybe you think of Gutenberg's printing press. Maybe you think of the steam engine. Maybe you think of electricity. And it's true; those innovations really change the course of human history. But what's interesting is that at their outset, the full impact of those innovations was not understood. You know, Gutenberg was broke within a few years. The first mass-produced newspapers that came from Gutenberg's printing press happened many, many decades later, or who could have predicted that the steam engine would lead to the Industrial Revolution or that the invention of electricity would one day lead to the Internet and to microwaves. At Google, we argue that that same dynamic may be happening in the personal computer space. You know, PCs are, depending on how you count, 27 to about 33 old. In other words, we are just in the third decade of the personal computer revolution. And it may be that only now have our eyes become open to what the possibilities may be. In fact at Google, we see three major trends converging in combination with mobile phones that enable new scenarios that absolutely excite us. Let me talk to you about those trends then I'll get in and show you some of those scenarios. Now, the first trend is Moore's Law or computing. Now, I am in my 40s. All I've ever known in my life is Moore's Law. Ever since I was a kid and I had electronic toys to the computers I used as a teenager, I knew one thing for sure; that next year whatever device I have would be better, faster, cheaper. It's the ubiquitous, pervasive law we know as Moore's Law. And it's a powerful trend in computing. That's trend number one, computing. The second trend is far more recent. That is the trend of connectivity. You know, think back just as recently as a decade and a half ago. Think back to 1995. If someone came to you in 1995 and said, "Look, there's a time that's coming that billions of devices, every device would be connected to every other device." Would you have believed that? You know, the more technical you are, the more cynical you likely would have been. And, you know, if you think in 1995, the best file servers of the day, things like Novell NetWare, you know, they connected a few thousand simultaneous connections and the idea that you have billions of connections seemed a little bit absurd. Today, of course, we take it for granted. The Internet has happened. You don't even bat an eye when someone next to you takes out their cell phone and controls their TiVo at home. It is an amazing change, this change of connectivity that has swept the world. Computing, connectivity--and the third trend is the most recent, and that is the emergence of powerful clouds. Now, when I say cloud, I mean huge amounts of compute resources that programmers have at their disposal, data centers the sizes of football fields that host massive amounts of computational power and the ability to manipulate huge data models. When you combine these three things: computing, connectivity, and the cloud, and then you think about what's happening with mobile devices, you get something very, very interesting. Let's talk about mobile devices. Let me go grab a mobile device here. This is an Android device, common Smartphone. And you think about something like this, it's got built-in sensors. It's got a camera. It's got a GPS chip. It's got a speaker. It's got maybe an accelerometer. You know, by themselves, these sensors are not that extraordinary. Some of you may say, "So what? This camera isn't very special. My Nikon or Canon camera from 10 years ago surpasses this in quality." And you could be right. Or the microphone--what's the big deal about the microphone? You know, the microphone I use in my church or synagogue, much better than this microphone. But not when you compare this in the context of computing, connectivity, and the cloud. You see, when you take that camera, and you connect it to the cloud, it becomes an eye. That microphone connected to a cloud becomes an ear. And in fact, what we're going to show you this morning is innovations that are the combinations of those three trends and mobile phones that allow you for the first time to do powerful new things including search by sight, search by location, and search by voice. Let's begin with search by voice. Now, some of you may remember that it was about a year ago that we introduced Google--Google Voice Search available for the iPhone in the Google Mobile App. And during 2009, we worked very, very hard to improve the accuracy and the recognition rates of that voice recognition. You might be surprised at how good it's gotten. Let me show you. This is Google Mobile App, and it's running this Voice Recognition Software, Google Search by Voice. You simply bring it to your ear and speak a query. So, let me try a query, something like pictures of Barack Obama with the French President at the G8 Summit. You know, unlikely that you would ever type in a query that long. And there you have it. Isn't that amazing? Now, as impressive as that query is, it becomes even more impressive when you think about what just happened. What happened was we took your voice file, the digital representation of your expression, sent it up to the Google Cloud where it was broken down into phrases, sentences, words, and those were compared against the billions of daily Google queries we get. A probability score was associated with all the potential textual representations, the textual expressions of what we thought you said, they were ranked, and then the best ones were sent back to your phone all within what you saw here in fractions of a second. Really amazing, amazing work. Now, we've done more than just work on the improvement on accuracy of Voice Search. We've also been adding languages. You heard Marissa talked about our focus on languages. Last month, we announced Mandarin Voice Search. Now, think about our customers in China, our users in China, who have struggled to enter in with a notoriously difficult Mandarin character set on their mobile phones now being able use just voice. Let me do a demo for you. Same app, Google Mobile App available from the App Store, except in this case, I'm going to show you the Mandarin version of this particular app. Okay. There you go. So you see that it's in Mandarin and I'll try my only Mandarin query that I'm capable of doing. [SPEAKS FOREIGN LANGUAGE] That was McDonald's in Beijing. Let's see if we get it. There we go, all the McDonald's in Beijing. Okay. Now, I'm very happy to announce that today, and joining English and Mandarin, we have a new language we'll be supporting and that is Japanese. Now, instead of me trying to say McDonald's in Japanese, we thought we'd have a native Japanese speaker, Toshi--Toshi, please come up on stage--and do a real demonstration for you. So, Toshi, thank you very much. So, you're going to do a query. What kind of query are you going to do? >> TOSHI: So, my first query, I want to search for the pictures of Kyoto's Kiyomizu Temple. >> GUNDOTRA: Okay. Let's see you do this query. >> TOSHI: Sure. Okay. Here's the screen. [SPEAKS FOREIGN LANGUAGE]. >> GUNDOTRA: Fantastic. Okay. So, that was a great query. But why don't you try a query, a longer query, a query that you probably wouldn't type? And why don't you tell us what it is before you do it. >> TOSHI: I want to try Voice Search for my favorite restaurant around the Google's Tokyo office. >> GUNDOTRA: Okay. Your favorite restaurant by the Google office in Tokyo. Okay. >> TOSHI: So, I'm going to search for the By Address. [SPEAKS FOREIGN LANGUAGE]. >> GUNDOTRA: Oh, wow. Fantastic. Thank you. Thank you, Toshi. You can see why I've--when I practiced trying to learn that, I gave up. So, thank you, Toshi. You know, as I mentioned at the beginning, we really do get the sense that we are just now beginning to sense the possibilities. In fact, our dreams at Google go way beyond what you just saw--as Marissa mentioned earlier, in addition to voice recognition, Google also has massive investments in translation from one language to another. In fact, we have huge compute infrastructure for translation. Imagine the following scenario. Imagine that you could speak to your phone in one language that Google could recognize what you said, take the text of what you said, feed it to our translation infrastructure, translate it to a different language, and have it then come back to the device using the text-to-speech engine on the device and have it play back in a different language. In other words, it would be almost a real-time interpreter. Would that be great? Let me show you a demo. What we're about to show you is a technological demonstration of a capability that we hope to deliver in 2010. There is a technological preview. Let me try something. Hi. My name is Vic. Can you show me where the nearest hospital is? English up to the cloud, translated back into Spanish. >> [SPEAKS FOREIGN LANGUAGE]. >> GUNDOTRA: Okay. You know what? I apologize, Randy. I didn't have this thing plugged in. Let me try this again, okay? Hi. Can you show me where the nearest hospital is? English up to the cloud, back to Spanish, back down to the device. >> [SPEAKS FOREIGN LANGUAGE]. >> GUNDOTRA: Okay. Now, that's just an early indication. You can imagine us working to get latency down even faster. We also demonstrated English, Mandarin, Japanese. In 2010, you will see us dramatically accelerate our efforts and support many, many more languages. Our goal at Google is nothing less than being able to support all the major languages of the world, okay? So that's Search by Voice. Let's talk about location. You know, we were reviewing some market research from Japan and we found the research quite surprising. What the research showed was that, in Japan, consumers were holding onto their cell phones 24 hours a day and keeping them as close as one meter away. And we thought, "Wow, those Japanese, they must really love their cell phones to have that cell phone with them all the time." Until I went home, fell asleep, rolled over on my bed, looked at my night stand, and right on my night stand was my phone. In fact, we're not that different than the Japanese. You know, I suspect, if you look around now, either you have the phone in your hand, it's in your purse, it's in your bag, you know, the phone, because its location is likely your location has become the most intimate and the most personal of all personal computers. At Google, in our engineering teams, we want to make location a first-class component of all the products we built. Why? Well, it speaks to the personalization point that Marissa talked about earlier. In addition to personalizing the results we bring back to you, we can also just deliver them faster. For example, we included My Location in Google Mobile Maps. Maybe some of you use Google Mobile Maps. Our data shows that, on average, we save about nine seconds per user because when you open up Google Mobile Maps, we know your location and render that location immediately. But we want to do so much more than that. In fact, let me show you some of the things that we're doing to use location in other properties. How many of you use Google Suggest? Okay. Lots of hands go up. Our data shows that over 40% of mobile queries that come into Google are queries that result from a user selecting a suggest item. Now, how could that be improved if we use location? Let me show you. I have two separate iPhones here that I'm going to do this demonstration on. Let me just get them started up. And in both cases, I'm going to show you a future version of the Google.com homepage that supports this suggest. Okay. So that's one. And that's two. Let me just make sure you can both see them. Now, we've done something here, we've hard coded, we've hard coded this phone to believe, as you can see that it's in Boston. And we've hard coded this phone to believe it's in San Francisco, okay? So, let's start doing a query. How about RE, what would you expect? Well, there comes the suggest options, Red Sox, Red Sox schedule, which makes total sense in Boston. Let's do exactly that same query in San Francisco. And what do you think will show up? RE, REI, one of the most popular retailers in San Francisco. Isn't that great? Customize suggest based on location. Of course, there's other things we can do, for example, product search. This is the time people do lots and lots of shopping. But maybe you're like me. Maybe, in addition to just shopping online, there are times that you wish you could actually know if that product happened to be available locally or could we at Google combine your location with inventory feeds from retailers and tell you if that product was available locally? That's exactly what a future version of Google product search will do. Let me do a search here for a product. How about something like a Canon Camera? How about a Canon EOS Camera? I'll do a search here. Let's wait for those results to come back. Now, you'll know that the top two results have something very special. You see that blue dot? That blue dot says those two products are in stock nearby. And if you select in stock nearby in combination with our partners who are sharing inventory data with us, you can see that Best Buy has this camera 1.3 miles away and it's available at Sears which is about 7 miles away, isn't that great? All right, let's go back, let's go back and talk about the Google.com homepage again. Have you ever asked yourself, what's nearby now? Maybe you get to an event a little early. You have a few minutes to spare; you're in a new place. And you go, I wonder what's nearby? That simple query is very hard to answer. You know, think about how you might solve it today. Maybe, you go into your car. Maybe, you'll use your car's navigation system and go through the rigid categories they provide. Maybe you'll type in an address onto a map; but query is so difficult because your location is the query. And what we're going to do is make location a first-class object right within the Google.com homepage. In fact, let me show you a brand new feature called "Near Me Now" on Google.com mobile. Right there, "Near Me Now" and I simply select "Near Me Now" and look at that, it knows I'm at the Computer History Museum and shows me other categories of areas or interests that I might have to go search for. But we can even do better than that; what we really want to do is take your location, send it up to the Google cloud, have the Google cloud reverse GO code your lat-long location, understand the business or place of interest that you're at, look around you for all other relevant places, rank them appropriately at--for their popularity and then send those back to you, back down to your phone in a fraction of a second. That would tell me what's nearby. But watch what will happen; I'm going to click that down arrow right by explore right here and boom. There you go, all the places right around the Computer History Museum. In this case, we've not only answered the question of what's near, nearby now, but we've also, if you look at the ratings, also answered the question of what's good nearby. Now, of course, we realize that you may do the search on more than just the Google.com homepage. And I'm very happy to announce that, today, we have a new version available today in the Android marketplace, a Google Mobile Maps for Android. And among the many features that are in this new product is also this--is this What's Nearby feature. You simply go wherever that you want to look at, any place on the map; it doesn't have to be any particular place. You just long press. So I'll just long press, select, and then, say what's nearby, and exactly that same feature is available today on Google Mobile Maps for Android. Now, let me switch and talk about search by site. You know, of the human senses, sight or vision is considered the most complex. It's estimated that up to two-thirds of our brain, of the human cortex, is involved in the processing of visual images. And it's widely believed by scientists that if we could figure out how the human brain processes images, that would be a monumental step forward in understanding in how the brain actually works. Well, I'm very happy to announce a new product available in Google Labs today, a product called Google Labs, I'm sorry, product called Google Goggles which represents our earliest efforts in the field of computer vision. Google Goggles allows you simply to take a picture of an item and use that picture, that picture of whatever you take as the query. Now, before I demonstrate the product, let me tell you a story on how I use the product. Obviously, we "Googlers" test product before we make them available. And I was doing my job testing the product. I had a friend, a couple, call me. They were scheduled to come over for dinner, but they called and said they were running late. They were stuck in traffic. In fact, they were stuck for traffic for one hour. So, when the doorbell rang and I opened the door, both the husband and wife had to use the bathroom desperately. And as I opened the door, they said, "Sorry Vic, we've been stuck in traffic, we need to go use the restroom." I very happily pointed to where the restroom was and the wife handed me the gift that they have brought, a bottle of wine. And she said, "While we're in the bathroom, open this bottle of wine." Well, I did what you would probably do. I pulled out Google Goggles and took a picture of the bottle of wine. Maybe, you've ever wondered, is this bottle of wine any good? Oh, you know, I got a result sent back; it said that the wine has hints of apricot and hibiscus blossom, just then the door opened and my friend came out of the restroom. Of course, she's not a "Googler," so I took the confidential product we were testing and I put it away. She said, "Please, please pour the wine," and so, I poured the wine. And she said, "It's my favorite wine, you know, what do you think of it?" And I tasted it and I said, "It's got hints of apricot and hibiscus blossom." She was blown away. They thought I had a great wine palate. I'll be honest with you, I don't even know what hibiscus blossom is. You know, let me show you a demonstration of that. I happen to have that bottle of wine, not that exact one, but the same bottle. Let's try it, let's take a picture. Let's launch Google Goggles and let's see what happens. Now, the lighting conditions are less than optimal. I'm getting a huge amount of glare, so I don't know how this is going to work, but we will try. So, let's see here, you--let me take an image, a picture of this bottle and then we'll bring it over here. Okay, oops, well, it helps if I actually get an image, so sorry, I moved that. Let's try that again. One more time, and I will try to hold my hand steady; surprising, holding my hand steady wasn't a problem during rehearsal. Okay, and then you see, it's scanning, attempting to recognize what it sees here and, in this case, it has gotten some of the bristling items. And, of course, as you use it and as we get more and more data, it'll become more and more accurate; pretty exciting work. Now, you may think, "Well, Vic, I'm not going to take a picture of a wine bottle, if that's all the Google Goggles does." Well, it does a lot more. It recognizes things like CD covers, movie posters, books, barcodes, lots and lots of categories. You know, for example, imagine that you're traveling in Japan and you come across this landmark. Now, you don't speak Japanese, but you do know that's a famous landmark. How would you go about and getting any information about that? Well, using Google Goggles, you could just take a picture of it, and ask Google to help you. Let's try that. Let me come over here and let me try to take a picture of that landmark. I'll pretend I was there. Okay, I got a picture. Let's come back over here and show you the results. It's analyzing and there we go. How about that? It accurately recognizes that landmark. Now, once again, it's incredibly impressive when you understand what's going on. In this case, those images are being screened to the Google or send to the Google Service, the Google clouds. There, our vision algorithms are analyzing the image and looking for objects that it can detect. Those objects have signatures that are created and then those signatures are matched up against an index that has over a billion images in it. The best matches are ranked and then sent back down to your device all in a fraction of a second, showing you really the power of devices that are connected to the cloud. Now, some of you may ask, "Why is this in labs? Why is this product in Google Labs and available today?" Well, we put it in Labs for two reasons; one, because of the nascent nature of computer vision. We really are just at the beginning here, and the technology is just getting under way. The second reason is because of the scope of our ambitions. Google Goggles today works very well on certain types of objects in certain categories. But it is our goal to be able to visually identify any image over time. Today, you have to frame a picture and snap a photo; but in the future, you will simply be able to point to it, as simple, as easy as pointing your finger at an object. And we'll be able to treat it like a mouse pointer for the real world. Of course, we're a long way from that today. But today marks the beginning of that visual search journey. And we strongly encourage those of you with Android phones to download that product, give us feedback and help us really grow. So let me wrap up here. You know, we really are at the beginning of the beginning. You know, if you think about the mobile announcements that we talked about today, everything from Japanese voice search to a new version of Google Mobile Maps that allows you to search nearby or the upcoming changes to the Google.com homepage or even something like Google Goggles, all of these are powerful demonstrations of what happens when you take a sensor-rich device and you connect it to the cloud. Yes, it could be that we are really at the cusp of an entire new computing era, an era where devices will help us explore the world around us that devices can understand our own speech or help us understand others, devices that may even augment our own sense of sight by helping us see further. We hope you're as excited as we are about these mobile announcements. While it is just the beginning, the possibilities ahead inspire us. So, thank you. Marissa, please. >> MAYER: So, you can see, search engines that understand where you are in the world, search engines that understand you when you talk to them, even search engines with eyes. These are the things that are going to change the interface for search fundamentally, as we move forward. The other thing that will fundamentally change the interface for search is media. Remember, media is what we refer to in terms of the richness of the Web. And the way that richness needs to be reflected on the results page. And it's not obvious, but media is really fundamentally a relevance problem. Can you find the right information? Can you rank it the right way? And if you look at the way media has evolved inside Google Search over the past 11 years, it's pretty interesting. We started with just Web sites, URLs, a list of 10 new URLs that you needed to be ranked. Then we evolved towards universal search bringing books, news, video, local information, images, products, blogs, all onto the results page. But there again, there's all kinds of interesting relevance questions: When should those genres appear at all? Where should they appear? Which one from those genres? Which item from those genres should be surfaced on the results page? It's a huge relevance challenge. And then think about the Web today; the Web is hugely participatory. It's hugely fresh. I heard this morning that in Tahoe, people woke up to two feet of snow. And I've heard that some of you were caught on it on your way home; I'm jealous. But what's interesting is was it two feet, was it one foot, was the snow blowing or not? And there are some official sources for that, but they don't always get it right. Yet, there are user updates out there that do get it right. But how do you find them? And how you cut through all of the updates that aren't relevant? That's why we have one of our foremost relevancy experts here in the company to talk about some of the challenges with media and relevancy. Please help me welcome Amit Singhal, Google fellow. >> SINGHAL: Thank you very much, Marissa. So, we are here at this wonderful Computer History Museum today. And before I get to today's big announcement, it's just fitting that we take a moment and talk about the history of information flow. Now, I've worked in the field of search for almost 20 years now. And what we are going to announce today is one of the most exciting things I have seen in my career. But let me first take a moment and talk about how information has flown over century. So thousands of years back, people got most of their information by word of mouth. Sitting around campfires, from their tribe, from their village, kids would walk up to the village elders who have all the knowledge and say, "Hey, grandpa, should I eat those beans?" And the grandpa would say, "No, no, no, them are poisonous, okay?" And the kids who listen eventually became grandpas and pass that knowledge along generations. And it took generations for knowledge to get from one point to another point geographically. And that was clearly not fast enough. Then Guttenberg invented the movable-type printing press and this process of information dissemination was parallelized. An author could write a book, thousands of books can be--books can be printed. And then, they were sent on horsebacks and by boats to around the world. Village elders around the world now had the power of that knowledge. And believe it not, some--up to some 20, 30 years back, that was the primary mode of information transfer. And even though great strides were made in the physical limits--in the printing technology and transportation technology, the physical limits of printing technology and transportation technology still made it so that for information to get from an author to their--to the consumers took weeks, months and sometimes even years and that was clearly not fast enough. And then came the Internet. And what I'm showing you here is one of the early Google servers that are displayed here at the museum. And, now, suddenly, the world changed because billions and billions of documents were available to millions and millions of users just through the Internet, through search engines. And that was a great revolution. In the early days of Google when I got here nine years back, we used to call that information every month, and we would put up a new index every month. People will call it Google Dance. And, clearly, a month was not fast enough. And then, the Web world, and then we started calling every few days then every day, and then, every few hours to now when we actually can call every few minutes. But, clearly, in today's world, that's not fast enough. In today's world, the world is producing information from around the globe every second by tweeting, by posting other updates, by creating Web pages, by writing blogs; you name it. Information is being created at a pace I have never seen before. And in this information environment, seconds matter. Let me just use a metaphor to explain my point, the old metaphor of a library. Imagine a library with billions and billions of books. The librarian, just when the librarian understood the billions and billions of books in his library, the librarian realizes that there are a hundred million new books coming in every day. Now, the librarian finally figures out how to deal with a hundred million new books arriving at his library every day so that he can tell the patrons what they should look for. And just when he mastered that process, the librarian realizes that there are a hundred million of people running around in his library, adding pages, deleting pages, writing notes, adding things to books and guess what? If they didn't find the book on what they were looking for, they wrote 140-character note and handed it to the librarian saying, "Not in your library." And that's what, that's the information environment today. And imagine in this library, a patron walks up to the librarian and says, "Hey, I need to learn about President Obama." And the librarian says, "You need that book, that book, that article, that image, that speech, that video. And, oh, by the way, 17 people just left me a note that President Obama has been talking to the Democrats in a special session on Sunday about health care and, by the way, he'll probably be in Oslo on Thursday receiving Nobel Prize," and so on and so forth. Now, imagine that librarian does all that in under half a second without breaking a sweat. As of today, that's what Google can do. We are here today to announce Google Real Time Search. Google Real Time Search is Google's relevance technology meeting the real-time Web. Now, I can't emphasize this enough--Marissa has said this right now--relevance is the foundation of this product. It's relevance, relevance, relevance. There's so much information being generated out there that getting to you relevant information is the key to success of a product like this and that's where we, as Google, come in because for 11 years, that's what we have done. Rather than talking about this more, let me just show you, I have with me Dillon Casey, the product manager for this product and Dillon, what's happening out there? >> CASEY: Well, as you mentioned, we weren't the only one's working yesterday, so one of the great things about this new product is we can actually see what people are talking about in regards to Obama right now. >> SINGHAL: So, Dillon types Obama into Google--Google Research Page comes in and, wow, look at that, results just came in Real-Time, this page has come to life. Do you see that? This is results coming into Google's Results Page. As they are produced on the Real-Time Web out there, okay, our users will get their results on the results page as they are produced in Real-Time out there. This is the first time ever any search engine has integrated the Real Time Web into the results page. So, let us show you a little more of this, Dillon, why don't we click on the latest result links out there. When Dillon clicks on the latest results link out there, he's taken to a full page of Real-Time Results as they are produced on the Web out there coming into this page, you see there's a twit that just came in. Here's another Real-Time page that we called from anwerstatyahoo.com seconds ago, another twit that just came in seconds ago and--hey, Matt--man--what are you twitting out there? Our good friend Matt just twitted and, guess what, it just showed up in Real-Time as he twitted. This is the power of Real-Time Search. Now, let me just take three examples to demonstrate what you have seen out here. So we have been testing this product internally with our Googlers and as we have been testing this product, I have received good examples from my friends about how they are experiencing this product. One time, one Googler had heard that GM's car sales were finally stabilizing, so she typed GM into Google and, of course, our results were right there--the biggest news of the day was indeed that GM's car sales have stabilized, however, she noticed this latest results for GM section while she was searching and read that GM CEO Fritz Henderson had just stepped down seconds ago, now this was the information she needed right then and this is the power of Real-Time Search. Now, on the results page, as I said, this is the first time we are presenting Real-Time Web on the results page--what you see in this Real-Time section is a scrollbar to the right, so that if a Real-Time result just scrolled past you, you can go back, go forward, in addition, what you see here is--seconds ago, we had called an article from Business Insider and that was presented to the Googler and there was also an update from Twitter.com, a twit from Twitter.com talking about Mr. Henderson stepping down. Now, what you observe here is this is the whole Real-Time Web; this is comprehensive Real-Time Web, with twits, with news articles, with blogs and so on and so forth. Let me take another example. One of the Googlers was searching for Bank of America, and, of course, our results were very relevant but the key part there was that the Real Time Web was buzzing about how Bank of America had decided to re-pay its TARP loan and all those results started showing up in this latest results section that I've been talking about; but if you click on the latest result for Bank of America, as Dillon did for the query Obama to take you to the second page, you are taken to this full page of Real-Time results and what you notice here is this special new link that we are launching today under our search options. We are very happy to launch this new search option called the latest results, which are available to you by clicking through with the latest result section or opening the search options on the top left of the result page and then clicking at this new feature. And once you click on this new feature, you are given information from the Real-Time Web, the comprehensive Real-Time Web from Twitter, from Wall Street Journal, from a Website, the Loop21.com, and so on and so forth--comprehensive Real-Time Web results coming to Google's results page in Real-Time. Now, let me take a third important example, one of the Googlers was visiting home for Thanksgiving break in Maryland and wanted to get H1N1 vaccine and heard--had heard that his whole high school was administering H1N1 vaccine, so he typed H1N1 vaccine Arundel, the high school's name, and was taken to the result page, which were very relevant results saying "Free vaccine to be distributed at schools." So, very fresh results around Thanksgiving telling the Googler that vaccines will be free and available at this high school but the Googler already knew that. He wanted to know how long the lines were, what else is happening. So the Googler clicked on this show options link that I just talked about and, having clicked on that, got the new options panel and on this options panel, we are very happy to announce today that we are adding a new update's link and by click on--by clicking on this--this updates link, you will get all the twits and other updates coming into Google system in Real-Time. In this particular case, when the Googler clicked at this update's link, the Googler got one very, very relevant twit saying that the high school had run out of vaccines--we saved him a trip and he was totally impressed. Now, for such hyper-local queries, we are--maybe one person is twitting on their cell phone or very few people are saying something, Real-Time Search becomes incredibly powerful because it shows you exactly what you need in your geography when you need it. This was one single twit and it became available to that Googler right on Google's page. So why I don't show you this page live. Hey, Dillon, what's happening out there man? >> CASEY: Well, actually I hate to admit, I've been up here kind of surfing around while you were talking but, you know, I don't know about you but I'm really excited about Google Goggles and I've been watching what people are saying about it, it's--it's so cool. So I just put the query in and I hit search and there's people talking about it right now, in fact... >> SINGHAL: Wow, look at that. Vic, dude, when did you announce it? How many minutes back? And here we are on Google's Result Page Real-Time Web brought to our users for something you heard from Vic right now. Now, that's incredibly exciting and, as you can see, we can go into all the full Real-Time page and you can see the entire Real-Time Web for Google Goggles being brought to our users right at Google's results page and--and the page that we have been talking about after that. Hey, Dillon--hey, man, what are you doing? We are in the middle of the most important launch of the year and you're playing with your cell phone, man? Grow up. >> CASEY: Right, sorry. >> SINGHAL: What's happening? >> CASEY: Sorry, Amit, a little guilty pleasure. You know, I've been kind of following the Tiger story and it turns out this also works on mobile. I'm getting updates from the Apps. >> SINGHAL: What? It does? Then why don't you share it with all of us? >> CASEY: Okay, forgive me but you can see right here. >> SINGHAL: Wow, look at that. This is Google's Real Time Search on mobile phones. So we are very happy to announce today that Google's Real-Time Search is available on iPhone and Android and anywhere you are--you need your information now, just pull out your Smartphone, type a query into Google.com and you will get the Real-Time Web in your palm right away. And that's the power of mobile Real-Time Search. At this point, I am also happy to announce that our Google Trends page is coming out of labs--it's leaving labs, it's graduating from labs. We are very happy. And we have added to it this new hot topic section that Dillon is showing you right now. On the hot topic section, you will see what the Real-Time Web is generating right now, what information is coming into from the Real-Time Web into Google systems and, by clicking on one of those queries, you will of course see Google's Real-Time Search Results. In addition, we have added a window down there where you can type your query and see Real-Time Search Results. Now, before we go there, let me just say one thing, we are rolling the Google Real-Time Search product over the next couple of days and over the next couple of days, some users may not have access to this product as we roll the binary out, however, you can always go to this new google.com/trends page and by clicking on one of the hot topics, you will get to see Google's Real-Time Results or you can type your own query into more hot topics. So Bernanke's speech is what's happening? >> CASEY: Yeah, yeah. While you were giving your speech, Bernanke just presented a huge, gigantic speech and the stock market went up. >> SINGHAL: Wow, it happened right now? >> CASEY: Yeah. >> SINGHAL: Man, I should call my broker. What's happening to my money now? Anyone knows? My broker is not here. So what you observed here is this new google.com/trends page which will take you to our new feature right away. Okay, so, we are all technologists here. I have been working in the field of search for almost 20 years. We all love technology, you love technology and I would be cheating you if I didn't tell you what went into building a product like this. Let me just tell you, we literally had to develop dozens of new technologies to make Google Real-Time Search as relevant as it is; technologies like language models, we had to model whether a certain twit was genuine information-carrying twit or was it just a weather buoy sitting out there twitting automatic twits. We had to develop query fluctuation models. If queries fluctuate, the volumes fluctuate at a certain rate, then something becomes eligible for you to see on your results page. We had to--we had to develop the Real-Time content generation fluctuation model. If there's suddenly a lot of content about Bernanke's speech or the stock market, something just happened. Now, this is--these are some of the most exciting technologies that we have developed to build a product like this. Today, we are processing over a billion documents a day generated by the Real-Time Web out there and within seconds, we have analyzed the documents and twits and updates and we have understood the documents and twits and updates and within seconds we have seen a user query, which may have--which we may have never seen before. We don't--we haven't seen one-third of the queries that we will see today ever before. So you take documents that you've never seen before, you take queries that you have never seen before and you merge them together and filter for relevance and bring it to the users within seconds and that's what Real-Time Search is all about. Now, at Google, we talked about the four pillars of search: comprehensiveness, relevance, user experience and speed. And let me tell you one thing; I've worked in this field for a long time--I've worked at Google for nine years and as the information world has exploded, as the amount of information at the level out there has exploded, we are getting hundreds of millions of new items every hour, the importance of relevance has gone through the roof. Everything is important. Comprehensiveness is important. User experience is important. Speed is important. Indeed, relevance is important but as the amount of information out there has grown, as much as it has much and it's growing at the pace at which it is growing, relevance has become the critical factor in building products like this. Now, let me just recap of what we just talked about. So today we are very proud to announce Real-Time Search, a new latest search option in the Google search options, a new update search option in the Google search options, a mobile version of our Real-Time Search and the new trends page living labs with a new hot topic section that would give you Real-Time Results. And I'm incredibly proud of what we have built. As I study this--information is now getting to you instead of--from in generations, instead of in years, instead of in months, instead of in days, instead of in hours, in minutes, it's getting to you within seconds and I'm incredibly proud of what we have built but at Google, we are never satisfied. It takes about one-tenth of a second for lights to go around the world and, at Google, we will not be satisfied until that is the only barrier between you and your information. Thank you. Let me hand it back to Marissa. Thank you, Dillon. >> MAYER: Thanks. So the first time that updates have been integrated into the search results and we have actually the most comprehensive set of updates, too. We're so excited about this product. With that said, we didn't want to rest on those laurels. We actually have two new exciting partner announcements. The first of those announcements is with Facebook. Facebook will be providing us with the feed of updates from their public profile pages, also known as Facebook Pages and these will be appearing in Google's Real-Time Search. The second new partner, we have to announce today is MySpace. MySpace will be providing us with feed of updates from all of their users on any updates that are public and these updates will also be appearing in Google's Real-Time Search. We have support from our partners here today. I want to thank Biz Stone and the--and the team from Twitter for being here today and also Jason Hirschhorn from MySpace, the chief product officer from MySpace and his team. Thank you very much for coming and supporting the launch today. Last year on--in conclusion, last year on our tenth anniversary, we published a blog on the future of search and that laid out the vision for modes, and media, and language, and personalization. And I think when you look at today's announcement--search engines that have eyes, search engines that can understand you when you talk, search engines that know where you are and search engines that know what is happening anywhere in the world and can bring it to you in Real-Time. It's amazing to see how far we've come on realizing that vision in just one short year. And with that, I'd like to welcome Amit and Vic back to the stage and we'll take some questions. Thanks. >> BENNET: And we're also going to take some questions here in a second from online as well while we set up. And if folks have questions here, feel free to step up. We have microphones just right here. But, we're going to start with an online question first that's coming from The Guardian in UK. This is for you Vic. We got to put you on the hot seat first. It says, "Given Google's acquisition of Neven, to what extent can Goggles recognize faces?" >> GUNDOTRA: It's a great question. Hartmut Neven is in the room, I believe. Hartmut, are you here? There he is. Hartmut Neven had a company that we acquired and he is the leader of this particular project concurrently at Google. His previous company did some pretty amazing work around voice--around face recognition. And the technology that we built with Google Goggles is very general. Of the billions of images in the index that we do recognize, "faces" is one of those objects. However, for this particular product, we made the deliver product decision not to do facial recognition. At Google, we deeply respect the user's privacy, and we still want to work through issues around user opt-in and control. And so while we have the underlying computer science technology to do facial recognition, we decided to delay that until we have more safeguards in place. >> BENNET: That's great. And again, if folks have questions here, just go forward, put up your hand. And if you could just introduce yourself, so that people who are listening in online can know where the question is coming from. >> TENAKA: Hi, this is Akito Tanaka from the NIKKEI. I had a question regarding the Real-Time Search. What is the advertisement opportunity in that area? >> SINGHAL: So, right now, we are concentrating on bringing the most value to our users with all the wonderful partnerships that Marissa just announced, and our partnership with Twitter. And I believe that this phase is very, very young. As time goes on, new models would develop, and all the companies that we are talking about are experimenting with multiple models of how to generate revenue from all this wonderful real-time information that the world is producing. I think all the companies like Twitter and others have added tremendous value to the world, because we can figure out what these key conditions are in Tahoe right now. And I can figure out how the traffic is like in Bangalore right now. And you name it, right? There's so much information out there. And as long as the product brings value to users, I think new models are recognizing and various of the revenue streams emerge. And you will see a revolution in that space over the next few years. >> WATERS: Rich Waters in the Financial Times. Can you tell us more about in Real-Time Search how many sources you're crawling? How often you're crawling? Are you taking feeds from Twitter and Facebook, and other places? And longer term, you know, how much real-time information you can be able to run? >> SINGHAL: That's a great question. So, we are crawling a lot of content. As I said, right about a billion pages a day. We are crawling everyday. We are crawling from many, many sources. We are going out to the web, and we are crawling all those sources, all the good sources out there. Definitely, all the new sources, but not just the new sources. If a company announces a new product into the press release, yes we will get it. If a blogger writes a blog about something, yes we will get it. And we will do that within seconds. So, we are crawling, we are casting a very wide net. The key here is comprehensiveness of real-time information, and integration with Google's search results. Those are the two keys--those are the two key design principles behind this product. And indeed, we are taking feeds from our partners, Twitter, and going forward very soon from MySpace and Facebook. And we would like to get as much information as there is out there via feeds or any other mechanism, because our objective is to build the most comprehensive Real-Time Search out there. >> MAYER: And I should also add that on our latest mode, we actually do have other update providers, including FriendFeed, Jaiku, Identi.ca, and TWiT Army. And as Amit said, we'll be working to bring Facebook and MySpace into that functionality over the next few weeks. >> BENNET: Well, we have further questions from here. Oh yeah, go ahead. >> HELEN: Hi, Helen Malbay [PH] with the Financial Times Germany. What about availability of Real-Time Search on your non-US site? >> SINGHAL: That's another great question. So, we at Google strongly believe in all of our products becoming international rapidly. This first launch is available in all English speaking locales. That would be the UK, Canada, India, Australia, and New Zealand, and so on and so forth. And very soon, some time in Q1, we are planning on launching many new languages. So, this is one of our top priorities in this project. Our first priority was launching it, stabilizing it, making of our infrastructure and relevance work, because that, as you can imagine, has been a hard challenge. And as I showed you with all the new technologies that we have had to develop, that was our first focus: building a product that would bring value to the users. And going forward very soon, we are going to rapidly internationalize this product. >> BENNET: So, we're going to take one more online question here. This is another one for you, Amit, about relevance. How do you prevent spammers from taking advantage of the Real-Time search results? This is from Steven Bivins [PH]. >> SINGHAL: This is, you know, this is something that I know something about having run Google search for about nine years. We have the best systems in place to prevent gaming of the system, okay? Our spam lead out here sitting with us, Matt Cutts. And Matt runs the best spam prevention team that there is out there. And we have had experience with this for so long, we have developed algorithms inside that can see what's going to happen next and counteract almost before it happens. Matt's team has developed some of those algorithms. And real-time is moving--for us, real-time is moving from minutes to seconds, and we are already in the game of running the system that's minutes fresh. And we do a great job. You find Google results very useful that we called a few minutes back. And Matt and his team and the rest of the team at Google are experts in this area. We know many things about it, and that's how our real-time product is already so relevant. >> CURTIS: David Curtis. A quick question on the customization, are you going to be able to integrate social memory counts so I can customize the prioritization of Real-Time search results by the people who I care about? So in other words, using Facebook connect or LinkedIn, or Twitter a lot to sort of prioritize or even let me selectively choose, like, the real-time results that I wanted to deal with. >> SINGHAL: You are just picking questions out of my mind. That's so wonderful. We are very excited about what's happening. And the key thing that we are excited about is we are just getting started. You mix with this real-time search all the key principles that Marissa talked about, and Vic talked about. Localization, personalization, and now you have a real-time product that everyone would like to use it. In addition, recently you have noticed we have been experimenting with social search in our labs. Marissa launched Social Search a few weeks back. And that is an angle of personalization where you see results from your social circle alongside the most relevant results from Google. Now, you can just imagine when you merge these multiple technologies that we have developed here from Real-Time Search to social search to job location-based search, what you will get is a product that you would love, and that's coming very soon to a Google near you. >> BENNET: See, Glenn up here in the front? >> GLENN: I'm really curious [indistinct]. Could you further expand on the distinction between Facebook's going to give you the public feeds, which the stuff that members are already designated, is okay to be seen by anyone, and then MySpace is giving you all the feeds, I mean instead of... >> MEYER: Right. >> GLENN: So, that's anything that a MySpace member will get, please. >> MAYER: So, Facebook has a product called Facebook pages, which are special public profiles for specific entities, their feed that they're providing us will have the updates that come from those pages. For MySpace, it covers all of their user's profile pages and any update that's designated as public. So if I were a user on MySpace and I saw that my updates could be public, those will all be coming to us in the feed. >> GLENN: [INDISTINCT] >> MEYER: That's right. So that users can decide what they'd like to see offered by this feed to Google and then search and how broad that they really want to share it on the social network, in general, by using the privacy controls available on each of those networks. >> MCCRACKEN: Harry McCracken with Technologizer. You talked about how you have these partnerships with a bunch of major sources of real-time content. You need those relationships and those feeds to do this or what would happen if there was something that people are excited about that you did not have a relationship with? >> MAYER: I think that overall our goal is always comprehensiveness. But our mission is the world's information making it universally accessible and useful. And we really do mean the world's information. As Amit can attest to, you get better results when you have more items to choose from, when you can analyze them, understand how they relate to each other. So the more comprehensive we can be, the better we can serve our users. And that's why we've had such a focus on making our real-time results that we're launching today already the most comprehensive available. And we're taking that even further with the MySpace and Facebook partnerships. In the future, if there was something else, obviously we'd want to partner with and include those sources as well. >> BENNET: We have a question from Danny up in the front here. >> DANNY: Can you go back and clarify with Twitter what financial deals are there, if there are any, and then the same for Facebook and MySpace. Are there ad deals that you're paying for this stuff or is it just the goodness of their hearts or what? >> MAYER: We cannot disclose the financial details of any of the deals. >> DANNY: [INDISTINCT] >> MAYER: I can't. I'm sorry. >> DANNY: [INDISTINCT]. If we go back to MySpace, we've got Murdoch saying that he wants you guys to pay him to carry his news results, right? But then, with MySpace, either they've just decided to give you this stuff because they think it make sense and that's his company, and apparently that information is free for them to hand out, or you're actually paying for it. So, it seems reasonable to ask whether or not somebody is getting something out of this financially. Even if you can't do the details, it's either they're doing it for free or they're not doing it for free. Can't you give us that? >> MAYER: Yeah, I'm sorry, we can't confirm. >> SINGHAL: We don't exclude any source, any source of real-time information we would really like to have integrate it with our system. And we let our algorithms for relevance decide what updates or tweets, all blogs, all news, all web pages to serve this to the user. So at Google, we are all about comprehensiveness and we will accept all sources of real-time information and we'll build the most comprehensive real-time search. >> But are those--I mean, are you applying the same relevance algorithms? You write the algorithms, so you know what they can do. And are you applying the same algorithms you applied to normal web search just on a faster--on a faster footing or how have you have to change it? >> SINGHAL: So as I was talking about earlier, we have had to develop at least a dozen new technologies to make Real-Time Search work as well as it does, because clearly in Real-Time Search, you need to have models of information fluctuation. As information fluctuates out there in the real-time world, all as information fluctuates in our query screen you have to react to that. And clearly, those fluctuations are useful for traditional web search. But for real-time search, they are just incredibly critical. So, I wouldn't say it's exactly the same algorithm that we can use, because to work with this amount and this base of information generation, you have to develop these new technologies. And what we have developed, some of these are just amazing technologies, right? I've work in this field so many years and I didn't think we would develop these technologies so fast at the rate that we did. So, I'm incredibly proud of how much relevance we have brought to the product based on the technologies we have developed using our experience with relevance. >> BENNET: So we're going to take one more question from online for Vic. The question comes from Luke Wilson with more--I don't know if it's the Luke Wilson, but from a Luke Wilson--with more Android phones and possibly even Google Android hardware, does Google intend to reduce support for non-Android devices? >> GUNDOTRA: Absolutely not. You know, our desire is to reach our customers on whatever platform they're on. And today, there are a variety of smart phone platforms like the iPhone in Apple is a strong partner of ours, like a BlackBerry, Nokia. At times, we choose different priorities in terms of which one we do first. But it is our goal to reach as many as possible based on the technical capabilities of that underlying platform. >> BARAK: Hi, Sylvie Barak from Hexus. I wanted to know, first of all, do you feel that your Real-Time Search will be the death of journalism? And second of all, everyone knows that knowledge is power. Does that make Google the most powerful company in the world? >> SINGHAL: So, let me answer it to the best of my ability. Your questions are clearly very loaded. So, journalism has its role and it always will have that role. Information is indeed power, and what you in this room are doing, are empowering the world, as we speak, with information about what's happening here right now. So, I can't even think about putting those two words together, death of journalism based on Real-Time Search, because you bring so much value to the world that this value has to be brought to the world. And regarding your second question, our goal at Google has always been to bring timely information to our users. And clearly we are empowering our users with the information that they need now. We have been in this business for 11 years. We get the information, we do our special relevance work, and we bring it to our user. So, I think it's all about user empowerment. I personally have felt empowered many times when I had the knowledge in my hand through Google and I walk into situations that I had to. So it's all about user empowerment. And Real-Time Search is the next step in that direction. >> MAYER: Yeah, I will just add that, you know, our purpose is really around facilitation and reference. Getting people to do their search and getting them off of our site and on to where the information exists in its native form as quickly as possible. And so from that, like I said, you take a little bit of exception with the question of we have the information, we don't. The web has information and we want to get the user's back out to the web as efficiently as possible. >> BENNET: It's probably worth knowing too that Google sends billions of clicks each month to news, publishers, and this will be yet another channel through which to send those clicks. Yeah. >> HELLER: Hi, Michael Heller. I'm with the Google Technology User Group. I just--my question is around the integration of real-time results with the rest of Google results. I saw, the latest within the web page. And I'm just wondering what your vision is for long term. From a user perspective, how much does the user thinking about is something real time result versus some other kind of result and where you're trying to go in terms of that direction? >> SINGHAL: Another great question. The power of our universal search is that users don't have to think about whether they should be searching here or there. They should be searching in Google search box, and all information that is relevant to them at that moment should surface on the search results. We have made great strives in universal search with integrating books, videos, images, news, blogs, and so on and so forth into Google Search. We don't think of it as this search or that search. We think of it as Google Search. Whatever needs to be seen by the user now should be integrated on the Google's results page. And Google Real-Time Search is just the newest feature of Google Universal Search because this is the genre that's very relevant in today's web and we have just brought it right to our users with our integration with Google Search. >> MAYER: Yeah, I would just add one example there, which I think drives on how much of a relevance problem this really is. So for example, think about when you're searching for a product. So, apparently, a few weeks ago there was a massive stroller recall, and what's nice is when we're actually dog-fooding this, this what we call to using it internally, when we're dog-fooding Real-Time Search, we did search for that stroller and only did you get places where you could go and buy it, but you also were alerted to the fact that there are been a recall, which I would argue for users. That is a very important piece of information. If you're about to go and make that purchase, you want to know that there's been a major news issue with this or potentially a major safety issue with it. And I think that that shows the power of the Real-Time Search. And in that case, that Tweet, that news article is incredibly relevant, and that's why we surface it on the same page as our search results. >> Marissa, you just mentioned that it's still Google's idea to get users as fast as they can to other pages on the web. We've seen some changes though from Microsoft and Yahoo that's in to think that creating pages from information around the web is a better way to go and gives people answers. You also mentioned answers. Can you give a sense of where Google could have lay ends on that kind of creating user interface or sort of still having algorithm being the king? >> MAYER: I think that our view is that we overall believe that the web thrives on openness. And so, the reason that we have this amazingly rich set of data to search and provide on our search results is because the web is open and there's like a huge amount of participation to say, "Oh, you now, we'll develop the most authoritative page on this." I think it's problematic, because it does make it a much more enclosed system and we want to be much more inclusive. That said certainly as we evolve search, there comes a point when you do want to be referring to things potentially more as entities. Now here's a restaurant. What can you tell me about it? Right? And we do that by offering heterogeneous search results. So this will really allow you to see not only the canonical page for that restaurant, but also reviews, et cetera. I think we'll play with the user interface with that, but again, the point is still that the best information, the richest information is out there on the web. We want to get people there faster, but we don't want to hold them on what we would call the most authoritative page or host that page. >> SHANKLAND: Yeah, this Stephen Shankland from CNET News. I appreciate your focus on relevance and recency. But in my observation, a lot of times those things-—they don't necessarily go together. Do you have any way of putting truth into the equation? There's a lot time where--a lot of situations where time goes by and the truth seeps out. So, is there a way that you can actually assure people that they're not getting connected to rumors and things that are potentially factually wrong very quickly? >> SINGHAL: No, this is a very good question and a very, very tough scientific problem that the research community is also thinking about and we are also thinking about. And right now, a straightforward answer to your question is we emphasize quality and relevance. And that often brings the truth out. I say often because there are maybe occasional chances when the truth is somewhat grey and not black and white. In which case, it can be debated. But it's a very hard problem because language understanding is still an unsolved problem. We have made great strides in language understanding at Google. However, it's still an unsolved problem. We are very excited about the algorithms we are developing to understand language, but what you are talking about is, in some sense, the grand challenge for language understanding. So, we are excited about the strides we have made but that's our ultimate objective years down the road to get to that point. And that's why after having worked in the field for 20 years, I come into work every morning like a kid going to candy store, because I get to work on all these things. >> MOORE: Patrick Moore. >> GARY: I'm sorry. I'm speaking. This is Gary. I'm with SearchWise [PH]. Actually we are hosting some Real-Time information on our Website. So, is there any way that we can submit the channels to the Google Real-Time Search, you know, where the resource? >> SINGHAL: Sure. Please, talk to Dillon, who's sitting here after the event. >> GUNDOTRA: Yup. >> GARY: That's also a general question for a lot of Real-Time hosting Websites. >> SINGHAL: Right. >> MAYER: I think it make sense at some point to have a standard API and a standard feed that we accept and I think we will be moving in that direction as we evolve the product. >> GARY: Okay, thanks. >> PATRICK MOORE: Patrick Moore. One of the questions that I've noticed come up on Twitter, yeah, I think it should be repeated is a lot of energy and effort is spent on PageRank. What does this do in the Real-Time Search? It seems like, you know, you got a real issue here, how are people going to deal with this as PageRank impact the Real-Time Search results? >> SINGHAL: So, Page Rank is a very important piece of our ranking relevance technology that we use for our entire Web search. And PageRank is indeed one of the hundreds of factors that we use in our ranking system. And for Real-Time Search, we have all those hundreds of factors. Some may not be as powerful in the context of Real-Time Search, some maybe more powerful. In addition to Page Rank and those numerous other signals, we have had to add these technologies that I talked about. Like, language modeling, right? How do you model language that it actually is finding good relevant information? And so, PageRank is always a very important piece of our technology including Real-Time Search. We have just had to develop many new things to make Real-Time Search as relevant as it is. >> MAYER: And I would add that Page Rank is really about finding authoritative pages. One of the more fascinating things that we've seen or we're beginning to see inside some of the real-time data is authoritativeness exist there as well and there are signals that indicated. So, for example, re-twits and replies and the structure of how the people in that ecosystem relate to each other, you can actually use some of our learnings from PageRank in order to develop a, say, you know, an updates rank and/or an updater rank for the specific people who are posting. So, this is something we're beginning to experiment with. It was interesting to see that same parallel where PageRank looks at links. You can actually look at the very mechanisms inside of these updates streams and searches and in a sense authoritativeness in the same way. >> BENNET: We're going to take one more online question from Eric Wester from wikiHow, who's asking: If there's an option to disable the scrolling feature, he says he realized there's a pause but what about disabling it all together? >> SINGHAL: So, we really have experimented tremendously with this user interface for Real-Time Search. And based on early positive and good feedback like this, we did introduce the "Pause" button. And after a lot of experimentation, I think the current interface is serving its purpose of conveying to you the Real-Time nature of your query and providing the Real-Time results. And all—we are always experimenting with everything that we do, including user experience, not just for Real-Time Search, but for everything else. And we would be working on the user experience for over the entire search system going forward, clearly Real-Time Search is the new feature we are very excited about, and we would do a lot more work in the user experience direction of Real-Time Search going forward. >> MAYER: And I think we also have to acknowledge this very early in the evolution of the feature, and we don't know that this is exactly the right user interface. It may ultimately change and so--and we also want to always to honor our user's preferences. So, if they have expressed a preference to us, we don't want to later say, "Well, hey, here's a whole new way the user interface could work and how do we interpret that option there." A few years ago when we first put PDFs into our search results, we had the PDFs somewhat imbalanced, never showing up too frequently. And lot's of users mailed it and said, "Can we—can I just turn off having PDFs in the search results?" And we also said, "Actually, you know, bear with us. The relevance of PDFs will get much better." And sure enough, it did. And we didn't want to have this sort of mechanism where you turn it off and then we have to re-introduce it later once it was better. We actually think that the Real-Time Search relevance is already very good, but that--we can anticipate changes to these search results. So, a "Pause" was the right compromise. >> KENNEDY: Nell Kennedy. I'm wondering about time to locate on the GeoSearch, can you talk about your coverage for non-orbital sources of field location data worldwide through your scanning MAC addresses and looking at cell towers right now? And have you done any--have you done any work looking at how Galileo could possibly change how you do search internationally? >> GUNDOTRA: When you say Geo Locate, you mean a feature like My Location? >> KENNEDY: Correct. >> GUNDOTRA: Okay. So, our time to locate varies depending upon the source of the geo data. So, on a cell phone, if it's GPS data, cell phones depending on make and model can sometimes take a very long time, up to 20 minutes to get their initial GPS fixed. In those cases, we fall back to using cell tower, and if the phone has an A-GPS chip, A-GPS gives us assist to GPS. That assist comes from using the unique identifier of the cell tower, and we're able because of a very large database to almost instantaneously give you a reasonably accurate fix until we can get the true lat-long once the GPS kicks in. So, that gives you that very, very fast experience in Google Mobile Maps. >> KENNEDY: And do you feel like you have good coverage internationally for those types of location sources, other than cell--other than GPS? >> GUNDOTRA: The coverage has grown by an order of 92 this year. And, obviously, the more and more phones that carry Google Mobile Maps increases our coverage. So, we're very happy, most places internationally, there are a few very rural areas that we continue to drive for a better coverage in. But the rate of growth is very encouraging. They will have broad coverage. >> KENNEDY: Will Google's Real-Time Search API gets supported in the Google Search API in the future? I would rather like see that in a long tail adaptation. >> SINGHAL: So, that's a great idea. We haven't yet looked into the details of that. And I assure you, we'll be looking at the details of that going forward. >> BENNET: Yeah, one more question online, which is just, when is all the stuff going to be live both in the mobile front end and on the Web front? >> GUNDOTRA: So, I'll take the mobile front stuff. Live today is Japanese Voice Search. Brand new in the Android marketplace is Google Mobile Maps Version 3.3. Or if it isn't live now, it will be live in the next few hours. And also, Google Goggles, as you saw from the Twit, is already available in the Android marketplace. Some of the new innovations that you saw on the Google homepage, like Near Me Now, those are in the coming weeks ahead. We can't exactly predict those dates because of some of the holidays, but it's in the very imminent, in the near future. Things like Product Search are probably a few more weeks beyond that. But that gives you a timeline for everything we discussed except one thing, that fascinating demo I showed where the devise was able to translate from English to Spanish. That was a technology concept demo. You'll see the first products from Google that start to include that technology some time in Q1. >> SINGHAL: So, for Real-Time Search, we are starting the rollout process today. By the end of the day, some percentage of Google users would start seeing Real-Time Search. And in the coming days, we will complete that rollout as our systems roll out to the various data centers that we have. For now, if you want to access Real Time Search, please go to google.com/trends and you can click on the "New Hot Topics" panel on the left, or type your query under that panel in the "New Search" window we have added and you will get Google's Real Time results now. >> You don't get at the Website [INDISTINCT]. SINGHAL: We can check that. We'll make sure. >> I have heard that... >> SINGHAL: No, I'm sure. You know, maybe some of the binary we just did needed. >> [INDISTINCT] before that. >> SINGHAL: Okay. We'll check it right away. Someone's checking it. >> JONES: Bobby Jones from the Guardian, again. I wanted to ask Vic about the visual search. You said there are a billion images already in the computer vision data. I'm just wondering who--whether Google owns, like, the canonical encyclopedic entry of what that image is or whether you determine what that is by, you now, a Web search algorithm. You know, is the Empire State building determined by everyone on the Web saying it's the Empire State building or you deciding that's what it is? >> GUNDOTRA: So I could give you a very compelling answer but I would be remiss as the person sitting directly to your left is the engineering lead behind it. And so, in this case, if you could hand over the microphone to Neven--no, Hartmut. Hartmut will give you the answer. >> NEVEN: Actually, one of the most interesting parts of our system is a technology called "Unsupervised Learning." So, essentially, the algorithms will go out and build a model for visual recognition in a completely unsupervised memory based on photos we find. And then, a model, let's say in your example, for the Empire State building, will emerge as a reflection of what's on the Web. >> JONES: Okay. My follow-up question then was: Would it be possible to Google-bomb visual search? >> NEVEN: In principle, yes, but, I mean, let's say, we have techniques to prevent things like this. >> JONES: Okay. >> BENNET: So, I think we have time for just a couple more questions. One over her in the back, yeah. >> MACENA: So, currently--I'm Chris Macena. I'm sorry. You're doing a lot with text-based status updates. I'm curious if you're looking to expand that to other types of activities that people are doing on the Web that are being recorded through other types of social networking services and systems, again, just beyond status updates. >> SINGHAL: Indeed. We are very excited about what the future holds. Today, we are starting our new Real-Time Search with text-based status updates and the rest of the Real-Time Web. And we are learning about how to do relevance in this world and we have done a very good job of getting you relevant real-time results. As time progresses, we have image search technology, we have video search technology, and we will be accepting all those forms of real-time information. Some of the greatest information I've seen real-time is held in Twitpic, for example, right? The cable broke on the Bay Bridge or something like that. There's excellent information there and, indeed, we will be integrating that going forward. >> MAYER: As also, MySpace, one of our partners, is already looking at how they can take some of the non-textual updates and make them available to us. >> RIPER: Hi. I'm Van Riper. I am actually one of the co-leaders of the Silicon Valley Google Technologies User Group. But in my day job, I work for Crillon (ph), which is a product-based local search in real-time inventory look-up. So I was kind of curious--I should be looking for another job. Could you say a little more about the availability of the integration of product availability in your results that you mentioned early on before all these real-time stuff? It's just very exciting. Did you have a question? Early on--yeah, early on demo and something about it being able to get real-time product availability, product inventory. >> GUNDOTRA: Yes. >> RIPER: And when you're--when you then answered the follow-up question about availability of stuff, you didn't even mention that so I was just curious. >> GUNDOTRA: Oh, I'm sorry. That will be integrated into our product search sometime in Q1. The partners I demonstrated there were Sears and Best Buy. And, obviously, we're working with many other retail partners to get inventory data so that we can combine that with the user's location and deliver that experience. >> RIPER: Got it. Thanks. >> GUNDOTRA: Yeah, sure. >> SINGHAL: Okay. Let me just take a moment. And Dillon tells me that google.com/trends should be working now. Maybe we hit a minor glitch so we will check that again. But please test it. >> SINGLE: Ryan Single from Wire.com, again. Could you give a sense of what relationship there is between the Real-Time Search Index and the Google Web Index, and whether one is feeding into the other or not? >> SINGHAL: Another great architectural question. Google Search Index over the years has evolved to be updated every few minutes, even like, you know, within a few seconds. And Real-Time Search Index is just a continuation of that technology that we have been building for many years now to do things like Google New Search, or even calling the Web very fast. The new thing that we have added into this is update receiving and indexing and merging with the index. And that has been a great new technology because, at that end, we have built some of the technologies I showed you for modeling how information is flowing in the system. >> BENNET: So I think that's all the time we have for now. We'll stick around if you all have questions here. And thanks so much for coming. >> SINGHAL: Thank you. [END]
Info
Channel: Google
Views: 1,151,539
Rating: 3.7439981 out of 5
Keywords: Google, Real, Time, Search, Event
Id: oXHHkROejik
Channel Id: undefined
Length: 93min 26sec (5606 seconds)
Published: Tue Dec 08 2009
Related Videos
Note
Please note that this website is currently a work in progress! Lots of interesting data and statistics to come.