Skip to content
Home » Google I/O 2014 Keynote Transcript

Google I/O 2014 Keynote Transcript

Yesterday Google kicked off this year’s annual developer conference Google I/O 2014 at Moscone Center in San Francisco. We produce here the transcript of the opening keynote of the event…

TRANSCRIPT:

Operator

Ladies and gentlemen, please welcome Senior Vice President, Android, Chrome, and Apps, Sundar Pichai.

Sundar Pichai – SVP, Android, Chrome, and Apps, Google

Thank you, everyone. It’s great to see all of you. Welcome to Google I/O. Every year, we look forward to this date. We’ve been hard at work since last I/O evolving our platforms so that developers like you can build amazing experiences. So thank you for joining us in person.

I/O is a pretty global event. We have viewing parties in over 597 locations in 85 countries, in six continents. And there are over one million people watching this on the live stream today. Let’s say hello to a few locations.

London. Hello, London.

Let’s say hello to Brazil. Everyone is talking about Brazil today. So if it weren’t for I/O, I would be there for the World Cup. I’m tempted to shout “goal!”

Finally, let’s go to Nigeria, where we are thrilled to have an all-female developer group in Nigeria. We are working hard to elevate women in computer science. So look forward to seeing what they develop one day. In fact, at I/O this year, we are very excited there is over 20% female participation, which is up from 8% last year.

And even more excited, we are joined by over 1,000 women in this room today. So thank you.

Of course, I/O is when we talk about our two large computing platforms, open platforms, Android and Chrome, which are built from the groundup for developers like you. Today, we are going to give you an update on the momentum we are seeing in mobile. We are living in amazing times. So we want to talk about the mobile momentum we see and how we are evolving our platforms to support that momentum. And more importantly, we are beginning to think and evolve our platforms beyond mobile. You will hear about that from us today.

And, finally, we want to talk to you as developers as to how you can achieve success on top of our platforms, including an update on Google Cloud platform and Google Play. So let’s get started.

If you look at global smartphone shipments, the numbers are stunning. The industry shipped over 300 million phones last quarter, so they’re on track to ship well over a billion phones each year.

So how is Android doing in the face of this momentum? So in the past, we have talked about cumulative activations of Android. We are switching and focusing on 30-day-active users, users who are currently using their Android devices globally. And you can see the number has been doubling every year. We have gone from 220 million to over 530 million as of last year’s I/O. We are very excited as of this year’s I/O, we are over one billion 30-day-actives users.

The robot is pretty happy as well. So let’s internalize what one billion users actually mean. Android users on a given day sent over 20 billion text messages each and every day. More importantly, perhaps, they take around 93 million selfies every day. The team tells me about 31 million of these are duck faces. We estimate Android users take around 1.5 trillion steps per day, and they pull out their phones and check it over 100 billion times each day. Important use cases which we are working on addressing and you’ll hear about it later today.

Developers are building profound experiences on top of smartphones. Stories we hear every day. A few examples. In Kenya, 40% of Kenya’s GDP flows through M-PESA, giving unbanked people access to financial transactions throughout the country. NETRA-G, a company uses a smartphone and just off the shelf accessories to measure your eye prescription. And they are as accurate as $50,000 equipment you find in optometrist offices, providing very, very affording care to many people.

And finally, University of Michigan, they are using for their patients, they monitor subtle changes in voice quality using their smartphone to detect early signs of bipolar disorder. So the kind of experiences we are seeing on top of these phones are amazing.

So far, I’ve been talking about phones. Let’s shift to Tablets.

Tablets

We are seeing amazing growth in Android Tablets as well. So there is tremendous adoption of these devices. And if you look at how we are doing vis-à-vis the overall market, Android Tablets accounted for 39% of all shipments two years ago. That number increased to 46% as of last year’s I/O.

As of this year’s I/O, Android Tablets account for 62% of the overall market. We don’t include other variants of Android like Kindle. If you add that, it would go up a few percentage points. You know, these are shipment numbers. Again, we care about usage, so we view these as leading indicators of where usage would be.

If you take a look at tablet usage, we’re going to use YouTube as a way — as a proxy to understand usage. A year ago, the total tablet viewership of YouTube, 28% was from Android. That number has gone up, again, to 42%. So we are seeing usage track shipments, and we are very excited people are adopting these devices as well.

Another metric of engagement is app installs. App installs just this year alone on tablet is up by over 200%. So people are really engaging with these devices.

So we are very excited we have a billion users. But we talked about this at last year’s I/O. Our goal is to reach the next five billion people in the world. If you look at a map of the world today, all the regions in blue, emerging markets, majority of users don’t have a smartphone.

When I go back home to India and other countries like that — thank you — it is exciting to see the impact phones have on people’s lives. But it’s disappointing that only less than 10% of the population have access to smartphones. We want to change that. So we’ve been working hard with our ecosystem for — on a very important initiative which we call Android One.

So let me talk to you about Android One.

Android One

what we are doing for the first time, if you look at all the OEMs in these countries, each of them has to reinvent the wheel, and in the fast-paced mobile industry, they have to build a new smartphone within nine months. So we want to pool resources and help everyone. So we are working on a set of hardware reference platforms. We identify the components which go into a next-generation smartphone. These are high-quality, affordable smartphones. We qualify vendors so that we provide a turnkey solution for OEMs to more easily build a smartphone.

In addition to hardware, we are working on software as well. So the software on Android One is the same software you see running on Stock Android, Nexus phones and Google Play edition phones. In addition, through Play, we allow OEMs and carriers to add locally-relevant applications on the device which users have full control over.

And finally, we provide full automatic updates, all the software in Android One comes from Google , so we will keep them updated, just like we do with Nexus and Google Play edition phones.

Let’s take a look at one example device which we are working on.

so this is a device with Micromax. You can see it has a 4.5-inch screen. It has features which matters to a country like India, Dual Sim. Removable SD cards, and FM radio. I’m used to cutting-edge phones, and I’ve been using this phone for a while. And it is really good. And it costs less than $100.

We are working with many partners, we are going to be launching this around the world. But we start this journey in India. And we are launching this with three OEMs in India in the fall of this year: Micromax, Karbonn, and Spice. We are also working with carriers in these markets to provide affordable connectivity packages with these devices.

What we are excited is this is a leveraged turnkey solution so that at scale, we can bring high-quality, affordable smartphones so that we can get the next billion people access to these devices. And we can’t wait to see the impact that it would have.

So we have talked about the momentum in mobile. The next thing we want to talk to you is about how we are evolving our platforms, Android and Chrome. And today, for the first time since we launched Android with Open SDK, we are going to give you a preview of the upcoming “L” release. You will be able to download this later on your development devices. We’ve been working very hard. This is one of the most comprehensive releases we have done. It has over 5,000 new APIs, and we are thinking about “L” release not just for mobile, but for form factors beyond mobile.

One of the things, as we thought about “L,” we wanted to take a radical new approach to design. User experiences are evolving rapidly, and we wanted to rethink the user design experience in Android to have a fresh, bold, and new look.

To talk about the design for “L,” let me invite Matias Duarte.

Matias Duarte – VP of Design, Google

Thank you, Sundar. Design is essential in today’s world. It defines your experiences and your emotions. So we challenged ourselves to create a design that was not just for Android phones and tablets. We worked together — Android, Chrome, and across all of Google — to craft one consistent vision for mobile, desktop, and beyond. We wanted a design that was clear and simple and that people would intuitively understand.

Philosophy of Material Design

So we imagined, what if pixels didn’t just have color, but also depth? What if there was an intelligent material that was as simple as paper but could transform and change shape in response to touch? And this led us to a way of thinking that we call material design.

[Video Presentation]

We drew inspiration from paper and ink. However, unlike real paper, our digital material can expand, reform and reshape intelligently. Material has physical surfaces and edges because the human mind is wired at its most primitive level to instinctively understand objects and their relationships. These scenes and shadows provide meaning about what you can touch and how it will move.

In the real world every small change in position and depth creates subtle, but important changes in lighting and shadows. So as part of the “L” preview, we’ll now allow app developers to specify an elevation value for any UI surface and the framework will render it in correct perspective with virtual light sources and real-time shadows.

Material design is beautiful and bold because clean typographic layouts are simple and easy to understand. Your content is the focus. So the “L” preview will allow app developers to easily colorize all framework elements in your app to match the theme to your brand and we’re previewing a new support library that we call palette to easily extract colors from images and really put those vivid pictures front and center.

We’re giving designers familiar tools like baseline grids that work across screens, grids ensure apps have a consistent rhythm and character and this will allow you to start with a design on a phone and logically and easily bring that same design to tablets and laptops.

Now one design doesn’t mean one size fits all. Our guidelines allow you to appropriately adapt the UI, so your users will already know their way around your app no matter what screen they use it on.

And we’ve also updated our system font Roboto so that designers and developers can use one type face, designed and optimized for every screen from your watch to your laptop to your television.

Animation

So now let’s talk about Animation. It’s delightful when your touch is rewarded with motion and material surfaces slide around with the physics of card stock, but they respond to touch with splashes of virtual link that are like ripples in a pond.

As part of the “L” preview, all of your applications UI building blocks have been updated to include rich, animated touch feedback. And no detail is too small to bring a smile to your face like when the reload button loops around or the playback controls can change.

Finally, in the real world, nothing teleports from one place to another and that’s why it’s so important to animate every change on screen in a way that makes sense. In the “L” preview, Android developers will now be able to create seamless animations from any screen to any other between activities and even between apps.

So you’re probably wondering how this looks like in practice. We’re going to give you a sneak peek at one of our familiar Google applications in the new material design.

Here you can see step by step how we update the design. The typography, the grid changes, and finally the surfaces and bold colors.

And a few small changes make a really big difference. And you can also see how easy it is to take a same design to different screens. Now, I’ve talked about only a few of the highlights of material design and just some of the APIs that you can try out in the Android “L” preview.

But as you know, people spend an enormous amount of time on the web and especially the mobile web. Last year at I/O, we announced Polymer, which is a powerful new UI library for the web. Today we’re bringing you all of the material design capabilities to the web through Polymer. As a web developer, you will be able to build applications out of material design building blocks with all of the same surfaces, bold graphics and smooth animations at 60 frames per second.

So between the “L” preview and Polymer, you can bring the same rich, fluid material design to every screen. And to help you take full advantage of this framework, we’ve also completely redesigned and created one unified set of style guidelines for every screen and for all devices. These guidelines will help designers and developers alike understand best practices and build consistent, beautiful experiences.

We’re releasing the first draft of these guidelines as part of our preview today at Google.com/design.

Now that you’ve seen our new look and feel, I’d like to invite Dave Burke to show you some of the new features in the Android “L” developer preview.

Dave Burke – Director of Engineering, Android

All right. So over the last eight months, our team has been busy cooking up the biggest release in the history of Android. And as Sundar mentioned, we’ve added over 5,000 new APIs touching nearly every aspect of the system.

Now we don’t have time to even come close to covering everything in “L” today. So instead, what I like to do is walk you through some of the highlights of the tremendous steps we’re taking on the user experience and on the performance of the Android platform.

so let’s start with user experience. Now, bringing material to “L” is of course a big part of what we’re trying to do here. We’ve added a new material theme, so it’s a new style for your application that includes new system widgets, transition animations and animated touch feedback.

We’ve also added new animation support, so a new drawable for animated ripples, a reveal animator to animate a clipping circle to reveal views and we’ve extended views to not just have an X and Y component, but also a Z component to provide elevation. So you can float elements of your UI and the framework will cast a real-time shadow for you.

My favorite feature that we’ve added in support material is the ability to customize activity, entry and exit animations. You can even include [share here] elements. For example, an image that starts in one activity and animates seamlessly through translation and scaling into another.

So let’s take a look at this in practice. Let’s have a look at now, we’re all familiar with, which is the phone dialer. Thanks, Marcelo.

Okay. So the first thing you will notice when you fire up the phone dialer are those bold material colors and shadows. And you will see the ripple touch effect as I touch each of the tabs and you get a more subtle material touch effect on the recent calls. You’ll see that the dialer button has is elevation sets, so it’s floating above the UI and as I tap it you get these really nice delightful animations.

Now another feature we added to support material is something we call nested scrolling and the idea is as I scroll we propagate the scroll events up the view hierarchy and different parts of your views can respond differently. So for example, as I scroll upwards here you will notice that the recent call to Marcelo will start to shrink and disappear, then the search box will start getting pushed upwards and the tabs will lock into place. It’s a really nice effect.

So let’s go over to the dialer. So it turns out my mom’s a big fan of material design. I need to go call her up and tell her about how to set elevations on her views. I know she loves that. So let’s get start dialing. You will see that ripple touch effect again emanating out from the buttons. Then when I go to place a call, you will see a view animator and will animate into the — in call screen like so. So it’s a really nice effect.

Okay. So that’s a quick taster of material in “L” and what you’re seeing here is really a sneak peek of work in progress. We wanted to give you guys early access so you could start bringing material to your apps.

And we also recognize the change in the UI in such a big way will take some time. So we started with a dialer as a showcase. Over the coming summer months, we’ll be extending material to all aspects of our apps and the system and the result is going to be a dramatically enhanced, fresh user experience.

Okay. So another area where we’ve improved the user experience on “L” is around notifications. One of the most frequent reasons we all take our phone out of our pocket everyday is to respond to incoming notifications. We all do this dozens and dozens of times a day. So we wanted to streamline the process. Everything from the phone buzzing to you acting on the notification.

In “L” we give you instant interactive access to notifications right from the lock screen. So now you can read, open and dismiss in seconds.

So let’s take a look at my device. The first thing you’ll see are all my top notifications on the lock screen and we’re rendering them as sheets of material, they animate really beautifully. If I touch them you can see that material touch effect.

Now in “L”, we’ve improved the way Android organizes and prioritizes notifications by analyzing this behavior to make sure only the most useful, relevant notifications are presented to you. I can swipe down and I get my full list of notifications. And we’ve done a clever thing here where we’ve merged the notification shade, so something that’s been in Android since 1.0 with the lock screen.

And so from here I can double tap on a notification to launch the corresponding app, or if there is something I don’t need, I can just dismiss with a single swipe. And to unlock the phone, well this is just the notification shade, so you just swipe it away and you’re straight into the device. Fast and simple.

We’ve also introduced a new type of notification in “L” that we call the heads-up notification. And this can be used to let you know about something urgent without interrupting what you’re doing. So let’s say I’m playing any new favorite game, Piano Tiles, and I’m going along here, I’d get my highest score ever, and then all of a sudden I get a call from Marcelo. So from here I can keep going or — if I want to act on that like non-script or if I am busy, swipe it away. And then I can go back to my game and get the highest score I ever got in public. Yeah. All right. That’s actually my worst score I’ve ever got.

Okay, anyway, let’s move on. So while we’ve made the notifications more powerful, if you’re one of the approximately 15% of people who has a pin or pattern lock you waste many minutes a day cumulatively on that thinly task of entering your pin. So we figured there’s got to be a better way.

In “L” we’re introducing a new concept, we call, personal unlocking. And personal unlocking enables the device to determine if it’s in a trusted environment, say in the owner’s hand or beside the owner on a table. Person unlocking uses signals such as locations you designate, Bluetooth devices that are visible, even your unique voice print. So for example, let’s have a look at this device. Thanks Marcelo.

So I currently have a pattern unlock on this device, but because I’m wearing a Bluetooth watch my phone knows it’s me who is present and so it doesn’t challenge me with an unlock. So for example, if I just swipe up, the phone will unlock just like that.

Now, let me just reset that. If I take any watch off, let me just hand it to Marcelo. Okay. So now my phone can no longer see the watch. And because of that my phone cannot ascertain that’s me who is present. As a result, my phone will lock down its security. So now when I go to unlock the device it presents me with a pin lock. It’s a really great feature.

So that’s a few of the user experience improvements we’ve made to support material and notifications.

Another area of “L” where we’re significantly improving the user experience is around how we’ve integrated the mobile web into the platform. So to learn more, let me invited up Avni Shah to the stage.

Avni Shah – Director, Product Management, Google

Thanks, Dave. A core part of your experience with mobile devices is the mobile web. Just to get a sense of the growth that we’ve been seeing, at the beginning of last year, we had 27 million monthly active users of Chrome on mobile. Today we have more than 300 million. That’s 10x growth. Yeah, it’s awesome. It’s 10x growth in just the last year alone.

What that means for us is that we need to make the mobile web work well for our developers and our users. Today, I’m going to talk about three ways we’re going to do that.

We’re enabling material design experiences on the mobile web. We’re redesigning recents to help you multitask, and we’re extending our capabilities of app indexing to help people get to where they want to go faster.

Material Design

First let’s talk about material design. One of the big parts of your experience with the mobile web is, obviously, the web sites themselves. They need to work well. They need to look great. They need to be fun to use. You heard Matias earlier talking about the philosophy of material design, a bold, consistent, intuitive experience that just works across screens. We’ve been working really hard at making those experiences not just possible, but the new standard for the mobile web.

To show you what this looks like, my good friend Tom here is going to walk us through an exploration of google.com search results on the mobile web re-envisioned with material design.

So, Tom, let’s go ahead and do that search for a “starry night.”

Now, the first thing that you see here is that this panel is rendered as a beautiful material-style card. You notice the use of color, the title is on a blue background. That was actually programmatically matched to the painting. If Tom clicks on expand the card, you’ll notice it filled the screen with a continuous animation. If he scrolls, the header will shrink. It won’t pop into place, but it has a smooth animation that just makes sense.

Now let’s go ahead and click on the suggestion at the bottom to get more of Van Gogh’s artwork. You’ll see those search results also smoothly animated into place. Tom is going to continue to give us some demo eye candy over here.

And while this is just an exploration that you’re seeing, I want to mention that this is fast, fluid, continuous animation at 60 frames per second. This thing was — just wasn’t possible a year ago.

We’ve been working really hard at improving the performance and predictability of the platform to make things like this possible. For example, this demo shows off the work that we’ve done on touch latency, giving you as a developer notice the touch events earlier in the frame so you have more time to act.

As Matias mentioned earlier, with Polymer — our UI toolkit for the web — all of you can build web experiences that feel as awesome as this.

Multitask

The next big area we’ve been thinking about is how to help you multitask. And we think the recents feature in Android is one way we can actually make this easier, especially as tasks across the web and apps, as they often do. So once again, Tom is going to walk us through the changes here.

So, Tom, let’s go ahead and click on the recents icon in the lower right. Now, as Tom scrolls through, the first thing that you will notice is recents has also been grounded in material design. You’ll notice the overlapping cards have been rendered with realistic shadows and perspective.

But there is another thing going on here that may not be immediately apparent. Tom’s Chrome, they also sit here as well. He’s been researching restaurants to go in SF. So he has articles from the New York Times and SF Chronicle, here is individual items. You’ll notice the site icons and the fav icons there.

AS he scrolls back of it further, you’ll notice he has been researching in the Urbanspoon app, he has docs up opening where he’s been collaborating with some friends.

So let’s go ahead and click on that doc and see what your friends have to say.

I’ve heard great things about State Bird Provisions. Let’s check out that article here. Now, what you see here as this loads, this is actually loading as a web site in Chrome. You’ll notice the URL up at the top. Now Tom pops it back into recents, that page is now listed there, along with all of his other open stuff.

I want to point out what the big difference is here. This is a view you couldn’t get before today. If you wanted to get to all your open websites, you’d have to go into Chrome and kind of flip through them there. But by bringing all of your individual Chrome tabs here and listing them in your recents view, we’re making it really easy for you to move between the web and apps, making multitasking just that much easier.

And last, but not least, this change to Chrome is actually built on top of a new API in “L” that allows apps to populate multiple items in recents. So for all you app developers, if this kind of thing makes sense for you, you can make use of it as well.

Going a step further, we’re also making it easy for you to find content using Google search, whether it’s deep in the web or deep in an app. So last fall, we announced app indexing. As a developer, this capability lets you get to app content, lets you get your users to app content, directly from the search results page. Since then, we’ve been working on a ton of UI improvements and extending some APIs to make this more powerful.

But let me just give you a quick refresher of what this capability enables. So let’s go ahead and do a search for Waterbar Restaurant, I’ve heard good things about it, over by the Embarcadero. As Tom scrolls through the search results, you’ll see close to the top there’s a link for the home page to Waterbar. And near the bottom of the screen, there’s — oh, actually, in the middle of the screen, there’s a result for OpenTable.

Now, what’s different about this UI is it’s linked to OpenTable instead of going to the website, it’s actually going to take us to the OpenTable app, because Tom happens to have the app installed.

So let’s go ahead and click on that link. And you’ll see it takes us directly to Waterbar within the OpenTable app. Up until now, this was only available to a few applications. But today we’re opening it up to all Android app developers globally, along with some tools to get you started.

And going further, if your site requires — sorry, if your app requires your users to sign in, you’ll be able to use Google+ sign-in in the coming months to have your public content show up in search as well. You know, we thought this would be even better if we could help your users rediscover content that they’ve already found in your apps. So we’re adding a new API in Google Play services to do just that.

So let’s quickly show you how this works. Tom found this really cool 3D tour of the Ferry Building earlier and he wants to get back to it. So starting with the search box on his home screen ,he’s going to do a search for Ferry Building. And what you will notice at the bottom of the screen is there are search suggestions for Ferry Building marketplace in the Google Earth app. And this is there because this was the app that he was using when he found that tour before, even if he himself didn’t remember. With a single click, he’ll get taken directly to the tour of the Ferry Building within the Google Earth app.

Now, this is possible because the app is making its content available based on its user’s previous actions. We just showed you this with Google Earth. But any app that utilizes this new API will have the same capability. For developers, we think this is a great way for you to help your users rediscover content right when they’re looking for it.

With that, I’ll hand it back to Dave who is going to take you through more of the enhancements you can look forward to in “L.”

Dave Burke – Director of Engineering, Android

Thanks, Avni. So we’ve covered some of the highlights of the user experience. But there’s lots of other user experience improvements in “L,” for example, the new keyboard UI, Do Not Disturb mode, your quick settings and much, much more.

Android Virtual Machine

But in the interest of time, let’s move on to the second major theme of “L.” And that’s performance. Let’s start with the Android Virtual Machine. So you might remember that we made a very early version of our new runtime, ART (Android Runtime), available as a developer option in KitKat. But we got some really great feedback from you guys, as well as some excellent open source contributions from ARM and Intel and MIPS. And I’m excited to say that we’re finally ready to pull the trigger on this bad boy, because the “L” release runs exclusively on the new ART runtime.

So we wrote ART from the groundup to support a mix of ahead of time (AOT) compile, just-in-time compile and interpreted 6code. It’s truly cross-platform, so it supports ARM x86, and MIPS. We put a lot of effort into optimizing ART’s back-end compilers. And this has resulted in a 2x improvement performance over Dalvik. And best of all, this one is on us. You don’t have to make a single change. All of your app code just gets the performance improvement for free.

ART also has a brand-new garbage collector and memory allocator. So this dramatically reduces the number of pauses and duration of pauses associated with a garbage collection event. As a result, your app runs more smoothly. So if you take a look, for example, at Google Maps on both Dalvik and ART, first you’ll notice the number of pauses have reduced from two to one. But also the pause duration has reduced from roughly 10 milliseconds down to about 2 milliseconds to 4 milliseconds. So now it fits comfortably in a desync window. No more application stutters.

And there’s more. ART doesn’t just bring better performance, it’s also more memory efficient. So it’s smart about when the app is put into the background, in which case we’ll apply a slower but more intensive moving collector to save anything from hundreds of kilobytes to many megabytes.

And finally, ART is fully 64-bit compatible. In fact, we’ve adapted and optimized the entire platform to take advantage of the new 64-bit architectures. So now you can benefit from larger number of registers, newer instruction sets and increased memory addressable space.

So to take advantage of 64-bit, we’ve added support for new APIs in the NDK, so ARMv8, x86-64 and MIPS64 and of course if your app is written in JAVA, then it will work with absolutely no modification on new 64-bit hardware.

Okay. So that’s CPU performance. The other side of the coin is GPU performance, graphics. And I’m really excited about some of the things that we’re doing in “L” in this area.

So historically, mobile graphics has lagged to desktop by virtue of the fact that mobile GPUs are smaller and more power-constrained. But that’s changing quickly. Mobile GPU performance is catching up with console graphics and even PC graphics. So in “L,” we specifically wanted to close the gap between desktop DX11-class graphics capabilities and mobile. And we’re doing that with something we call Android Extension Pack.

So we set out to work with GPU vendors, including NVIDIA, Qualcomm, ARM, and Imagination Technologies, and together we define the Android Extension Pack. So it’s a set of features that includes things like tessellation, geometry shaders, computer shaders, and advanced ASTC texture compression.

So let’s take a look at the Android Extension Pack in action. And what you’re about to see is Epic’s Unreal Engine 4 desktop rendering pipeline running on Android “L” on the latest NVIDIA tablet hardware. Now the Android Extension Pack enables much more advanced shaders, so we can have more realistic environments, more realistic characters, and vastly improved lighting.

So let’s go start this off. We have a bit of a flicker here.

[Video Presentation]

Well, if you’re squeamish, you might not want to look now.

So you have to excuse some of the flickers. It’s a vertical electromagnetic storm up here.

Okay, so as I mentioned, this isn’t just cut-scenes, actually live and we can play it through the world. Hopefully we can get that up in a moment. Some of the rendering that you saw that it was truly incredible. So there were really amazing reflections in the water, lighting effects. Tessellation were being used for the smoke effects.

And starting with the “L” release in the fall, you’re going to see new high-end tablets and phones shipping on Android with this level of graphics capability. So quite literally, this is PC gaming graphics in your pocket.

The last performance enhancement I want to take you through is on battery. And we’ve worked hard to make sure that the battery keeps up with the performance. And of course, there are variety of systems and components that tax the battery on a modern phone or tablets. So Wi-fi radios, cell radios, GPS, GPU, et cetera.

And you might remember we’ve had some previous efforts to improve quality on other releases, so Project Butter for UI smoothness and Jelly Bean, Project Spell for memory for print in Kitkat.

Well in the same team, and brought to you by those same project naming geniuses, we have Project Volta. And the goal of Project Volta is to optimize how the extensive subsystems of the device are used and to improve overall battery life. So the first thing we did was improve our instrumentation of battery data. You can’t improve unless you can measure. So we created the tool that we call Battery Historian. And it helps you visualize on a time axes the battery usage information.

Now you can correlate battery discharge with what was happening to the device at the time. So here’s an example of Battery Historian from a real device. And on the top graph, you can see an issue where the radio is waking up approximately every 20 seconds, but Battery Historian helps us quickly identify the issue so we could fix it and therefore improve battery life. In fact, we’re using this tool to make this system in Google apps more efficient so — and we can expect a significant battery improvement in “L.”

We’ve also added a new JobScheduler API to help you optimize power consumption in your apps. So using the JobScheduler API, you can make your application more efficient by allowing the platform to call less non-urgent network requests from multiple apps. As a result, the platform can keep the radio asleep a higher percentage of the time, thus saving significant power.

So you can use the JobsCheduler API for your app to schedule a maintenance task while the phone is connected to the charger, for example. Or also even just to download application updates. I’m sure you’ve all had that experience where you’re out and about, your battery is about 4%, and then applications start updating themselves. It’s like, “No!.” And then finally, we’ve added a new Battery Saver Mode in “L.”

So Battery Saver allows you to clock down CPU, the refresh rate, even turn off background data to conserve battery. And you can trigger it manually or configure it to come on automatically when the battery level is low. So Battery Saver is really great if you’re about to embark on a long hike or maybe a long protest and, say, you want the battery to last even longer.

So on the Nexus 5 running in Battery Saver mode, you can extend your battery life by up to 90 minutes of usage within a typical single day’s use.

So I just gave you a quick whirlwind tour of some of the highlights of “L,” how we’re improving the user experience through steps like improved design, smarter notification, and intuitive authentication. And also the enhancements on the performance side, so faster run time, better graphics, and stronger battery performance.

But I only scratched the surface of “L.” And as I mentioned at the start, this is our biggest release to date. You’re going to find things like better multitasking, Bluetooth 4.1, burst mode camera API, USB audio support and much much more.

Tomorrow morning, we’re going to be making the “L” developer preview SDK available from developer.android.com and also posting early system images for the Nexus 5 and Nexus 7 so you can start developing for “L” today.

So with that, let me hand back to Sundar. Thank you.

Sundar Pichai – SVP, Android, Chrome, And Apps, Google

Thank you, Dave. As Dave said, the “L” release with 5,000 new APIs is one of our most comprehensive, and we are very excited we are sharing it today.

We have a whole new design with “L,” tons of UX features, and a whole slew of performance improvements. When you take a step back and you look at what we are doing with Android, the approach we are taking is very unique and distinctive. We aren’t building a vertically integrated product. What we are doing is building an open platform at scale. We work with hundreds of partners globally to bring the product and the platform that touches billions of people. And we want to do it in a way in which we are innovating at a very, very fast pace.

If you take a look at the innovation that’s happening in Android and if you look at some of the recent announcements from others, you can see that things like custom keyboards, widgets, those things happened in Android four to five years ago. We are working very, very hard to bring an open platform and innovate on it at an unprecedented scale. We want to make sure we ship these features to users as fast as possible. That’s where 3google Play Services come in.

Google Play Services ships every six weeks. And 93% of our users are on the latest version of Google Play Services across all versions of Android. In fact, by shipping every six weeks, we in many ways can iterate faster than typical OS release cycles. While it’s open platform and we want to innovate fast, we want to make sure it’s very, very secure as well. So we take security very seriously.

Security

Let’s take an example at malware production. In Google Play, we automatically scan every single application for malware. And if users opt in, we even scan applications from outside of Google Play to make sure they are malware-free. Given the popularity of Android, there’s a whole vested industry, given there’s a lot at stake around security. But based on every data we see, well, well less than half a percent of users ever run into any malware issues. And increasingly, we are pushing security updates through Google Play. So any security updates related to Google Server communications, we are now pushing those updates through Google Play Services so that we can get them to users within six weeks.

With “L,” we are also launching factory reset production so that if your phone gets stolen, users have full control to disable their phones.

Privacy

Finally, privacy is an important part of security. So with “L” release, for the first time, we have a centralized setting, what we call universal data controls, where users can go and manage their important privacy protections. They can control data that is shared from the device, like location history, et cetera. And so we are doing that in “L” as well.

So far, we have been talking about “L” release in the context of mobile phones and tablets. But users increasingly are living in a multiscreen world. You are using other connected devices: the television in your living room, you’re increasingly wearing things on your body, you are — when you get into your car, you expect a connected experience. We want to work to create a seamless experience across all these connected devices.

So with “L,” as well with Chrome, we started laying some foundational principles on how we evolve our platforms to support these new connected experiences. So here are a few principles.

We are making everything contextually aware. We want to understand whether you’re at home with your kids and you want to be entertained, or you’re at work trying to be productive, or maybe you are traveling. We want to bring the right information to you at the right time. We want the experience to be voice-enabled. We are building the most advanced voice recognition infrastructure in the world. And we want to help users interact with computing devices in an intuitive way. For example, when they are driving or cooking, we want voice to be a major source of input. We want the experience to be seamless. It shouldn’t matter which device you were using before. We want to pick up where you left off.

And, finally, users always have their smartphone. So we want to make sure all these connected experiences work based on your smartphone, be it your wearables, be it your car, or like we have shown with Chromecast, via television.

So both with “L” release and Chrome, we are bringing a whole set of new experiences to many connected screens around you. The first area we are going to talk to you about is wearables.

Wearables

About three months ago, we launched our preview of Android Wear. We announced a developer SDK, and the reception has been very, very positive. To give you a further update, I’m going to invite David Singleton onto the stage.

David Singleton – Director of Engineering, ANDROID

We’re right at the beginning in a new phase in the miniaturization of technology, which means that it’s finally possible to make a powerful computer small enough to wear comfortably on your body all day long. And there’s a huge opportunity to bring rich user experiences to these devices. And that’s why we’re building Android Wear as our platform for wearables based on Android.

Android Wear makes it easy for developers to reach users on this new form factor using precisely the same tools we’re already familiar with on Android phones and tablets. People will be wearing these small, powerful devices, so style is important. And that’s why Android Wear supports both square and circular screens. And we think that there will be a wide variety of fashionable designs.

Sensors will help them understand your context, so they can provide useful information when you need it and help you reach your fitness goals. And as the device that you always have with you, your watch will also provide intelligent answers to spoken questions. And as Dave showed us earlier, act as your key in a multiscreen world.

Across the world, people check their Android phones an average of 125 times every day. And that’s why we’ve designed Android Wear to quickly show you relevant information and make sure you never miss an important message, while letting you stay engaged with the people that you’re actually with. We do this by working to understand the context of what you care about while enabling very brief interactions with the device.

Here’s a live demo on the LG G watch. You can see that it has an always-on display, that at any given time shows you the most important thing we know for you. So, Jeff, it looks like your flight to Brazil for the World Cup is on time. I guess you do deserve a break after this big demo. And if Jeff wanted to see more, he can simply raise his watch or tap the screen to switch into vibrant, full color that you’re already seeing here.

Throughout the day, if Jeff receives a notification which buzzes his phone, his watch will vibrate on his wrist and show him what’s up at a glance, so he won’t miss an important message like this one. Swiping up and down navigates you through this stream of cards, which includes information from Google Now, apps running on Jeff’s phone, and apps running directly on the wearable itself. And when there’s a page indicator, Jeff can swipe horizontally to see more details.

You can see that we’ve applied material design here. The cards float above beautiful, textured backgrounds. And just like in your phone’s notification shade, you can swipe a card away to remove it from your screen. Let’s take a look at Jeff’s phone. And that notification has disappeared.

Back at the watch face, pressing and holding lets you choose a different one. You can see that there’s a broad selection of analog and digital designs in a variety of styles to suit your tastes. Okay. Now that we’re acquainted with the overall UI model, let’s see how Android Wear can work for you.

Imagine that Jeff has just got up in the morning. He swipes up and sees the weather forecast for the day. His commute’s not looking too bad. And, oh, look, Jeff, I guess you need that package for your trip to Brazil. You better not forget to pick it up.

Jeff: Okay, Google, remind me to check my mailbox when I get home.

David Singleton: If we can see Jeff’s phone at the same time, you’ll see this is immediately synced across. And in this case his watch was smart enough to know where home is. A little later on, as Jeff is arriving at the office, his watch vibrates again with a chat message from one of the team. He can see who it’s from and what he’s saying, all without having to fumble around and get out his phone. Your watch and your phone stay in sync. When you swipe away a notification on the watch it disappears from the phone as Jeff is showing now.

It’s super convenient and apps will stay in sync too. While Jeff is making his lunch, he finds it really hard to get the last peanut butter out of the jar and he comes up with an idea, possibly a very brilliant one that he wants to remember.

“take a note, double sided peanut butter jar

Watch his phone. We’ll try that one more time.

“take a note. Double sided peanut butter jar”.

I guess the universe doesn’t want there to be a double sided peanut butter jar after all.

So moving along that note would have been saved immediately to his phone and he can get on with his lunch. In the evening Jeff is having dinner with a friend at a restaurant. If he’s unfamiliar with one of the ingredients on the menu he can just say…

“what is limburger?”

So it looks like limburger is a type of cheese. Jeff is lactose intolerant, so he better order something different or this dinner could go wrong. And when Jeff receives a phone call his watch will vibrate and he can see who is calling at a glance. It’s another one of Jeff’s co-workers. Now Jeff could get out his phone to answer, but since he’s busy he can either swipe to reject the call from his wrist or swipe up to choose from one of these SMS replies. His phone sends the SMS and he’s done.

Sometimes you’re enjoying dinner so much that you want to avoid any more interruptions. And for that you can set do-not-disturb with a single downward swipe from the top of the screen.

And now Jeff’s watch won’t buzz again until he wants it to. Later that night Jeff arrives home. Oh, that’s right, your package is here. Now that he’s at home, the reminder that Jeff created this morning has triggered.

You can also use Android wearables to control other devices around you. Let’s listen up with a bit of music.

“play some music”.

Now you will see that Jeff has music controls on his watch. He can see what song is playing, he can pause the music or skip to the next track. And while it’s playing the album art is beautifully displayed right there on his wrist.

Finally at the end of the day it’s time for bed.

“set an alarm for seven a.m.”

With glanceable notifications and quick voice actions, Android Wear gives you what you need, right when you need it. Let’s take a closer look at some of the contextual information that Android Wear provides when you’re traveling. So Jeff’s about to leave on that big trip to the World Cup. It’s the morning of his flight so his phone is already displaying relevant information for his trip. He can see his flight status, and even show his boarding pass. His hotel address will be there when he needs it and he knows whether or not he will need to pack an umbrella. It does look like it’s going to rain in Brazil on Friday.

And once he’s in Brazil, Android Wear continues to give him useful, timely information at a glance, whether it’s his restaurant reservation, the time back at home so he knows when to call his family or the local bus schedule. And while he’s walking around the city, Jeff can see how many steps he’s taken today along with a step count history for the week. On devices that support it he can even check his heart rate after a jog.

So we’ve shown you what Android Wear can do out of the box. We’re even more excited to see what developers build on top of this platform. The Android Wear SDK enables you to build glanceable, contextual apps for this new category of device. Let’s talk through the capabilities it gives to developers and then we’ll show you some examples.

Right off the bat, Android Wear automatically bridges notifications from your Android phone or tablet directly to your watch. Now, Android’s notification APIs already allow you to build beautiful user interfaces with big pictures, actions and more. And there are hundreds of thousands of apps delivering billions of these notifications everyday. And now they’re available on your wrist.

Back in March, we released a developer preview enabling apps running on the phone to add things like voice replies, have several pages and group notifications in bundles. With these additions you can begin to provide a tailored experience for wearables and we’ve used these features to add wear support to Google apps like Hangouts and Gmail and there’s been a huge response from developers.

The very best wearable apps respond to the user’s context, put glanceable cards on the screen and allow the user to take direct actions in just a few seconds. Here is one of my favorite examples. With Pinterest you can follow other people’s pins. Pinterest app will let you know when you’re near a place that has been pinned by someone you follow. So Jeff’s friend Susie loves Korean barbecue and she’s somewhat of an authority on the best restaurants in San Francisco. So when Jeff is in the city, Pinterest can notify him that he’s near one of Susie’s pinned restaurants. The notification will appear on his wrist just like this and it uses pages allowing him to quickly glance at the details, then swipe to see a map. And if he likes it he can start navigation right from his wrist.

This is using Google Maps for mobile which gives you turn by turn navigations on your watch. And it’s particularly useful when you are walking and it works with all Android Wear devices. In addition to what’s possible with notifications bridged from the phone, today we’re making a full Android Wear SDK available which enables you to write code — it’s pretty great. It enables you to write code that runs directly on the wearable itself and almost all the APIs that you’re already familiar with on Android are available here. That means that you can present fully customized UI, read sensors directly and much, much more.

We’re also introducing a new set of APIs in Google Play services that makes it easy for your app to send data between a phone or tablet and a wearable. And we’ve road tested these APIs with some developers over the past few weeks. Let’s take a look at examples of what they built.

Eat24 is an app that makes food ordering both fun and easy. Watch this. Hopefully I’m going to order a pizza in 20 seconds. When it comes to takeout I’m a creature of habit and Eat24 has recognized this and takes advantage of that contextual stream. Around the same time I 2made an order last week it posted a notification suggesting I order again. I can tap on the notification and launch into their full screen UI. And here I’m presented with a beautiful interface that lets me confirm the kind of food I’d like today. Let’s stick with pizza. And then I can quickly swipe to see and repeat my last order. Just one more tap to pay. And the pizza is on its way.

I think that clocked in under 20 seconds. Now, you might be wondering how this app got to my watch. Well, all I had to do was install the Eat24 app from the Play Store on my phone. When a watch is connected, the wearable portion of the app is automatically installed and kept up to date on that watch.

I mentioned the new wearable APIs for easy communication between phone and watch. All The Cooks is a social recipes app which has made really great use of these APIs. I don’t know about you, but I find it really hard to follow recipes especially when it gets to the tricky bits where everything is happening at the same time. So wouldn’t it be more convenient if I could just look down at my watch and see what to do next?

With the All The Cooks app I can choose a recipe — let’s go into my favorites and choose beef brisket chili. The recipe will immediately appear on my watch so it’s always right there with me. Let’s get started.

I’ve got all the ingredients, so let’s start following the steps and now watch the phone carefully. As I move from step to step the phone stays in sync too. And if you’re wondering whether or not it’s safe to wear your watch while cooking, it’s great to know that all the devices we’re talking about today are water resistant.

And with All The Cooks, whenever a recipe calls for a timer like this four hours in the oven, I can do that right away on my wrist. So no more burnt dinner.

We saw some great examples of voice actions earlier today. And we believe voice actions will be most useful when they can invoke the best service in any app. We’re just getting started with this, but we’re making voice available for some key actions on the wearable today and we’ll be adding more over the coming months.

Lyft is a transportation service and ride-sharing app that allows you to request a car to pick you up at your exact location. Lyft have implemented our call a car intent so it’s really easy to just walk outside and say —

“Okay, Google, call me a car”.

You’ll see that lyft is able to determine Jeff’s exact location from his phone and presents this confirmation screen so he can verify his address. The app has also made great use of notifications in the stream. You can see when your car has arrived, keep up to date throughout the journey and even rate your driver right from your wrist when you’re at your destination. Thanks to all our developers.

Now, we showed a preview of a couple of watches we were working on with our partners back in March. The LG G watch will be available to order later today on the Play Store.

In addition, you might have caught a glimpse of a new device during the demos. We’re very happy that Samsung is joining the Android Wear family with the Samsung Gear Live. And the Samsung Gear Live is also available to order later today.

The Moto 360 is the first watch to adopt the round Android Wear UI and it will be available for sale later this summer. Those are just the first three watches. There are many more on the way and we’re thrilled to enable developers across the world to build apps for what we believe will be a revolutionary new form-factor.

And now I’d like to invite Patrick Brady on stage to tell you about how we’re bringing Android to the car.

Android in The Car

Patrick Brady – Director of Engineering, Android

Thank you, David. Isn’t that great? Android Wear creates a seamless experience by connecting your Android smartphone to a wearable device. And the result is truly amazing. Wouldn’t it be great if all of your devices were this connected? For many of us cars are an integral and essential part of life. They bring us to the grocery store and take us on weekend trips. They bring us to work and take us home.

In fact, in the United States the average commuter spends over one hour in the car everyday. In many ways our cars keep us connected to the physical world around us, but they remain disconnected from our other devices and our digital lives.

So what have drivers done to bridge this divide? Well, even though it’s unsafe and in many cases illegal, people use their phones while driving. And reports show that 25% of accidents in the U.S. are caused by people fumbling with gadgets behind the wheel. There’s got to be a better way?

So back in January, we announced the Open Automotive Alliance to address this problem and make the connected car a reality. We’d like to show you what we’ve all been working on. And today we’re happy to announce Android Auto.

We’ve redesigned the Android platform for automotive, making it easier and safer to use the connected apps and services drivers want in the car. We looked at what people do with their phones in the car today. And these things stood out to us. Navigation, communication, music and other forms of streaming media. Android Auto puts this front and center so you don’t have to go hunting through the grid of icons to find the apps that are most important to you when you’re in the car.

Android Auto is contextually aware to give you the right information right when you need it. And most importantly Android Auto is completely voice enabled so that you can keep your hands on the wheel and your eyes on the road. You know, we really wanted to drive a car up here on stage and show you this live in action, but apparently there are these regulations and logistics that make driving a vehicle in a room packed with 6,000 people a very hard thing to do. So we set one of our engineers on the problem and apparently this is what happens when engineers have access to a blowtorch.

So we’re down one test car, but we have a great demo cockpit to show you. And now I am happy to introduce Andrew Brenner, our product manager who will literally drive this demo.

So to start, Andy connects his Android phone to the car and the phone casts Android Auto experience to the car’s screen. Andy can now put his phone down and use the familiar car controls, steering wheel buttons, console dials, and touch screens, to control Android Auto. It looks and feels like it’s part of the car. But all the apps we see here are running on Andy’s phone, which means that the experience gets better when Andy updates his apps or gets a newer, faster phone. This also means that Andy has a personalized experience that he can bring with him into any compatible car.

The first thing Andy sees is the overview screen, which shows personal and contextually relevant destinations, reminders, contacts, and music from Google Now and other apps. One tap and he is navigating or listening to his favorite road trip mix. Andy, why don’t you play us some music.

[Music]

Let’s look for a seconds at Play Music. It has been adapted to have simple, glanceable controls for the car. Andy has access to all of his curated play lists, radio stations, albums, and artists, and to all the key features in Google Play music. He can also use voice or the steering wheel controls to control the music in the car, keeping his hands on the wheel. Fantastic. Of course, Android Auto needs great maps and navigation. So let’s show you Google Maps.

We all love Google Maps because it’s fast, accurate, updated, and it seems to know where everything is. In Android Auto, drivers have access to all of their favorite maps features: great local search, personalized suggestions, live traffic, and, of course, turn-by-turn navigation. And Google Maps for Android Auto is even more powerful because it is completely voice-enabled.

Andy, why don’t you take us for a drive?

“How late is the De Young museum open today?”

Google: De Young museum is open from 9:30 a.m. to 5:15 p.m. on Wednesday.

“Good, I can go there. Navigate there”.

“Google: Navigating to De Young museum… head toward Fourth Street Northeast on Main Street and 600 feet, use any lane to turn right onto Fourth Street”.

So Andy was able to start navigation without ever entering an address or taking his hands off the steering wheel. During navigation, instructions are spoken, as you heard, and displayed on the screen in a material card that floats above the map. Great. So that’s music and navigation.

What’s next?

Let’s show you voice-enabled 5messaging.

Google: New message from Hiroshi Lockheimer. Here it is: Andy, are we there yet?”

As you can see, incoming messages show up as heads-up notifications. So Andy can still see the upcoming 1turn in maps. When he’s ready, he can just use the steering wheel voice button to reply.

“Andy Brenner: Reply.”

“Google: What’s the message?”

“Andy Brenner: I have no wheels”.

“Google: Here’s your message to Hiroshi Lockheimer: I have no wheels. Do you want to send it?”

“Andy Brenner: Sure”.

“Google: Sending message”.

So we’re really excited to bring these great experiences into the car. But we also want you, our developers, to come along for the ride. We know it’s not easy to build apps for cars today. There are dozens of different car platforms, input controls, and user interfaces. There is no central way to distribute your app or keep it updated.

Wouldn’t it be great if building an app for the car was just like building an app for your smartphone or tablet? Well, we have good news for you. The road ahead is brighter, and today we’re announcing the Android Auto SDK so that you — we thought you’d like that. — so that you can just focus on making great apps for the car.

We’re starting with a full set of APIs for audio and messaging applications. First, let’s talk about audio. We worked with a great set of developers on a prerelease version of the Android Auto SDK to develop some great audio streaming apps that let you listen to music, internet 9radio, news, sports, and podcasts on the go. You can try these apps out live in our demo cars right outside.

Next, let’s talk about messaging apps. Andy showed us earlier how he can send text messages using Android Auto completely with his voice. Well, we’re opening this up to your messaging apps. So using these APIs, your apps can notify users of incoming messages and allow them to respond using voice. And this is the same API we’re using for notifications and remote reply on Android Wear, with just a few lines of code, you can let users know on their wrist or in their car. It’s really, really powerful.

So we’re really excited about Android Auto, and we think we’ve found that better way. But I know what you’re all thinking: when does the rubber actually meet the road? Well, we’re happy to say that you won’t have to wait long. The Android Auto SDK will be published soon, and the Android Auto experience will be available to users with the public “L” release later this year. And the excitement in the auto industry has really been growing.

Today, we’re happy to announce that over 40 new partners have joined the Open Automotive Alliance. And over 25 car brands have signed up to ship Android Auto in the near future. What’s more? The first cars with Android Auto will be rolling off dealer lots before the end of this year. So that’s just a peek at Android Auto, an Android experience that’s been redesigned for the car, with all of the apps drivers know and love, through an interface that’s built for driving.

Now I’d like to welcome Dave Burke back on stage to tell us about android in the living room.

Android in The Living Room

Dave Burke – Director of Engineering, Android

Thanks, Patrick. It’s pretty cool to see what you guys are doing in autos. But some of us don’t actually have a car in our living room, wheeled or not. So I’m going to talk about a different form factor, and that’s TV.

So TVs are fast becoming smarter and more connected. And really they’re becoming computing devices in their own right. So we see a great opportunity to bring some of the strong capabilities of Android, such as voice input, user experience, and content, to the largest screen in your house.

Now, in some ways, TV space is not too dissimilar to the mobile space in 2006. Each TV manufacturer has a different OS with different APIs and programming model, often with limited developer tools. And the cost and friction to developer service to run across multiple TVs is too expensive. As a result, smart TVs are typically limited and not competitive with their mobile cousins. So we wanted to go and change that.

Today, we’re announcing Android TV. So this isn’t a new platform. That’s kind of the point. We’re simply giving TV the same level of attention as phones and tablets have traditionally enjoyed. We want you to be able to leverage your existing skills and investment in Android and extend them to TV.

There’s now one Android SDK for all form factors. Now, remotes are a core part of the TV experience. And Android TV requires just a d-pad with voice input. And that can be made available as a hardware remote control, as a game controller, or even a virtual controller on a phone or tablet.

So today, I’m going to use the Android TV app on my phone. And the best way to understand Android TV is to just see it in action. So what you’re about to see is hot off the press and really just an early look at the TV support that we’re adding to the “L” developer preview.

Okay. So let’s start with the most integral part of television, Live TV. So in “L,” we added what we call the TV input framework to Android. So it enables Android-based TVs to handle video from sources such as HDMI, TV tuners and IPTV receivers. And the UI provides a unified view of your channels in a familiar channel-hopping UI with the channel information on the top.

Now if you want to do something different, just like every other Android device, you press home. And you’ll notice that home overlays on top of the content. So I can keep watching while I browse. But unlike phones and tablets, where the behavior is more task-based, we designed home to be a super simple lean-back experience, because TVs, unlike computers or mobile devices, they’re primarily entertainment interfaces. Users don’t expect or want complexity from their TV.

So the launcher presents you with a set of content recommendations at the top. Floating over the UI, using that familiar material theme. As I scroll down, you get immediate access to your applications ordered by how often you use them. Scroll down again, you get access to your apps, also ordered by usage order.

Now if I scroll back up to the content recommendations, you can see this is a quick way for me to watch content, the recommendation system is completely open. Any app can publish to it and it’s ranked according to your usage patterns. So for example, I am currently binge watching Game of Thrones. I am actually and it will automatically be presented with a recommendation for the next episode like so. So that’s the home.

Let’s talk about search. Today, people regularly take out their phone to search for something to watch. With Android TV, we decided to build a core search functionality directly into the experience, powered by voice. So, for example, I can just simply say, Breaking Bad. Google will interpret the results, get me a result for the popular TV show. Now, with one click, i can then watch it in Google Play, maybe some TV, or any other service that I have installed.

If I scroll down, you’ll see information on cast members. Scroll down again, I get related search terms. Also YouTube clips at the bottom. Now, I can also pivot on cast members. So, for example, I can click on Anna Gunn. I’ll get that nice material transition, I get information on the actress. Scroll down, I can get movies and TV shows that she stared in, even YouTube clips of interviews with the actress.

Now the power of Google comes into its own for more abstract queries. For example, I can just say Oscar-nominated movies from 2002. Google will interpret that query, and, of course, get me all my Oscar-nominated movies.

So from here, of course, one click, I can watch it. Now, search of course is backed by Google’s Knowledge Graph. So I can ask a question. So, for example, I could say something like who played Katniss in the Hunger Games? And, of course, the answer is.

“Google: Jennifer Lawrence played Katniss Everdeen in the Hunger Games”.

And again I get maybe some TV shows she’s involved in, other related queries and YouTube clips. So that’s Google search tailored for your TV.

Let’s now take a look at some of the applications. So let me launch Google Play movies in TV first. Now developing for TV means creating a ten-foot user experience, because that’s typically how far you’re standing or sitting from the screen. In “L,” we’ve expanded the platform to support a lean-back experience with new framework classes that help you to quickly and easily build fluid cinematic ten-foot user experiences. Our play movies team was able to take their existing tablet app for Android, quickly add the lean-back classes on top of it to produce a great TV experience. They now have the same APK for TV, phones, and tablets.

So what you see here is the browse fragment, part of the lean-back classes, you got nice, slick animations, you got those bold, lean-back — excuse me, material colors. If I then dig into a TV show, I can get more information. This is showing our details view. By the way, if you don’t have your remote handy, you can always use your Android Wear watch as a d-pad. For example, let’s try this out.

So we’ve created a little wearable app. So from here, I can actually — I can interact with the TV. So I can go back. Let’s go from the watch. Now you see me. I think that we could show. So I click that. And then we can start watching it. Okay. So that’s not loading. That’s fine.

Okay. So let’s move on. So to wrap up, all of these lean-back building blocks are ready for you to use and customize in your own applications as you see fit.

Okay. So next, let’s talk about games. People who have traditionally not thought of themselves as gamers frequently download and play games on their phones and tablets. And, in fact, three out of four Android users are playing games, which has helped make the Play Store one of the biggest catalogues of games in the world.

With Android TV, we enable you to take your games to the biggest screen in your house. And the games are getting really good. So, for example, let me fire up one of my favorites, which is Leo’s Fortune, with this game pad. And this is a really fun game. It’s kind of typical of a modern Android game, it has got gameplay, it’s got great graphics, fast and fluid. I also like it because our lead UX designer’s name is Leo. He’s not quite as cuddly as this guy, but I can definitely see him in a handlebar mustache.

Anyway, so another advantage of Android TV is you can tap into Google Play games to share achievements and leaderboards with friends. You can even play multiplayer games with your friends playing from any device. So for example, I can launch NBA Jam, which is a really great game, and play a multiplayer game with my friend Allen here 4on the sofas. Let’s try it out. So he’s playing on his tablet. I’ve got my game controller. Let’s try this. So we have a bet on the first person to score is me. Okay. First one to 2. Okay. Let’s go. So the first person to do has to buy the other person a beer tonight. Let’s go. Hopefully, I can actually score. I’m from Ireland, so I don’t really understand this game. Let’s try this. Oh! Okay. Did you get 2?

Okay. Last one. Last one. Bear with us. I want free beer tonight. Oh! Yeah! All right. That’s a good thing there’s a free bar at the after party, Allen. Okay. I hope.

Okay. So that’s games. Now, sometimes you just want to cast or send content from your movie — such as movies or music from your phone. So Android TV includes full Google cast support. So you can use it just like a Chromecast. In fact, Android TV enables us to bring Google cast to more TVs.

So, for example, imagine Allen’s visiting me house and he wants to share his new favorite Jam with me. He can fire up play music on his phone or tablet and cast it directly to my TV. And, hopefully, his music will appear. I’m always really nervous about what he’s actually going to play. So I’m not sure I even want to see this. Okay. That’s pretty innocuous. That’s cool. Great. So that’s cast.

Now, to distribute your applications, we’ve designed a very TV-centric play store experience. I’ll just fire it up here. And the “L” developer preview comes with a sneak peek of TV apps and games to showcase the platform and you’ll recognize big names like Netflix and TED as well as some great casual and multiplayer games. The store will open officially in the fall with the launch of “L,” packed with some of the best content available today for Android, tailored, of course, for TV. So that’s a quick overview of Android TV.

Android TV is ideal for multiple device types, so everything from built-in televisions to set-top boxes to streaming boxes and gaming consoles. We’re working with silicon vendors across the industry, including everyone from Marvel to Intel and more.

Today I’m super excited to announce of the entire 2015 HD and 4K smart TV ranges from Sony and the 2015 ranges of Sharp and TPVision Philips will run on Android TV.

We’re also seeing activity in the Pay TV space with Wix, SFR, LGU Plus adopting Android TV, and we’ll also be seeing streaming boxes powered by RAZR, Asus and others launching this fall. And the Unreal Engine demo that I showed you earlier, that was running on Android TV with Nvidia’s TK1 reference design capable of console style gaming and we expect to see TV products based on that hardware in the fall, too.

Finally, to bootstrap the ecosystem today, we’re making a development kit we call ADT1, available through a sign-up page to application developers like yourselves, so you can start developing for TV today. In fact, all the demos you saw were running on ADT1.

So to learn more about Android TV, you can visit developer.android.com/tv. And I mentioned earlier that Google Cast is a core part of the Android TV experience. And as cast gets better, so does Android TV. So to learn more about some of the exciting new developments in Google Cast and Chromecast, let me hand over to Rishi. Thank you.

Rishi Chandra – Product Manager, Chromecast

Thanks, Dave. So I’m going to give an update on Chromecast. We launched Chromecast last July to deliver a new TV experience that was both simple and powerful. By using devices you already know and love, your phone, your tablet, and your laptop. And for the first time, you had an experience that just worked. All it took was a simple press of a button from your favorite app.

And we’ve been really happy about the positive reaction we’ve gotten from both the press and consumers. We’ve sold millions of devices and consistently outsell all other streaming devices combined in major retail channels like BestBuy.

And recently, we’ve been able to replicate that success in 18 countries. Today, Chromecast is a top-selling electronics device on Amazon in the U.S., UK, France, Japan, and Canada.

And as sales have ramped up, usage per device has increased 40%. Already, YouTube sees more active engagement on Chromecast than any other TV streaming product. The model is working.

Now, Chromecast is just the beginning. We want to build an ecosystem of Google Cast-ready apps and Google Cast-ready devices. So that’s why we’re excited that Google Cast support is coming to Android TV devices as devices roll out later this year.

So let’s talk a little bit about developers. We launched Chromecast with five content apps. And since then, we’ve brought on board many of the top content apps from around the world, including the BBC iplayer and the Watch ESPN app, which came just in time for the World Cup to kill all productivity of the Chromecast team. Luckily, we have a Brazilian VP.

Google Cast is designed to work with the most popular devices you find in a home today. In fact, in the last 30 days, almost 50% of Chromecasts were used by devices for multiple platforms. That’s why in February, we launched the Google Cast SDK across Android, iOS, and Chrome. So now any developer can take their existing mobile or web app and extend it right to the TV.

In just a few months, we have over 6,000 developers registered who are actively building over 10,000 Google Cast apps. So starting to see a lot of momentum. Now, as more and more apps are coming on board and integrating with the SDK, we want to make it even easier for consumers to find your Google Cast-ready app. So today we’re announcing new discovery experiences on Android, iOS, and Chrome. Consumers can find these at chromecast.com/apps or from the Chromecast app on your phone, tablet, or laptop. This is one of many improvements we’re making to Google Cast to make it even easier to use. For example, one of my favorite features of Chromecast is that it works with my friends and family, because they can use their own devices.

Well, today, we’re announcing a new feature to make that even easier by allowing others to cast through to your TV without needing to be on the same Wi-Fi network. So I have AK here on stage with me. And let’s say AK is over at my house and he wants to use his phone to share or cast one of his favorite YouTube videos to my TV. Now, normally, I need to share my Wi-Fi password with him. But sometimes they’re long and complicated or I don’t remember it. In this case, I just don’t trust him with the password. But I still wanted to see his great video.

Well, with this new feature, AK can still take out his phone and open up the YouTube app. And you’ll see on the top right he’s only connected to the cellular network. But you still see the cast button. Simply press the button, and it will allow him to connect to your devices. Now, in a few seconds, his phone will connect to Chromecast through the cloud. And that’s it. So now he can control the video, pause the video, play the video, just as if he was on my same wi-fi network.

So how do we do it? We’re using a variety of different technologies which allows us to authenticate users in the same room as the Chromecast. And if for whatever reason we can’t automatically detect or authenticate you, we’ll ask for a pin that will always be present on the screen. Check this part out. This is my favorite part of the video. That’s pretty cool.

So now you can invite all your friends over to your house, kickback, and let of them cast to your TV without any friction or hassle. It makes Chromecast an even more awesome experience. Now this is an opt-in feature. So you always have control over who can cast to your TV. We’ll be rolling this out to all Android users later this year. And developers will get this for free via the Google Cast SDK.

So we also spent a lot of time thinking about new use cases for the television. For example, today, lots of people talk about the five hours per day people watch TV. What about the 19 hours per day your TV is just a blank, empty screen? With Chromecast, we want to make use of this large, beautiful canvas. So we started with something really simple, a feed of scenic images that’s been gradually expanding over time. Today, we’re announcing the new Google Cast ambient experience. And we call it backdrop. It just takes a minute to set up, and you can do it from any iOS or Android device. In this case, we’re going to use an iPhone 5 for the demo. And we have the feed accelerated just for the demo purposes so you can see more images.

So let’s go ahead and open up the Chromecast app. And you’ll see backdrop’s a new option in the app drawer. And from here, you can personalize the feed to match your own interests and tastes. One of the top feature requests we’ve gotten is the ability to add your own personal photos to the ambient feed. Well, now you can.

So if we turn on photos, you can select from one or multiple Google+ photo albums. And in a few seconds, you’ll start seeing my personal photos show up right on the TV screen. It’s that simple. Your TV is now the largest picture frame in the house. Grandparents everywhere are going to love this feature.

So we also have a lot more topics. We want to give you an infinite source of great and beautiful images. One of my favorites is one called Places, which brings geospatial images from Google Maps to your TV. These satellite photos give you a totally different perspective on how you can see the world. It’s really amazing. Leave it on for a really long time.

We also have other topics like news and lifestyle and Google+ featured photos. One of my other favorite topics is art, which includes artwork from famous museums and collections from around the world, including the Getty Museum and The National Museum of Women in the Arts. And now my TV’s beautiful set of paintings.

Every topic is curated to make sure we’re showing high-quality and safe images. And every user has the ability to control which topics show up on the TV. Now, generally, I’m going to leave this in the background of my house. But let’s say one of the images catches my eye and I want to learn more — a little bit more info about it. I can use Google voice search to always learn more.

“Google: What’s on my Chromecast?”

For every topic, we’ll show a synchronized card. We’ll show a synchronized card that will show you relevant information and relevant actions. So in this case, I can learn more about any artists of the painting. Backdrop brings your TV to life even when you’re not actively using it. And we’re looking forward to working with third-party developers to actually bring in their own topics into the feed. Backdrop will roll out to all Chromecast users later this summer.

So finally one of the big advantages of the Google Cast model is that we can start thinking beyond traditional use cases of the TV like video. For example, your phone and your tablet have unlimited possibilities, and there’s many times you want to bring those experiences and extend them to the TV. Well, to help accelerate that shift, we’re launching a new Google Cast feature that allows you to mirror any Android device to your television.

So to start mirroring, just go to your Chromecast app on your Android device and select cast screen. And from here everything on your phone will just show up on the TV. It’s that simple. No cables or wires needed.

Now we built our own protocol to reduce latency and frame drops, so experience feels really natural and smooth. And now you can share anything on the big screen. So let me give one example, a real life example I had a few months ago, using Google Earth.

I was planning a family trip to Maui later in the year and I wanted to show my four-year-old daughter where Maui is on the map. So instead of huddling around my small phone I ended up opening Google Earth and projecting it on the TV. So let’s type in Molokini, one of my favorite snorkeling destinations in Maui. And as you can see it’s a totally different experience to navigate the world with my daughter on the biggest screen in the house. You can play with this for hours.

So we’ll do one more fun thing. We’re actually going to open up the camera app and we’re actually going to mirror in live action, which is a little nervous, in live action, all of you on the big screen. It’s a lot of fun. So we’re working with a large variety of devices from Samsung, from Nexus, HTC and LG, with many more devices coming soon.

We’ll be rolling out this initially in beta as devices get updated to the latest version of Google Play services, which will happen over the next few weeks. As you can see, we’re really excited to show how the cast model can change how we think about entertainment in the home and we look forward to innovating all of you to make that a reality.

Now I’ll pass it back to Sundar.

Sundar Pichai – SVP, Android, Chrome, And Apps, Google

Thanks, Rishi. So we have talked so far about wearables, your car and the television in your living room. Another very important device in our lives is our laptop. Our journey here began with Chromebooks. We are seeing tremendous momentum here. We started the journey with one reference device, the CR-48, which we launched about three years ago.

A year later with Samsung and Asus, we launched two devices in two countries. Fast forward to today, we have eight OEMs making 15 devices with many more on the way in 28 countries. Users really love the fundamental insight behind Chromebooks, speed, simplicity and security. All 10 of the top 10 highest rated laptops today in Amazon are all Chromebooks.

We are seeing tremendous traction in education as well. Just this year alone the number of Chromebooks sold to K-12 schools in the United States has grown by 6x and we are investing a lot more in this area. So let me talk about how we are evolving the Chromebook experience.

As I said earlier, users almost always have a smartphone with them, including when they are using a Chromebook. So we want to increasingly connect these experiences together so that they have a seamless experience across their devices. Let’s take a look at what we can do. Dave Kan is going to help me through with some demos. Dave talked earlier about how with the L release you can unlock your phone knowing that it is with you based on what we are doing with the APIs in “L.”

Well if we can unlock your phone we can also unlock your Chromebook. So every time you approach your Chromebook and your phone is with you, we will automatically unlock your Chromebook and sign you into your favorite applications and services so it works seamlessly.

We have already added Google Now notifications, so the same Google Now cards which you see on your phone you see them on your Chromebook and we’re adding a few more things. Let’s say you get an incoming phone call. You will start seeing those incoming call notifications on your Chromebook. If you get a text message, you would see those text messages on your Chromebook as well. And I recently had this experience, my phone was in my pocket and running out of battery and my Chromebook popped up a notification and said, your phone is running out of your battery. Simple, delightful experiences to connect your phone and your Chromebook.

As we started working on this, one of the things that struck us is wouldn’t it be nice if you could get some of your favorite Android applications which you love on your phone on your Chromebook? So this is a difficult challenge technically, so we’ve been working on this project for awhile. Our goal is to bring your favorite Android applications in a thoughtful manner to Chromebooks. We want this to be intuitive for users. These applications were built for Android for the phone so we want them to work when there is a mouse, keyboard and touch event et cetera.

For developers we want this to work with as little

modifications as possible. So we’re in the early days and we’re going to show you a preview today and Kan is going to help me with this. So Kan is going to pull his tablet up and what you’re seeing on his tablet is Evernote. This is one of his favorite applications and I think he’s planning a birthday party there.

Let’s switch over to the Chromebook and the same Android application is now available in the launcher on your Chromebook. So you can click that and you get the exact same application. And because you’re on a Chromebook you can start making changes to it, you can go to the web, copy, paste, everything just works. We have ported that Android application to run within your Chromebook.

Let’s give you one more example. We’ll pull a canonical application, an application you see on your phone, like Vine. We now have that application running on your Chromebook Pixel as well. Kan is going to browse through the World Cup channel and you can see how the experience feels native and intuitive on a Chromebook and because the app has access to some of the underlying device APIs, it has access to the camera API, Kan can actually take a clip of himself just like a Vine user and post it straight from his phone.

The final application we want to show to you is Flipboard. It’s a beautiful immersive experience and the Android version of the app now comes alive on Chromebooks and you can see how beautiful and immersive this experience is. We are very, very excited about bringing important favorite Android applications for users straight on their Chromebooks so that they can get an even more connected experience. And we are working on bringing our experiences across both Android and Chrome together so that it works in a delightful way for our users.

So far we have talked about bringing our platforms Android and Chrome to all the connected devices in your life so that we can create a great experience across all of them. But there is another very important environment we want to talk to you about as well.

Most of you spend a lot of time in your workplace. We’ve always had this funny insight at Google that it is the same person at work who is — and it’s the same person at home. That’s a picture of Matt, which is his work badge. In his spare time he’s a lonely planet traveler and he writes books for them. So you can see people have very, very different contexts, yet your computing experience is very fragmented. The way it worked in the PC world is companies gave you a separate laptop for work and a laptop for — you had your own personal laptop. The experience was disconnected. And that model starts breaking down when you start thinking about phones.

No one wants to carry two phones around. So what we are doing is with the “L” release in Android, we are adding a whole set of APIs to unify this experience. So we’re bringing both your experiences so that as a user you can have one experience and both your personal applications and corporate applications can live on the same device. We are doing it thoughtfully by providing underlying data separation so all your personal data is isolated from your corporate stuff and vice versa. So we are providing full data isolation and security which enterprises care about.

As developers, there is no modification needed of your existing apps. All your apps will be available through Google play and companies can buy them in bulk and deploy so you can reach many more users. These are all available in “L,” but we are wrapping up many of these features as a separate app so that it will also work on prior versions of Android.

And finally, Samsung has done a lot of important work in this area with Knox. And we really want to thank Samsung, they’re contributing all of their work in Knox to the Android platform so that we have one consistent story for surprises across Android.

Can I get the next slide? So we’re also working with major partners, all major OEMs, names you would recognize, in the fall, to have a certified Android for work program by which we can bring these devices to companies with guaranteed updates and full security. So we are very excited by this journey.

As we bring Android for work, one of the important use cases we care about is productivity. Documents collaboration, which is why we’ve invested a lot in Google Docs and the whole suite of editors. We’ve always had great mobile webs for Google Docs and Sheets and today we’re also announcing Google Slides so that you can create and share presentations straight from your mobile devices.

One of the common use cases we run into in companies when people use this, is they run into Office files. It’s a common experience for all of us and we want to make sure as we bring Android for work, Office files work seamlessly. So we acquired QuickOffice and we’ve been hard at work integrating QuickOffice into Google Docs and today we are announcing native Office editing built within the Google Docs suite of editors.

Let’s take a look. Kan is going to show how this works. I’m sure you’re all familiar with getting emails in which you get a Word file. And so if you get a Word file in your email, you can click on it. In the past we would convert it into a Docs file, but we don’t anymore and so what you see is a native Word file handled straight within Google Docs and this works for Sheets and Presentations as well. And Kan can make edits to the document and most importantly when he saves it, it saves back as a Word file so that he can send it back to people. So it works seamlessly.

You can always up convert this to Google Docs so that you get advantage of world class collaboration features. One of the common features we get requests on is red lining. So we have done great work to bring modern collaborative approach to red lining and the feature is called Suggested Edits. So just like Google Docs always does, people can add their comments and very easily you can review and accept changes.

So with Suggested Edits, we have modern collaborative features and with built-in native Office, we think our productivity suite is great for the workplace. All our productivity files live in Google Drive. We launched Google Drive about two years ago and today we are very excited to announce Google Drive has over 190,000,030-day active users. These are not registered users, these are 30-day active users and the number has grown over 85% just last year alone. We are bringing all of Google Drive for work functionality to companies as well. And what we are doing is we are encrypting the data both during transit and at rest on our servers. We are providing enterprises full APIs, audit and analytics APIs based on how their employees are using their data across their company.

And finally, it is hassle-free. Unlimited storage for just $10 per user per month. So the combination of Android for work, our Google Doc suite and Google Drive for work really offers a comprehensive suite for companies and joins Google apps and Chromebooks. And as companies are thinking about moving away from traditional Windows architecture we are seeing tremendous momentum. 67 of the top 100 startups have gone Google. 58% of Fortune 500 companies have gone Google. And 72 of the top 100 universities have gone Google.

So we have talked about how we are bringing our platforms Android and Chrome across all your devices not just for your personal life, but also for your workplace. We’re going to switch gears now and talk about how you all as developers can build success on top of our platforms.

In fact, we hear amazing stories everyday from corporate startups to students who act as developers and create amazing experiences on top of our platforms. We’ve put together a video so that you can take a look.

[Video Presentation]

I was incredibly moved when I saw that video the first time and even more excited. All the folks on the video, including those middle school students from Resaca School in Texas are joining us in the audience today.

Thank you. We know these examples are just the tip of the iceberg and we care deeply about evolving our platform so that you can continue to create these amazing experiences. So we’re going to talk about that next.

The first is how you can build and scale your applications on top of Google Cloud platform. We’ve been, within Google, building world-class infrastructure and running large scale services like search, maps, YouTube and Gmail and we are very excited, we’re bringing the full power of the Google Cloud platform to developers like you.

To talk about that I’m going to invite Urs Holzle.

Urs Holzle – SVP of Technical Infrastructure, Google

Thank you. Thank you, Sundar. Hi everyone. I am Urs Holzle and my team and I build the Google Cloud platform. And everyday we go to work excited to see what you all are doing with it. And literally everyday there’s hundreds of thousands of developers building applications that scale to hundreds of millions of users. And they keep their teams small and they can focus on what they do best because we run the rest of the infrastructure for them.

Let’s have a quick look on what about — what the Google Cloud platform is about. First of all, of course we have VMs with compute engine and they are best in class performance and price and you can run whatever code you want. But if you don’t want to bother administering machines we also have app engine and app engine makes it incredibly easy to write really high scale applications. That’s how Snapchat got to incredible scale without having a single back end developer.

On storage, we offer many options, including of course SQL as a fully managed service. NoSQL, which we invented by the way. It is running billions of queries every hour. And then of course object storage that scales to exabytes. But you cannot just store data, you also want to analyze it and in our platform you have many tools that make it easy for you to analyze data sets without worrying about scalability. For example, with BigQuery you can stream hundreds of thousands of records per second into the cloud and query them interactively with SQL.

And beyond technology we also lead the industry in price and performance. And as hardware gets cheaper we pass on these savings to you so you see Moore’s law in the cloud. And moreover you get great sustained usage discount without having to sign up for contracts, without having to make upfront payments or without having to forecast your utilization for the next three years just to get great pricing.

But of course what really matters to us is seeing what you developers do with our platform. And whether it’s Netflix, for storage, Wix, Web Filings, Khan Academy who run their entire business on our cloud, there’s a lot going on. And I am going to tell you just two things about two recent new customers.

First, Secret. Secret where a single back end developer, built an application that in two months saw over a thousand fold increase in traffic. And they came to Google because they needed a platform that can handle this kind of hyper scale growth without much hassle. And that’s exactly what they got. Or maybe you’ve seen the debut of Rising Star on ABC last Sunday. It’s an interactive music competition where millions of TV viewers vote in real time on the winner. Screenz, the company behind Rising Star is using compute engine and BigQuery for the instant voting app and it’s been battled tested to run 1.3 million queries per second.

And at the beginning of each application is a developer who wants to get their idea out to the market as quickly as possible. No matter what it is. But that’s why developer productivity is actually very important to us. We try to make our cloud not just highly functional, but also really easy to use. And as you’ve seen the best mobile apps come with an intelligent back end that’s in the cloud. And there’s no better place to build that back end than on Google because we have the performance and we have the tools that make it really easy for you to build those applications with small teams and little focus on operations.

And today I’m very excited to show you four new tools, very, very cool new tools that make it really easy to understand your server side application. So please welcome Greg DeMichillie, who is going to show you some demos.

Greg DeMichillie – Director of Product Management on Google Cloud

Thanks, Urs. As part of showing you some of the new features that we are showing for the Cloud platform, I’ve built a sample application here and this is about as peaceful an application as I know how to build. It allows me to record walks that I’m talking. So this application allows me to record a walk I took through the city, share it with my friends, let them experience the same walk I took and then share comments.

Now since this whole purpose of this application is to share information, obviously I need to store that data somewhere other than my device. And how do I do that? The answer is Cloud Save. Cloud Save is a new simple API that lets me save and retrieve per user information. Now that could be application data, it could be user preferences or setting and I do this with no server side coding. I just do a few lines of client side. And with that my data can be stored in the cloud, it can be retrieved and synchronized to other devices, it can be available offline so that my user doesn’t have an internet connection, they can still walk through this walk.

And I’m a cloud guy, so what really gets me excited is that this data is stored in Cloud Data Store which means I have the full access to the cloud platform. I can query that data with BigQuery. I could build web applications that use that data using App Engine or Compute Engine and that’s in fact what I’ve done in this case.

Let me show you the web version of walk share accessing the same walk data. So this is the web version. I took a nice walk out by the Golden Gate Bridge yesterday and recorded it. I can expand the window and I can see the comments. Now if you look at the comments right away you can see that there’s something a little funky going on. I have code here to replace common character sequences with emoticons, but if you look at that they’re sort of running amuck. I’m putting emoticons where they don’t belong. So I’ve got some sort of bug that I need to find.

So I’m going to flip over here and look at the source code from GitHub for this application. Now this application is running in production. It’s not running on one server, it’s running on tens or hundreds of servers. And there’s no way to debug an application that’s running in that kind of environment. At least not until today. So I’m going to click into the debug mode, I am going to click down here and I am going to put a watch point on this line of code.

Now the platform is watching all those servers until one of them actually hits that line of code. So at this point, there’s people hitting the traffic. It’ll take a few seconds. Take a few seconds. Okay. We’ll go and refresh it just to make sure it hits it.

Oh, no. Let’s try it again. I’m determined on this one. We’ll find that line of code. Going to debug. Set the watch point. Nope. Oh, there it went. I just had to wait long enough. Patience pays off. You’ll notice, however, I didn’t hit the right comment, because this doesn’t show the bug. So I’m actually going to use a conditional watch point and say, “comment.commenter name is Rachel.” And now I’m going to wait a few seconds. And we’ll actually hit the breakpoint for that comment, call stack local window. And sure enough, this is the one that’s having the problem.

So to fix it, I’m going to go up, and I see here’s the code that was being called. I can switch into edit mode and very quickly change this regular expression to be the correct one and learn that I should always have Urs do my code reviews, because he never makes regular expression errors. So that’s cloud debugging.

But when I was looking at that application, I also noticed that the application seemed a little slow to me. It seemed like it wasn’t really responding right. How do I find out what’s happening in production there? Another new feature of the platform is Cloud Tracing.

Cloud Tracing gives me a tracing view of all of the various requests my application was processing and how long they took. I’m going to drill in for comments, and sure enough, when I look at this, I’ve got some queries here that are taking 200 milliseconds. That really seems long. I’m going to drill into one of them. And I get a view of all the service calls that went into that request. So what appears to be happening here is that I’m doing a bunch of data store operations in sequence instead of doing them in parallel. And that’s not really the best practice data store.

So to fix that, I’ll switch over to the code for that. I’ll edit it. Sure enough, this whole block of code is in a for loop. That’s not really the best practice for getting the best performance out of data store. So we will delete all that. And I’ll paste in the proper version of the code, which actually, then does all of that as one operation. I can commit the change. It gets rebuilt, I’m in production. I’ve fixed my performance problem.

Now, I fixed it for one user. How do I know I fixed it for everybody? Well, it turns out Cloud Tracing also gives you the ability to have reports. So this is a report that I’ve done that compares the latency before and after that change. The blue part is before I made the change. The orange part is after I made the change. So, sure enough, I can see that my curve shifted to the left. My application is in fact faster, not just for one user, but for every user.

So the last thing I want to show you is Cloud Monitoring. Last month, the team from Stackdriver joined Google. And I’m really happy to show you some of the integration we’ve been working on. One of the hard parts about operating a service in production is building up the monitoring and alerting that you need in order to stay on top of things.

With cloud monitoring, you, of course, get the basic infrastructure type monitoring you’d expect: disks, VMs, that sort of thing. But what’s really powerful is, you also get service-level monitoring. We automatically detect what services you’re using and give you default dashboards and monitoring. And it’s not just for Google services. We automatically detect over a dozen open source packages, such as Redis, in this case and automatically give you intelligent default monitoring for those. So with no additional work, I get a dashboard that helps me stay on top of what’s happening in my app in production.

Finally, as everybody who runs a service knows, the last thing you want is your customers to be the first ones to find out that you’ve got a problem. You want to have alerting in the event of problems. Cloud monitoring includes alerting. So I can go and set custom alerts on any of a variety of metrics whether it’s on app engine or compute engine or Redis. In this case, I have an alert set up to tell me if Redis exceeds its memory threshold for more than five minutes and I have a choice of how I want to be notified. Do I want to be emailed or paged or sent a text message? So that’s just a very fast preview of four new developer productivity features coming to the cloud platform.

And with that, I’ll turn it back over to Urs.

Urs Holzle – SVP of Technical Infrastructure, Google

Thank you, Greg. So you just saw four new features that make it much easier to build and understand your back-end, Cloud Save securely saves and synchronizes across devices. No service-side code, just a few lines of client code. Cloud debugger gets you desktop debugging in your server. Debug a live production app with live traffic. Get local stack traces and so on and so on. How cool is that?

Cloud tracing then shows you your latency statistics across different groups. You can compare before and after. And cloud monitoring gets you intelligent monitoring with almost no setup, including, for many popular open source packages.

But now let’s go from code to data. Information is being generated at an incredible rate. And, of course, we want you to be able to analyze that information without worrying about scalability.

And today, even when you’re using MapReduce, which we invented over a decade ago, it’s still cumbersome to write and maintain analytics pipelines. And if you want streaming analytics, you’re out of luck. And in most systems, once you have more than a few petabytes, they kind of break down. So we’ve done analytics at scale for a while, and we’ve learned a few things.

For one, we don’t really use MapReduce anymore. It’s great for simple jobs, but it gets too cumbersome as you build pipelines. And, really, everything is an analytic pipeline these days. So what we needed was a new analytic system that scales to exabytes that makes it really easy to write pipelines that optimizes these pipelines for you and that lets you use the same code for both batch and streaming analytics.

Cloud Dataflow

And today we’re announcing just that with Cloud Dataflow. Cloud Dataflow is the result of over a decade of experience in analytics to simply fully manage service, so no machines to worry about. You can create data pipelines for ingesting, transforming, and analyzing arbitrary — arbitrarily large data sets, both in batch and in real time mode. And to see Dataflow in action, I have asked no other than our very own Eric Schmidt to help me look at something that’s happening right now, namely, the World Cup. So please welcome Eric Schmidt.

Eric Schmidt: Thanks, Urs. This demo is about performing sentiment analysis of World Cup matches with Cloud Dataflow. We’re going to analyze millions of millions of tweets per match and calculate negative or positive sentiment per team and correlate the sentiment to match data. I’d like to thank the developer relations team at Twitter for the support in using the Twitter and data APIs for this demo.

Cloud Dataflow is an SDK and a managed service for building big and fast parallelized data analysis pipelines. You write a program as a logical set of data transformations to specify your analysis. You then submit the pipeline to the Dataflow service, and that handles all the optimization, deployment of VMs, scheduling, and monitoring for you.

Now, here’s the code for my sentiment pipeline. The Dataflow API provides a simple mechanism for you to add one to many transforms to your data. The first transform extracts a real-time stream of JSON data from Cloud Pub Sub. Now this pipeline is running in streaming mode. However, Dataflow can also be run in batch mode, meaning that I can point this exact same pipeline of – at archive data, say, in BigQuery, and produce the exact same analysis. One pipeline, batch or stream.

Now, let’s drill into the second transform, Tweettransformer. Tweettransformer is responsible for the core transformation and mapping of my data. I did serialize the stream of JSON into a tweet object. I translate it, if it’s needed, using Google Translate API. And I score the sentiment with a third-party service from Alchemy API.

Now, the part of the syntax that you see here is what parallelizes the processing for each step. This parallelization is optimized for you by the Dataflow service across processes and machines. Now, finally, I apply another transform to calculate the average for all the tweets in a three-minute sliding window. Dataflow provides powerful built-in primitives for doing MapReduce-like and continuous computation operations.

Two lines of code, two lines of code to create a sliding window and to average all the tweets within that window, reducing a massive collection of data down to one record per minute per team. Dataflow handles all of the aging out of my data, shuffling, et cetera. I don’t have to worry about that. So, awesome, we have a pipeline.

Let’s switch gears and take a look at what this would look like in production. What you see here is my deployed pipeline shown in the Dataflow monitoring UI built right into the cloud developers console. I’m going to go ahead and click in. And you’ll notice that the graph correlates one to one to the code that I just showed you, making it easy for you to understand — easy for you to understand your processing topology.

If I scroll down further, I can see other information, for example, the total records of processed. So I started this pipeline this morning right before I went on stage, we processed around 5 million, 5.2 million tweets. And right now I’m running at about 406 records a second. We have lots more headroom with this pipeline. But I wanted to show you that my data is actually flowing.

Now, our goal is to make dataflow monitoring valuable and integrated with the rest of your cloud development pipeline — sorry, cloud development experience.

Now, let’s take a look at what we could do with this data. This is a replay of the opening match between Brazil and Croatia. This represents an analysis of millions and millions of tweets and tens of thousands of touch information represented in the match. Now, if I scroll back and forth, I can see something interesting happen at the 71st minute. Brazil scores a goal. But you’ll notice in orange, Brazil’s sentiment goes down. So this is odd. Typically, the team that scores a goal will have positive sentiment.

Now, if I look in the timeline, there’s also some information that’s been injected for me by my pipeline that says, “controversial goal — controversial call related to the goal.” So this is giving me some insight as to why there may be some negative sentiment towards Brazil, but I still don’t trust my algorithm. So fortunately I’ve had Dataflow stream all of my raw data into BigQuery so that it can easily perform interactive analysis over very, very large data sets.

Now here is some of the filtered raw data around that 71st, 72nd minute. There you have it. The fans are upset with Brazil related to a bad call by the referee. Now I think I understand my correlation a little bit better. I absolutely love the World Cup. So thanks, Dataflow.

With Clouds Dataflow, you have a fully managed service and unified programming model for classic ETL and continuous analysis over simple or highly complex pipelines in batch or streaming mode.

Back to you, Urs.

Urs Holzle – SVP of Technical Infrastructure, Google

Thank you, Eric. And I’m sure, as all of you know, Switzerland is playing — my own country, Switzerland, is playing at 1:00 p.m. today. And I hope at the end of the day my sentiment is going to be 100.

But to go back to the technical part, I think — i hope you understand now why we stopped using MapReduce years ago. Cloud Dataflow really does for entire pipelines what MapReduce did for a single step, namely, it just makes it very easy. You don’t have to worry about scalability. You don’t have to worry about parallelism. And it will run faster and scale better than pretty much any other system out there, because we needed it too to solve our own problems.

So whether it’s data or code on server or on a mobile device, we try to make your life easier as a developer, more productive, while giving you the best price and the best performance of any cloud. We released hundreds of new features in just the last year, and we’re seeing incredible growth, thanks to all of you who are using the platform. And I’m very excited to see what you can do next year with the power of Google behind you. Thank you.

And now I would like to introduce Ellie to talk about Google Play. Ellie?

Ellie Powers – Product Manager, Google Play

Hi, everybody. I’m Ellie. And I’m absolutely thrilled to be talking to you about a few things that we’re doing to help you all create amazing user experiences and also grow your businesses.

Now, Google Play is the key to getting your app into the hands of millions of users globally, more users than on any other platform. And google offers app developers a wide range of cross-platform tools, and we build on this every year.

Now, you just heard about the cloud platform. And now I want to talk about a few ways that Google Play is helping app developers differentiate their applications and accelerate momentum. We’ve been making aggressive investments across development tools, ways to distribute and engage, and new mechanisms to generate robust revenue for your business.

So I’m going to talk today about a few of our efforts in each of these areas, what we’re doing to help you develop, what we’re doing on distribution, and what we’re doing to help you monetize your apps. So let’s talk first about development, building a great application.

A key part of this is testing. You’ve told us that testing can be painful. And we want it to be easy. Today, we are completely thrilled to announce that the Appurify team is joining Google. Appurify offers the most sophisticated mobile device cloud testing service, and more importantly, they’re just as passionate as we are about delivering high-quality user experiences. Appurify is leading the way in replicating how your app behaves in the real world and we’re excited to help them further scale and bring their expertise to your app development process.

Appurify allows you to test your apps across a wide range of devices. And this is critical if you want to be sure that your app produces consistently amazing results on every type of device, on iOS and Android all over the world. But ensuring quality means more than just device testing. Appurify service can simulate a specific mobile network and it can even simulate what happens if the connection is weak or drops out completely. Appurify also gives you detailed log data on device performance, network issues, power consumption, and stability. And this means your engineering team gets all the information they need to fix the problem, without needing to repro on a local device.

Now, developers tell us that testing is the cornerstone of creating high-quality experiences. And we want to make testing available to as many developers as possible. That’s why Appurify will continue to be cross-platform on both iOS and Android and available as a freemium service.

Next, let’s talk about building apps that help users get the most out of wearable devices. This is a new area that we’re really excited about, as you can tell. Today, we’re announcing a platform preview of Google Fit. This is an open platform designed to help users. Yeah! Thank you.

We want to help users keep better track of their fitness goals. So we’re providing a single set of APIs to manage fitness data, from apps and sensors, on both cross-platform devices and on wearables.

Now, before Google Fit, I was trying to track and monitor my bike rides through my bike computer and then my weight training through a specialized app, and it was a huge hassle. The information was way too siloed to actually help me. Fit takes away the complexity of handling multiple sources, giving you a unified view of a user’s fitness activity. And this helps you create more comprehensive apps.

So if a user grants permission, apps can have access to a user’s entire fitness stream to give better recommendations through this additional context. So, for example, Noom is a weight loss coach app and they’ve been an early partner on Google Fit.

So let’s just take a second to point out how the platform has helped Noom to enhance their app. So you’ll see that Noom is able to combine my workouts, nutritional information, and my weight. And because it can talk directly to [my life] in scale, it can let me know when my daily cookie habit gets just little bit overboard. So to centralize everything, Google Fit APIs allows fitness apps and brands to share your fitness activity, but only with your explicit permission. That’s the key, you are in control, you can choose who you share what with and you can delete your fitness activity whenever you want.

Google Fit APIs are also opening up new access to data coming in from hardware from top fitness brands. For example, Adidas has a collection of smart sensors that they’re opening up to developers for the first time. And we’re thrilled to announce that Nike is allowing other apps and fitness devices to integrate with Nike Fuel through this API. Nike will be publishing Nike Fuel to the Fit platform, meaning that your app can use it to give better insights into a user’s fitness. And, of course, there are many, many, many partners that are joining the Google Fit ecosystem. More partners in the program means you can all create more meaningful experiences for users.

So we’re incredibly excited about the potential here for a platform approach in this area. We want to make it so that you can build great fitness experiences with just one API. The platform preview SDK will be available in just a few weeks. So stay tuned.

So next, let’s talk about increasing distribution for your apps, growing your user base. And that’s where Google Play comes in. So last year, we announced our cross-platform service Google Play Games. And we have been delighted by the response from users and developers.

Google Play Games is now the fastest-growing mobile game network of all time. We have activated over 100 million new users in just the past six months. And Google Play Games connects your game to a concentrated network of people who love games. It makes gaming more fun through services like achievements, leaderboards, multiplayer, game gifts and cloud save. Developers use these to make awesome games and bring players back more frequently boosting their success.

And today, we’re announcing new experiences and game services to help you further enhance game play. So first step, we have the new game profile. In the new play games app, your game profile changes automatically based on the games you play and your achievements in each game. This new profile means that each player expresses their own gaming identity and it makes playing games with friends more fun.

Since the launch last year, users have loved saving their play games progress in the cloud. And we’re evolving that now into saved games. Users will be able to see bookmarks of their progress in the play games app. So Dave had shown us Leo’s Fortune earlier. This is a really fun game and I just started playing it last week. So I can see here in the play games app my saved game with a screen shot of me playing level 3. And we’re rolling out a new feature called quests. So quest is an online time-based goal that you can set up in your game, such as collecting a bunch of in-game items on a specific day.

Now we’re offering a set of APIs that can run these events for your players and reward them all without you needing to update your game. We have some early games that have started integrating quests and saved games, and we can’t wait to see how you are going to integrate these into your games. These features will roll out soon in the next update of Google Play Services and the Play Games app.

Okay. So we’ve had a look at a few ways that you can build high-quality, differentiated experiences and distribute them to a broad audience. Now let’s talk about what Google Play is doing to help you monetize your business. So one popular way that users pay on Google Play is through direct carrier billing. This means charging Google Play purchases directly to the mobile phone bill. Now, we’ve been working hard, rolling out direct carrier billing to our fastest-growing markets. It’s taken off really quickly. We’ve just expanded coverage with seven new markets for a total of 25 countries.

We’re happy to announce that direct carrier billing is now going to be available on more devices, on tablets. So if you have a phone and you’ve already set up carrier billing, you’ll be able to pay for apps, games, movies, music, books, and other content on your tablet, all still paid through that monthly phone bill. It’ll even work on tablets that are wi-fi only.

Okay. So there we are, app testing, the Google Fit platform, new Play Games capabilities, and direct carrier billing for tablets. These are four new ways that we can help you to create uniquely delightful experiences for your users. Thank you.

All right. Back to you, Sundar.

Sundar Pichai – SVP, Android, Chrome, and Apps, Google

Thanks, Ellie. We are seeing tremendous momentum in Google Play. And we really take this seriously, because it translates to success for you all. In fact, since last year’s I/O, we have paid out over $5 billion to developers on top of Google Play. It’s not just the volume of this number, but the rate at which this number is growing. It’s increased two and a half times, from $2 billion the year before. So we are seeing tremendous momentum. We are very excited, because it directly translates to developers building their livelihood on top of our platforms.

Of course, you all don’t just come to I/O to hear stats like this. You also come because you get your hands on cool new gadgets. So we’re going to give you some.

The first is an interesting one. A set of engineers in their 20% time surprised all of us with what you can do just with an off-the-shelf cardboard and your smartphone and the combination takes you into a very, very immersive experience. We’re going to hand each and every one of you a cardboard as you walk out from the keynote. And please share your thoughts on hashtag cardboard.

Next, we are very excited about Android Wear. So we are going to give each and every one of you either the LG G watch or the Samsung Gear Live. These are great devices, and I hope you enjoy them. We don’t want you to just create experiences for square-faced watches. We want to make sure you think about circular ones as well. So we will give each and every one of you a Motorola 360 as soon as it is available.

It’s an incredible time in personal computing. You saw our journey today across the platforms, across all these computing devices which people are using. It’s tough to believe that personal computing started only a few decades ago. We feel humbled to be part of this journey with you all. And we look forward to building more amazing experiences with you. Thank you. Have a great conference.

Related Posts

Reader Disclosure: Some links on this Site are affiliate links. Which means that, if you choose to make a purchase, we may earn a small commission at no extra cost to you. We greatly appreciate your support.

Multi-Page