Lens Fest 2021

Video Statistics and Information

Video
Captions Word Cloud
Reddit Comments
Captions
- We're rolling. - Action. Hi, everyone. Hi. - Hi, everyone. - Hi. - You want me to start? - Are you ready? Should I actually do the thing? - Take one. - Let's go. I love the future so much. - Empowerment. - Innovation. Limitless creativity. Like this? Peekaboo. Do I do this? Hi everyone and welcome to Lens Fest. We are so excited to be here with all of you. Lens Fest is our annual global event that highlights the innovation and creativity from every corner of the Snap AR community. It's been amazing to see how Lens Fest has evolved over time. Four years ago, we gathered a couple dozen Lens creators for a day of celebrating and building Lenses together in L.A. and now we're hosting thousands of attendees virtually from around the world. Whether you're just getting started with developing AR experiences, looking to create with even more advanced tools, or you're building a business with Snap AR, we are here to help you learn, get inspired, reach your goals, and celebrate the community's ingenuity in advancing the way we experience the world through augmented reality. This community is made up of creators, developers, brands, and partners who use Lens Studio to build, publish, and promote Lenses. With Lens Studio, you can build powerful AR experiences that enhance the way we communicate, get things done, shop, and learn, bringing your imagination to life. More than 200 million people engage with Lenses on Snapchat every day, and now there are more places than ever to experience Lenses. From video chats through Snap Camera, to immersive AR on Spectacles, and in partners' apps through Camera Kit, the possibilities are endless to fuse fun and utility in our lives. No matter where you love to use Lenses, it's so exciting to see how the community taps into cutting edge AR technologies and brings them to life in innovative and creative ways. Over the last year, our team has introduced a suite of advanced new features like garment and skin segmentation, 3D body mesh, and voice ML, all to help AR creators and developers build more engaging, realistic, and accessible experiences. There are countless examples of incredible Lenses made by our global creator community that are helping people to express themselves, try on a new look, learn something new, or experience art. Augmented reality is enabling new ways for us to interact with each other and the world around us. Snapchat was founded on the idea that the camera a could be used to, not just to document moments in time, but to communicate in real time. Lenses make this even more meaningful, bringing new imagination into your perspective. Excuse me. Today, there are 306 million Snapchatters around the world using our camera to create an average of five billion Snaps every day. Now, Snapchatters play with AR an average of more than six billion times per day. Augmented reality grounds us with a view of the physical world and brings all that's available in the digital world to us. Lens creators and developers are bringing this new world to life, enhancing our view with informative, relevant, and fun experiences. We are so excited to have you here as part of our Snap AR community and we cannot wait to invent the future alongside you. So let's start at the foundation of the Snap AR platform. Please welcome Trevor Stephenson to share the latest on Lens Studio. - Go. - Okay, so you want me to go. And then, you want me to start? Thanks Bobby. My name is Trevor Stevenson, and I lead the engineering team behind Lens Studio. All Lens creation starts with Lens Studio. Whether you're publishing a Lens to Snapchat, Snap Camera, Spectacles, or into your own app through Camera Kit, with all these ways to experience Lenses, we're evolving Lens Studio to be the most advanced, intuitive tool for creators to build AR for camera enabled devices everywhere. From advancements in machine learning through Snap ML, to 3D hand and body tracking, to scans, object recognition technology, the Snap AR platform puts the most innovative capabilities right into your hands. Today, we're launching Lens Studio 4.10, empowering creators to build the most entertaining, realistic, and helpful experiences yet. This first update might strike a chord. Last fall, we launched Sounds, giving Snapchatters a new way to express themselves by adding music to their Snaps. Our library now includes tracks from major labels and artists, like Olivia Rodrigo and J. Baldwin, along with rising stars, so Snapchatters can discover and share new music with their real friends. Sounds have been a hit. More than 520 million Snaps have been created with the feature. Now we're bringing the sounds library into Lens Studio so creators can add our licensed music and audio clips to their own Lenses, search through millions of songs, or find inspiration from a featured playlist in the asset library, and easily add a soundtrack to your Lens for a whole new dimension of expression. We're also introducing new features so AR elements can interact with the world in the same physical way that objects do. The world is governed by the laws of physics, so in order for AR elements to look incredibly real, they need to react to forces like gravity and collide with each other. Our new feature, called Real World Physics, makes that possible. Depth is another key feature of the physical world. For example, there's a person standing behind a camera right in front of me, so the camera is occluding the person. Before, state of the art depth have only worked on higher end devices, requiring creators to limit their audience or build the same experience multiple times for different devices. Now, our new World Mesh feature lets Lens creators, use depth information and world geometry understanding to create experiences that look and feel like they're part of the real world. Just build a world Lens with depth once, and reach far more Snapchatters than ever before across operating systems and a wide range of smartphone models. Not only are we bringing AR to the physical world, we're also bringing a bit of the real world and real time information into Lenses. Add real time data into Lenses through the new API library, making it possible to create smarter and more useful experiences. Monitor the US stock market up to the minute by adding Alpaca's market API data to a Lens, or build Lenses that dynamically change, based on current local weather conditions from AccuWeather. We're starting with four partners in the API library today, and we can't wait to see what you create. With Lenses becoming more robust, we want to help creators better understand who Lenses are resonating with and how they're being used. We're adding new metrics to Lens analytics so creators can get even more insights and consider refining their experiences, netting a better return on their investment, their time, and talent, into building Lenses. You can find Lens analytics in the My Lenses section of Lens Studio. We also want to share a sneak peek at one more thing coming soon to Lens Studio. We launched Land Markers in 2019, and local Lenses in 2020, layering augmented reality into dozens of popular destinations around the world. But, we know that, especially these days, not everyone can visit Paris's Eiffel Tower, or the Gateway of India, or London's Carnegie Street, and discover these experiences so easily. Soon, you'll be able to build and experience AR mapped to your own local landmarks with custom land markers, right in your city, or even in your own backyard. Here's how it'll work. Using a mobile device with a LIDAR scanner, map the area, using our mapping Lens to create a 3D model of your custom land marker. Then open Lens Studio to access your model and explore the creative possibilities. By choosing real world locations to build AR experiences , creators, brands, and partners will soon be able to bring new life to physical places, making your space into your canvas. On behalf of the team behind Lens Studio, we can't wait to see how you'll use these new features to build even more expressive, lifelike, useful Lenses for Snapchatters. Looking ahead, we want to help you innovate faster by delivering cutting edge capabilities, making it easier to collaborate, and expanding Lens Studio to be even more accessible for our growing global creator community. Now, please welcome Sophia Dominguez to share more about the growth of the community and how AR creators are building their careers and businesses through Snap AR platform. Thanks, Trevor. Not only are we amazed by the creativity and technical prowess of the community, but we're also excited about how creators and developers have found success with the Snap AR platform. From Michael Nickel building Blink into an AR powerhouse for the music industry, to Hardik Shah, of Super Fan, evangelizing AR for clients across India, we're honored to help fuel your growth. The Snap AR community has grown in so many ways. Today, there are more than 250,000 Lens creators from more than 200 countries and territories. In fact, more than 80% of Lens creators are based outside the United States. They've made more than two and a half million Lenses that have been viewed by Snapchatters, now, more than 3.5 trillion times. It's not just a select few whose Lenses have taken off. 300 creators have a Lens that have been viewed by Snapchatters more than one billion times. It's important to us to help the community succeed with Snap AR, empowering each creator, developer, team, and company, to define what it means to them. The Snap Lens Network is our foundational program to help some of our top creators and developers find success by providing early access to Lens Studio features, a verified profile on Snapchat, and new opportunities to earn money building Lenses. There are three key ways that we aim to help creators thrive: innovation, recognition, and monetization. We're pushing what's possible in AR alongside developers and creators, by encouraging and funding innovation all year long. Earlier this year, we introduced Ghost, our AR innovation lab, for developers and small teams to explore the technical and creative limits of AR. 20 Ghost fellows are working closely alongside Snap engineers and designers to build some of the most impressive, useful Lenses we've ever seen, from accessibility to fashion. Take Ecobatics, a Boston based team that aims to make the world more accessible for the visually impaired using computer vision. They're exploring how, through Spectacles, people with low vision can navigate indoor spaces and interact touch screens, using text recognition and text to speech. And the team behind Metail has a vision to create more realistic try on experiences. Through machine learning, they're making AR clothing drape and crease on the shopper's body, and accurately represent the color and material no matter where you are. Now, it's your turn. We're accepting new Ghost applications on a rolling basis and are awarding grants of up to $150,000 for each project. With early access to new features and support from the Snap team, you can explore and build new use cases that will help define the feature of AR. AR developers and creators can experiment building new kinds of experiences for Spectacles through Ghost or our Spectacles creator program. Spectacles are our first pair of glasses that bring augmented reality to life. We've partnered with dozens and dozens of creators around the world to explore new ways to fuse fun and utility through immersive AR. They've already made hundreds of Lenses with the world as their canvas. We've been listening to the feedback from the creator community and from those insights, we're already developing advanced, new capabilities that creators can tap into. We've brought Connected Lenses to Spectacles, so multiple people wearing them can interact with each other through Lenses. We've added Location Triggers, allowing creators to adapt their Lens based on the specified GPS radius, like a museum or a local neighborhood where the wearer is located. We've also introduced a new endurance mode, which lets creators turn off the display while their Lens is running, optimizing battery life for longer running Lenses. We're so excited by what creators and developers have already been building with Spectacles, and we know there's so much more collaboration, experimentation, and innovation to come. If you're looking to reimagine what's possible with AR, apply to join us on the Spectacles journey. We're building a bright feature of AR alongside creators, so we aim to help the community's Lenses shine. From bringing creator made Lenses to the Carousel, to new categories and favorites in Lens Explorer, and even showcasing creators in global campaigns, there are so many ways for Snapchatters to discover your work. Creators are at the heart of Snap AR and why we come together for Lens Fest. So this year we'll recognize the best AR creators and experiences through the brand new Lens Fest awards. Over the last few weeks, a panel of Snap's own AR experts have reviewed thousands of Lenses and evaluated creators' contributions to the Snap AR community. In our Lens Fest closing ceremony, we'll reveal the recipients across these categories. We can't wait to announce the award recipients at the Lens Fest award show on the final day of Lens Fest. The common focus across our AR creator and developer programs is to support your efforts to build a career or business making Lenses. We're exploring a number of pathways designed specifically for this community, especially as more Snapchatters, brands, and developers are increasingly finding value in the Snapchat camera. The demand for extraordinary AR experiences will only accelerate. Any Lens creator can work with businesses and brands who want to build Lenses for Snapchat, as well as developers, who've integrated Camera Kit and need Lenses to feature in their own app. The Snap Lens Network can also help get your AR career off the ground, with opportunities to earn through calls to creation and in the Creator Marketplace. And we're exploring a few more ways to support Lens creators' careers through audience engagement on Snapchat. Our new Lens Call To Action feature lets creators include a link on Lens, making it easy to drive Snapchatters to a destination, like a merch shop. We're so excited by the early successes from those who've tapped into this feature, like CyreneQ, who CTA linked to her store, driving more than $10,000 in merch sales. Another monetization opportunity for Lens creators in the Snap Lens Network is through gifting. A gift is a show of support that these Lens creators can receive through Story replies from their subscribers. Now, when a Snapchatter sees Snaps that brought delight to their day, they can send gifts and kickstart conversations with the Lens creators they love. Lens creators earn a share of the revenue from the gifts received through story replies. Gifting is available now for Lens creators in the Snap Lens Network across the US and India, and will roll out to additional countries soon. As Snap AR platform grows, we're focused on deepening the ways that creators and developers can grow their careers and businesses. We're so excited to be on this journey to build a whole new AR creator economy and learn along the way together. Now, I'll turn it over to Carolina Arguelles Navas to share more on how our partners are building their businesses with augmented reality on Snapchat. Thanks, Sophia. Hi everyone. I'm Carolina. I lead the product marketing and strategy for Snapchat's augmented reality products. As our AR creator tools have become more advanced. It's been thrilling to watch the creator community expand and take advantage of these powerful new technologies. Now at the same time that we've been improving AR tools for creators, we're also seeing phenomenal impact with our augmented reality offerings for brands. From product try on to virtual store environments, AR is proving to be an impactful, efficient growth vehicle. Now over the last two years, AR business experiences have undergone a major evolution, largely driven by the Snapchat community. Gen Zs and Millennials have grown up with a camera in the palm of their hands. They're used to using it every day, throughout the day. A large part of that engagement is with augmented reality. For Snapchatters, AR is an established behavior, as much a part of their daily routine as texting is for older generations. We now see over six billion AR Lens plays every day. As a platform, Snapchat enables to reach consumers where they naturally choose to spend their time. This makes AR on Snapchat not just a powerful strategy for businesses, but an essential one. We conducted a study with Foresight Factory earlier this year, which found that in the US, one in four shoppers say they would prefer to use augmented reality to try on clothes virtually, rather than visit a store to do so in person. For consumers, AR is already the expectation. According to Deloitte, today, there are more than a hundred million consumers shopping with AR online and in stores. By 2025, nearly 75% of the global population, and almost all smartphone users, will be frequent AR users. This explosion in AR interest makes it imperative for businesses to have a long term camera strategy. Whether the goal is engagement, product purchases, app downloads, or brand awareness and favorability, augmented reality can accomplish your goals in a way that is compelling and efficient. It's immersive, connected, fun, and useful, and it helps create relationships with customers, build brand loyalty, and can even drive business outcomes. For example, last year, American Eagle's virtual popup store, which Snapchatters could engage with an AR, generated $2 million in incremental sales. Earlier this year, Zenni Optical drove a 7.9x return on ad spend with their shoppable AR Try On Lens. Advances in augmented reality technology itself are also increasingly unlocking new opportunities for cutting edge virtual experiences. The addition of Vertebrae, an AR commerce company that recently joined the Snap family, will enable brands to create and manage realistic 3D versions of their products and easily extend those models into Snapchat or their own application and website to enable visualization and Try On. Innovations like 3D body mesh and realtime cloth simulation, which we recently brought to Lens Studio, enable more realistic clothing try on. Immersive, realistic AR Try On can benefit retail businesses in a number of ways. In a study conducted by Alter Agents, we found that two in three consumers are more likely to purchase after encountering a branded AR experience. AR Try On can also help reduce the number of returns from online purchases, which across 12 global markets, equates to $7.5 billion in value. For customers, augmented reality is the future of shopping. It can help visualize not just how a pair of sunglasses looks, but how it looks on you. Right now, more than three in four of all consumers believe AR technology will play a role in how people shop in the next five years, because it's the most personal way to shop and the most personal way for brands to make an impact. In a world so focused on the feed, a feed of content, a feed of products, the camera allows us to recenter shopping around the shopper, evolving from a feed to a mirror, the most personal experience someone can have with your brand. This why we are constantly working to improve our AR experiences to get as close to trying on the real products as possible. From reimagined retail experiences, to trying on beauty products, there are countless ways brands can use augmented reality to strengthen connections with their consumers. AR is becoming a true utility and a must have as a part of a future forward business strategy. As we head into the future, we believe augmented reality is the greatest opportunity to improve the way people shop, are entertained, search for products, discover new travel destinations, and so much more. We're incredibly excited about the world of possibilities, and businesses should be excited too. Now I'll turn it over to KP to share more about our offerings for app developers. Thanks Carolina. I'm KP and I support platform partnerships and the developer community here at Snap. Like this? Peekaboo. Our goal is to bring together both digital and physical experiences through the camera and build platforms that deliver value for people and for developers. But we can't do this alone. The success of AR is dependent on the diverse developer community to push the boundaries of this technology and the things that we can do with it. Last year, we introduced Camera Kit to make it possible for developers to use the power of Snap's camera to build unique experiences for their respective communities. This powerful SDK has enabled our partners to discover innovative ways in which AR can foster creativity and enhance the world around us. Millions of people are already interacting with AR experiences powered by Camera Kit every single day. Over the past year, we have seen incredible new AR use cases through a variety of apps through our closed beta program. With Camo, you can use a phone as a webcam for higher quality video calls, whether you're using Zoom, Meet, Slack, or other platforms. Camo introduced Lenses into their app so that you can be more creative and expressive, no matter what tool you're using for visual meetings and hangouts. Over in India, people use Moj to make short form videos and complete daily challenges. Since introducing Camera Kit, nearly half of the camera content posted to Moj includes a Lens. Beyond communication and expression, DRESSX is an app that lets you experience and shop digital clothing. They're using Camera Kit to make it possible for their community to try on AR clothing from selected designers and create videos wearing virtual clothes. Camera Kit helps our partners accelerate their time to market, tap into the AR ecosystem, and increase the discoverability of their Lenses through their public profile on Snapchat. This all starts with Lens Studio. Partners can build their own Lenses, or like Moj, partner with local Lens creators, or even use the creator marketplace to hire Lens creators to develop Bespoke experiences for their communities. For Lens creators, Camera Kit provides a new way to expand the reach Lenses beyond Snapchat and the opportunity to work on projects that push the boundaries of Lens innovation. We want to make it even easier for Lens creators and Camera Kit partners to work together. Soon we are rolling out Lens Packs on Creator Marketplace. 575 00:26:47,892 --> 00:26:50,459 Lens creators can feature a categorized group of Lenses on their profiles for Camera Kit developers to purchase. This opportunity provides a new way for Lens creators to monetize, while making it easy for Camera Kit partners to quickly bring their integration to life and iterate on experiences for their communities. We are so inspired by the creativity and innovation from our Camera Kit partners and Lens creator community. Today, we are excited to share that we are planning to expand the reach of our beta program in the coming months. We are so excited to work with even more partners to build new AR experiences that we would never have imagined. 592 00:27:37,498 --> 00:27:41,065 Snap AR makes it possible to bring new AR experiences to highly engaged communities around the world and help developers leverage the intelligence and power of the Snap camera for the apps. Thank you all so much for joining us to celebrate the amazing work that has come to life through Snapchat, Lens Studio, Spectacles, and Camera Kit. There are so many exciting things for us to build together in an amazing community to serve. We can wait to help you create, build, and explore. Hi, I'm Olha, AR engineer here at Snap, and today I'm excited to tell you about Sounds, a whole new section of the Lens Studio asset library that allows you to browse through sounds from the most popular musical artist in the world and bring them into your Lenses. Once you open the corresponding section of the asset library, you can see the list of featured audio tracks. You can play them, pressing on the play button, and find the information of the availability in different countries. Once you've made your choice, press the input button. At this point, you will see the new asset added to your resources panel, and we can play it and any other audio track by using audio component. You can also use it together with Spatial Audio by enabling corresponding check boxes and adding in the Audio Listener component to your project. You can use the updated audio analyzer template that allows you to work with different types of audio tracks, using the license sound, file audio track, microphone audio, or even your own buffer of generated audio samples. So I'm going to import my sound once again and just set it as the input of the helper script. I'm going to match sample rate. The template also allows you to write audio samples into the procedural texture and combine inspiration from the sound from your favorite artist together with such powerful tools as Material Editor or the Effects Editor and create great effects like this one, or you can even build a music video like this one, by utilizing building blocks from the asset library and helper scripts. โ™ช Gotta know what's it like, yeah yeah โ™ช With Sounds in Lens Studio, popular music from around the world is now available to help you be even more expressive and creative. I remember Evan coming to me with this concept of sending photos that disappear, and I don't think I was initially bought in, but I liked that it was a very different idea. I remember presenting the prototype to my class, and just before class, the prototype wasn't working. I'm messaging Bobby, "Dude, we got to get this thing working." I go, I present to the class, and I'm so excited about this idea of sharing photos that disappear, and everyone's like, "That's the dumbest idea ever. Why would you send photos that disappear?" The very first prototype, it was just images, pre captioned. We would write little notes on Post its, or type things out on the keyboard, take photos and send them back and forth. But, it was super, super fun. I just distinctly remember that initial feeling of, "Holy cow, this is just amazingly fun to send photos." I think over time we really kind of developed this broader idea around communicating much more naturally as humans in a digital form and responding to what other companies had done before us with more permanence. I think people felt really liberated. I mean, at the time, all of our friends were trying to get jobs and were worried about what they had posted on social media, and everyone was talking about that. So I think there was this feeling with Snapchat that was just more free. It wasn't about posting the perfect photo online. It was whatever was funny or random that you could just share with your friend, and then it would just be gone. I think like that lightness was so different to what was on the internet at the time. I remember hearing, and being surprised to hear that, one of the early Snapchat trends was, "Make the ugliest face you can," and friends were just sending hundreds of Snaps back and forth making ugly faces. That was a really cool thing to hear. It seems super silly, but definitely indicative of kind of the early behavior on Snapchat and the culture that we were starting to build. Oh. Hey everybody. This is Cassidy from Smule. I'm here to talk to you about how we integrated Snap Camera Kit into our apps. For those of you who don't know, Smule is an app where you can sing with people on the internet, and a lot of times that can be pretty scary. So enabling people to put a Snap Lens on their face is a great use case for us because sometimes when you're singing on the internet you just want to be a goldfish, or you might want to be a pineapple, or if you're like me, you're going to want to be an astronaut. So I'm going to talk to you today as an astronaut and tell you about how we're using Snap Camera Kit and Smule together. We have a unique use case where we allow two different people to use two different Lenses at the same time. For example, the person on the left side of the screen might be wearing aviator sunglasses, and the person on the right side of your screen might be wearing retro sunglasses. We power this with Snap Camera Kit. If we were to build it ourselves, it would be incredibly expensive and difficult for us to do. Another great use case is the green screen functionality. We have people singing across the globe in different locations, but we want to make them look like they're in the same location. The green screen functionality lets us do this, where we can apply a common background across two video streams, make two people look like they're singing in the same location, even when they're on opposite sides of the world. We're always looking for fresh AR content to showcase in our apps and particularly Lenses that allow users to connect and have fun in ways that they never thought was possible. So we can't wait to see our users singing with your Lenses. Signing off from outer space. The inspiration for my Hot Potato connected Lens game came from the game that we all played when we were kids. I was really inspired to take this brand new technology and bridge it with something that was familiar to a lot of us so that when this game was introduced, people would generally exactly how to play when they opened it. I thought that was important when introducing people to this whole new world of augmented reality. Hi. I am Rajni Jaques, the Global Head of Fashion and Beauty Partnerships at Snap. We are in the age of augmented reality or AR, and the power of the Snapchat camera is this generation's conversation catalyst. AR on Snapchat is evolving how people shop, engage with art and music, experience fashion and beauty, and so much more, by transforming in real life experiences into immersive digital moments that spark conversation and encourage Snapchatters to see the world in new and interesting ways. Fashion and beauty brands have been harnessing the power of AR through the Snapchat camera for some time now, unlocking a range of new experiences for Snapchatters to engage with products, including the ability to virtually try on anything. I'm talking shoes, sunglasses, makeup, nail polish, and from anywhere. By bringing millions of people, the ability to explore products without ever having to set foot in a store, AR is evolving into a true utility that can drive product consideration and purchase, in addition to brand awareness and favorability. AR on Snapchat is where people come to play with the brands that they love and experience or discover new ones. This isn't the future. This is right now. We are already seeing thousands of brands using AR in their campaigns on Snapchat, and we're excited to continue the momentum with our new and existing partners for the Snapchat generation to experience first hand. Honestly, all of this is pretty cool. This is so cool. My name is Jeff Miller, and I'm the Senior Director of Creative Strategy at Snap, Inc. And action. As we think about the many ways that augmented reality can propel forward the world of art, we have to start with the collaboration that we developed with LACMA. That was a focus on how do we make monuments more reflective of diverse stories and not wait for physical monuments to be built. In that collaboration, we worked with a diverse group of artists and Lens creators to bring these physical stories into life in the real world. Take our work with Alex Israel. We had an opportunity to sit down with him in his studio to understand how he wanted to translate his physical forms into this new augmented reality space. For Art Basel this year, we created two new experiences with Alex directly. The first one has one of his signature characters, the avocado, dancing alongside you and your friends. The last one, you see Alex himself originate from with inside his installation and then paint the outline of his own head. These types of experiences are only possible because of AR. For example, I could be in my living room in Los Angeles, California, and immediately be taken to Paris, where I can see the final Christo, and see directly the vision that Christo himself was working on for 60 years to wrap the Arc de Triomphe. Now that is magical. I am so proud of our collaborations with artists today, those that are established, as well as those that are still young in their careers. So much of the great work that originates, starts when you have an artist with the clear vision operating in new formats. But what has me most excited, are the stories that we don't even know about yet, the artists that are still trying to find their voice. AR is such a great medium, democratized to help them frame their stories. I can't wait to see what the community develops next. Peace out. - That was great. - Thanks. One of my favorite releases was when we added video because I feel like that a big step change for Snap, and we invented that super cool way of having to take a photo and holding for video, which at the time, was a totally new thing. We had to figure out the timing, basically, for how quickly you released to make it a photo, versus holding it a little bit longer to start the video timer and everything. It was really cool. Yeah, we've done some amazing things with the company that have all been very, very exciting to release and get out there. For me, anytime we've been able to really surprise our community with something we've been able to do, that's always super, super rewarding. The first Our Stories that we did were just super fun. We launched at EDC. Then from then on, we started doing these kind of regional stories around the world. Just to see this on the ground, real content from people submitting public videos, getting that put together in the way that we did, was really amazing and totally unprecedented. It's exciting to see that work kind of, now, carry forward into Spotlight. We've had some amazing hit Lenses over the years, particularly the ones around machine learning, rendering, and generative networks, anime, cartoon Lenses. Recently, those have been big hits and just endlessly impressive from a technical feat. I mean, I think the lesson, always, for us has been just to follow the community, follow what people are doing. If you just pay attention and look at human behavior, all of these ideas that seem innovative and radical are just lying all over the place in front of you. You just have to listen, really, to people and pay attention. Whew. Okay, ready? Go. Just open it up and close it. Hi, I'm Aradhana Vaidya. I'm the Product Manager for Lens Studio at Snap. Lens Studio is a desktop software to create Lenses. It is your gateway to create unique and powerful AR experiences. Over the past few years, we have seen our community create immersive experiences, both for fun and utility. Lens Studio is the 3D tool which enables this creation. It has all the bells and whistles of a professional 3D creation software, and what's special is that it's tailored towards the AR creator. In Lens Studio, we have state of the art graphics and rendering. We have a material editor that lets you create custom materials and modify their properties and parameters to create materials with the precise look and feel that you are going for. We also have a VFX editor, which is our node-based visual effects engine. With it, one can create and simulate particle systems, for instance, to create effects such as rain, fire, sparkles, confetti, et cetera. In the past few months, we have released some truly exciting features which enable Lens creators to create innovative AR experiences with Lens Studio. We have accurate face, body, and hand tracking, through our Snap ML technology. You've probably seen a vast number of Lenses on Snapchat that have been created using Snap ML. We released a feature called Upper Body Skin Segmentation, with which we've seen our community make tattoo Lenses or Lenses with animated textures and effects applied to their face, body, and arms. Then, there is the Try On feature, where you can take an external mesh, like clothing, and apply it to a body. This clothing mesh does not have to be a rig mesh. With the try on mechanism, it attaches itself and deforms according to the pose and shape of the body that it's applied on. It opens up endless possibilities. We have seen Lenses with clothing, different kinds of garments, accessories, and people expressing themselves in different ways. We recently introduced voice capabilities as a part of Lens Studio. This is our Voice ML technology. With this, one can create voice-based Lenses. There is speech and command recognition, with which speech can be used to trigger AR effects. This incorporates transcription and keyword detection, and it brings natural language understanding into your Lenses. There is also text to speech, which converts text into human-like speech. Also with system voice commands, you can build Lenses to accept commands, like, "Take a Snap," or "Record a video." The great news is that we have templates in Lens Studio, which demonstrate each of these functionalities, and using templates can help you learn these new features easily. So now, let's talk about some new features in Lens Studio that we are planning to release soon. - Is it awkward yet? - Yeah. Music integration. You now have access to a catalog of licensed music in Lens Studio. In the asset library, go to the music section, and you'll see a large catalog of tracks to browse and select from. You can now make new and exciting Lenses with these cool music tracks. Simply add a track from the library to your Lens, and use it to bring that special magic to your AR Lens. We can't wait to see how creative our community gets with this added music functionality. Another feature that we are truly excited about in Lens Studio is Real World Physics. The physics engine enables objects in the scene to behave and interact with forces in the real world, such as gravity, velocity, or acceleration. It can help you add realism to your Lenses. With physics, we have added the functionality for collisions, rigid body simulation, and constraints. You can now create scenes with bouncing balls, toppling boxes, or even falling debris, and make these interactions look real. We will continue to take this further with additional functionality, like the ability to script, soft body simulation, terrain, dynamic colliders, and even making improvements to our hair and cloth simulation. Not only are we innovating with features that let you create cool Lenses for the face and body, we're also building features to create world facing AR. One such feature is our new cross platform, World Mesh solution, that works on both AR Kit and AR core devices. This allows developers to create more realistic world AR experiences without a hardware sensor and without having to create different content for different devices. World Mesh understands the geometry of the world around you, reconstructing the environment as a metric scale track mesh with semantic labels. It also provides a depth texture with a four pixel depth estimate for the current camera image. With this, developers can create AR content that looks and feels like its part of the real world. Another feature in this category is Custom Landmarkers. With Custom Landmarkers, for the first time, we'll be allowing creators to choose their own locations to build AR on. Through creative feedback, we have seen the desire for more location AR experiences, and we are excited to provide these capabilities to our community. With this feature, you'll have the ability to scan a structure or a building with LIDAR and load it into Lens Studio. You'll then be able to author AR against that structure and publish a Lens with a Custom Landmarker. We are really excited to see what our community builds with this. Some examples could be an AR walking tour, where the content interacts meaningfully with the surrounding environment, or a digitally enhanced storefront. Again, the possibilities here are endless. Now let's talk about API Library. This will allow Lens developers to build Lenses with first and third party APIs using Lens Studio. You can now power unique AR experiences with data that these APIs could fetch. For example, weather data, market data, or sports data. Take a look at our asset library to see the full set of APIs that we are providing through the API Library. Show us what you can do with these APIs, and give us your feedback to let us know what other APIs you'd like to see. For those of you who have been using Lens Studio for a while, you know that we've been releasing Lens Studio every two months. But as we develop more features or improve workflows, we'd like to provide newer versions of the software, at a faster pace. Starting early next year, you'll see a new Lens Studio version every two weeks. Lenses you create with Lens Studio are usually published for Snapchat on your mobile device, but you could expand your Lenses' reach by also pushing your Lenses to Spectacles. You could also use our Camera Kit SDK to develop and build Lenses that are cross platform. You got to do it harder than that. Okay, ready? Cool. I'm ready. Thank you Aradhana for showcasing the powerful Lens Studio. I am confident developers around the world are excited to build creative and powerful AR experiences with the released and upcoming features of the platform. At Snap, we believe in cultivating a self-sustaining ecosystem, which enables AR developers to build creative experiences, innovate, monetize, and connect with one another. We do this by engaging and activating our AR developer ecosystem through our foundational program, the Snap Lens Network. The Snap Lens Network underpins all our AR developer relation efforts where our developers get access to unreleased Lens Studio features, paid opportunity, receive a verified profile on Snap, so that the Lenses and work at broader distribution on the platform. The program caters to three distinct groups of developers. We have the Lens partners. Lens partners consist of agencies and studios. These are professional companies or teams that develop professional AR experiences, apps, or games, and earn majority of their revenue through brand work. Lens developers. These are individual developers or teams who build products and complex AR experiences. They occasionally produce work for brands, but typically, focus on driving innovation on our platform. And finally, we have Lens creators. These are individuals who create artistic, fun, and shareable Lenses, that often go viral, and sometimes partner with our developers on technical projects. While in a Snap Lens Network, qualifying members have the opportunity to further monetize and innovate with us through our auto-developer programs and initiatives. Let's start with the Creator Marketplace. Select Snap Lens Network members have the opportunity to be added to the Creator Marketplace, which enables businesses and companies to reach out for Bespoke Lens creation. The Creator Marketplace enables businesses to discover and partner with Snap's creator community in a scaled and seamless way. For agencies and studio partners, we have a dedicated program to help them monetize further through the AR Partner Program. The partner program was created with a goal of democratizing the creation of branded augmented reality. Through the Snap Lens Network, agencies and studio have the opportunity to be invited or considered to join our AR Partner Program. Ghost is Snap's AR innovation lab, designed to spark the next generation of camera products built on our platform. Through Ghost, we provide funding for developers and studios to build Lenses in the form of products that leverage our most advanced AR technology and deliver value beyond communication. Visit the Snap AR site for more information about how to apply. The new Spectacles are designed as a development tool for developers looking to push the boundaries of AR. Although they're not for sale, if you are a developer, you can apply using our creative form on the Spectacles website. Those selected for the program, will get a pair of next generation Spectacles for their work, dedicated support from our product team, access to weekly team office hours to get feedback and share best practices, as well as access to a dedicated email address for their work. Finally, you also get opportunities to connect and learn from other gifted developers in the program. Whew. Okay. Ready? Go. Yellow is a collection of programs that support founders building companies at the intersection of creativity and technology. Yellow supports founders in their entrepreneur journey, through events, networking, community, education and opportunities for investment, through the Yellow Accelerator program. Visit the Yellow program's website for more information. Finally, as a way of say thanks to our most engaged community members, we would like to formally recognize those in our platform who have sustained a warmth and friendliness of our current developer community by being highly active, helpful on the forum, creating resources that support community development and engaging and thoughtful conversations around advanced AR topics. With our new Snap AR Lens Ambassador program, we would like to support those in the Snap Lens Network who have been passionately sharing their knowledge of AR with their local and global community. In addition to recognizing the Lens ambassadors on the forums as stewards of AR, we will feature them and their work within our blogs, banners, and social channels. We hope their kindness, intelligence, and creativity, will inspire developers for years to come. At Snap, we are committed to the success of AR developers, and we are thrilled to be on this journey with you. To learn more about the Snap Lens Network and our other AR developer programs in our developer ecosystem, kindly follow this link. Hello everyone. I'm Yevhenii. Thank you for joining this event. In this video, I'm so happy to share with you my tips and tricks on optimizing projects in Lens Studio. I want to start with the fact that project optimization is quite an important stage when working on Lenses, not only when we need to fit into the required size, but even when this wait of the Lens fits into the wait budget. Plus, fast loading of the Lens is a very good step towards the user because the sooner they load your Lens, the faster they can enjoy it and, let's be honest, we don't like to wait for loading. So without further ado, let's get started. Let's start with the simplest. You may have guessed, but for such an effect as vignette, you don't need a large texture at all, but only a quarter of such texture is enough, which can be, then, reproduced with the mirroring function. Sometimes, you don't need texture at all. For instance, in this simple frame, it can be created from screen images. I can create screen image of whatever size I want, remove the sprite component, and the object becomes just a container for other sprites. Next, I will create this sprite inside, with just basic material. Now let's talk a little about the formats, JPEG and PNG. They have their own strengths and weaknesses. Let's take this picture. In PNG, it takes like this, while if you save it in JPEG without compression, then it takes like this. So as you can see, the same picture can take several times less, with practically the same result. Using the built in compression tool in Lens Studio, you can also find that, in this case, the JPEG format still takes up less. However, with this picture where there are few color or grayscale transitions, PNG wins noticeably. The key thought in this experiment is that there is no single rule of where and when to use these formats. Try to experiment with saving in different formats and you yourself will learn to predict in which cases which format is better to use. The last trick is to use the same object but with different color. For example, if we take this texture, then with help of combination of base color and recolor, we can achieve a huge variety of colors just from one texture and this will significantly save space. Now let's move on to creating simple and optimizing effects using the node editor. With one texture like this, you can create a lot of beautiful things. Let's add some heat to this video and create fire from a static texture. We have a texture with slightly different noise and RGB channels, which is styled. Let's separate them by channel first. Add, remap, then merge them back through the blend node with screen node. Then I'll add to each channel movement by the scroll cords and also scale cords for each channel, and I will make small changes in the parameters. Now I will make a mask. For this, we will create a radio gradient subgraph and slightly correct the values. Let's apply this mask through the construct node. And fine. Now let's add some randomness. There is a bit of magic going on now, so watch my moves. I will duplicate this area with the nodes and just add it as a modifier for the UV of the same first channel. I will remove the intensity a little through multiply. We can also slightly change the values for the speed and scale of this. Let us call it turbulence. Doesn't coincide with the main movement. Then duplicate this turbulence for the other to noise in the channel, and change their values quite a bit. We can also use this technique for the mask. This time, I'll do it through the mix node. I think two channels will be enough. Now let's create something like a glow on top of this noise. You can easily make such a glow from the UV. We need ingredients like remap, distance, and one minus. Connect it all, and the glow is ready. I will combine this node into subgraph for convenience and add this entire effect after the construct node. I will also use the clamp node so that the balance did not exceed zero and one. All that remains is the paint, and in this example, I will just apply the lookup texture. Also, of course, there are many ways to paint it without it. Slightly reduce the influence of this lookup through the mix node, and also slightly dim the dark areas in the front of lookup using another mix. Wow, we made an animated fire from the static texture, but I'll tell you more. We made a super constructor, and the real magic is just beginning. So fasten your seatbelt. Make a copy of the fire material and apply it to another sprite. Next, let's reduce the effect of textures on UV a little. Add spiral coils to make it move a little rotated. Remove a couple of unnecessary details, and tint. Put this sprite somewhere here, duplicate it, and voila. The fog is ready. Next, let's duplicate the material again, assign it to another sprite, add a one minus node to make the effect darker, change the blend mode to multiply, and slightly adjust the color values. Maybe I'll add some more contrast with green map, and voila again. The smoke is ready. Let's place it somewhere here. Duplicate the sprite too, and randomize it a bit. Now great. Next. 1339 01:06:26,172 --> 01:06:27,005 No problem. Still using the constructor. I'll make a copy of the fire material and leave here only one channel, and turbulence for one channel. Let's change the direction for this texture movement here. Probably, we will make the influence of turbulence a little less, and all this can be applied as an alpha for a color, using the construct node. Let's do the placement trick again, and that's it. Super easy. Let's add some cold post effect for showiness. For example, frost sounds like something cold, and probably a little ss. Good. Well, to make it really beautiful, I add some post distortion effect and change the texture to our noise. We're using one set of texture for animation, and let's bring the face back a little so that it doesn't get very distorted. Finally, we did it. As you can see, you don't need a lot of texture and very complex shaders. Start creating Lenses now. It's super easy, fun, and entertaining. Thanks. See you. My name is Olesia, and I work as an AR engineer. Favorite part of my job is getting to help our amazing Lens creators building in templates and example projects for them to help bring new ideas to life. In this demo, I'd like to talk about how to combine additional visual effects to achieve a more cohesive, polished look. Let's take a look at the example Lens project. As we can see, main Lens parts have been already added, but I'll go through some steps that could be final steps before publishing. Let's take a look here. Main idea of this Lens is just having cute peach stickers on the face. Let's say you already have your texture here, you already added it to face mesh, you created material, and everything is set up. Then, on top of that, what my idea was to add a bunch of makeup, oops. A bunch of makeup, like lookup and additional bloom effect on top of the stickers to make it just come together more and look more naturally and nice. Let's start with this project here, that has only these peach stickers. Now let's see what we can add on top of here. I already have resources for this prepared for me in this folder. I think I'd like to start with some eyeshadow, something to help make the peach stickers less obvious and less stand out on the face. So let's say I'll add this one here. As you can see, it's already a face mask object, so I already have texture in here. I already set up coordinates and how it's aligned to the face. It's usually what I'd like to do instead of setting up textures and how it looks on the face every time from the beginning. I just do it once and then save it as an export object summary in my folders, ready for me to use in different Lenses. Let's create same object and call it, "Makeup," just to nicely organize it in here, and put eye shadow. Then, let's check that we don't have any random layers. Sometimes, you get leftover layers. Let's see if it's working. If you can't tell if your effect is working, sometimes it's good just to use a static image when model is not moving, so it's going to be easier to see. Okay, now let's add a couple more stuff. This one. This one adds more definition to the outer corners of eye, and just add some very smooth eye line, which is always nice to have. Then, let's see, maybe I want to use some eyelashes also. This one, lower only, for example. Then, I want to use for one also. Don't forget to fix layers for all of those. Also, as you can see, there is one little nice optimization. Instead of using duplicated texture for both of eyes, we're only using half of the face. Then, in material, we set up in a way that it just been mirrored to the second part of the face. So you get two size smaller texture, which is always nice. As you can see, already, when you have more makeup on eyes, stickers already doesn't look so foreign on a face, if that makes sense. But, let's say, now, I want to add maybe a cute blush to go underneath the stickers, to blend it even more. I have one pretty cool one, because as you can see, it kind of goes over nose and it's like additional shadow and depth underneath our stickers. Okay. Now I think I'm done with the makeup for now. But now I just want to add something on top of both stickers and makeup to blend it together more. So, ideally, I would use some lookup like color correction, that is an appropriate, warm tone. Let's see. Let's try this one. Yeah. As you can see, it adds more saturation, adds more warm colors. If I had stickers, it would be cool toned, like some pink, blueish purple. I would probably use like a cooler tone lookup for this. If I think it's too intense, I can just tone it down a little bit. Yeah. Also, one material I really like to use is this bloom one. It just adds a nice, soft glow to the whole picture. Yeah, as you can see, adding all of this helps still preserve your idea of having these cute stickers, but just adds more to the overall picture and how it looks together nicely. Yeah, I think that's it. Thank you so much for watching. I hope you enjoyed it and get inspired to create more. Okay, thank you.
Info
Channel: Snap AR
Views: 12,071
Rating: undefined out of 5
Keywords: Snapchat, Lens Fest, Augmented Reality, AR
Id: XG31NacaT1E
Channel Id: undefined
Length: 89min 10sec (5350 seconds)
Published: Tue Dec 07 2021
Related Videos
Note
Please note that this website is currently a work in progress! Lots of interesting data and statistics to come.