Generative AI ,an Inflection Point for Human Creativity | Mike Dougherty | TEDxPaloAlto

Video Statistics and Information

Video
Captions Word Cloud
Reddit Comments
Captions
Transcriber: Hubert Stasik Reviewer: Zsuzsa Viola We are at an inflection point for human creativity. These happen every so often in our history, usually caused by some sort of technology innovation. The innovation this time is the artificial intelligence, or generative AI, and it’s going to change the way we tell stories. I know this personally because it already changed me. In the last six months, I’ve made over 2000 images and videos from AI. I was not a creator before this. Perhaps I’m a little addicted to it. My family wonders sometimes when I share with them the next image. You know, I kind of think they want to see the next one, but... but along the way, I learned a lot about how to make things with AI tools and also a little bit about what it might mean for creativity. We have such strong imaginations when we are young. When I was a child, I loved to read. I loved when my parents read to me. I loved playing with Legos, all the good stuff. And I remember when my parents brought home a home computer, 16-color graphics. I was in heaven, and I could play this game where I was Gandalf the Wizard. Later in my life, I never really learned to paint. You know, video photography was a little too expensive for me to do, so I didn’t become a creator. But I think there was something inside me during that time. And I think it’s inside all of us. It’s been there since the beginning. Humanity has been telling stories for 100,000 years. And about 70,000 years ago, we learned how to record them, cave paintings. Eons later, we invented the printing press, learned how to distribute these ideas. That was one of the inflection points. The camera. The camera allowed us to record ideas instantly. And then, broadcast technology, one to many became one to millions. The next big inflection point has been recent. That was mobile and the internet. Where all of us, it democratized creativity, it democratized us as creators, we were able to use our camera phones and capture photos, capture videos, and share them with everybody. And we did, trillions of photos, trillions of videos have been created, more media than ever in the history of our species. The next inflection point is now. By the way, it can be scary. Hopefully, after the end of this presentation, you’ll see that it can also be uplifting. Generative AI takes the camera out of the picture. And allows you to start with an idea. An idea that you put into text and create a text prompt into an AI engine. And that comes back with a turn paper, or a tweet thread, or programming code, or now images and video. Image creation has really rapidly and radically advanced in the last six months. About six months ago, if you would have made something, maybe it had an extra finger or an extra arm or something, and now it’s becoming more photorealistic. The way the technology works is fascinating. In a diffusion model, which is a machine learning model, we’ll start with a latent space, which is essentially the image, but use random. It makes random noise there. It starts with a single point randomly each time and builds out mathematically based on your prompt as a guide, along with the training data to generate an image in seconds. Each time that you regenerate, it will start with a different point, a different seed. And that is why our images look different every time we create something with AI. And that’s why it’s not copying. It’s generating something new. So the technology is the most advanced technology we’ve ever seen as humans. But it’s actually pretty easy to use. Let's do a text prompt together and see what happens. So, there are some building blocks that a good text prompt for imaging will have. Specifically, first, a medium. It’s like a photograph or a painting. Second, a subject - a man talking. An action on a setting, a Ted stage, and then, finally, the style. This would be the creative direction. These words here represent a sort of photorealism. So, let’s see what happens. Normally, this would take about 15 to 20 seconds. And here we are, a Ted guy on a stage talking. So, it’s pretty cool, but it’s really fast. So, let’s iterate a bit and have fun. Let’s put our favorite Gandalf into the image and see what happens. So, let’s give them some grotesque clothing as well, just to see what that looks like. So, here we go. And here’s Gandalf, kind of, right? He’s a little funky; I like him, but he is on a TEDx stage, so that’s pretty cool. Let's do one more thing with Gandalf. Let’s give him some special effects behind him. He’s a cool guy. And let’s use different styles. So, the style would be 80s sci-fi illustration, and see what happens. (Laughter) Okay, now that’s a wizard, right? Okay. It’s pretty good. Now, you might be an artist, and you might say, I could make that. I couldn’t make that. And if I could, or if you could, it would take you days because of the detail in this image. It took us seconds. Not every prompt has to be elaborate to be powerful. This is a prompt study that was done by user researcher, Lilly Godbole. And she took this prompt of a girl and just incrementally aged her, just changing the age in each prompt to see what would happen. Five years old. Gradually getting older. 15 years old. A 20-year-old woman. A 30-year-old woman. A 40-year-old woman. A 50-year-old woman and a 60-year-old woman. So, the prompt study quickly for me moved from a prompt state to a study on aging and a study on bias on how the training data perceives 40 years old. Also notice the fact that the old woman defaulted to Caucasian. So, there is likely a whole Ted talk on bias in training data and bias on engine. But it does show you that by changing the words a little bit, you can get a very different outcome. So, as I did more creations, that addiction I have. There’s some hacks I learned. This is a little one I’ll share with you. Take an iconic photo or just a cool photo that you like, and you can put it in an AI engine and use the describe function. This Elvis picture, it’s iconic of Elvis on stage. This is what the engine thinks the prompt would be to create that. These are four prompts that it comes back with. Now, if you do a close reading of some of the words up there, you’ll notice that the syntax is not that syntax that any of us would create to make this image. [Piles/stacks] - I do not see those in this image. I wouldn’t come up with that myself to come up with a prompt to make that. “Lovecore”. I’m not sure that’s a real word, so AI is creating words. Actually, the style, it’s in the engine. The style is lovecore. But the point is, it’s not a word a person would ever come up with to describe this image. My favorite on there is kombucha punk. (Laughter) Seriousl, well, it’s awesome. I was like: “This is cool. I can work with this.” So, I said: “Look, if our goal is to get the coolest image to come out of AI, it’s like, work with the AI engines and give it what it wants. Let’s give it back these sorts of phrases.” So let’s try one right now. Let’s take the post-70s ego generation, which is a style that came up a lot with some of the describes I did. Turns out that’s actually an artistic movement of Chinese artists. Notice the images are all Caucasian again. So again, not to make this the bias talk, but it's interesting. Let’s combine that post... [but their cool]. Let’s combine that post-70s ego generation with one of those funny words that we saw. This was one of the funniest that I ever found in the six months. Mommy’s on the phonecore. (Laughter) I literally do not know what that means, but I was like, I got to see what it makes. So, let’s mash the two together, post-70s ego generation and mommy’s on the phonecore, and see what happens. Boom. There you go. (Laughter) So, I got the 70s vibe - a Caucasian woman again. And the telephones are not something mommy’s going to want to spend too much time on. (Laughter) But her hair is looking great today, and she’s feeling good about that. (Laughter) You can kind of have fun with these. As you can tell, these are your Legos. Put in some kombucha punk next time, put in a brooding mood. See what happens. Video is next. So six months ago, this was the first video I made with a prompt. And you can see how it's warping. And its video is rapidly and radically advancing as well. In fact, this week OpenAI announced some amazing things with video. I’m guessing that this talk is going to be out of date very quickly, giving how fast AI is moving. But I think the point is that it’s moving forward. A realism is coming to video, a wide range of styles and our ability to tell stories in seconds with a prompt, with an idea is coming, is now. I think what’s going to happen with this innovation again, I hope in two years it's a lot more. Those dance videos that we all see on TikTok that we’re sharing right now, the performances are going to become more videos of ideas, more videos of creation. So, the next thing you can try, if you’d like, is music. So, there’s this music engine that you can put a prompt in a text prompt again. It will create a music bed, it will create lyrics, and it’ll have someone sing it, all artificial in seconds. So let’s put our favorite post-70s ego generation mommy on the phonecore, prompt into this track and see what happens. (AI music text) “Mommy’s sitting on the couch, Scrolling through her phone, she’s always in her own world. Will never want to let go. Snapchat, Instagram and Facebook. So she’s gotta get that perfect angle for the likes. It’s true. Mommie’s on the phone. She’s always on the phone. Will never get her attention. She’s always in the zone.” So, that song. (Laughter) That song. (Applause) That’s not a good song, everybody knows. I like, honestly, it was going through my head, and I had to practice, and it’s like, I can’t get the song out of my head. (Laughter) What's weird about that song, though, is that that little mommy on the phonecore created a whole story about a child trying to get the attention of the mom who’s addicted to social media. I mean, like, so this stuff is amazing. And it’s fun, and it’s uplifting when you get to use it a little bit. Okay. So, storytelling is the next big thing. And this is a masterpiece. It's a painting by Hieronymus Bosch, The Garden of Earthly Delights. You’re probably familiar with it. There’s a lot going on in this painting. And I said: “Well, what happens if we could ask the AI to kind of give us more of the story, kind of get in there and tell us what’s going on?” So I asked for an extreme close-up of this painting, and this is what it came back with. These are not close-ups. These are not blown up-pictures. These were generated. This is what the AI thought would be happening in that painting. So that, to me, was fascinating. It was sort of like fan fiction for art geeks, you know? And so where researchers and everyone are going now is this notion of character consistency, this idea that to tell stories, we need to have the character, not warp on you by the end of the video. With celebrities, this is a lot easier because they’re embedded in the training data. But for you and I, it’s a little harder. But that’s changing. Now we can be part of these stories too. This is me. Not me now. This is me from college, because I’m not stupid. I used the good photo from that. Right? (Laughter) But I never played football. So, this is me from college with a weird prompt where I’m playing football with these uniforms that would never be on the football field. This is me in a 19th-century painting. (Laughter) This is me in a movie. This is me in a music video. (Laughter) As you can tell, I had way too much fun with all of these places to put my favorite character, myself, in things. And so, it’s really interesting to think about where this is going. This is me asking the engine, what do I look like older as a video. In 18 to 24 months, I believe that we’re going to be able to make videos of ourselves, and what if our lives and all sorts of things. AI has been trained, as we just discussed, on hundreds of millions of images, more than any one of us can see in our entire lifetimes. But each of us has a brain, the original AGI. And we've been trained on our lives deeply for a long time. I believe there's an opportunity for us to take the influence that we have, have the influences, and the influence that AI has and combine them to create new things, new ways to create new ideas, new artistic directions. And it’s certainly the right time. AI is hitting an inflection point. The early adopters are now moving into mainstream. There are 7 billion smartphones. Billions of computers. Everyone’s going to have access to this technology. We are at an inflection point for human creativity. The “aha moment” that I had. I can be a creator. Maybe you’re having one now, too. You can be a creator. Maybe you will be the next Scorsese. Maybe you'll be the next Mozart. Maybe the next Picasso. I think we can all create new things together, and I can’t wait to see what you create. Thank you. (Applause)
Info
Channel: TEDx Talks
Views: 6,614
Rating: undefined out of 5
Keywords: AI, Art, Creativity, English, TEDxTalks, Technology, [TEDxEID:55171]
Id: 5-BXY7rxKUc
Channel Id: undefined
Length: 16min 1sec (961 seconds)
Published: Fri Mar 08 2024
Related Videos
Note
Please note that this website is currently a work in progress! Lots of interesting data and statistics to come.