- Hello everyone, and welcome to AWS Innovate,
Generative AI and Data Edition. My name is Tom Godden. I'm a director of
enterprise strategy at AWS and former Chief Information
Officer at Foundation Medicine. Before we get started, I want
to thank you for taking time to come and learn with us. From model customizations in Bedrock to vector databases in FMOps, today's event is packed
with some amazing sessions designed to help you
build with generative AI. And today I'm excited to share
with you how you can innovate with generative AI. Wherever you are, whether you have a
generative AI application in production today, or
you're trying to move quickly to transform the way you
engage with your customers or the way your employees get work done, you have the opportunity to
change the way you do business. Generative AI has taken the world by storm because we've been able to, through customer facing applications, experience the most powerful and latest machine learning models. While a lot of attention's been given to how consumers are using generative AI, we think there's an
even bigger opportunity in how businesses will use it
to deliver amazing experiences for their customers and their employees. We believe the true power of generative AI goes beyond a search engine or a chat bot. It will transform every
aspect in how companies and organizations operate. In fact, Goldman Sachs
forecasts that generative AI could lead to a $7 trillion
increase in global GDP and lift productivity growth
by 1.5 percentage points over a 10 year period. Gartner goes even further to say that generative AI is likely to be one of the most disruptive innovations yet encountered in the digital workspace, and that they expect it to impact 80% of jobs to some extent, with the information workers'
job changing the most quickly and the most dramatically. Customers around the world
have deployed a wide range of generative AI applications and are now seeing the
benefits of this technology and becoming more efficient in transforming the customer experience. Customers like Intuit. Over the past decade, Intuit
has partnered with AWS, first to migrate their
applications to the cloud and now improving their
customers' experience with generative AI. Intuit's AI-driven strategy has led to the creation of GenOS, a proprietary generative AI
operating system built on AWS. GenOS facilitates the rapid development and deployment of personalized
financial solutions, and it does this in two ways, by providing the ability
to harness underlying data and by providing access to
a multitude of third party and large language models. These can be scaled out
to customers with ease, all while balancing costs using Amazon SageMaker and Amazon Bedrock. Intuit uses AWS to process
an immense volume of data. Their systems are making more than 65 billion machine
learning predictions a day. With GenOS, Intuit is
building new applications like Intuit Assist for
TurboTax, Credit Karma, QuickBooks, and MailChimp. In short, Intuit Assist is a
generative AI-powered assistant that offers personalized insights to help users make smart
financial decisions. Another example is Adobe, where they're unleashing
a new era of creativity with the development of Adobe Firefly. Since launching their
first production model in March of 2023, users of Adobe Firefly, the company's family of
generative AI models, have generated over five billion images, enabling users around the world access to powerful generative AI tools. Adobe moved quickly with
generative AI, prioritizing speed and harnessing the power of their data. The professional quality content trained on Adobe stock assets and openly licensed public domain content has been an immense success with their Photoshop generative fill, where they saw 10X adoption compared to a typical new feature. Success stories like Adobe and Intuit are just the beginning. Artificial intelligence
is changing at a pace like we have never seen before, with new discoveries and
innovations happening each and every day. At AWS, we've made it
our goal, our obsession to lead the way for customers. Things like cutting edge
innovations with our custom silicon to bring you the best infrastructure
for your AI workloads, purpose-built managed services that provide a new wave of capabilities, and access to high performing
models to make it even easier for you to securely build and scale generative AI applications. And of course, continued investment in
training and resources, highlighted by our commitment
to train 29 million people with free cloud computing
skills training by 2025 through over 100 AI and
machine learning courses. Pushing the edge of this new era requires a constant willingness
to learn and take action. And in working with our customers, we've identified four foundations needed for you to help innovate
with generative AI. The first being, you gotta
choose the right use case and you need to move quickly. Generative AI can accelerate productivity and transform your business
operations in a variety of ways. When thinking about use cases, there are so many to choose from, but I like to think of
these as front office and back office use cases. Front office use cases are those that directly impact the
experience of your customers. Here's where you're using generative AI to reinvent the way they
interact with your company and enhance their experience. The back office use cases,
the ones behind the scenes, where generative AI is
boosting productivity and creativity of employees or optimizing and driving
higher efficiencies and lower costs with
your backend processes. But what does this look
like in production today? The possibilities for generative AI to revolutionize your
business and industry are nearly endless, with
hundreds of use cases and opportunities at your fingertips. AWS customers of all sizes and industries are innovating in so many unique ways. Customers are continually working to enhance their customer experience, like the PGA Tour, who
began collaborating with AWS and Amazon Bedrock
during the summer of 2023 to delve into the possibilities of what generative AI can offer. The ongoing work is leading the PGA Tour towards the development
of exceptional experiences across a diverse range of platforms, unlocking additional value for the tour, the players, and their fans. Customers are also boosting employee productivity and creativity. Ryanair crew scheduling is
currently managed by a team of operation planners at their
headquarters, who are tasked with ensuring that planes
and employees depart and return on time, taking into
account training, holidays, and weather disruptions. In partnership with
AWS, they created an app that helps Ryanair cabin
crew manage their work lives in one place, using an uncomplicated tool in the palm of their hand. And companies like Adidas
China are using generative AI for process optimization. Adidas China wanted to improve
their inventory management with visual components
like backgrounds or models, so they looked degenerative
AI for a solution. Together with AWS, they
utilized product data to create a virtual try-on solution, generating lifelike models
and proper backgrounds for Adidas products
based on Stable Diffusion and controllable generative AI tools. But with so much choice,
so much opportunity, how do you choose? How do you know where to start? When it comes to prioritizing use cases, the most important thing is to find what you can implement
quickly and get building. You need to weigh the risks and requirements of each
opportunity to get early wins with the technology. By implementing these solutions, you can kickstart a flywheel of innovation within your organization. Early success drives
excitement and buy-in, which makes more complex and more unique and innovative use cases
all that more achievable. And that brings us to the
second generative AI foundation, which is the importance of using your data to customize your generative AI solutions. When you want to build
generative AI applications that are unique to your business needs, your organization's data
will be your differentiator. When you think about it,
every company has access to the same foundation models, but companies that will be successful in building generative AI applications with real business value are those that will do
so using a diverse set of robust data. You see, data is the difference between generic generative AI applications and those that know your business
and your customers deeply. But how can you put this data to work? What data foundations
do you need in place? First and foremost is
having a data strategy for your organization. 93% of CDOs acknowledged the importance of an end-to-end data
strategy, and its role in making their generative
AI initiatives a success. But along with the data strategy is also the quality of the data. The quality of that data
matters in generative AI, because higher quality
data improves the accuracy and the reliability of the model response. In a recent survey of Chief Data Officers, almost half of CDOs viewed data quality as one of their top challenges to implementing a generative AI strategy. And data quality and data
strategy are key because, well, data is growing
at an incredible rate, powered by consumer activity,
business analytics, sensors, and so many other drivers. That data growth is driving
a flywheel for generative AI. Foundation models are
trained on massive data sets, and then companies are using smaller private
enterprise data sets for additional customizations of foundation model responses and learning and creating new intermediate data sets. These customized models will in turn drive more generative AI applications, which through more interactions
creates even more data and even more data for that flywheel. And that customization, using your data to develop
generative AI applications for your customers and employees is vital. Let's talk about three popular approaches for building generative AI
solutions with your own data, and understanding which
one is right for you and your use case. Things like speed, accuracy,
cost, and complexity are all factors. And you'll need to
consider the business value for each situation and
weigh the benefits of each to determine what makes the most sense. These three techniques are commonly ranked by their level of complexity from the easiest to the most complex. Many customers use prompt engineering. It is a simple, cost-effective process that lets you refine your
inputs for generative AI so that you can get on-target
outputs and optimal results. You can customize the
outputs of an existing model with retrieval-augmented generation, sometimes referred to as RAG, without the need for
retraining that model. With RAG, the external data used to come from your augmented prompts can come from multiple data sources, including document repositories,
databases, and APIs. RAG helps the model to adjust its output with data retrieved as needed from these knowledge libraries. Next is fine tuning. When you fine tune an
existing foundation model, you're using a smaller sample from your own domain-specific data, basically creating a new model
with your prepared dataset. And finally, continued pre-training. Continued pre-training means you pick up where the foundation
model provider left off, training the model on data
sets in your enterprise to extend both the generalized and specialized knowledge of that model. And if you haven't tried
using Amazon Bedrock to customize foundation
models, well, you should. It has a lot of capabilities and is evolving at a super rapid rate. And its supports all three
of these capabilities to customize your model
responses with your own data. With use case selected and data prepared, that brings us to the third foundation, building with the most
comprehensive set of capabilities for generative AI. From startups to enterprises,
organizations of all sizes are getting started with generative AI. They want to take the
momentum they're building with early experiments and turn it into real
world productivity gains and innovations. At AWS, we're ready to help
you reinvent with generative AI because we think differently
about what it takes to meet your needs. And we reinvent again and again and again to help you deliver. We think about generative AI
as having three macro layers, and they're all equally important and we are investing in all of them. The bottom layer is the infrastructure used to train foundation models and run these models in production. The middle layer provides access to all of the large language models and foundation models you need, and to the tools you need to build and scale generative AI applications. At the top layer, we
have applications built leveraging foundation models, so you can quickly take
advantage of generative AI without any specialized knowledge. Let's start with the bottom
layer, the infrastructure. With foundation models, there are two main types of workloads, training and inference. Training is to create and
improve foundation models by learning patterns from
large amounts of training data. And inference uses those models to generate an output such
as text, images or video. These workloads consume massive
amounts of compute power. To make generative AI use
cases economical and feasible, you need to run your
training and inference on incredibly performant,
cost-effective infrastructure that is purpose built for machine learning and artificial intelligence. GPUs are chips that can
perform a high volume of mathematical
calculations simultaneously, making them popular for workloads like machine learning
simulations and 3D rendering. If you're building your own models, AWS is relentlessly focused on providing everything you need, the best chips, the most
advanced virtualizations, powerful petabyte scale
networking capabilities, hyperscale clustering, and the right tools to help you build. Along with the infrastructure, you need to have the tools to build with large language models
and foundation models. That's why we're investing in
the middle layer of the stack. We know many of you need it to be easier to access a powerful and diverse set of large language models and other foundation models, and then to quickly build
applications with them, all while maintaining
security and privacy. And that's why we built Amazon Bedrock. Bedrock is the easiest way to build and scale generative AI applications with large language models
and other foundation models. Customers in every industry
are already using Bedrock to reinvent their user
experiences, their products, and their processes, and to bring AI into the
heart of their business. Why Bedrock? You enjoy the broadest choice
of models, many of which are available first or only on Bedrock. You can add your own business
context quickly, easily, privately, and with the broadest selection of customization options. And you get enterprise
grade security and privacy because we designed it
that way from day one. Customer excitement has been overwhelming. AWS is investing in the middle layer of the generative AI stack to make it easier for
organizations to access powerful and a diverse set of large language models and other foundation models, and to quickly customize those models, all while maintaining
security and privacy. We see a huge opportunity
to help with that by infusing generative AI
into systems people use in their daily lives. We believe generative AI should help everyone at work seamlessly. That means helpful, relevant assistance, whether or not they know the first thing about foundation models, RAG,
or any of the rest of it. And that brings us to
the top of the stack, applications that you use that
are powered by generative AI, leveraging foundational models. We believe generative AI
has the potential over time to transform virtually every
customer experience we know. It can pour through the nooks
and crannies of a system and find data you never
would've known had existed, and it can help you put it to optimal use. It can generate insights that
save hours of manual work, and it can take projects that were slogs and make them snaps. To bring generative AI
to everyone at work, last year we released Amazon Q, a new type of generative
AI-powered assistant designed to work for you at work. Q lets you get answers quickly with natural language interactions. You can easily chat, generate content, and take actions, all
informed by an understanding of your systems, your data repositories, and your operations. And of course, we know how
important rock-solid security and privacy are to your business. Amazon Q can understand and respect your existing identities, roles, and permissions. If a user does not have permission to access something without Amazon Q, they can't access it using Q either. We have designed Amazon Q to meet enterprise customers'
stringent requirements from day one. And we never use business
customers' content from Amazon Q to train underlying models. We've also announced Amazon CodeWhisperer. CodeWhisperer uses generative
AI to allow developers to build faster, while freeing them up to focus on more creative
aspects of coding. It uses a foundation model to radically improve
developer productivity. It generates code suggestions in real time based on a developer's
comments in a natural language. I have to say, there's few things as CIO that have made me stop and say, wow. But the productivity gains from
CodeWhisperer do just that. It's unbelievable. As we continue to reinvent for
our customers, we've learned what is needed to bring generative AI to your customers and employees. You need the right capabilities to build performant,
cost-effective infrastructure. You need a secure,
private, easy way to build and scale powerful new applications
with foundation models. You need generative AI
applications with capabilities that could be enriched with
the context of your business. And it's all on AWS,
built with our high bar for giving you broad and
deep capabilities, choices, and enterprise readiness. It's still so early in the game. And we're incredibly excited
about what we can do together With the tools available to
start building, it brings us to the fourth and final foundation for all of those using this technology. You need to take the steps
to innovate responsibly with generative AI. Our commitment to develop
generative AI in a responsible way is integral to our approach at AWS, and to do so, we focus on four key areas. First, we want to help
transform responsible AI from theory into practice,
and help operationalize it across key elements of responsible AI. Second, responsible AI is an integral part of an entire end-to-end life
cycle of a foundation model, including design and
development, deployment, and ongoing use. It is not something that
can be done in a silo, but instead it must be
integrated across the lifecycle with a commitment to test, test, test, for accuracy, fairness, and other key responsible AI dimensions across all of your models. Third, you need to prioritize education around how generative AI works
and what its limitations are. Finally, advance the science behind developing
generative AI responsibly. At AWS, we know that
generative AI technology and how it will be used
will continue to evolve, posing new challenges. Together with academic, industry,
and government partners, we are committed to the
continual development of generative AI in a responsible way. Altogether, we want to build
generative AI that is safe, trustworthy, and a force for good. AWS deeply believes in responsible AI, and you'll see that
belief in our products. I'll touch on just a few examples here. First, transparency. Transparency means explicitly
sharing the details and context to responsibly
use your service. AI service cards are a form of
responsible AI documentation that provide customers with a single place to find information on
the intended use cases and limitations, responsible
AI design choices, and the deployment and performance
optimization best practices for our AI services. We recently announced
six new AI service cards, including Amazon Titan
and AWS HealthScribe. And as part of responsible AI, AWS is providing intellectual
property indemnity coverage for outputs of Amazon Titan models and Amazon CodeWhisperer. So if you use one of the
generative AI applications or models from AWS that
I just talked about and someone sues you for IP infringement, AWS will defend that lawsuit, which includes covering
any judgment against you or settlement costs. It's part of responsible AI,
and it's part of how we work. We are also investing in
automating the adoption of responsible AI practices. We've announced a preview of
Amazon Bedrock Guardrails, which is a new capability in Bedrock. It makes it easy to implement
application-specific safeguards based upon your responsible AI practices. Guardrails lets you
specify topics to avoid and automatically filter
out queries and responses in restricted categories. For example, an online banking
application can be set up to avoid providing investment advice and limit inappropriate
content such as hate speech, profanity, or violence. Another way to automate responsible AI is to build it into your
foundation model selection. With model customization
in Amazon Bedrock, you can evaluate and select
your foundation model based upon key responsible AI dimensions like accuracy, robustness, and toxicity. Customers can also use the same
model evaluation capability in Amazon SageMaker Clarify. And that automated control also applies to our generated content from our Amazon Titan
Image foundation model, which will contain an invisible watermark. A huge factor to enabling
responsible AI is you. Your goal is to be your
own toughest auditor, so you're prepared for
whatever comes your way for compliance in the future. AWS offers services like AWS CloudTrail, Amazon CloudWatch, Amazon DataZone, and Amazon OpenSearch Service. They're all designed to help you establish
the right governance, the right auditing structure, and the right processes
for your organization. Lastly, and maybe most
importantly, when you use your data to customize generative
AI for your business, your generative AI tools and
services should be architected so your most valuable IP, your data, remains protected and private when you use customized foundation models to meet the demands of your organization. For example, Amazon Bedrock
makes a separate copy of the base foundation model that is accessible only to the customer, and trains this private copy
of the model with your data. This means your data is never exposed or used to train the original
base foundation models. Your data stays contained in your own isolated virtual private
cloud environment, and it will be encrypted. And it means that you will have
the same AWS access controls that you have with any other AWS service, with over 300 security services and features built to support you. We have covered a lot of ground here today on how you can find success
on your generative AI journey. There is still so much more to learn and we are at just the
start of this journey. I think it's important to
note that it's not just about how these models learn about our businesses
through our data. It is how we as individuals and
organizations learn as well. It's about learning
generative AI techniques, learning to navigate the
world of data transparency and responsible AI, and incorporating those learnings
about this new technology to help our businesses, our
users, and our communities. At Amazon, we have a leadership principle called Learn and Be Curious. It says that we are never done learning. We're always seeking to improve ourselves. And we are curious about new possibilities and act to explore them. This event was made for
those of you who are curious, those of you who want to learn, and those of you who are ready
to unlock the power of data and generative AI. We're excited to see what you build and AWS is with you every step of the way. Thank you.