🎁Amazon Prime 📖Kindle Unlimited 🎧Audible Plus 🎵Amazon Music Unlimited 🌿iHerb 💰Binance
Video
Transcript
joining us today.
Please welcome to the stage Sam
Altman.
[music]
[cheers and applause]
Good morning.
Welcome to our first ever
OpenAI DevDay.
We’re thrilled that you’re here
and this energy is awesome.
[cheers and applause]
And welcome to San Francisco.
San Francisco has been our home
since day one, the city is
important to us and to the tech
industry in general.
We’re looking forward to
continuing to grow here.
So we’ve got some great stuff to
announce today, but first, I’d
like to take a minute to talk
about some of the stuff that
we’ve done over the past year.
About a year ago, November 30th
, we shipped ChatGPT as a
low-key research preview, and
that went pretty well.
[laughter]
In March we followed that up
with the launch of GPT-4,
still the most capable model out
in the world.
[applause]
And in the last few months, we
launched voice and vision
capabilities so that ChatGPT
can now see, hear, and speak.
[applause]
There’s a lot, you don’t have to
clap each time.
[laughter]
More recently we launched DALL•E
3, the world’s most advanced
image model.
You can use it, of course,
inside of ChatGPT.
For our enterprise customers, we
launched ChatGPT Enterprise,
which offers enterprise grade
security and privacy, higher
speed GPT-4 access, longer
context windows, a lot more.
Today, we’ve got about 2 million
developers building on our API
for a wide variety of use cases,
doing amazing stuff.
Over 92% of Fortune 500
companies building on our
products, and we have about 100
million weekly active users now
on ChatGPT.
[applause]
And what’s incredible on that
is, we got there entirely
through word of mouth.
People just find it useful and
tell their friends.
OpenAI is the most advanced and
the most widely used AI platform
in the world now.
But numbers never tell the whole
picture on something like this.
What’s really important is how
people use the products, how
people are using AI.
So I’d like to show you a quick
video.
I actually wanted to write
something to my dad in Tagalog.
I want a nonromantic way to tell
my parent that I love him and I
also want to tell him that he
can rely on me, but in a way
that still has the respect of,
like, a child-to-parent
relationship that you should
have in fill I teen zero culture
and in taking a long.
I love you very deeply and I
will be with you no matter where
the path he leads.
I see so many possibilities,
I’m like, who he, sometimes I’m
not sure about some stuff, and I
feel like the actual
ChatGPT – just thinking about
giving it more confidence.
The first thing that blew my
mind was that it levels with
you.
That’s something that a lot of
people struggle to do.
It opened my mind to just what
every creative could do if they
just had a person helping them
out who listens.
So this is to represent
circulating hemoglobin –
And you built that with
ChatGPT.
ChatGPT built it with me.
I started using it for daily
activities like, hey, here’s a
picture of my fridge, can you
tell me what I’m missing because
I’m going grocery shopping and I
really need to do recipes that
are following my Vegan diet.
As soon as we got access to
Code Interpreter, I was like,
wow, this thing is awesome.
It can build spreadsheets.
It can do anything.
I discovered about – on my h
birth date.
Very friendly, very patient,
very knowledgeable, and very
quick.
It’s been a wonderful thing.
I’m a 4.0 student but I also
have four children.
When I started using ChatGPT,
I realized I could ask ChatGPT
that question, and not only does
it give me an answer, but it
gives me an explanation.
Didn’t need computer go as
much.
It gave me a life back.
I gave me time for my family and
time for me.
I have a chronic nerve pain
on my whole left half of my
body, nerve damage.
I had like a spine – brain
surgery.
I have limited use of my left
hand.
Now you can just have the
integration of voice input, and
the newest one where you can
have the back-and-forth
dialogue, that’s just like
maximum best interface for me.
It’s here!
[music]
[applause]
So we love hearing the stories
of how people are using the
technology.
It’s really why we do all of
this.
Okay, so now on to the new
stuff, and we have got a lot.
[cheers and applause]
First, we’re going to talk about
a bunch of improvements we’ve
made, and then we’ll talk about
where we’re headed Next.
Over the last year, we spent a
lot of time talking to
developers around the world.
We’ve heard a lot of your
feedback.
It’s really informed what we
have to show you today.
Today, we are launching a new
model.
GPT-4 Turbo.
[cheers and applause]
GPT-4 Turbo will address many
of the things that you all have
asked for.
So let’s go through what’s new.
We’ve got six major things to
talk about for this part.
Number one, context length.
A lot of people have tasks that
require a much longer context
length.
GPT-4 supported up to 8k and
in some cases up to 32k
context length but we know that
isn’t enough for many of you and
what you want to do.
GPT-4 Turbo supports up to
128,000 tokens of context.
[cheers and applause]
That’s 300 pages of a standard
book, 16 times longer than our
8k context.
And in addition to longer
context length, you’ll notice
that the model is much more
accurate over a long context.
Number two, more control.
We’ve heard loud and clear that
developers need more control
over the model’s responses and
outputs, so we’ve addressed that
in a number of ways.
We have a new feature called
JSON load which ensures that the
model will respond with valid
JSON.
This has been a huge developer
request, it will make calling
APIs much easier.
The model is also much better at
function calling.
You can now call many functions
at once.
It will do better at following
instructions in general.
We’re also introducing a new
feature called reproducible
outputs.
You can pass the seed parameter
and it will make the model
return consistent outputs, which
gives you a higher degree of
control over model behavior.
This rolls out in beta today.
[cheers and applause]
And in the coming weeks, we’ll
roll out a feature to let you
view log probs in the API.
[cheers and applause]
Number three, better world
knowledge.
You want these models to the
access better knowledge about
the world, so do we.
We’re launching retrieval in the
platform.
You can bring knowledge from
outside documents or databases
into whatever you’re building.
We’re also updating the
knowledge cutoff.
We are just as annoyed of all of
you, probably more than, that
GPT’s knowledge of the world
ended in 2021.
We will try to never let it get
that out of date again.
GPT Turbo has knowledge of the
world up to April 2023 and we
will improve that over time.
Number four, new modalities.
Surprising no one, DALL•E 3,
GPT-4 Turbo with Vision, and
the new text-to-speech model are
all going to into the API
today.
[cheers and applause]
We have a handful of customers
that have just started using
DALL•E 3 to programmatically
generate images and designs.
Today, Coke is launching a
campaign that lets its customers
generate Diwali cards using
DALL•E 3, and our safety systems
help developers protect their
applications against misuse.
Those tools are available in the
API.
GPT-4 Turbo can now accept
images as inputs via the API,
can generate captions,
classifications, and analysis.
For example, Be My Eyes uses
this technology to help people
who are blind or have low vision
with their daily tasks like
identifying products in front of
them.
And with our new text-to-speech
model, you’ll be able to
generate incredibly natural
sounding audio from text in the
API with six preset voices to
choose from.
I’ll play an example.
Did you know that Alexander
Graham bell, the eminent
inventor, was enchanted by the
world of sounds?
His ingenious mind led to the
creation of the graphophone,
which etched sounds onto wax,
making voices Whisper through
time.
This is more natural than
anything else we’ve heard out
there.
Voice can make apps more natural
to interact with.
It unlocks a lot of use cases,
like language learning and voice
assistance.
Speaking of new modalities,
we’re also releasing the next
verse of our open source speech
recognition model, Whisper V3,
today, and it will be coming
soon to the API.
It features improved performance
across many languages and we
think you’re really gonna like
it.
Okay.
Number five, customization.
Fine-tuning has been working
really well for GPT-3.5 since we
launched it a few months ago.
Starting today, we’re going to
expand that to the 16k version
of the model.
Also starting today, we’re
inviting active fine-tuning
users to apply for the GPT-4
fine-tuning, experimental access
program.
To fine fine an API is great for
adapting our models to achieve
better performance in a wide
variety of applications with a
relatively small amount of data,
but you may want to model to
learn a completely new knowledge
domain or to use a lot of
proprietary data.
So today we’re launching a new
program called Custom Models.
With Custom Models, our
researchers will work closely
with a company to help them make
a great custom model, especially
for them, and their use case,
using our tools.
This includes modifying every
step of the model training
process, doing additional
domain-specific pre-training, a
post-training process tailored
to a specific domain.
We won’t be able to do this with
many companies to start, it will
take a lot of work and in the
interest of expectations, at
least initially it won’t be
cheap, but if you’re excited to
push things as far as they can
currently go, please get in
touch with us and we think we
can do something pretty great.
Okay.
And then number six.
Higher rate limits.
We’re doubling the tokens per
minute for all of our
established GPT-4 customers,
so that it’s easier to do more.
And you’ll be able to request
changes to further rate limits
and quotas directly in your API
account settings.
In addition to these rate
limits, it’s important to do
everything we can do to make
it – you successful building on
our platform.
We’re introducing Copyright
Shield.
Copyright Shield means that we
will step in and defend our
customers and pay the costs
incurred if you face legal
claims around copyright
infringement, and this place to
both ChatGPT Enterprise and
the API.
And let me be clear.
This is a good time to remind
people, we do not train on data
from the API or ChatGPT
Enterprise ever.
All right.
There’s actually one more
developer request that’s been
even bigger than all of these.
So I’d like to talk about that
now.
And that’s pricing.
[laughter]
GPT-4 Turbo is the industry
leading model.
It delivers a lot of
improvements that we just
covered, and it’s a smarter
model than GPT-4.
We’ve heard from developers that
there are a lot of things that
they want to build, but GPT-4
just costs too much.
They’ve told us that if we could
decrease the cost by 20, 25%,
that would be great, a huge leap
forward.
I’m super excited to announce
that we worked really hard on
this, and GPT-4 Turbo, a
better model, is considerably
cheaper than GPT-4 by a factor
of 3X for prompt tokens –
[applause]
And 2X for completion tokens,
starting today.
[cheers and applause]
So the new pricing is 1 cent per
thousand prompt tokens and $0.03
per thousand completion tokens.
For most customers that leads to
a blended more than 3.75%
cheaper to use.
We worked super hard to make
this happen.
We hope you’re as excited about
it as we are.
[cheers and applause]
So we’ve decided to prioritize
price first because we had to
choose one or the other but
we’re going to work on speed
next.
We know speed is important,
too.
Soon you will notice GPT-4
Turbo becoming a lot faster.
We’re also decreasing the cost
of GPT-3.5 Turbo 16k, also input
tokens for 3X less, and output
tokens are two, less.
Which means 16k is now cheaper
than the previous model.
Running a fine-tune GPT-3.5 16k
version is also cheaper than the
old version.
We just covered a lot about the
model itself.
We hope these changes address
your feedback , we’re really
excited to bring all of these
improvements to everybody now.
In all of this, we’re lucky to
have a partner who’s
instrumental in making it
happen.
So I’d like to bring out a
special guest, Satya Nadella,
the CEO of Microsoft.
Music.
[cheers and applause]
Welcome.
Thank you so much.
Thank you.
Satya, thanks so much for coming
here.
It’s fantastic to be here,
and, Sam, congrats.
I’m really looking forward to
Turbo and everything else that
you have coming is, it’s been
just fantastic partnering with
you guys.
Awesome.
Two questions, I won’t take too
much of your time.
How is Microsoft thinking about
the partnership currently?
First –
[laughter]
We love you guys.
Look, it’s been fantastic for
us.
In fact, I remember the first
time I think you reached out and
said, hey, do you have some
Azure credits, we’ve come a long
way from there.
Thank you for those.
That was great.
You guys have built something
magical.
There are two things for us when
it comes to the partnership.
The first is, these workloads
and even when I was listening
backstage to how you’re
describing what’s coming even,
it’s just so different and nut.
I’ve been in the infrastructure
business for three decades –
No one has seen
infrastructure like this.
The workload, the pattern of
the workload, the training jobs
are so synchronous and large and
data parallel.
The first thing we’ve been doing
is building in partnership with
you the system all the way from
thinking from power to the DC to
the rack, to the accelerators,
to the network, and just really
the shape of Azure is
drastically changed.
And it’s changing rapidly in
support of these models that
you’re building.
And so our job number one is to
build the best system so that
you can build the best models,
and then make that all available
to developers.
The other thing is, we ourselves
are developers, building
products.
My own conviction of this entire
generation of foundation models
has completely changed.
The first time I saw GitHub
Copilot on GPT.
And so we want to build our
Copilot, GitHub Copilot, all
as developers on top of OpenAI
APIs so we’re very, very
committed to that.
What does that mean to
developers?
I always think of Microsoft as a
platform company, a developer
company, and a partner company,
and so we want to make – for
example, we want to make
GitHub available –
GitHub Copilot available,
the enterprise he diddation to
all the attendees here so they
can try it out.
That’s awesome.
We’re very excited about
that.
[applause]
And you can count on us to build
the best infrastructure in Azure
with your API support, and bring
it to all of you, and then even
things like the Azure
marketplace, building out
products here to get to market
rapidly.
That’s sort of really our intent
here.
Great.
How do you think about the
future?
Future of the partnership or
future of AI or whatever.
[laughter]
Anything you want.
You know, like, there are a
couple of things for me that I
think are gonna be very, very
key for us.
One is, I just described how the
systems that are exceeded as you
aggressively push forward on
your roadmap, requires us to be
on the top of our game, and we
intend fully to commit ourselves
deeply to making sure you all,
as builders of these foundation
models, have not only the best
systems for training but the
most compute so you can keep
pushing forward.
We appreciate that.
On the frontiers because I
think that’s the way we’re going
to make progress.
The second thing I think we both
care about, in fact, quite
frankly, the thing that excited
both shades to come together is
your mission and ours.
Our mix is to empower every
person and organization on the
planet to achieve more, and
ultimately AI is only going to
be useful if it does empower.
I saw the video you played
earlier.
That was fantastic to see
those – hear those voices
describe what AI meant for them
and what they were able to
achieve.
So ultimately it’s about being
able to get the benefits of AI
broadly disseminated to
everyone, I think is going to be
the goal for us.
The last thing is we’re very
grounded in the fact that safety
matters and safety is not
something that you care about
later but it’s something we do
shift left on and we’re very,
very focused on that with you
all.
Great.
I think we have the best
partnership in tech, I’m excited
to be working together.
Thank you for coming.
Thank you.
[applause]
Okay.
So we have shared a lot of
great updates for developers
already, and we’ve got a lot
more to come, but even though
this is a developer conference,
we can’t resist making some
improvements to ChatGPT.
So, a small one, ChatGPT now
uses GPT-4 Turbo with all the
latest improvements including
the latest knowledge cutoff,
which we’ll continue to update,
that’s all live today.
It can now browse the web when
it needs to, write and run code,
analyze data, take and generate
images and much more, and we
heard your feedback that model
picker was extremely annoying,
that’s gone starting today.
You will not have to click
around the dropdown menu.
All of this will just work
together.
[cheers and applause]
ChatGPT will just know what to
use and when you need it.
But that’s not the main thing.
And neither was price, actually
the main developer request.
There was one that was even
bigger than that.
And I want to talk about where
we’re headed and the main thing
we’re here to talk about today.
So we believe that if you give
people better tools, they will
do amazing things.
We know that people want AI that
is smarter, more personal, more
customizable, can do more on
your behalf.
Eventually you’ll just ask a
computer for what you need and
it will do all of these tasks
for you.
These capabilities are often
talked in the AI field about as
agents.
The upsides of this are going to
be tremendous.
At OpenAI we really believe that
gradual iterative deployment is
the best way to address the
safety challenges with AI.
We think it’s especially
important to move carefully
towards this future of agents,
it’s going to require a lot of
technical work, and a lot of
thoughtful consideration by
society.
So today we’re taking our first
small step that moves us towards
this future.
We’re thrilled to introduce
GPTs.
GPTs are tailored versions of
ChatGPT for a specific
purpose.
You can build a GPT, a
customized version of ChatGPT,
for almost anything, with
instructions, expanded
knowledge, and actions, and then
you can publish it for others to
use.
And because they combine
instructions, expanded
knowledge, and actions, they can
be more helpful to you.
They can work better in any
context and they can give you
better control.
They’ll make it easier for you
to accomplish all sorts of tasks
or just have more fun and you’ll
be able to use them right
within ChatGPT.
You can in effect program a GPT
with language just by talking to
it.
It’s easy to customize the
behaviors so that it fits what
you want.
This makes building them very
accessible and it gives agency
to everyone.
So, we’re going to show you what
GPTs are, how to use them, how
to build them, and then we’re
going to talk about how they’ll
be distributed and discovered.
And then after that, for
developers, we’re going to show
you how to build these
agent-like experiences into your
own apps.
First, let’s look at a few
examples.
Our partners at Code.org are
working hard to expand computer
science in schools.
They’ve got a curriculum that is
used by tens of millions of
students worldwide.
Code.org crafted Lesson Planner
GPT to help teachers provide a
more engaging experience for
middle schoolers.
If a teacher asks it to explain
four loops in a creative way, it
does just that.
In this case, it will do it in
terms of a video game character,
repeatedly picking up coins,
super easy to understand for an
eighth grader.
As you can see, this GPT brings
together Code.org’s extensive
curriculum and expertise and
lets teachers adapt it to their
needs quickly and easily.
Next, Canva has built a GPT that
lets you start designing by
describing what you want in
natural language.
If you say, make a poster for a
DevDay reception this afternoon,
this evening, and you give it
some details, it will generate a
few options to start with by
hitting Canva’s APIs.
This concept may be familiar to
some of you.
We’ve evolved our plug-ins to be
custom actions for GPTs.
You can keep chatting with this
to see different iterations, and
when you see one you like, you
can click through to Canva for
the full design experience.
So now, we’d like to show you a
GPT live.
Zapier has built a GPT that lets
you perform actions across 6,000
applications to unlock all kinds
of integration possibilities.
I’d like to introduce Jessica,
one of our solutions architects,
who is going to drive this
demo.
Welcome, Jessica.
[cheers and applause]
Thank you, Sam.
Hello, everyone.
Thank you all.
Thank you all for being here.
My name is Jessica Shay, I work
with partners and customers to
bring their product to life.
Today I can’t wait to show you
how hard we’ve been working on
this, so let’s get started.
To start, where your GPT will
live is on this upper left
corner.
I’m going to start with clicking
on the Zapier AI actions.
And on the right-hand side, you
can see that’s my calendar for
today.
So it’s quite a day.
I’ve used this before so it’s
connected to my calendar.
To start, I can ask what’s on my
schedule for today.
We built GPTs with security in
mind, so before it performs any
action or shares data, it will
ask for your permission, so
right here I’m going to say
allowed, so GPT is designed to
take in your instructions, make
a decision on which capability
to call to perform that action,
and then execute that for you.
So you can see right here, it’s
already connected to my
calendar, it pulls in my
information, and then I’ve also
prompted it to identify
conflicts on my calendar.
You can see right here it
actually was able to identify
that.
So it looks like I have
something coming up.
So what if I want to let Sam
know that I have to leave
early?
Right here I say, let Sam know I
gotta go, chasing GPUs.
[laughter]
With that, I’m going to swap to
my conversation with Sam, and
then I’m going to say, yes,
please run that.
Sam?
Did you get that?
I did.
[applause]
Awesome.
So, this is only a glimpse of
what is possible, and I cannot
wait to see what you all will
build.
Thank you and back to you, Sam.
[cheers and applause]
Thank you, Jessica.
So those are three great
examples, in addition to these,
there are many more kinds of
GPTs that people are creating
and many, many more that will be
created soon.
We know that many people who
want to build a GPT don’t know
how to code.
We’ve made it so that you can
program the GPT just by having a
conversation.
We believe that natural language
is going to be a big part of how
people use computers in the
future, and we think this is an
interesting early example.
So I’d like to show you how to
build one.
All right.
So I’m going to create a GPT
that helps give founders and
developers advice when starting
new projects.
I’m going to go to create a GPT
here.
And this drops me into the GPT
builder.
I worked with founders for years
at YC and still, whenever I meet
developers, the questions are
always about how do I think
about a business idea, can you
give me some advice.
I’m going to see if I can build
a GPT to help with that.
So to start, GPT builder asks me
what I want to make, and I’m
going to say I want to help
start-up founders think through
their business ideas and get
advice.
After the founder has gotten
some advice, grill them –
[laughter]
On why they are not growing
faster.
[laughter]
All right.
So to start off, I just tell the
GPT a little bit about what I
want here, and it’s going to go
off and start thinking about
that, and it’s going to write
some detailed instructions for
the GPT.
It’s also going to ask me about
a name.
How do I feel about start-up
mentor?
That’s fine.
That’s good.
So if I didn’t like the name, of
course I could call it something
else but it’s going to try to
have this conversation with me
and start there.
And you can see here on the
right, in the preview mode, that
it’s already starting to fill
out the GPT, where it says what
it does, it has some ideas of
additional questions that I
could ask.
It just generated a candidate.
Of course I could regenerate
that or change it but I sort of
like that, so I will say, that’s
great.
And you see now that the GPT is
being built out a little bit
more as we go.
Now, what I want this to do, how
it can interact with users, I
can talk about style here but
what I’m going to say is, I am
going to upload transcripts of
some lectures about start-ups I
have given.
Please give advice based off of
those.
All right.
So, now it’s going to go figure
out how to do that, and I would
like to show you the configure
tab so you can see some of the
things that were built out here
as we were going by the builder
itself and you can see there’s
capabilities here that I can
enable.
I could add custom actions.
These are all feign to leave.
I’m going to upload a file.
Here’s a lecture that I gave
with some start-up advice, and
I’m going to add that here.
In terms of these questions,
this is a dumb one.
The rest of those are
reasonable.
And very much things founders
often ask.
I’m going to add one more thing
to the instructions here, which
is be concise and constructive
with feedback.
All right.
So, again, if I had more time,
I’d show you a bunch of other
things but this is like a decent
start, and now we can try it out
over on this preview tab.
So I will say – what’s a common
question?
What are three things to look –
what are three things to look
for when hiring employees at an
early stage start-up?
Now, it’s going to look at that
document I uploaded.
It will also have all of the
background knowledge of
GPT-4.
That’s pretty good.
Those are three things that I
definitely have said many
times.
Now, we could go on and it would
start following the other
instructions and grill me on why
I’m not growing faster, but in
the interest of time, I’m going
to skip that.
I am going to publish this only
to me for now.
I can work on it later, I can
add more content, I can add a
few actions that I think will be
useful, and then I can share it
publicly.
So that’s what it looks like to
create a GPT.
[applause]
Thank you.
By the way, I always wanted to
do that, after all of the YC
office hours, I thought, some
day I’ll make a bot that can do
this and that will be awesome.
[laughter]
With GPTs we’re letting people
easily share and discover all
the fun ways that they use
ChatGPT with the world.
You can make private GPTs like
I just did.
Or you can share your creations
publicly with a link for anyone
to use.
Or if you’re on ChatGPT
Enterprise, you can make GPTs
just for your company.
And later this month, we’re
going to launch the GPT Store.
You can list a –
[applause]
Thank you, I appreciate
that.
[applause]
You can list a GPT there, and
we’ll be able to feature the
best and the most popular
GPTs.
Of course, we’ll make sure that
GPTs in the store follow our
policies before they’re
accessible.
Revenue sharing is important to
us.
We’re going to pay people who
build the most useful and the
most used GPTs a portion of
our revenue.
We’re excited to foster a
vibrant ecosystem with the GPT
Store just from what we’ve been
building ourselves over the
weekend, we’re confident there’s
going to be a lot of great
stuff, we’re excited to share
more information soon.
Those are GPTs, and we can’t
wait to see what you’ll build.
But this his a developer
conference and the coolest thing
about this is we’re bringing the
same concept to the API.
[applause]
Many of you have already been
building agent-like experiences
on the API.
For example, Shopify Sidekick,
which lets you take actions on
the platform, Discord’s Clyde,
lets Discord moderators create
custom personalities for, and
Snap’s My AI, a custom island
chatbot that can be added to
group chats and make
recommendations.
These experiences are great but
they have been hard to build,
sometimes taking months, teams
of dozens of engineers, there’s
a lot to handle to make this
custom assistant experience.
So today we’re making it a lot
easier with our new Assistants
API.
[cheers and applause]
The Assistants API includes
persist tents threads so they
don’t have to figure out how to
deal with long conversation
history, built-in retrieval,
Code Interpreter, a working
Python interpreter in a sandbox
environment, and of course the
improved function calling that
we talked about earlier.
So we’d like to show you a demo
of how this works and here is
Romain, our head of developer
experience.
Welcome.
[music]
[applause]
Thank you, Sam.
Good morning.
Wow.
It’s fantastic to see you all
here.
It’s been so inspiring to see so
many of you infusing AI into
your apps.
Today, we’re launching new
modalities in the API, but we
are also very excited to improve
the developer experience for you
all to build assistive agents.
So let’s dive right in.
Imagine I’m building Wanderlust,
a travel app for global
explorers and this is the
landing page.
I’ve actually used GPT-4 to
come up with these destination
ideas, and for those of you with
a keen eye, these illustrations
are generated programmatically
using the new DALL•E 3 API
available to all of you today.
So it’s pretty remarkable.
But let’s add a very simple
assistant to it.
This is the screen, we’ll come
back to it in a second.
I’m going to switch over to the
assistants playground.
Creating an assistant is easy,
you give it a name, some initial
instructions, the model, GPT-4
Turbo, and I’ll go ahead and
select tools.
I’ll turn on Code Interpreter
and retrieval and save.
And that’s it.
Our assistant is ready to go.
Next I can integrate with two
new primitives of this
Assistants API, threads and
messages.
Let’s take a quick look at the
code.
The process here is very
simple.
For each new user, I will create
a new thread, and as the users
engage with their assistant, I
will add their messages to the
threads, very simple.
And then I can simply run the
assistant at any time to stream
the responses back to the app.
So we can return to the app and
try that in action.
If I say, hey, let’s go to
Paris, all right.
That’s it.
With just a few lines of code,
users can now have a very
specialized assistant right
inside the app.
And I’d like to highlight one of
my favorite features here,
function calling.
If you have not used it yet,
function calling is really
powerful.
As Sam mentioned, we’re taking
it a step further today.
It now guarantees the JSON
output with no added latency and
you can invoke multiple
functions at once.
If I say, what are the top 10
things to do, I’m going to have
the assistant respond to that
again.
And here what’s interesting that
the assistant knows about
functions, including those to
annotate the map that you see on
the right, and now all of these
pins are dropping in real-time
hire.
[cheers and applause]
It’s pretty cool.
And that integration allows our
natural language interface to
interact fluidly with components
and features of our app, and it
truly showcases now the harmony
you can build between AI and UI
when the assistant is actually
taking action.
But let’s talk about retrieval.
And retrieval is about giving
our assistant more knowledge
beyond these immediate user
messages.
I got inspired and already
booked my tickets to Paris so
I’m going to drag and drop this
PDF.
While it’s uploading I can sneak
peek at it, typical united
flight ticket, and behind the
scene here, what’s happening is
that retrieval is reading these
files and, boom, the information
about this PDF appeared on the
screen.
[cheers and applause]
And this is of course a very
tiny PDF but assistants can
parse from documents, from
extensive texts to intricate
product specs depending on what
you’re building.
I booked an AirBNB so I’m going
to drag that over to the
conversation as well.
We’ve heard from so many of you
developers how hard that is to
build yourself.
You typically need to compute
your on biddings, set up
chunking algorithm, now all of
that is taking care of.
There’s more than retrieval.
With every API call, you usually
need to resend the entire
conversation history, which
means, you know, setting up a
key value store, that means
handling the context windows,
serializing messages and so
forth.
That complex it now completely
goes away with this new stateful
API.
Just because OpenAI managing
this API does not mean it’s a
black box.
In fact, you can see the steps
that the tools are taking right
inside your developer
dashboard.
So here if I go ahead and click
on threads, this is the thread I
believe we’re currently working
on, and these are all the steps,
including the functions Building
Coded with the right parameters
and the PDFs I’ve just
uploaded.
Let’s move on to a new
capability that many of you have
been requesting for a while.
Code Interpreter is now
available today in the API as
well.
That gives the AI the ability to
write and execute code on the
fly but even generate files.
So let’s see that in action.
If I say here, hey, we’ll be
four friends staying at this
AirBNB.
What’s my share of it plus my
flights?
All right.
Now here what’s happening is
that Code Interpreter noticed
that it should write some code
to answer this query so now it’s
computing, you know, the number
of days in Paris, number of
friends, it’s doing some
exchange rate calculations
behind the scenes to get this
answer for us.
Not the most complex math but
you get the picture.
Imagine you’re building a very
complex finance app that’s
crunching countless numbers,
plotting charts, so, really, any
task that you’d normally tackle
with code, Code Interpreter will
work great for you.
I think my trip to Paris is
sorted.
To recap here, we’ve seen how
you can quickly create an
assistant that manages states
for your user conversations,
leverages external tools like
knowledge and retrieval and Code
Interpreter and I know vocation
your own functions to make
things happen.
But there’s one more thing I
wanted to show you to really
open up the possibilities using
function calling defined with
our new modalities that we’re
launching today.
While working a DevDay, I’ve
built a small custom assistant
that knows everything about this
event.
But instead of having the chat
interface while running around
all day today, I thought, why
not use voice instead.
So let’s bring my phone up on
screen hear so y so
you can see it.
On the right you see a simple
swift app that takes microphone
input.
I’m going to bring up my
terminal log so you can see
what’s happening behind the
scenes.
Let’s give it a shot.
Hey there, I’m on the keynote
Stage Right now, can you greet
our attendees here at DevDay?
Hey, everyone, welcome to
DevDay, it’s awesome to have you
all here.
Let’s make it an incredible
day.
[cheers and applause]
Isn’t that impressive?
You have six unique voices to
choose from in the API, each
speaking multiple languages so
you can find the perfect fit for
your app.
On my laptop, you can see what’s
happened behind the scene.
I’m using whisper to converts
the voice input into text, and
the new SSI to make it speak.
Function calling, things get
even more interesting when the
assistant can connect to the
internet and take real actions
for users.
So let’s do something A even
more exciting here.
How about this?
Assistant, can you randomly
select five DevDay attendees
here and give them $500 in
OpenAI credits.
Yes.
Checking the list of attendees.
[laughter]
Done.
I picked five DevDay attendees
and added $500 of API credits to
their account.
Congrats to: (Reading name).
If you recognized yourself,
awesome, congrats.
That’s it, a quick overview
today of the new Assistants API
combined with new tools and
modalities we launched, all
starting with the simplicity of
a rich text or voice
conversation for you end users.
We really can’t wait to see what
you build and congrats to our
lucky winners.
Actually, you know what?
You’re all part of this amazing
OpenAI community here so I’m
going to talk to my assistant
one more time before I step off
the stage.
Hey, assistant, can you actually
give everyone here in the
audience $500 in OpenAI
credits?
[cheers and applause]
Sounds great.
Let me go through everyone.
[cheers and applause]
All right.
That function will keep running,
but I’ve run out of time so
thank you so much, everyone,
have a great day.
Back to you, Sam.
[cheers and applause]
Pretty cool, huh?
[cheers and applause]
So that Assistant API goes into
beta today and we’re super
excited to see what you all do
with it.
Anybody can enable it.
Over time, GPTs and assistants
are precursors to agents, are
going to be able to do much,
much more.
They’ll gradually be able to
plan and to perform more complex
actions on your behalf.
As I mentioned before, we really
believe in the importance of
gradual iterative deployment.
We believe it’s important for
people to start building with
and using these agents now to
get a feel for what the world is
going to be like as they become
more capable.
And as we’ve always done, we’ll
continue to update our systems
based off of your feedback.
So, we’re super excited that we
got to share all of this with
you today.
We introduced GPTs, custom
versions of ChatGPT that
combine instructions, extended
knowledge and actions.
We launched the Assistants API
to make it easier to build
assistive experiences with your
own apps.
These are our first steps
towards AI agents and we’ll be
increasing their capabilities
over time.
We introduced a new GPT-4
Turbo model that delivers
improved function calling,
knowledge, lowered pricing, new
modalities and more.
And we’re deepening our
partnership with Microsoft.
In closing, I wanted to take a
minute to thank the team that
creates all of this.
OpenAI has got remarkable talent
density but it takes a huge
amount of hard work and
coordination to make this
happen.
I truly believe I’ve got the
best colleagues in the world.
I feel incredibly grateful to
get to work with them.
We do this because we believe AI
is going to be a technological
and societal revolution, will
change the world in many wakes,
and we’re happy to get to work
on something that will empower
all of you to build so much for
all of us.
We talked about earlier how if
you give people better tools,
they can change the world.
We believe that AI will be about
individual empowerment and
agency at a scale that we’ve
never seen before, and that will
elevate humanity to a scale that
we’ve never seen before,
either.
We’ll be able to do more, to
create more and to have more.
As intelligence gets integrated
everywhere, we will all have
superpowers on demand.
We’re excited to see what you
all will do with this
technology, and to discover the
new future that we’re all going
to architect together.
We hope that you’ll come back
next year.
What we launch today is going to
lack very quaint to what we’re
creating for you now.
Thank you for all that you do.
Thank you for coming here