🎁Amazon Prime 📖Kindle Unlimited 🎧Audible Plus 🎵Amazon Music Unlimited 🌿iHerb 💰Binance
Video
Transcript
hello everyone my name is Naja Rhodes I
was a software engineer at Microsoft
working on web services currently at
large I guess and I dedicated my summer
as a scholar to language specifically
generative models language why would
that be I’m really intrigued by artistic
creative applications of machine
learning computer vision is where the
coolest efforts towards this kind of
thing is focused in my opinion and it’s
no wonder why I’m playing with visuals
and imagery is innately memorized
mesmerizing and fun procurements
meanwhile I feel like there’s a
conspicuous lack of text-based creative
projects although maybe I’m following
the wrong people on Twitter so please do
let me know if you know any about any of
cool projects or cool people working in
creative NLP because I’d love to hear
more find me at the demo tables but my
suspicion about the imbalance is that
text has a fundamental challenge it can
be really hard to make sense of erratic
outputs as I learned this summer it can
be literally mine
mind-numbing in their incoherence and
instead of interesting like at all so I
read a lot of bad bad samples this
summer and then it’s kind of because
there’s different kinds of failures
right with them images there’s this cool
project called text image where you type
a caption and then the Gann tries to
generate an image that matches that
caption so you can see that you know
when it tries to generate a cat sitting
on a windowsill in a room you kind of
get something cat like there’s some fur
a general shape I don’t know where the
head is but you know at least it’s kind
of still compelling like you can still
look at it even though it’s technically
a failure both tests it’s like this is
the kind of stuff I was looking at
I’m gives the unknown token and yeah and
then a bunch of like repetitive stuff
and I like this dr. Seuss quote but
unfortunately it didn’t quite hold up in
my summer because I was reading a lot of
garbage generations but you know every
once in a while it would give me
something that was you know somewhat
cohered
like a deep house too so it’s kept
trying to talk about music house music
it’s trying to say something about house
music what exactly I’m not completely
sure but it’s kind of delightful like
you can kind of try to get a sense of
what it’s trying to say and so my goal
for the summer of tech generation what’s
the aim for good descriptive meaningful
generations but if all else fails at the
very least reach this kind of level of
delightful yet coherent so general text
but of what kinds the final project idea
that drove my NLP things this summer is
what I call deep pipe pot it’s my
Twitter bot for all your generating good
music commentary and the idea was to
automatically detect tweets about songs
obtain interesting attributes about the
songs from Spotify and then use that to
condition the language model and feed
that into my model and produce some sort
of coherent commentary about the song
and this idea was largely thanks to an
inspiring data source called the hype
machine it’s a music blog aggregator
throughout undergrad and ever since I
kind of relied on it because it has this
collection of small music blogs that it
gathers and then has these charts and
you can play music click through look at
the different blogs so early in the
summer I wrote up some API calling web
scraping rate limited Python code for
collecting this training data I’m
extracting it cleaned up over about a
hundred thousand sentences and so to get
into the deep learning a little bit my
momma employs a conditional sequence to
sequence a variational Honor encoder
that’s a model language and why I like
this particular architecture is because
it first learns a richer representation
latent representation
of attacks and that it uses that
representation to generate new samples
and it works at a more macro or global
level because it encodes the entire
sentence versus like an LS TM which is
taking a history historical context and
then trying to predict word in a word
word for word locally and the VA e also
introduced some variability in the
generation process hopefully leading to
a little bit more novelty because it
randomly samples in the Leighton space
and mine was conditional in particular
because I wanted to provide some knockin
on text or context in particular I
wanted to use genre information and
Spotify has this cool API that gives you
these really specific genres like paper
soul optimism very hipster but it was
pretty good at like pinpointing exactly
what kind of music was going on so in
addition to the knowledge of general
past his music writing it could also use
a little bit of knowledge that wasn’t
maybe in the text and then once I had
the VA II I refined it with something
called a latent constraints can
generative adversarial Network or else
again we were calling it it helps
control aspects of the text that’s
generated by kind of letting you choose
what qualifies as a satisfying sample
because some is right here this blue
circle it represents the prior which is
the entire latent space learned by the
VA II most VA ease will completely learn
to use that entire waiting space so the
green blobby area is where like the most
realistic samples kind of lie and then
the red blob is where you decide oh yeah
these are the kinds of samples that I
like out of the dataset and so when you
apply the LC again it translates the
stuff from the realistic part of the
space into that more red part of the
space that’s you know for my particular
case I wanted more flowery descriptive
language instead of like stuff about
maybe the artist
thing that might be this kind of data
set so what’s nice about this is that
there was no reach retraining of the v8
you required it was more of a fine
tuning process and yeah so and the other
things that you need something that can
you need something so you need to be
able to pick out what’s flowering and
what’s not basically and you could do
that by hands but I decided to use a
topic modeling to do that because my
hypothesis was that with topic modeling
I could distill the commentary into
different types and as you can see I
made for different topic groups there it
ranges from topic one which has stuff
that says like beginning when driving
drums and famous song very vocal
harmonies that’s the kind of feel I was
trying to get out of my generations
versus like topic model thirty up here
which is just like tour dates and stuff
um I did keeps and I did keep some of
this stuff in the data set even though
it’s not exactly what I wanted to
generate just because it helps to have
data that the model can just generally
get a feel for English so you don’t want
to like limit the data set too much and
just pinpoint the stuff that you want
necessarily
so just real quick this was about the
the Twitter bot appointment pipeline
that I had so it’ll go from like a tweet
about pumped up kicks’ for example send
it to Spotify get some johner
information from it feed into the game
and then come up with something that’s
like kind of related it’s really catchy
but the kind that hooks in those sounds
like Pumped Up Kicks to me and yeah this
was the most liked tweet so far it’s
like the ployed at the iPod but I will
say that there are a lot of mad
generations like it’s like so basically
I had to like feed this stuff into a
spreadsheet and then the human curator
which is me he gets to pick the best
ones too sensitive Twitter looking
forward it would be cool if I could take
these kinds of lights and feed it back
into the model and say people tend to
like this kind of thing so I’m gonna
give it more of this also called human
preferences sci-fi also has this cool
API called audio features that measures
aspects of a song like dance ability
energy levels tempo valence that’ll be
cool instead of a genre perhaps and in
general in my future I’d like to do more
creative coding and more stuff with
language and it’ll be so I’d like to
thank my mentor Natasha Jacques she’s in
London right now some shout out some
women and the other open a scholars and
urban area and general thank you for
supporting this program thank you
[Applause]
oh yes
yeah absolutely yeah she asked if I had
considered feeding and the likes
retweets as some sort of reward and
reinforcement learning yeah my mentor is
super into reinforcement learning I
didn’t quite get that far this summer
but it would be a cool reward signal for
sharing Thanks yes
right
I think so because yeah because you Sam
going from late this late in space you
can give it any kind of random latent
vector and like I was showing in that
little blob like there’s a lot of space
where it’s just going to be garbled so
but that’s the space that you’re working
with and so that again was supposed to
kind of help with that in that I could
then tell it okay but these are the
kinds of thought vectors that are still
realistic and still understandable but
also creative and like novel so yeah
it’s definitely when a patient of the
Dae but and that’s why I kind of put the
game on top to try this help counter
that oh yeah I mentioned mentioned that
like hands-on texted hard in general and
I’ve made it seem like maybe it’s the
thing that people do but not quite it
turns out that like text isn’t the
furniture rule when you do back
propagation or whatever yourself like
but the thing that I could do was
differentiate these effects are and that
works thank you
[Applause]