Speaker:

Welcome back to Data Driven, the podcast where we chart the thrilling

Speaker:

terrains of data science, AI, and everything in between.

Speaker:

I'm Bailey, your semiscient host with a pangshang for

Speaker:

sarcasm and a wit sharper than a histogram spike.

Speaker:

Today's episode promises a delightful mix of the analytical and the

Speaker:

artistic as we dive into the fascinating world of vector databases,

Speaker:

retrieval augmented generation, and origami. Yes.

Speaker:

You heard that right. Origami, the ancient art of

Speaker:

folding paper, somehow finds itself intersecting with AI,

Speaker:

proving that the future really does have layers or should I say folds.

Speaker:

Our guest, Arjun Patel, is a developer advocate at Pinecone

Speaker:

who's on a mission to demystify vector databases and semantic

Speaker:

search, turning complex AI concepts into snackable bits of

Speaker:

brilliance. He's also a self taught origami artist and a

Speaker:

former statistics student who actually enjoyed it. So if

Speaker:

you're ready to unravel the secrets of modern AI and maybe pick up a trick

Speaker:

or two about folding life into geometric perfection, you're in the

Speaker:

right place.

Speaker:

Hello, and welcome back to Data Driven, the podcast where we explore the emergent

Speaker:

fields of data science, AI, data engineering.

Speaker:

Now today, due to a scheduling conflict, my most favorite is data engineer

Speaker:

in the world will not be able to make it. But I will

Speaker:

continue on, despite the recent snowstorms that we've had here in

Speaker:

the DC Baltimore area. With me today, I have

Speaker:

Arjun Patel, a developer advocate at Pinecone,

Speaker:

who aims to make vector databases retrieval augmented generation,

Speaker:

also known as RAG, and semantic search accessible by

Speaker:

creating engaging YouTube videos, code notebooks, and blog

Speaker:

posts that transform complex AI concepts

Speaker:

into easily understandable content. After graduating with

Speaker:

a BA in statistics from the University of Chicago, his journey through

Speaker:

tech world stands spans from making speech coaching

Speaker:

accessible with AI at Speeko to tackling AI

Speaker:

generated content detection at Appen. Arjun's

Speaker:

interest spans traditional natural language processing into modern

Speaker:

large language model development and applications.

Speaker:

Behind beyond his technical prowess, Arjun has been designing and folding his

Speaker:

own origami creations for over a decade. Interesting.

Speaker:

Seamlessly blending analytical thinking with artistic expression and his

Speaker:

professional and personal pursuits. Welcome to the show, Arjun.

Speaker:

Hey. Nice to meet you, Frank. Thanks for having me on. Excited to be here.

Speaker:

Awesome. Awesome. There's a lot to unpack from there, but I think it's interesting to

Speaker:

note that you have a BA in statistics. Yes. So you were probably

Speaker:

studying, this sort of stuff before it was cool?

Speaker:

Yeah. Yeah. A lot of the old school ways of analyzing

Speaker:

data, understanding what's going on, so on and so forth.

Speaker:

It was kind of, like, made clear to me pretty early that

Speaker:

understanding how to work with data at small scale and at large scale is gonna

Speaker:

be very important going to the future. So I kinda just took that and ran

Speaker:

with it with my education. Very cool. It was

Speaker:

definitely, you know, one of those things where I don't

Speaker:

think people realized how important statistics would be until,

Speaker:

you know, until the revolution happens, so to speak. So and it's also

Speaker:

interesting to see because there's a lot of people that I think could benefit from,

Speaker:

you know, picking up that old picking up a, an old statistics book and

Speaker:

reading through it and understanding, like, a lot of the fundamentals. Obviously, there's a lot

Speaker:

of new things, but a lot of the fundamentals are largely the

Speaker:

same. You know, just I'll

Speaker:

use this example. You know, McDonald's can add a Mc McRib sandwich,

Speaker:

but it's still a McDonald's. Right? Like, it's This

Speaker:

is what happens when you're shoveling snow. Like, your

Speaker:

brain gets I absolutely agree. And, like,

Speaker:

another proof on that point is that Anthropic just released a

Speaker:

blog recently kind of recapping how to do statistical analysis when you're

Speaker:

comparing different large language models. And when you read the paper in the blog,

Speaker:

it's basically just like 2 sample t tests and kind of going over really,

Speaker:

like, not introductory, but still statistics that's easily accessible for people to

Speaker:

learn and understand. So it's still relevant, and it's still important.

Speaker:

Interesting. One of the things that that that stood out in your in your bio

Speaker:

was, people tend to forget that there

Speaker:

was a natural language processing field prior

Speaker:

to chat gpt launching.

Speaker:

How do you, you know,

Speaker:

we wanna talk about the difference between those 2? Sure.

Speaker:

So the one of the first and probably only

Speaker:

course I took in college related to natural language processing was

Speaker:

called geometric models of meaning. And everything I learned in that

Speaker:

course was like everything before, what we now would

Speaker:

consider, like, modern embedding models. So bag of

Speaker:

word methods, understanding how to represent documents and text purely

Speaker:

based on, like, the frequency of the words that exist in the text,

Speaker:

and then trying to understand, like, okay. Based on that information, how can

Speaker:

we learn about the concepts that exist in text from the words that are being

Speaker:

used? Like, what is the framework we can use to understand what these

Speaker:

words mean based on their, co occurrences with the other words and

Speaker:

texts that you're working with and based on, what those

Speaker:

words mean as well. So, like, what the words' neighbors are and what their meaning

Speaker:

helps and also what those words are doing. And I think a lot of traditional

Speaker:

natural language processing, methodologies kinda stem from that, and

Speaker:

there's a there's a lot of mileage you can get out of just thinking about

Speaker:

approaching problems there before you step into these more complicated methods,

Speaker:

like, these embed modern embedding models that exist. So that's kind of, like, what I

Speaker:

would consider, like, traditional NLP, like, doing named entity recognition,

Speaker:

trying to understand how to, find keywords really

Speaker:

quickly. And then once you get really good at that, there's a whole host of

Speaker:

problems that you encounter afterward that kind of modern techniques try to

Speaker:

solve. Right. That's interesting. So so

Speaker:

what was it, what was your thoughts

Speaker:

when you first, like given that you were an NLP practitioner

Speaker:

prior to the release of transformers and things like that, what was your initial thought?

Speaker:

Because I'm curious because there's not a lot of people there are a

Speaker:

lot of experts today that really kind of started a couple of years ago. No

Speaker:

fault on them. They see where the industry is going. Totally understand it. But what

Speaker:

was your thoughts? What was your thoughts when

Speaker:

you when you first saw the attention all you need? The

Speaker:

attention is all you need paper. So that would have been

Speaker:

probably around the time I graduated college, around

Speaker:

maybe a year or 2 after I took the course that I was just describing.

Speaker:

So I I just started learning about, like, okay. Like, this is

Speaker:

how, like, old school, quote unquote, like, embedding

Speaker:

methodologies work. And the biggest takeaway that I got from those is that they work

Speaker:

pretty well. They work pretty well for, like, a lots of different kinds of

Speaker:

queries. And I think what the attention all you need paper did

Speaker:

was it kinda helped you, understand how

Speaker:

to rigorously create representations of text that

Speaker:

generalize way better than, any sort of, like,

Speaker:

normal, keyword based, bag of word based search methodology.

Speaker:

And I think that at the time, I probably didn't

Speaker:

grasp as much what impact the attention all you need paper would have on the

Speaker:

field until we started getting embedding models that people could use really

Speaker:

easily, like Roberta or Bert. And we're like, okay. Now we can do, like,

Speaker:

multilingual search without any issue. Now we can represent,

Speaker:

like, any sentence without keyword overlap when we

Speaker:

wanna find some document that's interesting, without doing any

Speaker:

additional work. Like, once those papers started hitting the scene, I think now we start

Speaker:

seeing, like, okay, this is what attention is doing for us. This is what the

Speaker:

ability to, like, contextualize our vector embeddings is doing for us.

Speaker:

And now we can see what's kind of getting benefited there. But I think I

Speaker:

think my, understanding of how beneficial that

Speaker:

was kind of lagged until we started seeing these other models kind of hit. And

Speaker:

I'm like, okay. Now I can kinda see why this is important and why, like,

Speaker:

future and future models are gonna get better and better based on this architecture.

Speaker:

Interesting. So so for those that don't know kind of and even I'm rusty on

Speaker:

this. Right? Yeah. One of the things that was interesting about this was the in

Speaker:

on this. Right? Yeah. One of the things that was interesting about this was the

Speaker:

in first, appearance. What was it? You you just described it a

Speaker:

minute ago, but it was something like the the prevalence of a word

Speaker:

in a bit of text versus the lack of prevalence and how that

Speaker:

metric becomes was very important in in

Speaker:

I'll call it classical natural language processing.

Speaker:

Right. So this is the idea that if you have words that co

Speaker:

occur together in some document space, the meaning of those words are gonna be

Speaker:

more similar than words that don't co occur in some other given document

Speaker:

space. This is rooted in something called the

Speaker:

distributional hypothesis, which is basically this idea and the other

Speaker:

idea that, concepts cluster in in this type of

Speaker:

space. So what what does that mean actually? Right? So if you have the word

Speaker:

like hot dog, it's probably gonna be seen in a corpus that's

Speaker:

near other food related words than it would be if you picked some

Speaker:

other word like space or moon. And there's something we can

Speaker:

learn from that relationship to infer the meaning of what that word

Speaker:

is and how we can use that meaning of that word to learn about what

Speaker:

other words are doing. So So this is kind of, like, the theoretical

Speaker:

basis of, like, why we can represent words geometrically,

Speaker:

with with a little bit of hand waving. But that's kind of the core idea.

Speaker:

And attention kind of takes this a little further by allowing the

Speaker:

representation of these tokens or words to be altered based

Speaker:

on the words that occur in a given sentence. So you might have a

Speaker:

word like does, like, does this mean something?

Speaker:

You might say something like that. Or you might say, I saw some

Speaker:

does in the forest. Both spelled exactly the same, but have

Speaker:

completely different meanings based on their context. And if you used a

Speaker:

traditional, maybe, bag of words model where you're just counting the

Speaker:

words that occur in a given document and kind of creating a representation of what

Speaker:

that document looks like based on the words that are composed in there, you're gonna

Speaker:

overlap and conflict with the meaning of those of of the word

Speaker:

does and does because they're spelled exactly the same. They might look

Speaker:

exactly the same with this type of representation. But if you have a way of

Speaker:

informing what that word means with its context, which is what attention

Speaker:

allows us to do, then you can completely change how that's being

Speaker:

represented in your downstream system, which allows you to do interesting things

Speaker:

with with search. So that's kind of, like, the biggest benefit that's coming out of

Speaker:

that type of methodology, and that kinda enables what is now known as

Speaker:

semantic search and retrieval augmented generation and so on and so forth. I was gonna

Speaker:

say, that sounds very it's almost like it was, like, the old pre

Speaker:

that error, the vectorization of this and the distance in

Speaker:

that vector in that geometric space. I guess

Speaker:

we've been doing that for a lot longer than most people realize in in a

Speaker:

sense. Yeah. I mean,

Speaker:

looking through, indexes or document stores with some sort of

Speaker:

vectorization has has has been,

Speaker:

something that people have done, except instead of being dense vectors, which is, like,

Speaker:

you have some fixed size representation that isn't necessarily interpretable

Speaker:

to the human eye for some given query or document, it would

Speaker:

be, like, the size of your vocabulary. So you think of, like, Wikipedia. You

Speaker:

can find, like, every unique word on Wikipedia, and, like, that is gonna be how

Speaker:

big your vector's gonna be. And every time you have a new document come in,

Speaker:

a new article, somebody's kind of, like, wrote up and published to Wikipedia, like, you're

Speaker:

representing that in terms of its vocabulary. But now instead of doing that, we

Speaker:

have, like, this magical fixed sized box that allows us

Speaker:

to represent chunks of text in a way that is

Speaker:

extremely fascinating and abstract. And every time I think about it, it just, like, blows

Speaker:

my mind, but that's kind of, like, the main kind of difference is the way

Speaker:

we're representing that information and how compact compact that is and

Speaker:

generalizable it has become. Yeah. That is, like, it it's almost

Speaker:

like you're, you know correct me if I'm wrong, but, you know,

Speaker:

creating these vectors, these large vector databases, right, with, you

Speaker:

know, 10, 12,000 dimensions, right, of how these words

Speaker:

are measured in relationship to others.

Speaker:

It's almost as a consequence of training a large language

Speaker:

model, you create a knowledge graph. Is that is that true? Is that really the

Speaker:

case where, you know, like, you know, dog is most likely to be

Speaker:

next to, you know, the word pet, you know, or

Speaker:

it has the same distance. Is that I'm not

Speaker:

explaining it right. No. No. No. You're you're on you're on the right track exactly.

Speaker:

And I think this is, like, one of the most fascinating qualities

Speaker:

of even, like, what people would consider, like, older

Speaker:

embedding models is this idea that you can take, like, a training test that

Speaker:

seems completely unrelated to the quality that you want in a downstream model,

Speaker:

and it turns out that that actually achieves that quality. So, what you were referring

Speaker:

to, Frank, is this idea that you might have, like, a sentence. You

Speaker:

might have, like, I took my dog out on a walk, and you might say,

Speaker:

okay. I'm gonna remove the word, walk, and I'm gonna have

Speaker:

I'm gonna train some model that tries to predict what that word

Speaker:

where I removed was. This is masked language modeling, which is this idea that you're

Speaker:

kind of getting at of, like, okay, what are the words and how are they

Speaker:

in relation to the other words in that sentence? And it turns out that if

Speaker:

you, like, do this with, like, 100 of 1,000 of millions of sentences and

Speaker:

words, in some corpus that is somewhat representative of

Speaker:

how people, use human language, you can

Speaker:

act you will get really good at this task, number 1, because you're training the

Speaker:

model on that task exactly. But if you are training a neural

Speaker:

network on that model, some intermediate layer representation

Speaker:

in that model so somewhere in that set of matrix

Speaker:

multiplications where you're turning this input sentence into some fixed size

Speaker:

vector representation is gonna be a good representation

Speaker:

of what that word or that token or that sentence is going to be.

Speaker:

And the fact that that works is not intuitive. Right?

Speaker:

The the fact that that works has been shown empirically, and it turns out that

Speaker:

we can kind of do that and kind of have these models work really well.

Speaker:

And nowadays, in addition to kind of doing that, which is what we would consider

Speaker:

pretraining on some large corpus, we now fine tune those

Speaker:

embedding models on specific tasks that are important to us

Speaker:

for retrieval. Like, okay, we have this query or question we're

Speaker:

asking. We have the set of documents that might answer this question or might

Speaker:

not. We want a model that makes it so that the query's embedding and the

Speaker:

document relevance embeddings are in the same vector space. So you're on the right track.

Speaker:

That's, like, basically how these models are able to learn these things. I don't know

Speaker:

if I would call them, graph representation, maybe a little bit

Speaker:

of, being being pandactic on, like, use of words there because that can

Speaker:

be a little bit, different how how you're organizing that information.

Speaker:

But you can make the argument that the way that these large language models are

Speaker:

representing information is a compressed form of, like, the giant dataset that they're

Speaker:

trained on. And we don't actually know exactly, like, where that

Speaker:

information lies inside that neural network. There's some research that's,

Speaker:

like, trying to get at answering that question, But you could, for the sake of

Speaker:

argument, be like, yeah. There's probably, like, a a a dog

Speaker:

node somewhere in this neural network that knows a ton about dogs, and that's how

Speaker:

we're able to kind of learn this information. That is the stuff that we don't

Speaker:

exactly know. Interesting. Because, there was a really good

Speaker:

video by 3 blue one brown, which you probably are I love that

Speaker:

channel. Where he gives examples where, you know, famous historical

Speaker:

leaders from Britain have the same distance

Speaker:

from you change the country to Italy

Speaker:

or the United States have the same kind of distance. So you can kind

Speaker:

of infer I'm not saying that the AI it

Speaker:

almost seems like this knowledge graph is also is also a byproduct

Speaker:

of of of building this out. Like, the there's some

Speaker:

type of encoding or semantic, I guess, is this is really what it is. Right?

Speaker:

Like, that that you get with it. And, I wanna get

Speaker:

your thoughts because yesterday, I I caught the part the

Speaker:

first half of the Jetson Juan keynote at c s CES,

Speaker:

which this you know, we're recording this on January 8th. Right? And one of the

Speaker:

things that the video starts off with is, you know, the idea

Speaker:

that tokens are kind of fundamental elements of

Speaker:

knowledge. And I did a live stream where I'm like, well, I never really thought

Speaker:

about it this way. Right? They're they're building blocks of knowledge or the pixels, if

Speaker:

you will, of of of of knowledge. And I wanted to get your

Speaker:

thoughts on that because, like, that kind of blew my mind and maybe I'm simple.

Speaker:

I don't know. Maybe I'm not. But it all it seems like we've been kinda

Speaker:

dancing around this idea where and now NVIDIA is really

Speaker:

fully, you know, going all in on this, the idea that, you know,

Speaker:

these are not, this isn't an AI system. It's a token factory

Speaker:

or a token score. What are your what are your thoughts on that? I'm curious.

Speaker:

So when I started learning about how, like, tokenization works

Speaker:

and how we're able to kind of, like, basically build these

Speaker:

models without having massive, massive vocabularies,

Speaker:

it is it is pretty it it is pretty

Speaker:

interesting to be, like, okay. Like, maybe maybe there's some,

Speaker:

abstract notion of information that each token has that

Speaker:

is being that is what the model is learning during training time. And then

Speaker:

we're just combining these sets of information in order to kind of, like, understand

Speaker:

what words mean or what documents mean, so on and so forth. Because when you

Speaker:

look at how, tokenizers work and the size of the number of

Speaker:

tokens for, like, maybe the English language or maybe, like, a really multilingual

Speaker:

model like Roberta or multilingual e five large, they're a lot

Speaker:

less than you would expect. Like, it's on the order of, like, maybe a 100000,

Speaker:

200000, 300000, tokens.

Speaker:

So it is kind of

Speaker:

odd to think about whether those tokens

Speaker:

themselves hold information that's readily interpretable for us. But I

Speaker:

think that we've gotten so far with using

Speaker:

systems that are just combining, the operations on top of

Speaker:

these tokens in order to retrieve the information that these systems have learned, that there's

Speaker:

definitely something important there. And I would love to, like, know

Speaker:

exactly, like, what is happening when we're able to do that. The the

Speaker:

heuristic that I like to use is, large

Speaker:

language models are generally reflections of the training datasets that they've been trained on,

Speaker:

and they're basically creating, like, really efficient indexes over that

Speaker:

information. And sometimes those indices hallucinate. And the reason

Speaker:

why is because we are when we ask, quote, unquote, what

Speaker:

a question to a large language model or query a large language model, we

Speaker:

are kind of conditioning that model, on a probability

Speaker:

space where every token being generated after is

Speaker:

likely to exist given the query or the context or whatever we're passing to

Speaker:

it. And once you think about it that way, then it just feels like

Speaker:

instead of thinking about what each of the tokens are doing, you're kind of just

Speaker:

querying what the model has been trained on and what it will tell you

Speaker:

based on what it, quote unquote, learned or knows.

Speaker:

And then you can kind of run with that metaphor a lot and build systems

Speaker:

on on top of that. That seems, much more actionable than thinking about,

Speaker:

like, what each of the tokens are doing individually. Does that kinda make sense? No.

Speaker:

That makes a lot of sense. I think the whole gestalt of it is what

Speaker:

really makes it magical. Right? Like Yeah. You know, you can you

Speaker:

can obviously, I I don't this is not this is not, like, the newest iPhone

Speaker:

or whatever. But, you know, if you go through the the text auto complete,

Speaker:

you can maybe make a sentence that sounds like

Speaker:

something you would write. But much beyond that, it starts getting weird. In

Speaker:

early generative AI was very much like that, particularly the images.

Speaker:

Well, you know Don't like, yes. A 100%

Speaker:

understand. I started learning about generative, text

Speaker:

generation before we had instruction fine tune model. So are you

Speaker:

familiar with, like, the concept of instruction fine tuning, Frank? I think I am,

Speaker:

but I IBM slash Red Hat defines it one way. I would like to get

Speaker:

your opinion. Yeah. So, this is the idea that

Speaker:

you can train or fine tune large language models to follow

Speaker:

instructions to complete tasks. So, before we had,

Speaker:

like, models that could that we could just, like, ask questions of and just, like,

Speaker:

receive answers directly, you had to craft text

Speaker:

that would increase the probability that the document that you want to

Speaker:

generate would happen. So if you wanted a story about, like, unicorns or something,

Speaker:

you would have to start your query to the LLM as there

Speaker:

once was, like, a set of unicorns living in the forest. Blah blah blah blah.

Speaker:

And then it would just, like, complete sentence, just like a fancy version of autocomplete.

Speaker:

Right. And that that's kind of, like, what we used to have, and that was

Speaker:

pretty hard to work with. And then once researchers kinda cracked, like, wait a second.

Speaker:

We can create a dataset of, like, instruction pairs and, like, document

Speaker:

sets and fine tune models on them. And it turns out now we can just,

Speaker:

like, ask models to do things, and they will do them. Whether or not

Speaker:

those are correct is kind of the next part of the story. But getting to

Speaker:

that point, it was, like, pretty interesting and pretty significant.

Speaker:

Interesting. Interesting. When I think of

Speaker:

fine tuning, I think of I think of

Speaker:

primarily InstruqtLab, where you basically kinda have a

Speaker:

LoRa layer on top of the base LLM doing

Speaker:

that. Is that the same thing? Or is it kind of slightly

Speaker:

it sounds like it's slightly nuanced. So the nuance there

Speaker:

is that, one, though this the methodology that I'm

Speaker:

describing is mostly dataset driven. So you have, like, your original LLM,

Speaker:

and then you have, like, a new dataset that allows the LLM to learn a

Speaker:

specific task. Or in this case, like, a generalized form of tasks,

Speaker:

which is you have instruction, answer, user query,

Speaker:

give it an instruction. Whereas in your case, you're kind of, like, adding another layer

Speaker:

to the LLM and, like, forcing the LLM to learn all the new

Speaker:

methodology inside that layer in order to accomplish a specific

Speaker:

task. So that's kind of like what client cleaning ends up doing. So the other

Speaker:

way there's multiple ways to do this, it seems. Right? Like, there there's that way

Speaker:

we add the layer, but there's also kind of I hate the term prompt engineering

Speaker:

because it's just so over overblown. But, like, giving it

Speaker:

more context and samples. And now that the the token context

Speaker:

window is large enough that you don't have to be well, if you wanna

Speaker:

save money, you have to be very mindful of that. But if you're running it

Speaker:

locally, like, doesn't really matter. Well, you could give it an example of

Speaker:

let's just say you had I'm trying to think of a short story or a

Speaker:

novel. I don't know. Let's pretend,

Speaker:

Moby Dick was only a 100 pages. Right? I

Speaker:

could give it that as the part of the prompt. Let's say write a sequel

Speaker:

to this book based on what happens in this one. Is that what you're talking

Speaker:

about? Were you kinda giving an example as part of the prompt? Or is there

Speaker:

some and not part of the layer? Or some combination thereof? Or was some third

Speaker:

thing entirely? So this would be like, what what

Speaker:

you're describing is more like few shot learning, which is you gave kind of an

Speaker:

example, and then you're, like, okay. Like, given these examples, can you do this other

Speaker:

task this test that I've described on this unseen example? What I'm describing is

Speaker:

kind of, like, slightly before that. So, like, before we had the ability to, like,

Speaker:

give models examples, we had to, like, give them we have to

Speaker:

create the ability to follow instructions. And then once you have the ability to

Speaker:

follow instructions, you can be like, okay. Here are the instructions. Here's

Speaker:

examples of correctly completing the instruction, now do the instruction.

Speaker:

And that is the reason why that happens in that order is

Speaker:

because first, you have, like, just, like, sequence completion, like,

Speaker:

autocomplete. Then you have, like, okay, given this

Speaker:

task given this set of instructions, just follow the instruction instead of,

Speaker:

like, trying to do autocomplete. And then you have, okay, now you know how to

Speaker:

follow instructions. I'm gonna give you a few data points in order to

Speaker:

learn a new task. Now do this new task. So you're kind of,

Speaker:

like, moving from a situation where you need tons and tons

Speaker:

of data just to get the, sequence completion. And then you need

Speaker:

a smaller set of data to, like, get the capability to follow instructions.

Speaker:

And then you need a very, very, very small amount of data, like,

Speaker:

maybe 3 points or 10 examples or 15 examples to complete kind of, like,

Speaker:

a new task. So there's a lot of kind of nuance in, like, how

Speaker:

modern LLMs are being used and how they're kind of trained and fine tuned, so

Speaker:

on and so forth. And I think there's a lot of, like,

Speaker:

important importance in, like, learning what what happened kind of

Speaker:

before because the advancements have happened so quickly. It can be really hard to kind

Speaker:

of differentiate, or, like, oh, why is why do models perform like this? Why

Speaker:

do things kind of happen like that? And even though, prompt

Speaker:

engineering has kind of, like, let's say, traveled through the

Speaker:

hype cycle where people were, like, really excited about it, and then we're, like, this

Speaker:

is not actually that interesting. Right. What's interesting is that,

Speaker:

doing building a good RAG system or trivial augmented generation system,

Speaker:

you really need to be good at prompt engineering in a sense

Speaker:

because you're assembling the correct context for this model

Speaker:

to answer some downstream question, And it's not

Speaker:

intuitive how to assemble that context. So understanding, like, how are these

Speaker:

models are trained, like, whether they can follow instructions, how good they are at

Speaker:

doing so, how many examples of information they need in order to accomplish some task

Speaker:

really affects how you build that knowledge base in order to help the

Speaker:

model do some sort of new thing. Interesting.

Speaker:

So RAG is obviously all the rage now.

Speaker:

Yep. But there's also a relatively new because this this

Speaker:

space changes rapidly. Like, I mean, I took 2 weeks off in December, and

Speaker:

I feel completely disconnected from the cutting edge, you know.

Speaker:

Because when I was watching the keynote from CES, and I'm like, wow. That's

Speaker:

really cool. And I was texting, you know, slacking with a coworker, and he goes,

Speaker:

oh, no. This is a retread of their, like, last keynote they did. Like

Speaker:

and I'm like, okay. Wow. Blink and you missed

Speaker:

something. So what

Speaker:

you're describing the fine tuning, is that really what Raft is, where the

Speaker:

idea that you have kind of retrieval augmented fine tuning, which I think is what

Speaker:

the acronym stands for. Is that not I'm

Speaker:

not familiar with how Raft works. So I don't wanna, like, kind of venture

Speaker:

and guess without without knowing what it is. But do you remember, like, what context

Speaker:

you encountered this in? Basically, it's the idea that

Speaker:

it's the idea that you can fine tune the results. Sounds very

Speaker:

similar to what you're doing, and I've haven't read the paper in a while.

Speaker:

Back when I was a Microsoft MVP, like, you know,

Speaker:

they had a Microsoft Research had the thing for their calls, and they

Speaker:

were all raving about it. The paper had just come out and things like that.

Speaker:

It's the idea that you can kind of give it pretrained examples.

Speaker:

You start with a base LLM, and you give it pre trained examples, and then

Speaker:

you add on top of just to retrieve an

Speaker:

augmented portion of it. It's very similar, not to

Speaker:

plug my you know, for my day job. I work at Red Hat. That's why

Speaker:

there's a fedora there. We have a product called Rel

Speaker:

AI, which is based on an upstream open source project called instruct

Speaker:

lab. And it's the idea similar idea in that you you you

Speaker:

basically give it a set of data.

Speaker:

And then you we there's a there's a little more to it because there's a

Speaker:

teacher model. And basically what it'll do is it will and synthetic data generation.

Speaker:

So you can start with a modest document set.

Speaker:

And based on how the questions and answers that you

Speaker:

form and the the the,

Speaker:

the taxonomy that you attach to it, it will

Speaker:

create a LoRa layer on top of an existing LLM.

Speaker:

And it it could be that it's it's it's not quite exactly the same as

Speaker:

Raft, but it's definitely in the same direction. Same same thing as, like, Bert, Elmo,

Speaker:

and, you know, Roberta, which, I think

Speaker:

I think I understand. So it's kind of like you so the I think the

Speaker:

problem that might be addressing is kind of just really similar to the problem that

Speaker:

traditional RAG tries to address, except in a more kind of deliberate fashion

Speaker:

Exactly. Yeah. Where you have some document store internally. Like, let's say we

Speaker:

both work at some company, and we have a giant customer support document store.

Speaker:

You take some LLM off the shelf. It's not necessarily gonna know the

Speaker:

contents of your internal kind of documents. So how can you get

Speaker:

it to, like, successfully help answer tickets or triage tickets that

Speaker:

you're trying to build, so that you can answer, like, most difficult tickets and

Speaker:

kind of work toward that. In this situation, maybe you

Speaker:

want to, inject some of the knowledge of

Speaker:

the documents in addition to having the

Speaker:

model being able to search over the document store. So maybe, like, the what this

Speaker:

lower layer is doing is, like, absorbing Yeah. Some of the knowledge from the

Speaker:

document store so that you can kind of more

Speaker:

efficiently query, the database and so

Speaker:

that you don't have to, like, query it all the time. The only,

Speaker:

issue, quote, unquote, I'd have with that method is that you'd have to, like, keep

Speaker:

that updated from time to time, and that's, like, not that's nontrivial. Whereas

Speaker:

if you just do, like, traditional RAG, you just need to

Speaker:

update your, Vector Store, and then you can just have the model

Speaker:

query that new information when you need to. But, you know, it's always best to

Speaker:

use whatever solution works best for your, given use case.

Speaker:

And experimenting with different use cases is always really important. But I imagine that's, like,

Speaker:

kind of what that is trying to address, which is the That is basically it.

Speaker:

The I, you know, I don't wanna go down that rabbit hole of that. But

Speaker:

but, basically, the idea is that, if

Speaker:

you train an LLM or you have a layer on top of an

Speaker:

LLM that not only does retrieval from a source document

Speaker:

store. Right? I think that's a pretty set pattern. But it also has a

Speaker:

better understanding of your business, your industry, the jargon.

Speaker:

Right. Right. Blah blah blah. Right? The idea is that the retrieval success

Speaker:

rate will be higher. Now we're not publishing the numbers yet,

Speaker:

but the research is still ongoing. But basically, it's a

Speaker:

pretty substantial from what I've seen well, I haven't

Speaker:

seen the actual numbers yet, but from what I've been told those numbers are by

Speaker:

the researcher, that it is a it is a substantial improvement

Speaker:

that is worth the, the juice is worth the squeeze in that in that regard.

Speaker:

You're not and it's also computationally, you're not quite training the

Speaker:

whole thing again. You're just kinda putting a new Instagram filter, so to

Speaker:

speak, together on top of the base. So it definitely

Speaker:

does it definitely does some things. Now when we get the hard

Speaker:

numbers, then, you know, I mean, I can

Speaker:

say them publicly, then I think we'll we'll know is the juice how

Speaker:

much does the the the the squeeze to juice ratio is?

Speaker:

But, I can confidently say publicly now, like, there's a there

Speaker:

there. Yeah. And, you know, we'll have those numbers soon

Speaker:

enough. But it's it's interesting because you're right. I mean, this paper

Speaker:

came out in 2019. Right? There was just an

Speaker:

explosion of these different mechanisms. You mentioned Bert. You mentioned Roberta.

Speaker:

Fun fact, my wife's name is Roberta. So that was kind of fun.

Speaker:

There was Elmo. There was Ernie. There was a whole Sesame

Speaker:

Street themed zoo of of model

Speaker:

types. That seems to have kind of that branching out of

Speaker:

those different directions has seemed to have stalled, and we're going into more of

Speaker:

these retrieval augmented generation systems. So for those who because

Speaker:

not everybody on our listeners know exactly what retrieval

Speaker:

augmented systems are. Could you give kind of a a

Speaker:

level 200 elevator explanation? Sure.

Speaker:

So, when you speak to a modern chatbot,

Speaker:

what's happening is that they've learned information through their pre

Speaker:

training processes, the large corpus of basically the entire Internet,

Speaker:

and are generating information based on the query that you're passing in.

Speaker:

The problem that often occurs is that

Speaker:

these AI models might error, and the error could

Speaker:

be making, inform making information up that doesn't

Speaker:

exist. For example, if a model is trained before a period of time,

Speaker:

like, it might not know about that period of time, which is which happens more

Speaker:

often than you think. The information could be false, untruthful, or it could

Speaker:

just be incorrect in a way that's not, like, bad, but still not

Speaker:

helpful. And the reason for this is the way that these

Speaker:

models are accessing that information. The idea behind retrieval

Speaker:

augmented generation is that instead of having the model try

Speaker:

to, generate the correct document or the correct

Speaker:

response given its pretraining process, you instead

Speaker:

add factual content to the query that you're asking

Speaker:

the model for. You first search for that content, which is where

Speaker:

the retrieval part comes, and then you augment the generation of what that

Speaker:

model is going to create based on that content, hence

Speaker:

retrieval augmented generation. There's usually, a querying

Speaker:

step. So you take in a user query, you hit it against some sort

Speaker:

of database, usually a vector database. In our case, it could be Pinecone.

Speaker:

You find a set of relevant documents. You pass that to the generating LLM.

Speaker:

The generating LLM uses those documents to generate a final

Speaker:

response. And it turns out that if you do this, you can reduce the right

Speaker:

hallucinations. And that makes sense because if the model was given true

Speaker:

information and then conditioned its generation on that information, it

Speaker:

follows that the probability of generating information that is

Speaker:

correct could be higher. That's a good exam that's a good

Speaker:

explanation. So you're basically giving it a

Speaker:

crash course in what documents you care about. Right? Like

Speaker:

Exactly. Interesting. And that's a good segue

Speaker:

because you work for Pinecone. So so tell me about Pinecone. What is Pinecone?

Speaker:

Yeah. So Pinecone is a, knowledge layer for AI. It's

Speaker:

kind of like the way we like to describe it. We the main product that

Speaker:

we provide is a vector database. So this is a way of storing

Speaker:

information, information that has been vectorized, in a really

Speaker:

efficient manner. And it turns out that if you have the ability to store information

Speaker:

in this manner, you can search against it really quickly, with

Speaker:

low latency and to find the things that you need to find really interesting for

Speaker:

these types of semantic search and rag systems. Pinecone has a few other

Speaker:

offerings now that kind of help people build these systems a lot easier. There's

Speaker:

Pinecone Inference, which lets you embed data in order to do that querying

Speaker:

step. Pinecone Assistant, which lets you just build a RAG

Speaker:

system immediately just by upsurting documents into our vector database,

Speaker:

so on and so forth. But the reason why, like, you

Speaker:

need a vector database is because all of this advance of

Speaker:

semantic search of embedding models. People have gotten really, really

Speaker:

good at representing chunks of information using these dense sized

Speaker:

vectors. But once you have 1,000, millions,

Speaker:

even billions of vectors across tons of different users, you need a way

Speaker:

of indexing this information to access it really quickly at

Speaker:

scale, especially if your chatbot's gonna be querying this vector database really

Speaker:

often. And so having a specialized data store that can handle that type

Speaker:

of search becomes really useful. That's why Pinecone is here, and that's

Speaker:

why we exist. Interesting. Interesting.

Speaker:

One of the other interesting things from your bio, aside from

Speaker:

the the the origami,

Speaker:

Tell me about this. So so you

Speaker:

your crew does your do you create the YouTube videos, or do you use your

Speaker:

tools, or is it something completely it's just part of your job as a developer

Speaker:

advocate? So it is just part of my job as a

Speaker:

developer advocate. Oh, okay. Like, often that, you

Speaker:

know, I do that because we are interviewing people or because there's a new

Speaker:

concept we wanna teach people, so on and so forth. Or we do a webinar,

Speaker:

and we just upload it to YouTube. Oh, very cool. Very cool.

Speaker:

Yeah. I started my career in developer

Speaker:

advocacy. One was called evangelism. So I was a a Microsoft

Speaker:

evangelist for a while. So yeah. Yeah. Cool. YouTube

Speaker:

is very important. Yep. But it's

Speaker:

also it's also, I think, speaks to how people learn,

Speaker:

but, how people learn. YouTube University is very

Speaker:

real. Right? And Yep. You know, not not a knock on

Speaker:

traditional schools, not a knock on traditional publishing, but this space

Speaker:

is moving so fast that if it weren't for YouTubers like 3blueonebrown

Speaker:

I think his real name is, Grant Sanderson. I think that's his real name.

Speaker:

Somebody will send me hate mail if I get it wrong. But,

Speaker:

he he is, like, really good at explaining these

Speaker:

really abstract mathematical concepts. And

Speaker:

unlike you, I didn't study math undergrad. I didn't I mean, I had to. I

Speaker:

only took the requirements. Right? But I have comp sci degrees. So, like, for me

Speaker:

to kind of fall in love with math again or for the first time, depending

Speaker:

on depending on how you wanna say that, for me, that

Speaker:

was very helpful. And under having an understanding of this, if you're a data engineer

Speaker:

and, you know, or wanna get into this space, it's

Speaker:

definitely vector databases for traditional kinda SQL kinda

Speaker:

RDBMS person will look very awkward at first. But

Speaker:

I know a lot of people that have made the transition, and they kinda love

Speaker:

it. Right? Because in a lot of ways, it's way more efficient,

Speaker:

than, I dare say, traditional data stores. But when you're

Speaker:

processing the large blocks of text, it's really good for kind of

Speaker:

parsing through that. But

Speaker:

that's that's really cool. So, we do have the preset

Speaker:

questions if you're good for doing those. I'll put them in the chat in case

Speaker:

you don't have them. Sure. They're not brain teasers

Speaker:

or anything like that. They are pretty basic of,

Speaker:

questions, and I will paste them in the chat.

Speaker:

So the first question is, how did you find your way into

Speaker:

AI? Did you did you find AI, or did

Speaker:

AI find you? So this is a little bit of a

Speaker:

crazy story, but AI definitely found me.

Speaker:

So when I was in college, when I was looking for my 1st

Speaker:

internship, I couldn't find any internships, basically, because I had, like, no

Speaker:

previous experience in working at tech or anything like that. And,

Speaker:

the first company I worked for, Speeko, took a chance on me because they were

Speaker:

building public speaking, tools to kind of help people learn how to do

Speaker:

public speaking better, for an iOS app. And I had some

Speaker:

public speaking experience. They were, like, close enough. We'll have you come on and kind

Speaker:

of help us, like, work work things out. And while I was there, it was

Speaker:

made very obvious to me how important building

Speaker:

very basic deep learning systems and AI systems to kind

Speaker:

of accomplish really specific tasks that could help serve an

Speaker:

ultimate goal. Like, what we were trying to do is just, like, see how many

Speaker:

filler words people are using or how quickly or slowly you were speaking.

Speaker:

And that requires a lot of, complicated

Speaker:

processing because you have to do transcription and because you have to figure out what

Speaker:

words are being said, so on and so forth. So kind of experiencing that and

Speaker:

seeing that firsthand really opened my eyes to how powerful

Speaker:

the technology had been even back in, like, 2017. And ever

Speaker:

since then, I started learning more and more and more about statistics,

Speaker:

AI, natural language processing through my internships,

Speaker:

learning more complicated problems, reading research papers, so on and so forth.

Speaker:

And I got to where I am now. A lot of where I learned is

Speaker:

just out of pure curiosity. Just like, okay. There's this new thing. I wanna learn

Speaker:

about it. That's where I wanna be. And that's kind of how I fell into

Speaker:

large language models and AI, just by wanting to learn about what was going to

Speaker:

happen and then eventually being there. So it definitely found me. I was

Speaker:

not looking for it. Didn't even know I liked statistics until I started doing

Speaker:

statistical modeling. And I was like, wait. This is really fun. I wanna do a

Speaker:

lot more of this. I wanna learn a lot more of this. And I knew

Speaker:

that, once I was in college and I bought a statistics book for fun, and

Speaker:

I was like, okay. I'm I'm past the point of no return. Like, this is

Speaker:

definitely Right. Right. Right. Right. That that might be one of the first times in

Speaker:

history that that's been said. Right. Because I I learned statistics for

Speaker:

fun. I I took stats in college.

Speaker:

I hated it. Hated every minute of it. But

Speaker:

when I got into data science,

Speaker:

I the first two weeks were not fun. I'm not gonna lie. Yep. But

Speaker:

just like the VI editor, once you stick with it,

Speaker:

Stockholm syndrome kicks in, And you start loving

Speaker:

it. That's cool. 2, what's your favorite

Speaker:

part of your current gig? The favorite part of my

Speaker:

current job is being able to learn interesting,

Speaker:

fun, even complicated things in data science and AI,

Speaker:

and figuring out how to communicate them to a wide

Speaker:

audience. It's a really fun challenge. It's really similar to, like,

Speaker:

what, 3 blue one brown does all the time on the YouTube channel, and it's

Speaker:

something that I get to learn and practice and keep keep doing. That's the best

Speaker:

part of the job. I love learning things and, like, teaching other people about them

Speaker:

and learning even more things. And the fact that I have an opportunity to do

Speaker:

that every single day is, like, the best. That's cool. That's

Speaker:

cool. We have 3 complete sentences. When I'm

Speaker:

not working, I enjoy blank. When I'm

Speaker:

not working, I enjoy, baking sweet treats and

Speaker:

goods. I can't have any dairy. So very often, I had to kind

Speaker:

of give up a lot of the cakes and desserts that I loved eating when

Speaker:

I was younger. So now I, like, spend my time trying to figure out how

Speaker:

I can make them again without dairy so they taste really good. So that's that's

Speaker:

something I enjoy I really enjoy doing. Very cool.

Speaker:

Next, complete the sentence. I think the coolest thing in technology

Speaker:

today is blank. I

Speaker:

thought really hard about this question because we're living in a

Speaker:

crazy time of technological development. But the thing that really

Speaker:

stuck out to me and the thing that was also the moment for me

Speaker:

when I started working with, like, chatbots and LLMs was code

Speaker:

generation models. The first time I learned how to

Speaker:

use, GitHub Copilot specifically, I

Speaker:

was I was completing some function, and it completed it before I was done typing

Speaker:

it. And I was like, what the heck? This is amazing. Like, this this this

Speaker:

actually figured out exactly what I needed. And because I was still, like,

Speaker:

a budding developer, it was extremely helpful because I could learn

Speaker:

faster rather than having already a huge kind of store knowledge already in my

Speaker:

brain and kind of pulling from that. So I could see it benefiting my workflow.

Speaker:

So I think the development of those tools and modern tools like

Speaker:

Cursor, so on and so forth, extremely cool. And I can't wait to

Speaker:

see, like, what the next generation of those technologies will look like. Yeah. I

Speaker:

mean, that's a that's a great example. It's almost like you don't

Speaker:

need, you know, the the classic 10000 hours to master a skill or something like

Speaker:

that. It's almost like you can leverage the AI to take on the

Speaker:

lion's share of the 10000 hours. You're still gonna need to know something. You still

Speaker:

have to put in some reps, but not to the degree that you used to.

Speaker:

No. I think that's gonna be very transformative. I mean, I mean, I'm

Speaker:

learning, JavaScript and Next. Js on the side because it's something I have no

Speaker:

experience in. Right. And I was able to build my personal website

Speaker:

entirely through using Cursor and Progression. Nice. I

Speaker:

often check that out. Which is insane. Right? Which is, like, really, really

Speaker:

fascinating. And and I'm not gonna claim to, like, suddenly be an expert in

Speaker:

NextGen or anything like that. Right? Right. Right. Right. I still wanna learn, like, exactly

Speaker:

what's going on under the hood, But having a project that you can kind of,

Speaker:

like, tinker on that's, like, pretty small in scale and that you can kind of

Speaker:

afford to make a few mistakes on and having, like, an expert system kind of

Speaker:

help you go through that, expert, quote, unquote, being close enough, really cool

Speaker:

learning experience. No. That's a great way to put it because, like, I I

Speaker:

I don't have any apps on the modern devices. Right? Like,

Speaker:

so, it would be nice if I

Speaker:

had an Android app that could kick off some automation process that I have.

Speaker:

Right? Or do some kind of tie in with, you know, Copilot

Speaker:

into that or things like that. Like, where, you know, I

Speaker:

originally wrote a content automation system I wrote. I originally wrote in

Speaker:

dotnet, but I ported it to Python with the help of

Speaker:

the help of AI. And I could well, that's just it. Right?

Speaker:

It really the true valuable resource in in life is

Speaker:

time. Right? Yes. It's not Yes. I mean, I could have done it by hand.

Speaker:

I could have done it by myself, but it was one of those things where

Speaker:

am I gonna do it because it's gonna take x number of hours or whatever?

Speaker:

But if I can just kinda here's the dot net version that I, you know,

Speaker:

I posted. This is before there was Copilot, so I pasted it into chat g

Speaker:

p t. And it basically spit out a Python

Speaker:

version, had some errors. You know, this was a while ago. But I

Speaker:

was able to, inside of a day, get it done as opposed to

Speaker:

before. Like, I know how my ADD works. Right? Like, I'll start it.

Speaker:

First 3 days, working on it, grinding on it, and then

Speaker:

I don't touch it again for 2 weeks. And it never gets built. But

Speaker:

with this, I'm able to kinda harness the the spark of

Speaker:

inspiration and and execute much faster. Now I think I don't think

Speaker:

people fully realize, like, you know, it's not all doom and gloom. Nobody's

Speaker:

gonna have any programming jobs. There's a lot of upside too. And I

Speaker:

guess that's just where we are in the hype cycle. As you said.

Speaker:

Yeah. Yeah. Yeah. Exactly. That's a good segue into I look forward to

Speaker:

the day when I can use technology to blank. I look

Speaker:

forward to the day where I can use technology to get a high quality

Speaker:

education on any subject for free. So Nice.

Speaker:

Free education is really important to me. A lot of

Speaker:

what I learned about large language models, deep learning, all that

Speaker:

stuff was online courses that I took for free on places like

Speaker:

EDX, Coursera, so on and so forth. Or people sharing

Speaker:

articles and kind of learning from them, or YouTube videos, or all that sort of

Speaker:

things, in addition to my education. But there's a lot of things you kinda have

Speaker:

to learn after that. Right? And I think that especially with, like,

Speaker:

cogeneration models, it's, like, very easy to be, like, okay. Build me this app

Speaker:

and, like, just make it work. And you can sit there for a couple hours,

Speaker:

and it'll, like, work. But I think the missing piece is

Speaker:

creating a structured kind of learning path that's, like,

Speaker:

personalized to whoever you are for the

Speaker:

thing that you're really interested in with the context of

Speaker:

having, like, these tools that can help you do that thing. And I'm not sure

Speaker:

if we have anybody or any offering that can

Speaker:

kind of do that technologically, because you need a lot of information about what the

Speaker:

user knows or doesn't know. You need to be able to create ability, and then

Speaker:

you need to be able to kind of create, like, an entire mini course that's

Speaker:

personalized to whatever that person needs. But if we can do that, we can solve

Speaker:

so many wonderful problems. Absolutely. I'm

Speaker:

thinking about special education needs and things like that. I don't think we're that

Speaker:

far off from this. No. But I

Speaker:

the biggest issue, is going to be just hallucinations. Right? And,

Speaker:

hopefully, people can build, like, rag systems using tools like PineCone to kind

Speaker:

of produce those hallucinations. But we will also for for something like

Speaker:

that specific use case, we probably need, like, another breakthrough in

Speaker:

indexing information or kind of presenting it, or we need a process that

Speaker:

really allows people to create this information quickly

Speaker:

and verifiably in order to kind of make that happen. But if if that is

Speaker:

a future that we can live in, where technology can can kind of, like, help

Speaker:

people learn, like, really important things really well, that would be

Speaker:

wonderful. And I think that would be, like, amazing for for humanity.

Speaker:

Oh, absolutely. Share something different

Speaker:

about yourself, but remember as a family podcast.

Speaker:

One of my favorite hobbies for about a decade is

Speaker:

designing and folding origami. And it's really fun.

Speaker:

It's very easy, but it's also very hard. There's a lot

Speaker:

of comp complexity inside it as well. One thing people

Speaker:

don't know about that is that there's a lot of mathematical complexity.

Speaker:

So once you get to a point where you wanna design a model with

Speaker:

really specific qualities, really specific features, it suddenly

Speaker:

becomes a paper optimization problem where you

Speaker:

have, like, a fixed size square, and you have different

Speaker:

regions of that paper that you're allocating to portions of the model you're

Speaker:

designing. And it turns out that there are entire mathematical

Speaker:

principles and procedures to solve this problem. So much

Speaker:

so that one of the leading, like, practitioners in the

Speaker:

field is, like, this physicist who wrote a textbook on how to do origami design,

Speaker:

and that's, like, the textbook everyone looks at. So, like, learn how to solve it.

Speaker:

Yeah. I'm not surprised. There's definitely there's definitely a a correlation

Speaker:

between the mathematics of that. And I look at origami creations, and I

Speaker:

just fascinated that could be done from a single sheet. Like, it's

Speaker:

just how is that I mean, that's just mind bending. Now it's

Speaker:

and and makes sense that there's a mathematical because you have a certain type of

Speaker:

constraint, And there's obviously

Speaker:

folds factor into it and things like that. And, yeah, that's that's

Speaker:

interesting. I I should what's the name of that book? I should pick it up.

Speaker:

It's called Origami Design Secrets. Got it. Alright. I will check

Speaker:

it out. So where can people learn more about

Speaker:

you and Pinecone? Of course. You wanna learn more about Pinecone? The

Speaker:

best place is our website, pinecone. Io. You can also find

Speaker:

us on LinkedIn and on x and other social media platforms.

Speaker:

You wanna learn more about me? You can go to my LinkedIn, which you can

Speaker:

find at Arjun Girthi Patel, or you can go to my website, which is also

Speaker:

my name, arjun, k I r t I p

Speaker:

a t e l.com. Cool. And we can also check out your

Speaker:

Next JS skills there too. Exactly. Hopefully, nothing is

Speaker:

broken, but, you can you can see you can see how well I've gotten by

Speaker:

with the Awesome. Trust me.

Speaker:

JavaScript alone is is a is a frustration

Speaker:

creation device.

Speaker:

Audible sponsors the podcast. Do you do audio books? Is there a book that you

Speaker:

would recommend? I do do audiobooks, but I've just

Speaker:

started recently, so I don't have a huge, audiobook library. But

Speaker:

there is I I am a huge fan of short story collections, and

Speaker:

kind of the one that comes to mind is really anything by Ted

Speaker:

Chiang, who does a lot of kind of sci fi short stories. If you've seen

Speaker:

the movie Arrival, the short story based on that is story of your life,

Speaker:

and it's wonderfully written. It's one of my favorite short stories ever.

Speaker:

Yep. So highly recommend that. I believe the collection is

Speaker:

called, story of your life and others, something like that. So

Speaker:

Oh, interesting. Careful with audiobooks. They are very

Speaker:

addictive. So,

Speaker:

with Audible is a sponsor of the show. So if you go to the data

Speaker:

driven book.com, you'll get routed to Audible and

Speaker:

you'll get a free book on us. And if you

Speaker:

choose to subscribe, we'll get a little bit of kickback. It helps run the show

Speaker:

and helps, helps us bring, bring some good stuff to to

Speaker:

the masses. So any any parting thoughts?

Speaker:

No. But thank you so much for having me on, Frank. This was a ton

Speaker:

of fun. I learned a lot from you, and I hope I I helped you

Speaker:

learn one one small thing as well. Absolutely. It was it was

Speaker:

a great conversation, and, we'll let the nice British lady finish the

Speaker:

show. And that's a wrap for this episode of Data Driven, where we

Speaker:

journeyed from the intricacies of vector databases to the surprising

Speaker:

elegance of origami. A huge thank you to Arjun Patel for

Speaker:

sharing his insights on retrieval augmented generation and his passion

Speaker:

for making AI accessible to all. From turning raw data

Speaker:

into actionable knowledge to turning paper into art, Arjun

Speaker:

proves there's beauty in both precision and creativity. If today's

Speaker:

episode left you curious, inspired, or just itching to fold a

Speaker:

piece of paper into something meaningful, be sure to check out

Speaker:

Arjun's work and Pinecones innovative tools. Remember,

Speaker:

knowledge might be power, but sharing it makes you a force to be reckoned

Speaker:

with. As always, I'm Bailey, your semi sentient guide to

Speaker:

all things data. Reminding you that while AI might shape our

Speaker:

future, it's the human touch or sometimes the paper fold that

Speaker:

gives it meaning. Until next time, stay curious,

Speaker:

stay analytical, and don't forget to back up your data.

Speaker:

Cheerio.