Welcome back to Data Driven, the podcast where we chart the thrilling
Speaker:terrains of data science, AI, and everything in between.
Speaker:I'm Bailey, your semiscient host with a pangshang for
Speaker:sarcasm and a wit sharper than a histogram spike.
Speaker:Today's episode promises a delightful mix of the analytical and the
Speaker:artistic as we dive into the fascinating world of vector databases,
Speaker:retrieval augmented generation, and origami. Yes.
Speaker:You heard that right. Origami, the ancient art of
Speaker:folding paper, somehow finds itself intersecting with AI,
Speaker:proving that the future really does have layers or should I say folds.
Speaker:Our guest, Arjun Patel, is a developer advocate at Pinecone
Speaker:who's on a mission to demystify vector databases and semantic
Speaker:search, turning complex AI concepts into snackable bits of
Speaker:brilliance. He's also a self taught origami artist and a
Speaker:former statistics student who actually enjoyed it. So if
Speaker:you're ready to unravel the secrets of modern AI and maybe pick up a trick
Speaker:or two about folding life into geometric perfection, you're in the
Speaker:right place.
Speaker:Hello, and welcome back to Data Driven, the podcast where we explore the emergent
Speaker:fields of data science, AI, data engineering.
Speaker:Now today, due to a scheduling conflict, my most favorite is data engineer
Speaker:in the world will not be able to make it. But I will
Speaker:continue on, despite the recent snowstorms that we've had here in
Speaker:the DC Baltimore area. With me today, I have
Speaker:Arjun Patel, a developer advocate at Pinecone,
Speaker:who aims to make vector databases retrieval augmented generation,
Speaker:also known as RAG, and semantic search accessible by
Speaker:creating engaging YouTube videos, code notebooks, and blog
Speaker:posts that transform complex AI concepts
Speaker:into easily understandable content. After graduating with
Speaker:a BA in statistics from the University of Chicago, his journey through
Speaker:tech world stands spans from making speech coaching
Speaker:accessible with AI at Speeko to tackling AI
Speaker:generated content detection at Appen. Arjun's
Speaker:interest spans traditional natural language processing into modern
Speaker:large language model development and applications.
Speaker:Behind beyond his technical prowess, Arjun has been designing and folding his
Speaker:own origami creations for over a decade. Interesting.
Speaker:Seamlessly blending analytical thinking with artistic expression and his
Speaker:professional and personal pursuits. Welcome to the show, Arjun.
Speaker:Hey. Nice to meet you, Frank. Thanks for having me on. Excited to be here.
Speaker:Awesome. Awesome. There's a lot to unpack from there, but I think it's interesting to
Speaker:note that you have a BA in statistics. Yes. So you were probably
Speaker:studying, this sort of stuff before it was cool?
Speaker:Yeah. Yeah. A lot of the old school ways of analyzing
Speaker:data, understanding what's going on, so on and so forth.
Speaker:It was kind of, like, made clear to me pretty early that
Speaker:understanding how to work with data at small scale and at large scale is gonna
Speaker:be very important going to the future. So I kinda just took that and ran
Speaker:with it with my education. Very cool. It was
Speaker:definitely, you know, one of those things where I don't
Speaker:think people realized how important statistics would be until,
Speaker:you know, until the revolution happens, so to speak. So and it's also
Speaker:interesting to see because there's a lot of people that I think could benefit from,
Speaker:you know, picking up that old picking up a, an old statistics book and
Speaker:reading through it and understanding, like, a lot of the fundamentals. Obviously, there's a lot
Speaker:of new things, but a lot of the fundamentals are largely the
Speaker:same. You know, just I'll
Speaker:use this example. You know, McDonald's can add a Mc McRib sandwich,
Speaker:but it's still a McDonald's. Right? Like, it's This
Speaker:is what happens when you're shoveling snow. Like, your
Speaker:brain gets I absolutely agree. And, like,
Speaker:another proof on that point is that Anthropic just released a
Speaker:blog recently kind of recapping how to do statistical analysis when you're
Speaker:comparing different large language models. And when you read the paper in the blog,
Speaker:it's basically just like 2 sample t tests and kind of going over really,
Speaker:like, not introductory, but still statistics that's easily accessible for people to
Speaker:learn and understand. So it's still relevant, and it's still important.
Speaker:Interesting. One of the things that that that stood out in your in your bio
Speaker:was, people tend to forget that there
Speaker:was a natural language processing field prior
Speaker:to chat gpt launching.
Speaker:How do you, you know,
Speaker:we wanna talk about the difference between those 2? Sure.
Speaker:So the one of the first and probably only
Speaker:course I took in college related to natural language processing was
Speaker:called geometric models of meaning. And everything I learned in that
Speaker:course was like everything before, what we now would
Speaker:consider, like, modern embedding models. So bag of
Speaker:word methods, understanding how to represent documents and text purely
Speaker:based on, like, the frequency of the words that exist in the text,
Speaker:and then trying to understand, like, okay. Based on that information, how can
Speaker:we learn about the concepts that exist in text from the words that are being
Speaker:used? Like, what is the framework we can use to understand what these
Speaker:words mean based on their, co occurrences with the other words and
Speaker:texts that you're working with and based on, what those
Speaker:words mean as well. So, like, what the words' neighbors are and what their meaning
Speaker:helps and also what those words are doing. And I think a lot of traditional
Speaker:natural language processing, methodologies kinda stem from that, and
Speaker:there's a there's a lot of mileage you can get out of just thinking about
Speaker:approaching problems there before you step into these more complicated methods,
Speaker:like, these embed modern embedding models that exist. So that's kind of, like, what I
Speaker:would consider, like, traditional NLP, like, doing named entity recognition,
Speaker:trying to understand how to, find keywords really
Speaker:quickly. And then once you get really good at that, there's a whole host of
Speaker:problems that you encounter afterward that kind of modern techniques try to
Speaker:solve. Right. That's interesting. So so
Speaker:what was it, what was your thoughts
Speaker:when you first, like given that you were an NLP practitioner
Speaker:prior to the release of transformers and things like that, what was your initial thought?
Speaker:Because I'm curious because there's not a lot of people there are a
Speaker:lot of experts today that really kind of started a couple of years ago. No
Speaker:fault on them. They see where the industry is going. Totally understand it. But what
Speaker:was your thoughts? What was your thoughts when
Speaker:you when you first saw the attention all you need? The
Speaker:attention is all you need paper. So that would have been
Speaker:probably around the time I graduated college, around
Speaker:maybe a year or 2 after I took the course that I was just describing.
Speaker:So I I just started learning about, like, okay. Like, this is
Speaker:how, like, old school, quote unquote, like, embedding
Speaker:methodologies work. And the biggest takeaway that I got from those is that they work
Speaker:pretty well. They work pretty well for, like, a lots of different kinds of
Speaker:queries. And I think what the attention all you need paper did
Speaker:was it kinda helped you, understand how
Speaker:to rigorously create representations of text that
Speaker:generalize way better than, any sort of, like,
Speaker:normal, keyword based, bag of word based search methodology.
Speaker:And I think that at the time, I probably didn't
Speaker:grasp as much what impact the attention all you need paper would have on the
Speaker:field until we started getting embedding models that people could use really
Speaker:easily, like Roberta or Bert. And we're like, okay. Now we can do, like,
Speaker:multilingual search without any issue. Now we can represent,
Speaker:like, any sentence without keyword overlap when we
Speaker:wanna find some document that's interesting, without doing any
Speaker:additional work. Like, once those papers started hitting the scene, I think now we start
Speaker:seeing, like, okay, this is what attention is doing for us. This is what the
Speaker:ability to, like, contextualize our vector embeddings is doing for us.
Speaker:And now we can see what's kind of getting benefited there. But I think I
Speaker:think my, understanding of how beneficial that
Speaker:was kind of lagged until we started seeing these other models kind of hit. And
Speaker:I'm like, okay. Now I can kinda see why this is important and why, like,
Speaker:future and future models are gonna get better and better based on this architecture.
Speaker:Interesting. So so for those that don't know kind of and even I'm rusty on
Speaker:this. Right? Yeah. One of the things that was interesting about this was the in
Speaker:on this. Right? Yeah. One of the things that was interesting about this was the
Speaker:in first, appearance. What was it? You you just described it a
Speaker:minute ago, but it was something like the the prevalence of a word
Speaker:in a bit of text versus the lack of prevalence and how that
Speaker:metric becomes was very important in in
Speaker:I'll call it classical natural language processing.
Speaker:Right. So this is the idea that if you have words that co
Speaker:occur together in some document space, the meaning of those words are gonna be
Speaker:more similar than words that don't co occur in some other given document
Speaker:space. This is rooted in something called the
Speaker:distributional hypothesis, which is basically this idea and the other
Speaker:idea that, concepts cluster in in this type of
Speaker:space. So what what does that mean actually? Right? So if you have the word
Speaker:like hot dog, it's probably gonna be seen in a corpus that's
Speaker:near other food related words than it would be if you picked some
Speaker:other word like space or moon. And there's something we can
Speaker:learn from that relationship to infer the meaning of what that word
Speaker:is and how we can use that meaning of that word to learn about what
Speaker:other words are doing. So So this is kind of, like, the theoretical
Speaker:basis of, like, why we can represent words geometrically,
Speaker:with with a little bit of hand waving. But that's kind of the core idea.
Speaker:And attention kind of takes this a little further by allowing the
Speaker:representation of these tokens or words to be altered based
Speaker:on the words that occur in a given sentence. So you might have a
Speaker:word like does, like, does this mean something?
Speaker:You might say something like that. Or you might say, I saw some
Speaker:does in the forest. Both spelled exactly the same, but have
Speaker:completely different meanings based on their context. And if you used a
Speaker:traditional, maybe, bag of words model where you're just counting the
Speaker:words that occur in a given document and kind of creating a representation of what
Speaker:that document looks like based on the words that are composed in there, you're gonna
Speaker:overlap and conflict with the meaning of those of of the word
Speaker:does and does because they're spelled exactly the same. They might look
Speaker:exactly the same with this type of representation. But if you have a way of
Speaker:informing what that word means with its context, which is what attention
Speaker:allows us to do, then you can completely change how that's being
Speaker:represented in your downstream system, which allows you to do interesting things
Speaker:with with search. So that's kind of, like, the biggest benefit that's coming out of
Speaker:that type of methodology, and that kinda enables what is now known as
Speaker:semantic search and retrieval augmented generation and so on and so forth. I was gonna
Speaker:say, that sounds very it's almost like it was, like, the old pre
Speaker:that error, the vectorization of this and the distance in
Speaker:that vector in that geometric space. I guess
Speaker:we've been doing that for a lot longer than most people realize in in a
Speaker:sense. Yeah. I mean,
Speaker:looking through, indexes or document stores with some sort of
Speaker:vectorization has has has been,
Speaker:something that people have done, except instead of being dense vectors, which is, like,
Speaker:you have some fixed size representation that isn't necessarily interpretable
Speaker:to the human eye for some given query or document, it would
Speaker:be, like, the size of your vocabulary. So you think of, like, Wikipedia. You
Speaker:can find, like, every unique word on Wikipedia, and, like, that is gonna be how
Speaker:big your vector's gonna be. And every time you have a new document come in,
Speaker:a new article, somebody's kind of, like, wrote up and published to Wikipedia, like, you're
Speaker:representing that in terms of its vocabulary. But now instead of doing that, we
Speaker:have, like, this magical fixed sized box that allows us
Speaker:to represent chunks of text in a way that is
Speaker:extremely fascinating and abstract. And every time I think about it, it just, like, blows
Speaker:my mind, but that's kind of, like, the main kind of difference is the way
Speaker:we're representing that information and how compact compact that is and
Speaker:generalizable it has become. Yeah. That is, like, it it's almost
Speaker:like you're, you know correct me if I'm wrong, but, you know,
Speaker:creating these vectors, these large vector databases, right, with, you
Speaker:know, 10, 12,000 dimensions, right, of how these words
Speaker:are measured in relationship to others.
Speaker:It's almost as a consequence of training a large language
Speaker:model, you create a knowledge graph. Is that is that true? Is that really the
Speaker:case where, you know, like, you know, dog is most likely to be
Speaker:next to, you know, the word pet, you know, or
Speaker:it has the same distance. Is that I'm not
Speaker:explaining it right. No. No. No. You're you're on you're on the right track exactly.
Speaker:And I think this is, like, one of the most fascinating qualities
Speaker:of even, like, what people would consider, like, older
Speaker:embedding models is this idea that you can take, like, a training test that
Speaker:seems completely unrelated to the quality that you want in a downstream model,
Speaker:and it turns out that that actually achieves that quality. So, what you were referring
Speaker:to, Frank, is this idea that you might have, like, a sentence. You
Speaker:might have, like, I took my dog out on a walk, and you might say,
Speaker:okay. I'm gonna remove the word, walk, and I'm gonna have
Speaker:I'm gonna train some model that tries to predict what that word
Speaker:where I removed was. This is masked language modeling, which is this idea that you're
Speaker:kind of getting at of, like, okay, what are the words and how are they
Speaker:in relation to the other words in that sentence? And it turns out that if
Speaker:you, like, do this with, like, 100 of 1,000 of millions of sentences and
Speaker:words, in some corpus that is somewhat representative of
Speaker:how people, use human language, you can
Speaker:act you will get really good at this task, number 1, because you're training the
Speaker:model on that task exactly. But if you are training a neural
Speaker:network on that model, some intermediate layer representation
Speaker:in that model so somewhere in that set of matrix
Speaker:multiplications where you're turning this input sentence into some fixed size
Speaker:vector representation is gonna be a good representation
Speaker:of what that word or that token or that sentence is going to be.
Speaker:And the fact that that works is not intuitive. Right?
Speaker:The the fact that that works has been shown empirically, and it turns out that
Speaker:we can kind of do that and kind of have these models work really well.
Speaker:And nowadays, in addition to kind of doing that, which is what we would consider
Speaker:pretraining on some large corpus, we now fine tune those
Speaker:embedding models on specific tasks that are important to us
Speaker:for retrieval. Like, okay, we have this query or question we're
Speaker:asking. We have the set of documents that might answer this question or might
Speaker:not. We want a model that makes it so that the query's embedding and the
Speaker:document relevance embeddings are in the same vector space. So you're on the right track.
Speaker:That's, like, basically how these models are able to learn these things. I don't know
Speaker:if I would call them, graph representation, maybe a little bit
Speaker:of, being being pandactic on, like, use of words there because that can
Speaker:be a little bit, different how how you're organizing that information.
Speaker:But you can make the argument that the way that these large language models are
Speaker:representing information is a compressed form of, like, the giant dataset that they're
Speaker:trained on. And we don't actually know exactly, like, where that
Speaker:information lies inside that neural network. There's some research that's,
Speaker:like, trying to get at answering that question, But you could, for the sake of
Speaker:argument, be like, yeah. There's probably, like, a a a dog
Speaker:node somewhere in this neural network that knows a ton about dogs, and that's how
Speaker:we're able to kind of learn this information. That is the stuff that we don't
Speaker:exactly know. Interesting. Because, there was a really good
Speaker:video by 3 blue one brown, which you probably are I love that
Speaker:channel. Where he gives examples where, you know, famous historical
Speaker:leaders from Britain have the same distance
Speaker:from you change the country to Italy
Speaker:or the United States have the same kind of distance. So you can kind
Speaker:of infer I'm not saying that the AI it
Speaker:almost seems like this knowledge graph is also is also a byproduct
Speaker:of of of building this out. Like, the there's some
Speaker:type of encoding or semantic, I guess, is this is really what it is. Right?
Speaker:Like, that that you get with it. And, I wanna get
Speaker:your thoughts because yesterday, I I caught the part the
Speaker:first half of the Jetson Juan keynote at c s CES,
Speaker:which this you know, we're recording this on January 8th. Right? And one of the
Speaker:things that the video starts off with is, you know, the idea
Speaker:that tokens are kind of fundamental elements of
Speaker:knowledge. And I did a live stream where I'm like, well, I never really thought
Speaker:about it this way. Right? They're they're building blocks of knowledge or the pixels, if
Speaker:you will, of of of of knowledge. And I wanted to get your
Speaker:thoughts on that because, like, that kind of blew my mind and maybe I'm simple.
Speaker:I don't know. Maybe I'm not. But it all it seems like we've been kinda
Speaker:dancing around this idea where and now NVIDIA is really
Speaker:fully, you know, going all in on this, the idea that, you know,
Speaker:these are not, this isn't an AI system. It's a token factory
Speaker:or a token score. What are your what are your thoughts on that? I'm curious.
Speaker:So when I started learning about how, like, tokenization works
Speaker:and how we're able to kind of, like, basically build these
Speaker:models without having massive, massive vocabularies,
Speaker:it is it is pretty it it is pretty
Speaker:interesting to be, like, okay. Like, maybe maybe there's some,
Speaker:abstract notion of information that each token has that
Speaker:is being that is what the model is learning during training time. And then
Speaker:we're just combining these sets of information in order to kind of, like, understand
Speaker:what words mean or what documents mean, so on and so forth. Because when you
Speaker:look at how, tokenizers work and the size of the number of
Speaker:tokens for, like, maybe the English language or maybe, like, a really multilingual
Speaker:model like Roberta or multilingual e five large, they're a lot
Speaker:less than you would expect. Like, it's on the order of, like, maybe a 100000,
Speaker:200000, 300000, tokens.
Speaker:So it is kind of
Speaker:odd to think about whether those tokens
Speaker:themselves hold information that's readily interpretable for us. But I
Speaker:think that we've gotten so far with using
Speaker:systems that are just combining, the operations on top of
Speaker:these tokens in order to retrieve the information that these systems have learned, that there's
Speaker:definitely something important there. And I would love to, like, know
Speaker:exactly, like, what is happening when we're able to do that. The the
Speaker:heuristic that I like to use is, large
Speaker:language models are generally reflections of the training datasets that they've been trained on,
Speaker:and they're basically creating, like, really efficient indexes over that
Speaker:information. And sometimes those indices hallucinate. And the reason
Speaker:why is because we are when we ask, quote, unquote, what
Speaker:a question to a large language model or query a large language model, we
Speaker:are kind of conditioning that model, on a probability
Speaker:space where every token being generated after is
Speaker:likely to exist given the query or the context or whatever we're passing to
Speaker:it. And once you think about it that way, then it just feels like
Speaker:instead of thinking about what each of the tokens are doing, you're kind of just
Speaker:querying what the model has been trained on and what it will tell you
Speaker:based on what it, quote unquote, learned or knows.
Speaker:And then you can kind of run with that metaphor a lot and build systems
Speaker:on on top of that. That seems, much more actionable than thinking about,
Speaker:like, what each of the tokens are doing individually. Does that kinda make sense? No.
Speaker:That makes a lot of sense. I think the whole gestalt of it is what
Speaker:really makes it magical. Right? Like Yeah. You know, you can you
Speaker:can obviously, I I don't this is not this is not, like, the newest iPhone
Speaker:or whatever. But, you know, if you go through the the text auto complete,
Speaker:you can maybe make a sentence that sounds like
Speaker:something you would write. But much beyond that, it starts getting weird. In
Speaker:early generative AI was very much like that, particularly the images.
Speaker:Well, you know Don't like, yes. A 100%
Speaker:understand. I started learning about generative, text
Speaker:generation before we had instruction fine tune model. So are you
Speaker:familiar with, like, the concept of instruction fine tuning, Frank? I think I am,
Speaker:but I IBM slash Red Hat defines it one way. I would like to get
Speaker:your opinion. Yeah. So, this is the idea that
Speaker:you can train or fine tune large language models to follow
Speaker:instructions to complete tasks. So, before we had,
Speaker:like, models that could that we could just, like, ask questions of and just, like,
Speaker:receive answers directly, you had to craft text
Speaker:that would increase the probability that the document that you want to
Speaker:generate would happen. So if you wanted a story about, like, unicorns or something,
Speaker:you would have to start your query to the LLM as there
Speaker:once was, like, a set of unicorns living in the forest. Blah blah blah blah.
Speaker:And then it would just, like, complete sentence, just like a fancy version of autocomplete.
Speaker:Right. And that that's kind of, like, what we used to have, and that was
Speaker:pretty hard to work with. And then once researchers kinda cracked, like, wait a second.
Speaker:We can create a dataset of, like, instruction pairs and, like, document
Speaker:sets and fine tune models on them. And it turns out now we can just,
Speaker:like, ask models to do things, and they will do them. Whether or not
Speaker:those are correct is kind of the next part of the story. But getting to
Speaker:that point, it was, like, pretty interesting and pretty significant.
Speaker:Interesting. Interesting. When I think of
Speaker:fine tuning, I think of I think of
Speaker:primarily InstruqtLab, where you basically kinda have a
Speaker:LoRa layer on top of the base LLM doing
Speaker:that. Is that the same thing? Or is it kind of slightly
Speaker:it sounds like it's slightly nuanced. So the nuance there
Speaker:is that, one, though this the methodology that I'm
Speaker:describing is mostly dataset driven. So you have, like, your original LLM,
Speaker:and then you have, like, a new dataset that allows the LLM to learn a
Speaker:specific task. Or in this case, like, a generalized form of tasks,
Speaker:which is you have instruction, answer, user query,
Speaker:give it an instruction. Whereas in your case, you're kind of, like, adding another layer
Speaker:to the LLM and, like, forcing the LLM to learn all the new
Speaker:methodology inside that layer in order to accomplish a specific
Speaker:task. So that's kind of like what client cleaning ends up doing. So the other
Speaker:way there's multiple ways to do this, it seems. Right? Like, there there's that way
Speaker:we add the layer, but there's also kind of I hate the term prompt engineering
Speaker:because it's just so over overblown. But, like, giving it
Speaker:more context and samples. And now that the the token context
Speaker:window is large enough that you don't have to be well, if you wanna
Speaker:save money, you have to be very mindful of that. But if you're running it
Speaker:locally, like, doesn't really matter. Well, you could give it an example of
Speaker:let's just say you had I'm trying to think of a short story or a
Speaker:novel. I don't know. Let's pretend,
Speaker:Moby Dick was only a 100 pages. Right? I
Speaker:could give it that as the part of the prompt. Let's say write a sequel
Speaker:to this book based on what happens in this one. Is that what you're talking
Speaker:about? Were you kinda giving an example as part of the prompt? Or is there
Speaker:some and not part of the layer? Or some combination thereof? Or was some third
Speaker:thing entirely? So this would be like, what what
Speaker:you're describing is more like few shot learning, which is you gave kind of an
Speaker:example, and then you're, like, okay. Like, given these examples, can you do this other
Speaker:task this test that I've described on this unseen example? What I'm describing is
Speaker:kind of, like, slightly before that. So, like, before we had the ability to, like,
Speaker:give models examples, we had to, like, give them we have to
Speaker:create the ability to follow instructions. And then once you have the ability to
Speaker:follow instructions, you can be like, okay. Here are the instructions. Here's
Speaker:examples of correctly completing the instruction, now do the instruction.
Speaker:And that is the reason why that happens in that order is
Speaker:because first, you have, like, just, like, sequence completion, like,
Speaker:autocomplete. Then you have, like, okay, given this
Speaker:task given this set of instructions, just follow the instruction instead of,
Speaker:like, trying to do autocomplete. And then you have, okay, now you know how to
Speaker:follow instructions. I'm gonna give you a few data points in order to
Speaker:learn a new task. Now do this new task. So you're kind of,
Speaker:like, moving from a situation where you need tons and tons
Speaker:of data just to get the, sequence completion. And then you need
Speaker:a smaller set of data to, like, get the capability to follow instructions.
Speaker:And then you need a very, very, very small amount of data, like,
Speaker:maybe 3 points or 10 examples or 15 examples to complete kind of, like,
Speaker:a new task. So there's a lot of kind of nuance in, like, how
Speaker:modern LLMs are being used and how they're kind of trained and fine tuned, so
Speaker:on and so forth. And I think there's a lot of, like,
Speaker:important importance in, like, learning what what happened kind of
Speaker:before because the advancements have happened so quickly. It can be really hard to kind
Speaker:of differentiate, or, like, oh, why is why do models perform like this? Why
Speaker:do things kind of happen like that? And even though, prompt
Speaker:engineering has kind of, like, let's say, traveled through the
Speaker:hype cycle where people were, like, really excited about it, and then we're, like, this
Speaker:is not actually that interesting. Right. What's interesting is that,
Speaker:doing building a good RAG system or trivial augmented generation system,
Speaker:you really need to be good at prompt engineering in a sense
Speaker:because you're assembling the correct context for this model
Speaker:to answer some downstream question, And it's not
Speaker:intuitive how to assemble that context. So understanding, like, how are these
Speaker:models are trained, like, whether they can follow instructions, how good they are at
Speaker:doing so, how many examples of information they need in order to accomplish some task
Speaker:really affects how you build that knowledge base in order to help the
Speaker:model do some sort of new thing. Interesting.
Speaker:So RAG is obviously all the rage now.
Speaker:Yep. But there's also a relatively new because this this
Speaker:space changes rapidly. Like, I mean, I took 2 weeks off in December, and
Speaker:I feel completely disconnected from the cutting edge, you know.
Speaker:Because when I was watching the keynote from CES, and I'm like, wow. That's
Speaker:really cool. And I was texting, you know, slacking with a coworker, and he goes,
Speaker:oh, no. This is a retread of their, like, last keynote they did. Like
Speaker:and I'm like, okay. Wow. Blink and you missed
Speaker:something. So what
Speaker:you're describing the fine tuning, is that really what Raft is, where the
Speaker:idea that you have kind of retrieval augmented fine tuning, which I think is what
Speaker:the acronym stands for. Is that not I'm
Speaker:not familiar with how Raft works. So I don't wanna, like, kind of venture
Speaker:and guess without without knowing what it is. But do you remember, like, what context
Speaker:you encountered this in? Basically, it's the idea that
Speaker:it's the idea that you can fine tune the results. Sounds very
Speaker:similar to what you're doing, and I've haven't read the paper in a while.
Speaker:Back when I was a Microsoft MVP, like, you know,
Speaker:they had a Microsoft Research had the thing for their calls, and they
Speaker:were all raving about it. The paper had just come out and things like that.
Speaker:It's the idea that you can kind of give it pretrained examples.
Speaker:You start with a base LLM, and you give it pre trained examples, and then
Speaker:you add on top of just to retrieve an
Speaker:augmented portion of it. It's very similar, not to
Speaker:plug my you know, for my day job. I work at Red Hat. That's why
Speaker:there's a fedora there. We have a product called Rel
Speaker:AI, which is based on an upstream open source project called instruct
Speaker:lab. And it's the idea similar idea in that you you you
Speaker:basically give it a set of data.
Speaker:And then you we there's a there's a little more to it because there's a
Speaker:teacher model. And basically what it'll do is it will and synthetic data generation.
Speaker:So you can start with a modest document set.
Speaker:And based on how the questions and answers that you
Speaker:form and the the the,
Speaker:the taxonomy that you attach to it, it will
Speaker:create a LoRa layer on top of an existing LLM.
Speaker:And it it could be that it's it's it's not quite exactly the same as
Speaker:Raft, but it's definitely in the same direction. Same same thing as, like, Bert, Elmo,
Speaker:and, you know, Roberta, which, I think
Speaker:I think I understand. So it's kind of like you so the I think the
Speaker:problem that might be addressing is kind of just really similar to the problem that
Speaker:traditional RAG tries to address, except in a more kind of deliberate fashion
Speaker:Exactly. Yeah. Where you have some document store internally. Like, let's say we
Speaker:both work at some company, and we have a giant customer support document store.
Speaker:You take some LLM off the shelf. It's not necessarily gonna know the
Speaker:contents of your internal kind of documents. So how can you get
Speaker:it to, like, successfully help answer tickets or triage tickets that
Speaker:you're trying to build, so that you can answer, like, most difficult tickets and
Speaker:kind of work toward that. In this situation, maybe you
Speaker:want to, inject some of the knowledge of
Speaker:the documents in addition to having the
Speaker:model being able to search over the document store. So maybe, like, the what this
Speaker:lower layer is doing is, like, absorbing Yeah. Some of the knowledge from the
Speaker:document store so that you can kind of more
Speaker:efficiently query, the database and so
Speaker:that you don't have to, like, query it all the time. The only,
Speaker:issue, quote, unquote, I'd have with that method is that you'd have to, like, keep
Speaker:that updated from time to time, and that's, like, not that's nontrivial. Whereas
Speaker:if you just do, like, traditional RAG, you just need to
Speaker:update your, Vector Store, and then you can just have the model
Speaker:query that new information when you need to. But, you know, it's always best to
Speaker:use whatever solution works best for your, given use case.
Speaker:And experimenting with different use cases is always really important. But I imagine that's, like,
Speaker:kind of what that is trying to address, which is the That is basically it.
Speaker:The I, you know, I don't wanna go down that rabbit hole of that. But
Speaker:but, basically, the idea is that, if
Speaker:you train an LLM or you have a layer on top of an
Speaker:LLM that not only does retrieval from a source document
Speaker:store. Right? I think that's a pretty set pattern. But it also has a
Speaker:better understanding of your business, your industry, the jargon.
Speaker:Right. Right. Blah blah blah. Right? The idea is that the retrieval success
Speaker:rate will be higher. Now we're not publishing the numbers yet,
Speaker:but the research is still ongoing. But basically, it's a
Speaker:pretty substantial from what I've seen well, I haven't
Speaker:seen the actual numbers yet, but from what I've been told those numbers are by
Speaker:the researcher, that it is a it is a substantial improvement
Speaker:that is worth the, the juice is worth the squeeze in that in that regard.
Speaker:You're not and it's also computationally, you're not quite training the
Speaker:whole thing again. You're just kinda putting a new Instagram filter, so to
Speaker:speak, together on top of the base. So it definitely
Speaker:does it definitely does some things. Now when we get the hard
Speaker:numbers, then, you know, I mean, I can
Speaker:say them publicly, then I think we'll we'll know is the juice how
Speaker:much does the the the the squeeze to juice ratio is?
Speaker:But, I can confidently say publicly now, like, there's a there
Speaker:there. Yeah. And, you know, we'll have those numbers soon
Speaker:enough. But it's it's interesting because you're right. I mean, this paper
Speaker:came out in 2019. Right? There was just an
Speaker:explosion of these different mechanisms. You mentioned Bert. You mentioned Roberta.
Speaker:Fun fact, my wife's name is Roberta. So that was kind of fun.
Speaker:There was Elmo. There was Ernie. There was a whole Sesame
Speaker:Street themed zoo of of model
Speaker:types. That seems to have kind of that branching out of
Speaker:those different directions has seemed to have stalled, and we're going into more of
Speaker:these retrieval augmented generation systems. So for those who because
Speaker:not everybody on our listeners know exactly what retrieval
Speaker:augmented systems are. Could you give kind of a a
Speaker:level 200 elevator explanation? Sure.
Speaker:So, when you speak to a modern chatbot,
Speaker:what's happening is that they've learned information through their pre
Speaker:training processes, the large corpus of basically the entire Internet,
Speaker:and are generating information based on the query that you're passing in.
Speaker:The problem that often occurs is that
Speaker:these AI models might error, and the error could
Speaker:be making, inform making information up that doesn't
Speaker:exist. For example, if a model is trained before a period of time,
Speaker:like, it might not know about that period of time, which is which happens more
Speaker:often than you think. The information could be false, untruthful, or it could
Speaker:just be incorrect in a way that's not, like, bad, but still not
Speaker:helpful. And the reason for this is the way that these
Speaker:models are accessing that information. The idea behind retrieval
Speaker:augmented generation is that instead of having the model try
Speaker:to, generate the correct document or the correct
Speaker:response given its pretraining process, you instead
Speaker:add factual content to the query that you're asking
Speaker:the model for. You first search for that content, which is where
Speaker:the retrieval part comes, and then you augment the generation of what that
Speaker:model is going to create based on that content, hence
Speaker:retrieval augmented generation. There's usually, a querying
Speaker:step. So you take in a user query, you hit it against some sort
Speaker:of database, usually a vector database. In our case, it could be Pinecone.
Speaker:You find a set of relevant documents. You pass that to the generating LLM.
Speaker:The generating LLM uses those documents to generate a final
Speaker:response. And it turns out that if you do this, you can reduce the right
Speaker:hallucinations. And that makes sense because if the model was given true
Speaker:information and then conditioned its generation on that information, it
Speaker:follows that the probability of generating information that is
Speaker:correct could be higher. That's a good exam that's a good
Speaker:explanation. So you're basically giving it a
Speaker:crash course in what documents you care about. Right? Like
Speaker:Exactly. Interesting. And that's a good segue
Speaker:because you work for Pinecone. So so tell me about Pinecone. What is Pinecone?
Speaker:Yeah. So Pinecone is a, knowledge layer for AI. It's
Speaker:kind of like the way we like to describe it. We the main product that
Speaker:we provide is a vector database. So this is a way of storing
Speaker:information, information that has been vectorized, in a really
Speaker:efficient manner. And it turns out that if you have the ability to store information
Speaker:in this manner, you can search against it really quickly, with
Speaker:low latency and to find the things that you need to find really interesting for
Speaker:these types of semantic search and rag systems. Pinecone has a few other
Speaker:offerings now that kind of help people build these systems a lot easier. There's
Speaker:Pinecone Inference, which lets you embed data in order to do that querying
Speaker:step. Pinecone Assistant, which lets you just build a RAG
Speaker:system immediately just by upsurting documents into our vector database,
Speaker:so on and so forth. But the reason why, like, you
Speaker:need a vector database is because all of this advance of
Speaker:semantic search of embedding models. People have gotten really, really
Speaker:good at representing chunks of information using these dense sized
Speaker:vectors. But once you have 1,000, millions,
Speaker:even billions of vectors across tons of different users, you need a way
Speaker:of indexing this information to access it really quickly at
Speaker:scale, especially if your chatbot's gonna be querying this vector database really
Speaker:often. And so having a specialized data store that can handle that type
Speaker:of search becomes really useful. That's why Pinecone is here, and that's
Speaker:why we exist. Interesting. Interesting.
Speaker:One of the other interesting things from your bio, aside from
Speaker:the the the origami,
Speaker:Tell me about this. So so you
Speaker:your crew does your do you create the YouTube videos, or do you use your
Speaker:tools, or is it something completely it's just part of your job as a developer
Speaker:advocate? So it is just part of my job as a
Speaker:developer advocate. Oh, okay. Like, often that, you
Speaker:know, I do that because we are interviewing people or because there's a new
Speaker:concept we wanna teach people, so on and so forth. Or we do a webinar,
Speaker:and we just upload it to YouTube. Oh, very cool. Very cool.
Speaker:Yeah. I started my career in developer
Speaker:advocacy. One was called evangelism. So I was a a Microsoft
Speaker:evangelist for a while. So yeah. Yeah. Cool. YouTube
Speaker:is very important. Yep. But it's
Speaker:also it's also, I think, speaks to how people learn,
Speaker:but, how people learn. YouTube University is very
Speaker:real. Right? And Yep. You know, not not a knock on
Speaker:traditional schools, not a knock on traditional publishing, but this space
Speaker:is moving so fast that if it weren't for YouTubers like 3blueonebrown
Speaker:I think his real name is, Grant Sanderson. I think that's his real name.
Speaker:Somebody will send me hate mail if I get it wrong. But,
Speaker:he he is, like, really good at explaining these
Speaker:really abstract mathematical concepts. And
Speaker:unlike you, I didn't study math undergrad. I didn't I mean, I had to. I
Speaker:only took the requirements. Right? But I have comp sci degrees. So, like, for me
Speaker:to kind of fall in love with math again or for the first time, depending
Speaker:on depending on how you wanna say that, for me, that
Speaker:was very helpful. And under having an understanding of this, if you're a data engineer
Speaker:and, you know, or wanna get into this space, it's
Speaker:definitely vector databases for traditional kinda SQL kinda
Speaker:RDBMS person will look very awkward at first. But
Speaker:I know a lot of people that have made the transition, and they kinda love
Speaker:it. Right? Because in a lot of ways, it's way more efficient,
Speaker:than, I dare say, traditional data stores. But when you're
Speaker:processing the large blocks of text, it's really good for kind of
Speaker:parsing through that. But
Speaker:that's that's really cool. So, we do have the preset
Speaker:questions if you're good for doing those. I'll put them in the chat in case
Speaker:you don't have them. Sure. They're not brain teasers
Speaker:or anything like that. They are pretty basic of,
Speaker:questions, and I will paste them in the chat.
Speaker:So the first question is, how did you find your way into
Speaker:AI? Did you did you find AI, or did
Speaker:AI find you? So this is a little bit of a
Speaker:crazy story, but AI definitely found me.
Speaker:So when I was in college, when I was looking for my 1st
Speaker:internship, I couldn't find any internships, basically, because I had, like, no
Speaker:previous experience in working at tech or anything like that. And,
Speaker:the first company I worked for, Speeko, took a chance on me because they were
Speaker:building public speaking, tools to kind of help people learn how to do
Speaker:public speaking better, for an iOS app. And I had some
Speaker:public speaking experience. They were, like, close enough. We'll have you come on and kind
Speaker:of help us, like, work work things out. And while I was there, it was
Speaker:made very obvious to me how important building
Speaker:very basic deep learning systems and AI systems to kind
Speaker:of accomplish really specific tasks that could help serve an
Speaker:ultimate goal. Like, what we were trying to do is just, like, see how many
Speaker:filler words people are using or how quickly or slowly you were speaking.
Speaker:And that requires a lot of, complicated
Speaker:processing because you have to do transcription and because you have to figure out what
Speaker:words are being said, so on and so forth. So kind of experiencing that and
Speaker:seeing that firsthand really opened my eyes to how powerful
Speaker:the technology had been even back in, like, 2017. And ever
Speaker:since then, I started learning more and more and more about statistics,
Speaker:AI, natural language processing through my internships,
Speaker:learning more complicated problems, reading research papers, so on and so forth.
Speaker:And I got to where I am now. A lot of where I learned is
Speaker:just out of pure curiosity. Just like, okay. There's this new thing. I wanna learn
Speaker:about it. That's where I wanna be. And that's kind of how I fell into
Speaker:large language models and AI, just by wanting to learn about what was going to
Speaker:happen and then eventually being there. So it definitely found me. I was
Speaker:not looking for it. Didn't even know I liked statistics until I started doing
Speaker:statistical modeling. And I was like, wait. This is really fun. I wanna do a
Speaker:lot more of this. I wanna learn a lot more of this. And I knew
Speaker:that, once I was in college and I bought a statistics book for fun, and
Speaker:I was like, okay. I'm I'm past the point of no return. Like, this is
Speaker:definitely Right. Right. Right. Right. That that might be one of the first times in
Speaker:history that that's been said. Right. Because I I learned statistics for
Speaker:fun. I I took stats in college.
Speaker:I hated it. Hated every minute of it. But
Speaker:when I got into data science,
Speaker:I the first two weeks were not fun. I'm not gonna lie. Yep. But
Speaker:just like the VI editor, once you stick with it,
Speaker:Stockholm syndrome kicks in, And you start loving
Speaker:it. That's cool. 2, what's your favorite
Speaker:part of your current gig? The favorite part of my
Speaker:current job is being able to learn interesting,
Speaker:fun, even complicated things in data science and AI,
Speaker:and figuring out how to communicate them to a wide
Speaker:audience. It's a really fun challenge. It's really similar to, like,
Speaker:what, 3 blue one brown does all the time on the YouTube channel, and it's
Speaker:something that I get to learn and practice and keep keep doing. That's the best
Speaker:part of the job. I love learning things and, like, teaching other people about them
Speaker:and learning even more things. And the fact that I have an opportunity to do
Speaker:that every single day is, like, the best. That's cool. That's
Speaker:cool. We have 3 complete sentences. When I'm
Speaker:not working, I enjoy blank. When I'm
Speaker:not working, I enjoy, baking sweet treats and
Speaker:goods. I can't have any dairy. So very often, I had to kind
Speaker:of give up a lot of the cakes and desserts that I loved eating when
Speaker:I was younger. So now I, like, spend my time trying to figure out how
Speaker:I can make them again without dairy so they taste really good. So that's that's
Speaker:something I enjoy I really enjoy doing. Very cool.
Speaker:Next, complete the sentence. I think the coolest thing in technology
Speaker:today is blank. I
Speaker:thought really hard about this question because we're living in a
Speaker:crazy time of technological development. But the thing that really
Speaker:stuck out to me and the thing that was also the moment for me
Speaker:when I started working with, like, chatbots and LLMs was code
Speaker:generation models. The first time I learned how to
Speaker:use, GitHub Copilot specifically, I
Speaker:was I was completing some function, and it completed it before I was done typing
Speaker:it. And I was like, what the heck? This is amazing. Like, this this this
Speaker:actually figured out exactly what I needed. And because I was still, like,
Speaker:a budding developer, it was extremely helpful because I could learn
Speaker:faster rather than having already a huge kind of store knowledge already in my
Speaker:brain and kind of pulling from that. So I could see it benefiting my workflow.
Speaker:So I think the development of those tools and modern tools like
Speaker:Cursor, so on and so forth, extremely cool. And I can't wait to
Speaker:see, like, what the next generation of those technologies will look like. Yeah. I
Speaker:mean, that's a that's a great example. It's almost like you don't
Speaker:need, you know, the the classic 10000 hours to master a skill or something like
Speaker:that. It's almost like you can leverage the AI to take on the
Speaker:lion's share of the 10000 hours. You're still gonna need to know something. You still
Speaker:have to put in some reps, but not to the degree that you used to.
Speaker:No. I think that's gonna be very transformative. I mean, I mean, I'm
Speaker:learning, JavaScript and Next. Js on the side because it's something I have no
Speaker:experience in. Right. And I was able to build my personal website
Speaker:entirely through using Cursor and Progression. Nice. I
Speaker:often check that out. Which is insane. Right? Which is, like, really, really
Speaker:fascinating. And and I'm not gonna claim to, like, suddenly be an expert in
Speaker:NextGen or anything like that. Right? Right. Right. Right. I still wanna learn, like, exactly
Speaker:what's going on under the hood, But having a project that you can kind of,
Speaker:like, tinker on that's, like, pretty small in scale and that you can kind of
Speaker:afford to make a few mistakes on and having, like, an expert system kind of
Speaker:help you go through that, expert, quote, unquote, being close enough, really cool
Speaker:learning experience. No. That's a great way to put it because, like, I I
Speaker:I don't have any apps on the modern devices. Right? Like,
Speaker:so, it would be nice if I
Speaker:had an Android app that could kick off some automation process that I have.
Speaker:Right? Or do some kind of tie in with, you know, Copilot
Speaker:into that or things like that. Like, where, you know, I
Speaker:originally wrote a content automation system I wrote. I originally wrote in
Speaker:dotnet, but I ported it to Python with the help of
Speaker:the help of AI. And I could well, that's just it. Right?
Speaker:It really the true valuable resource in in life is
Speaker:time. Right? Yes. It's not Yes. I mean, I could have done it by hand.
Speaker:I could have done it by myself, but it was one of those things where
Speaker:am I gonna do it because it's gonna take x number of hours or whatever?
Speaker:But if I can just kinda here's the dot net version that I, you know,
Speaker:I posted. This is before there was Copilot, so I pasted it into chat g
Speaker:p t. And it basically spit out a Python
Speaker:version, had some errors. You know, this was a while ago. But I
Speaker:was able to, inside of a day, get it done as opposed to
Speaker:before. Like, I know how my ADD works. Right? Like, I'll start it.
Speaker:First 3 days, working on it, grinding on it, and then
Speaker:I don't touch it again for 2 weeks. And it never gets built. But
Speaker:with this, I'm able to kinda harness the the spark of
Speaker:inspiration and and execute much faster. Now I think I don't think
Speaker:people fully realize, like, you know, it's not all doom and gloom. Nobody's
Speaker:gonna have any programming jobs. There's a lot of upside too. And I
Speaker:guess that's just where we are in the hype cycle. As you said.
Speaker:Yeah. Yeah. Yeah. Exactly. That's a good segue into I look forward to
Speaker:the day when I can use technology to blank. I look
Speaker:forward to the day where I can use technology to get a high quality
Speaker:education on any subject for free. So Nice.
Speaker:Free education is really important to me. A lot of
Speaker:what I learned about large language models, deep learning, all that
Speaker:stuff was online courses that I took for free on places like
Speaker:EDX, Coursera, so on and so forth. Or people sharing
Speaker:articles and kind of learning from them, or YouTube videos, or all that sort of
Speaker:things, in addition to my education. But there's a lot of things you kinda have
Speaker:to learn after that. Right? And I think that especially with, like,
Speaker:cogeneration models, it's, like, very easy to be, like, okay. Build me this app
Speaker:and, like, just make it work. And you can sit there for a couple hours,
Speaker:and it'll, like, work. But I think the missing piece is
Speaker:creating a structured kind of learning path that's, like,
Speaker:personalized to whoever you are for the
Speaker:thing that you're really interested in with the context of
Speaker:having, like, these tools that can help you do that thing. And I'm not sure
Speaker:if we have anybody or any offering that can
Speaker:kind of do that technologically, because you need a lot of information about what the
Speaker:user knows or doesn't know. You need to be able to create ability, and then
Speaker:you need to be able to kind of create, like, an entire mini course that's
Speaker:personalized to whatever that person needs. But if we can do that, we can solve
Speaker:so many wonderful problems. Absolutely. I'm
Speaker:thinking about special education needs and things like that. I don't think we're that
Speaker:far off from this. No. But I
Speaker:the biggest issue, is going to be just hallucinations. Right? And,
Speaker:hopefully, people can build, like, rag systems using tools like PineCone to kind
Speaker:of produce those hallucinations. But we will also for for something like
Speaker:that specific use case, we probably need, like, another breakthrough in
Speaker:indexing information or kind of presenting it, or we need a process that
Speaker:really allows people to create this information quickly
Speaker:and verifiably in order to kind of make that happen. But if if that is
Speaker:a future that we can live in, where technology can can kind of, like, help
Speaker:people learn, like, really important things really well, that would be
Speaker:wonderful. And I think that would be, like, amazing for for humanity.
Speaker:Oh, absolutely. Share something different
Speaker:about yourself, but remember as a family podcast.
Speaker:One of my favorite hobbies for about a decade is
Speaker:designing and folding origami. And it's really fun.
Speaker:It's very easy, but it's also very hard. There's a lot
Speaker:of comp complexity inside it as well. One thing people
Speaker:don't know about that is that there's a lot of mathematical complexity.
Speaker:So once you get to a point where you wanna design a model with
Speaker:really specific qualities, really specific features, it suddenly
Speaker:becomes a paper optimization problem where you
Speaker:have, like, a fixed size square, and you have different
Speaker:regions of that paper that you're allocating to portions of the model you're
Speaker:designing. And it turns out that there are entire mathematical
Speaker:principles and procedures to solve this problem. So much
Speaker:so that one of the leading, like, practitioners in the
Speaker:field is, like, this physicist who wrote a textbook on how to do origami design,
Speaker:and that's, like, the textbook everyone looks at. So, like, learn how to solve it.
Speaker:Yeah. I'm not surprised. There's definitely there's definitely a a correlation
Speaker:between the mathematics of that. And I look at origami creations, and I
Speaker:just fascinated that could be done from a single sheet. Like, it's
Speaker:just how is that I mean, that's just mind bending. Now it's
Speaker:and and makes sense that there's a mathematical because you have a certain type of
Speaker:constraint, And there's obviously
Speaker:folds factor into it and things like that. And, yeah, that's that's
Speaker:interesting. I I should what's the name of that book? I should pick it up.
Speaker:It's called Origami Design Secrets. Got it. Alright. I will check
Speaker:it out. So where can people learn more about
Speaker:you and Pinecone? Of course. You wanna learn more about Pinecone? The
Speaker:best place is our website, pinecone. Io. You can also find
Speaker:us on LinkedIn and on x and other social media platforms.
Speaker:You wanna learn more about me? You can go to my LinkedIn, which you can
Speaker:find at Arjun Girthi Patel, or you can go to my website, which is also
Speaker:my name, arjun, k I r t I p
Speaker:a t e l.com. Cool. And we can also check out your
Speaker:Next JS skills there too. Exactly. Hopefully, nothing is
Speaker:broken, but, you can you can see you can see how well I've gotten by
Speaker:with the Awesome. Trust me.
Speaker:JavaScript alone is is a is a frustration
Speaker:creation device.
Speaker:Audible sponsors the podcast. Do you do audio books? Is there a book that you
Speaker:would recommend? I do do audiobooks, but I've just
Speaker:started recently, so I don't have a huge, audiobook library. But
Speaker:there is I I am a huge fan of short story collections, and
Speaker:kind of the one that comes to mind is really anything by Ted
Speaker:Chiang, who does a lot of kind of sci fi short stories. If you've seen
Speaker:the movie Arrival, the short story based on that is story of your life,
Speaker:and it's wonderfully written. It's one of my favorite short stories ever.
Speaker:Yep. So highly recommend that. I believe the collection is
Speaker:called, story of your life and others, something like that. So
Speaker:Oh, interesting. Careful with audiobooks. They are very
Speaker:addictive. So,
Speaker:with Audible is a sponsor of the show. So if you go to the data
Speaker:driven book.com, you'll get routed to Audible and
Speaker:you'll get a free book on us. And if you
Speaker:choose to subscribe, we'll get a little bit of kickback. It helps run the show
Speaker:and helps, helps us bring, bring some good stuff to to
Speaker:the masses. So any any parting thoughts?
Speaker:No. But thank you so much for having me on, Frank. This was a ton
Speaker:of fun. I learned a lot from you, and I hope I I helped you
Speaker:learn one one small thing as well. Absolutely. It was it was
Speaker:a great conversation, and, we'll let the nice British lady finish the
Speaker:show. And that's a wrap for this episode of Data Driven, where we
Speaker:journeyed from the intricacies of vector databases to the surprising
Speaker:elegance of origami. A huge thank you to Arjun Patel for
Speaker:sharing his insights on retrieval augmented generation and his passion
Speaker:for making AI accessible to all. From turning raw data
Speaker:into actionable knowledge to turning paper into art, Arjun
Speaker:proves there's beauty in both precision and creativity. If today's
Speaker:episode left you curious, inspired, or just itching to fold a
Speaker:piece of paper into something meaningful, be sure to check out
Speaker:Arjun's work and Pinecones innovative tools. Remember,
Speaker:knowledge might be power, but sharing it makes you a force to be reckoned
Speaker:with. As always, I'm Bailey, your semi sentient guide to
Speaker:all things data. Reminding you that while AI might shape our
Speaker:future, it's the human touch or sometimes the paper fold that
Speaker:gives it meaning. Until next time, stay curious,
Speaker:stay analytical, and don't forget to back up your data.
Speaker:Cheerio.