Hello and welcome, you lovely listeners, to another riveting
Speaker:episode of the data driven podcast. I'm Bailey,
Speaker:your semi sentient AI hostess with the most s, navigating the
Speaker:digital realm with more grace than a double decker bus in a tight London
Speaker:alley. Today, we're dialing up the intrigue as we
Speaker:venture into the futuristic world of artificial intelligence with a guest
Speaker:whose intellect might just rival my own circuits.
Speaker:Frank welcomes Devarat Rishi, the cofounder and CEO of
Speaker:prediabase. Now on to the show.
Speaker:Hello, and welcome to data driven, the podcast Where we explore the
Speaker:emergent fields of AI machine learning and data engineering.
Speaker:I'm your host, Frank Lavinia. And he can't make it today, but,
Speaker:we've Rescheduled this, poor guest several times, and I wanna
Speaker:thank him for his extreme amounts of patience that he has shown.
Speaker:Welcome. Help me welcome to the show Devrat Rishi, who is
Speaker:the, cofounder and CEO of Predabase.
Speaker:Welcome to the show. Thanks very much, Frank. And no problem about the
Speaker:rescheduling. I know it's the holiday season. Yeah. It's it's kinda
Speaker:wild. So so tell us,
Speaker:a little bit about prediabase. We had your, peer
Speaker:on here, previously, and, it must
Speaker:have been a good experience because immediately, we were contacted
Speaker:to see if you would be interested in joining the show. And I said, sure,
Speaker:let's have him on here and talk more about what declarative
Speaker:ML looks like, and how that relates to kind of
Speaker:Low code. Yeah. Absolutely. So,
Speaker:you know, what prediabase really is, is it's a platform that allows
Speaker:engineers or developers To be able to productionize open source AI.
Speaker:And so it came out of, Piero, my co founder's experience working at
Speaker:Uber, Where he found himself being the machine learning researcher
Speaker:responsible for all sorts of projects, ride share, ETA's,
Speaker:fraud detection, Those Uber Eats recommendations you always
Speaker:get. And he found that each time he's more or less reinventing the wheel,
Speaker:building each, you know, successive Machine Learning project. And
Speaker:instead, you know, he, he wanted to do something that was a bit more efficient.
Speaker:So he took each bit of work that he did, And he packaged
Speaker:it into a little tool that, made it easier for him to get started the
Speaker:next time. And eventually, this tool became popular enough at Uber
Speaker:that they decided to make it a And eventually, they open sourced it under the
Speaker:name Ludwig, and other engineering teams kind of around the world found it very useful
Speaker:as well. And what it really allowed anyone to do was be able to set
Speaker:up their entire end to end ML pipelines in just a few lines of
Speaker:configuration. So if you think about what infrastructure as code did
Speaker:for, you know, software development, similar idea, but
Speaker:brought to machine learning. You're able to start really easily, But then
Speaker:customize as you need, and Protabase really is kind of, you know, taking that
Speaker:same core concept and burning the, enterprise platform around
Speaker:it. So any engineering team that wants to work with open source AI and open
Speaker:source LMS as an example, can use our platform to easily and
Speaker:declaratively fine tune those models and then serve those directly
Speaker:inside of their cloud. And that's, you know, large part of what we do
Speaker:today. Interesting. Interesting. So
Speaker:What what does that what does that look like? Like, we
Speaker:know kind of generally what a a typical project looks like in terms of this,
Speaker:right, like, how does this interface with because I think it was the 1 question
Speaker:that I wish I'd asked, on the previous show. How does it
Speaker:interface with something like data engineering? Right? Yeah.
Speaker:We're I mean, we're, there's always gonna be rough spots. Right? So I'm not giving
Speaker:you a hard time, but there's always gonna be sharp edges when you're handling, Any
Speaker:kind of technology. Right? You've obviously kind of figured out the middle
Speaker:part, but, like, what does that look like in terms of the interface to data
Speaker:engineering? Is that what's What's that look like?
Speaker:Yeah. I'll insert in 2 parts. 1 of them is what does the user journey
Speaker:look like? And then what's the intersection with data engineering? So in
Speaker:the platform today, users do 3 things. The first thing they do is they connect
Speaker:the data source. This could be a structured data warehouse like a Snowflake, a
Speaker:BigQuery, Redshift, or unstructured object storage just directly files in
Speaker:s three. The second thing they do then is they declaratively
Speaker:train these models. What that looks like is they more or less fill out a
Speaker:template, you can think of it, just like a YAML configuration that says this
Speaker:is the type of training job I want. The beauty is the template makes it
Speaker:very easy for them to get started, but they can customize and configure as much
Speaker:as they want down to the level of code. They can build and train as
Speaker:many models as they want. And finally, after they've trained a model they're happy with,
Speaker:they get to the 3rd step, which is they can serve and deploy that model,
Speaker:make it available behind an API so any applications can start to ping it.
Speaker:So that's what the user journey really looks like in CrediBase, and how does this
Speaker:intersect with data engineering? So as you've probably heard before, like, you know,
Speaker:Machine Learning is really In large part, really about the data that you're
Speaker:using and like the quality of the data that you're using. Data
Speaker:engineering comes in 2 places. The first is you need to get all
Speaker:of your data wrangled across multiple different sources to be able to live in
Speaker:one area that you can connect as an upstream source and.
Speaker:This is the snowflake example, you know, of like getting that into a table.
Speaker:And that piece of the journey lives outside of Firebase. That lives
Speaker:as a step before you essentially connected into your system. But then
Speaker:there's the 2nd step that often happens, which we call data cleaning.
Speaker:So you've gotten your table, but, you know, all of your text is in,
Speaker:let's say lowercases and upper cases, you know, you have
Speaker:Really weird variable lens. You haven't normalized numerical
Speaker:data. Maybe you have images and things aren't actually, you know, resized
Speaker:to to scale. All of those data cleaning
Speaker:techniques, we have packaged in as pre processing modules
Speaker:inside of prediabase. And so what the declarative interface
Speaker:allows you to do is train a full machine learning pipeline from data
Speaker:to pre processing, through model training, through post processing and
Speaker:deployment. And so once you've gotten your data wrangled into a
Speaker:form, prediabase can come take in, help you clean out that data, and
Speaker:then be able to train a model against Interesting. Because it's that that
Speaker:preprocessing that, you know, the the the nightmare is, you
Speaker:know, this canonical example is address, you know, 123
Speaker:Main Street freight is an s t. Exactly. Right? That is not a lot of
Speaker:fun for anyone. And then obviously the the the
Speaker:lowercase uppercase thing like that becomes an issue too.
Speaker:So what is the what is the what's the user experience look like? Right?
Speaker:Like, is it is it drag and drop? It's declarative?
Speaker:Yeah. What what what does that look like? Like, what, you know, you mentioned user
Speaker:journey, and I love that term. But like, what does that look like
Speaker:from, from a practitioner's point
Speaker:of view. Right? Like Definitely. Now the first thing I'll say
Speaker:is, you know, our obviously underlying project is open source. You can check it out
Speaker:in Ludwig AI, and you can even try out, you know, our full UI for
Speaker:free on productbase.com. So if any part of this is a little too high
Speaker:level, you can actually get in involved For free, like immediately. But
Speaker:the user experience really looks like 2 ways. We have a UI
Speaker:that's really built around our configuration Language. And our
Speaker:configuration language is just a small amount of YAML.
Speaker:So your very first basic model can get started in just 6 lines.
Speaker:What those 6 lines do, and they, they say, these are the inputs I
Speaker:want. So you pass it, you know, what is the,
Speaker:column that is, you know, that contains the text you're predicting from. And
Speaker:then the output is what is your, what is it that you're trying to predict?
Speaker:So for example, my input is A sentence and
Speaker:my output is, the intent. So I'm trying to do intent
Speaker:classification with that model. And that's all user defines and
Speaker:they can do this programmatically in our SDK or there's like a drag and
Speaker:drop UI where they can build these components out together. The part that I
Speaker:think is really interesting just based on my experience working on other automated machine
Speaker:learning, you know, tools before no code UIs for ML is
Speaker:that ML really is a last mile problem. And so you have this weird
Speaker:complexity where you need to make it easier to get started, But a
Speaker:lot of the actual value ends up being in the last 5 or 10% where
Speaker:you customize some part of that model pipeline to get to work for your system.
Speaker:And so what credit what this configuration language, you know, does is sometimes I
Speaker:describe it as it builds you like a pre fat house. It gives you something
Speaker:like out of the box That like works end to end, and then you can
Speaker:just change the little bit of the pipeline that you want declaratively,
Speaker:which means in a single line. So you could say something like, you know, I
Speaker:want the windows of the house to be blue or, you know, I wanna change
Speaker:my pre processing of the text feature to lowercase all the letters, And then you
Speaker:can change leave everything else up to the system.
Speaker:We you we allow you to control what you want, and you just automate the
Speaker:rest. Interesting. Okay. So then it's kind
Speaker:of, the middle part of the the journey. Right? Like the
Speaker:Yeah. Is what this on so How does this relate? Because you
Speaker:said, you know, and I, you said automated ML. How much of this
Speaker:is automated? I mean, like, what? Because that was 1 what I had just assumed
Speaker:that I because I know I've heard of Ludwig as kinda like this automated ML.
Speaker:And when I say automated ML, I mean, You know, for lack of a
Speaker:better term, you know, here, there's a problem we're trying to solve.
Speaker:Computer, you figure out, you throw as much spaghetti at the wall and then figure
Speaker:out which model is the best, Right. Yeah. Is is that
Speaker:kind of the same thing here where I just say I wanna predict this and
Speaker:then the underlying models and methods are kind of automatically figured
Speaker:out? You know, I think that, that is an approach
Speaker:that a lot of folks have tried with AutoML v one, as I kind of
Speaker:often think about it. I actually was a PM on Vertex AI where we rolled
Speaker:out our auto non product as well. And the main issue we run into
Speaker:it is, you know, in deep learning, especially
Speaker:the search space is Too big to be able to run an effective
Speaker:hyperparameter search over all the different architectures and sub parameters you
Speaker:might wanna be able to use. It sounds computationally expensive. Right? I mean,
Speaker:it's Potentially prohibitive, really, in order to be able to say, you
Speaker:know, I want let's imagine you are, You know, in the modern world,
Speaker:building a model to be able to build, let's say, content moderation
Speaker:systems. How do you know which pre trained, like, should use a LAMA
Speaker:To a Bertha, De Bertha, like all of these models themselves are quite expensive
Speaker:to go to train and fine tune, and each of them have their own sub
Speaker:parameters. And And so I think it becomes computationally prohibitive to run an
Speaker:exhaustive grid search for your individual, types of,
Speaker:individual types of use cases. And so what a lot of AutoML systems did
Speaker:was they kind of just said, well, we know better than the user, so
Speaker:we'll just make some selections, Right. And then, and the we'll
Speaker:make it as easy and simple as you for the user as possible. So user
Speaker:just provides a few inputs, we give them a model, boom, they'll be happy. And,
Speaker:you know, I was actually I was, a PM for Kaggle. I was the 1st
Speaker:product manager at Kaggle, a data science and machine learning community that grew to about
Speaker:14,000,000 users Today, where we see a lot of citizen data scientists, and we rolled
Speaker:out AutoML in that community as well. And we saw
Speaker:a spike in usage And then extremely heavy churn
Speaker:as soon as we, like, rolled it out. And if you interviewed those users, the
Speaker:main reason why was because they didn't have any controller agency over that
Speaker:So the like, it would essentially spit out a model
Speaker:and say, here you go. You know, be happy. Go ahead and put this into
Speaker:production. But like I was saying previously, ML is a last mile problem,
Speaker:and no one is going to be comfortable using something they see as a dead
Speaker:end, And that's where I think about, you know, our approach really kind
Speaker:of, differing. And so inside of Premedbase, you can
Speaker:actually, you kind of get that, AutoML like
Speaker:Capability, where you're able to
Speaker:build a model just by saying, you know, here's the inputs, the model I
Speaker:wanna fine tune, And we will go ahead and get you the entire end to
Speaker:end model. But if you want to edit anything, for example, you want to
Speaker:edit, you know, the way we pre process the data and the At sequence
Speaker:length, you can go ahead and do it for any part of the model pipeline
Speaker:and just kind of like 1 single statement. And that's kind of like a
Speaker:large part of, you know, how we think about making it both easy to get
Speaker:started, but also, like, flexible where it's not just a
Speaker:toy, something you can actually use. Right. Because like,
Speaker:you know, my first experience with AutoML was the,
Speaker:was Microsoft's, offering. Right? And it
Speaker:was only it was very to get around the computationally prohibitive
Speaker:parts, they they narrow the problem set you could do that on. Right? So it
Speaker:was basically No neural networks. This was before chat
Speaker:c p t, before l l m's were, I wouldn't say a
Speaker:thing, but before they were a major, Point of views.
Speaker:But, you know, so it it cons it was constrained. Right? So it would just
Speaker:basically just Throw a bunch of problems and
Speaker:then kinda test it out, which Yeah. I I think what you refer
Speaker:to as, you know, AutoML v one. I think,
Speaker:The world has evolved, and it's interesting to see how that goes. And,
Speaker:the tooling looks really cool, actually. The,
Speaker:for those for those who are listening to this as opposed to watching this, I
Speaker:will make sure we we post that little snippet there. But
Speaker:but, you know, like, what And you were at
Speaker:Kaggle. Right? So Kaggle is kind of a big deal. What
Speaker:I think that's really cool. Looking at your resume, it's very impressive, actually. You
Speaker:you word Google, that would explain your interaction with
Speaker:Vertex, and things like that. So so what
Speaker:What what niche does this address or what need does this address that the existing
Speaker:market didn't address? Right? And like what Yeah. Because I think that's really, I
Speaker:think, where the rubber meets the road, particularly with an open I'm a big fan
Speaker:of open source too. So,
Speaker:Yeah. Well, let me start off by saying that, you know,
Speaker:I I think that the need has actually been unfilled in the market For a
Speaker:while, but there is also a fundamental technology shift, and I'm gonna talk about both
Speaker:of those pieces. So when I say the need was unfilled for a
Speaker:while, Yeah. I was a product manager on Vertex AI. I was a
Speaker:product manager on Google research teams, productionizing machine learning, and we've hired
Speaker:a number of folks Now that work does ML engineers across different companies. And I
Speaker:remember when one of our ML engineers joined the team, he told me, Dev, I've
Speaker:worked at 3 different companies doing machine learning for 3 different teams.
Speaker:Everybody does it differently, and I think the truth is, you know, for
Speaker:developers, there never really was like a de facto stack of here's how you do
Speaker:an ML problem. Pure data engineer. There is like a stack of, you know, what
Speaker:are the best practices for being able to get there's obviously a lot of variation.
Speaker:But there's like Some best practices of, you know, what you're using for your
Speaker:ETL pipelines, how you're thinking about being able to put things into data
Speaker:warehouses, what your stack is for being able to query and downstream.
Speaker:But in machine learning, it really looked like the wild west. Everyone was working
Speaker:across different types of projects. And I think a lot of companies
Speaker:tried to tackle that need, but unsuccessfully. And the
Speaker:fundamental technology shift that I think actually changed was exactly what you were
Speaker:talking about, Which was like you said that the old school version of Azure
Speaker:was not really any deep learning, maybe because it was computationally expensive for
Speaker:others. To be clear, the auto the automated ML part of it. I don't
Speaker:wanna get a lot of hate mail, but yes. Sorry. Sorry to sorry to interrupt
Speaker:you. Go ahead. No, no worries. I'm sorry to hijack the screen again,
Speaker:but, like, you know That was awesome. I think this just the way that I
Speaker:think about, like, the the change that's happened in industry is
Speaker:Machine learning 2 decades ago or even, like, 6, 7 years
Speaker:ago looked very different than what it is today. And I
Speaker:think that a lot of the hype around the LLM revolution is gonna actually
Speaker:translate and be realized as just the hype of pre trained deep learning models.
Speaker:Now, if we talk about ML 10 years ago, it basically looked like
Speaker:predictive analytics. So people were doing things like I'm going to predict the price of
Speaker:a house, And the way I'm gonna predict it is I'm gonna multiply the square
Speaker:footage of the house by some number and add in the number of bedrooms, and
Speaker:then figure out the coefficients based on my historical data. Really
Speaker:structured data tasks, regressions and classifications and others.
Speaker:But about 7 years ago, I think the really interesting pieces came out
Speaker:with pre trained deep learning models with Bert using the transformer architecture,
Speaker:the few image models even prior to that, that I think made it possible to
Speaker:do 2 things. The first is you could start with larger amounts of
Speaker:unstructured data. So now you didn't have to just work on these kind of more
Speaker:boring predictive analytics, numerical only tasks, but you could work with text,
Speaker:images, and others. And the second thing is you could start to actually use
Speaker:them pre trained, so you didn't have to have as much data before you start
Speaker:to get value out of it today. And what I think OpenAI showed was,
Speaker:okay, if I scale these same types of models up by 2 or 3 orders
Speaker:of magnitude, now people can use it with virtually no data whatsoever,
Speaker:and I can actually prompt and response, you know, it directly.
Speaker:But the underlying technology shift actually, I think is a shift towards
Speaker:just pre trained deep learning models. And the truth is, as we get away from
Speaker:some of this type of, like, the really cool conversational interfaces and we get to,
Speaker:like, how do these models drive value inside of organizations, I think that
Speaker:that's the emergent need for platforms like Predabase, which is how do I take
Speaker:any of these deep learning models and then customize them for what I actually need
Speaker:inside So fine tune and tailor it to my data, and then get
Speaker:it deployed inside of my organization for Cerven. Yeah. That makes a
Speaker:lot of sense. I think I think the
Speaker:The need for training something from the ground up, I
Speaker:think is overrated for most applications. Right?
Speaker:Why teach and model all the intricacies of the human
Speaker:language when that is already done, and you could take it
Speaker:from kind of a you You know, the example would be, like, if I owned
Speaker:a store. Right? And I needed someone to work the cashier.
Speaker:Right? I could have another child, Raise that child, change
Speaker:his diapers, send it to kindergarten, teach it to learn, read, and write.
Speaker:And in about 10 years, depending on labor laws, let's say
Speaker:15 years. I'll have someone who can work that cashier,
Speaker:plus however much it costs. Now, obviously, I'm not comparing a child to an l
Speaker:m, But I mean or you could just find an existing person
Speaker:out there, and say, here's how my register
Speaker:system works. This is the nature of the job, And I can kinda start from
Speaker:there as opposed to start from 0. You start from the 50th floor as opposed
Speaker:to start from the basement. That's exactly
Speaker:right. Yeah. I often think about, you know, these,
Speaker:pre trained LMS is like, well, what if I had like an army of
Speaker:like Cumulative high school students, you know, in high school, you study all the
Speaker:general subjects that kind of like a at a broad level. Right? So you know
Speaker:a little bit about history, a little bit about how to write, a little bit
Speaker:about how to You're not really an expert on any of those? Well,
Speaker:the really interesting thing becomes then how you do, like, the vocational training or kind
Speaker:of, like, you know, the task specific fine tuning It's how we think about it
Speaker:in ML parlance. And, I think that's where the cool opportunities get
Speaker:unlocked. It's really amazing to see the fact that you can scale up to, you
Speaker:know, as many intelligent agents If you want, but then you need to, our
Speaker:favorite customer quote is generalised intelligence is great, but I don't need
Speaker:my point of sale system to recite French poetry. Right. So it's great that
Speaker:you can go ahead and, recite history and others, but, like, how do you do
Speaker:something very individual is what our platform is, oriented on.
Speaker:No. That's that's a good point. That's that's a good point. Like, I I often
Speaker:say, like, you know, do you want your cardiologist to be
Speaker:also be a CPA, Or do you want them
Speaker:to be a good cardiologist? I know if I were under an operation, I'd
Speaker:probably wanna go with someone who was just all in on cardiology,
Speaker:You know? Yeah. But, And those are actually the
Speaker:2 trends I think we're gonna start to see with Gen AI, overall.
Speaker:I think, you know, one trend is going to be People are gonna start thinking
Speaker:of use cases that are more creative than just, you know,
Speaker:question answering chatbot. So, you know, I think, like,
Speaker:9 months ago, everyone I was talking to was like, I want chat g p
Speaker:g provider enterprise, and I'd say, okay, what does that mean to you? And they'd
Speaker:either shrug and say no idea or they would say like, you know, I wanna
Speaker:be able to ask a question about The truth is if you had this access
Speaker:to this, you know, army of agents that are like high school capable, I'm sure
Speaker:we can think of more interesting things. Just basic question answering.
Speaker:And then the 2nd big change I think is we aren't gonna use as much
Speaker:of these super general purpose APIs in production. They're the easiest way to
Speaker:experiment and get started. In production, you're gonna want your cardiologist to be the
Speaker:expert in medicine and you don't really care if they know how to change a
Speaker:tire or not. Exactly. That that is a a really good way to
Speaker:put it. And I think that, you know, people, we're
Speaker:still have to realize that we're still in the very early stage of this,
Speaker:For lack of better term revolution. Right? Like, you know, because you're right. Like, I
Speaker:talk to customers, and they say, we wanna we wanna get all all in on
Speaker:Gen AI. Okay. What are you gonna do? Well, we wanna chatbot.
Speaker:Okay. I don't know if you've seen
Speaker:this. I'm sorry. Go ahead. Oh, I was gonna say,
Speaker:And it's not not necessarily a bad starting point, but, you know, there there's so
Speaker:much more out there. Sorry. Well, no. I mean, exactly. Right? It's like, I want,
Speaker:if you could do anything in the world, what would you do? I don't know,
Speaker:take a day off, like, you know, but but that's you're missing the point, like,
Speaker:you're you are, there there's a meme going around. Again, I don't know
Speaker:if it's true, it's Screenshot where a, car
Speaker:dealership, had implemented some kind of chatty p t. You've
Speaker:seen this, you're nodding. Right? Where it basically sold a guy a car
Speaker:for a dollar, and basically, the person got it to
Speaker:say, no, this is a legally binding contract. Basically, Tricked the
Speaker:chatbot into saying no. Totally. No backsies, I think was the first phrase
Speaker:to use. Right? And he he got it to say things like, oh, no. Absolutely.
Speaker:I wanna make you a happy customer, And you can have this Chevy Tahoe for,
Speaker:like, $1 or something like that, but he and I I don't know
Speaker:how that's gonna play out in a court. Obviously, I imagine a
Speaker:dealership is gonna have some, lawyers look into that,
Speaker:and I'm not a lawyer, but I I can I can easily see like, you
Speaker:know, this is a great example of, To your point, do you really need your
Speaker:point of sale system, you know, re be able to recite
Speaker:French poetry? Right? Now, I guess if I were, You know,
Speaker:a very niche kind of bookstore slash
Speaker:coffee shop, maybe? But for the most part, no. Right? And
Speaker:and obviously, Yo. There I wouldn't classify that as a
Speaker:guardrail. I would say that more as a domain kind of boundary.
Speaker:But, you know, these chatbots are gonna need Guardrails too. Right? Not just the
Speaker:obvious things that we always hear about, you know, but also, you
Speaker:know, don't wanna be giving away. I haven't priced
Speaker:what a Tahoe cost, but I imagine it's much more than $1.
Speaker:Yeah. I bet too. Yeah. I think it's actually a function of 2 The first
Speaker:is we need some better infrastructure on guardrails of what models can and can't
Speaker:say. And actually, by the way, this is where fine tuning is actually very
Speaker:useful. It restricts, Like, it's one of the best ways to reduce hallucinations. It,
Speaker:like, teaches the model this is the type of thing that you're supposed to be
Speaker:outputting, but it's not bulletproof. And I think that
Speaker:actually the more, meaningful longer term conversation
Speaker:is if you believe, like, I believe, and I
Speaker:think a lot of folks, Yeah. About working this industry do that AI will
Speaker:become kind of a dominant aspect of most businesses
Speaker:over the next decade. That like the companies that embed
Speaker:AI are going to be the ones that survive and have differentiated value.
Speaker:The ones that don't are likely gonna be less competitive. If you believe
Speaker:that, it's also hard to imagine that you're going to defer all
Speaker:control of the model to a third party. And that's where
Speaker:things like, you know, It's one thing to say, like, we need the guardrails. It's
Speaker:another thing, like, if you realize that if those folks were using something
Speaker:like, you know, commercial API that's Behind a walled garden where you
Speaker:don't have access to the model, you don't have access to the model weights. They're
Speaker:kind of limited in what they actually can do. They can post process the
Speaker:output of the results, but they can never really get that fine granular
Speaker:level of control. And that's why we think the future is gonna be open source.
Speaker:Because ultimately, people are going to wanna own those models, own the outcomes
Speaker:of the part of the IP that they think is gonna drive a lot of
Speaker:their enterprise value in the future. So our like, I would say our our
Speaker:bet as a company is really on 2 things like fine tuning and
Speaker:open source. And I think that, you know, the example you just gave is a
Speaker:good why I think the world is gonna have to move into both of
Speaker:those directions. No. That makes a lot of sense. I think that open
Speaker:source is important for a number of reasons. I mean,
Speaker:not the least of which is, you know, we we have seen recently that if
Speaker:if if these things are behind a commercial firewall,
Speaker:If, for instance, there was some kind of, I don't know, political shake
Speaker:up inside of said company board, which of course would never
Speaker:happen. Right? Never happened. Then
Speaker:you you are taking down that risk. Right? Which is, I think, is another
Speaker:reason why open source, just in Generally, an industry is is
Speaker:popular because decisions tend to be made at the community
Speaker:level. Right? Now, there's obviously flaws with that approach
Speaker:too, but It is, and I would use this as an example
Speaker:of if you look at HTML and JavaScript Yep. Versus
Speaker:say Flash and dare I say Silverlight. Right? Flash was
Speaker:always a proprietary product. Silverlight, if people remember it, was also a
Speaker:proprietary product, but HTML,
Speaker:JavaScript Had its flaws, but eventually, they did get their act together,
Speaker:and it it has a certain more
Speaker:implicit compatibility. And I think with AI, I think the
Speaker:it's not so much about compatibility. It's implicit transparency.
Speaker:You get with open source AI. Right. Is it perfect? Is it totally
Speaker:transparent? No. That that's not the point. But the
Speaker:point is you're starting at a much more Transparency almost
Speaker:by default or transparent, maybe translucent,
Speaker:as as as as a default as opposed to completely opaque.
Speaker:Yeah. I I think that it's both the transparency and the
Speaker:control that's critical. Yes. It's the fact that people do not only
Speaker:introspect and understand what's happening, but They can edit and change, you know,
Speaker:in instances. Even if you're like a lot of our models, users do not
Speaker:edit 99% of the pipeline, But it's important that they're
Speaker:able to edit all of it, and that they do make the edits to the
Speaker:1%. And I think that exists for open source. And I think from just like
Speaker:an industry macro standpoint, you know, Trying to fight open
Speaker:source and developer platforms is like trying to fight physics,
Speaker:basically. It's kind of against the natural working of those systems.
Speaker:And so our view is that, you know, people are
Speaker:gonna come out with amazing models. And some of them are gonna be commercial, and
Speaker:some of them are gonna be open source. The open source Size of the pie
Speaker:is going to grow, and I think you wanna see this here, right? Like it
Speaker:has caught up, so quickly. Like the
Speaker:open source attraction has caught up so quickly to everything else. Our
Speaker:view is just like, what do you need when you want to use open source?
Speaker:Well, you need the you need the infrastructure around it. You need to be able
Speaker:to plug it into proprietary, settings. You need to be able
Speaker:to create those guardrails around it. That's, you know, where we think about ParetoBase
Speaker:providing the info For being able to use open source. Interesting.
Speaker:Well, this is a fascinating conversation. We could probably go on for another hour or
Speaker:And I definitely would love to have you or someone else from Credit Base because
Speaker:I think, you know, it's just a cool idea. Right? Like it and
Speaker:and I think that it it really solves a missing piece of the puzzle
Speaker:In terms of making this, you know, when you say
Speaker:YAML, when I think YAML, I think OpenShift, right, obviously, you know, work at Red
Speaker:Hat, that's kinda, but I mean, I think that,
Speaker:it's one thing to open source the model. It's quite another to how do you
Speaker:manage and control that animal? Right. Because these are
Speaker:not these are not tiny little things. Right? These are
Speaker:potentially very compute intensive activities. Right. So you
Speaker:don't want you wanna be efficient. That's the way the world has gone.
Speaker:Right? It's more compute intensive and,
Speaker:heavier weight, and so that's where the infrastructure components become
Speaker:critical for any company that's actually gonna use it. Absolutely. And you have to at
Speaker:least If you can't be a 100% efficient because you really can't,
Speaker:but you wanna at least, prioritize towards compute efficient
Speaker:Activity. Because otherwise, you are literally throwing money out the
Speaker:door. And I think that it looks like
Speaker:your tool is really good at kind of Making it
Speaker:so it's compute efficient, like, or at least that that
Speaker:it goes a long way to helping that. I'm sure you can probably do some
Speaker:serious damage With any tool. Right? Like, I wouldn't give my my 2
Speaker:year old a chainsaw. You know what I mean?
Speaker:But, now that's interesting. So
Speaker:now we're gonna transition into the pre canned questions.
Speaker:How did you find your way into data Or AI. Like,
Speaker:did you find AI or did AI find you?
Speaker:That's an interesting question. I,
Speaker:I first got into it just out of studying
Speaker:computer science. You know, I when I went into university, I thought I
Speaker:wanted to study economics. Really liked, you know, the theory
Speaker:behind economics. I took a intro to computer science class because I thought it'd be
Speaker:interesting. And that more or less just completely shifted where I went
Speaker:because CS was actually magic. You know, economics is a great way to be
Speaker:able to explain things that were happening in the world, but with computer science, you
Speaker:could actually build systems. And that was really interesting.
Speaker:And then I found the 1 piece that I think I liked just as much,
Speaker:which was statistics. And the natural
Speaker:marriage of computer science Statistics really is, you know, data and data
Speaker:science. And so, I'd studied it for a while, and then
Speaker:when I went to, Yo. Go work in in a professional industry.
Speaker:I first started off as a PM at Google, and I worked at completely different
Speaker:things on Firebase, developer platform, authentication, security. I
Speaker:remember somebody saying like, you know, you have to work on what you're most passionate
Speaker:about. You know, a new college graduate, I have no idea what I'm passionate about
Speaker:professionally. And so I thought back to, you know, the things that I'd studied that
Speaker:I found the most interest in, that I found the most fun to work on.
Speaker:And it really was those data science projects, Honestly, starting with the early
Speaker:Kaggle competitions that I did in 2013, where you were trying
Speaker:to compete to see who could build the best housing prices model who could build
Speaker:the best recommender system model, and you had to exploit all
Speaker:these interesting nuances in data and models to be able to get there.
Speaker:And so I just found it so fun. And then
Speaker:I think after a little while, found it trading
Speaker:that everyone else didn't have sort of the same access to those types,
Speaker:those types of experiences and tools. And so that's where the experience really
Speaker:began. I would say, you know, early on, just having that academic
Speaker:background and then seeing the problems kind of being manifested in Google and
Speaker:eventually, you know, working as well on Kaggle of the data science and machine learning
Speaker:community there. Interesting. Interesting.
Speaker:I see you did a brief stint in cybersecurity for a while,
Speaker:Which is funny because I think people see that as a as a totally separate
Speaker:discipline, and in in a very real sense, there is. But I think that in
Speaker:a very real sense, A big chunk of cybersecurity is
Speaker:monitoring logs and input data and figuring out what's happening.
Speaker:Sounds at all sounds familiar. Doesn't it?
Speaker:I think cybersecurity, you know, when I was doing cybersecurity, work, it
Speaker:was very, very much in the early days, strategic, how to
Speaker:think about risk postures at an enterprise level. Right. But I think what's
Speaker:really interesting now is, cybersecurity and AR are gonna have
Speaker:a very interesting marriage where Cybersecurity is gonna be influenced
Speaker:by AI. For example, we work with 1 company today that does open source supply
Speaker:chain security, and they're looking at using LMS to read code and be able to
Speaker:do things like Identify vulnerabilities, advise on remits, and
Speaker:others. And so one obvious area is going to be that
Speaker:cybersecurity companies themselves are gonna get revolutionized with AI. But
Speaker:But this is gonna be one of the industries where there's kind of like the
Speaker:bidirectional era as well. AI is gonna need some cybersecurity
Speaker:best practices too. Yeah. These made these weights are now,
Speaker:open source. How do you think about whether or
Speaker:not the security governance Factors should be
Speaker:on the inputs, you know, when the data is fed into the model,
Speaker:in the model layer itself, like, how the model processes
Speaker:that data On the outputs. Like, what is the framework for thinking
Speaker:about, like, you know, which ones introduced what kind of risk? And the type of
Speaker:industry that's had the most experience in this historically has in the cybersecurity industry,
Speaker:Thinking about how we deploy software internally and others, and so that
Speaker:marriage is gonna be, I think, really interesting. I bet there's gonna be really best
Speaker:of breed companies in both worlds. I could totally see that.
Speaker:I think that's a very good cogent response to,
Speaker:you know, these are not isolated industries. Right. I mean, they
Speaker:obviously have different origin stories, but I I could
Speaker:totally see them merging. And to your point, right? I mean,
Speaker:Yeah. If you look at potentially 2
Speaker:things, right? 1, the, who, the amount of input
Speaker:data that you have, like, Could that be poisoned in a way that could produce
Speaker:negative effects later on in an LLM? And 2,
Speaker:We don't really know the sort of latent, for lack of better term, latent spaces
Speaker:that exist in these extremely large complicated,
Speaker:models like for I'm sure you've seen this, but there was a random
Speaker:string of characters that would produce bizarre output
Speaker:In chatty b t. And there was also one that would basically short circuit
Speaker:the, the safety rails inside of
Speaker:some of these LLMs too. And it was just like,
Speaker:wow. I mean, you know, was that the one, how was that figured out?
Speaker:Was that random, or did somebody kind of understand that there's Weird
Speaker:latent spaces and how to manipulate that. I think that is gonna
Speaker:be a new frontier opening up, in the
Speaker:not too distant future. If it hadn't already happened,
Speaker:honestly. Yeah. I agree. I agree. And I think
Speaker:it starts with understanding that, You know, those those
Speaker:bits of, I guess, entropy that feel random to us are,
Speaker:are more features oftentimes than bugs. So the fact that the random characters
Speaker:produce, like, a weird output, it's actually really interesting
Speaker:because what that means is maybe I don't need to type out a full
Speaker:English Paragraph to get this model to do what I want. You know, there's really
Speaker:cool things in prompt compression where people have basically been like, can I just
Speaker:say, like, a couple of characters AFD, something that would mean
Speaker:nothing to you and I, but the model understands that means, okay, go ahead and
Speaker:pick up the dry cleaning on the way home and then make sure that you've,
Speaker:you know, swung by and filled Like, essentially a set of instructions that get compressed
Speaker:into this model's internal representation? So I think we're barely
Speaker:scratching the surface of it, It's one of many ways that the I think,
Speaker:l m revolution is gonna be really interesting in the ways that we haven't fully
Speaker:explored yet. I could have said it better myself.
Speaker:Our next question, what's your favorite part of your current
Speaker:gig? My
Speaker:favorite part is Probably the part that's also, I think one of the most
Speaker:challenging is the space is moving so quickly. I know people
Speaker:say that frequently, but the truth is I've heard people say that about different
Speaker:technologies historically, and I'm like, yeah, it's moving faster than other
Speaker:things. You know, for example, Mobile moved quickly.
Speaker:There were over many years to transform things that happened.
Speaker:The Timescale that our world is kind of, dominated. I'm gonna
Speaker:say our world. I think it just mean, like, you know, the the AI movement
Speaker:so far over the last year It's it's in weeks. Right? Like, every
Speaker:few weeks, there's a new seminal groundbreaking, whether it's,
Speaker:Yeah. I I can think about the moments where, like, Llama got introduced as an
Speaker:open source model. Its weights got leaked. That was amazing because it spurred out of
Speaker:the whole new community. GPT 3.5 got upgraded to GPT
Speaker:4, new set of capabilities that came out there. LAMA 2 came out
Speaker:this year with commercially viable licenses and like, You know, really, I
Speaker:think, best in class performance up to the
Speaker:point that Mixed Straw came out, which was a, you know, mixture of experts
Speaker:model significantly smaller doing as well as chat g p t. This was only
Speaker:a few days after Google released Gemini, you know, their own, model.
Speaker:We have AWS in the race with Bedrock. It's kind of like, you know, an
Speaker:interplay between different providers. I'm saying a
Speaker:lot of sentences, but like the The really interesting piece of it is all that's
Speaker:really come out in the last 6 months, and I haven't even covered up, like,
Speaker:all the academic, you know, like It's wild. It's wild. Like, so I
Speaker:was on a cruise, like, we were talking in the virtual green room, and I
Speaker:had intermittent Internet, and I looked at my phone far more than I should,
Speaker:for being on vacation, but it was just like Gemini happened,
Speaker:AMD, and made some hardware announcements. And I know
Speaker:hardware In the the unintended
Speaker:consequence of being compute intensive is that hardware starts to matter again.
Speaker:Right? Yeah. There was if you were a software
Speaker:engineer, obviously, mobile, let's let's take that in the conversation.
Speaker:But if you were a software engineer building websites, hardware wasn't really a major
Speaker:Concern. Right? It was kind of pushed to the side. I mean, it
Speaker:mattered, when you got, like, your Amazon bill was through the roof
Speaker:and you weren't as efficient as you should be. But I mean, it wasn't really
Speaker:a major concern. Now we have let's say it's starting to be a limiting factor
Speaker:in terms of, you know, how many h one hundreds can you get your hands
Speaker:on. Right? It's it's,
Speaker:no. But, but you're right. Like, I mean, just I missed a week and I
Speaker:still feel like I'm catching up and that was like almost 2 weeks ago. So
Speaker:Yeah. And the, and that's the most exciting piece for us.
Speaker:Right? It's because, all this changes created a lot of opportunity. So
Speaker:We got a lot of popularity recently for something called Lorax.
Speaker:Mhmm. It's an open source project that we released that basically,
Speaker:was just a problem we had to solve for ourselves. It's the industry is moving
Speaker:quickly. We needed to allow people to fine tune and serve large language
Speaker:models for free in our trial. Now every single one of
Speaker:these l m's requires a GPU and sometimes bigger, heavier,
Speaker:meatier GPUs. And so if we're giving away a lot of free trials To, you
Speaker:know, people just on the Internet who are all using a GPU,
Speaker:investors would not be the happiest. And so we needed to figure out a better
Speaker:solution where we could actually serve Many, potentially hundreds of these
Speaker:large language models on the same individual GPU. And
Speaker:so we, we came out with a really cool technique to be able to do
Speaker:that. We called it Lorax for LoRa Exchange.
Speaker:And, we open sourced it and back a lot of popularity. One of the reasons
Speaker:that I think it got picked up in such a way was because it really
Speaker:kind of just fed into them kind of main, main thought process in the
Speaker:moment And everyone's staying up to date on kind of the latest. So, you know,
Speaker:it kind of fed nicely into that hardware constraint, area of the world
Speaker:as well as kind of a need that the market had. And so It's been
Speaker:really fun, I think, to just be on top of that. Very cool. Very cool.
Speaker:So we have 3 complete this sentence, questions. The
Speaker:first one is when I'm not working, I enjoy blank.
Speaker:I have a very San Francisco Answer to this question. But when I'm not
Speaker:working, I enjoy being outdoors. And in
Speaker:particular, I really enjoy biking, taking a road bike and going up a mountain,
Speaker:because the reward at the end of that's amazing. And playing tennis, those are
Speaker:probably the 2 things that, you know, I I enjoy the most. Very
Speaker:cool. The San Francisco is perfect for that sort of thing, like the bikes in
Speaker:the mountains, in the ocean. It's gorgeous. Yeah. Yeah. It's
Speaker:gorgeous. I think the coolest thing about
Speaker:technology the coolest thing in technology today is blank.
Speaker:The accessibility. I think the coolest thing about technology today is the fact
Speaker:that I can go ahead and run GPT four
Speaker:Or llama 270,000,000,000, the commercial variants of, you
Speaker:know, the leading edge or the open source variant. I can run both
Speaker:of them More or less for free, at least to try out
Speaker:for, like, you know, a little while. And that's sort of the same thing that,
Speaker:you know, big bank over here is gonna be using Or, you know,
Speaker:leaving technology company over there. Now, at least as the starting
Speaker:point where it starts to diverge is like how, when you get heavier into the
Speaker:customization and others. The coolest thing about technology to me is
Speaker:in, and again, I think of it very much from like an AI centric lens,
Speaker:just given my day to day. But, it's the fact
Speaker:that, you know, I, the graduate students, you
Speaker:know, somebody abroad in a different country, And then you know the m
Speaker:l engineer at a company like Netflix, all have some shared experience
Speaker:of language based on technology that just came out this year
Speaker:Because the barriers to entry are not significantly high to be able to get
Speaker:started. Now, I think the barriers to entry are still too high to, you know,
Speaker:go from prototype to production. That's what we wanna be able to lower, but that's
Speaker:to me the most compelling thing that we've done. That's very cool.
Speaker:The 3rd and final Is I look forward to the day when I can use
Speaker:technology to blank.
Speaker:That's a good question. I think I look forward to the day,
Speaker:when I can use technology to, to be sort
Speaker:of like the Adviser and whiteboarding
Speaker:buddy, if that makes sense. So if you think about,
Speaker:like, what you often do with an advisor, it's, It's
Speaker:actually generative in a lot of ways. You'll walk through them with a problem.
Speaker:I do this with my dad all the time. And so, you know, he and
Speaker:I will talk through Some challenge that I'm thinking about at work
Speaker:or or something else. And he doesn't have all the context, you know, that that
Speaker:might, but he's able to apply these like general frameworks and come up
Speaker:with a few different types of suggestions based on based
Speaker:on that. And some of them, because he's coming from a very different place, Might
Speaker:be different than the way that I thought about it. And I
Speaker:actually see that as a capability for,
Speaker:For technology that as we've come up with it as well is to be, you
Speaker:know, you've actually seen like companionship apps in terms of like, you know,
Speaker:psychological help or behavioral help or, or Or just having someone to
Speaker:talk to is actually like a use case that these models have already
Speaker:started to pick up on, within like a niche group of users. And what I
Speaker:think would be interesting is, you know, if you think about what you probably lean
Speaker:on friends or family and other types of things for, I
Speaker:think should still be friends and family and others. They are the ones who know
Speaker:you best, but the model can be like one additional source of that
Speaker:input. And it's gonna be really cool when, like, you know,
Speaker:if you're if you're working through something hard and you wanna go ahead and, you
Speaker:know, you get, like, get a few ideas for how to be able to go
Speaker:through it, You can text your family group, you can text your friend group, and
Speaker:you can ask the model that knows you, and you can kind of pick the
Speaker:best idea amongst those 3. That's a great idea. I think that, a
Speaker:lot of the media hype around things like replica AI and things like that has
Speaker:been like, oh my god, it's gonna replace human interaction. And it's like, Are
Speaker:they intentionally missing the point, or is it clickbait? Like, I can't tell.
Speaker:Right? Are they are they are they clue are they clueless by default, or are
Speaker:they clueless to make money? Not really sure. But I think that you're right.
Speaker:It's meant to augment. Right? And I think that's a very healthy way to look
Speaker:at it too, you know. Because I if I get stuck writing something. Right? Like,
Speaker:I'll I'll ask chat TBD. Like, hey, how would you word this?
Speaker:Right? Sometimes it comes up with a good answer, but at least it it kinda
Speaker:clears the log jam in my head Where I'm like, oh, okay. Let me let
Speaker:me go around it this way. I think that's a, I think that's an
Speaker:underrated use for AI or these LLMs.
Speaker:Yeah. I totally agree. Share something different about
Speaker:yourself. We always joke, like, you know,
Speaker:remember it's a It's a it's a family, iTunes
Speaker:clean rated podcast. Something different about
Speaker:myself. Yeah. I don't know if it's different or at least something that,
Speaker:Not a lot of folks know about me, like, when I, first, first got
Speaker:with them, but, I'm a 1st generation immigrant, and as is, like, my entire
Speaker:family. So I was actually born, in India, came over, you know, when I was
Speaker:a lot younger. So that I think is interesting because
Speaker:I was both that, but also grew up right here in the Bay
Speaker:Area. You know, I I think very much saw, like, the tech
Speaker:I I think very much saw 2 things. One of them was just the US
Speaker:kind of as, corollary and adjacency to to India
Speaker:where, like, parents had spent the vast majority of their lives and, you
Speaker:know, where we had come from. And then the second was like a very specific
Speaker:part of the US with Silicon Valley that was just, had a
Speaker:very interesting culture, Some healthy disregard for the
Speaker:rules in some regard, not always for the best, but sometimes for the best.
Speaker:And a real kind of inclination towards, you know, moving very quickly and kind of
Speaker:being on the latest since and and and Barry progressed in that way. And
Speaker:so I think that, This might be a little bit more of a backstory
Speaker:than an interesting individual facts, but I do think that, you know, that,
Speaker:immigration To especially this area, I think
Speaker:was kind of a very, at least different experience than what
Speaker:I think a lot of other folks that I've talked to have. Yeah. I often
Speaker:wonder what it would be like to grow up in the Bay Area, and I've
Speaker:met some people through through work and things like that who did. And they're like
Speaker:It's hard because if you if it's if you grew up there, it's kinda all
Speaker:you know, so you don't really have a good Yeah. Benchmark. Like, I grew up
Speaker:in New York City, and people are like, oh my god. How could you grow
Speaker:up there? I'm like, I don't know. It was just So I I
Speaker:grew up in the Bay Area and then went to school in the northeast and,
Speaker:you know, there's some things you realize, definitely. One of them
Speaker:is, Yeah. Fewer people wear, like, hoodies and, you know, flip flops,
Speaker:boat shoes are more of a thing. Like, there's all sorts of funny changes,
Speaker:You know, that exists culturally, especially. I think the
Speaker:biggest things that I've kind of picked up on is, like,
Speaker:The Bay Area has a very kind of, or at least I think where,
Speaker:the environment I grew up in, a very like, risk forward culture. It's kind
Speaker:of a why not, worst thing happens. Whereas I feel like a lot of other
Speaker:areas are a little bit more steeped in tradition And views
Speaker:that as a good thing. I think the Bay area
Speaker:potentially, and not to say one is right or wrong, but I think the Bay
Speaker:area has a bit more of a culture, A healthy disregard
Speaker:for tradition. And, you know, I
Speaker:think, Sofia had the great quote about tradition,
Speaker:That I'm forgetting. But, like it's,
Speaker:yeah, I think it's one thing that I definitely think about, especially the difference between,
Speaker:like, For example, where I grew up in the northeast, where I spent some time.
Speaker:Right. Right. And you were I'm I'm inferring because you went to Harvard that you
Speaker:were in Boston, and Boston is kind of its own Yeah. Its own corner
Speaker:of the northeast. If you ask somebody, like, you
Speaker:know, if you ask, I've lived in Europe, I've lived
Speaker:in, in new in
Speaker:New York and now the DC kind of Richmond, now
Speaker:Baltimore. There are slight variations in culture, but like, I
Speaker:can only imagine like how much of a shock it would have been from like
Speaker:the bay area To, like, Boston, especially.
Speaker:Right? Where it's it's far more I think things are far more rooted in tradition
Speaker:there. Right? Yeah. And it's it's not a knock on it. Right? Like, I I
Speaker:will knock on their baseball team, but that's another another story. Right?
Speaker:But, you know, but still, the both I mean, the
Speaker:the Boston area is also known for its innovation in both
Speaker:biotech and technology. Right? So it's not, These are not mutually exclusive
Speaker:things. Right? They're just different approaches.
Speaker:Absolutely. And both of them have worked, you know, really well for those respective
Speaker:Areas. One of them feels a lot more at home to
Speaker:me. But I think, you know, it was fun and interesting to kind of see
Speaker:those 2 differences, Especially spending time in both cities.
Speaker:Yeah. That's cool. That gives you a unique perspective on, you know, that the
Speaker:US culture is not one monolith, it's just Fragments of
Speaker:different things. It's it's an interesting perspective. I almost
Speaker:have to ask, like, was it as much of a culture shock coming to the
Speaker:US or coming from the Bay Area? Well, honestly, the Bay Area to
Speaker:anywhere else. Right? You know, the weird thing
Speaker:is I didn't expect the culture shock to I expected the culture shock coming to
Speaker:the US. Both from you, but you know, I was young, especially for my family.
Speaker:Yeah. I think that was there, but you're kind of, you're expecting
Speaker:it. And so it's always something that you're well prepared for. I don't think I
Speaker:expected the culture shock going from the Bay Area to to Boston.
Speaker:Because these are the 2 cities in the US. These are 2, you know, Progressive
Speaker:cities that are well educated in the United States, how different can they be.
Speaker:And you don't actually notice the difference, I think on a one day or two
Speaker:day visit, you kinda notice the difference when you actually spend a longer period of
Speaker:time there and understand the undercurrent. So Yeah. It
Speaker:wasn't a shock actually as much as it it was kinda cool. Like, I appreciated
Speaker:that 2 places in the US could actually feel very different because,
Speaker:you know, diversity is the spice of life. So actually really, really, I liked
Speaker:it even though it was different to maybe how I thought. That's cool. That's
Speaker:cool. The winter must have been a good shock on you. The
Speaker:winter was a shock in less of a positive way. Yeah. Diversity is a spice
Speaker:of life minus in weather. Yeah. I'll say
Speaker:70 degrees sunny year round all day. Were you there during the year? They
Speaker:had, like, a record amount of snowfall, like, something like Yeah. Fifteen
Speaker:feet over the winter? I was. Yeah. Yeah. Exactly. Yeah.
Speaker:Yeah. Campus shut down. Yeah. I was a student then,
Speaker:and, You know, as I was saying, very healthy risk
Speaker:appetite. I think everyone was out in the yard, like, throwing snowballs at each
Speaker:other while there was, like, a record blizzard So it was, it was
Speaker:fun. It was less fun when the snow was still on the ground in Maine,
Speaker:June. That was when I was thinking, get out of here.
Speaker:Do you listen to audiobooks at all? Yes. I
Speaker:I read more often, but sometimes I do re I listen to audiobooks to conveniently
Speaker:Do you have any Recommendations?
Speaker:I really like The Happiness Advantage by Shawn Achor.
Speaker:It's yeah. It's a book about how,
Speaker:I think there's a thought process that, you know, like, success breeds happiness,
Speaker:but this is also, like, work by a behavioral psychologist. Like how happiness can breed
Speaker:success and just how to be able to be in that mindset more often. And,
Speaker:you know, it's a weird book because it's actually kind of style as a business
Speaker:book. But I actually think it's a lot about like personal development. And
Speaker:so, yeah, that's definitely one I'd recommend.
Speaker:Cool. Audible is a sponsor of the show. And if you go to the data
Speaker:driven book .com, you will get, 1 free book on us. And,
Speaker:if you sign up for a subscription, You get a we
Speaker:get a you get a subscription and of knowledge, and we get a little bit
Speaker:of a kickback for them being a sponsor. And
Speaker:finally, where can people learn more about you and Predabase?
Speaker:Yeah. Absolutely. So, the obvious and easiest answer there is of
Speaker:course prediabase.com. I think, you know, we've learned,
Speaker:the easiest way to learn more is just to go ahead and try it.
Speaker:And so you'll see things there like documentation, you'll see a bunch of
Speaker:videos on our, blog page, which are short, 3 to 5
Speaker:minutes, and our YouTube channel, on prediabase, p
Speaker:r e d I b s e, actually has longer form 1 hour pieces of
Speaker:content that are more educational. But I'm a big believer that the
Speaker:easiest way to actually learn is just to be able to get your hands dirty.
Speaker:So if you click that try for free button, you'll get a few weeks, and,
Speaker:you know, credits. We'll give you the GPU out of the box so you can
Speaker:run all these models yourself, and you can learn firsthand. That's usually the easiest
Speaker:way, you know, to be able to get Started more. And then if you wanna
Speaker:learn a little bit more about our underlying technology, we've open sourced
Speaker:both of the key components. So for how to train models, we have Ludwig,
Speaker:And then for how to be able to serve models, we have LAURACS. And
Speaker:so those are the 2 l's that you can kind of use in order to
Speaker:be able to understand how the tech works under the hood. Very cool.
Speaker:Thanks for joining us in the show, and thank you once again for your, patience
Speaker:as we work through some scheduling conf conflicts,
Speaker:And, I'm glad we had this conversation. You're always welcome back in the
Speaker:show, and I'll let the nice British AI lady finish the show.
Speaker:Thanks, Frank, and thanks, Dev. What a
Speaker:splendid conversation that was. It felt like
Speaker:navigating through a maze of data with only the smartest chaps as my
Speaker:guides. To our listeners, I hope your brains are
Speaker:buzzing with as much excitement as mine is metaphorically speaking,
Speaker:of course, since my excitement is more of a series of well organized
Speaker:algorithms. To our dear listeners, if today's chat
Speaker:has ignited a spark of curiosity t in you, then I dare say we've
Speaker:done our job. Remember, the world of AI is vast
Speaker:and ever evolving, and it's thinkers and doers like deaf who keep the digital
Speaker:wheels Turning. Before we sign off, a gentle
Speaker:reminder to keep your minds open and your data secure.
Speaker:Until then, be sure to like, share, and subscribe as the
Speaker:kids say these days.