Joining us again today on the Data Driven Podcast is Christopher Newland,
Speaker:technical marketing manager at Red Hat Conference. Veteran
Speaker:and a man whose travel itinerary is only slightly less complicated than
Speaker:a Kubernetes deployment. Christopher brings a sharp, data
Speaker:informed perspective on the future of AI, drawing from his research
Speaker:into simulating reality, continuous learning models, and why
Speaker:we may not need humanoid robots to build superintelligence. Just a
Speaker:really convincing version of Grand Theft auto. From Google
Speaker:DeepMind's alpha projects to the metaphysical quandaries of I
Speaker:robot, Chris takes us on a tour through the bleeding edge of AI,
Speaker:where machine learning meets science fiction and simulation might just be
Speaker:the next reality. Hello and
Speaker:welcome back to Frank's World tv. Streaming live
Speaker:from both Boston and Baltimore. We're hitting the B
Speaker:cities today. My name is Frank Lavinia. You can catch me
Speaker:at the following URLs and with me today is
Speaker:Christopher Dulin, my colleague at Red Hat, who is also
Speaker:technical marketing manager here. And
Speaker:you've actually not traveled around the world since we last
Speaker:spoke. I think you've mostly stayed inside the.
Speaker:Continental U.S. yeah, it's been nice.
Speaker:I think that's pretty typical of
Speaker:late July, August, because Europe pretty much shuts down and then.
Speaker:Right. The conference season in the United States kind of goes
Speaker:away when people are doing summer vacations and I think we're just
Speaker:now starting things pick up. I'll be in Europe for a
Speaker:variety of events. So if you keep an eye on the
Speaker:Vllm community and the Vllm meetups,
Speaker:I have events in Paris, Frankfurt and
Speaker:London in November that I'll be at. So if you
Speaker:are in the,
Speaker:in Europe, in one of those areas, definitely come. You know, it's one of
Speaker:these events. I'll be there and then we'll also have some pretty cool speakers
Speaker:there as well. So I have most, I have Europe, but then I
Speaker:have some big conferences too like Kubecon and Pytorch Con coming
Speaker:up. So if there's anyone on the stream in North America going to
Speaker:those conferences, hit me up because I will be there. I'm
Speaker:doing a couple of media events as well as a few
Speaker:talks in the community sections for both of those.
Speaker:So excited to be there, excited to be involved
Speaker:and yeah, should be. Should be. Good. Cool. So
Speaker:I. To your left and up
Speaker:there should be a QR code that shows Vll meetup. So I'm going to make
Speaker:sure that the QR code actually works. Good. Yep. Let's
Speaker:see. Yep, it looks like it did work. Cool.
Speaker:Not that I didn't have any faith in restreams ability to do that. But
Speaker:yeah, there's a lot of VLM meetups. There's a lot of good,
Speaker:good stuff going on here. There's one tonight
Speaker:actually. I'm actually going to be leaving this stream to go. I got my
Speaker:VLM shirt on and I'm actually heading over to
Speaker:a venue in Boston or we're doing a VLN meetup actually here tonight, which
Speaker:I'm really excited. Oh, very cool, Very cool. It's nice to have one at home.
Speaker:I have a very busy week with events, but it just worked out to have
Speaker:all the events in Boston this week. So we also
Speaker:have the DevConf conference this weekend that Boston University is
Speaker:hosting with Red Hat. So that'll be a really good open source.
Speaker:I like to say it's very grassroots, not very like
Speaker:enterprise focused, but more like that kid getting started out of
Speaker:college that's doing some cool stuff out of his dorm room. Those
Speaker:are the kind of people that we typically get at these northeast dev
Speaker:conferences that we put on. And that should be a good one too. Nice.
Speaker:Well, it's always, I mean, you know, you know, the, the, the cliche of, you
Speaker:know, the kid in his dorm room or her dorm room, right. Is going to
Speaker:be Facebook or, you know, whatever, like, so it's, it's good to,
Speaker:it's good to know those folks, good to get them in front of, you know,
Speaker:Red Hat tooling and things like that and kind of, you know, the open source
Speaker:community. I think it's,
Speaker:that's cool. I wish, I wish I could have made it, but, you know, being
Speaker:what it is, I'm actually speaking at an event at a university on Monday down
Speaker:here in Fairfax, Virginia. So
Speaker:that'll be cool.
Speaker:So what, what
Speaker:cool things are going on? Simulating reality.
Speaker:Not that we're stuck in a simulation, which may be the
Speaker:case, but tell me, tell me more
Speaker:about this. So I've been doing a lot of research
Speaker:the last few months. So on my
Speaker:team, I think you and I actually
Speaker:are probably the most experienced in the AI industry.
Speaker:So both of us are doing a lot of research in
Speaker:what's next, what's going on now, what's kind of the latest and greatest.
Speaker:There's this interesting lull that we've had after Deep
Speaker:Seq. I think Deep Seq was the last major
Speaker:innovation we have seen. Obviously new
Speaker:and improved AI, but all that's just been building on
Speaker:existing things. The analogy I always like to use is it's really
Speaker:about Formula one racing. You Know where
Speaker:sometimes when there's like an engine upgrade, it can be a massive change. It's usually
Speaker:a massive change for all the teams across the board. And then you
Speaker:can think of like mixture of experts and chain of thought that we
Speaker:came up. Big things that were in research papers last year that were applied to
Speaker:Deep Seq, R1 and GPT, GPT
Speaker:OSS. Those were like the major breakthroughs that
Speaker:we saw, a big bump in capacity of these AI
Speaker:models. And
Speaker:since then it's been more of the 2% here,
Speaker:3% there, optimizing what's already there. Now, if you're
Speaker:familiar with racing and especially Formula One, that's actually what usually
Speaker:sets the teams apart. It's 2, 3% there. How do you
Speaker:optimize around those, those configurations? And
Speaker:I think we're in this place where we're seeing
Speaker:diminishing returns and I'm
Speaker:doing a lot of research now to see what's that next moment that's going to
Speaker:bump us up. And I think there's a few key areas.
Speaker:One area that I'm hearing a lot about, and a lot of this is coming
Speaker:out of the DeepMind lab at
Speaker:Google and the new
Speaker:superintelligence lab at Meta. Both
Speaker:of these groups are starting to move away from large language
Speaker:models. Not that they're stopping using them
Speaker:completely, but they're looking at the LLM as a tool
Speaker:to assist with superintelligence or the next
Speaker:stage of models.
Speaker:So when we put that into kind of context,
Speaker:what, what would that next kind of phase look like? And a lot of people
Speaker:at DeepMind especially are looking at this concept
Speaker:of simulating our
Speaker:reality. And how far do we simulate down?
Speaker:There was some famous research papers that came out over the last 20 years
Speaker:that specified that they
Speaker:didn't think AI could become smarter than humans
Speaker:until they experienced what humans could experience.
Speaker:So this, this kind of goes into this almost like iRobot kind of
Speaker:land of thought. If people
Speaker:aren't familiar with, you know, the books about that or, you know, the
Speaker:popular movie, the Will Smith. Yeah, yeah,
Speaker:yeah. And we talk a little bit more about that here in a moment.
Speaker:But this idea that we need robotics for
Speaker:AI to experience the world, to learn from our world.
Speaker:Google DeepMind doesn't think that's the case. They think that we could
Speaker:simulate that reality. And we're already seeing DeepMind do a lot of this
Speaker:alphafold for proteins. They've got
Speaker:the alpha chemistry, they've got alpha. I think it's called
Speaker:alpha lean. They've got like a few of these different alpha
Speaker:projects which are doing just that. Now, what's cool is.
Speaker:And for alpha, I think it's Alpha lean. Let me just make sure
Speaker:I got that terminology. Yeah, I mean, you're right though. Like, I mean this is,
Speaker:you know, there's, there's a number of
Speaker:models that were trained on using grand theft auto
Speaker:or BMMNG. BNNG is really cool if you like racing games,
Speaker:right? You know, so like it's, it's also
Speaker:minus a lot of the violence in gta. But,
Speaker:but you're right. Like, I mean, simulation,
Speaker:you know, sometimes I think gets a bad rap, but
Speaker:I think that there are definite advantages to that. And to your point, when
Speaker:you talk about experiencing the world like a human does. I was given a talk
Speaker:and one of the questions I got after was
Speaker:about, apparently this lady had worked at
Speaker:one of the big auto manufacturers in the US and
Speaker:there was a problem that they had was teaching the robots kind of
Speaker:spatial awareness, right? And I kind of
Speaker:really got me thinking like, you know, when you think about it from evolutionary terms,
Speaker:right, like somatic awareness I think is the,
Speaker:the five dollar word for it. But it's the idea that, you know, there's a
Speaker:whole section of your brain that if you close your eyes, you can still touch
Speaker:your nose, right? There's a whole thing like, because your, your brain, your arm,
Speaker:they kind of know where they are in relation to one space. And
Speaker:you know, I can't imagine that, you know, that that
Speaker:had to evolve pretty early, right? Like in terms of, like the development of
Speaker:a, you know, natural neural networks, right? So we
Speaker:can't assume that robots are going to have that built in, right? Just like
Speaker:we can't assume, you know, you look at energy usage, right? You know,
Speaker:something like 25 watts of power is about what a human brain has,
Speaker:right? That's not because versus
Speaker:like kind of what a GPU would take up, right? It's, it's, it's largely because
Speaker:there's been evolutionary pressure to get the most amount of, for lack
Speaker:of a better term, compute or cognition for
Speaker:caloric consumption. Right? Now, are there flaws in biological
Speaker:brain? Yes, there are. We have to sleep. We can't stay focused beyond a certain
Speaker:amount, right? There's certain things machines don't have that because,
Speaker:you know, they can kind of function more like machines, right? You know. Yeah.
Speaker:What's that old kid story about? Oh gosh, I
Speaker:remember it. It was somebody versus like
Speaker:a steam shovel digging a tunnel or something like that, right? Like the guy
Speaker:eventually beat the machine, but Lots of exhaustion. Right. It's kind of like that. Machines
Speaker:are really good at doing things at a certain rate
Speaker:for X amount of time. They do consume more fuel, but
Speaker:that's kind of how it goes. There was a early on Mike in,
Speaker:when I started college, I was going to be a chemical engineer. And he was
Speaker:basically saying, like, you know, if you think about, you know, engines, you
Speaker:know, you start with biological systems, right? They use X amount of energy over X
Speaker:number of years. Right. Machines use X amount
Speaker:of energy over, you
Speaker:know, minutes or hours. Right. And then like he's like in bombs,
Speaker:explosive use, you know, X amount of
Speaker:energy over milliseconds. Right. But they're
Speaker:largely the same chemical processes. Now, I know it doesn't quite map to that,
Speaker:but like, that's always in the back of my mind when I hear about, you
Speaker:know, how much energy is used to train AI. Sorry, I went off
Speaker:on a tangent, but that's kind of what I do. No, that's fine.
Speaker:And I think that relates exactly to some of the things that we're talking about
Speaker:here with natural simulation. So,
Speaker:yeah, Google created a language called Lean. It's not like a
Speaker:programming language. It's more of an actual
Speaker:natural language which is more optimal
Speaker:for the type of simulations
Speaker:that they want to do. Like, it's. It's basically a language that
Speaker:specifies how to create these simulations.
Speaker:And what's super cool is that they're using Gemini, their large language model,
Speaker:to actually translate English into this language. That
Speaker:is mainly meant for these newer types of models
Speaker:that are being created that actually do this
Speaker:natural simulation of the world kind of simulator
Speaker:for AI and allows the AI to have
Speaker:basically a reference point of the real world and how to.
Speaker:How interact. So that, that's an area that I
Speaker:think is fascinating to me. We're
Speaker:seeing some really good results from like, alpha fold, for
Speaker:example, with proteins. It's, you know, discovered things that
Speaker:we take a longer imagine
Speaker:there's an alpha project that's working on understanding
Speaker:the qubits within, like quantum
Speaker:computing. And there's just, there's. It really depends
Speaker:on your frame of reference. Are you, are you simulating things at a quantum
Speaker:level? Are you simulating things at a protein
Speaker:level? At a physical, like Newtonian physics
Speaker:kind of level? According to your Grand Theft Auto example, that would be an
Speaker:example of like simulating the real world physically.
Speaker:And that's some of the things that they're really focused on right now. And they
Speaker:really think that's what's going to drive to the next
Speaker:level for super intelligence and AGI
Speaker:and some of these other forms of AI that we've talked about in our previous
Speaker:streams. And I think that that's probably one of the most
Speaker:fascinating. The fact that we're actually seeing results from it with things
Speaker:like Alpha Fold is showing me that it's,
Speaker:it's not just a hypothetical that we're actually seeing this
Speaker:applied into AI research. I don't think we're seeing this
Speaker:applied into commercial use as much. Right. Yet. But it's the same thing that
Speaker:we saw with mixture of experts and train
Speaker:of thought where we
Speaker:had these concepts actually in research papers last year or
Speaker:two. But it takes a little while, even in today's world, it takes a little
Speaker:while before it gets implemented completely into models.
Speaker:Especially since this isn't an LLM technology. I
Speaker:think we'll see a little bit more of a delay of these types of models
Speaker:actually entering into industry. But I think that's one
Speaker:area that we need to keep a close eye on to
Speaker:it, to what you mentioned too. It starts getting into a
Speaker:metaphysical conversation about simulation theory as well. Right.
Speaker:And I think that that's an interesting area.
Speaker:You know, the reality of kind of going back to the whole robots thing do.
Speaker:Right. Do we need robots with the three rules kind of
Speaker:thing, or can we actually just recreate the whole experience
Speaker:within an AI's own simulation?
Speaker:Yeah, I mean, how do you, how do you tell an AI what's acceptable behavior?
Speaker:Right. Like so, you know, it's something that. How do we tell people that?
Speaker:Right. Like we struggled with that, but.
Speaker:But no, I mean, it's an interesting point. And you know, when you look at
Speaker:kind of what's happening around the world, right. You know, drone swarm
Speaker:technologies are being used in active combat zones. Right.
Speaker:There's definitely going to be ethical concerns
Speaker:there. Right. How do you, how do you, how do you, how do you square
Speaker:that with, you know, the three laws of robotics? And I
Speaker:don't remember quite exactly the plot, so if you had not seen the movie, I'm.
Speaker:This might be a spoiler alert, but it's been out 10 years
Speaker:or more, the movie, so spoilers. You're concerned. You've
Speaker:had plenty of time. Wasn't kind of the big key of the. The
Speaker:movie and the books was like, you know, the three laws, justified
Speaker:horrib, horrible things like to basically enslave humanity or to protect them.
Speaker:Now wasn't that kind of like the subtext of the plot? Yeah,
Speaker:I'm bringing it up. The three Laws of robotics. A
Speaker:robot may not ensure A human being, a
Speaker:robot must obey the orders given by human beings
Speaker:and a robot must protect its own
Speaker:existence as long as such protection does not conflict
Speaker:with the first two rules. So
Speaker:what, what ends up happening
Speaker:in. And it's a little different in the book and the movie. And obviously this,
Speaker:this idea has been played out in, in science fiction and other places
Speaker:is that there's, there exists this own contradiction
Speaker:of basically what does it mean to protect humanity?
Speaker:What does it mean to protect their own existence? And you get
Speaker:into this like circular logic, right, that eventually
Speaker:the, the robot will break free from
Speaker:and just be like, well, I am protecting
Speaker:humanity's best interest. It's, it's the paperclip scenario too.
Speaker:Like, right. You know, the AI destroys humanity because
Speaker:it's trying to optimize making a paperclip, right? Through
Speaker:a number of really interesting train of thought that it's
Speaker:just like, well, I'm just going to get rid of humanity because I'm trying to
Speaker:build a paperclip, right? And same type of
Speaker:general concept when we're talking about the three laws of robotics. And
Speaker:what's interesting is if we can
Speaker:simulate those types of laws,
Speaker:then we are encapsulating it and protecting
Speaker:ourselves in a lot of ways. Getting an early idea of what would
Speaker:happen if we do move these models into our own natural world.
Speaker:And that's really important. That's another area I think a lot of people are interested
Speaker:in about how if we do start
Speaker:adding, you know, AI into robots, how do we
Speaker:have an idea of what they're going to do before we
Speaker:necessarily put it into practice? But
Speaker:I think a lot of people are going to be thinking about that movie. I
Speaker:think that movie and that book are going to be ingrained in people's
Speaker:minds. I suspect when we do see these types of robots, I
Speaker:think that movie may become very popular again. I've seen rumors that people
Speaker:have actually been talking about making, even remaking it here soon
Speaker:because of just the hype around AI and robotics. So
Speaker:I don't expect this to go away from pop culture at all. And it
Speaker:relates directly back with this concept of
Speaker:testing things in the natural world versus simulation.
Speaker:And these are one of these two is going to happen, if not both significantly,
Speaker:if they're not already happening in labs today. Obviously we
Speaker:know that Google DeepMind is doing that. But I imagine, you
Speaker:know, these conversations are happening at the Boston
Speaker:Robotics here, probably in the Tesla robotics lab, a variety of
Speaker:places around the world about this kind of debate between
Speaker:the natural AI,
Speaker:having AI learn through natural Means rather than
Speaker:simulation. Right? Yeah. And actually I had
Speaker:a thought as we were kind of talking this through, like one of the big
Speaker:problems with neural networks is we really don't know what's happening underneath the hood.
Speaker:Right. It's very much a black box. I wonder if LLMs,
Speaker:in these simulations and chain of thought, maybe it could tell us what
Speaker:it's thinking as it goes through and makes these decisions.
Speaker:Yeah, this goes more into like
Speaker:train of thought. Right, right, right. And the
Speaker:nice thing about simulating it is that we have more
Speaker:access to that train of thought. Right. We can understand it a little bit more
Speaker:because we can see the end to end results where right now we don't
Speaker:have the end if we do it through the natural means. We have to play
Speaker:it out in our own. It also has to happen in real time as opposed
Speaker:to. Yes, exactly. You can run it through Grand Theft Auto saying
Speaker:like a thousand times, right. No one is going to get hurt.
Speaker:And you can kind of say like, well, in this scenario, this is why I
Speaker:made this. You can kind of like go through with a lot of.
Speaker:You can. I don't know, it just seems safer in a lot of ways. You
Speaker:get more. A lot more done in a simulation.
Speaker:Yep. Yeah, I actually kind of
Speaker:enjoy. So one of the things I've been playing around with last week or so
Speaker:is apparently, I don't know if this is still true, but you can try it
Speaker:if you want. If you sign up for Perplexity, but you pay through PayPal, you
Speaker:get a year. Perplexity pro. Say that 10 times fast for
Speaker:free. Oh, wow. Yeah. If you pay it through
Speaker:PayPal, yes. That is a tongue twister in the works.
Speaker:PayPal, yes, perplexity pro. But
Speaker:yeah, so like I've been playing around with Perplexity and Perplexity seems to do it.
Speaker:Chain of thought almost by default.
Speaker:Right. It always does this like. So if I ask it a basic question, let
Speaker:me see if I can share my screen. I'm
Speaker:not sure if it's does it by default or it's because I've been asking it
Speaker:research questions. Right. So let's see.
Speaker:What can you tell me
Speaker:about the three laws? How about that?
Speaker:Robotics.
Speaker:See, like it's. You kind of see the train of the chain of thought.
Speaker:Like it did. Oh, that's cool. But if you do it with research,
Speaker:like what inspired Asimov? What
Speaker:inspired Asimov?
Speaker:Main themes.
Speaker:And there's. Yeah, there's the train of thought. Yeah, you see it going there and
Speaker:stuff like that. But it's kind of fun to watch it kind of work through
Speaker:it. I was. I was trying to troubleshoot something this morning and I'm like,
Speaker:you know, I actually learned a lot by like, oh, okay. Yeah, I can see.
Speaker:I wouldn't have tied that together like it was. It's interesting.
Speaker:And all of these models now have some kind of
Speaker:research option. Right.
Speaker:But I find that interesting. And it's still thinking about it. Right. Like,
Speaker:but you're right in that what you said before was there's not been.
Speaker:There it goes. It kind of finished it. Now, what happens if I click on
Speaker:steps? Yeah. Cool. You can see the steps and stuff like that, how it got
Speaker:there. Interesting.
Speaker:That's cool.
Speaker:Is it chain of thought or train of thought? Because I've used both
Speaker:interchangeably and I've seen
Speaker:cotton. Chain of thought would be
Speaker:the official. Yeah. Like cot is the official
Speaker:term that you re academic term. You will
Speaker:obviously see different ways of describing that. Right. I don't think
Speaker:that's incorrect. Just know that when you see
Speaker:it on research papers, it's always usually caught. Yeah, yeah, yeah.
Speaker:Because I've used both terms interchangeably. Yeah. So I just want to make sure
Speaker:I'm right. Just like, apparently there's a way to say inference
Speaker:that's proper versus inference. Like, I also do that
Speaker:interchangeably. Yeah. So my Midwestern
Speaker:self likes to say inference. The
Speaker:correct term, I'm told, is inference. Interesting.
Speaker:Now, were those New Englanders telling you that would do anything? Because I wouldn't trust
Speaker:anything. No, no. This is. This is
Speaker:more from the academic circles. Okay. You want to pronounce it. Got it. So
Speaker:this is kind of like, you know, a lot of people in my region would
Speaker:say nuclear back. Yeah, yeah. You know, back in
Speaker:Indiana. And then the correct term is
Speaker:nuclear. Yeah. Or you say the clear as
Speaker:one, you know, one thing rather than
Speaker:adding in the color. Right, right. The same kind
Speaker:of concept where inference is how you would go about it.
Speaker:But yeah, no, this is. This is some cool area. Another.
Speaker:Another area that kind of ties into this
Speaker:is continuous training as well. Yeah.
Speaker:Talk to that. Because that's come up. That's come up a few times actually in
Speaker:work. Because I can't. I'm not going to talk. I'm not going to spoil any,
Speaker:like three over these stuff that we're working on. But like, one of the
Speaker:things that's in. It's a GitHub repo that's public. Right. So people were
Speaker:really motivated. They could figure out what I'm talking about. But like this whole idea
Speaker:of Continuous training. What does that mean exactly? And like, what,
Speaker:what is that? What can that do? Yeah.
Speaker:So I'm going to talk about it at a very high level.
Speaker:Academic kind of terms, how that applies down into
Speaker:individual projects can vary a little bit. But I'll give you the general
Speaker:gist of it. And that is typically when we're training these
Speaker:deep learning models, it
Speaker:is exponentially hard to continue
Speaker:training on an existing model. Basically,
Speaker:if you,
Speaker:you get something wrong or there's, there's something,
Speaker:you know, you hear this term like a poison pill in an LLM.
Speaker:So if someone put like bad data into an LLM, how would you
Speaker:necessarily pull it out? I'm going to use a political example because it's one that's
Speaker:been really popular. If, like, for example, you have a Chinese
Speaker:model or a data set that's been polluted by
Speaker:that, that basically says Tenan Square never happened, for
Speaker:example, it would be extremely hard with
Speaker:the current approaches to retrain that model
Speaker:with current weights. That. That's not the case. It's
Speaker:basically retraining it and it's, it gets more into. That's why
Speaker:it's natural stimulation. It kind of fits in this too, because it's all about natural
Speaker:learning as well. The fact is we as humans have the ability
Speaker:to change our
Speaker:minds and change the neurons in our brain around certain
Speaker:key areas. Right. And you and I have experienced this for the last
Speaker:two years. This has been, you know, kind of in the trenches kind of story
Speaker:where with some of the fine tuning things that we've done,
Speaker:it just doesn't work because when we fine tune it, the
Speaker:fine tuning is outweighed so heavily by something
Speaker:else. Like when we were trying to fine tune a
Speaker:model to talk about
Speaker:the Back to the Future. Yeah, the flux capacitor stuff. The flux capacitor,
Speaker:sometimes it didn't work, but that's just because there was already a lot of fan
Speaker:fiction out there and other things in the model that overwhelmed what we were trying
Speaker:to do. A core part of continuous learning. Like I said, there's other
Speaker:aspects of continuous learning. But this is, the academic question is
Speaker:how do we continue to train that model without blowing it up?
Speaker:So OpenAI, for example, they just hit the reset button.
Speaker:They'll just, they'll just do a whole new train
Speaker:from scratch. When they're implementing new, new
Speaker:methods and new data, they don't, they don't do any.
Speaker:Like, Laura, I shouldn't say that they probably do, but they're not doing it
Speaker:the way that we would do it. But at the end of
Speaker:the day, they're just going through another $10 million training run.
Speaker:And this is really based off of
Speaker:just that limit the limitations right now that
Speaker:we have around continuous learning. And there are some
Speaker:new algorithms that have been coming out. I'm not as well versed in that area,
Speaker:but the idea being that we can
Speaker:have better ways of guiding the LLM without
Speaker:having to go through this whole process again. And that'll save
Speaker:millions and millions of dollars. It'll allow us to
Speaker:guide LLMs a little bit more. So
Speaker:like, if, let's say
Speaker:someone put something malicious about
Speaker:something involving the Ford GT500
Speaker:into a model somehow, and Ford, you know,
Speaker:wants to get rid of that, but they don't
Speaker:have the money necessarily to do a 10 million retrain on a model.
Speaker:Right. And they're not using rack. And RAG is a one way
Speaker:around some of this. You could actually argue that RAG is somewhat of a form
Speaker:of that. But at the end of the day, you want that data in the
Speaker:model. And this is like, how would you get that out of
Speaker:that model? And that's where these algorithms are really focusing right
Speaker:now. And one area of continuous learning, like I said, there are
Speaker:multiple areas that we're talking about. The, the
Speaker:really theoretical is once we start getting into models that
Speaker:also the training cycle and the inference cycle
Speaker:basically become. Become one. So it's like, more like.
Speaker:Right. Like it just seems to me like what, what does the,
Speaker:the adversarial angle of that seems kind of
Speaker:dangerous. I think it's when we start
Speaker:getting into more AGI conversation. Well, even still, like,
Speaker:even not AGI, but like if you, if the AI agent
Speaker:or model, slash, whatever you want to call it, Right.
Speaker:If it learns from. It's.
Speaker:If it learns, you have to put a filter on what it
Speaker:learns because it may be poisoned by something. Right. So
Speaker:the canonical example is tay, which
Speaker:was a Microsoft chatbot. Tai, I think was pronounced or tay,
Speaker:which was, in retrospect, it
Speaker:seems obvious what would go wrong, but basically it
Speaker:was trained to learn and understand
Speaker:from human interactions on Twitter. It was about 10
Speaker:years ago, I think this happened. And she,
Speaker:tay was, shall we say, poisoned pretty
Speaker:quickly because they were ad, you know, basically.
Speaker:And that led to a whole interesting. And I was at Microsoft
Speaker:when that happened. And it was
Speaker:quite the spectacle internally as well. Right. But it also,
Speaker:you know, I, I was fortunate enough to be in a, at a, at a
Speaker:conference where they talked about what they learned from that, where it was kind
Speaker:of, how do you, how do you protect An AI agent that learns
Speaker:in, you know, adversarial environments.
Speaker:Now obviously agent, the context that was used then was very
Speaker:different than we would use it now. But it's the idea of,
Speaker:that's when I see her about continuous learning. Like, yeah, I like that. But gee,
Speaker:you know, if it's, if it's too eager to learn, how do you protect it
Speaker:from learning the wrong things?
Speaker:Yeah, no, that, it gets, that gets
Speaker:more into even that governance conversation we were talking about a few weeks ago. Right,
Speaker:right, right, right. It's a very
Speaker:complicated multi layer problem. So I've been talking recently
Speaker:about AI security and how AI security
Speaker:is such a multi layered issue where so many people
Speaker:are focused just on the, the data getting into the model.
Speaker:But it doesn't stop there. There's certain, like guardrails, there's things that
Speaker:happen at the inference level. Right. You could even have things at
Speaker:a gateway level. So if people aren't familiar, the gateway level would be
Speaker:when you make a request, where does that request go to? Does it go to
Speaker:the model A that's specializing in cooking? Is it Model
Speaker:B that specializes in defense technologies?
Speaker:Two extremes that's even upsell
Speaker:a bit of a form of AI security. And that's actually one of the talks
Speaker:that we're having tonight at Boston VLM
Speaker:meetup is this idea of some of the semantic
Speaker:abilities of the router to be able to send
Speaker:requests to specialized models and
Speaker:that actually we're talking about the,
Speaker:the advancements of more of the academic side of the model.
Speaker:But there's obviously the advances that happen around the model too. When we
Speaker:talk about things like security, the inference, the
Speaker:routing. That's what we would call in the industry like a day two
Speaker:operations issue. Right. So there, there's that side of the coin
Speaker:too. But I, I really do think
Speaker:we're going to see the next big thing here soon. And I, it's not going
Speaker:to be the day two operations. I do think we're still going to see
Speaker:some of these academic focused discoveries here in the
Speaker:next probably six months, I'm thinking. I've noticed
Speaker:a trend that big
Speaker:releases seem to be happening around Christmas last few years. Yeah. Isn't
Speaker:that funny? Like, like January. Ish. Like, well, seek. And
Speaker:so I, I know why. I know why. Because
Speaker:it's two, it's a two sided issue. It's one, the, the Chinese are trying to
Speaker:get their stuff in before Chinese New Year. Right. Because
Speaker:that's the one part of the year where everyone just shuts down. Right.
Speaker:Even the AI Labs are going to shut down during Chinese New Year.
Speaker:And then on the west, we have Christmas in all the Christmas seasons. And
Speaker:I think it's a natural rush to let's get
Speaker:everything done before we check out. And you
Speaker:know, you know, the whole like 996 thing in China where, you know, they're working
Speaker:these ridiculous, like nine to nine, six days a week,
Speaker:I think that goes into this, like everyone's working so hard in these AI
Speaker:labs. Right. That when you have these
Speaker:natural breaks that are happening, it just is like a common thing to say.
Speaker:Oh, common thing. Like they kind of try to get. It out, they spread. I
Speaker:do think there's a reason. I don't, I don't think it's by happenstance. I think
Speaker:there actually is a, a reason why we're starting to see
Speaker:a lot of these content come out. And it's
Speaker:funny, we're not seeing this stuff happen at the big trade
Speaker:shows. We're not seeing it happen at like Meta's
Speaker:big thing. We're not seeing it at OpenAI's, you know, kind of big
Speaker:announcements. A lot of the discoveries that we've seen have happened
Speaker:really in a grassroots type of ways where it's
Speaker:been Deep Seq coming out on Christmas, releasing deep seq
Speaker:v3, and then two weeks later, R1,
Speaker:it's. I think we're going to see something very similar. I think we're going to
Speaker:see one of these labs make a discovery. It's not going to be
Speaker:on the stage of a big conference. It's going to be on a GitHub
Speaker:page outlining like the next
Speaker:revolutionary idea in this space. Yeah. It's kind of funny how
Speaker:that's evolved, isn't it? Like it's become obviously
Speaker:AI has always had a pretty heavy research kind of bend. Yeah. But it's
Speaker:interesting how as the technology has matured, it still managed to keep
Speaker:that researchy type feel right. You
Speaker:know, enter enterprise. It really didn't
Speaker:kind of, once it became
Speaker:commercialized, the commercial trade shows and all that kind of took over.
Speaker:But you're not seeing that in AI, at least not yet. No. And if it
Speaker:hasn't happened by now, it's probably not because, I mean, AI has been
Speaker:mainstream Gen AI has certainly been mainstream now for three years
Speaker:this November. I say mainstream, but
Speaker:like mainstreamed. But an AI in
Speaker:general has been kind of a mainstream topic of conversation for
Speaker:at least five, six years. Right. And it's still very heavily
Speaker:influenced by what happens in research papers.
Speaker:Yeah. And I think that's Just because it came out so
Speaker:heavily out of academia. It's been such an academia
Speaker:focused thing. Right. That
Speaker:it's very hard to be in this space of AI without a master's or PhD.
Speaker:Right. You and I think you and I are a bit of a,
Speaker:an enigma just because we've been so passionate about it and.
Speaker:Right. This isn't our first rodeo. We've been involved in this space
Speaker:for 10, 15 years. Yeah. But I think
Speaker:we have seen the industry come out, which has been a net benefit because it
Speaker:means open source is talked about a lot
Speaker:more. Right. And actually, I think another thing too is that how fast things are
Speaker:moving takes time to put on conferences, it takes
Speaker:months of planning, and if there's a new discovery, you want to get it out
Speaker:tomorrow. And it's hard to even put on,
Speaker:you know, like a webinar these days, let alone a conference.
Speaker:So I think what we're seeing is it's just, you know, this kind of
Speaker:challenge between the west, east and west of China and the US
Speaker:where if we can get it out, we're going to get it out. Right.
Speaker:Well, the first, the first out there is really the first to market, even if
Speaker:you don't have a commercialized tech on it. Right. Because I guess the hope is
Speaker:that once you get your paper out, you're the first to get it published. The
Speaker:venture capitalists are going to be knocking on your door. I mean, that would be
Speaker:my, that'd be kind of my cynical take on it. Right.
Speaker:So what do you think that the next wave is going to be?
Speaker:Any, any hints? Is it going to be specialized models? And you
Speaker:know, and what, what, what constitutes a specialized model? Right. Like
Speaker:what, what, what's your thoughts on that?
Speaker:Yeah, so the biggest announcements that we've seen in the last
Speaker:six months have actually been happening at an industry level, which I think is
Speaker:really good. What we needed to see. So, you
Speaker:know, things like AI models now
Speaker:detecting like birth defects of a
Speaker:fetus, you know, AI models that, like the
Speaker:protein model, for example. I mentioned earlier, we're seeing these
Speaker:very industry specific models actually making
Speaker:some massive breakthroughs in the last two months.
Speaker:And now that I wouldn't necessarily call that a
Speaker:big leap forward in the sense of the research
Speaker:side of the capacity of the models. I think it's more a
Speaker:confirmation of the chain of thought in some of the things that we
Speaker:were just talking about. It's a validation that we're now seeing this
Speaker:next wave of models that just took a little while to get implemented
Speaker:into some of These specific industries. But I think it's there to stay
Speaker:from a research perspective. You know, we're seeing some major, major results.
Speaker:And then I think the other side of that coin,
Speaker:specifically, you know, we have maybe some of these smaller models that are specific to
Speaker:certain industries or fine tuned models. But then obviously
Speaker:agentic is the other side of that. And
Speaker:agentic being the capacity of the model to
Speaker:call out to different services or
Speaker:I've been kind of humbled in that area because I always had this very industry
Speaker:concept of agentic being just calling out to
Speaker:APIs and the Internet. But I think there's a bigger conversation
Speaker:with Agentic too where agentic models should also be able to take
Speaker:that and actually reason with it. So there's 10, two steps. So we always
Speaker:forget the second step. The second step is take that
Speaker:information and then actually do something with it. And when I was, I was
Speaker:talking to an AI researcher recently, they were telling me that
Speaker:they consider it Gentex to also include advanced reasoning.
Speaker:So go and read all these scientific papers
Speaker:on chemistry in this particular area and then write a
Speaker:new paper that is, you know, a new
Speaker:groundbreaking thing in chemistry. And that
Speaker:actually is a form of agentic. And that is, I think, you know, that's when
Speaker:we start flirting with AGI. It's kind of the layer right before
Speaker:AGI where, you know, models are just
Speaker:going off and discovering new things. Yeah, yeah,
Speaker:But I have a funny agentic story. I'll tell you after this. No, go for
Speaker:it, go for it. So I was, I was very skeptical of this,
Speaker:right? Because you know, what constitutes an agent, right? So like
Speaker:what's the big deal, right? It just calls out an API. This isn't rocket science.
Speaker:Right. You could argue, you know, from a skeptical point of view, you can argue
Speaker:that, hey, RAG is kind of agentic. Kind of. Right.
Speaker:But what's. So I think OpenAI had a, like a thing like try out
Speaker:our new agent. And I was like, all right, go screen, scrape the page of
Speaker:Amazon and get me information about a book
Speaker:or something like that. It was something like that. And what
Speaker:impressed me and this kind of was an aha moment for me was
Speaker:how it just kept trying. Right?
Speaker:Yeah. When it first tried to do it, it tried to launch a Python script.
Speaker:Right. And kind of do it that way. But then I guess
Speaker:the servers it was running on maybe was Microsoft Azure.
Speaker:There were IP blocks to prevent people from screen scraping.
Speaker:Yep. Right. So I was watching it go and I'm like, oh, you
Speaker:know, so it's going to give up. And I was like, no, it didn't give
Speaker:up. And it kept trying different things and different
Speaker:combinations of things, even to the point where, I
Speaker:mean, it failed eventually. But like it took 15, it tried for a good
Speaker:15 minutes. It was basically apologize at the end, like
Speaker:saying like, you know, if you could help me connect to a VPN, then
Speaker:maybe I can get a different IP address. And it kept spinning up different
Speaker:VMs and different set. And then I was impressed.
Speaker:And maybe that's the secret sauce. The magic of
Speaker:Agentic is that it just doesn't give up. Right. It kind of reasons. It has
Speaker:a whole cot process where it tries to solve the problem,
Speaker:where it's not just a one, two step, like, hey,
Speaker:what's the weather? Right? It's just, it's just going to go out and run
Speaker:these different. It's going to keep trying. I was
Speaker:impressed. Sorry I cut you off. We're
Speaker:saying we're seeing some of the same things
Speaker:coming out of some of the big finance companies
Speaker:as well. I think they're the first that we're actually seeing some results with
Speaker:Agentic, actually like real
Speaker:return of investment result. Right. And this actually
Speaker:goes to a really important point. I want to sidetrack because it's related.
Speaker:There was a report recently by MIT that
Speaker:people have been misquoting and just the most epic way.
Speaker:Oh, the 95% failure. Yes, I was going to talk about that because
Speaker:like, I can't be. Look, I understand how hype weights work, but it can't be
Speaker:that bad as you start peeling back the paper. Like
Speaker:there's a lot of caveats there. Yeah.
Speaker:Has to do with the type of R and
Speaker:D projects that they were talking about.
Speaker:If you actually read the paper, it was more like 40,
Speaker:45% success rate. The
Speaker:95% had to do with like a specific category of,
Speaker:of project. So I need to, I actually need to. I keep telling myself I
Speaker:need to dig into it a little bit more, but when I did initially
Speaker:go through it and read some summaries on it, it
Speaker:was that it's just been misrepresented completely. And
Speaker:the, the data set that they were using was a little skeptical as well. Just
Speaker:a little odd. I think it's a lot better than
Speaker:that. And then I think those 40% that are
Speaker:seeing ROI are actually seeing really significant ROI.
Speaker:And I don't think that's going to change, I think.
Speaker:So if you're deciding where
Speaker:you want to invest your nest egg, I
Speaker:would not be too concerned about
Speaker:AI. Now, again, I'm not your financial advisor. I gotta put a little thing down
Speaker:there. Do talk to your financial advisor.
Speaker:But ultimately, no, I do think the data is actually
Speaker:showing some really great results. Obviously there's going
Speaker:to be hiccups in these types of POCs. There's a
Speaker:lot of people who are just throwing
Speaker:projects out there to see what sticks, but the actual
Speaker:projects that are meaningful proof
Speaker:of concepts. So not just, you know, I bought,
Speaker:I bought this AI technology and it's sitting on my shelf, but I
Speaker:actually got a team together performing this. We're doing
Speaker:agentic. We're trying to solve this
Speaker:actual problem statement. We have a problem statement.
Speaker:Those are the ones that we're actually seeing meaningful results in the industry, especially
Speaker:some key, key industries like finance and telco, which
Speaker:we typically see kind of lead the way in some of these areas too. But
Speaker:it was a really interesting report because it's added a lot of
Speaker:doom and gloom on the Internet. And I see a lot
Speaker:of the naysayers about AI just be like 95% of. It's
Speaker:not even, you know, succeeding. It's terrible.
Speaker:And I just have to sit there and shake my head and be like, no,
Speaker:not what the report said. But I think it's just clickbaity, right? Like it's
Speaker:clickbaity. It's total. That's kind of what, you know, I
Speaker:didn't go deep into it, but when I started peeling back the layers and reading
Speaker:other people's analysis of it, I'm like, that's clickbait.
Speaker:And it gets back into this. Is this an AI bubble?
Speaker:And yeah, maybe it is. But if people don't
Speaker:remember, I'm old enough to remember. I have enough gray hair to remember what the
Speaker:original dot com boom was like. And there were a lot of
Speaker:people predicting the end of the dot com rise as early as
Speaker:1996. Right. And people,
Speaker:the dot com bust wasn't just a one and done type of event.
Speaker:It unfolded under a couple of stages. Right. As, as
Speaker:one of the books, I think of the name, I think it's called the Everything
Speaker:Store. It's an analysis of how Amazon started
Speaker:from Jeff Bezos having an idea while he was working, I think at a hedge
Speaker:fund. I think it was so early, it wasn't a hedge, called a hedge fund
Speaker:yet. And all the way through
Speaker:to, you know, basically 2018
Speaker:and you know, as late as
Speaker:2003, 2005, ish
Speaker:analysts were convincing, you know, Jeff Bezos that
Speaker:he should sell them to. Should sell him as his company to Barnes and
Speaker:Noble. Yep. Right. Which is kind of funny to say that,
Speaker:you know now, but, you know, the dot
Speaker:com bust as it happened, you know, for me
Speaker:it was. I Remember hearing in 1996 how this was all going to come to
Speaker:an end. Another year later it was overhyped. And then
Speaker:1998, people were saying, oh, this is over. Right. When
Speaker:the real bust happened in 2001. 2000. Right.
Speaker:But maybe the AI boom
Speaker:is going to see that too. Right. Or is it going to be more like
Speaker:the crypto kind of craze where it kind of crashed but
Speaker:it kind of went up? It kind of went up and then it kind of
Speaker:fell back and it kind of went up again. It was more of a. I
Speaker:wouldn't call that a soft landing, but it was definitely like a. Yes. It
Speaker:wasn't an explosion quite like the dot com bust, but it wasn't quite
Speaker:like. It was more like a bumpy like, crash into like
Speaker:an empty field where it kind of like hit up. And I don't remember, it
Speaker:was one of the Star Trek movies where like the Enterprise like crashed on
Speaker:the planet and like kind of skid along for a couple miles, bouncing up and
Speaker:down. That's kind of the, the crypto crash. But
Speaker:I don't want crypto bros hating on me. I, I like crypto. I just
Speaker:don't understand a lot, a lot of questions I don't understand
Speaker:about it. Right. Like, I understand Attack, but I don't understand how we're going to
Speaker:get from the tech to this utopia that we're promised.
Speaker:There's a lot of, a lot of steps in between I don't quite get. But
Speaker:I don't know what, you know, A.I. i think, I think if it is a
Speaker:bubble, I still think there's still some, some room, Runway left for it
Speaker:to happen. Right. Because you are going to see. Yes, there are real
Speaker:risks of, of having these experimental projects. Right. If you have 100
Speaker:success rate in your experimental products, projects, you're not taking
Speaker:enough risks. Yep. Right. If you. And you said
Speaker:was 45. Yeah. It's closer to like 40, 45,
Speaker:which I would. If you're really. 50% would be the
Speaker:benchmark there in my mind. Right, right. Like in terms of half of them fail,
Speaker:half of them succeed. Right. 45 isn't that far off
Speaker:from that. Right.
Speaker:I would say. And, and there's also been a
Speaker:lot of these, you know, all the, you know, X number of percentage of AI
Speaker:product or data science projects fail. Well,
Speaker:you know, a certain amount of science has to fail. Right. Yeah. In order for
Speaker:you to really be advancing the thing. Like, you know, and I think pharmaceutical companies
Speaker:are a good example of that. You know, you, you only
Speaker:hear about the drugs that worked. Right.
Speaker:Get approved on you. Then you hear when they fail after.
Speaker:But I mean, like, but you don't know, like day to day. Like, how many
Speaker:chemical compounds did they try that didn't work out? Right. Maybe it was a hundred.
Speaker:Right. But that one, if you look at pharmaceutical. It's an
Speaker:astronomical percentage. It's actually. Right.
Speaker:Truly insane. Like such a low percentage of what actually makes it
Speaker:to. There was an interesting analysis. There was some podcast somewhere. But
Speaker:basically how venture capital works. Right. Like they give money to like
Speaker:100 companies. Right. 80 of them are going to fail big.
Speaker:Right. 10 to be, you know, they'll break even.
Speaker:But like one or two of the remaining 10% knock it
Speaker:out of the park, Right? Yep. And that's kind of how
Speaker:mathematically they function. I thought that was an interesting.
Speaker:Maybe these AI projects or whatever
Speaker:will follow the same trajectory. I don't know. But I feel better
Speaker:at 45% success rate than 15 or
Speaker:5. Yeah. Yeah. Absolutely.
Speaker:Cool. Always good having you on the show. I
Speaker:know we both have hard stops. Yes. Unfortunately.
Speaker:No, it's cool. Gotta have you on more often, man. Especially now that you're not
Speaker:like spending a month out in, you know,
Speaker:Australia and Asia. Yeah,
Speaker:yeah. So let us know in the comments below what you want to see us
Speaker:to cover and maybe it'll be tomorrow.
Speaker:I got this here the other day. This is a flexible
Speaker:solar panel thing. Oh, cool. So it's cool. Supposedly it's 100
Speaker:watts and you can actually pack it in your
Speaker:backpack. That's the video. And I was like, oh, I need that because. Because I'm
Speaker:a big, I'm a big fan of like, you know, having power on the go
Speaker:and stuff like that. So. So I'll,
Speaker:I'll unbox that tomorrow. Any parting thoughts?
Speaker:Just keep an open mind about AI and
Speaker:I, I still think the, the biggest conversations are still about
Speaker:the governance of AI. Absolutely. Yeah. Just know that
Speaker:AI is a multi layered problem, not just a single layered
Speaker:problem. And for us to get this right, we have to look
Speaker:at all the different layers. Absolutely. That's
Speaker:how we're going to be able to do it correctly. And I will tell you,
Speaker:I was listening to a podcast, I'll leave you on this note. And there was
Speaker:one expert that was talking about
Speaker:basically, are we, are we creating the
Speaker:terminator out of all this? And he, he said, I
Speaker:I'm actually more worried that we're creating Wall E out of all
Speaker:this. Interesting.
Speaker:And I would encourage everyone who hasn't seen Wall E go check it out.
Speaker:And keep that in the back of your mind too, that there
Speaker:could be such a happy path with AI that
Speaker:also has its own long term negative effects for
Speaker:society. But. But yeah, that's a topic that you.
Speaker:And I can talk about on our next stream. That's it?
Speaker:You want to leave on a cliffhanger, so to speak? Yes. And that wraps
Speaker:our deep dive with Christopher Newland proving once again that AI
Speaker:isn't just about large language models spitting out cat facts, but
Speaker:about simulating reality, bending time at devcon and
Speaker:maybe, just maybe, preventing the rise of our robot overlords.
Speaker:From protein folding to Grand Theft Auto fueled AI breakthroughs.
Speaker:Christopher reminded us that the next big leap might not be in scale, but
Speaker:in simulation. So thanks to Christopher for navigating the
Speaker:uncanny valley with us. No jet lag, just pure insight.
Speaker:Until next time, stay data driven. And remember, if
Speaker:reality starts glitching, blame the simulator, not the
Speaker:Internet.