In this 349th episode of data driven, we are pleased
Speaker:to interview Pavel Goldman Khaledin, where he's the head of artificial
Speaker:intelligence and machine learning at Sumsub.
Speaker:Sumsub isn't your average AI startup. They're
Speaker:globally recognized for their work in k y c, AML,
Speaker:and anti fraud technologies. Our guest is the
Speaker:wizard behind the curtain, crafting tech to outsmart financial
Speaker:fraud does and deep fake artists. Quite the
Speaker:digital Sherlock Holmes, if you will. Now here are
Speaker:Frank, Andy, and Pavel.
Speaker:Hello, and welcome to Data Driven, the podcast where we explore the emergent
Speaker:Fields of data science, artificial intelligence, and,
Speaker:of course, data engineering, which is basically the underpinning of it
Speaker:all. And with me on this, journey is my favorite data
Speaker:engineer of them all, Andy Leonard. How's it going, Andy? Good, Frank.
Speaker:How are you? I'm doing alright. We we were recording this, the day
Speaker:after we did a 2 hour show,
Speaker:Kinda by accident, don't I see our guest in, it look kinda had this
Speaker:look of, uh-oh. No. It's not gonna turn. I can't do that today.
Speaker:But we are very excited here to in spite of our issues with Microsoft
Speaker:Bookings, in spite of our crazy hectic schedules, And in
Speaker:spite of your allergies and, really tasty jelly jam and
Speaker:and and biscuits Really sorry about that. No.
Speaker:I I don't know what it is on the East Coast this week, man. It's
Speaker:it's well below freezing, and I'm sneezing. Oh, that rhymed.
Speaker:Allergy station should be over for me. I don't know what's going on. For real.
Speaker:But our guest is actually, from Berlin,
Speaker:and one of my favorite cities in the world. In fact, they were singing the
Speaker:virtual green room. Had I lived in Berlin instead of Frankfurt, I probably
Speaker:never would have come back to New York,
Speaker:or the US, but he is
Speaker:our guest today is Pavel Goldman Kaledin. Hopefully, I said that
Speaker:right. He is the head of AI and ML
Speaker:at Sumsub, a global know your customer anti
Speaker:money laundering, anti fraud company, and,
Speaker:we're we're welcome to we're happy to have him. Although, I don't think he's in
Speaker:Berlin today. I think he's somewhere a bit warmer. Welcome to the show, Pavel.
Speaker:Yeah. Hi, guys. Happy to be here. Good. Good. So I have
Speaker:a lot of questions. You know,
Speaker:first off,
Speaker:I think I can kinda see the map, but What's the
Speaker:connection between know your customer, KYC,
Speaker:anti money laundering, and anti fraud? I think I think
Speaker:I see it, but I wanna hear you you kinda walk me through it because
Speaker:I haven't had enough coffee either today. So so what's the, like, what's
Speaker:the common thread? Because, like, because I I've not seen those 3
Speaker:kinda put together in kinda 1,
Speaker:sentence, but I can kinda see why. But
Speaker:I I I I can try to explain. But the thing is and we actually
Speaker:this is what we focus on. So we try to secure as a company.
Speaker:We try to secure the whole customer journey from
Speaker:onboarding. So this is the first step of when, for instance, like, I'm in a
Speaker:bank. So So I want to onboard some of my customers, and I want to
Speaker:make sure that this has real persons, for instance, that are not fraudsters.
Speaker:So I want to onboard them, make sure they are,
Speaker:that person, they actually pretend to be. And then and
Speaker:here's the thing. If I can, for instance, like, I'm a Journey
Speaker:person. But a month later. There could
Speaker:be some, you know, strange patterns of, you know,
Speaker:financial transaction happening. So probably, there are some sort of a pattern of
Speaker:money laundering. So this is where transaction monitoring comes.
Speaker:So you can actually this is a person. So this is but knowing customers are
Speaker:very simple. You can actually I mean, you
Speaker:can So basic basic attack is to be just pretend to be,
Speaker:a person. You you are not, basically. But then even if I'm
Speaker:not, I'm just a real person, I can actually, yeah, come up with some sort
Speaker:of, you know, few things to
Speaker:do. And then where just we try to monitor it, and then from a permit,
Speaker:make sure that, Okay. We can actually flag the transaction and
Speaker:then make sure it's it's it's getting looped. And then, I mean, there is a
Speaker:flag raised, and then, Probably, we can do
Speaker:something about that. This is just, like
Speaker:this. If we're talking about anti fraud, and here's the
Speaker:thing. Sometimes it's very easy to see that something fish is
Speaker:happening. So for instance, like, A very like, 2 years ago, it
Speaker:was a very typical attack. So I tried to, you know, open a bank
Speaker:account or, like, remotely, And I actually, I'll leave somewhere
Speaker:else, or I don't I I use a stolen document. What
Speaker:what I can do To do that, I can actually just print out the
Speaker:image of a person and just try to make sure that actually the
Speaker:KFC provider like us Tried to make us
Speaker:believe that I'm a real person. That was a very, you know, typical attack 2
Speaker:years ago. Now it's very easy to detect. Still peep some people use
Speaker:it. And that's it. And that's that for us. It is very easy to do
Speaker:that. But probably, I mean, this is not a real person. Some of you trying
Speaker:to use the printed out images. This is
Speaker:Fraud. We can actually or reject it or or ask a person. Can
Speaker:you well, I mean, we need your real real pay real real image.
Speaker:Or we can just tell our customers that, you have to take a look because
Speaker:there was something fishy going. And then it goes and goes and goes. And the
Speaker:whole customer journey, We try to make sure that the fraud is not happening. This
Speaker:is basically it. So
Speaker:fraud is kind of, I think, Cyber fraud or whatever the cool
Speaker:kids call it, I think is has has infected
Speaker:every industry. I mean, if I just I
Speaker:mean, I I get 2 factor authentication logging in the
Speaker:roadblocks, like, for my kids. Right. And I'm like,
Speaker:they'll they'll they'll they'll get in front of their device, and they'll be like, can
Speaker:you tell me what the passcode is that they texted you? Like, Sometimes
Speaker:some days it's the only way I see 1 of my kids. But,
Speaker:has the because I I wonder, like, has the pandemic kind of Accelerated
Speaker:kind of virtual fraud, or is that just independent?
Speaker:I think it I think it is. Because it, right now, it's but it's not
Speaker:Related to fraud. Exactly. But the thing is is that now
Speaker:people are used to actually work
Speaker:remotely, Or it's so it's not that common for you
Speaker:to go to bank in person. So you just call there. You just I
Speaker:mean, use over the internet, basically. It's like easier
Speaker:So and now you can actually, there is no way, you can actually verify
Speaker:that this is the only person. Right. Yep. And this is a final thing because
Speaker:for instance, in Germany, where I reside, most of the
Speaker:time. There is a regulation called it's called
Speaker:video ident. So for in Germany, in order For for
Speaker:me, if you are going to open an account, anyway, I really have to call
Speaker:a person, a live in person operator, And talk to him, and he makes
Speaker:sure that or she makes sure that, a a million person. But everybody
Speaker:do not like it, basically. Because, I mean, it it takes time. You have to
Speaker:talk, talk to a person. I I just want to open an account. So it's
Speaker:it's it's it's fast as I'm but but except Germany, all of the rest
Speaker:of European Union, I think across the world as well. It's, I mean, you
Speaker:just Send your image or video, some of your documents,
Speaker:and then the the account is up. So it's very easy. And people get you,
Speaker:getting used to it. And that's why it's easier to to to actually,
Speaker:do fraud because it's, I mean, it's it's a soldier to trade off,
Speaker:Make it easier, and then it's easier for fraudsters to actually do their business. So
Speaker:that's that's the thing. Gotcha. Do you see,
Speaker:you mentioned you see, Like, new scams, people
Speaker:are running as well. And you also mentioned a lot
Speaker:of what I I thought would be pretty effective ways to to
Speaker:combat those scams, without really
Speaker:giving anybody any ideas. Are there, like, brand new
Speaker:scams that have happened maybe in in the very recent past
Speaker:that, you're still working on ways to combat?
Speaker:I must say that, there is there will always
Speaker:be some sort of, you know, arms, right.
Speaker:Competition? Yeah. So you have to say or. There will always be,
Speaker:like, a new prod Of yours.
Speaker:And then we have to actually deal with that. But I can tell you a
Speaker:story. So for instance, like, so we asked him so not a big company. Yeah.
Speaker:The technology team is not that So big, we have to move fast. But
Speaker:in my team, the AI slash, ML, it's not
Speaker:anti money laundering, but artificial intelligence slash machine learning. We have
Speaker:a very small department aimed at creating defects.
Speaker:So we do not detect defects. We have to actually learn how to create
Speaker:them So you actually know how I mean, how people actually read Oh,
Speaker:that makes sense. So synthetic data. Interesting. Yeah. Yeah.
Speaker:And this is at and I can also tell you that I mean, and this
Speaker:is for me, it was, like, so sorry if, you know, a surprise because,
Speaker:Most of the like, let's talk about defects. So, yes, then what what is like
Speaker:recent type of fraud? Deepest, for sure. We had a report. I
Speaker:I think it, We published it 3 years 2 days ago or like
Speaker:yesterday on friends. So what's actually happening right now?
Speaker:And the thing is that deep fakes, They use usage of
Speaker:defects for fraud. It maybe it rest
Speaker:like 5 times. So like 2 years ago, like nobody actually
Speaker:knew so About defects. But now it's it's very easy to craft. It's
Speaker:very easy to craft. I mean, people like I mean, you are a fraudster. You
Speaker:have to actually, it's very rare
Speaker:prefer for you to just craft just 1 defect. It's usually something
Speaker:we call the serial fraud. You create like hundreds of defects. So now it's easy,
Speaker:very easy to create them. So now it's like a craft, like, hundreds
Speaker:of identities. And then I tried to bypass our security checks. So that's why this
Speaker:is like the recent trend. I mean, as so it's on the news,
Speaker:basically. And then we have to actually try to make sure that our solution,
Speaker:can detect it. And it's not sometimes, it's not that easy. Well,
Speaker:it sounds like, you know, there's there's stuff that people used
Speaker:years ago, and you've got that figured out. And it's probably not being used
Speaker:as much, at least alone. But now you've got,
Speaker:people coming up with, first, new ideas, and then second, they're
Speaker:doing combinations new plus older ideas. Is that
Speaker:accurate? But but, it is actually. And the thing is Okay. So,
Speaker:these are also like, Okay. Just imagine. We have a very
Speaker:sophisticated deep fake detector. So I I'm pretty sure that our, like,
Speaker:models are more or less, good. So, like,
Speaker:I mean, it's not 100% for sure. Mhmm. But what
Speaker:happens next? So can I actually, I mean, combat defects, 5
Speaker:years later? Maybe it's I'm so advanced. I so make like,
Speaker:our customers, like, ask us about it, like, once in a
Speaker:month. So what do you actually what is your plan, to talk about defects
Speaker:in 2 years. Right. Because now, you know, AI is like, it's very hard problem
Speaker:to solve. But here's also problem. There is a thing
Speaker:called mules. Have you heard about mules or money
Speaker:mules? This is, the the thing is
Speaker:that you actually go, hire a person.
Speaker:Usually, buy, pay some €50. And then
Speaker:actually this person passes a KVST check for you.
Speaker:And then Oh, wow. The person just sells sells here his or her
Speaker:account to you. And then this is a real person. I mean, it's not a
Speaker:defect. I found it that I could defect. Wow. It's not obvious and not defect.
Speaker:Yeah. But that well, this is that looks suspicious. But
Speaker:but I if I'm in a bank, I'm in a I'm a bank, for me,
Speaker:it's like a real person just trying to open up in a bank account. Yeah.
Speaker:And now we actually have to look around. So that's why so I
Speaker:like working with Deepgrams. I mean, it's very, you know, cool technology. You have to,
Speaker:like Yeah. It's technology. But Now you actually have
Speaker:to look around. You have to make sure what is, I mean, the
Speaker:pattern. What are the devices do you use? It's like lots
Speaker:of small Features or, signals, you have to actually
Speaker:combine or merge them altogether and then make a decision. Is it, like,
Speaker:specia or suspicious sorta? And this is like, but this is fun. This is
Speaker:like, you have to really look around, look collect lots of data, and then try
Speaker:to find, you know, your way into making a decision.
Speaker:Interesting. It's it's it's a fascinating the simple things are no
Speaker:longer simple. Right? Just signing up for an account, You know,
Speaker:it's just now it's become like this massive multinational worldwide
Speaker:cyber Security kind of exercise. It's a
Speaker:fascinating, Yes. For a
Speaker:customer, it is it must remain easy. Yes. I don't know like
Speaker:I mean, since, like even, you know, the really, really
Speaker:typical KBC check is includes recording your
Speaker:video. You usually have to do something like, you know, turn your head
Speaker:or something. I mean, if you have this experience. People do not like it. For
Speaker:them, it's like, why do you have to do this? That's it's it looks strange.
Speaker:I mean, just can I just open an account? And then it's like so it's
Speaker:also trade off unless you have to be simultaneously
Speaker:secure and busy. And this is Yeah. Those those are
Speaker:those are very much contradictory, forces. Yeah.
Speaker:Well, the other thing too, like, if I'm if I'm If I'm an average
Speaker:customer or paranoid me. Right? Like, if I go to a
Speaker:thing and they want me to look this way, look that way, Am I training
Speaker:their deep fake model of me? Do you know what I mean? Like, I mean,
Speaker:I'm kinda like, you know, obviously, I've done a lot of live streams and stuff
Speaker:like that, so I shudder Better to think what you know, where that could lead.
Speaker:But, what are your thoughts on that? Like, I mean, are do do you have
Speaker:people who are Do savvy customers
Speaker:do they get a little suspicious? Like,
Speaker:what are your thoughts on I'm not. I I
Speaker:must said that I mean, the defects that we see, they they
Speaker:can be crafted just for 1 1 image. Right. So like,
Speaker:here's the problem. So so like, there are, none of that, I mean,
Speaker:you can see them, but Usually, people send, you know,
Speaker:low quality images. So it's even harder for us to see it. Even harder for
Speaker:for human person for human to see that this is a problem.
Speaker:But there is also, I think, if I find a story that I
Speaker:know, that some of our models
Speaker:actually detect defects better than humans. So
Speaker:it's actually easier for a fraudsters to treat a leading
Speaker:person than a model. This model, like, can look back from certain artifacts with
Speaker:eyes or just, like, some sort of, you know, glitches.
Speaker:It's easy. But for person, especially the quality of the image is It's bad.
Speaker:It's like there is no way anybody can actually spot this is the
Speaker:problem. And this is great. It it is a problem. I I I must I
Speaker:must admit this is, I think, this is what we
Speaker:actually have to be have to hear
Speaker:about about creating deep fakes. I know that that
Speaker:is a very interesting thing. So, you know, about I mean, there are lots of
Speaker:things happening, around AR regulations, Especially in the
Speaker:European Union. Sure. And then so we actually tried to follow and then to
Speaker:make sure that everything is compliant. And actually, I wanted to say that we touched
Speaker:upon k y c KYT, which is know your
Speaker:transaction. There was also KYB and all your business, which is basically, you
Speaker:know, how we make sure that the company you work with is is
Speaker:I know fraudsters. And there is also a thing called k y
Speaker:a I, know your AI. And it says about transparency.
Speaker:So many people out there want to be to know actually how AI is used.
Speaker:So the k l it's it's a very new trend, I think. You have never
Speaker:heard about it because, I mean, it was going to be a week ago. Since
Speaker:I like, I want to actually know what's happening with all of this model of
Speaker:error, not just about touch prod, ground everywhere. But back
Speaker:to the problem with defects. The thing is,
Speaker:what to to say that,
Speaker:Oh, sorry. I lost the my my train of thought. But this is the all
Speaker:the time. Yeah. We I was just about to say that. But what you know,
Speaker:one solution to this, I I think, Pavel, would be
Speaker:if people did something, you know, like, I don't know, colored their
Speaker:hair Or grew a cool beard. I'm just
Speaker:throwing that out and with apologies to people listening and not
Speaker:watching. No. You know? I'm just
Speaker:saying. But but if you did but if you did
Speaker:grow a beard, would would or or or change your hair color or
Speaker:altered their face? Like, I know that, like, facial most facial recognitions
Speaker:use landmarks on, like, the eye sockets. Right. The a lot harder to change I
Speaker:was joking. Didn't mind. But, like, would it would it would that
Speaker:I don't know. Like, does that have any impact on these kind of systems or
Speaker:are they more like facial recognition systems? They are,
Speaker:it's, so we operate on the if you're talking about defect detectors or
Speaker:defect, models for defect detection. Yeah. There are
Speaker:some, I can't say that I face recognition. The
Speaker:models, they mostly focus on artifacts. So so for
Speaker:instance, like, a defect of a year ago, usually,
Speaker:had problems with eyes. Your eyes of a defect, they usually are
Speaker:very, you know, not really human.
Speaker:So it will be changed. It will be like as as as the technology,
Speaker:is getting more advanced. But like a few years ago, you can actually just crop
Speaker:Eyes of an image of a person, pretending to be a human person, then they'd
Speaker:make sure that this is actually a defect. Also I must say that
Speaker:Yeah. So a video is is is easier to detect because you can actually
Speaker:so, there is a thing called, I don't like the term in blindness because
Speaker:No, but nobody actually know what Linus is, but Linus is a detection.
Speaker:Linus detection is detection. If this is a
Speaker:leading person or not. And before, like, 5 years ago, it was
Speaker:mostly a distinction between, a video of a person or
Speaker:a printed out image. Now it's a detection of an image,
Speaker:defect, and the linear person. And at that time,
Speaker:you actually there are 2 types of fly misses. One tool that's passive,
Speaker:and we actually use also sometimes our customers actually ask us for
Speaker:pacifying. Let's adjust 1 image. But it's easier for
Speaker:us and for everybody else to ask a person to actually do something.
Speaker:And for defects, for instance, like, if I ask them to rotate, Sometimes some
Speaker:artifacts can appear. Some artifact. And then you can actually see that probably. I
Speaker:mean, this is not the only person. There are some sort of problems with visual
Speaker:artifacts. So it is it is like this.
Speaker:Also, I must say that there was also a challenge for us because there
Speaker:are, certain cameras. They have some sort of a
Speaker:beautifiers. So I'm pretty sure as I'm calling from my,
Speaker:my computer, and then my camera actually
Speaker:Advances my image. So my image is a little bit, better
Speaker:than I'm in the real life. So my my skin is is is a little
Speaker:bit better. So it's it is actually, Embedded into
Speaker:hardware. And for us, it looks like, some sort of, you know so there is
Speaker:a signal for us. It does some sort of, you know it's Oh, I see.
Speaker:So It's hard. You know? And you have to make sure that make sure that,
Speaker:okay, it's not defect. It's just the person using that, camera off my,
Speaker:computer. It's like, you know, you have you have to be really, a
Speaker:yellow error. Apple, I mean, installs
Speaker:another camera, and then you have to be actually tune your models to make
Speaker:sure that you actually do not penalize people from with
Speaker:I think about that. Yeah. The cameras are gonna behave differently if you use different
Speaker:cameras. So I'm here using my 4 k,
Speaker:camera. Kind of an outdated one, but it's still it does the job. But what
Speaker:if I pick up my droid Or, you know, my wife
Speaker:my wife, you know, she's the the device. She's got an
Speaker:iPhone. And if I'm trying to log in through her device, That would be different
Speaker:images, and it may change. You know, it may tell me, nope. That's not
Speaker:you. Those are gonna be different artifacts. That's fascinating. And I also
Speaker:think it's funny that you have an old four k camera, which
Speaker:is a pretty funny thing to say. Like For for podcasting, I won't
Speaker:No. I know. I don't wanna throw back to, theme from yesterday's
Speaker:2 hour show, but I'll just make this note. We we
Speaker:learned that we're in the top 2 a half percent of podcasts.
Speaker:So now I feel like I should have, I don't know, 16 k studio
Speaker:and Yeah. I should have a lot of time like Joe Rogan has in a
Speaker:brick wall. Exactly. Right. I don't I need something better than this
Speaker:old four k camera. But
Speaker:if all of a sudden You just want to open a bank account right
Speaker:now. Yeah. It looks strange because, I mean, a typical person is like you
Speaker:use your iPhone or you're like a regular computer. Like, with 4 k or 16
Speaker:k camera, it's like very strange. It's some something, you know. It's it's a signal
Speaker:for for every model and make sure that It's an outlier. Right? And
Speaker:it sounds like a big this is still obviously, there's way
Speaker:more complicated things than what you do, But outliers
Speaker:detecting outliers is probably 1 1 big tool in your tool belt.
Speaker:It is. Yeah. That's very hard if you have a Genuine person,
Speaker:and you are an outlier somehow. I mean, everybody can be an
Speaker:outlier in some sense. It's very hard because, yeah,
Speaker:So this is hard. So, like, at some point, yeah, colored hairs
Speaker:can be also an outlier. I don't No. It's just interesting. So I imagine, like,
Speaker:Instagram filters and things like that probably also cause
Speaker:chaos and things like that. Yeah. Of course. But, yeah, I
Speaker:mean So usually use, yeah, filters,
Speaker:a strong signal for us. I mean Right. And also I must I must have
Speaker:this defects. So going back, thing with defects is that
Speaker:it's not, like, specifically use the fraudsters. Here's the
Speaker:problem. You know, there are lots of cool things for defects. You can press
Speaker:advertising. Right. I don't know what what else. But, usually,
Speaker:you can actually adopt a person to, like, Replaced an
Speaker:actor in the movie. This is also a defect. It's a very cool defect, very
Speaker:sophisticated defect, very high quality defect. Still a defect. So those
Speaker:are our usage is actually for for that, I mean, not just for fraud.
Speaker:And then going back to our problems, it's like, I mean, And the
Speaker:even even that and even that from that, I like this example,
Speaker:but, the guys from the,
Speaker:I mean so we focus on financial fraud. Yeah. So it's more or less like
Speaker:people trying to actually sue money on, like, take over your account, something like
Speaker:that. But the thing is the defects, they are mostly created
Speaker:not for that. And this is a very interesting thing, I think. They are created.
Speaker:And, actually, I didn't know about that, but we actually knew that When they started
Speaker:to try and to create our Deepak's. So we went, you know, to the Internet,
Speaker:some strange forms to make sure what what people actually use
Speaker:What they create deep eggs for. And they create
Speaker:deep eggs for porn. It's like 98%, 89%
Speaker:Deepex, I slide 4. And this is also a problem because in in there is
Speaker:a thing called nonconsensual port. Deepex are used for that, And this
Speaker:is also a problem. So it's not our business, but the thing is that the
Speaker:same technologies is there. And you actually I mean, if you,
Speaker:I mean, work in the area, you can actually so the same model can actually
Speaker:be applied to detect, this type of defects. Right. So it's
Speaker:different, but, I mean yeah. Yes. It's, That was expressed to
Speaker:me maybe a year ago. It's fascinating how
Speaker:quickly this space is just Evolving or
Speaker:devolving, I guess, depending on your point of view. Yeah.
Speaker:But, no, you're right. Like, most of it is
Speaker:Those a lot of the deep fake kind of work is done
Speaker:for adult content. And, you know, and it's there
Speaker:the The legislation around this is gonna vary
Speaker:widely from place to place. But, like, you know,
Speaker:revenge porn laws don't apply. And there. I I think that was a big thing
Speaker:in, and there was a controversy somewhere. I think it
Speaker:was New Jersey, Where somebody had
Speaker:created deep fake images of either high
Speaker:school or middle school girls, which adds an extra level of
Speaker:legal Concern I have a whole lots of extra
Speaker:levels of concern. Let's be honest. But, like, you know and and and and
Speaker:there was this, you know, the big debate. And my first reaction was, I'm
Speaker:actually kinda surprised it took this long for that to happen,
Speaker:which is a very cynical take, I'll admit. But I can tell I I can
Speaker:tell you the reason. The thing is that Technology moves so fast. Yes. And
Speaker:legislation actually is always, like
Speaker:so even with with EAU, AI act,
Speaker:those I mentioned defects just a little because they started working on
Speaker:the regulations 2 years ago. And 2 years ago, it was not a problem.
Speaker:And now it's, like, all over, you know, the Internet, and then you have to
Speaker:actually tweak the, wording,
Speaker:but it takes time. Well, even still, like, you know, like, there's,
Speaker:a few months ago, they had these fake commercials that were created by with
Speaker:combination of 11 Labs and A few other companies to name them, so I
Speaker:forget. But, you know, they had a picture of Elon Musk, you
Speaker:know, eating spaghetti, and it looked weird. But you can easily see,
Speaker:like, You know, I was messing around with v q early versions of v
Speaker:q grant d q GANs in early
Speaker:2022, And that stuff looked
Speaker:weird, and it it really evolved. And this morning, I saw
Speaker:Pika AI, I guess, just went Yeah. Yeah. Yeah. Went to a wider beta.
Speaker:And, yeah, released and and and, like, I'm seeing what's created with that,
Speaker:and, you know, it still looks weird, it still looks cartoonish,
Speaker:but it's not The fact that we've gone that far in the span
Speaker:of, you know, less than 2 years, like, I think says something, like and to
Speaker:your point, legislation Usually takes years, to
Speaker:make. So, like, by the time these laws are written, they may not be valid.
Speaker:In the case of New Jersey, I think there's some debate over,
Speaker:does what sorts of laws that applies to? Because
Speaker:the the original, The faces
Speaker:were mapped on to something else, but that the
Speaker:something else I'm trying to keep our clean rating here. The something else were
Speaker:people over 18, but the bases were mapped onto it. So there's
Speaker:some debate over, do existing laws cover that?
Speaker:I'm not a lawyer. Don't look at me, and I'm not. But,
Speaker:it's just fascinating to your point. Like, this is moving quickly.
Speaker:Yep. It's definitely complicated. So we've
Speaker:reached the point in our show, Pavel, where we, like to
Speaker:ask a set of questions. They're in the chat. And
Speaker:I'll start out, with the, the very first question.
Speaker:How did you find your way into this field? Did this field find you,
Speaker:or did you find it? Yeah. I must say I have a
Speaker:story to tell. I just studied yeah. Studied computer
Speaker:science at, university And I actually worked as a software engineer
Speaker:at Motorola. You may remember this company, with
Speaker:HQ in Chicago back then, for 5 years.
Speaker:And then it was, 2011, which is, like, long time
Speaker:ago, the very first, massive
Speaker:online courses appeared. There was a one called AI class,
Speaker:and it later turned out to be a Udacity. And there
Speaker:was also a m l called ML class. It's a ML class. And
Speaker:this now this Coursera. It's like 10 years ago. And I was like, okay.
Speaker:Cool. I enrolled and actually, I pushed because it is like it was it was
Speaker:hard. It was like, you have to really, be involved. And
Speaker:then I felt like, okay, this is a cool thing. This is like a next
Speaker:big thing for me and, like, for everybody else. It was like
Speaker:12 years ago. So I quit my job, and I actually, so
Speaker:at the same time, I started to try to run a small startup with my
Speaker:friend, failed miserably. But I take, took my time, studied,
Speaker:for maybe half a year, and then joined a small data
Speaker:startup as a data scientist. And then it just
Speaker:started there. So it's I think I I find, my way into
Speaker:data. But Yeah. I don't know. So You want to
Speaker:I'm sorry. Go ahead. I just I just say it sounds like you were very
Speaker:intentional about finding your way into it. So that's cool. Yeah.
Speaker:That's cool. And I see you were You were at VK for a while too,
Speaker:which I've never seen VK, but I hear it's like a like
Speaker:a Russian language version of Twitter slash Facebook. It used
Speaker:to be. Yes. Yeah. Yeah. I don't I yeah. Obviously, now things are different, but
Speaker:yeah. Yeah. Yeah. Yeah. I worked there for 5 years, a long time ago. Oh,
Speaker:interesting. And, you know, if you're talking about the data, I mean,
Speaker:the, where it's like the the place where you can
Speaker:actually play with data. You can actually cool do many cool things.
Speaker:Oh, yeah. Nice. Nice. And he's being modest. According to LinkedIn, he was director
Speaker:of AI research, so he's super smart.
Speaker:But, what's your favorite part of
Speaker:your current job? Oh, I can't say it
Speaker:could create some defects, but, it's not
Speaker:it. I think
Speaker:no. I mean, I would say that what I like is, they,
Speaker:the the Samsung, Samsung is is now it's it's a product or any company. So
Speaker:have our own own products, whether, like, a technology company, yet we have our
Speaker:own product. And having that,
Speaker:actually, our own product, Actually helps us, you know, I know what our
Speaker:customer wants. Wonderful. I know the
Speaker:data. So it's like, you know, I mean, you have to actually so you have
Speaker:to look around. Okay. There is a problem with defects. I have to,
Speaker:like, make sure that I mean, I had, I actually have to understand this. This
Speaker:is a problem. And for many of our customers, I mean, I
Speaker:don't I would not like to say that we have to educate them or actually
Speaker:make make sure that they understand this is a problem with defects. And now we
Speaker:have when they understand, we can actually help them with their their,
Speaker:safety and security. One thing that this is, like, a little bit, I
Speaker:mean, Clumsy answer, but I'm sorry if you know.
Speaker:Yeah. Being closer to the product is is is is fun.
Speaker:Oh, sorry. Cool. So we have 3 complete
Speaker:sentence. And the first one is when I'm not
Speaker:working, I enjoy blank.
Speaker:Okay. Okay. Let me think for a while. There are many things I can
Speaker:say. No. I can say no. This is I think of this as I can,
Speaker:I can share? No. I I I I run or I can see job.
Speaker:Mhmm. Oh, cool. Cool. I run-in the the ring marathon.
Speaker:This is my Nice. There are Major Martins, like, 5,
Speaker:6 Martins across the world. So that's New York, Paris,
Speaker:London, Tokyo, Berlin, and,
Speaker:London. Nice. Like, 6 so that Very So Berlin was my 1st major
Speaker:marathon. So I ran it, this this September, and it was great. No. That's
Speaker:awesome. That's awesome.
Speaker:When you said Berlin, the first thing that popped in my mind was, Berliner
Speaker:Kendall wrote, which is like this local kinda drink.
Speaker:Yeah. Yeah. Yeah. Yeah. I know. That's like Yeah.
Speaker:Yeah. But I prefer there is a it's a vehicle. It's
Speaker:like a craft. Oh, yeah. From Berlin.
Speaker:Right. But I talking about Berlin, so I run. It was
Speaker:super fun, but, on my finishing picture, so
Speaker:it's my me, Ryan. So close to Bernsberg. It's a
Speaker:very central grid. Mhmm. And there is also a guy in the
Speaker:bottle question. And and I
Speaker:wasn't it was not slow. I wasn't slow. Yeah. There was a guy in a
Speaker:huge ball, like, I still running, like, finishing with me. Like, so it was, Oh,
Speaker:that's funny. That's fun. It's that's fun. That's funny. Very cool.
Speaker:Next, complete the sentence. I think the coolest thing in
Speaker:technology today is
Speaker:blank. Oh, it's it's it's hard to say. Let me I'll just
Speaker:think for a while. But, I mean,
Speaker:I think that so my my area
Speaker:seems like I expert a personally specified natural
Speaker:language processing. So I know about language models. And,
Speaker:actually, we had papers on language models, like, before they they
Speaker:were super big. So, like, on tuning language models. Yes. I
Speaker:found it really, really exciting that it in a
Speaker:year, it went from, you know, research
Speaker:Prototypes to, like, everyday product. This is Yeah. This was
Speaker:a compelling. So, like, my parents used Chargebee PCs. Like, I mean, this
Speaker:is like this is like a mobile phone. This is I mean, this is what,
Speaker:like, some sort of a milestone, last year.
Speaker:I think this is this is it. And he is that the actual unit
Speaker:for main things. You can build products on on language models. And
Speaker:this is also like. It's wild, isn't it? Like, you know,
Speaker:and and it's captured everybody's imagination in in good and bad ways.
Speaker:But, like, my father-in-law, you know, So he used to
Speaker:say Frank works with computers. Now he says Frank works in AI.
Speaker:Okay. You know? That's good.
Speaker:But I also like we used to say machine learning. So now you have to
Speaker:say AI. That's right. That's right. You have to say that data mining
Speaker:core something. So it's like, you know That's right. It definitely would.
Speaker:I wonder what it'll be next year. Who knows? Gen AI probably.
Speaker:Probably. So our next one, complete this Regulate, I think. Oh, that's
Speaker:right. Regulation. That's right. Regular. Our our last completes the
Speaker:sentence is I look forward to the day when I can use technology
Speaker:to blank. Uh-huh. I
Speaker:can't it's hard to answer because, I mean, like, I
Speaker:can't say it would be cool If I can, you know,
Speaker:develop drugs. And then there are very cool startups for drug design
Speaker:with AI. Yet, I mean, Just imagine we have
Speaker:a a a cure for cancer, but Right. We have so
Speaker:many diseases to care to cure. So let's say, I think I
Speaker:hope Once we fix anything, then there is gonna
Speaker:be a next, you know, next milestone for us to look forward. So I'm sorry
Speaker:if, you know, there's never I hope there will be
Speaker:no such date, I can say. Right. Right. That's
Speaker:a good one. I'm pretty sure you will agree with me. Like Yeah.
Speaker:Especially work with the technology. I mean So true. For sure.
Speaker:The next question, share something different about yourself, but remember, It's a family
Speaker:oriented well, not family oriented, but we like we we like it so that
Speaker:you can list it with your kids in the in the car. Right? Like, That's
Speaker:kind of a Yeah. Yeah. Yeah. And, yeah, and I live in Berlin across, very
Speaker:close. There's a very, how to say, kinky club, which is Berlin.
Speaker:Was that the the the tier garden? It's it's
Speaker:it's a it's a it is family friendly. It's it's like the most family
Speaker:friendly place in in in Berlin. You got some. Yeah. No. It's it's
Speaker:called KitKat. Yes. What I can say.
Speaker:I have, purple hair. Since last month.
Speaker:I don't know. So I can say that I speak
Speaker:a few languages, all of that. But, no, I'm I'm
Speaker:joking. So I speak Japanese. I don't I don't Japanese, for a
Speaker:long time. So I I can speak Japanese. I speak
Speaker:English, obviously, Russian. My parents are from
Speaker:Russia. And I also speak German. So I actually Studied
Speaker:German for 2 years. So I actually studied right now. So I had, like, my
Speaker:German classes 3 or 4 times per week, which is
Speaker:let me just go. Sorry. So I hope in a year, I will be able
Speaker:to do a podcast in German as well. Oh, Wendeschon. That is
Speaker:not
Speaker:Yeah. Yeah. And we just lost, like, We we just
Speaker:looked at our analytics, and, like, most of our listeners are from English language countries.
Speaker:So I think we just lost them. Maybe we
Speaker:can attract new listeners. Oh, I like it. I like the way you think. We
Speaker:wanna we wanna get to the top 2.4% now.
Speaker:Our new goal. So,
Speaker:Audible is a sponsor of the show, and I'm not sure if
Speaker:Audible is big in Europe. I think it is because I've seen a lot of
Speaker:German language audiobooks. It is a no. Okay.
Speaker:So do you do you listen to audiobooks? And if so, you have a good
Speaker:recommendation. Otherwise, we'll take a recommendation on the regular good Fashion
Speaker:paper dead tree book. No. I have a couple. I think I can
Speaker:give you a couple of examples. This is like,
Speaker:I like this was the most, you know so so I'm so my
Speaker:background is from many, places,
Speaker:since Israel, Russia, and Germany in some extent. So
Speaker:I would recommend, there is a very Good book. It is
Speaker:in my opinion, this is very known, but not many people know about it for
Speaker:some reason. It's called the good soldier's make. Okay.
Speaker:Like it said, didn't Not heard of. About the, sort of
Speaker:third world war by Oh, interesting. But this is
Speaker:it's very good. Like, you can actually learn a lot about
Speaker:Czech Republic, Germany, Austria in the beginning of
Speaker:the, Last century. Oh, interesting.
Speaker:Especially now, it's the very thing. It's called in the park. This is a very
Speaker:good thing too. And it's very funny. It's like one of the funniest, books
Speaker:ever written. And also the the second one, I have 2.
Speaker:This called Arc of Triumph, by remark. Okay.
Speaker:This is also about the pre war Europe, pre second World War
Speaker:Europe, like, Southeast, years of the
Speaker:last century. And this is also very, like, you know, you
Speaker:really you really feel like what what was the I mean, living in
Speaker:Germany and, France, during that time, it's very, very interesting.
Speaker:So one of my favorites. So I can definitely recommend both of these
Speaker:videos. Very cool. So audible detecting I'm sorry. I'm
Speaker:detecting a history theme. Yes. Yeah.
Speaker:Yeah. Yeah? Cool. There's a really good book. Since you live in Berlin,
Speaker:you might like it. It's called Faust's Metropolis, and
Speaker:it's about the history of Berlin from, like, you know, Almost
Speaker:stone age time till Okay. Cool. You know, the 20
Speaker:you know, early 21st century is kind of like And the basic
Speaker:gist is, like, you know, a lot has happened in Berlin. Good.
Speaker:Sure. Yeah. We all know the bad. Right? But, like, some good things have
Speaker:happened, kinda everything in between. It's kind of it's an interesting look at, like, the
Speaker:history of the city and how it apparently was built on a swamp or something
Speaker:like that. Like Yeah. It's just, it's it's
Speaker:interesting. And Audible is a sponsor of
Speaker:Data Driven. If you go to the data driven book .com,
Speaker:I think even the data driven book .com might work. Uh-huh. That was
Speaker:a pronunciation joke. You'll get a free, on 1 free
Speaker:audiobook on us, and And we'll get a kickback if you sign up for a
Speaker:subscription. And finally, where can
Speaker:folks find out about you, more about you, and what you're up to at Sumsub
Speaker:And, some of the other things you you're up to.
Speaker:What's up? My my connection was, Oh, where can folks find out
Speaker:more about you and what you're up to? Oh, yes. It's,
Speaker:yes. It's, it's a company. It's called Samsung. So Samsung dot com.
Speaker:Also, like, what we have is, today is
Speaker:with anti fraud. And you have to I mean, It's not about
Speaker:all the product. It's actually about making people helping people
Speaker:learn about, security. So how they can actually navigate the Internet or,
Speaker:like, their life More safely. So we have a portal called
Speaker:some suburb where we actually post a lot
Speaker:of stuff on Making your Internet life,
Speaker:can I say like this, safer? So, actually, I I advise you
Speaker:to take a look, and then probably you'll find something interesting there.
Speaker:We definitely will. And, any parting thoughts before
Speaker:we end the show? Any final thoughts? I just
Speaker:want to say, yeah, Just I was very happy to, to be here
Speaker:and hope, it was Cool. Interesting. This is a great show. It's always good to
Speaker:it's always good to kinda understand The the the intersection
Speaker:of of AI data and security because some people still see
Speaker:those as separate things. But I think as time goes on,
Speaker:we're gonna I'm gonna we're gonna wonder how we ever saw it as separate
Speaker:things. There are so many things to talk about that. Yeah. Yeah. Yeah.
Speaker:Yeah. Well, awesome. Any parting thoughts, Andy?
Speaker:No. Just a great show. Pavel, thank you for, for joining us.
Speaker:It was our honor. Yes. Likewise. And we'll let
Speaker:Bailey finish the show. That was some show.
Speaker:We appreciate you listening to Data Driven. We know you're
Speaker:busy and we appreciate you listening to our podcast. But
Speaker:we have a favor to ask. Please rate and review our
Speaker:podcast on Itunes, Stitcher, or wherever you subscribe to
Speaker:us. You have subscribed to us,
Speaker:haven't you? Having high ratings and reviews helps us
Speaker:improve the quality of our show and rank us more favorably with the search
Speaker:algorithms. That means more people listen to us,
Speaker:spreading the joy. And, can't the world use a little
Speaker:more joy these days? So go do your part to
Speaker:make the world just a little better and be sure to rate and review the
Speaker:show.