Deepfake Spoofing & Liveness Detection | BioID Biometrics Presentation @CODE2021 | Deepfakes Attack - YouTube

Channel: unknown

[4]
So thank you for the introduction and also the聽 presentations we heard before as Dr Wolf said,聽聽
[10]
i'm coming from or we are coming from a different聽 perspective, because we are a solution provider聽聽
[15]
in the field of identity verification and so I聽 will be speaking about the use case and also the聽聽
[21]
challenges concerning liveness detection and聽 deep fakes if it comes to a remote identity聽聽
[27]
verification. So to begin with i think we already聽 know everyone sees it and we've also heard it聽聽
[35]
today, that digital identity verification is on聽 the rise on the one hand this has been a trend聽聽
[40]
since globalization and due to globalization but聽 of course Covid-19 has actually increased the聽聽
[48]
remote verification necessity by a large scale.聽 I'm just thinking about one of our customers who聽聽
[56]
provides identity verification to the British聽 government and they used to have a solution聽聽
[62]
for more than one and a half years but then聽 last year during the first lockdown they had聽聽
[67]
to scale up their service enormously, because聽 all the branches and for the British government聽聽
[73]
services were closed, so everything had to move聽 online and I think we see that all over the world聽聽
[78]
and especially in Europe. I think remote identity聽 verifications have risen enormously and now we see聽聽
[87]
a diversified market as not only we have聽 the video ident services with video agents聽聽
[95]
verifying identities remotely but also automated聽 systems and this is also where BioID comes from聽聽
[102]
and this is the reason I will be focussing on聽 automated systems today, but all the things I聽聽
[108]
will be saying are also applicable to analyzing聽 videos and video agent services of course.
[118]
For an automated identity verification of course聽 you need to perform a few steps. One of them is to聽聽
[124]
make sure that the idea is authentic, we've heard聽 that in the first presentation as well, it could聽聽
[129]
be manipulated the picture could be manipulated聽 or the ID could be stolen. So sometimes Interpol聽聽
[136]
databases have to be checked and things like that聽 so a user trying to verify their identity auto or聽聽
[142]
remotely they would be asked to take a picture of聽 their ID so that it can be analyzed by algorithms.聽聽
[149]
The second step is that it has to be made聽 sure that this ID actually belongs to the聽聽
[154]
person who submits it and for that biometric聽 face matching is performed. So we need a very聽聽
[161]
accurate mechanism to make sure the ID picture is聽 the same as the person looking into the camera.聽聽
[168]
Which brings us to the third step for remote聽 identity verification, that is liveness detection.聽聽
[175]
It's also called presentation attack detection and聽 it means that we need to be sure that the images聽聽
[181]
are actually derived from a live person and聽 not from some type of spoof and then if all聽聽
[186]
those three checks were successful, we can聽 say that verification has taken place and聽聽
[191]
we have trusted credentials,. But of course聽 the question comes up what about security,聽聽
[198]
because there are different steps as I said聽 that can all be manipulated, so there need to be聽聽
[203]
security mechanisms making sure for instance the聽 holographic is genuine and the ID is not stolen.聽聽
[211]
That the face is really matching the one on the聽 ID and as I said that no spoofing is taking place.
[219]
So this brings me to the main topic of the聽 presentation, which is liveness detection.聽聽
[224]
What does liveness detection have to do in the聽 use case scenario that we are now speaking of,聽聽
[230]
which is remote identity verification. First it聽 has to reliably detect life versus fake input,聽聽
[237]
but it also has to do that in real time and聽 easy to use and also hardware independently聽聽
[244]
because if we think of the British聽 government or the British citizens,聽聽
[248]
we never know which kind of device they use,聽 in which scenario they are, we don't know聽聽
[253]
if the old lady maybe doesn't have so much聽 experience with doing things like that. So we need聽聽
[259]
to be able or the algorithm needs to be able to聽 use any device, any camera, any lighting and still聽聽
[265]
be very reliable with its decision for whether聽 or not the images were coming from a live person
[276]
So which spoofing attacks need to be addressed?聽 Of course we heard some today already and saw some聽聽
[283]
pictures. 2D photos and things like image swapping聽 with 2D pictures need to be detected. Also 2D cut聽聽
[291]
outs, like I have one here for instance. I think聽 you can see that the eyes are open. So spoofing聽聽
[299]
with things like these masks need to be rejected聽 of course. Then on the other hand, we also have聽聽
[305]
3D attacks, so different types of 3D masks,聽 wax heads and sculptures and things like that,聽聽
[312]
but also a very big challenge are 3D paper聽 masks and I thought i'd bring you this one.
[321]
Of course we have a lab here and we manufacture聽 spoofing attempts to make sure our system is able聽聽
[327]
to cope with them and yeah, so this is one of the聽 more sophisticated 3D paper masks and you can see,聽聽
[335]
it kind of looks like Trump, I would say and聽 also it has some skin close texture and it's聽聽
[342]
three-dimensional, so things like that have聽 to be addressed and also have to be detected,聽聽
[347]
because although it's not easy to build these聽 masks, it's still available for people, if they聽聽
[353]
well if they try little harder and then the third聽 thing we need to address are video replays and聽聽
[360]
video projections on different materials, because聽 that's actually the most common type of fraud we聽聽
[365]
see in the real life scenarios. That people try to聽 use either grandma's picture or a wedding picture聽聽
[372]
or some kind of video from social media from聽 the person like now this video for instance is聽聽
[378]
recorded from me and so that means it will be聽 available on social media and people could use聽聽
[384]
it to try to spoof systems using this video of me聽 and of course we have to be able to detect these.
[391]
But instead of just speaking about it, i thought聽 it would be nice to show a demo. Um excuse me.聽聽
[399]
So you can see that this person is actually聽 detected a fake and the reason for that is聽聽
[405]
that it was a silicon mask and when I took off聽 the mask I was detected as live in real time.聽
[412]
This was a live demo by the way here聽 in our office, maybe half a year ago,聽聽
[417]
so that's what it when it was recorded yeah聽 and you see that i try to bend the picture and聽聽
[423]
try to overcome the system by kind of聽 making realistic features available but聽聽
[430]
yeah in this case of course it worked to detect聽 a live person as live and the fake person even聽聽
[437]
here with the video where it looks like me and聽 it moves like me it's still detected as a fake.
[451]
And the question is, how do we do that? One of聽 the steps so we combine two different elements聽聽
[457]
and one of the steps is 3D structure analysis,聽 meaning that we perform optical flow analysis on聽聽
[464]
2D pictures, so we need to work with standard聽 cameras, as I said for this real life use case聽聽
[470]
scenario of remote identity verification, we聽 don't have any depth information or 3D cameras,聽聽
[476]
so we need to be able to be sure that it's a 3D聽 geometry we see with just using 2D rgb standard聽聽
[484]
cameras or pictures and we do that by asking the聽 person to slightly move and then from the slide聽聽
[491]
movement we can calculate this 3D geometry. This聽 is a very good way to prevent 2D spoofing attacks,聽聽
[500]
but then we also use texture analysis and AI and聽 we on the one hand have handcrafted features for聽聽
[506]
the texture analysis, but on the other hand聽 we use deep convolutional neural networks,聽聽
[511]
which are trained with our data to make sure,聽 that the live people are as I said detected as聽聽
[517]
live and the fakes are rejected. Of course聽 this is a very complicated problem, because聽聽
[522]
for a two class problem like live or fake you聽 kind of need all types of fakes in the fake聽聽
[530]
class for training and you are never sure you can聽 never be sure that your system generalizes well聽聽
[537]
enough to also cope with unseen attacks. So this聽 is an ongoing research and development project,聽聽
[543]
but it's already in use and very reliable,聽 but it has to be modified all the time.
[552]
So this brings us to the topic聽 of new unseen attacks or new聽聽
[557]
kinds of fraud attempts, which is deep fakes.聽 So deep fakes, as we already heard, are becoming聽聽
[566]
more challenging, because they are difficult to聽 distinguish from real videos, because they are聽聽
[572]
getting so good in quality so the human eye really聽 has problems to distinguish them. Especially if聽聽
[577]
they are not trained and of course the algorithms聽 also can have challenges with deep fakes.聽聽
[586]
So what is the status quo if it comes聽 to deep fake detection? And here聽聽
[592]
the thing is, that deep fakes typically, at least聽 for this use case that i'm speaking of, which is聽聽
[597]
remote identity verification, here deep fakes are聽 presented on a display so for instance, a video or聽聽
[605]
a deep fake, a recorded deep fake, is presented on聽 a cellphone to a camera which tries to verify the聽聽
[611]
identity of a person by sending the pictures聽 to us for instance to the solution provider聽聽
[618]
so if deep fakes are presented at the sensor level聽 then the good thing is that our standard liveness聽聽
[624]
detection and the texture analysis and the AI are聽 able to detect this, because as i said, they are聽聽
[630]
presented on a display, they are projected to some聽 kind of material and our liveness detection would聽聽
[636]
detect all of these materials and displays. So聽 the good news is, that for this use case scenario聽聽
[645]
a secure app preventing video injection聽 solves the problem of deep fakes,聽聽
[650]
because if deep fakes can't be injected into聽 the application, but are only presented,聽聽
[655]
then our liveness detection already is able聽 to cope the situation, cope with the situation
[663]
But still we need to speak about video injections聽 of course. So what about deep fake attacks at the聽聽
[668]
application level? So here, if we are speaking聽 about a pre-recorded video feed, a pre-recorded聽聽
[675]
deep fake, then challenge response mechanisms聽 can be used, so the user would be asked to聽聽
[682]
move in a certain randomized direction,聽 like up and then to the side for instance聽聽
[687]
and it's very unlikely that an attacker really has聽 a video that where or a deep fake video where the聽聽
[693]
person moves in this prompted way. So this can be a way to聽聽
[699]
detect deep fakes, even if they are, if聽 the attack is at the application level.聽
[705]
But what about live manipulated video feeds, so聽 especially for all the use cases, for all the聽聽
[712]
scenarios where we're not in the used case of聽 identity verification, but maybe a politician,聽聽
[718]
a video of a politician needs to be analyzed for聽 instance. We are also we also want to be able to聽聽
[725]
detect whether it was a deep fake or not and live聽 manipulated video feeds that are injected into an聽聽
[731]
application, like identity verification of course聽 are a challenge and future research is relevant.
[739]
Which brings me to the FakeID project, that we聽 have already heard about today. So our future聽聽
[746]
research aim is to develop methods to detect deep聽 fakes in the video and photo material directly,聽聽
[754]
this is in order to make identity聽 verification more trusted and be聽聽
[758]
sure about the credentials that are derived聽 from such automated processes and support all聽聽
[764]
the different cases where we want to analyze聽 video material and be sure that it's genuine
[772]
And to sum it up, I would just like聽 to say a few words about BioID.聽聽
[777]
So we are a company focused on research and聽 development which comes from our history.聽
[782]
We've been founded in 1998 as a spin-off from聽 the Fraunhofer IIS here in Erlangen and we are,聽聽
[790]
we have our headquarters here in Nuremberg聽 and are a German company with 100% proprietary聽聽
[797]
software and we are proud to say that we have 23聽 years of experience in biometrics, so we've been聽聽
[802]
on the topics from the beginning um yeah and are聽 happy to be able to bring the technology into the聽聽
[810]
use cases now, that we've been speaking about. So thank you for your attention and I actually聽聽
[817]
hope there are some questions and if not now,聽 then please don't hesitate to contact me later on,聽聽
[823]
after the conference and聽 yeah thank you for having me!