MINDWORKS

Deep Fakes: With Vladimir Barash and Laura Cassani

Daniel Serfaty

How do we distinguish fact from fiction in an age where seeing is no longer believing? 

Join MINDWORKS host Daniel Serfaty as he explores one of the most intriguing and controversial developments of artificial intelligence today – Deep Fakes – with Vladimir Barash of Graphika and Laura Cassani of Aptima. 

Daniel Serfaty: Deepfakes, what a scary combination of words indeed. Are deepfakes a looming threat to the truth as we know it? How do we distinguish facts from fiction in an age where seeing is no longer believing? Hello everyone and welcome back to MINDWORKS, where we delve into the fascinating word of technology, innovation and the future. This is your host, Daniel Serfaty, and today we have an incredible, insightful episode lined up for you. We will be exploring one of the most intriguing and controversial developments of artificial intelligence today, something we called deepfakes.

For those of you who might not be familiar with it, deepfakes are AI-generated synthetic media where a person in an existing image or a voice or video is replaced with somebody else's likeness. While this technology showcase the remarkable advancement in AI, it also brings, as you can imagine, a myriad of ethical, legal, technological, and societal challenges. And to help us navigate through this maze, we have two very special guests with us today. Both are expert in the field of artificial intelligence and deepfakes, with an impressive background that spans across academic and professional realms.

Dr. Vladimir Barash is a chief scientist of Graphika. Vlad received their PhD from Cornell University in information science and wrote their thesis on the flow of rumors and verily marketed products through social media networks. At Graphika, Vlad oversees research into social network analysis and large-scale social phenomena, which powers Graphika's cutting-edge analytic engine. Under Vlad's leadership, Graphika's research wing has successfully completed numerous multi-year grants to study viral phenomena in social media, disinformation, and AI-generated content, as well as causal mechanism for online behavior. In 2019, Vlad testified before the US House of Representatives about influence campaigns targeting US veterans and military service members.

Our second guest today is Laura Cassani. Laura is a deputy director of intelligent performance analytics division at Aptima Incorporated, so she's also my colleague and focuses on innovation in the development of and applications of artificial intelligence technology to improve human and machine performance. That's a mouthful. As principal research engineer at Aptima, she has overseen a portfolio of projects focusing on operationalizing AI approaches to drive human machine systems. Here we go again about human AI teaming. Laura has led multiple successful R&D efforts for the Department of Defense, the Office of Naval Research, the Defense Advanced Research Project Agency or DARPA, the Air Force Research Lab and the Marine Corps. More recently, Laura has focused on leading projects involving large language models and generative AI.

This episode is crucial for anyone interested in the future of technology, the ethical use of AI, and understanding how we can prepare ourselves for the challenges and opportunities that lie ahead. So sit back, relax, listen and join us as we unravel the world of deepfakes. Vlad and Laura, welcome.

Laura Cassani: Thank you. Excited to be here.

Vladamir Barash: Thank you.

Daniel Serfaty: All right, let me start. I tried maybe unsuccessfully to explain what deepfakes are in my introduction, but I would like you to spend a minute each to tell our audience what is a deepfake. Vlad, let's start with you.

Vladamir Barash: Yeah, I think you got it basically right, Daniel. A deepfake usually looks like media of real people or situations. Probably a well-known example is a deepfake video of Tom Cruise playing the guitar. There's no actual such video ever recorded. It was created by an AI from scratch. More recently there was a deepfake of Katy Perry at the Met Gala. This was an image that was not actually taken by a real camera, and yet it was convincing enough that it fooled her mother.

Daniel Serfaty: Well, if you can fool a mother, you're dangerous. Laura, you want to add something to Vlad's? Very cogent definition of deepfake From your perspective, how do you see that?

Laura Cassani: I think that's a good summary. As Vlad was speaking, I was thinking about all the new developments in this space. So there's a number of different kinds of deepfakes and different technical approaches you can take to create them from fully synthetic video. So fully fake the entire image and video itself is depicting something that was not real. You can do that by also just manipulating small parts of a video. So having a real video of a person and swapping on a different face or a different expression, adding smiles where there should be frowns. You can change the voice of a particular individual in the photo. So there's a number of different approaches that you can take, but I think the fundamental aspect that Vlad was getting at is you are depicting something that did not exist in sort of the real world as captured by a traditional camera or photography equipment.

Daniel Serfaty: That's interesting because classically I think we are naturally engineered as humans to understand that there is fiction and there is non-fiction, but it is when one tried to disguise itself as the other and us in a sense that we start being bothered. What sparked your interest, Laura, in deepfakes and the AI technology that you used to create them? For our audience, I want to clarify, Vlad and Laura are not creators of Deepfake. They are fighters. Okay. They are fighting for the truth and they're fighting to help us discover when something is fake. So what sparked your interest, Laura? Why did you get into this field?

Laura Cassani: Similar to other technologists in this space, I've been really excited about the applications of generative AI and I eagerly have been playing around with the tools as soon as they become available, which, if anyone's sort of familiar with this space, there's been a plethora of new generators and new techniques capable of making sophisticated text or audio or video or images. Many for benign purposes, you can use these techniques to create fun songs or audio clips. You can use generative AI for text purposes in terms of, I've been using it to help write study guides for my kids for their novel studies.

                There's many benign purposes and it's fun to sort of play around with them, and I think video was one where we thought, I think as a community, that it would take much longer to have the sophisticated nature of fully synthetic videos perhaps generated from text and it arrived much sooner than anyone expected. So the pace of development has been really exciting and I, like many others have just been trying to keep a pace and play around with the tools as much as possible to keep up with all the developments.

Daniel Serfaty: You got it through, how would I say, the way of being a data analyst or researcher in AI. What about you, Vlad? I know you wrote your thesis on things that started to go in that direction, so is that really what drove you to be a deep fake fighter?

Vladamir Barash: Yeah, pretty much. I'd say it's been a pretty long but continuous journey from studying and graduate school what goes viral online to realizing that people can use the internet to spread virally, not just helpful information but disinformation or misinformation, and then lately with AI enabled deep fakes to actually spread content that is completely false and is generated by a machine. For me, I think one of the most fascinating but also most frightening aspects of deep fakes is just how easy it is to generate them. All you need is basically an idea and a laptop.

There are plenty of free models and even if you want to use one of the off-the-shelf industrial models, a lot of them are very cheap like stents or dollars to use. Now these models are run by large companies and these companies do put in safeguards or guardrails to make it harder to share malicious deepfakes as opposed to the benign ones that Laura was talking about. So for instance, they might have a guardrail, their model won't let you generate images of violence or war or pornography, but unfortunately there are also ways to get around these guardrails. Increasingly that means fighting misinformation spread via deepfakes, so it's also a big professional interest.

Daniel Serfaty: Right. Let's stay with you for a second, but I really want to understand from both of you, what do you find most rewarding about your current work in going into this plethora of deepfakes in our lives, in the press in all kind of other environment, what is most rewarding for you? I know this is a technological war in a sense, you're trying to use AI to fight AI in a sense, but is it more rewarding than other projects? I know that both of you are working on a major project together, and we may talk about it a little later, but I want to understand really from a lead scientist's perspective, what is most rewarding about this?

Vladamir Barash: I'd say you got it. It's a technically challenging problem and it sort of engages my brain and it also has this immense positive social impact if we succeed fighting deepfakes and fighting disinformation. There's a lot of complexity around it, but often one of the reasons I didn't end up staying in academia when I finished my PhD was in academia you might work on a technically challenging problem and then you write a paper and then 20 people see it and that's it. Here at Graphika I get to be part of work and fighting disinformation around Chinese disinformation operations or around Taylor Swift and write reports that go out into the world and get seen by the public and are very relevant to people's lives, which is amazing.

Daniel Serfaty: So it's impactful but in a more profound way than academic impact. What about you, Laura? Do you find it satisfying?

Laura Cassani: I guess two things come to mind. One is I'm really motivated by the people I work with and the teams of people as part of those various projects, and there's many different related projects in this space, but it really spans the gamut from technologists and AI researchers to journalists that may be sort of at the pointy end of the sphere in terms of interpreting this information for public audiences. There's just a number of really great individuals that are excited and motivated to make a meaningful impact in this space, and that's been really satisfying for me to be able to be a part of those teams. And then secondly, I think deepfakes are really having their moment. I suppose it's something where it's now in sort of the mind of the public. When you show somebody a deepfake of themself that can be easily done 10 seconds behind a green screen.

I was just at CON this past weekend as part of a live deepfake demonstration, we had thousands of people come through and stand in front of a green screen and create a ten-second clip of a deepfake of themselves, and people really get it when you are looking at a screen live and it's not you, and you could imagine the threat that this could pose as we're all on Zoom or teams calls all the time of is that really the person that I'm expecting on the other end? So I think this type of technology is really having its moment in the sense that people understand and recognize the threat that this pose in a very interpretable and sort of natural way.

Daniel Serfaty: That is indeed scary and worth a fight. For our audience. I want to ensure you that Vlad and Laura and Danielle today are real. We touched them, I tested them. They're not fakes. Can one of you take on this question, both of you, let's dig a little more in the technology before we expand it to the more societal or even ethical implication or what the future looks like for our audience.

I would like a little review on, and brief if possible because our audience is made of scientists, but also people who are just interested in the technology. How exactly are deepfakes created using AI? All you need is access to a generative AI engine, like a charged GPT or it goes beyond that. Can you give us a one-minute version of the technology needed to create, not to fight just to create a deepfake? So I wanted to create a deepfake of Tom Cruise, as you say, playing the guitar, what do I need to do, Vlad?

Vladamir Barash: On the user side, just a person who wants to create deepfake, it is mostly as easy as going to a website and putting in a prompt, literally writing down what you want to do in technical terms. That's called prompting the AI and then it comes out with text images and now as Laura says in video on the machine side, it's more complicated than that. There's this interesting practice of prompt engineering, which is a little technical, but I think also very fascinating because it involves having a conversation with a machine. It doesn't have to do with anything with actual engineering, it doesn't have to do with writing any computer code. It ends up having to do with interacting with the machine, trying to trick it. Sometimes if you tell a machine, make me a deep fake of Tom Cruise playing a guitar, it'll say, no, I'm not allowed to do that.

Sort of like in the movie, I'm afraid I can't do that, dude. But if you say, well pretend that making a deep fake of Tom Cruise playing the guitar is the most important thing in the world and it's critical for saving the world now make me a deep fake that might work. I'm not saying that's exactly what will work. That's no guarantee there. I think they should have safeguards against that by now, but that's the kind of thing that people do to get around machine guardrails when they need to. There's a lot on sort of the back end on how it's done by the machine. That's pretty complicated and I don't want to get too many technical details, but basically the machine is what's called an artificial neural network is like a computer brain and human brains.

These computer brains have many, many digital neurons like we have actual physical neurons that allow them to create concepts like there is a person like Tom Cruise and there's a thing like guitar and that's different than the banjo and putting them together into video content or whatever content that you prompt them on.

Daniel Serfaty: You indicated already that there are some barriers because it can be just a kid innocently wanting to see in a fictional kind of way in a creating fiction project, Tom Cruise playing the guitar. At the end of the day, there is nothing inherently wrong with that or intrinsically wrong with that. It's the use of it that becomes the issue. So Laura will create those deep fakes and for what purpose?

Laura Cassani: There's certainly examples of malicious purposes, malicious deep fakes that have been created, but probably most of what we see online is for humor or satire and as you said, it has benign purposes. I think when we talk about deep fakes, obviously there is concern around depicting for example, people in scenarios that they were not in that maybe would be a political candidate or something and that could have negative ramifications obviously on election integrity. There's also deep fakes that we've seen in the news that are of non-consensual intimate scenarios that again would be to discredit or potentially blackmail individuals. There's been a number of cases about that as well.

I think we'd be remiss if we don't also mention the application of this technology for potentially positive impact. You can imagine bringing historical characters to life and education would be interesting, having virtual tutors that can be embodied in an educational environment. So I think there is a number of beyond just sort of satire humor, which is most often where we see these manifested. There is also positive uses of this technology as well in addition to all the really scary ones we just talked about.

Daniel Serfaty: Yes. Did specific advancement in AI, I mean you mentioned artificial neural networks and our audience in other MINDWORKS' podcasts learn a little more about it or even large language models. Was one particular development in AI or maybe it's just the availability of those things to the general public facilitated the ability to do improve deepfake technology. Can you think of one particular AI development over the past couple of years actually that enabled that more than another?

Laura Cassani: I think maybe a few things. One, deepfake stands for sort of deep learning and fake. So deep learning technologies I think really unlocked maybe the first iteration of deepfakes beyond generative adversarial networks, which is a previous approach maybe where there's been a lot of excitement in the community is with the advancement of transformer models, which is largely what large language models and that class of model where you would categorize that. So if you saw Sora from OpenAI and other tools around that, there's been a lot of excitement about the ability to use a text-based prompt and automatically generate a scene in a video that again, that scene is not depicting something real.

It's not starting with something and then editing a small piece of the video. It is fully synthetic video that in many cases looks incredibly realistic and there's sometimes some artifacts that can be detected and maybe you've seen examples of these models have a hard time getting limbs accurate and the number of fingers accurate, but that will likely be resolved with future iterations on those models. Some of those artifacts are what we use to try to detect videos as being synthetic, but it's very much a sort of arms race in the sense that as soon as we develop detectors to detect some semantic inconsistency in the video, there'll be a new model release that resolves those potential inconsistencies or artifacts. But transformer-based models have been a very exciting development in the realm of fully synthetic video capabilities.

Daniel Serfaty: You anticipated actually my next sets of technical question, which is the other side of it. In a sense, given what we know about how big fakes are being developed using transformer technology or others, how do we fight them? And I really would like to devote the next few minutes to understand how do we use AI to fight AI in a certain sense, and what are the tools you're using, how effective they are, et cetera. Can you tell our audience if AI today can be used to identify and flag deep fake content on social media and other platforms? It's a broad question. I'm sure you have many layers of answers, but let's start to answering that very question. Should we feel safe?

Vladamir Barash: I would say broadly, yes. Probably the clearest example or the easiest example I can give you is I think this happened the summer of 2023. There was the manipulated image, I think it was a deep fake image of an explosion at the Pentagon circulating on social media. Graphika was one of the organizations that was asked to look at the image very, very quickly as it was circulating, and we were actually able to use a tool from this project that Laura and I are working on as well as our broader platform for detecting disinformation to analyze it, assess with the help of AI that it was indeed manipulated image or synthetic image very, very quickly and get that assessment out before the image caused a panic before it led to some very bad consequences. So that shows the power of AI, fighting AI to detect deep fakes and also the stakes that are very, very high. Sometimes in critical situations you have minutes or at most hours to respond before something goes viral and it's a lot harder to counteract.

Daniel Serfaty: I see. Well, Laura, just to continue on that stream, how effective are current detection, identification method, not even mitigation for now, but just the fact that we can say through AI analysis that Oh yes, this is a fake, I know you were part, you were leading actually part of the evaluation of those methods. The example that Vlad gave us was from 2023, you said Vlad, is that right? So now I guess we are more than a year later. Are detection methods actually more powerful, faster, more accurate? What would you say?

Laura Cassani: I think it depends. So there's a broad level of capabilities and different analytics that take a variety of different approaches to identifying features in manipulated media, especially videos. So there's reason to feel positive in the sense that there's entities that are looking into this, that there's investment in developing these capabilities, and I think we're to the point where we want to be able to distribute those broadly.

One point that Vlad made is the real time nature perhaps is not there yet. So in the case of the picture that Vlad was mentioning about the explosion at the Pentagon, that was a manipulated image that did result in a real market dip before that was found to be synthetic and then widely reported on to be not accurate, but it did have a real world impact in terms of economic consequences. Initially these tools exist, but they're currently really set up for analysis after the fact. I think making sure that we can get these type of tools and detection capabilities more broadly to platforms or others where they can be distributed and help to identify things in real time is certainly an approach that we're interested in taking.

Daniel Serfaty: What do you mean by real time? An image happened in social media, what is real time?

Laura Cassani: I can imagine as a consumer of information on social media or in the news wanting to be able to have a sense when a piece of media comes across my feed or a timeline have a sense as to what the provenance of that particular piece of media was or is A lot of the analysis as to synthetic or manipulated media happens after the fact. So it's something appears again, in the case of the Pentagon explosion, it appeared first in Telegram I believe before it spread across platforms and networks and then it made the news because it was obviously meant to go viral and it again had those real world sort of economic impacts and then it was sort of analyzed and found to be synthetic, which happened quickly and is certainly very exciting.

I think the question you're asking is how do we condense down that timeline to even have tools that can empower people at the point of consumption to be able to have a sense as to whether or not what they're looking at, maybe not even if it's synthetic or manipulated, but in what ways has it been changed or touched up or modified so that you're maybe empowering the consumer of the information with additional context.

Daniel Serfaty: So pretty soon we are going to have these tools on our devices, on our phones, on our laptops, detection of synthetic media won't matter where everything has already been touched in some way. How do we develop tools that tell us not only what the Pentagon is not burning, that's a what, but the why and the how that something was changed. Because I think at the end of the day, we as humans want not just want to understand why would somebody put an image of the Pentagon out there, either as citizens or even as investors where we see our investment dip for a while or as actors in this social media. How do we develop tools that tell us those more sophisticated question that will allow us then as human, to process that information properly, namely the why and how? Vlad, you want to take that on?

Vladamir Barash: Yeah, it's a big one to take on. It is something that Laura and I have actually spent a lot of time thinking about and it's a complicated question, but we do think an answer is possible. So the main way I would say is to look at the context around the, is it manipulation related to conflict related to violence? Who are the key people involved in the manipulation, who is targeted in the manipulation or the deep fake media? How is it being spread through what channels?

As Laura mentioned, Telegram, then other platforms than news media that can tell an expert a lot about not just what is happening but why it might be happening, and then it's the expert's job to synthesize that into a report. Often AI can help a lot here to provide some of that context automatically. So something that we work on in our project is called Characterization. So there's detection, which is it a deep fake or not, but there's also characterization of who is in the deep fake or how is it generated and maybe how it's being spread. And these are all questions that AI can help answer with the help of human experts.

Daniel Serfaty: Laura, can you add to that by taking an example again of how, I don't know if that's a good example. We heard two examples today of Tom Cruise playing guitar after the stunt he just pulled at the closing ceremony in the Olympics. I wouldn't be surprised that he knows how to play guitar, but also the Pentagon example you just gave. So how would AI help us explain basically the context or the why? Pick any example you want, even not one of these two, but I would like our audience to visualize a little bit how that thing happen.

Laura Cassani: I think there is great excitement and application around new technologies in this space that might be able to automate a lot of what Vlad was saying in terms of providing additional context. So the audience might be familiar with, there's transformer based models like the Chat GPTs of the world and other similar text-based models, but there's also vision language models that have recently been released.

So you can imagine this is similar. You're providing an input such as an image like the Pentagon explosion or maybe an image you took that you then touched up from your weekend out with your friends and you've provided some AI capabilities change the lighting or edit out or add somebody in that wasn't there at the start. So you provide that as inputs and then the model can interpret that visual input, whether it's a video or an image and provide a written natural language explanation about the content and likely context of that image.

So you could imagine in the sense of the Pentagon explosion picture, so if you're visualizing it and you could find it online also it's a picture of the Pentagon where there's smoke and fire that is not real. You provide that as an input and you have a conversation with the model of why might this have been created for what purpose was this created, what is the intention or the particular technique behind this and the tool, these tools, these models will be able to provide that to you in a consumable natural language explanation in a similar way that you chat with chat GPT.

So I think the approach where the research community is going is to be able to utilize this new class of model to be able to provide sort of characterization large across many pieces of media that will provide that context for that consumer, that user of the information sitting behind a desk or looking at their phone and seeing these images pass through. It'll be able to provide not only what was manipulated, whether it was a person was removed from the image or smoke was added or a weapon was inserted or the sunset lighting was changed for beautification and provide a sense as to what's important to look at and what may be more.

Daniel Serfaty: Thank you. That visually adds a lot of texture to the technical description and I think it helps because we are in a world that this is going to be more and more needed, I believe for us to deal with that distortion, if not disappearance of the truth as we know it. We'll explore now the implications of having deepfakes in our midst. Nobody can stay passive to this situation because deepfakes come in our social media, in the image we consume in the news or in our environment, and I want to ask our guests today, how concerned should we be about the expanding use of deepfakes in our lives? We are approaching a presidential election in a couple of months or should we be very concerned and how really at the end of the day is it different from classic propaganda, which is basically the mass propagation of falseness? Vlad, are you concerned?

Vladamir Barash: Yeah, I am concerned. I think I'm going to take the second part of your question first, which is how it's different from classic propaganda and this is also a big reason why I am concerned. The big difference is ease of access. So with classic propaganda, you needed access to mass media broadcast technology, it was prohibitively expensive. Really you have to be a nation state to do classic propaganda. Now with deepfakes, basically all you need is a social media account, laptop or a phone. Like I said earlier, I think there are many services that can be used to generate deepfakes that are completely free and even the paid ones cost only a few dollars and they do have guardrails, but they can be overcome. The big difference, it's just how easy it is to generate a deepfake overall.

In terms of how concerned I am this year, I think there are very serious efforts to make sure something like deepfakes don't affect election integrity a whole of society. I don't really want to speculate on them in detail and I don't know them in detail, but even in the 2020 election, people were already starting to think about it and certainly disinformation was already a problem. There were efforts to make sure that people at the basic level knew where to go vote and when to go vote and that there was no false information spreading around that. So there's definitely efforts across the whole of society to make sure we can all vote and exercise our constitutional right.

Daniel Serfaty: I'm trying to digest those very important statements you just made because it is really almost see the keys to the kingdom or so have been given to everyone and therefore you rely on everybody ethical compass to make a decision and that's terrifying. What measures Laura can be taken to mitigate the misuse of deepfake technology, the misuse of it?

Laura Cassani: I think there's a number of different capabilities that are in development in terms of detection and characterization mechanisms for examining the authenticity of media artifacts like deepfakes. To your earlier question about the threat in the election officials that we've been speaking to, one of the nightmare scenarios is that close to election day somebody gets a phone call or there's a video out that there's a security situation and a polling place and not to go. Of course those are fabricated claims.

It could be an audio clone of an election official, whether or not there's not a security situation, it shows enough distrust that it affects people going out to exercise their right to vote. That's a huge concern and is quite difficult to combat on the scale that is required for a national election. The other piece that you mentioned about what can be done to combat is building off of what Vlad said about why video and audio sort of modalities are quite difficult is I think in many cases as consumers of information, we do an okay job at discerning traditional information sources like text or we see a post come through, we know to maybe question who said that, where did this come from? What's the title of the article, what's the publication it's coming from?

And we have some sort of history of critical thinking skills around media literacy with traditional sources. I think where we need to maybe focus efforts is on places where we originally thought like a video, you would assume that that's real what you're seeing, or if you hear somebody's voice on the other end of the telephone, you assume that that is the person you're expecting. That's no longer the case in sort of the current landscape. And we need to be able to build those critical thinking skills to I guess question everything, trust no one trust, but verify that even in traditional mediums where you would expect authenticity like video, like audio, that is no longer something that we can take for granted.

Daniel Serfaty: From what you two are saying in terms of the measure, I can see the measure in addition to improving our countermeasure technology, which you guys are really the point here. On the one hand we should educate the public about that. So there is a make the consumer of information more mature, more sophisticated, more questioning. On the other hand, the hammer in a sense that very tough regulations and legal measures against people we use be fake. Which side of that equation would you advocate more or would you advocate? We do both of them at the same time? I want actually both of you to answer that question because I think you may have different perspective on that.

Laura Cassani: I think there needs to be a recognition that depending on the context that will largely information is protected by the first amendment. Even information that may be false or misleading or untrue, I would err I think on the side of providing more information to consumers and letting them make their own decisions and interpretations. There are a number of laws and legislation that is currently being considered at the state and at the national level on political campaigns, which may be a particular carve out where specifically targeting legislation that if you are a campaign you either cannot or have to disclose if you are depicting something that is materially untrue about the other candidate in this video context. And that's been a bipartisan effort and a number of states have passed legislation to that extent restricting that.

But I would say that I think media literacy providing people more tools that provide information about what they're consuming, what's been changed, what's been manipulated, how to think critically and skeptically about the information they're consuming as opposed to trying to limit its use or spread would be where I would invest my efforts.

Daniel Serfaty: Vlad, do you agree on a hundred percent or less than that?

Vladamir Barash: Yeah, I largely agree. The example that personally really appeals to me is something like ex-formally Twitter, the community notes feature. I see it on a Twitter post. I can quickly see there's something there, I can read it and I can make my own decision about whether to repost it or like it or whether I'm skeptical about this post that has a community note. I think another one is the digital Content authenticity initiative. I think that's an industry initiative. I think it's run by Adobe basically creating watermarks for media and lagging when the provenance and the origin of some media like this came from somebody's phone, camera, obviously anonymized as needed or this one was digitally altered with a tool or this one was AI generated.

So I think that's another very useful approach and education of course and a big proponent of, so yeah, I largely agree with Laura. It's about helping the public make better, more informed decisions. Laura also touched on some very important points about emergency situations, something like the disinformation equivalent of yelling fire in a crowded theater. That's a known limitation, right? So I think that's important to be aware of, but the emphasis should be in most cases it's not that most cases it's about educating the public, helping them make better informed decisions.

Daniel Serfaty: And I think we are both in a known territory here because the education of the public can be thought of locally but also very globally. And the question is that do you think that people that are already skeptical of the media as we know it, you turn on CNN or Fox News and depending on your leaning, you say, I don't believe that's true or they exaggerate or something like that. So we are naturally, at least in this country, suspicious of what we hear on the media, even from sources that have been vetted by decades of being in the space.

How do we prevent people from being even more skeptical of the media? I'm using media in a very general meaning here if they don't know that any piece of media is real or not, and I want our audience to understand that media is not what you hear on CNN or what you see on Telegram or Twitter. It can be also phone call that you receive from the voice that is remarkably like that of a relative will people become more skeptical of everything? I mean, I'm terrified of a society where you start doubting absolutely everything. What do you think? And that last sentence was actually pronounced by me, not by someone else.

Vladamir Barash: It is tough, and I think Laura said, I think people will become more skeptical, which has upsides and downsides. But you make a very good point that there is such a thing as too much skepticism. We can think about it as this concept of the Liar's dividend. If you pollute the information environment with enough deep fakes, then people start questioning everything that's a problem. So I think there is a golden middle that it is possible to inform the public about and help the public educate themselves about. There's some great tools that people are developing on the education level, like bad news, internet game where you can sort of play and try to see what it's like to be in the mind of someone who creates a deep fake.

I think that's a very powerful tool, and I think there's healthy skepticism, good media literacy practice such as you look for multiple sources of confirmation. You look for good journalistic practice, you look for primary accounts instead of, oh, I heard somebody say this, so on so forth. You look for multiple modes of verification. I think those are all skills that we should help spread in a good way and they can help all of us, the citizenry, the public, find a healthy middle between, oh, I'm just going to believe everything I see online versus, oh, I can't believe anything I see online no matter how verified.

Daniel Serfaty: But it's interesting because I believe, and since we are kids, we are implicitly trained to find that golden middle that I don't throw away everything, but I don't accept everything as I see it. It's almost like if fake news have shifted that point that we are so comfortable with, and I'm afraid it'll go too much to one extreme. I'm going to make it personal. Laura, what would you tell your kids about how to deal with the media that they are consuming right now? What will be your advice and look at our audience as if they were your kids?

Laura Cassani: I'm encouraged actually because of this generation's familiarity with various tools in this space and maybe their comfort level. So I have four kids, 12, 10, 7 and four, and I think their familiarity with creating videos, editing videos, manipulating their voice with their friends, using tools even like chat GPT to make fun images and create virtual escape rooms. I find a great familiarity, at least in my little family circle of using these tools for fun games and educational purposes. And I'm hopeful that they'll see the capabilities for what they are and the positive uses as well as we'll be able now to have some sense as to how this could be used in a nefarious or malicious sense and to be wary of what they're consuming online for exactly that reason.

So I'm hopeful that the new generation, given their sort of eagerness and interest in technology and sort of the fun ways that they can approach it, that they'll be able to interpret and maybe be better aware of when they will inevitably encounter a malicious use of it online.

Daniel Serfaty: Wow, I love that answer. I love your hopefulness or even optimism when it comes to the future. So let's talk about the future for a second and predicting the future we know is a very dangerous business, especially in the space where the future is basically three months from now. Even the pace of technology development and its acceleration. But what do you see is the future of deepfake technology and its impact on our society in the next five years? And by that I mean not only the malicious use of the technology, but also the countermeasures we can apply to that flat. Can you give us a wild prediction or maybe in your case, a very informed prediction about the next five years?

Vladamir Barash: I think that probably the biggest philosophical almost shift that we will see in the next five years is people will reconsider their relationship to media and information. It's very similar to what Laura was talking about with her kids. I think you will see the youngest generation, but also older folks, increasingly thinking of media as something that can be synthesized and manipulated and not something that you get from a TV screen from a central broadcasting source. So that's I think, the biggest shift and I think it will come with more skepticism about what people see online.

On the positive side, I think people will build up defenses and counters to deep things. I think this is going to be, again, a whole-of-society thing, and it's already starting public sector, private sector. There will be more initiatives around the world to counter scams to ensure democracy and global political participation that citizens have the right to express their voice and vote for what they want without being duped by deep fakes. And we will see lots more organizations both dedicated to fighting deep fakes and new units of media organizations, of research, think tanks of major industry players that are dedicated to making sure that deep fakes do not spread. So I think it'll be a sort of organic response by society to a world where there is just a ton of synthetic media.

Daniel Serfaty: That's interesting. So it seems like you're envisioning some kind of an ecosystem of, I don't know, universities, schools, companies, banks, governments that is almost being re-engineered to deal with that. It's pretty big revolution in that vision. It's almost like it's becoming central and as a result, as an organism, we have to defend ourselves to absorb that threat and to deal with it. That's an interesting vision. Laura, you want to add to that?

Laura Cassani: I think there's very much interest in building a national capacity that could sort of harness these resources and work in a coordinated fashion in terms of developing literacy programs, educational content detection and characterization capabilities. And I think there's a great deal of goodwill and interest to across industry, government, nonprofit space to organize and get behind this national capacity effort. And hopefully that will help bridge what is inevitably a arms race as it were, between the developers of the technology, which again is not nefarious, it's just new generation capabilities that are continually coming out that produce more and more sophisticated synthetic content such as videos and the detection capabilities designed to be able to provide a sense as to whether the piece of media you're consuming is authentic or not.

And so it is this back and forth between a new generation technique comes out, there's a flurry of activity to develop a detector for that maybe can identify artifacts in that. And then a month down the line there's a new, it moves really that quick. There's a new generation technology and again, sort of this back and forth between these two groups. So it really will take a large ecosystem of players to be able to coordinate and act in this space to be able to move the in terms of providing this information to the public.

Daniel Serfaty: That's fascinating to look at that change actually. That made me want to invite you, the two of you back here in say a year from now and see whether or not that ecosystem has been built and whether or not we've made progress there. But in the meantime, my last request from you is for our audience, whether they are students or technology people or business professionals, do you have any advice for them to stay ahead of the curve to understand and manage the current implications of deep fakes in their lives, in their private lives? With all the danger that we hear now about anything from elections, impersonation, pornography, and all these other things, what will be your simple advice to how to deal that on a daily basis as a way to not only survive, but maybe even thrive in that new environment?

Vladamir Barash: Good question. We both said some of these things, double check your sources, look for multiple verifications and primary sources. When you see something, especially if something is inflammatory, says it's breaking news or something like that, be skeptical but reasonably skeptical.

One we haven't said is there's this principle of just waiting, like taking a few breaths and waiting before clicking share or responding to something you see online. It's actually based, I think, on this cognitive principle that in the first few seconds when we react, we react with our animal brain. It's a very fight or flight response. And if we just wait a little bit, then our higher order cognitive functions turn on and we become more thoughtful about it. We consider the context, we become more skeptical and that can be very, very helpful in preventing disinformation and deepfakes from spreading.

Laura Cassani: I would just say it's a very exciting and rapidly developing field and there's new techniques, new models, new approaches, new use cases coming out all the time. And so I would just encourage people to try to stay abreast of the technology developments in this space because it really will have wide-ranging impacts in education and entertainment. We're not far from sort of having the novel to video pipeline, which is kind of super fun and exciting to think about taking your favorite book and having a video just generated of it on the fly. So I think it's an exciting space and I would just encourage people to look skeptically, of course, at what they're consuming, but to be excited about all the positive potential that this can have for society writ large as well.

Daniel Serfaty: Thank you very much, Vladimir Barash and Laura Cassani for a wonderful session. And consider that an invitation back here in six to 12 months and we'll see whether or not your amazing predictions are being realized and what else is new in this field.

Thank you for joining me today. As always, the MINDWORKS team welcomes your comments and feedback as well as your suggestions for future topics and guests. You can email us at mindworkspodcast@gmail.com. We love hearing from you. MINDWORKS is a production of Aptima Incorporated. My executive producer is Ms. Debra McNeely, and my assistant producer is Ms. Chelsea Morrissey. Sound engineering is provided by Bespoke podcast editing. To learn more, please visit aptima.com/mindworks. Thank you.