But Duplex calling 911 is the least of your worries as more and more AI products reach the public. What you really have to worry about is fake video of you doing things you didn't do and saying things you didn't say. You will also have to worry about fake video implicating the wrong person in a crime or showing a guilty man lounging on the beach 500 miles away rather than robbing your local bank. We are about to enter an era where all video and audio evidence may be more suspect than it's ever been.
Fortunately, the body camera and in-car video companies long ago incorporated technology that prevents anyone from altering a video. Unfortunately, a percentage of the public inclined toward taking anti-police propaganda as gospel already believes that police departments doctor videos to get the evidence they need to exonerate officers after a controversial shooting. So imagine what will happen if in the very near future every controversial police shooting is posted on the Web as multiple videos showing totally different actions
For example, last month in Chicago, a police officer shot a barber after seeing the man allegedly reach for his waistband toward a concealed firearm. Officers had originally stopped the man because they suspected he was carrying because of a visible bulge at his waist. He reportedly resisted, pulled away from the officers, and then the sequence of events that led to his fatal shooting occurred. This shooting led to some unrest, including some "violent protesting," and some officers were injured by thrown objects. As of this writing, the city is still tense. But imagine how tense it would be if the police video showed the man was armed, which he was, and an equally convincing video showed he was unarmed.
Seamless video manipulation using AI algorithms is about to become a major challenge in many fields, including law enforcement. Crude apps are already available that can do some pretty remarkable video manipulation. For example, people have manipulated porn videos swapping the faces and voices of the actress for those of mainstream actresses. They've also created a bogus video of Barack Obama spewing obscenities about his successor. These videos are called "deep fakes."
It's pretty easy to determine that the current generation of deep fakes has been manipulated. But the thing about AI-based software is that by its very definition it learns, it gets better with use. So those telltale blurs and other artifacts that now reveal these things are phony are going to go away. And it's estimated that very soon even top experts in the field of video editing will have trouble determining what is real and what is fake.