An Information Apocalypse

Moderators: Elvis, DrVolin, Jeff

An Information Apocalypse

Postby seemslikeadream » Mon Feb 12, 2018 3:15 pm

He Predicted The 2016 Fake News Crisis. Now He's Worried About An Information Apocalypse.
“What happens when anyone can make it appear as if anything has happened, regardless of whether or not it did?" technologist Aviv Ovadya warns.

Posted on February 11, 2018, at 7:45 p.m.

Charlie Warzel
BuzzFeed News Reporter
In mid-2016, Aviv Ovadya realized there was something fundamentally wrong with the internet — so wrong that he abandoned his work and sounded an alarm. A few weeks before the 2016 election, he presented his concerns to technologists in San Francisco’s Bay Area and warned of an impending crisis of misinformation in a presentation he titled “Infocalypse.”

The web and the information ecosystem that had developed around it was wildly unhealthy, Ovadya argued. The incentives that governed its biggest platforms were calibrated to reward information that was often misleading and polarizing, or both. Platforms like Facebook, Twitter, and Google prioritized clicks, shares, ads, and money over quality of information, and Ovadya couldn’t shake the feeling that it was all building toward something bad — a kind of critical threshold of addictive and toxic misinformation. The presentation was largely ignored by employees from the Big Tech platforms — including a few from Facebook who would later go on to drive the company’s NewsFeed integrity effort.

Aviv Ovadya, San Francisco, Calif. Tuesday, February 1, 2018.

“At the time, it felt like we were in a car careening out of control and it wasn’t just that everyone was saying, ‘we’ll be fine’ — it’s that they didn't even see the car,” he said.

Ovadya saw early what many — including lawmakers, journalists, and Big Tech CEOs — wouldn’t grasp until months later: Our platformed and algorithmically optimized world is vulnerable — to propaganda, to misinformation, to dark targeted advertising from foreign governments — so much so that it threatens to undermine a cornerstone of human discourse: the credibility of fact.

But it’s what he sees coming next that will really scare the shit out of you.

“Alarmism can be good — you should be alarmist about this stuff,” Ovadya said one January afternoon before calmly outlining a deeply unsettling projection about the next two decades of fake news, artificial intelligence–assisted misinformation campaigns, and propaganda. “We are so screwed it's beyond what most of us can imagine,” he said. “We were utterly screwed a year and a half ago and we're even more screwed now. And depending how far you look into the future it just gets worse.”

That future, according to Ovadya, will arrive with a slew of slick, easy-to-use, and eventually seamless technological tools for manipulating perception and falsifying reality, for which terms have already been coined — “reality apathy,” “automated laser phishing,” and "human puppets."

Which is why Ovadya, an MIT grad with engineering stints at tech companies like Quora, dropped everything in early 2016 to try to prevent what he saw as a Big Tech–enabled information crisis. “One day something just clicked,” he said of his awakening. It became clear to him that, if somebody were to exploit our attention economy and use the platforms that undergird it to distort the truth, there were no real checks and balances to stop it. “I realized if these systems were going to go out of control, there’d be nothing to reign them in and it was going to get bad, and quick,” he said.

"We were utterly screwed a year and a half ago and we're even more screwed now"


Today Ovadya and a cohort of loosely affiliated researchers and academics are anxiously looking ahead — toward a future that is alarmingly dystopian. They’re running war game–style disaster scenarios based on technologies that have begun to pop up and the outcomes are typically disheartening.

For Ovadya — now the chief technologist for the University of Michigan’s Center for Social Media Responsibility and a Knight News innovation fellow at the Tow Center for Digital Journalism at Columbia — the shock and ongoing anxiety over Russian Facebook ads and Twitter bots pales in comparison to the greater threat: Technologies that can be used to enhance and distort what is real are evolving faster than our ability to understand and control or mitigate it. The stakes are high and the possible consequences more disastrous than foreign meddling in an election — an undermining or upending of core civilizational institutions, an "infocalypse.” And Ovadya says that this one is just as plausible as the last one — and worse.

Worse because of our ever-expanding computational prowess; worse because of ongoing advancements in artificial intelligence and machine learning that can blur the lines between fact and fiction; worse because those things could usher in a future where, as Ovadya observes, anyone could make it “appear as if anything has happened, regardless of whether or not it did.”

"What happens when anyone can make it appear as if anything has happened, regardless of whether or not it did?"
And much in the way that foreign-sponsored, targeted misinformation campaigns didn't feel like a plausible near-term threat until we realized that it was already happening, Ovadya cautions that fast-developing tools powered by artificial intelligence, machine learning, and augmented reality tech could be hijacked and used by bad actors to imitate humans and wage an information war.

And we’re closer than one might think to a potential “Infocalypse.” Already available tools for audio and video manipulation have begun to look like a potential fake news Manhattan Project. In the murky corners of the internet, people have begun using machine learning algorithms and open-source software to easily create pornographic videos that realistically superimpose the faces of celebrities — or anyone for that matter — on the adult actors’ bodies. At institutions like Stanford, technologists have built programs that that combine and mix recorded video footage with real-time face tracking to manipulate video. Similarly, at the University of Washington computer scientists successfully built a program capable of “turning audio clips into a realistic, lip-synced video of the person speaking those words.” As proof of concept, both the teams manipulated broadcast video to make world leaders appear to say things they never actually said.

https://www.youtube.com/watch?v=MVBe6_o4cMI

University of Washington, computer scientists successfully built a program capable of “turning audio clips into a realistic, lip-synced video of the person speaking those words.” In their example, they used Obama.
youtube.com / Via washington.edu
University of Washington, computer scientists successfully built a program capable of “turning audio clips into a realistic, lip-synced video of the person speaking those words.” In their example, they used Obama.
As these tools become democratized and widespread, Ovadya notes that the worst case scenarios could be extremely destabilizing.

There’s “diplomacy manipulation,” in which a malicious actor uses advanced technology to “create the belief that an event has occurred” to influence geopolitics. Imagine, for example, a machine-learning algorithm (which analyzes gobs of data in order to teach itself to perform a particular function) fed on hundreds of hours of footage of Donald Trump or North Korean dictator Kim Jong Un, which could then spit out a near-perfect — and virtually impossible to distinguish from reality — audio or video clip of the leader declaring nuclear or biological war. “It doesn’t have to be perfect — just good enough to make the enemy think something happened that it provokes a knee-jerk and reckless response of retaliation.”

"It doesn’t have to be perfect — just good enough"
Another scenario, which Ovadya dubs “polity simulation,” is a dystopian combination of political botnets and astroturfing, where political movements are manipulated by fake grassroots campaigns. In Ovadya’s envisioning, increasingly believable AI-powered bots will be able to effectively compete with real humans for legislator and regulator attention because it will be too difficult to tell the difference. Building upon previous iterations, where public discourse is manipulated, it may soon be possible to directly jam congressional switchboards with heartfelt, believable algorithmically-generated pleas. Similarly, Senators' inboxes could be flooded with messages from constituents that were cobbled together by machine-learning programs working off stitched-together content culled from text, audio, and social media profiles.

Then there’s automated laser phishing, a tactic Ovadya notes security researchers are already whispering about. Essentially, it's using AI to scan things, like our social media presences, and craft false but believable messages from people we know. The game changer, according to Ovadya, is that something like laser phishing would allow bad actors to target anyone and to create a believable imitation of them using publicly available data.


“Previously one would have needed to have a human to mimic a voice or come up with an authentic fake conversation — in this version you could just press a button using open source software,” Ovadya said. “That’s where it becomes novel — when anyone can do it because it’s trivial. Then it’s a whole different ball game.”

Imagine, he suggests, phishing messages that aren’t just a confusing link you might click, but a personalized message with context. “Not just an email, but an email from a friend that you’ve been anxiously waiting for for a while,” he said. “And because it would be so easy to create things that are fake you'd become overwhelmed. If every bit of spam you receive looked identical to emails from real people you knew, each one with its own motivation trying to convince you of something, you’d just end up saying, ‘okay, I'm going to ignore my inbox.’”


Via YouTube
That can lead to something Ovadya calls “reality apathy”: Beset by a torrent of constant misinformation, people simply start to give up. Ovadya is quick to remind us that this is common in areas where information is poor and thus assumed to be incorrect. The big difference, Ovadya notes, is the adoption of apathy to a developed society like ours. The outcome, he fears, is not good. “People stop paying attention to news and that fundamental level of informedness required for functional democracy becomes unstable.”

Ovadya (and other researchers) see laser phishing as an inevitability. “It’s a threat for sure, but even worse — I don't think there's a solution right now,” he said. “There's internet scale infrastructure stuff that needs to be built to stop this if it starts.”

Beyond all this, there are other long-range nightmare scenarios that Ovadya describes as "far-fetched," but they're not so far-fetched that he's willing to rule them out. And they are frightening. "Human puppets," for example — a black market version of a social media marketplace with people instead of bots. “It’s essentially a mature future cross border market for manipulatable humans,” he said.

Ovadya’s premonitions are particularly terrifying given the ease with which our democracy has already been manipulated by the most rudimentary, blunt-force misinformation techniques. The scamming, deception, and obfuscation that’s coming is nothing new; it’s just more sophisticated, much harder to detect, and working in tandem with other technological forces that are not only currently unknown but likely unpredictable.

For those paying close attention to developments in artificial intelligence and machine learning, none of this feels like much of a stretch. Software currently in development at the chip manufacturer Nvidia can already convincingly generate hyperrealistic photos of objects, people, and even some landscapes by scouring tens of thousands of images. Adobe also recently piloted two projects — Voco and Cloak — the first a "Photoshop for audio," the second a tool that can seamlessly remove objects (and people!) from video in a matter of clicks.

In some cases, the technology is so good that it’s startled even its creators. Ian Goodfellow, a Google Brain research scientist who helped code the first “generative adversarial network” (GAN), which is a neural network capable of learning without human supervision, cautioned that AI could set news consumption back roughly 100 years. At an MIT Technology Review conference in November last year, he told an audience that GANs have both “imagination and introspection” and “can tell how well the generator is doing without relying on human feedback.” And that, while the creative possibilities for the machines is boundless, the innovation, when applied to the way we consume information, would likely “clos[e] some of the doors that our generation has been used to having open.”


Images of fake celebrities created by Generative Adversarial Networks (GANs).

In that light, scenarios like Ovadya’s polity simulation feel genuinely plausible. This summer, more than one million fake bot accounts flooded the FCC’s open comments system to “amplify the call to repeal net neutrality protections.” Researchers concluded that automated comments — some using natural language processing to appear real — obscured legitimate comments, undermining the authenticity of the entire open comments system. Ovadya nods to the FCC example as well as the recent bot-amplified #releasethememo campaign as a blunt version of what's to come. "It can just get so much worse," he said.

“You don't need to create the fake video for this tech to have a serious impact. You just point to the fact that the tech exists and you can impugn the integrity of the stuff that’s real.”
Arguably, this sort of erosion of authenticity and the integrity of official statements altogether is the most sinister and worrying of these future threats. “Whether it’s AI, peculiar Amazon manipulation hacks, or fake political activism — these technological underpinnings [lead] to the increasing erosion of trust,” computational propaganda researcher Renee DiResta said of the future threat. “It makes it possible to cast aspersions on whether videos — or advocacy for that matter — are real.” DiResta pointed out Donald Trump’s recent denial that it was his voice on the infamous Access Hollywood tape, citing experts who told him it’s possible it was digitally faked. “You don't need to create the fake video for this tech to have a serious impact. You just point to the fact that the tech exists and you can impugn the integrity of the stuff that’s real.”

It’s why researchers and technologists like DiResta — who spent years of her spare time advising the Obama administration, and now members of the Senate Intelligence Committee, against disinformation campaigns from trolls — and Ovadya (though they work separately) are beginning to talk more about the looming threats. Last week, the NYC Media Lab, which helps the city’s companies and academics collaborate, announced a plan to bring together technologists and researchers in June to “explore worst case scenarios” for the future of news and tech. The event, which they’ve named Fake News Horror Show, is billed as “a science fair of terrifying propaganda tools — some real and some imagined, but all based on plausible technologies.”

“In the next two, three, four years we’re going to have to plan for hobbyist propagandists who can make a fortune by creating highly realistic, photo realistic simulations,” Justin Hendrix, the executive director of NYC Media Lab, told BuzzFeed News. “And should those attempts work, and people come to suspect that there's no underlying reality to media artifacts of any kind, then we're in a really difficult place. It'll only take a couple of big hoaxes to really convince the public that nothing’s real.”

Given the early dismissals of the efficacy of misinformation — like Facebook CEO Mark Zuckerberg’s now-infamous statement that it was "crazy" that fake news on his site played a crucial role in the 2016 election — the first step for researchers like Ovadya is a daunting one: Convince the greater public, as well as lawmakers, university technologists, and tech companies, that a reality-distorting information apocalypse is not only plausible, but close at hand.

"It'll only take a couple of big hoaxes to really convince the public that nothing’s real."
A senior federal employee explicitly tasked with investigating information warfare told BuzzFeed News that even he's not certain how many government agencies are preparing for scenarios like the ones Ovadya and others describe. “We're less on our back feet than we were a year ago," he said, before noting that that's not nearly good enough. “I think about it from the sense of the enlightenment — which was all about the search for truth,” the employee told BuzzFeed News. “I think what you’re seeing now is an attack on the enlightenment — and enlightenment documents like the Constitution — by adversaries trying to create a post-truth society. And that’s a direct threat to the foundations of our current civilization."

That’s a terrifying thought — more so because forecasting this kind of stuff is so tricky. Computational propaganda is far more qualitative than quantitative — a climate scientist can point to explicit data showing rising temperatures, whereas it’s virtually impossible to build a trustworthy prediction model mapping the future impact of yet-to-be-perfected technology.

For technologists like the federal employee, the only viable way forward is to urge caution, to weigh the moral and ethical implications of the tools being built and, in so doing, avoid the Frankensteinian moment when the creature turns to you and asks, "Did you ever consider the consequences of your actions?"

"I’m from the free and open source culture — the goal isn't to stop technology but ensure we're in an equilibria that's positive for people. So I’m not just shouting ‘this is going to happen,' but instead saying, ‘consider it seriously, examine the implications," Ovadya told BuzzFeed News. “The thing I say is, ‘trust that this isn't not going to happen.’”

Hardly an encouraging pronouncement. That said, Ovadya does admit to a bit of optimism. There’s more interest in the computational propaganda space then ever before, and those who were previously slow to take threats seriously are now more receptive to warnings. “In the beginning it was really bleak — few listened,” he said. "But the last few months have been really promising. Some of the checks and balances are beginning to fall into place." Similarly, there are solutions to be found — like cryptographic verification of images and audio, which could help distinguish what's real and what's manipulated.

Still, Ovadya and others warn that the next few years could be rocky. Despite some pledges for reform, he feels the platforms are still governed by the wrong, sensationalist incentives, where clickbait and lower-quality content is rewarded with more attention. "That's a hard nut to crack in general, and when you combine it with a system like Facebook, which is a content accelerator, it becomes very dangerous."

Just how far out we are from that danger remains to be seen. Asked about the warning signs he’s keeping an eye out for, Ovadya paused. “I’m not sure, really. Unfortunately, a lot of the warning signs have already happened.”
https://www.buzzfeed.com/charliewarzel/ ... .kl8wNx6q6
Mazars and Deutsche Bank could have ended this nightmare before it started.
They could still get him out of office.
But instead, they want mass death.
Don’t forget that.
User avatar
seemslikeadream
 
Posts: 32090
Joined: Wed Apr 27, 2005 11:28 pm
Location: into the black
Blog: View Blog (83)

Re: An Information Apocalypse

Postby Elvis » Mon Feb 12, 2018 4:52 pm

We are so screwed it's beyond what most of us can imagine


Thanks—I don't think the significance of what Ovadya is saying can be overstated!

seemslikeadream » Mon Feb 12, 2018 12:15 pm wrote:
He Predicted The 2016 Fake News Crisis. Now He's Worried About An Information Apocalypse.
“What happens when anyone can make it appear as if anything has happened, regardless of whether or not it did?" technologist Aviv Ovadya warns.
“The purpose of studying economics is not to acquire a set of ready-made answers to economic questions, but to learn how to avoid being deceived by economists.” ― Joan Robinson
User avatar
Elvis
 
Posts: 7433
Joined: Fri Apr 11, 2008 7:24 pm
Blog: View Blog (0)

Re: An Information Apocalypse

Postby dada » Mon Feb 12, 2018 6:07 pm

I think the spin of the article makes things sound a lot worse than they actually are.

In its own way, it attacks truth. The agenda is pretty transparent. "You, apathetic spectator, are drowning in a flood of artificial post-reality. The very civilizational institutions you hold dear are being undermined and upended. Trust the enlightened technocratic experts when we tell you, the truth is beyond your ability to comprehend."
Both his words and manner of speech seemed at first totally unfamiliar to me, and yet somehow they stirred memories - as an actor might be stirred by the forgotten lines of some role he had played far away and long ago.
User avatar
dada
 
Posts: 2600
Joined: Mon Dec 24, 2007 12:08 am
Blog: View Blog (0)

Re: An Information Apocalypse

Postby Rory » Mon Feb 12, 2018 6:16 pm

Dada, that's pretty much my read of this. Who the fuck does he think he is, wannabe arbiter of 'Truth'

And Buzzfeed as a vector for this Important Message? Gtfooh

More shite from professional fecal alchemists
Rory
 
Posts: 1596
Joined: Tue Jun 10, 2008 2:08 pm
Blog: View Blog (0)

Re: An Information Apocalypse

Postby Elvis » Mon Feb 12, 2018 8:56 pm

dada » Mon Feb 12, 2018 3:07 pm wrote:I think the spin of the article makes things sound a lot worse than they actually are.

In its own way, it attacks truth. The agenda is pretty transparent. "You, apathetic spectator, are drowning in a flood of artificial post-reality. The very civilizational institutions you hold dear are being undermined and upended. Trust the enlightened technocratic experts when we tell you, the truth is beyond your ability to comprehend."


Oh dear!
“The purpose of studying economics is not to acquire a set of ready-made answers to economic questions, but to learn how to avoid being deceived by economists.” ― Joan Robinson
User avatar
Elvis
 
Posts: 7433
Joined: Fri Apr 11, 2008 7:24 pm
Blog: View Blog (0)

Re: An Information Apocalypse

Postby elfismiles » Thu Jun 07, 2018 10:14 am

#DeepFake #DeepVideo

Deepfake, a portmanteau of "deep learning" and "fake", is an artificial intelligence-based human image synthesis technique. It is used to combine and superimpose existing images and videos onto source images or videos. Deepfakes may be used to create fake celebrity pornographic videos or revenge porn.
Deepfake - Wikipedia
https://en.wikipedia.org/wiki/Deepfake


You Won't Believe What Obama Says In This Video! - YouTube
Video for deepfake obama jordan peele ▶ 1:13
https://www.youtube.com/watch?v=cQ54GDm1eL0


Jordan Peele uses AI, President Obama in fake news PSA - YouTube
Video for deepfake obama jordan peele ▶ 2:32
https://www.youtube.com/watch?v=bE1KWpoX9Hk
Apr 18, 2018 - Uploaded by Good Morning America
Jordan Peele produced the video, which uses artificial intelligence, with Buzzfeed to warn about the future of ...

Deep Video Portraits - SIGGRAPH 2018

https://www.youtube.com/watch?v=qc5P2bvfl44

AI Can Now Manipulate People’s Movements In Fake Videos
Christian Theobalt / Youtube / Emily Cho
by Dan Robitzski June 6, 2018 Artificial Intelligence

There are already fake videos on the internet, manipulated to make it look like people said things (or appeared in porn) that they never did. And now they’re about to get way better, thanks to some new tools powered by artificial intelligence.

Instead of just moving a source video’s lips and face, an artificial intelligence-powered system can create photorealistic videos in which people can sway, turn their heads, blink their eyes, and emote. Basically, everything that an actor does and says in an input video will be translated into the video being altered.

According to the research, which will be presented at the VR filmmaking conference SIGGRAPH in August, the team ran a number of tests comparing its new algorithm to existing means of manipulating lifelike videos and images, many of which have been at least partially developed by Facebook and Google. Their system outperformed all the others, and participants in an experiment struggled to determine whether or not the resulting videos were real.

The researchers, who received some funding from Google, hope that their work will be used to improve virtual reality technology. And because the AI system only needs to train on a few minutes of source video to work, the team feels that its new tools will help make high-end video editing software more accessible.

The researchers also know their work might, uh, worry some folks.

“I’m aware of the ethical implications of those reenactment projects,” researcher Justus Thies told The Register. “That is also a reason why we published our results. I think it is important that the people get to know the possibilities of manipulation techniques.”

But at what point do we get tired of people “raising awareness” by further developing the problem? In the paper itself, there is just one sentence dedicated to ethical concerns — the researchers suggest that someone ought to look into better watermarking technologies or other ways to spot fake videos.

Not them, though. They’re too busy making it easier than ever to create flawless manipulated videos.

https://futurism.com/ai-can-now-manipul ... ke-videos/
User avatar
elfismiles
 
Posts: 8511
Joined: Fri Aug 11, 2006 6:46 pm
Blog: View Blog (4)

Re: An Information Apocalypse

Postby elfismiles » Tue Dec 18, 2018 11:59 am

Past RI posts about NVIDIA:
search.php?keywords=nvidia

Nvidia’s Scary AI Generates Humans That Look 100% Real
Jesus Diaz · Freelance Writer / Updated Dec 17, 2018

Believe it or not, all these faces are fake. They have been synthesized by Nvidia’s new AI algorithm, a generative adversarial network capable of automagically creating humans, cats, and even cars.

Image

The technology works so well that we can expect synthetic image search engines soon — just like Google’s, but generating new fake images on the fly that look real. Yes, you know where that is going — and sure, it can be a lot of fun, but also scary. Check out the video. It truly defies belief:


https://www.youtube.com/watch?v=kSLJriaOumA

READ MORE HERE:
https://www.tomsguide.com/us/nvidia-ai- ... 28869.html



elfismiles » 06 Jan 2014 14:31 wrote:Just thought it's time may have come ...

* Electronic Art's Majestic ARG circa 2001
* El Centro / Ong's Hat / Incunabula (not a traditional ARG)
* Staged Protests for promoting EA's game Dante's Inferno circa 2009
* Vancouver Space Centre hoaxes baseball UFO sighting circa 2013

There are many more examples I might pull into this thread that have already been discussed on RI.

The latest appears to be this Cali Crop Circle...

Crop circle: No, it wasn't aliens, it was marketers [It was marketing for NVIDIA]
By Tom Watkins, CNN / updated 7:22 AM EST, Mon January 6, 2014
http://www.cnn.com/2014/01/06/tech/cali ... rcle-hoax/

<snip>


Re: An Information Apocalypse
seemslikeadream » 12 Feb 2018 19:15 wrote:
He Predicted The 2016 Fake News Crisis. Now He's Worried About An Information Apocalypse.
“What happens when anyone can make it appear as if anything has happened, regardless of whether or not it did?" technologist Aviv Ovadya warns.

Posted on February 11, 2018, at 7:45 p.m.

Charlie Warzel
BuzzFeed News Reporter
. . .
For those paying close attention to developments in artificial intelligence and machine learning, none of this feels like much of a stretch. Software currently in development at the chip manufacturer Nvidia can already convincingly generate hyperrealistic photos of objects, people, and even some landscapes by scouring tens of thousands of images. Adobe also recently piloted two projects — Voco and Cloak — the first a "Photoshop for audio," the second a tool that can seamlessly remove objects (and people!) from video in a matter of clicks.
. . .
Unfortunately, a lot of the warning signs have already happened.”
https://www.buzzfeed.com/charliewarzel/ ... .kl8wNx6q6
User avatar
elfismiles
 
Posts: 8511
Joined: Fri Aug 11, 2006 6:46 pm
Blog: View Blog (4)


Return to General Discussion

Who is online

Users browsing this forum: Belligerent Savant and 54 guests