[BreachExchange] Deepfake Audio Scores $35M in Corporate Heist

Terrell Byrd terrell.byrd at riskbasedsecurity.com
Wed Oct 20 16:00:57 EDT 2021


https://www.darkreading.com/attacks-breaches/deepfake-audio-scores-35-million-in-corporate-heist


A group of fraudsters made off with $35 million after using forged email
messages and deepfake audio to convince an employee of a United Arab
Emirates company that a director requested the money as part of an
acquisition of another organization, according to a US federal court
request filed last week.

The attack targeted a branch manager with emails that appeared to be from
the director and a US-based lawyer, who the emails designated as
coordinator of the acquisition. This attack is the latest to use synthetic
audio created using machine-learning algorithms, known as neural networks,
to mimic the voice of a person known to the targeted employee.

For that reason, deepfake audio and synthesized voices will likely become
part of cybercriminals' techniques in the future. A variety of open source
tools are available to allow anyone to create deepfakes, both video and
audio, says Etay Maor, senior director security strategy at network
security firm Cato Networks.

"If there is money to be made, you can be sure that attackers will adopt
new techniques," Maor says. "It's not super-sophisticated to use such
tools. When it comes to a voice, it is even easier."

The corporate heist is the second known attack using deepfake technology.
In 2019, a manager of a UK subsidiary of a German company received a call
from what sounded like his Germany-based CEO, who he had previously met. At
the fake CEO's request, he transferred €220,000 to a supposed vendor. The
manager did not become suspicious until the same person posing as the CEO
called again two days later, asking for another €100,000. He then noticed
that the phone number came from Austria, not Germany.

The success of these attacks goes back to trust, says Maor. A call from
someone you know asking for money is different than an email claiming to be
a Nigerian prince. An employee talking to a person, who they believe is
their CEO, will be more likely to transfer money.

The solution for most companies will have to go back to "trust, but
verify," he says.

"We are going to have to adopt some of the principles of zero trust into
this world of relationships," he says. "It does not have to be a
technological solution. A process of verifying may be enough."

The US Department of Justice filing has few details of the United Arab
Emirates investigation. A US-based lawyer allegedly had been designated to
oversee the acquisition, and the Emirati investigation tracked two
transfers totaling $415,000 deposited into accounts at Centennial Bank in
the United States.

"In January 2020, funds were transferred from the Victim Company to several
bank accounts in other countries in a complex scheme involving at least 17
known and unknown defendants," stated the request to the US District Court
for the District of Columbia. "Emirati authorities traced the movement of
the money through numerous accounts and identified two transactions to the
United States."

The request asked the courts to designate a DoJ lawyer to be the point of
contact in the US for the investigation.

While the technology to create realistic fake audio and video of people
using generative adversarial neural networks (GANs) has fueled fears of
deepfakes wreaking havoc in political campaigns, and of wrongdoers claiming
actual evidence was created by deep neural-network technology, so far most
examples have been proof of concepts, outside of an underground market for
fake celebrity pornography and revenge pornography.

Yet the technical requirements are no longer a hurdle for anyone who wants
to create deepfakes. Maor estimates it takes less than five minutes of
sampled audio to create a convincing synthesized voice, but other estimates
put the necessary raw audio at two to three hours of samples. Lesser
quality synthesis takes a lot less time. For many business executives,
attackers can pull the necessary audio from the Internet.

Companies do not need special technology to defeat deepfake-fueled business
process compromises. Instead, they need to add verification steps to their
accounting processes, says Maor.

"If you have the right processes in place, you can weed out these issues,"
he says. "At the end of the day, a simple phone call to verify the request
could have prevented this."
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.riskbasedsecurity.com/pipermail/breachexchange/attachments/20211020/9ebb7cf2/attachment.html>


More information about the BreachExchange mailing list