Artificial Intelligence
Deepfakes Could Pose A Threat to National Security, But Experts Are Split On How To Handle It
Experts disagree on the right response to video manipulation — is more tech or a societal shift the right solution?

June 3, 2021—The emerging and growing phenomenon of video manipulation known as deepfakes could pose a threat to the country’s national security, policy makers and technology experts said at an online conference Wednesday, but how best to address them divided the panel.
A deepfake is a highly technical method of generating synthetic media in which a person’s likeness is inserted into a photograph or video in such a way that creates the illusion that they were actually there. A well done deepfake can make a person appear to do things that they never actually did and say things that they never actually said.
“The way the technology has evolved, it is literally impossible for a human to actually detect that something is a deepfake,” said Ashish Jaiman, the director of technology operations at Microsoft, at an online event hosted by the Information Technology and Innovation Foundation.
Experts are wary of the associated implications of this technology being increasingly offered to the general population, but how best to address the brewing dilemma has them split. Some believe better technology aimed at detecting deepfakes is the answer, while others say that a shift in social perspective is necessary. Others argue that such a societal shift would be dangerous, and that the solution actually lies in the hands of journalists.
Deepfakes pose a threat to democracy
Such technology posed no problem when only Hollywood had the means to portray such impressive special effects, says Rep. Anthony Gonzalez, R-Ohio, but the technology has progressed to a point that allows most anybody to get their hands on it. He says that with the spread of disinformation, and the challenges that poses to establishing a well-informed public, deepfakes could be weaponized to spread lies and affect elections.
As of yet, however, no evidence exists that deepfakes have been used for this purpose, according to Daniel Kimmage, the acting coordinator for the Global Engagement Center of the Department of State. But he, along with the other panelists, agree that the technology could be used to influence elections and further already growing seeds of mistrust in the information media. They believe that its best to act preemptively and solve the problem before it becomes a crisis.
“Once people realize they can’t trust the images and videos they’re seeing, not only will they not believe the lies, they aren’t going to believe the truth,” said Dana Rao, executive vice president of software company Adobe.
New technology as a solution
Jaiman says Microsoft has been developing sophisticated technologies aimed at detecting deepfakes for over two years now. Deborah Johnson, emeritus technology professor at the University of Virginia School of Engineering, refers to this method as an “arms race,” in which we must develop technology that detects deepfakes at a faster rate than the deepfake technology progresses.
But Jaiman was the first to admit that, despite Microsoft’s hard work, detecting deepfakes remains a grueling challenge. Apparently, it’s much harder to detect a deepfake than it is to create one, he said. He believes that a societal response is necessary, and that technology will be inherently insufficient to address the problem.
Societal shift as a solution
Jaiman argues that people need to be skeptical consumers of information. He believes that until the technology catches up and deepfakes can more easily be detected and misinformation can easily be snuffed, people need to approach online information with the perspective that they could easily be deceived.
But critics believe this approach of encouraging skepticism could be problematic. Gabriela Ivens, the head of open source research at Human Rights Watch, says that “it becomes very problematic if people’s first reactions are not to believe anything.” Ivens’ job revolves around researching and exposing human rights violations, but says that the growing mistrust of media outlets will make it harder for her to gain the necessary public support.
She believes that a “zero-trust society” must be resisted.
Vint Cerf, the vice president and chief internet evangelist at Google, says that it is up to journalists to prevent the growing spread of distrust. He accused journalists not of deliberately lying, but often times misleading the public. He believes that the true risk of deepfakes lies in their ability to corrode America’s trust in truth, and that it is up to journalists to restore that trust already beginning to corrode by being completely transparent and honest in their reporting.
Artificial Intelligence
Bennet, Young, and Warner Propose Legislation to Enhance U.S. Technology Competitiveness
New bill seeks to supports U.S. role in emerging technologies, including artificial intelligence.

WASHINGTON, June 9, 2023 – Citing threats from China, two Democratic and one Republican senator have introduced the Global Technology Leadership Act that would create an Office of Global Competition Analysis.
The new office would be tasked with assessing U.S. leadership in science, technology and innovation in advanced manufacturing, workforce development, supply chain resilience and research and development initiatives.
“We cannot afford to lose our competitive edge in strategic technologies like semiconductors, quantum computing, and artificial intelligence to competitors like China,” said Sen. Michael Bennet, D-Colo., one of the three sponsors, together with Mark Warner, D-Virginia, and Todd Young, R-Indiana.
On a periodic basis, the director of the Office of Science and Technology Policy, presidential assistants for economic policy, national security and the heads of such agencies of OSTP and the White House deem appropriate would determine the priorities of the office.
Bennet said that the office’s assessments would inform policymakers and help enhance American leadership in strategic innovation.
Artificial Intelligence
U.S. Must Take Lead on Global AI Regulations: State Department Official
Call for leadership comes during pivotal time in AI development.

WASHINGTON, May 31, 2023 – A State Department official is calling for a United States-led global coalition to set artificial intelligence regulations.
“This is the exact moment where the US needs to show leadership,” Jennifer Bachus, assistant secretary of state for Cyberspace and Digital Policy, said last week on a panel discussing international principles on responsible AI. “This is a shared problem and we need a shared solution.”
She opposed pitting the U.S. and China against one another in the AI race, saying it would “ultimately always lead to a problem.” Instead, Bachus called for an alliance of the United States, the European Union, and Japan to take the lead in creating a legal framework to govern artificial intelligence.
The introduction of OpenAI’s ChatGPT earlier this year sent tech companies in a rush to create their own generative AI chatbot systems. Competition between tech giants has heated up with the recent release of Google’s Bard and Microsoft’s Bing chatbot. Similar to ChatGPT in terms of its vast language model, these chatbots can also access data from the internet to answer queries or carry out tasks.
Experts are concerned about the dangers posed by this unprecedented technology. On Tuesday, hundreds of tech experts and industry leaders, including OpenAI’s CEO Sam Altman, signed a one-sentence statement calling the existential threats presented by A.I. a “global priority” on par with “pandemics and nuclear conflicts.” Earlier in March, Elon Musk joined several AI experts signing another open letter urging for a pause on “giant AI experiments.”
Despite the pressing concerns about generative AI, there is rising criticism that policymakers are slow to put forth adequate legislation for this nascent technology. Panelists argued this is partly because legislators have difficulty understanding technological innovations. Michelle Giuda, director of Krach Institute for Tech Diplomacy, argued for a more proactive contribution from the academic community and tech firms.
“There is a risk of relying too much on the government to regulate ahead of where innovation is going and providing the clarity that’s needed,” said Giuda. “We all know that the government isn’t going to stay ahead of the innovation curve, but this is an ongoing dialogue between tech companies, governments and civil society.”
Microsoft’s Chief Responsible AI Officer, Natasha Crampton, agreed that developers and experts in the field must play a central role in crafting and implementing legislation pertaining to artificial intelligence. She did, however, mention that businesses using AI technology should also share part of the responsibility.
“It is our job to make sure that safety and responsibility is baked into these systems from the very beginning,” said Crampton. “Making sure that you are really holding developers to very high standards but also deployers of technology in some aspects as well.”
Earlier in May, Sens. Michael Bennet, D-C.O., and Peter Welch, D-VT. introduced a bill to establish a government agency to oversee artificial intelligence. The Joe Biden administration also announced $140 million in funding to establish seven new National AI Research institutions, increasing the total number of institutions in the nation to 25.
Artificial Intelligence
AI is a Key Component in Effectively Managing the Energy Grid
The ability to balance the grid’s supply and demand in real time will become extremely complex.

WASHINGTON, May 30, 2023 – Artificial intelligence will be required to effectively manage and optimize a more complex energy grid, said experts at a United States Energy Association event Tuesday.
Renewable energy technologies such as solar panels, electric vehicles, and power walls add large amounts of energy storage to the grid, said Jeremy Renshaw, senior technical executive at the Electric Power Research Institute. Utility companies are required to manage many bidirectional resources that both store and use energy, he said.
“The grid of the future is going to be significantly more complicated,” said Renshaw. Having humans operate the grid will be economically infeasible, he continued, claiming that AI will drastically improve operations.
The ability to balance the grid’s supply and demand in real time will become extremely complex with the adoption of these new technologies, added Marc Spieler, leader for global business development at AI hardware and software supplier, Nvidia.
Utility companies will need to redirect traffic in real time to support the incoming demand, he said. AI enables real time redirecting of traffic and an understanding of the capacity of the grid at any point, said Spieler.
Moreover, AI can identify what changes need to be made to avoid waste by over generating electricity and black outs by under generating, he said. AI also has the capability to predict and plan for extreme weather that can be hazardous to electrical infrastructure and can identify bottleneck areas where infrastructure needs to be updated, said Spieler.
Human management will still be required to ensure that systems are operated responsibly, said John Savage, professor of computer science at Brown University. Utility companies should avoid allowing AI to make unsupervised decisions especially for unforeseen scenarios, he said.
The panelists envision AI as a decision support mechanism to help humans make more informed decisions, agreed the panelists. The technology will replace jobs that deal with mundane and repetitive tasks but will ultimately create more jobs in new positions, said Renshaw.
This comes several weeks after industry experts urged Congress to implement federal AI regulation.
-
Open Access4 weeks ago
AT&T Closes Open Access Fiber Deal With BlackRock
-
Expert Opinion3 weeks ago
Scott Wallsten: A $10 Billion Broadband Black Hole?
-
Digital Inclusion4 weeks ago
Debra Berlyn: Creating a Path to Close the Digital Divide for Older Adults
-
5G4 weeks ago
Crown Castle CEO Says 5G Plus Fixed Wireless Can Rival Fiber Connections
-
NTIA3 weeks ago
NTIA Should Remove Letter of Credit Requirement in BEAD Program, Event Hears
-
Broadband Roundup2 weeks ago
‘Urgent’ Social Media Advisory, Tribal Broadband Awards, Permitting Reform Progress, BroadbandNow Podcast
-
Artificial Intelligence3 weeks ago
Senate Witnesses Call For AI Transparency
-
#broadbandlive3 weeks ago
Broadband Breakfast on May 31, 2023 – Tribal Broadband Deployment