Artificial Intelligence
As ChatGPT’s Popularity Skyrockets, Some Experts Call for AI Regulation
As generative AI models grow more sophisticated, they present increasing risks.

WASHINGTON, February 3, 2023 — Just two months after its viral launch, ChatGPT reached 100 million monthly users in January, reportedly making it the fastest-growing consumer application in history — and raising concerns, both internal and external, about the lack of regulation for generative artificial intelligence.
Many of the potential problems with generative AI models stem from the datasets used to train them. The models will reflect whatever biases, inaccuracies and otherwise harmful content was present in their training data, but too much dataset filtering can detract from performance.
OpenAI has grappled with these concerns for years while developing powerful, publicly available tools such as DALL·E — an AI system that generates realistic images and original art from text descriptions, said Anna Makanju, OpenAI’s head of public policy, a Federal Communications Bar Association event on Friday.
“We knew right off the bat that nonconsensual sexual imagery was going to be a problem, so we thought, ‘Why don’t we just try to go through the dataset and remove any sexual imagery so people can’t generate it,’” Makanju said. “And when we did that, the model could no longer generate women, because it turns out most of the visual images that are available to train a dataset on women are sexual in nature.”
Despite rigorous testing before ChatGPT’s release, early users quickly discovered ways to evade some of the guardrails intended to prevent harmful uses.
The model would not generate offensive content in response to direct requests, but one user found a loophole by asking it to write from the perspective of someone holding racist views — resulting in several paragraphs of explicitly racist text. When some users asked ChatGPT to write code using race and gender to determine whether someone would be a good scientist, the bot replied with a function that only selected white men. Still others were able to use the tool to generate phishing emails and malicious code.
OpenAI quickly responded with adjustments to the model’s filtering algorithms, as well as increased monitoring.
“So far, the approach we’ve taken is we just try to stay away from areas that can be controversial, and we ask the model not to speak to those areas,” Makanju said.
The company has also attempted to limit certain high-impact uses, such as automated hiring. “We don’t feel like at this point we know enough about how our systems function and biases that may impact employment, or if there’s enough accuracy for there to be an automated decision about hiring without a human in the loop,” Makanju explained.
However, Makanju noted that future generative language models will likely reach a point where users can significantly customize them based on personal worldviews. At that point, strong guardrails will need to be in place to prevent the model from behaving in certain harmful ways — for example, encouraging self-harm or giving incorrect medical advice.
Those guardrails should probably be established by external bodies or government agencies, Makanju said. “We recognize that we — a pretty small company in Silicon Valley — are not the best place to make a decision of how this will be used in every single domain, as hard as we try to think about it.”
Little AI regulation currently exists
So far, the U.S. has very little legislation governing the use of AI, although some states regulate automated hiring tools. On Jan. 26, the National Institute of Standards and Technology released the first version of its voluntary AI risk management framework, developed at the direction of Congress.
This regulatory crawl is being rapidly outpaced by the speed of generative AI research. Google reportedly declared a “code red” in response to ChatGPT’s release, speeding the development of multiple AI tools. Chinese tech company Baidu is planning to launch its own AI chatbot in March.
Not every company will respond to harmful uses as quickly as OpenAI, and some may not even attempt to stop them, said Claire Leibowicz, head of AI and media integrity at the Partnership on AI. PAI is a nonprofit coalition that develops tools recommendations for AI governance.
Various private organizations, including PAI, have laid out their own ethical frameworks and policy recommendations. There is ongoing discussion about the extent to which these organizations, government agencies and tech companies should be determining AI regulation, Leibowicz said.
“What I’m interested in is, who’s involved in that risk calculus?” she asked. “How are we making those decisions? What types of actual affected communities are we talking to in order to make that calculus? Or is it a group of engineers sitting in a room trying to forecast for the whole world?”
Leibowicz advocated for transparency measures such as requiring standardized “nutrition labels” that would disclose the training dataset for any given AI model — a proposal similar to the label mandate announced in November for internet service providers.
A regulatory framework should be implemented while these technologies are still being created, rather than in response to a future crisis, Makanju said. “It’s very clear that this technology is going to be incorporated into every industry in some way in the coming years, and I worry a little bit about where we are right now in getting there.”
Artificial Intelligence
Sam Altman to Rejoin OpenAI, Tech CEOs Subpoenaed, EFF Warns About Malware
Altman was brought back to OpenAI only days after being fired.

November 22, 2023 – OpenAI announced in an X post early Wednesday morning that Sam Altman will be re-joining the company that built ChatGPT as CEO after he was fired on Friday.
Altman confirmed his intention to rejoin OpenAI in an X post Wednesday morning, saying that he was looking forward to returning to OpenAI with support from the new board.
Former company president Greg Brockman also said Wednesday he will return to the AI company.
Altman and Brockman will join with a newly formed board, which includes former Salesforce co-CEO Bret Taylor as the chair, former US Treasury Secretary Larry Summers, and Quora CEO Adam D’Angelo, who previously held a position on the OpenAI board.
Satya Nadella, the CEO of OpenAI backer Microsoft, echoed support for both Brockman and Altman rejoining OpenAI, adding that he is looking forward to continuing building a relationship with the OpenAI team in order to best deliver AI services to customers.
OpenAI received backlash from several hundred employees who threatened to leave and join Microsoft under Altman and Brockman unless the current board of directors agreed to resign.
Tech CEOs subpoenaed to attend hearing
Sens. Dick Durbin, D-Illinois, and Lindsey Graham, R-South Carolina, announced Monday that tech giants Snap, Discord and X have been issued subpoenas for their appearance at the Senate Judiciary Committee on December 6 in relation to concerns over child sexual exploitation online.
Snap CEO Evan Spiegel, X CEO Linda Yaccarino and Discord CEO Jason Citron have been asked to address how or if they’ve worked to confront that issue.
Durbin said in a press release that the committee “promised Big Tech that they’d have their chance to explain their failures to protect kids. Now’s that chance. Hearing from the CEOs of some of the world’s largest social media companies will help inform the Committee’s efforts to address the crisis of online child sexual exploitation.”
Durbin noted in a press release that both X and Discord refused to initially accept subpoenas, which required the US Marshal Service to personally deliver those respective documents.
The committee is looking to have Meta CEO Mark Zuckerberg and TikTok CEO Shou Zi Chew testify as well but have not received confirmation regarding their attendance.
Several bipartisan bills have been brought forth to address that kind of exploitation, including the Earn It Act, proposed by Sens. Richard Blumenthal, D-Connecticut, and Graham, which holds them liable under child sexual abuse material laws.
EFF urging FTC to sanction sellers of malware-containing devices
The Electronic Frontier Foundation, a non-profit digital rights group, have asked the Federal Trade Commission in a letter on November 14 to sanction resellers like Amazon and AliExpress following allegations mobile devices and Android TV boxes purchased from their stores contain malware.
The letter explained that once the devices were turned on and connected to the internet, they would begin “communicating with botnet command and control (C2) servers. From there, these devices connect to a vast click-fraud network which a report by HUMAN Security recently dubbed BADBOX.”
The EFF added that this malware is often operating unbeknownst to the consumer, and without advanced technical knowledge, there is nothing they can do to remedy it themselves.
“These devices put buyers at risk not only by the click-fraud they routinely take part in, but also the fact that they facilitate using the buyers’ internet connections as proxies for the malware manufacturers or those they sell access to,” explained the letter.
EFF said that the devices containing malware included ones manufactured by Chinese companies AllWinner and RockChip, who have been reported on for sending out products with malware before by EFF.
Artificial Intelligence
Sam Altman to Join Microsoft, New FCC Broadband Map, Providers Form 4.9 GHz Coalition
After being fired on Friday by the board of OpenAI, former CEO Altman will join Microsoft and lead its AI.

November 20, 2023 – Microsoft CEO Satya Nadella announced in an X post Monday that former OpenAI CEO Sam Altman will be joining Microsoft after being fired from the machine learning company.
Over the course of the last four days, OpenAI has undergone several shifts in leadership, which includes OpenAI investor Microsoft acquiring OpenAI president and chairman Greg Brockman to lead an AI research team alongside Altman.
Brockman, who had been concurrently relieved from his role as chairman of the OpenAI board, announced his resignation Friday via X, upon learning that the board had decided to fire Altman.
OpenAI said in a blog post Friday that Altman “was not consistently candid in his communications with the board, hindering its ability to exercise its responsibilities.”
OpenAI then notified The Information Saturday that Emmett Shear, co-founder of streaming site Twitch, would serve as CEO after having CTO Mira Murati serve that role in the interim.
Following Nadella’s announcement Monday morning, nearly 500 of the 700 OpenAI employees were signatories to a letter threatening to leave their roles to work under Altman and Brockman at Microsoft unless all of the current board members resign.
As of Monday, OpenAI board member Ilya Sutskever posted a message of regret on X regarding the board decision to remove Altman and Brockman. The phrase “OpenAI is nothing without its people,” is now emerging from employee’s X accounts.
FCC announces new national broadband map
The head of the Federal Communication Commission announced Friday the third iteration of its national broadband map, showing just over 7.2 million locations lack access to high-speed internet.
That is less than the 8.3 million identified in May.
FCC Chairwoman Jessica Rosenworcel noted that map data continue to fluctuate less between iterations, showing improvements in map accuracy.
Previous iterations of the national broadband map had been criticized for not accurately depicting areas with and without service, with widespread concern that that would impact the allocation of Broadband Equity, Access and Deployment funding.
The map outlines where adequate broadband service is and is not available throughout the nation and provides viewers with information on the providers who service those areas and the technology used to do so.
Providers form spectrum advocacy coalition
A group of telecom industry players including Verizon and T-Mobile announced Thursday the formation of the Coalition for Emergency Response and Critical Infrastructure to advocate for select use of the 4.9 GigaHertz (GHz) spectrum band.
The coalition is in support of prioritizing state and local public safety agencies as main users of the 4.9 GHz band, while ensuring that non-public safety licensees operate on the band to avoid interference.
“Public Safety agencies have vastly different needs from jurisdiction to jurisdiction, and they should decide what compatible non-public-safety use means within their jurisdictions,” read the coalition’s letter.
In January of this year, the FCC adopted a report to manage the use of the 4.9 GHz band, while seeking comment on the role a band manager would play in facilitating license allocation between public safety and non-public safety entities.
It had proposed two methods of operation for the band manager in which it would either lease access rights from public-safety entities and then sublease that to non-public safety entities, or to facilitate direct subleasing between public safety operators and external parties.
In its letter to the FCC, the coalition announced support for the second of those methods stressing the fact that it will allow public safety license holders retain authority over who they sublease their spectrum to.
Artificial Intelligence
FCC Cybersecurity Pilot Program, YouTube AI Regulations, Infrastructure Act Anniversary
The FCC has proposed a pilot program to help schools and libraries protect against cyberattacks.

November 15, 2023 – The Federal Communications Commission proposed Monday a cybersecurity pilot program for schools and libraries, which would require a three-year $200 million investment in ways to best protect K-12 students from cyberattacks.
In addition to going in and assessing what kind of cybersecurity services are best suited for students and school needs, the program would also subsidize the cost of those services used in schools.
The program would serve as a separate Universal Service Fund program, separate from the existing school internet subsidy program called E-Rate.
“This pilot program is an important pathway for hardening our defenses against sophisticated cyberattacks on schools and ransomware attacks that harm our students and get in the way of their learning,” said FCC Chairwoman Jessica Rosenworcel.
The proposal would be a part of the larger Learn Without Limit’s initiative, which supports internet connectivity in schools to help reduce the homework gap by enabling kids’ digital access to digital learning.
YouTube rolling out AI content regulations
Alphabet’s video sharing platform YouTube announced in a blog post Tuesday it will be rolling out AI guidelines over the next few months, which will inform viewers about when they are interacting with “synthetic” or AI-generated content.
The rules will require creators to identify if the video is made of AI content. Creators who don’t disclose that information could see their work flagged and removed, and they may be suspended from the platform or subject to other penalties.
For the viewer, tags will appear in the description panel on videos indicating that if the video is synthetic or AI generated. YouTube noted that for videos dealing with more sensitive topics, it may use more prominent labels.
YouTube’s AI guidelines come at a time when members of Congress and industry leaders are calling for increased effort toward AI regulatory reform, and after President Joe Biden’s executive order on AI guidelines signed into effect in October.
Two-year anniversary of the infrastructure investment jobs act
Thursday marked the second-year anniversary of the Infrastructure, Investment and Jobs Act, which prompted a $400-billion investment into the US economy.
The IIJA pushed for a variety of programs and initiatives, with over 40,000 sector-specific projects having received funding – several of those working to improve the broadband sector.
$65 billion was invested by the IIJA into improving connectivity, which helped to establish the $14-billion Affordable Connectivity Program, which has so-far helped more than 20 million US households get affordable internet through a $30 and $75 subsidy per month.
Outside of ACP, the IIJA called on the National Telecommunications and Information Administration to develop the Broadband Equity Access Deployment program, a $42.5-billion investment into high-speed broadband deployment across all 50 states.
Currently, states are in the process of submitting their BEAD draft proposals, which all outline how states will administer the funding they receive as well as any funding they already have or how they will use broadband mapping data.
-
Fiber3 weeks ago
The High Cost of Fiber is Leading States to Explore Other Technologies
-
Funding3 weeks ago
NTIA Confirms Licensed-by-Rule May Apply for BEAD Funding
-
FCC4 weeks ago
‘It Was Graft’: How the FCC’s CAF II Program Became a Money Sink
-
Expert Opinion3 weeks ago
Ryan Johnston: What Happens to BEAD Without the Affordable Connectivity Program?
-
Digital Inclusion3 weeks ago
Federal Officials Agree: Infrastructure Alone Will Not Close the Digital Divide
-
Digital Inclusion2 weeks ago
Connect20 Summit: The Crucial Role of Digital Skills Training
-
Spectrum3 weeks ago
Biden Administration Announces Plan to Free Up Spectrum
-
FCC3 weeks ago
FCC Approves Strong Digital Discrimination Rules