Innovation
Blockchain Experts Debate Just How Much Internet Needs to Change
Congressional Internet Caucus panelists split over how prepared the internet is for changes from new technologies.

WASHINGTON, October 29, 2021 – Blockchain experts debated how prepared the internet is to meet the challenges of the modern world as new, so-called “Web 3.0” technologies continues to progress.
During a panel discussion on “Is The Past A Prologue To The Fight For Web3?” at a Congressional Internet Caucus Academy event on October 21, these experts said that the decentralized financial tools like blockchain are or will be an impetus for many new changes.
Blockchain provides decentralized recording of transactions across many computers for the records of the cryptocurrency bitcoin. Some say blockchain will change the way the internet runs and serve as the basis for Web 3.0, allowing better user security with regards to data tracking. To implement blockchain technology, more users and service providers will first need to show interest in the technology and an internet must be developed where computing is decentralized to individuals’ own consoles.
The Senate introduced controversial language on cryptocurrency to the Infrastructure Investment and Jobs Act last month that has spurred renewed discussion about the blockchain.
That bill includes tax reporting requirements for wallet developers and cryptocurrency miners that blockchain enthusiasts criticize as overly burdensome. There are different opinions about whether these provisions will be addressed via amendment, or included in the measure when it is ultimately teed up for passage by the House.
Panelists at the event reacted to this development by discussing whether proper groundwork is being laid to support future advancements in cryptocurrency.
Cleve Mesidor, founder of the National Policy Network of Women of Color in Blockchain, and Carlos Acevedo, senior director of sales at Brave Software, criticized the preparedness of current internet structures to handle new developments in cryptocurrency. Both emphasized access inequities present with internet structures which prevent certain demographics from accessing functional internet service, which in the case of blockchain computing presents itself as a learning curve for those at a knowledge deficit with regards to Web 3.0.
“Accessibility is something we should be talking about,” said Mesidor.
Kevin Werbach, professor at the University of Pennsylvania, and Andrew McLaughlin, president of Assembly OSM, gave slightly more credit to current in terms of their preparedness to handle cryptocurrency of the future while still acknowledging limitations. Werbach stated that it is important to realize that even with these limitations the internet has accomplished a lot, and even Web 3.0 would not be able to make current web systems completely decentralized and solve all issues at hand.
McLaughlin pointed out that most extreme moral panics concerning cryptocurrency are exaggerated, as dangerous activity such as sharing of child sex abuse imagery is not as widespread within cryptocurrency as it is made to seem.
The session was moderated by Danny O’Brien, senior fellow at the Filecoin Foundation for the Decentralized Web.
Artificial Intelligence
Experts Debate Artificial Intelligence Licensing Legislation
Licensing requirements will distract from wide scale testing and will limit competition, an event heard.

WASHINGTON, May 23, 2023 – Experts on artificial intelligence disagree on whether licensing is the proper legislation for the technology.
If adopted, licensing requirements would require companies to obtain a federal license prior to developing AI technology. Last week, OpenAI CEO Sam Altman testified that Congress should consider a series of licensing and testing requirements for AI models above a threshold of capability.
At a Public Knowledge event Monday, Aalok Mehta, head of US Public Policy at OpenAI, added licensing is a means to ensuring that AI developers put together safety practices. By establishing licensing rules, we are developing external validation tools that will improve consumer experience, he said.
Generative AI — the model used by chatbots including OpenAI’s widely popular ChatGPT and Google’s Bard — is AI designed to produce content rather than simply processing information, which could have widespread effects on copyright disputes and disinformation, experts have said. Many industry experts have urged for more federal AI regulation, claiming that widespread AI applications could lead to broad societal risks including an uptick in online disinformation, technological displacement, algorithmic discrimination, and other harms.
Some industry leaders, however, are concerned that calls for licensing are a way of shutting the door to competition and new startups by large companies like OpenAI and Google.
B Cavello, director of emerging technologies at the Aspen Institute, said Monday that licensing requirements place burdens on competition, particularly small start-ups.
Implementing licensing requirements can place a threshold that defines a set of players allowed to play in the AI space and a set that are not, said B. Licensing can make it more difficult for smaller players to gain traction in the competitive space, B said.
Already the resources required to support these systems create a barrier that can be really tough to break through, B continued. While there should be mandates for greater testing and transparency, it can also present unique challenges we should seek to avoid, B said.
Austin Carson, founder and president of SeedAI, said a licensing model would not get to the heart of the issue, which is to make sure AI developers are consciously testing and measuring their own models.
The most important thing is to support the development of an ecosystem that revolves around assurance and testing, said Carson. Although no mechanisms currently exist for wide-scale testing, it will be critical to the support of this technology, he said.
Base-level testing at this scale will require that all parties participate, Carson emphasized. We need all parties to feel a sense of accountability for the systems they host, he said.
Christina Montgomery, AI ethics board chair at IBM, urged Congress to adopt precision regulation approach to AI that would govern AI in specific use cases, not regulating the technology itself in her testimony last week.
Advanced Energy
Lawmakers Debate Clean Energy’s Role in Resolving Supply Chain Fragility, China Concerns
The U.S. accounts for just 10 percent of electric vehicle production and 7 percent of battery production.

WASHINGTON, May 23, 2023 — As historic federal investments in infrastructure, manufacturing and clean energy begin to take effect, House lawmakers on Tuesday clashed over the best response to shared concerns about competition with China and the instability of domestic electricity sector supply chains.
Tuesday’s hearing, convened by the House Oversight and Investigations Subcommittee, came just one day after the Department of Energy announced that it would no longer be awarding a $200 million grant to Texas-based battery company Microvast, following intense scrutiny from Republican lawmakers over the company’s ties to China.
Subcommittee Chair Morgan Griffith, R-Va., praised the decision but expressed continued skepticism about the department’s vetting processes — emphasizing the absence of Office of Manufacturing Director David Howell, who declined an invitation to testify.
By failing to appear at the hearing, “the Department of Energy not only refused to provide transparency to this committee, but they’re refusing to be transparent to the American people, who deserve every assurance that their tax dollars are not being funneled to China,” said Rep. Cathy McMorris Rodgers, R-Wash., chair of the full Energy and Commerce Committee.
But Committee Ranking Member Frank Pallone, D-N.J., argued that the award reversal was proof that the department “is taking its stewardship of taxpayer money very seriously.”
Pallone also blamed the Republican majority for Howell’s absence, claiming that the hearing had originally been planned for June and that department officials were working to coordinate the schedule “when all of sudden the date changed earlier this month to today.”
The proposed Microvast award would have been part of a $2.8 billion investment in battery manufacturing funded by the Infrastructure Investment and Jobs Act of 2021, intended to accelerate and strengthen a domestic supply chain. The United States currently accounts for just 10 percent of global electric vehicle production and 7 percent of battery production capacity.
By contrast, China produces three-quarters of all lithium-ion batteries and is home to the majority of production and processing capacity for several key battery components. China was responsible for half of the growth of the electric vehicle market in 2021, according to a recent International Energy Agency report.
While the Biden administration has taken steps to bolster U.S. manufacturing through the IIJA and the Inflation Reduction Act of 2022, Republicans have broadly criticized this approach — particularly the IRA’s focus on clean energy.
Because of the current manufacturing imbalance, the administration’s emphasis on electric vehicles will largely benefit China rather than the U.S., argued Diana Furchtgott-Roth, director of the Heritage Foundation’s Center for Energy, Climate and Environment.
“Heavily subsidizing renewable energy and shoveling money in the form of financial awards out the door is not the solution,” McMorris Rodgers said. “These policies undermine our energy security and financially burden Americans already struggling with high cost of living — and would leave us even more reliant on China.”
However, a study commissioned by Third Way and Breakthrough Energy indicated that these pieces of legislation, alongside the CHIPS and Science Act of 2022, are “fostering investments in and a reshoring of the clean energy supply chain and the associated manufacturing base,” said Ellen Hughes-Cromwick, senior resident fellow for the climate and energy program at Third Way.
Hughes-Cromwick urged lawmakers to allow time for supply chain restructuring, saying that policy stability over multiple years is crucial to allow companies to make sustainable adjustments.
In order to achieve domestic clean energy goals, Congress should first modernize permitting, said Jeremy Harrell, chief strategy officer at ClearPath.
“The single largest mover of private sector investment is regulatory certainty,” Harrell claimed. “Never has the phrase ‘time is money’ been more appropriate — developers can only build new energy infrastructure as fast as federal, state and local governments can permit them, and right now that’s not fast enough.”
Artificial Intelligence
Senate Witnesses Call For AI Transparency
Regulatory AI transparency will increase federal agency and company accountability to the public.

WASHINGTON, May 16, 2023 – Congress should increase regulatory requirements for transparency in artificial intelligence while adopting the technology in federal agencies, said witnesses at a Senate Homeland Security and Governmental Affairs Committee hearing on Tuesday.
Many industry experts have urged for more federal AI regulation, claiming that widespread AI applications could lead to broad societal risks including an uptick in online disinformation, technological displacement, algorithmic discrimination, and other harms.
The hearing addressed implementing AI in federal agencies. Congress is concerned about ensuring that the United States government is prepared to capitalize on the capabilities afforded by AI technology while also protecting the constitutional rights of citizens, said Sen. Gary Peters, D-Michigan.
The United States “is suffering from a lack of leadership and prioritization on these topics,” stated Lynne Parker, director of AI Tennessee Initiative at the University of Tennessee in her comments.
In a separate hearing Tuesday, CEO of OpenAI Sam Altman said that is is “essential that powerful AI is developed with democratic values in mind which mean US leadership is critical.”
Applications of AI are immensely beneficial, said Altman. However, “we think that regular intervention by governments will be crucial to mitigate the risks of increasingly powerful models.”
To do so, Altman suggested that the U.S. government consider a combination of licensing and testing requirements for the development and release of AI models above a certain threshold of capability.
Companies like OpenAI can partner with governments to ensure AI models adhere to a set of safety requirements, facilitate efficient processes, and examine opportunities for global coordination, he said.
Building accountability into AI systems
Siezing this moment to modernize the government’s systems will strengthen the country, said Daniel Ho, professor at Stanford Law School, encouraging Congress to lead by example to implement accountable AI practices.
An accountable system ensures that agencies are responsible to report to the public and those that AI algorithms directly affect, added Richard Eppink of the American Civil Liberties Union of Idaho Foundation.
A serious risk to implementing AI is that it can conceal how the systems work, including the bad data that they could be trained on, said Eppink. This can prevent accountability to the public and puts citizen’s constitutional rights at risk, he said.
To prevent this, the federal government should implement transparency requirements and governance standards that would include transparency during the implementation process, said Eppink. Citizens have the right to the same information that the government has so we can maintain accountability, he concluded.
Parker suggested that Congress appoint a Chief AI Director at each agency that would help develop Ai strategies for each agency and establish an interagency Chief AI Council to govern the use of the technology in the Federal government.
Getting technical talent into the workforce is the predicate to a range of issues we are facing today, agreed Ho, claiming that less than two percent of AI personnel is in government. He urged Congress to establish pathways and trajectories for technical agencies to attract AI talent to public service.
Congress considers AI regulation
Congress’s attention has been captured by growing AI regulatory concerns.
In April, Senator Check Schumer, D-N.Y., proposed a high-level AI policy framework focused on ensuring transparency and accountability by requiring companies to allow independent experts to review and test AI technologies and make results available publicly.
Later in April, Representative Yvette Clarke, D-N.Y., introduced a bill that would require the disclosure of AI-generated content in political ads.
The Biden Administration announced on May 4 that it will invest $140 million in funding to launch seven new National AI Research Institutes, which investment will bring the total number of Institutes to 25 across the country.
-
Broadband Mapping & Data3 weeks ago
Video of CostQuest CEO Jim Stegeman at Digital Infrastructure Investment Summit
-
Open Access2 weeks ago
AT&T Closes Open Access Fiber Deal With BlackRock
-
Broadband Roundup3 weeks ago
Starlink Likes FCC Direction on 12 GHz, Verizon & Comcast Urge ACP Funding, FCC Head on ACP Tour
-
Broadband Roundup4 weeks ago
Problematic ACP Qualification Standard, Macquarie Invests in Pavlov, Cogent Closes T-Mobile Wireline Deal
-
#broadbandlive4 weeks ago
Broadband Breakfast on May 10, 2023 – GUMBO and Louisiana’s Broadband Progress
-
Broadband Roundup3 weeks ago
New ACP Landing Page, Cellular Association Wants More Mid-Band Spectrum, New Ezee Fiber CEO
-
Expert Opinion2 weeks ago
Scott Wallsten: A $10 Billion Broadband Black Hole?
-
5G2 weeks ago
Crown Castle CEO Says 5G Plus Fixed Wireless Can Rival Fiber Connections