|
<html> |
|
<title> - ADVANCES IN AI: ARE WE READY FOR A TECH REVOLUTION?</title> |
|
<body><pre> |
|
[House Hearing, 118 Congress] |
|
[From the U.S. Government Publishing Office] |
|
|
|
|
|
|
|
|
|
|
|
|
|
ADVANCES IN AI: ARE WE READY |
|
FOR A TECH REVOLUTION? |
|
|
|
======================================================================= |
|
|
|
HEARING |
|
|
|
before the |
|
|
|
SUBCOMMITTEE ON CYBERSECURITY, INFORMATION |
|
TECHNOLOGY, AND GOVERNMENT INNOVATION |
|
|
|
of the |
|
|
|
COMMITTEE ON OVERSIGHT |
|
AND ACCOUNTABILITY |
|
|
|
HOUSE OF REPRESENTATIVES |
|
|
|
ONE HUNDRED EIGHTEENTH CONGRESS |
|
|
|
FIRST SESSION |
|
|
|
__________ |
|
|
|
MARCH 8, 2023 |
|
|
|
__________ |
|
|
|
Serial No. 118-7 |
|
|
|
__________ |
|
|
|
Printed for the use of the Committee on Oversight and Accountability |
|
|
|
|
|
Available on: govinfo.gov |
|
oversight.house.gov or |
|
docs.house.gov |
|
_________ |
|
|
|
U.S. GOVERNMENT PUBLISHING OFFICE |
|
|
|
51-473 PDF WASHINGTON : 2023 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
COMMITTEE ON OVERSIGHT AND ACCOUNTABILITY |
|
|
|
JAMES COMER, Kentucky, Chairman |
|
|
|
Jim Jordan, Ohio Jamie Raskin, Maryland, Ranking |
|
Mike Turner, Ohio Minority Member |
|
Paul Gosar, Arizona Eleanor Holmes Norton, District of |
|
Virginia Foxx, North Carolina Columbia |
|
Glenn Grothman, Wisconsin Stephen F. Lynch, Massachusetts |
|
Gary Palmer, Alabama Gerald E. Connolly, Virginia |
|
Clay Higgins, Louisiana Raja Krishnamoorthi, Illinois |
|
Pete Sessions, Texas Ro Khanna, California |
|
Andy Biggs, Arizona Kweisi Mfume, Maryland |
|
Nancy Mace, South Carolina Alexandria Ocasio-Cortez, New York |
|
Jake LaTurner, Kansas Katie Porter, California |
|
Pat Fallon, Texas Cori Bush, Missouri |
|
Byron Donalds, Florida Shontel Brown, Ohio |
|
Kelly Armstrong, North Dakota Jimmy Gomez, California |
|
Scott Perry, Pennsylvania Melanie Stansbury, New Mexico |
|
William Timmons, South Carolina Robert Garcia, California |
|
Tim Burchett, Tennessee Maxwell Frost, Florida |
|
Marjorie Taylor Greene, Georgia Becca Balint, Vermont |
|
Lisa McClain, Michigan Summer Lee, Pennsylvania |
|
Lauren Boebert, Colorado Greg Casar, Texas |
|
Russell Fry, South Carolina Jasmine Crockett, Texas |
|
Anna Paulina Luna, Florida Dan Goldman, New York |
|
Chuck Edwards, North Carolina Jared Moskowitz, Florida |
|
Nick Langworthy, New York |
|
Eric Burlison, Missouri |
|
|
|
Mark Marin, Staff Director |
|
Jessica Donlon, Deputy Staff Director and General Counsel |
|
Raj Bharwani, Senior Professional Staff Member |
|
Lauren Lombardo, Senior Policy Analyst |
|
Peter Warren, Senior Advisor |
|
Mallory Cogar, Deputy Director of Operations and Chief Clerk |
|
|
|
Contact Number: 202-225-5074 |
|
|
|
Julie Tagen, Minority Staff Director |
|
Contact Number: 202-225-5051 |
|
------ |
|
|
|
Subcommittee on Cybersecurity, Information Technology, and Government |
|
Innovation |
|
|
|
Nancy Mace, South Carolina, Chairwoman |
|
William Timmons, South Carolina Gerald E. Connolly, Virginia |
|
Tim Burchett, Tennessee Ranking Minority Member |
|
Marjorie Taylor Greene, Georgia Ro Khanna, California |
|
Anna Paulina Luna, Florida Stephen F. Lynch, Massachusetts |
|
Chuck Edwards, North Carolina Kweisi Mfume, Maryland |
|
Nick Langworthy, New York Jimmy Gomez, California |
|
Eric Burlison, Missouri Jared Moskowitz, Florida |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
C O N T E N T S |
|
|
|
---------- |
|
Page |
|
Hearing held on March 8, 2023.................................... 1 |
|
|
|
Witnesses |
|
|
|
Dr. Eric Schmidt, Chair, Special Competitive Studies Project |
|
Oral Statement................................................... 6 |
|
|
|
Dr. Aleksander Mafdry, Director & Cadence Design Systems |
|
Professor of Computing, MIT Center for Deployable Machine |
|
Learning & Massachusetts |
|
Institute of Technology |
|
Oral Statement................................................... 7 |
|
|
|
Dr. Scott Crowder, Vice President & CTO, IBM Quantum/IBM Systems, |
|
Technical Strategy, and Transformation |
|
Oral Statement................................................... 9 |
|
|
|
Ms. Merve Hickok, Chair and Research Director, Center for AI and |
|
Digital Policy |
|
Oral Statement................................................... 11 |
|
|
|
Written opening statements and statements for the witnesses are |
|
available on the U.S. House of Representatives Document |
|
Repository at: docs.house.gov. |
|
|
|
Index of Documents |
|
|
|
---------- |
|
|
|
|
|
* Questions for the Record: to Dr. Crowder; submitted by Rep. |
|
Mace. |
|
|
|
* Questions for the Record: to Dr. Crowder; submitted by Rep. |
|
Connolly. |
|
|
|
* Questions for the Record: to Dr. Schmidt; submitted by Rep. |
|
Mace. |
|
|
|
* Questions for the Record: to Dr. Schmidt; submitted by Rep. |
|
Connolly. |
|
|
|
* Questions for the Record: to Ms. Hickok; submitted by Rep. |
|
Connolly. |
|
|
|
* Questions for the Record: to Dr. Mafdry; submitted by Rep. |
|
Mace. |
|
|
|
* Questions for the Record: to Dr. Mafdry; submitted by Rep. |
|
Connolly. |
|
|
|
|
|
ADVANCES IN AI: ARE WE READY |
|
FOR A TECH REVOLUTION? |
|
|
|
---------- |
|
|
|
|
|
Wednesday, March 8, 2023 |
|
|
|
House of Representatives |
|
|
|
Committee on Oversight and Accountability |
|
|
|
Subcommittee on Cybersecurity, Information Technology, and Government |
|
Innovation |
|
|
|
Washington, D.C. |
|
|
|
The Subcommittee met, pursuant to notice, at 2:19 p.m., in |
|
room 2154, Rayburn House Office Building, Hon. Nancy Mace |
|
[Chairwoman of the Subcommittee] presiding. |
|
Present: Representatives Mace, Timmons, Burchett, Greene, |
|
Luna, Edwards, Langworthy, Burlison, Connolly, Lynch, Khanna, |
|
Mfume, and Gomez. |
|
Ms. Mace. All right. Good afternoon, everyone. The |
|
Subcommittee on Cybersecurity, Information Technology, and |
|
Government Innovation will come to order. |
|
Welcome and good afternoon to everyone who is here on both |
|
sides of the aisle. Without objection, the Chair may declare a |
|
recess at any time. I recognize myself for the purpose of |
|
making an opening statement, if I may. |
|
Thank you all for being here today, the time and the effort |
|
and commitment to this congressional hearing on our artificial |
|
intelligence. As Chair of this committee, I recognize myself |
|
for five minutes to provide an opening statement on this very |
|
important topic which many of us here today are extremely |
|
passionate about. |
|
The field of artificial intelligence is rapidly evolving, |
|
and one of the most exciting developments in recent years has |
|
been the emergence of generative models. These models have |
|
shown the ability to produce human-like language and even |
|
generate images, videos, and music. While the potential |
|
applications of generative models are vast and impressive, |
|
there are also serious concerns about the ethical implications |
|
of their use. |
|
As we explore the potential of AI and generative models, it |
|
is essential that we consider the impact they may have on |
|
society. We must work together to ensure that AI is developed |
|
and used in a way that is ethical, transparent, and beneficial |
|
to all of society. This will require collaboration between |
|
government, industry, and academia to ensure that the AI we |
|
develop is reliable, trustworthy, and aligned with public |
|
policy goals. |
|
Moreover, we must consider the operational legal |
|
responsibilities of companies that use these models. AI can |
|
help us make better decisions, but we must also ensure that |
|
those decisions are ethical, unbiased, and transparent. To |
|
achieve this, we need to establish guidelines for AI |
|
development and use. We need to establish a clear legal |
|
framework to hold companies accountable for the consequences of |
|
their AI systems. |
|
The Federal Government has an important role to play in the |
|
development and deployment of AI. As the largest employer in |
|
the United States, the government can use AI to improve |
|
operations and provide better services to the public. AI can |
|
help reduce costs, improve efficiency, and enhance the accuracy |
|
of decision-making, for example. AI can be used to analyze vast |
|
amounts of data to identify patterns and make predictions which |
|
can help government agencies make more informed decisions. |
|
As we move forward, we must also ensure that AI is used for |
|
the benefit of society as a whole. While AI has the potential |
|
to improve efficiency, increase productivity, and enhance the |
|
quality of life, it can also be used to automate jobs, invade |
|
privacy, and perpetuate inequality. We must also work together |
|
to ensure that AI is used in a way that benefits everyone, not |
|
just a privileged few. |
|
In conclusion, the emergence of generative models |
|
represents a significant step forward in the development of |
|
artificial intelligence. However, with the progress comes |
|
responsibility. We must ensure that AI is developed and used in |
|
a way that is ethical, transparent, and beneficial to society, |
|
and the Federal Government has an important role in this |
|
effort. |
|
I look forward to working with my colleagues on both sides |
|
of the aisle on this committee to ensure that the U.S. remains |
|
a leader in the development of AI technologies. Thank you for |
|
your time and attention. |
|
Now before I yield back, I'd like to note that everything I |
|
just said in my opening statement was, you guessed it, written |
|
by ChatGPT in AI. |
|
The advances that have been made just in the last few weeks |
|
and months have been radical, they've been amazing, and show |
|
the technology is rapidly evolving. Every single word up until |
|
this sentence was generated entirely by ChatGPT. And perhaps |
|
for the first time in a committee hearing--I know Jake |
|
Auchincloss said a statement on the floor a couple weeks ago, |
|
but I believe this is the first opening statement of a hearing |
|
generated by ChatGPT or other AI models. |
|
I now yield to the distinguished Ranking Member, Mr. |
|
Connolly, for your opening statement. |
|
Mr. Connolly. Thank you, Madam Chairwoman. And let me first |
|
thank you for reaching out on a bipartisan basis to talk about |
|
this Subcommittee and our agenda. I really appreciate that, and |
|
I wish more committees and subcommittees operated that way. And |
|
I think we had fruitful conversation. We actually had a meeting |
|
with certain cyber officials of the executive branch while we |
|
were in Munich at the Security Conference. And, again, I just |
|
appreciate your approach, and hope we can collaborate and make |
|
music together over the next two years. |
|
The Cybersecurity, Information Technology, and Government |
|
Innovation Subcommittee has dedicated its first hearing to |
|
examining advances in artificial intelligence and its |
|
revolutionary impact on society. This decision reflects our |
|
membership's interest in commitment of exploring, |
|
understanding, and implementing emergent technologies. |
|
Last Congress, Chairwoman Nancy Mace, Representative Ro |
|
Khanna, and I introduced the Quantum Computing and |
|
Cybersecurity Preparedness Act, which encourages Federal |
|
agencies to adopt post-quantum cryptography. I'm also pleased |
|
the bill was signed into law just a few months ago. I look |
|
forward to future bipartisan collaboration as we define the |
|
problem sets associated with AI design solutions and that |
|
promote innovation while simultaneously mitigating the dangers |
|
and risks inherent in AI technology. |
|
The Federal Government has a historic, necessary, and |
|
appropriate role guiding and investing research development for |
|
new and emerging technologies. The Defense Advanced Research |
|
Projects Agency, DARPA, the well-known research and development |
|
agency of the United States Department of Defense, is |
|
responsible for the development of myriad emerging |
|
technologies. |
|
One of the most famous successes includes the ARPANET, |
|
which eventually evolved into the internet which we know today. |
|
Other innovations include microelectronics, global positioning |
|
systems, infrared--inferred night imaging, unmanned vehicles, |
|
and what eventually became cloud technology. |
|
AI will require similar Federal investment and engagement. |
|
As stated in the January 2023 final report from the National |
|
Artificial Intelligence Research Task Force, the recent CHIPS |
|
and Science Act reinforces the importance of democratizing |
|
access to a national AI research cyber infrastructure. U.S. |
|
talent and frontier science and engineering, including AI, in |
|
the report calls for 2.6 billion over the next six years for |
|
the purpose of funding national AI research infrastructure. |
|
While government certainly plays a role in R&D, a very |
|
important role, it also has a regulatory role. Congress has the |
|
responsibility to posture careful and thoughtful discussions to |
|
balance the benefits of innovation with the potential risks of |
|
emerging technology. |
|
A recent National Bureau of Economic Research report found |
|
that AI could save the United States healthcare industry more |
|
than $360 billion a year and be used as a powerful tool to |
|
detect health risks. A GAO report predicts AI could help |
|
identify and patch vulnerabilities and defend against cyber |
|
attacks, automate arduous tasks, and expand jobs within the |
|
industry. |
|
As with all technologies, in the wrong hands, AI could be |
|
used to hack financial data, steal national intelligence, and |
|
create deep fakes, blurring people's abilities to certify |
|
reality, and sow further distress within our democracy. AI can |
|
cause unintentional harms. GAO found that certain groups, such |
|
as workers with no college education, tended to hold jobs |
|
susceptible to automation and eventually unemployment. |
|
Another concern relates to machine learning and data. ML, |
|
machine learning, uses data samples to learn and recognize |
|
patterns, such as scanning hundreds or thousands of pictures of |
|
lungs to better understand pulmonary fibrosis and revolutionize |
|
medical care. But what happens if those lung samples only come |
|
from a homogeneous portion of the population? And that medical |
|
breakthrough is inaccurately applied. When it comes to data, |
|
equity is accuracy and must ensure datasets include as much and |
|
as comprehensive a universe of data as possible. |
|
It is paramount that during this hearing we begin to create |
|
a flexible and robust framework, particularly for government's |
|
use of AI to protect democratic values and preemptively address |
|
social, economic, and moral dilemmas AI might raise. |
|
During the last Congress, this committee voted to pass the |
|
AI Training Act and the AI in Counterterrorism Oversight |
|
Enhancement Act, with bipartisan support. The committee is not |
|
entirely new to the AI space, and we look forward to continuing |
|
efforts to support transformative research. We also look |
|
forward to building on the Biden Administration's efforts such |
|
as the National Artificial Intelligence Resource Task Force. |
|
Just over a month ago, that task force released its report, |
|
providing a roadmap to stand up a national research |
|
infrastructure that would broaden access to the resources |
|
essential to AI. |
|
AI is already integrated within the world around us, and |
|
its growing use throughout society will continue to drive |
|
advancements. America must implement an aggressive, research- |
|
forward Federal AI policy to spur competition with other |
|
countries that have already established nationwide strategies, |
|
and additional supporting policy strategies might also include |
|
promoting open data, policies, or outcome-based strategies when |
|
assessing algorithms. |
|
Finally, and more importantly, our country needs the work |
|
force to properly develop, test, understand, and deploy AI. |
|
This work force of the future will include technologists who |
|
will help govern AI responsibly. |
|
I look forward to hearing from our witnesses today. I look |
|
forward to collaborating with you, Madam Chairwoman, on any |
|
subsequent legislation we might want to develop. |
|
I yield back. |
|
Ms. Mace. Thank you, Congressman Connolly. And I, too, |
|
agree, I hope and I believe we will make music together, |
|
continue to do that. Cybersecurity has been one of the few |
|
places in Congress where we have been able to be bipartisan and |
|
not crazy. And so, I appreciate the ability to work with folks |
|
on both sides of the aisle. |
|
I'm pleased to introduce our four witnesses today for this |
|
Subcommittee's inaugural hearing of the 118th Congress. Our |
|
first witness is Dr. Eric Schmidt, Chair of the Special |
|
Competitive Studies Project. Dr. Schmidt is a former Google |
|
executive, where he held multiple senior-level positions, |
|
working alongside founders Sergey Brin and Larry Page. |
|
Google literally changed the world, and it's a huge honor |
|
to have one of the godfathers of modern day technology here |
|
with us today talking about the advent of AI and what comes |
|
next, because I believe this will be one of the greatest |
|
technological revolutions of our lifetime and around the world. |
|
Dr. Schmidt is an accomplished technologist, entrepreneur, |
|
and philanthropist. Dr. Schmidt founded SCSP in 2021. This is a |
|
bipartisan, nonprofit initiative that works on issues relating |
|
to AI and other emerging technologies. Dr. Schmidt also co- |
|
authored a book in 2021 with Dr. Henry Kissinger and MIT dean, |
|
Dr. Daniel Huttenlocher, titled, ``The Age of AI: And Our Human |
|
Future.'' The book attempts to explain artificial intelligence |
|
while raising thought-provoking questions about the role of AI |
|
in topics such as security and world order. And there is a Wall |
|
Street Journal article that was an excerpt from the book that |
|
folks should pick up and read, ``ChatGPT Heralds an |
|
Intellectual Revolution.'' I'm going to encourage folks in this |
|
space to read it. |
|
Our second witness is Dr. Aleksander Mafdry, director of |
|
the MIT Center for Deployable Machine Learning. Dr. Mafdry is |
|
also a member of the MIT Computer Science and Artificial |
|
Intelligence Laboratory, Cadence Design Systems professor of |
|
computing, and co-lead of the MIT AI Policy Forum. Dr. Mafdry's |
|
research interests span algorithms, continuous optimization, |
|
the science of deep learning, and developing reliable, |
|
trustworthy, and secure machine learning systems. |
|
We look forward to hearing from you about the policy |
|
challenges and moral and ethical questions surrounding AI. |
|
Our third witness is Dr. Scott Crowder, vice president of |
|
Quantum Computing and IBM, and chief technology officer, IBM |
|
Systems, Technical Strategy and Transformation. Dr. Crowder's |
|
responsibilities include leading the commercialization effort |
|
for quantum computers and accelerating innovation within |
|
development through special projects. |
|
The Subcommittee is very interested in learning more about |
|
quantum AI and how quantum computing may some day change the |
|
way AI models can store, process, and even report data. |
|
Our fourth witness is Ms. Merve Hickok, Chair and research |
|
director for the Center for AI and Digital Policy. |
|
We welcome everyone who is here today, and we are so |
|
pleased to have all of you here this afternoon. |
|
Pursuant to committee rule 9(g), the witnesses, if you will |
|
please, stand up and raise your right hands. |
|
Do you solemnly swear or affirm that the testimony you are |
|
about to give is the truth, the whole truth, and nothing but |
|
the truth, so help you God? |
|
Let the record show that the witnesses all answered in the |
|
affirmative. |
|
Thank you, and you may be seated. |
|
We appreciate all of you being here today and look forward |
|
to your testimony. I want to remind the witnesses that we have |
|
read your written statements, and they will appear in full in |
|
the hearing record. Please limit your oral arguments to five |
|
minutes, initially. As a reminder, please press the button on |
|
the microphone in front of you so we can all hear you when you |
|
are speaking. |
|
When you speak--begin to speak, the light in front of you |
|
will turn green. And after four minutes, the light will turn |
|
yellow. And then the light--red light comes on after your five |
|
minutes has expired. And we would ask that you please try to |
|
wrap up your comments at that time so that all the Members who |
|
are here today as part of this Subcommittee will get a chance |
|
to speak and ask you all questions. |
|
I would like to first recognize our first witness, Dr. |
|
Schmidt, to please begin your testimony. |
|
|
|
STATEMENT OF DR. ERIC SCHMIDT, CHAIR |
|
SPECIAL COMPETITIVE STUDIES PROJECT |
|
|
|
Mr. Schmidt. Chairwoman and Ranking Member, thank you so |
|
much, all of you, for spending some time on this incredibly |
|
important issue. |
|
I've been doing this for 50 years, and I have never seen |
|
something happen as fast as this round. It took five days for |
|
ChatGPT to get to a million users, and now we have it being |
|
used here in Congress. And, if you look throughout the country, |
|
throughout America, throughout the world I live in, machine |
|
learning in the broad form has taken it by storm. I'm used to |
|
hype cycles, but this one is real in the sense that enormous |
|
amounts of money are being raised to implement and build these |
|
systems. |
|
The sense to me is that this moment is a clear demarcation: |
|
A before and an after. And in our book, ``Age of AI,'' which |
|
you kindly mentioned, we actually talk about this is actually |
|
more than just an industrial strategy, it is actually a new |
|
epic in human experience. The last epic, of course, was the age |
|
of reason 400 years ago which came from the century of the |
|
printing press and the Reformation and things like that. |
|
The ability to have nonhuman intelligences that we work |
|
with and occasionally have to deal with is a major change in |
|
human history and not one that we will go back to. And you can |
|
imagine, if you speculate 10, 20, 30 years from now, at the |
|
rate at which this innovation is going, what it would be like |
|
to having these nonhuman intelligences in the midst, right? A |
|
topic for another day. |
|
The two most interesting things that have emerged in the |
|
last year have been large language models. Large language |
|
models can be understood as a system that was originally built |
|
to predict the next word, the next sentence, the next |
|
paragraph. But if you make them big enough--and when I say big, |
|
I mean huge--to the cost of a hundred million dollars, 200 |
|
million to build them, they appear to have emergent properties. |
|
They have what is technically known as capability overhang. In |
|
other words, we don't know exactly what they know. Although we |
|
do know they know an awful lot of things that are wrong, but we |
|
also know that they have a lot of insights. |
|
This has spurred enormous industry and a set of competitors |
|
that will be emerging in the next month or two. It's literally |
|
that fast. So, boom, boom, boom. |
|
The other one is the term ``generative AI,'' which for me |
|
is largely focused on the ability to generate new language, new |
|
pictures, new videos, and so forth. It's reasonable to expect |
|
that, in the next few years, a great deal of the content that |
|
we consume will be generated for us. |
|
Now, these are very, very, very powerful technologies. And |
|
the impact on society is going to be profound, and I don't |
|
think any of us understand how broad and how deep it will go. |
|
When I look at some of the issues that you all should face, |
|
I think the most obvious one is, what do you do about how |
|
people interact with the platforms? And I'll offer three |
|
principles. |
|
One is the platforms need to know where the content came |
|
from and they need to be able to tell you--this is to avoid |
|
misinformation, Russian actors, that sort of thing. You need to |
|
know who the users are. Even if you don't tell the end user who |
|
they are, there needs to be some notion of who they are and |
|
where they came from. True anonymity hidden behind a paywall |
|
would allow nation-state attacks. And the third is that these |
|
systems have to publish how their algorithms work, and then |
|
they have to be held to how their algorithms work. Those simple |
|
principles, I think, will help us manage the extreme cases |
|
here. |
|
We all, everyone in this room, wants the U.S. to win in |
|
this. And, again, Ranking Member, you mentioned--Connolly, you |
|
mentioned this issue around the national resource. My colleague |
|
to the left can speak about what it's like to be in a |
|
university where you don't have access to these models. We need |
|
that, and we need the computing capability as it transforms, |
|
not just language, but also every aspect of science and health |
|
and biology and material science. |
|
We want democratic partners, that is other countries. This |
|
is something where the West can do this together, and we can |
|
beat China, who is my primary focus. And, obviously, we need |
|
more AI and software talent in the government. And we wrote a |
|
long report for you all called the NSCAI that goes into that in |
|
great detail. |
|
What I want you to do is imagine the alternative. China |
|
announced a couple of years ago that they are going to be the |
|
dominant force in AI in 2030. Can you imagine the technology |
|
that imbues how we think, how we teach, how we entertain, and |
|
how we interact with each other imbued with Chinese values, not |
|
American values, not the values and rules that we have in our |
|
democracy? It's chilling. |
|
The military consequences are also profound, as are the |
|
biological, which we can talk about if you're interested. But |
|
the most important thing to understand is that we need to win |
|
because we want America to win, and this is our best, great |
|
opportunity to create trillions of dollars of wealth for |
|
American firms and American partners. |
|
Thank you. |
|
Ms. Mace. Thank you, Dr. Schmidt. |
|
I would now like to recognize our second witness, Dr. |
|
Mafdry, for his opening statement. |
|
|
|
STATEMENT OF DR. ALEKSANDER MAfDRY, DIRECTOR, MIT CENTER FOR |
|
DEPLOYABLE MACHINE LEARNING, AND CADENCE DESIGN SYSTEMS |
|
PROFESSOR OF COMPUTING, MASSACHUSETTS INSTITUTE OF TECHNOLOGY |
|
|
|
Mr. Mafdry. Chairwoman Mace, Ranking Member Connolly, |
|
Members of the committee, thank you for inviting me to testify. |
|
Today, I want to make three points: First, AI is no longer |
|
a matter of science fiction, nor is confined to research labs. |
|
The genie is out of the bottle. AI is being deployed, broadly |
|
adopted, as we speak. |
|
The key factor that made recent AI tools so popular is the |
|
accessibility. Tools like ChatGPT can be directed using simple |
|
language commands. We can ask it to draft us a memo or a speech |
|
or summarize a movie in much the same way we would ask any |
|
human. No AI expertise required. |
|
As the barrier to adopting AI gets lower and lower, AI will |
|
spread across our economy and our society. It will assist us in |
|
mental and creative tasks, such as writing, visual design, and |
|
coding. It will bolster and expand our capabilities. It can |
|
even help us integrate our accumulative knowledge; for example, |
|
in healthcare, in science, and engineering. |
|
But along with these opportunities, AI also brings risks. |
|
OK. Its lack of reliability; its propensity for promoting bias |
|
and enhancing social inequities; its undermining of |
|
accountability; its facilitation of deep fakes and manipulated |
|
media; its ability to fuel personalized, online phishing and |
|
harassment at scale. |
|
It's critical we proactively identify these emerging risks |
|
and develop clear and actionable ways to mitigate them. While |
|
doing that, we need to recognize, though, all the positives of |
|
AI and balance them against the negatives. In the end, the |
|
impact of AI is not a foregone conclusion as much as rapid |
|
progress of AI might suggest otherwise. |
|
This brings me to my second point. As we engage with AI |
|
more directly, we expose ourselves to interactions that go |
|
against our intuition. Because AI exploits our cognitive |
|
biases, we are often too likely to accept its results as |
|
gospel. Indeed, as we are able to communicate with AI so |
|
easily, so seamlessly, it's natural for us to think of them as |
|
human, but this is a mistake. These tools aren't human. They're |
|
a simple computation executed at impressive scale. |
|
By creating--by treating them as human, we fool ourselves |
|
into thinking that we understand how AI tools behave. We fool |
|
ourselves into thinking that we can straightforwardly adapt |
|
policies designed for humans to work in AI-driven contexts. |
|
Indeed, our intuition often fails us. Take ChatGPT. Given a |
|
question, it can write a really convincing answer, even if |
|
everything it is writing is factually incorrect. It can trick |
|
us into thinking the answer is critical by using prose that |
|
sounds like human experts. Therefore, an unmitigated reliance |
|
on such tools in our day-to-day lives, or even worse in our |
|
education, can have disastrous consequences. It can erode our |
|
analytical and reasoning capabilities. |
|
The final point I want to make is that we also need to pay |
|
attention to how AI is deployed. That is, if we focus solely on |
|
what I have discussed so far, we'll have a major blindspot. A |
|
key feature of measuring AI systems is that they can be used as |
|
foundation on top of which other systems are being built, |
|
forming what I would call AI supply chain. |
|
The upstream of this chain are organizations that create |
|
the foundation AI tools, like ChatGPT. And here, very few |
|
players will be able to compete, given the highly specialized |
|
skills and enormous capital investments the building of such |
|
systems requires. In contrast, we should expect an almost |
|
Cambrian explosion of startups and new use cases downstream of |
|
the supply chain, all leveraging the capabilities of upstream |
|
AI systems. |
|
This leads to a couple of policy-relevant observations. |
|
First, the limited number of large upstream systems may |
|
introduce new challenges, such as hidden systemic fragilities |
|
or structural biases. Imagine, for instance, if one of these |
|
upstream models goes suddenly offline. What happens downstream? |
|
Second, AI system won't be developed by a single entity. |
|
They will be products of multiple AI systems grouped together |
|
each from a different place. These composite systems will |
|
become even harder to predict, harder to audit, harder to |
|
regulate. For instance, who will be responsible and legally |
|
liable when something goes wrong? |
|
Third, this AI supply chain can redistribute power, control |
|
over where, when, and how AI is used. This factor will be |
|
paramount from a societal standpoint, from a geopolitical |
|
standpoint, from a national security standpoint. |
|
To conclude, let me say, we are at an inflection point in |
|
terms of what future AI will bring. Seizing this opportunity |
|
means discussing the role of AI, what exactly we want it to do |
|
for us, and how to ensure it benefits at all. This will be a |
|
difficult conversation, but we do need to have it and have it |
|
now. |
|
Thank you for the opportunity to speak with the |
|
Subcommittee. I look forward to the questions. |
|
Ms. Mace. Thank you. |
|
And I would like to recognize our third witness, Dr. |
|
Crowder, for your opening statement. |
|
|
|
STATEMENT OF DR. SCOTT CROWDER, VICE PRESIDENT, IBM QUANTUM, |
|
AND CTO, IBM SYSTEMS, TECHNICAL STRATEGY AND TRANSFORMATION, |
|
IBM |
|
|
|
Mr. Crowder. Chairwoman Mace, Ranking Member Connolly, and |
|
distinguished Members of the Subcommittee, thank you for this |
|
opportunity to testify before you today. |
|
Today, I represent IBM Quantum where we have two goals: to |
|
bring usable quantum computing to industry and research and to |
|
make our digital infrastructure quantum safe. We have a network |
|
of over 200 industry and research partners exploring the use of |
|
quantum computing for business and science, and have developed |
|
technology to make the transition to quantum safe cryptography |
|
easier. |
|
There is a common perception that classical computers can |
|
solve any problem if they're just big enough. That is not the |
|
case. There is a whole class of problems that classical |
|
computers are not good at and never really will be. |
|
When I talk to leading U.S. companies about their unsolved |
|
problems that if solved could bring them huge economic |
|
benefit--these types of problems turn up everywhere. Some of |
|
these longstanding problems could be solved with a combination |
|
of quantum computing and artificial intelligence. |
|
Quantum computing is a rapidly advancing and radically |
|
different computing paradigm which could launch a new age of |
|
human discovery. Just seven years ago, the notion of a quantum |
|
developer didn't exist. IBM was the first to put a real quantum |
|
computer on the cloud; at the time it was just five qubits. |
|
Today, IBM has systems over 400 qubits. And if we continue on |
|
this technology roadmap, by the middle of this decade, we'll |
|
have 4,000 qubit systems and will demonstrate the first |
|
practical use of quantum computing. |
|
IBM alone has deployed over 60 systems, and our 500,000 |
|
registered users have published over 2,000 research papers. One |
|
key thread in this research is the application of quantum |
|
computation within artificial intelligence. Many of our |
|
partners have published research results using quantum machine |
|
learning techniques. Examples include financial institutions |
|
exploring quantum algorithms for improved fraud detection; |
|
Boeing exploring optimization of composite materials for better |
|
airplane wings; and CERN exploring applications in high-energy |
|
physics. |
|
One primary reason quantum computing has benefit for |
|
artificial intelligence is because it uses a different method |
|
to find patterns in data. For example, in fraud detection, a |
|
quantum algorithm may be better at detecting true fraud and |
|
reducing false positives. A data scientist may choose to use |
|
either a quantum fraud model or a classical AI fraud model or a |
|
combination for the best results. Put simply, quantum will be |
|
another computational tool to use to improve AI results. |
|
Generally, we see the future of computing as a combination |
|
of classical, specialized AI, and quantum computing resources. |
|
It will not be based solely on classical bits, but rather built |
|
upon bits and neurons and quantum bits, or qubits. This will |
|
enable the next generation of intelligent, mission-critical |
|
systems and accelerate the rate of science-driven discovery. |
|
Researchers, companies, and governments that leverage this |
|
technology will have a distinct competitive advantage. |
|
That leads to a critical point: When one examines the |
|
financial commitment other countries are making in quantum |
|
computing, our belief is the U.S. Government investment in |
|
driving this critical technology is insufficient to stay |
|
competitive. At its inception in 2018, the $1.7 billion |
|
National Quantum Initiative stood as a leading public |
|
investment. Today, the planned global public investment in |
|
quantum technology is estimated to exceed $30 billion, with |
|
China at $15 billion. It is critical that we not only |
|
reauthorize the NQI, but also increase its investment in the |
|
critical area of research of use of quantum computers for |
|
mission-critical applications. |
|
The same importance for ethical and trustworthy AI applies |
|
whether classical compute or quantum compute underpins the |
|
solution. We know that trustworthiness is key to AI adoption, |
|
and the first step in promoting trust is effective risk |
|
management policies and practices. Companies must have strong |
|
internal governance processes, including, among other things, |
|
designating a lead AI ethics official responsible for its |
|
trustworthy AI strategy, and standing up an AI ethics board as |
|
a centralized clearinghouse for resources to help guide that |
|
strategy. IBM has implemented both, and we continue to advocate |
|
others in the industry to do likewise. |
|
Additionally, it's important to establish best practices |
|
for AI bias mitigation, similar to BSA's framework published in |
|
2021. |
|
It's difficult to pinpoint the precise benefits and |
|
possible challenges presented by any new emerging technology. |
|
Quantum computing is no different. However, those countries |
|
that make investments in this transformative technology today |
|
will reap benefits in the years to come. Those countries that |
|
do not will be at a competitive disadvantage in the future. At |
|
the same time, countries will also need to invest time and |
|
energy in developing an appropriate regulatory environment that |
|
supports the adoption of trustworthy AI regardless of the |
|
underlying compute technology. |
|
Thank you again for inviting me to testify, and I look |
|
forward to today's discussion. |
|
Ms. Mace. Thank you. |
|
And I would like to recognize our fourth witness, Ms. |
|
Hickok, for your opening statement. |
|
|
|
STATEMENT OF MS. MERVE HICKOK, CHAIR AND RESEARCH DIRECTOR, |
|
CENTER FOR AI AND DIGITAL POLICY |
|
|
|
Ms. Hickok. Thank you so much. |
|
Good afternoon, Chairwoman Mace and distinguished Members |
|
of the committee. I'm Merve Hickok, Chair and research director |
|
for Center for AI and Digital Policy. It's an honor to be here |
|
today, and thank you for the opportunity to testify. |
|
CAIDP is a global research organization based in D.C. We |
|
educate and train future AI policy leaders, collaborate with AI |
|
policy experts around the world. We also publish AI and |
|
Democratic Values index, analyzing AI policies and practices |
|
across 75 countries. |
|
I also train and build capacity in organizations on |
|
responsible AI development and governance. And prior to CAIDP, |
|
I was in the corporate world as a senior leader at Bank of |
|
America Merrill Lynch, responsible for recruitment technologies |
|
internationally. |
|
I provide this background because we believe in the promise |
|
of AI. However, we also know that AI systems, if not developed |
|
and governed with safeguards, have negative impacts on |
|
individuals and society. We believe that AI should first and |
|
foremost serve members of the society, their rights, their |
|
freedoms; our social, moral, and ethical values. |
|
The title of the hearing today asks if you are ready for a |
|
tech revolution. My brief answer is no. We don't have the |
|
guardrails in place, the laws that we need, the public |
|
education, or the expertise in the government to manage the |
|
consequences of the rapid technological changes. |
|
Internationally, we are losing AI post leadership. |
|
Domestically, Americans say they are more concerned about-- |
|
concerned than excited by AI making important life decisions |
|
about them, knowing their behavior. AI systems now produce |
|
results we cannot assess or replicate. Opaque systems put |
|
governments, companies, and individuals at risk. AI expands our |
|
research and innovation capabilities; however, it also |
|
replicates existing biases in the datasets and biases in the |
|
choices of the developers, resulting in disadvantaging people |
|
with disabilities in hiring, for example; inaccurate health |
|
predictions for patients of color; offering women lower credit, |
|
lower home valuations; innocent people being arrested due to |
|
biased facial recognition. |
|
We are now debating generative systems which produce |
|
synthetic text, image, video, and audio. The systems will |
|
certainly add to our creativity, there is no doubt about it, |
|
but they're already impacting the original creators. They will |
|
also be used by malicious actors to fabricate events, people, |
|
speeches, and news for disinformation, cyber fraud, |
|
blackmailing, and propaganda purposes. |
|
I give this testimony on International Women's Day, when |
|
unregulated opaque AI systems deepen discrimination and online |
|
harassment against women. |
|
Both governments and private companies know that public |
|
trust is a must-have for further innovation, investment, |
|
adoption, and expansion. Companies, large and small, are |
|
calling for regulatory guidance. |
|
Administrations of both parties have called for trust for |
|
the AI. President Trump's Executive Order 13960 explained that |
|
ongoing adoption and acceptance of AI will depend significantly |
|
on public trust, and AI should be worthy of people's trust, and |
|
that this order signals to the world U.S. commitment to develop |
|
and use AI underpins by democratic values. The order |
|
characterized trustworthy AI as being lawful, respective of |
|
civil rights, accurate, reliable, safe, understandable, |
|
responsible, transparent, accountable, and regularly monitored. |
|
Office of Science and Technology has recently published |
|
AI--Blueprint for an AI Bill of Rights, a critical policy |
|
framework underscoring similar qualities for AI, emphasizing |
|
democratic values and civil rights. |
|
President Biden has called for bipartisan legislation to |
|
keep companies accountable, and reiterated the same principles: |
|
transparency, accountability, and safeguarding our values. |
|
We very much support this committee and its bipartisan |
|
nature, but there are real challenges ahead, and I will |
|
conclude with a few recommendations toward those. |
|
We really need the Congress to hold more hearings like |
|
this, explore the challenges, the risks and benefits, and hear |
|
from the public and those impacted. We need the Office of |
|
Management and Budget to move forward with the long-delayed |
|
rulemaking for the use of AI in Federal agencies as part of the |
|
executive order. We need to build multidisciplinary capacity in |
|
Federal Government to ensure the work force understands the |
|
benefits and risks of AI. We need the wider work force to |
|
understand benefits and risks of AI as well. We need R&D |
|
capabilities expanded beyond a handful of companies, campuses, |
|
and labs, and demand trustworthy AI with our research agenda. I |
|
urge you to act now to enact the legislation reflecting the |
|
bipartisan nature. |
|
Absent a legislative agenda or implementation of AI policy, |
|
American people, American companies, and allies are lost about |
|
U.S. AI policy objectives. |
|
Thank you. |
|
Ms. Mace. And thank you. |
|
And I think one of the things that sticks out to me today |
|
is, actually, this is the first AI hearing this Congress in the |
|
U.S. House of Representatives. But also, this same day, the |
|
U.S. Senate had their first hearing on AI on this subject |
|
matter; they beat us by four hours this morning at 10 a.m. |
|
But I would now like to recognize myself for five minutes |
|
for a few questions of our panelists. |
|
Thank you, Dr. Schmidt, for painting what I would describe |
|
as a very vivid picture of what is happening in this space, |
|
because I agree with you, it's been rapid and, in your words, |
|
epic. And I'm not sure that the world is ready for what is to |
|
come in the next few months, years, et cetera. And so, it |
|
reminds me of Einstein. He said: ``I never think of the future. |
|
It comes soon enough.'' And it is here. And it is moving faster |
|
than the speed of light. |
|
So, my first question today is for Dr. Schmidt. How can we |
|
ensure that AI technology is developed in a way that is safe, |
|
transparent, and beneficial for society without stifling |
|
innovation? |
|
Mr. Schmidt. I'm always worried about AI conversations, |
|
because everyone believes the AI that we are building is what |
|
they see in the Terminator movies. And we are precisely not |
|
working on those things. So we are clear, we are not doing---- |
|
Ms. Mace. Not yet. |
|
Mr. Schmidt. We are not doing it yet, and we are not likely |
|
to. But what we are doing is working on systems that will |
|
affect the way people perceive their world. And I think the |
|
best thing for America to do is to follow American values, |
|
which include robust competition with government funding of |
|
basic research, and using the innovators, including the folks |
|
to my left, to actually deliver on this. |
|
I think that one of the things that is not appreciated in |
|
the last 30 or 40 years of tech companies is--speaking as a |
|
person who is associated with a number of them--is how good |
|
they are as American exports of our values. So, I come back to |
|
a much simpler formulation that American ingenuity, American |
|
scientists, the American government, and American corporations |
|
invent this future and will get something pretty close to what |
|
we want. And then you guys can work on the edges where you have |
|
misuse. |
|
The alternative is, think about if it comes from somewhere |
|
else which doesn't have our values. And I really believe that. |
|
Everything that you can do to finish, to support that |
|
innovation cycle, the universities, the graduate students, |
|
getting foreigners who are high-skilled in to help us, building |
|
those corporations, creating shareholder wealth, hiring lots of |
|
people--it's the American formula, and it will work here too. |
|
Ms. Mace. And then, on that note, in terms of the |
|
personnel, the resources, training folks in the technology so |
|
that it can advance, having that innovation. And lot of it is |
|
on the software side, but how does hardware figure into that? |
|
CHIPS, for example. |
|
Mr. Schmidt. So, in our AI commission that you all |
|
commissioned a while ago, we spent a lot of time on this. We |
|
felt it was very important for America to retain its |
|
leadership, which of course we didn't have, we gave it to |
|
Taiwan. The best result was to basically get the Taiwanese |
|
firms, primarily TSMC, and Korean firms, primarily Samsung, to |
|
locate plants in the United States, which has occurred. |
|
The Trump and Biden administrations have done a good job of |
|
restricting some of the sales and access to these tools to the |
|
Chinese. But, fundamentally, this is a race, it's a |
|
competition, and we're not that far ahead. So, we have to keep |
|
innovating, which is why your support for the CHIPS Act was so |
|
helpful. And so, on behalf of the whole industry I'll thank all |
|
of you for doing that. That's a good example of the government |
|
getting ahead of a real problem. |
|
Ms. Mace. Thank you. |
|
And, Dr. Mafdry, my next question for you, do we need to be |
|
worried about too much advancement too fast in AI? Are we |
|
capable of developing AI that could pose a danger to humanity's |
|
existence all over the world, some of the things that people |
|
talk about out of fear in this conversation because of a lack |
|
of knowledge, or is that just science fiction? |
|
Mr. Mafdry. Well, it really depends on what do we view as |
|
this kind of catastrophic risk. So, a Terminator-style |
|
scenario, I'm not too worried about this, as Dr. Schmidt just |
|
said. What I am worried is about something more mundane but |
|
essentially very, very corrosive, right. So, we see how this |
|
works out in social media where, essentially, AI also runs in |
|
social media. That's what decides what we see. And we are |
|
seeing the effect of that. Well, this is kind of not really |
|
transparent, not really aligned with the societal goal. |
|
So, that is--now think about things like this new |
|
generative models developed essentially in a way when we just |
|
maximize the profit, we just try to get maximum adoption. I'm |
|
worried about that. |
|
Having said that, I do think we can figure it out how to |
|
not stifle innovation, just moderate it so we still can |
|
progress. But just, again, ensure that the companies that we |
|
talk to, they are not only--only driven by profit, but realize |
|
they have some responsibilities, and they need to acknowledge |
|
them. |
|
Ms. Mace. And I would agree. And I think, you know, we've |
|
talked about algorithms for years, like on social media, and |
|
the use of divisiveness of politics today, and each side |
|
getting the extreme of their side and getting fed more of that |
|
information. I sort of feel like it would be--we were putting |
|
it on steroids of the future, immediate future of what the |
|
advances in AI might be. What are your concerns there? |
|
Mr. Mafdry. Yes. So, essentially, I think saying that this |
|
might be like social media on steroids is very much--is very |
|
much justified. So, again, now I told you that ChatGPT will be |
|
so much more pervasive than social media. And, essentially, we |
|
don't exactly know what will be the effects on our thinking |
|
here or like the way our children learn to think. Like, do they |
|
just fully trust what ChatGPT tells them or do they learn how |
|
to reason? |
|
So, again, I'm really worried about this, but I think--and |
|
that is where the government really needs to step up. We can-- |
|
you know, with enough involvement with government, which I |
|
think might not be too much in context of social media, but |
|
here we have to do it differently, and I think that we will do |
|
it well. |
|
Ms. Mace. Thank you. |
|
I would now like to recognize my esteemed colleague, Mr. |
|
Connolly, for questions. |
|
Mr. Connolly. Thank you so much, Madam Chairwoman. |
|
Listening to Ms. Hickok and the potential of AI is actually |
|
really positive in terms of how it can complement the quality |
|
of life for humans and make things better and promote peace and |
|
harmony. But we know that technology can be used for good and |
|
evil. |
|
And I'm listening to what you just said, Dr. Mafdry, in |
|
terms of your hope for the government's role. And yet, if you |
|
look at social media and you look at technology in general, |
|
Congress has been very reluctant to get into the game of |
|
regulation. And as a result, awesome power has been developed |
|
by and deployed by entrepreneurs who became billionaires in, |
|
largely, Silicone Valley without any interference by the |
|
government. They make all kinds of massive decisions in terms |
|
of content, in terms of what will or won't be allowed, in terms |
|
of who gets to use it, et cetera. |
|
And so, why should we believe that AI would be much |
|
different in terms of its future in the hands of the Federal |
|
Government? |
|
Mr. Mafdry. Well, again, the hope here is that we will-- |
|
don't play the same playbook we played for social media. And in |
|
particular, I think the point of start here is before we go-- |
|
so, first of all, I want to say that I strongly believe that |
|
regulation is a very important tool to make sure that, you |
|
know, just certain technologies are aligned with like broad |
|
societal benefits, and they need to be used. |
|
Having said that, before we go to premature regulation and |
|
we kind of rush regulation, first of all, even the rush |
|
regulation might not be fast enough for AI because AI is a very |
|
fast moving target. But even we forget, I think what we need to |
|
start, we need to start ask questions. And, in particular, |
|
government needs to ask questions of this company saying, what |
|
are you doing? Why are you doing this? What are the objectives |
|
of the algorithms that you are developing? Why is there no |
|
objectives? How will we know that you are accomplishing these |
|
objectives? How can we have some input into what these |
|
objectives showed. |
|
I think this change of tone, together with the government |
|
recognizing that you cannot abdicate AI to the big tech, as |
|
capable as they are, that they have different use cases. They |
|
have different priorities. Like, that's what needs to change. |
|
If this doesn't change, I'm extremely worried. |
|
Mr. Connolly. Well, I just think, if we look at the past |
|
and we look at social media, I wouldn't bet the farm on any |
|
kind of rapid regulatory regime coming from the Federal |
|
Government. |
|
Mr. Mafdry. And just to clarify, that's what I--that's |
|
exactly what I'm worried about. So, let's have conversations we |
|
can have now. |
|
Mr. Connolly. Right. |
|
Mr. Mafdry. Hopefully, we'll learn from the mistakes. |
|
Mr. Connolly. Thank you. |
|
Dr. Schmidt, you want to see the United States get ahead in |
|
this lane of technology and to compete successfully against the |
|
Chinese. Can you talk a little bit about what is the nature of |
|
that threat? How well are they doing in this sphere, and what |
|
do we need to be concerned about? |
|
Mr. Schmidt. There are four or five companies, Mr. Leader, |
|
that are American or British firms that have extremely large |
|
language models. There's also at least one large one in Baidu |
|
in China. I was interested to note that the largest |
|
noncorporate such example in the world that is not owned by a |
|
corporation is also in Tsinghua University in Beijing. |
|
So, there's every reason to believe that the Chinese |
|
understand everything that we're talking about now extremely |
|
well. They've published their intent, so we can read about it. |
|
And I view it as a national emergency. This technology is so |
|
powerful in terms of its ability to transform science, |
|
innovation, biology, semiconductors, you name it--and along |
|
with quantum, I should add--that we need to get our act |
|
together to win and to win a competition. |
|
If we don't--let me give you some examples. AI can be used |
|
to generate good things in biology, but also lots of bad |
|
viruses. You all have created a Bioterror Commission, which I'm |
|
fortunate to serve on, to take a look at this and the impact of |
|
that. That's another example of national security. |
|
The issues of misinformation of the nation-state could be |
|
very significant. Think about the progress of war and conflict |
|
where decision-making can be done faster than the OODA loop or |
|
faster than human decision-making. These are all challenges, |
|
and our government is behind where it needs to be in the |
|
adoption of these technologies for national security as well. |
|
Mr. Connolly. I just would end by saying, I couldn't agree |
|
with you more. And I think really we need to be looking at sort |
|
of like, you know, the race to the moon kind of shot in, you |
|
know, quantum computing, AI, cyber, 5G. Because if the Chinese |
|
dominate those areas, the future is theirs. |
|
I yield back. Thank you, Madam Chair. |
|
Ms. Mace. Thank you, Mr. Connolly. |
|
I would now like to recognize a fellow South Carolinian, |
|
Congressman Timmons. |
|
Mr. Timmons. Thank you, Madam Chair. That's great to say: |
|
Madam Chairwoman. Congratulations on being the Chair. |
|
Ms. Mace. Thank you. |
|
Mr. Timmons. First up, thank you so much for your |
|
attendance here today. You all are experts in your field, and |
|
we really appreciate you taking the time to come and share your |
|
thoughts on this important topic. |
|
Congress is grappling with technology. Our country's |
|
grappling with technology. And we're doing our best to try to |
|
figure out a regulatory environment that fosters innovation and |
|
allows economic growth, while managing the potential adverse |
|
impacts that technology can have. |
|
Obviously, we are working on cryptocurrency and digital |
|
assets, and that's a major challenge for us. Congress is not |
|
the youngest, most tech savvy part of our society, and we are |
|
doing our best. |
|
But I do want to talk about AI's potential impact on our |
|
work force, particularly how tech can be leveraged to further |
|
individual efficiency rather than possibly displace workers. |
|
So, Mr. Crowder, I want to start with you. What are the |
|
most promising use cases for AI as a tool in the work force, |
|
and how do you anticipate AI will be--will influence industries |
|
such as the financial services sector? |
|
Mr. Crowder. Yes, I think it's going to be pretty broad. |
|
And one of the exciting things that we didn't really talk about |
|
is that leveraging some of the underlying technologies like |
|
base or foundation models can be applied to things, not just |
|
writing a haiku or coming up with a speech, but also, you know, |
|
looking at language-like things in other fields, like tabular |
|
data and finance, et cetera, et cetera. |
|
So, in addition to the kind of things I talked about fraud |
|
detection, I think we've all experienced, you know, maybe |
|
traveling abroad and having your credit card be declined, and |
|
that's bad for banks because they want that credit card money. |
|
So, even a small percentage improvement in false positives is a |
|
lot of money for our financial institutions. So, there's lots |
|
and lots of applications. |
|
But to your point, I mean, I think we need to look at AI as |
|
augmenting what humans can do as opposed to replace. And I |
|
think good utilization of AI is to make it--make our work force |
|
more efficient. And I would argue one of the things that we do |
|
a good job in the United States is funding basic science. But |
|
we also need to look at how do we encourage our work force to |
|
be able to use the technology as opposed to just develop the |
|
technology. Because I think the use of AI is going to be a |
|
differentiating factor on, you know, making the U.S. Government |
|
as well as our companies more effective and more competitive. |
|
Mr. Timmons. As businesses try to compete in the free |
|
market, they're inevitably going to try to cut costs and |
|
replace work force with technology. What--how are we going to |
|
manage that challenge? |
|
Mr. Crowder. That's a good question. I don't know if I have |
|
a perfect answer for it. But I think having a more productive |
|
work force that is focusing on value creation, I think at the |
|
end of the day is what really drives success in business. And |
|
the more that you can automate tasks that aren't really value |
|
creation so you can free up your workforce to create value, I |
|
think that is good. And I think that is a more positive way of |
|
driving additional productivity as opposed to thinking about it |
|
as removal of cost. |
|
Mr. Timmons. Sure, sure. Dr. Schmidt, what jobs do you |
|
think will be created in the wake of AI and what jobs do you |
|
think will be threatened? |
|
Mr. Schmidt. I think one of the general comments to make is |
|
I've spent 20 years listening to the theme that jobs will be |
|
displaced or lost because of technology. And today we have a |
|
huge shortage of workers to fill the jobs in America. The |
|
biggest category is truck drivers. Remember how truck drivers |
|
are going to be replaced by automation. |
|
So, it looks to me like the most likely scenario in the |
|
next 20 years or so is not enough people to fill these jobs. |
|
And the best way to get people who can fill those jobs, to have |
|
them have better tools, better education, better knowledge, and |
|
a partner, if you will--all of the evidence that I've studied |
|
indicates that having a digital partner increases your wage, |
|
right? Literally, when you are using a computer to help do the |
|
job, the job has a higher salary. |
|
So, it looks to me like as we get more diffusion of this |
|
technology, on average, jobs go up. There are jobs that are |
|
lost, there are jobs that are created. |
|
Mr. Timmons. Sure. I couldn't imagine my life without |
|
Google, Apple, and Amazon. I feel like I'm attached to my |
|
phone. And I haven't been to the grocery store in three years, |
|
and it is great. So, I'm sure that this is going to create |
|
additional opportunities to make my life more efficient and |
|
make me more capable of having a greater impact. So, I |
|
appreciate that. |
|
And thank you, Madam Chair. I yield back. |
|
Ms. Mace. Thank you. |
|
I now recognize Congressman Lynch for five minutes. |
|
Mr. Lynch. Thank you, Madam Chair. And congratulations to |
|
you and to the Ranking Member. |
|
Dr. Schmidt, in March 2021, the National Security |
|
Commission on Artificial Intelligence released its |
|
comprehensive, I think it was like 800 pages. It actually |
|
defended itself against the risk of being read by its sheer |
|
thickness. But right after your report came out, I was the |
|
Chair of the National Security Subcommittee, and we invited you |
|
to testify regarding that. I see your staff all nodding. They |
|
have painful memories of this, I'm sure. But we invited you in, |
|
and we went over the report. It had 16 major recommendations, |
|
and then probably another 20 other subsidiary ancillary |
|
recommendations. |
|
I'm wondering if you could talk about the progress, or the |
|
lack of progress, we've made over these two years now since you |
|
last testified before this committee about this issue. |
|
You had some very pointed recommendations, you know, for |
|
DARPA. You had recommendations and action items for Congress, |
|
for the executive, for this interagency task force that you |
|
envisioned. |
|
Can you talk a little bit about where you think--how much |
|
progress do you think we have made? And, you know, would you |
|
give us--what kind of grade would you give all of us. |
|
Mr. Schmidt. Well, in the first place, you guys give us the |
|
grade, and we are happy to serve. I would say about half of our |
|
recommendations have been adopted through the NDAA and other |
|
processes. We were kind enough to write the legislation for you |
|
as a hint, and you all were able to adopt it fairly quickly, |
|
and it worked. |
|
The area that I'm most focused on right now is basically |
|
the training problem. And I just don't see the progress in the |
|
government to reform the way it hires and promotes technical |
|
people. As part of the--part of the AI report, we proposed a |
|
civilian, essentially, trading academy for digital skills. And |
|
there are various different forms of this. But I don't think |
|
the government is going to get what it needs unless it has a |
|
program which is modeled on the military academies but for |
|
civilians, where civilians get trained in technology in return |
|
for serving in the government in the civilian capacity to help |
|
the government out. |
|
This is a young person's game. These technologies are too |
|
new. You need new students, new ideas, new invention. I think |
|
that is going to be the fastest way to get it. I don't think |
|
the government is going get there without it. That'd be my |
|
highest--I think that's the largest omission. |
|
Mr. Lynch. Is there a way--so I actually was confronted |
|
with this same problem in my district, where a lot of high-tech |
|
firms, biotech firms moving into the district, and I grew up in |
|
the local public housing projects and those kids--our kids |
|
weren't getting those jobs. So, I started a--I founded a |
|
charter school that focuses on STEM, you know, math, science, |
|
technology. And it's doing really, really well. But it's one |
|
school, you know, out of a hundred. |
|
And is there a way to--I'm not so sure if it is efficacious |
|
to try to take somebody who is coming out of high school or in |
|
college and then make them a tech person. I think it's a much |
|
longer runway and better chances of success if we start at a |
|
very early age. I mean, is there any thoughts about, you know-- |
|
I mean, you know, we are having problems with our public |
|
education system anyway. But is there a way to amp that up at |
|
an early age in early grades to produce the type of workers |
|
that you envision will be necessary to maintain our edge, not |
|
only in artificial intelligence, but everything else we have |
|
got to do. |
|
Mr. Schmidt. Israel does something interesting in this |
|
area. If you are 15 or 16 and a math prodigy, they actually put |
|
you in a special school that is combined with their mandatory |
|
military training. I'm not suggesting a mandatory military |
|
training for our math people. God knows how they would do. But |
|
the important thing is identifying the talent early and then |
|
getting it into the right track. And, again, the educators to |
|
my left can talk about this at more length. |
|
But I think that at a Federal level, the easiest thing to |
|
do is to come up with some program that is administered by the |
|
states or by leading universities. Every state has a big land |
|
grant university that is interested in this. And getting them |
|
money so that they can build these programs, and then they get |
|
paid back for that with service. I like those models, and that |
|
is a model that takes your idea and scales it. There is a lack |
|
of money to build these systems at scale, and that idea or some |
|
variant of it would do it. |
|
Aleksander? |
|
Mr. Lynch. Thank you. |
|
Yes. Mr. Mafdry. |
|
Mr. Mafdry. If I can just add, is that AI's technology is |
|
to learn best by applying it to used cases. And government--so |
|
I actually was discussing exactly this problem with DOD because |
|
they have exactly suffered from this--and instead of thinking |
|
of this is a weakness, this could be a strength. Once it is |
|
tried to apply AI internally to your problems, that's where |
|
people will learn. And this way actually people come, let's say |
|
to DOD, or to government programs for three, five years, and |
|
they come back to the civilian sector. And really like, also, |
|
well, we are lacking this talent also in our civilian economy |
|
too. So, I think that is the way to go, and the government |
|
could play a big role here. |
|
Mr. Lynch. Thank you. |
|
Madam Chair, I know I have another one witness need to |
|
answer the question, but I think I've run out of time. |
|
Mr. Burchett. Chairlady, why don't you let him go. I'm |
|
next, but go ahead. |
|
Mr. Lynch. OK. Ms. Hickok? |
|
Ms. Hickok. I just wanted to follow up with the last remark |
|
as well in terms of education. I echo the task force |
|
nationally. I researched task force reports and recommendations |
|
on democratizing the research and development capabilities |
|
within the Federal Government for the government as well. |
|
Sometimes our brightest minds are forced to go to a handful |
|
of companies and labs and campuses to do their research in |
|
areas that they are interested. But if you have this capacity |
|
within the government's infrastructure as well, that would also |
|
be another way to attract this work force. |
|
And I will expand also the education piece from the schools |
|
to consumers, and expand the education need from technology |
|
jobs to all the jobs. We need lawyers who understand these |
|
concepts. We need sociologists, anthropologists, ethicists, |
|
policymakers. We need understanding and capacity building in |
|
this topic across the whole domain and industry. |
|
Mr. Lynch. Thank you. |
|
Madam Chair, I yield back. And I thank you for the |
|
courtesy, and I thank the gentleman as well. Thank you. |
|
Ms. Mace. Thank you. |
|
I would now like to yield to Congressman Burchett from |
|
Tennessee. |
|
Mr. Burchett. I'd tell my friend across the aisle, you'll |
|
get nowhere calling me a gentleman; I just want you to know |
|
that. And I didn't miss my thought process when I came in here, |
|
and we have a lady who is chairman and it is international day |
|
of the woman. And I think that is pretty cool that you Chair. |
|
If my momma were alive, she would think that is very cool, too, |
|
because she was a pretty cool lady. |
|
Thank you all for being here. I'm probably the least |
|
qualified person of ever asking y'all questions, but as the |
|
435th most powerful Member of Congress, I feel very empowered |
|
today, and I'm kind of digging this subject matter. And I'll |
|
try to go through these quick. |
|
Mr. Schmidt, I did Google, brother. And I don't know what |
|
it is. I hit that button--you know, my mom and daddy would |
|
always say look it up. Now I tell my daughter, Google it, |
|
honey. You know, so I think it is pretty cool. |
|
But the development of AI, how will that impact our |
|
international relations specifically with China? I fear what |
|
they will do if they get control of it, as you have stated |
|
there. I think you mentioned the date that they said they were |
|
going to control it, and I would say they probably be doing |
|
that five years ahead of time. |
|
Go ahead, brother. |
|
Mr. Schmidt. Thank you, Congressman. I worry about the |
|
following scenario: In the future, there's a war. It's an |
|
attack by North Korea on the U.S. Sorry. China stops the war |
|
between North Korea and the U.S., and the entire war took one |
|
millisecond. |
|
And the reason I worry about that is I can't figure out how |
|
we are going to build offensive and defensive systems that are |
|
reliable enough to put them in charge of a war that occurs |
|
faster than human decision-making. That, to me, is the ultimate |
|
threat of this technology, that the things occur faster than |
|
humans can judge them. I don't have a good solution for that. |
|
My second observation is that China is very smart, and they |
|
have identified these areas as these underlying technologies to |
|
provide leadership that dominates industries. A good example is |
|
synthetic biology, which is an area which was imbedded in the |
|
United States, likely to be, again, trillions of dollars of |
|
wealth. China has now maximized its investment in this area. |
|
Not only is it good for their national security, but it's good |
|
for their businesses. |
|
So, when you have got a nation-state that's smart, |
|
technocratic, focused on its own defense and innovation, and |
|
proposing its own companies in the form of civil military |
|
fusion, we have got a serious competitor. That's why we have to |
|
act. |
|
Mr. Burchett. Are you aware--or maybe you are not. I'm |
|
not--of the Chinese infiltrating any of our AI companies? |
|
Mr. Schmidt. I am not. You must assume the Chinese have |
|
operatives pretty much everywhere, based on their history. |
|
Mr. Burchett. OK. Is there any way that we could |
|
proactively protect against AI-generated cyber attacks? |
|
Mr. Schmidt. Well, you defend against them. |
|
Mr. Burchett. Right. |
|
Mr. Schmidt. Technology--we looked a lot at this. The |
|
question is could you create the equivalent of a Manhattan |
|
Project that was secret and you'd keep it all in one place, in |
|
one location, New Mexico, what have you. The knowledge is |
|
moving too quickly. There's too many people globally. We are |
|
going to have to win by staying ahead, which means building |
|
powerful defensive systems. |
|
Mr. Burchett. OK. Thank you. |
|
Dr. Mafdry, what are some of the personal risks to personal |
|
privacy that are associated with the use of AI? |
|
Mr. Mafdry. Sir, could you clarify what kind of risk? |
|
Mr. Burchett. Well, I guess I should ask my research |
|
person. As I stated, it is a little out of my league. |
|
Mr. Mafdry. No, I just didn't hear. I just didn't hear. |
|
Mr. Burchett. No, I just said what are some of the |
|
potential risks? |
|
Mr. Mafdry. I see. So, again, there is many, and they |
|
really depend which sector you look at because there are |
|
different levels of credibility. So, one of them and like the |
|
one big risk, and that's something I research myself so I'm |
|
intimately familiar with, this technology is not fully |
|
reliable. It works most of the time, but not always. |
|
Mr. Burchett. It's not fully what? I didn't---- |
|
Mr. Mafdry. It's not fully reliable. So essentially, it |
|
works most of the time but not always. And then what is worse, |
|
you might not realize when it's not working. OK. So, for |
|
instance, in ChatGPT they hallucinate things sometimes, and you |
|
might not realize they are hallucinating things because it |
|
looks very convincing. |
|
The other aspect of this is, essentially, as the systems |
|
ingest our data, they can really essentially know us better |
|
than we do. And again, that was true also of Google, also of |
|
the social media, but I think with this next generation of |
|
models, this will become even more so. |
|
And then the third level of risk is exactly the one that |
|
Dr. Schmidt talked about. I'm actually really worried about |
|
that. Not even about the actual war happening, but us preparing |
|
for the war like something can go wrong. And it becomes like |
|
when things are happening within a millisecond, like, we have |
|
no good intuition or no good ways to actually figure out how to |
|
make it safe. |
|
Mr. Burchett. OK. Thank you. |
|
Running out of time but, Dr. Crowder, real quickly, how |
|
will the quantum computing impact the security of encrypted |
|
data? |
|
Mr. Crowder. In the long-term, quantum computers, someone |
|
proved on a blackboard, that a lot of our current cryptography, |
|
how we basically send keys around and how we digitally sign |
|
things, eventually will get broken by a quantum computer. The |
|
good news is that people have come up with algorithms that |
|
quantum computers are not good at solving and classical |
|
computers are not good at solving. |
|
So, our challenge is really transitioning from the |
|
cryptography we use today to that new form of cryptography. And |
|
we want to do that as quickly as possible once we have got |
|
really safe standards because we're worried that people will |
|
take all the data today and decrypt it later. So, for some |
|
applications, that doesn't matter, but for a lot of |
|
applications that makes a big deal. |
|
Mr. Burchett. OK. Thank you. I've run out of time. |
|
Chairlady, thank you, ma'am, very much. |
|
Ms. Mace. You did a great job. |
|
I would now like to recognize Congressman Mfume. |
|
Mr. Mfume. Thank you very much, Madam Chair. Again, my real |
|
thanks to you and the Ranking Member for having discussions |
|
that led us to this point. |
|
This, for lack of a better term, has scared the hell out of |
|
me. And I thought I knew something about AI. I'm bopping around |
|
on the campus talking to students in a classroom now and then |
|
teaching them, but what I have heard today is unlike anything I |
|
have ever heard, particularly in terms of our national |
|
security. |
|
I think the Chairwoman mentioned or quoted Einstein a few |
|
minutes ago. He also said that great minds have always |
|
encountered violent opposition from mediocre spirits. And I |
|
don't know if you are encountering violent opposition, but I |
|
think you are encountering a great deal of inattentive or |
|
unattentive population groups who just, for whatever reason, |
|
are not paying attention to what is going on. It is very scary, |
|
and I would strongly support, Madam Chair, any future hearings |
|
on this. I just don't that think we have much of a choice. It |
|
is that imperative. |
|
Dr. Schmidt, you said it was a national imperative, almost |
|
a national emergency. That got my attention, and it will keep |
|
my attention. |
|
I don't know that we can do enough to ring the bell on this |
|
so that our institutions, whether it's government or business |
|
or academia, all start paying the kind of attention that we |
|
really, really need. |
|
Dr. Mafdry, you, in your testimony, talked about the |
|
overarching points. And the fourth one you talked about was |
|
that we pay attention, critical attention, to the artificial |
|
intelligence supply chain, that it will structure the |
|
distribution of power in a new AI world. Could you take just a |
|
moment to explain that? |
|
Mr. Mafdry. Of course. So, essentially the way AI is being |
|
deployed right now is no longer just one entity who gathers the |
|
data, trains the model, and applies it and deploys it to a |
|
given task. The way things happen is that there is a supply |
|
chain, in particular with this new generative models, like, |
|
they are very expensive to train, but they are very capable. |
|
And essentially what happens is that, you know, one of these |
|
companies--there is very few companies that can afford training |
|
such a model--they essentially develop it and then let other |
|
companies build on top of it. |
|
So, think of this as just like having initial capability |
|
and just adjusting it. For instance, you have a model like |
|
ChatGPT. It is able to summarize texts and understand to some |
|
extent the texts. So, maybe you have a hiring tool that then |
|
you build on top of it that essentially uses this capability to |
|
screen resumes and something. There is many risks there, but |
|
this is just an example. So, we are heading this kind of supply |
|
chain of interdependencies. |
|
And again, we have upstream where there is very few |
|
players. There is very few critical important models on which a |
|
lot of the economy depends, and then there are all these |
|
interactions between these other things. So, this is something |
|
that now we have to think. There's an ecosystem. |
|
Mr. Mfume. I see. I see. |
|
This sort of boasting that China has been doing that, by |
|
2030, they will be the dominant player is scary also. And the |
|
fact that the universities in Beijing and elsewhere are openly |
|
sort of trying to develop this and develop thinking that way, |
|
and that it's only seven years from now makes me, again, very |
|
concerned. |
|
I want to talk about risk for just a moment, and then my |
|
time will have expired. This whole notion of a war and |
|
decisions being made with the use of AI in a millisecond that |
|
counter and then counter-counter the decision. I don't know to |
|
what extent the military establishment--I assume they're |
|
looking at this as much as you are, but it is interesting. |
|
Now, I'd like to ask also, there are, as you know, fallible |
|
algorithms. You know better than I. They are just misleading or |
|
they are incorrect. What happens in consumerism, in business, |
|
in law enforcement, in military context that frighten you the |
|
most as risk as a result of an infallible algorithm? Any of |
|
you? |
|
Mr. Schmidt. Quickly, the biggest issue, as I mentioned, is |
|
the compression of time. Let's assume you have time. Then the |
|
question is, who gets to decide between the system and the |
|
human? And I'm very concerned about a misalignment of interests |
|
where the human has one set of incentives, the computer has |
|
been trained against a different set of outcomes, and the whole |
|
society wants an even different goal. |
|
And I spent lots of time with the military, who I'm really, |
|
really proud of and fond of, and they all want systems that |
|
help them automate their decisions. In practice, their use of |
|
technology will be largely to replace boring and uninteresting |
|
jobs, like watching TVs and things like that. These are things |
|
like Project Maven and its successors and so forth. So, I think |
|
at the moment, the government at the military level is going to |
|
use these more for sensing and analysis and not decision- |
|
making. Just to make it very clear, I think we would all agree |
|
it is not ready to make a final life critical decision. It may |
|
never be, but it is certainly not now. |
|
Mr. Mfume. Yes. Thank you. My time is expired. |
|
Thank you, Madam Chair, appreciate it. |
|
Ms. Mace. Thank you. Great questions. |
|
I would now like to recognize Congressman Burlison. |
|
Mr. Burlison. Thank you, Madam Chair. |
|
I, for one, am not afraid of the advent of AI. In fact, I |
|
want to welcome our future overlords. But I will say, I do see |
|
a lot of promise. You know, working in healthcare technology, |
|
we see an amazing opportunity to be able to comb the data |
|
records of patients, be able to use--be able to take that data |
|
and be able to accurately diagnose better than probably any |
|
medical professional possibly ever could to a greater degree of |
|
accuracy what you might be facing. To me, there is tremendous |
|
opportunity, but I also do recognize some of the threats, |
|
obviously. |
|
To that end, my question for you, Dr. Schmidt, first is |
|
that, given the state--the size and scope of the equipment |
|
that's necessary today, we're limited to what actors do have |
|
the ability to use AI, right? So, at least we know who has |
|
access to it, who's utilizing it. It's not like we have people |
|
in a Nigerian criminal syndicate using AI at this point. Is |
|
that correct? |
|
Mr. Schmidt. I can assure you it is coming because of |
|
diffusion. Basically, the models are trained very expensively, |
|
but when they're used for inference, which is where they answer |
|
questions, it's quite simple. So, I would expect us to see |
|
terrorists, other bad actors begin to use this technology in |
|
ways that we should anticipate. |
|
Mr. Burlison. And they--but at this point, they would have |
|
to access it on another platform. Someone would have to spend |
|
the resources to develop the tech--to house the data, et |
|
cetera? |
|
Mr. Schmidt. And Aleksander can--Professor Mafdry can help |
|
me here because we work together. The simplest way to |
|
understand it is the training part is really expensive, but you |
|
can take the trained information and put it on a laptop, and |
|
then it can be used. So unfortunately, in this scenario, all |
|
you need is a computer or a phone to do your evil acts. |
|
Mr. Burlison. OK. Dr. Crowder, my question to you relates |
|
to the quantum computers. These too, these are machines that |
|
you just couldn't walk around, handheld devices, right? Can you |
|
walk us through what it takes to--what the environment |
|
requirements are, what it takes to have a quantum computer? |
|
Mr. Crowder. Yes. I mean, the--we deploy them right now in |
|
regular data centers, but they are not laptops. They are not |
|
mobile phones. They are large, complex systems, and they are |
|
very, very hard to calibrate and manage, and that is a major |
|
trade secret of how to keep them up and running. That's |
|
probably going to be true for quite a while. |
|
So, you know, right now, we don't actually sell systems. We |
|
sell cloud access, because there's a small number of people who |
|
know how to actually keep them running, operating, et cetera, |
|
et cetera. So, you know, obviously that has some benefits from |
|
an IP protection security point of view as well. |
|
Mr. Burlison. And so, this is being used at what scale? How |
|
many businesses or---- |
|
Mr. Crowder. So, we have over 200 partners, and we |
|
carefully select what regions of the world we do business in |
|
and carefully select who we partner with. But we have got over |
|
200 industry and academic and research partners who are |
|
leveraging--we've got about 26 computers right now that are |
|
accessible via the cloud. |
|
Mr. Burlison. To that end, I know that from other testimony |
|
from other hearings, we have been hearing that the Chinese |
|
Government has a pattern of sending students to American |
|
universities, who then are able to glean data in working in |
|
cooperation on some of these projects. |
|
Are you aware of that activity? |
|
Mr. Crowder. Not personally aware of that activity. We |
|
obviously are very--for business reasons for our, whatever you |
|
want to call them, crown jewels or most protected IP, we are |
|
very careful not to--we are very careful who is part of that |
|
work. |
|
Mr. Burlison. And then again, my last question to you is, |
|
in the subject of quantum entanglement, has that had any real |
|
world applications or potential real world applications? |
|
Mr. Crowder. Right now, nobody from our perspective have |
|
proved that there is practical use, which means it is better |
|
than simulating it or just using classical computers. But we |
|
think that is going to change in the next very soon amount of |
|
time. The computers are getting rapidly advancing, and we think |
|
by the middle of this decade, there will be practical use. |
|
We are working with a lot of, you know, U.S. companies on |
|
applications today. I had mentioned a couple of them before, |
|
like Boeing on looking at better airplane wing optimization |
|
materials, fraud detection for banks, looking at medical health |
|
records and trying to predict more efficient treatment for |
|
patients with healthcare, like sciences companies. We have a |
|
big partnership with Cleveland Clinic looking more broadly |
|
across five sciences. But it is not practical today to be |
|
better than what we have classical. But we think that's going |
|
to come in the next couple of years. |
|
Mr. Burlison. Thank you. |
|
Ms. Mace. Thank you. |
|
I would now like to recognize Congressman Gomez for five |
|
minutes. |
|
Mr. Gomez. Thank you, Madam Chair. |
|
Before we begin, I want to--I was thinking about this: How |
|
do we rank AI in the history of development in humankind? It is |
|
something that I believe is--could be extremely startling. It's |
|
one issue that I have random people bring up to me on the |
|
streets. Some people compare it to the invention of the |
|
television or computer, or the internet, and I think it's |
|
beyond that, because this is something that--it makes it hard |
|
to discern something that you are looking at, a photo, a video, |
|
or even words on a piece of paper if it was actually written or |
|
developed by a human. |
|
And that is something that I think most people are trying |
|
to wrap their minds around. How is this revolutionary |
|
technology going to fundamentally change the way we live our |
|
lives, the way we interact with one another, the way we |
|
interpret information that is coming in? Because when you can't |
|
discern what is actually created by a person and what is |
|
developed by a computer program, then people start questioning |
|
all sorts of things. |
|
And that's one of the challenges. Maybe that is a |
|
philosophical challenge. Maybe it is a real life government |
|
regulatory challenge, but it is something that really, I think, |
|
is at the heart of it, when people start to question what is |
|
real and what is not. |
|
But I recognize it has--AI has a lot of great potential. |
|
Everything from predicting new variants of COVID to detecting |
|
certain types of cancers that doctors miss. The potential is |
|
staggering. But people want guardrails on this new technology. |
|
If not, it can and will be misused. |
|
When I first got elected, one of the--somebody ran a test |
|
of Members of Congress, and there was about--I think it was |
|
about 28 of us that got matched with people that had committed |
|
crimes. And this was under the best circumstances. They were |
|
using our photos from our websites that were taken with the |
|
best lighting and the best quality. So, AI has a potential also |
|
to have inherent bias built into it, and it often |
|
disproportionately impacts people of color, women, and that is |
|
a concern. |
|
So, how do we address those limitations on AI? How do we |
|
safeguard against the violation of people's civil liberties? |
|
And this is something that even my--when Mark Meadows was on |
|
this committee, him and I and others agreed that this was a |
|
problem. We just couldn't figure out a solution. |
|
So, Ms. Hickok, how can Congress best help address AI's |
|
racial bias? What can we do as a body and the Federal |
|
Government do to protect individual's civil liberties and, at |
|
times, their right to privacy as well? |
|
Ms. Hickok. Thank you for the question. And I'm really |
|
thankful that you mentioned the civil liberties, because with |
|
AI systems, as I mentioned earlier, you're talking about every |
|
single industry and domain that is going to be--is going to be |
|
even further impacted by this. |
|
You talk about civil liberties and access to resources and, |
|
unfortunately, that spans from anything from housing to |
|
employment to education to insurance loan to policing and |
|
criminal justice decisions and judiciary decisions. My concern |
|
is, as AI also, if you don't have the guardrails now and these |
|
systems are imbedded in the public sector services as well as |
|
private services, that they are also going to eventually |
|
connected. |
|
So, one erroneous decision from one system is going to be |
|
the input to another system, and we are going to completely |
|
lock people out of resources and opportunities, and that is |
|
going to widen the gap between haves and have nots. And it is |
|
also going to widen the gap within the society that we are all |
|
trying to narrow. |
|
How can we narrow that? How can we keep the systems |
|
accountable? It is really about the people and organizations |
|
and how we use them that we should be focusing on. Putting the |
|
civil liberty is putting the freedoms and rights at the center |
|
of it, and making sure that these systems--the systems that we |
|
use, especially that impact the resources and the rights, are |
|
built accountably, transparently, replicable. We heard from my |
|
co-panelists and witnesses that a lot of the times the systems |
|
are opaque. We don't know how they work, and we also cannot |
|
replicate the decisions. |
|
So, you might be denied a credit. You might be denied |
|
insurance or a job. It might not be--if you are trying to keep |
|
the organization accountable, we will not be able to trace back |
|
and keep them liable as well. So, we need to make sure from the |
|
start, from the very start, from the design data and design |
|
stages, that we put those guardrails in place, and we keep |
|
organizations and the users accountable. |
|
Mr. Gomez. Thank you. And my time has expired, but the |
|
question is: Is it too late to put those guardrails on? |
|
Ms. Hickok. It is not at all. In fact, at CAIDP, our |
|
students, especially our law students, asked that question last |
|
week, is it too late, is it inevitable, has the ship sailed? |
|
No, it is not. The humans, organizations, lawmakers, the |
|
humans, the users behind it hold the power. |
|
Mr. Gomez. Thank you so much. I yield back. |
|
Ms. Mace. Thank you, Congressman. |
|
I would now like to recognize Congresswoman Greene. |
|
Ms. Greene. Thank you. I think this is a very important |
|
hearing to have as AI is progressing and working in a lot of |
|
different sectors. And I really appreciate each of you being |
|
here. |
|
I'm definitely not an expert in AI, but I would like to |
|
talk about the fears and concerns that people in my district |
|
and people all over the country have when it comes to AI. |
|
We certainly don't like the idea of AI replacing humans and |
|
replacing people, especially when it comes to jobs. And so, |
|
when there's headlines like Alphabet announcing 12,000 job cuts |
|
globally while chief executive officer singled out AI as the |
|
key investment area, that is what people start to think about. |
|
Or when Microsoft announces its $10 billion investment in |
|
OpenAI just days after saying it would lay off 10,000 |
|
employees, those are the kinds of things people think about. |
|
Now, there is a difference between robotics and AI. |
|
Obviously, robotics are a good thing. For example, tightening |
|
bolts or moving heavy objects like in manufacturing, we really |
|
appreciate that. But when it comes to AI being able to be |
|
smarter than humans or replace humans on the job, I think that |
|
is a major concern, especially for a country that's over $30 |
|
trillion in debt and an economy that is struggling. |
|
This is something also concerning for education, learning |
|
that ChatGPT scored higher than many people on the MBA exam |
|
that was administered at Penn's elite Wharton School. That's |
|
definitely concerning, especially when thinking about how that |
|
could affect education. Just recently, ChatGPT was--is |
|
currently banned in New York City schools over cheating |
|
concerns, but then you think about what would this look like if |
|
AI became teachers, especially after the devastation caused to |
|
children's education levels, but also more importantly, kids |
|
being taught at home on computers. That was more devastating to |
|
them. |
|
The idea that AI could replace software engineers, |
|
journalists, graphic designers, that is also extremely |
|
concerning. So, I think these are important conversations to |
|
have. |
|
But something that happened, I just learned about in |
|
researching for this hearing, was that there are scams that |
|
happen to people where AI is so intelligent, it is able to |
|
imitate people's voices and images. And there is been people |
|
taken advantage of in horrible ways where they have gotten |
|
phone calls from who they thought was their loved ones but was |
|
not. And their loved one, which was really an artificial |
|
intelligence voice, mimicking their loved one was calling for |
|
help in serious distress, and then they got scammed out of a |
|
lot of money. That's terrifying and concerning that that can |
|
happen. |
|
But another thing that happened recently was when San |
|
Francisco officials voted in December against a controversial |
|
measure that would have allowed police to deploy robots to use |
|
lethal force in extreme situations, but this happened after the |
|
San Francisco Board of Supervisors came a week--it was a week |
|
after the board voted to approve the policy to allow what |
|
people called killer robots. But this is what people think |
|
about when they think of AI. They think of a robot that has the |
|
artificial intelligence to replace the police officer. |
|
But then the application to the military is where I thought |
|
was pretty concerning. |
|
Dr. Schmidt, I wanted to ask you, because I took to Google |
|
on this issue. And I wanted--I saw a headline that said ``AI's |
|
impact on warfare will be as big as nuclear weapons.'' And I |
|
also saw another headline that said ``Eric Schmidt Is Building |
|
the Perfect AI War-Fighting Machine.'' So, I thought you would |
|
be the perfect person to ask about this. Could you explain a |
|
little bit? |
|
Mr. Schmidt. Let me be clear, that's for the benefit of the |
|
United States. |
|
Ms. Greene. Only if it is in the United States' hands, |
|
though, Dr. Schmidt. |
|
Mr. Schmidt. And it will be. |
|
The trends in the military are fundamentally autonomy and |
|
drones and intelligence sensing gathering. The military spends |
|
most of its time looking at things and trying to analyze it. |
|
So, in the near term, the benefits to the military are |
|
profound. It allows the service people who we have trained |
|
exquisitely who are watching dumb screens to use their higher |
|
factory skills and have the computer say, hey, look, this tank |
|
moved or, hey, this thing happened over here, can you analyze |
|
it. |
|
I think a better framing for your constituents' fear is to |
|
say that AI will make the people much more successful in what |
|
they do, and that will drive higher incomes, higher jobs. And I |
|
think that that's the best, at least in the next 20 years, |
|
narrative about AI. It is true in the military. It is true in |
|
civilians as well. |
|
Ms. Greene. One more question. My time is expired, but how |
|
do we--with China and their ability to constantly spy on us and |
|
steal our technology and information, how could we prevent |
|
China from stealing this type of artificial intelligence with |
|
our military? And thank you. |
|
Mr. Schmidt. Of course. The bad news is that these research |
|
ideas are in the public domain and international, so we can't |
|
prevent China from getting it. The Trump and Biden |
|
administrations have done a good job of restricting access to |
|
hardware, which is helpful. So good job, all of you. |
|
With respect to software, the biggest answer is more |
|
software people, trained in the West, trained under our values, |
|
building systems that you as our Representative have some level |
|
of regulatory control over. When they do it in China, you can't |
|
pass a law to change that, but you can in the United States. |
|
Ms. Mace. Thank you. |
|
All right. I would like to now toss it over to Congressman |
|
Khanna for five minutes. |
|
Mr. Khanna. Thank you, Madam Chair, and thank you for your |
|
leadership, for your bipartisan cooperation and collaboration |
|
on the quantum bill, and the approach you have taken to work |
|
across the aisle. |
|
Dr. Schmidt, I respect your leadership in Silicon Valley. |
|
There's a paradox in my mind that I would love your insight. On |
|
the one hand, DARPA in the Department of Defense gave us the |
|
internet, as you know, with Vinton Cerf, gave us GPS, gave us |
|
the drone, gave us the mouse, probably the most innovation in |
|
the history of the 20th century, defense technology. And yet, |
|
now it seems there is this problem of the adoption of |
|
innovative technology. |
|
Why is the model that gave us all of this revolutionary |
|
technology not working? |
|
Mr. Schmidt. Thank you, Congressman. And you have really |
|
helped in a lot of these areas. |
|
If you go back to Vannevar Bush, the National Science |
|
Foundation, and DARPA, those are the engines that got us all |
|
here. We are all here fundamentally because of early decisions |
|
made by the Federal Government to invest in researchers who |
|
then we built on top of. So, I'm incredibly grateful to them. |
|
In the case of the government, and particularly the |
|
military, those innovations go into a bureaucracy that is not |
|
organized in a way to take them. And a simple example is |
|
software. The military is organized around procurements of a |
|
15-year cycle and complicated bidding among a small number of |
|
contractors. That is not how software works. And a number of us |
|
have worked hard to get software treated more as a continuous |
|
process. But the military, for example, would benefit by a |
|
large expansion of a number of more software people, just |
|
fixing stuff, making things work, making them smarter. That is |
|
a simple thing that you could do. |
|
Mr. Khanna. To that end, what do you think about an actual |
|
service academy around technology, cyber, AI? |
|
Mr. Schmidt. We looked hard at creating a military service |
|
academy when I was doing the AI commission. And the military |
|
has really, really good people in their academies. And what |
|
they do is, because of the way military promotions work, you |
|
take some brilliant person, you make them go stand guard duty |
|
for a while, which is stupid. Sorry to be blunt. It is much |
|
better to change the HR policies, which the military is trying |
|
to do now. In particular, Secretary Brown in the Air Force is |
|
trying to create a technical path to keep these people. That is |
|
how you solve that problem. And let me give the rest of my time |
|
to Aleksander. |
|
Mr. Mafdry. Yes. So, I just wanted to add because it is a |
|
very important question that you ask. So, I actually happen to |
|
co-lead and codevelop at MIT an executive education class, AI |
|
for national security leaders, which essentially hosts a number |
|
of general offices from Pentagon and other places to come and |
|
learn about AI. And this is a three-day program. Half of this |
|
program is not about AI; it's about organizational management |
|
aspects. |
|
So, this is what you recognize. There's a lot of |
|
frustration in DOD in your top military leaders that the |
|
technology is developed. DARPA did their part, although they |
|
should do more particularly in generative languages. But then |
|
we hit the bureaucracy. And there is just a lot of |
|
organizational problems that are kind of silly that the DOD is |
|
completely crippled in terms of adoption of AI. So, that is |
|
where we need the attention. |
|
Mr. Khanna. Well, I look forward to working with you |
|
[inaudible] with Representative Mace and Representative |
|
Gallagher. |
|
One other question--I mean, I'm back home in Silicon |
|
Valley. It seems the new thing there is everyone is doing AI. |
|
I'd be curious, Dr. Schmidt and Dr. Mafdry, how do you see-- |
|
will Silicon Valley lead the world in AI? How are we doing |
|
compared to China? |
|
And then one comment from my own version of American |
|
exceptionalism, it drives me crazy when Europeans are lecturing |
|
us about AI and technology. You know, I don't see Google, |
|
Apple, Tesla. I get they say they are going to innovate in |
|
policy, they are going to also innovate in technology. How are |
|
we compared to Europe as well? |
|
Mr. Schmidt. My cynical answer about Europe is that Europe |
|
is going to lead in regulation and, therefore, not lead in |
|
anything else. Their efforts do not appear to be successful, as |
|
you have pointed out. |
|
The reason we are so excited about AI is that anything that |
|
makes humans smarter and makes algorithms smarter and makes |
|
discoveries quicker is a horizontal technology that is |
|
transformative. The opportunities to make basic advancements |
|
outside of language models, right, are profound in terms of |
|
science, materials, plastics, every kind of logistics, every |
|
kind of analytical problem, as has been summarized by the |
|
panel. |
|
So, I think that AI is here to stay. It is the next big |
|
wave. I don't know when it will end, but we are still very |
|
early. Remember, we still don't understand exactly how these |
|
algorithms work. We also don't understand how big the models |
|
have to be. At some point, we'll know. But we are not anywhere |
|
close to being able to answer those questions. |
|
Mr. Mafdry. If I can just add very quickly because you |
|
asked the question about Silicon Valley. Silicon Valley is |
|
doing great. They will do great job. They are clearly |
|
harnessing this progress, but we as a country should not |
|
abdicate the progress on the strategically important technology |
|
just for Silicon Valley. Again, they will do great, but we |
|
should be doing more, and the U.S. Government should be doing |
|
more. |
|
Mr. Schmidt. Speaking as a professor at MIT. |
|
Mr. Mafdry. Yes. But I like Silicon Valley. |
|
Ms. Mace. In closing this afternoon, first of all, I just |
|
want to thank you all, all of our panelists for your time and |
|
your talent and everything that you have shared with us today. |
|
This will be, Congressman Mfume, the first of a series of |
|
hearings that I hope that we'll have on AI. I don't think that |
|
we are ready for what is going to happen in a very short period |
|
of time. And I think, if it is not happening already, it will |
|
be in the next five years where AI will be programming AI, and |
|
then what's next? |
|
And so, this was a great first discussion to start this |
|
conversation about what needs--what we need to be talking about |
|
in regards to this. |
|
So, in closing, I want to thank all of our panelists once |
|
again for your insightful testimony today. You have given us a |
|
tremendous amount to think about, and AI was created by humans, |
|
but it doesn't mean that it is going to be easy for all of us, |
|
especially up here on the Hill, to grasp what is before us and |
|
what is imminently coming. We appreciate the panel's expertise |
|
and ability to shed light on the state of the science and the |
|
broader societal implications that policymakers must consider. |
|
And I would like to yield to the Ranking Member, |
|
Congressman Connolly, for your closing remarks. |
|
Mr. Connolly. Thank you so much, Madam Chairwoman. |
|
And I found this an intriguing conversation, but wanting |
|
more. And like Mr. Mfume, I think we have opened the door to a |
|
lot further in-depth exploration, hopefully by this |
|
Subcommittee and by the Congress, because there are lots of |
|
issues we have to face. |
|
And while you may be right, Dr. Schmidt, about dismissing |
|
the Europeans as regulators but not innovators, on the other |
|
hand, given what we heard from Ms. Hickok and Dr. Mafdry about |
|
the need for some Federal intervention here, there have to be |
|
guidelines and guideposts so that we are off on the right foot |
|
and not facing profound issues later on where the technology is |
|
advanced and we never either anticipated it or addressed it. |
|
Maybe there are things we can learn from the Europeans in the |
|
regulatory guidelines, either things not to do or things to do. |
|
But any rate, I just think there is a lot more for us to |
|
explore, and I really appreciate this being the first of a |
|
series of hearings. |
|
Thank you, Madam Chairwoman. I yield back. |
|
Ms. Mace. Thank you. And I look forward to working with |
|
everyone on both sides of the aisle on this issue. It is very |
|
important. |
|
With that and without objection, all Members will have five |
|
legislative days within which to submit materials and to submit |
|
additional written questions for the witnesses which will be |
|
forwarded to the witnesses for their response. |
|
If there is no further business, without objection, my |
|
first Subcommittee stands adjourned. |
|
[Whereupon, at 3:57 p.m., the Subcommittee was adjourned.] |
|
|
|
[all] |
|
</pre></body></html> |
|
|