davidshapiro_youtube_transcripts
/
Escaped Sapiens David Shapiro on AI alignment safety and the future of humanity_transcript.csv
text,start,duration | |
is there anything that makes humans,0.06,3.0 | |
Irreplaceable,1.86,3.3 | |
so are we getting into a stage where,3.06,4.38 | |
there's nothing that AI that humans can,5.16,4.8 | |
do that AI won't be able to do,7.44,5.279 | |
uh from a functional standpoint from a a,9.96,5.759 | |
an objective standpoint I don't think so,12.719,4.98 | |
um and that actually begs a very deep,15.719,3.421 | |
philosophical and spiritual question,17.699,2.821 | |
which is what is the point of living,19.14,4.5 | |
what is the point of being a human,20.52,5.16 | |
um and that uh is something that I've,23.64,4.559 | |
done some work on I've wrote a paper or,25.68,5.58 | |
a short book called post nihilism where,28.199,5.581 | |
what I suspect is that we are barreling,31.26,4.92 | |
towards uh what I call a nihilistic,33.78,3.54 | |
crisis or actually we're in the middle,36.18,3.059 | |
of a nihilistic crisis and it actually,37.32,3.68 | |
started with the Industrial Revolution,39.239,4.921 | |
if you look at a lot of poetry and,41.0,5.98 | |
literature works of fiction during the,44.16,5.28 | |
the rise of the Industrial Revolution a,46.98,4.079 | |
lot of people had a lot of existential,49.44,3.299 | |
anxiety about what was the point of,51.059,4.801 | |
Being Human in an era of machines and,52.739,4.621 | |
this kind of pops up every now and then,55.86,4.26 | |
right same thing happened with computers,57.36,5.539 | |
um with the Advent of you know uh,60.12,5.34 | |
high-speed computers nuclear weapons so,62.899,5.021 | |
on and so forth technological advances,65.46,3.839 | |
advancements tend to give us some,67.92,2.66 | |
existentialing,69.299,3.961 | |
but to your question about like okay,70.58,5.38 | |
what is the benefit of being a human in,73.26,4.62 | |
a world where from a product,75.96,3.78 | |
productivity standpoint or an economic,77.88,3.48 | |
standpoint machines can do everything,79.74,3.48 | |
that we can do better faster and cheaper,81.36,4.98 | |
what's the point and so that is where we,83.22,5.219 | |
have to change our orientation towards,86.34,4.919 | |
how we value our own life and our own,88.439,4.5 | |
subjective experience so that's a deeply,91.259,3.661 | |
deeply philosophical and religious,92.939,4.561 | |
perspective or a question and it's it's,94.92,5.04 | |
really interesting because depending on,97.5,4.2 | |
someone's spiritual upbringing or,99.96,4.5 | |
spiritual disposition the question lands,101.7,5.279 | |
very differently because many uh,104.46,4.199 | |
religious doctrines around the world,106.979,3.721 | |
basically say that humans have a soul,108.659,4.861 | |
and that sets us apart and so whether or,110.7,5.76 | |
not that's true uh people have a model,113.52,5.279 | |
for just saying my subjective experience,116.46,6.299 | |
of being is Matt is very meaningful and,118.799,7.741 | |
it is unique and so part of overcoming a,122.759,6.301 | |
nihilistic crisis is we all have to face,126.54,4.62 | |
that whether or not not we believe in,129.06,4.2 | |
souls or God or whatever,131.16,5.159 | |
and we have to kind of go back to basics,133.26,5.76 | |
and look at the subjective experience of,136.319,4.261 | |
our own being and so back to your,139.02,3.42 | |
question earlier about children I,140.58,3.42 | |
suspect that children who grow up with,142.44,4.26 | |
AI they will just intrinsically know oh,144.0,4.08 | |
yeah my experience is different from,146.7,3.179 | |
this machine and that's okay and that,148.08,3.299 | |
they won't have any existential anxiety,149.879,4.261 | |
about it I hope at least do you have are,151.379,4.381 | |
you hopeful for the future or do you,154.14,3.84 | |
have this anxiety,155.76,5.58 | |
um no I I am uh I think I'm biologically,157.98,5.46 | |
programmed to be optimistic I just I,161.34,3.899 | |
can't be cynical,163.44,5.04 | |
um and I I,165.239,5.881 | |
part of that is that I've done a,168.48,4.08 | |
tremendous amount of work to understand,171.12,3.6 | |
what the dangers and risks are and I've,172.56,4.44 | |
also tried to contribute to coming up,174.72,4.86 | |
with a more optimistic outcome the,177.0,4.379 | |
machines so we we all learned this from,179.58,3.42 | |
watching Scooby-Doo,181.379,3.841 | |
um the monsters are always humans right,183.0,4.26 | |
there's no such thing as as an evil,185.22,3.9 | |
monster out there the problem is always,187.26,4.92 | |
humans and so this is this is a big,189.12,5.24 | |
reason that I've done my work is because,192.18,4.62 | |
you know it's not it's not that a,194.36,4.36 | |
machine is going to replace you and,196.8,3.359 | |
that's a bad thing right we all,198.72,3.599 | |
fantasize about like hey I want to you,200.159,3.72 | |
know go live in the countryside and just,202.319,3.481 | |
go fishing every day we all know what we,203.879,4.261 | |
want to do if we don't have to work what,205.8,5.4 | |
we are truly afraid of is not being able,208.14,5.22 | |
to take care of ourselves is that if the,211.2,4.14 | |
machine takes our job we're gonna go,213.36,3.9 | |
hungry we're gonna lose our home we're,215.34,4.02 | |
gonna end up lonely and and whatever,217.26,4.8 | |
that's the actual fear,219.36,4.799 | |
um nobody actually wants to keep working,222.06,4.38 | |
right nobody want like I remember one of,224.159,4.921 | |
the advertisements for um for a like,226.44,5.1 | |
health insurance here in America was you,229.08,3.9 | |
get to keep your health insurance you,231.54,2.82 | |
like your health nobody likes health,232.98,4.259 | |
insurance it's a necessary evil right,234.36,6.959 | |
jobs occupations are a necessary evil of,237.239,6.301 | |
the economic environment that we're in,241.319,4.14 | |
and the technological limitations that,243.54,3.96 | |
we're in and so as these things progress,245.459,3.481 | |
this is this is I'm basically just,247.5,3.599 | |
unpacking why I'm optimistic as these,248.94,3.78 | |
things progress I hope that we're all,251.099,3.36 | |
going to be able to have kind of a Back,252.72,3.54 | |
to Basics moment where it's like you,254.459,3.12 | |
wake up one day and it's like how do you,256.26,2.58 | |
actually want to live right if you want,257.579,2.94 | |
to go fishing every day do it if you,258.84,3.54 | |
want to focus on being an opera singer,260.519,4.381 | |
go do that you know there's we all have,262.38,4.74 | |
stuff that we want to do but that we,264.9,4.44 | |
sacrifice for the sake of earning enough,267.12,4.2 | |
money to take care of ourselves and that,269.34,5.34 | |
is the reality for most of us today,271.32,4.8 | |
one of the reasons why we have this,274.68,3.36 | |
worry is because currently we live in,276.12,3.48 | |
sort of a negotiated environment right,278.04,4.02 | |
the successive labor movements was,279.6,4.92 | |
because labor was needed when humans are,282.06,4.699 | |
no longer needed,284.52,4.92 | |
there's sort of a worry that,286.759,4.121 | |
we're not going to have the opportunity,289.44,2.64 | |
to go fishing,290.88,2.46 | |
right where they're going to have,292.08,3.3 | |
nothing and I guess that's that's the,293.34,3.48 | |
the worry that you're pointing at what,295.38,3.24 | |
what do you think the first jobs are,296.82,3.9 | |
that are going to go,298.62,3.66 | |
well there's already been quite a few,300.72,2.64 | |
layoffs,302.28,3.96 | |
um various uh communities on Reddit or,303.36,4.98 | |
private communities on Discord,306.24,4.56 | |
um so for instance my uh fiance's uh,308.34,4.38 | |
were both writers but she's on a few uh,310.8,3.899 | |
private writing communities,312.72,3.479 | |
um copywriters have already been laid,314.699,4.201 | |
off and replaced by AI,316.199,4.741 | |
um uh marketing teams have been notified,318.9,3.6 | |
that you know they've got a year until,320.94,2.88 | |
they're all going to get laid off and,322.5,3.66 | |
replaced by you know AI generated images,323.82,4.86 | |
and AI generated emails,326.16,4.62 | |
um so it's happening,328.68,5.04 | |
um yeah that's that that's where we're,330.78,6.24 | |
at now I guess to to your your larger,333.72,5.28 | |
point of you know if we're all,337.02,3.959 | |
replaceable you know what's what's the,339.0,4.08 | |
bottom line and the fact of the matter,340.979,4.381 | |
is from a corporate perspective from uh,343.08,3.899 | |
from the perspective of neoliberalism,345.36,4.559 | |
human labor is one of the most expensive,346.979,5.461 | |
aspects of productivity and it's also,349.919,4.861 | |
the biggest constraint you look at a,352.44,4.5 | |
population decline in places like China,354.78,4.02 | |
and Japan because China just crested,356.94,3.96 | |
right so from here on out China's,358.8,3.54 | |
population is going down for at least,360.9,3.239 | |
the next Century Japan has been in,362.34,3.78 | |
Decline for a couple decades now uh,364.139,4.201 | |
ditto for Italy and a few other nations,366.12,4.799 | |
so their labor force is Contracting,368.34,5.34 | |
right and from an economic perspective,370.919,6.241 | |
that's really bad for for for Nations so,373.68,6.0 | |
AI hopefully will actually Shore up,377.16,4.86 | |
those Labor uh labor markets and,379.68,5.7 | |
actually replace lost human labor now,382.02,5.28 | |
because humans are so expensive right,385.38,4.62 | |
you can pay uh 20 a month for chat gbt,387.3,4.56 | |
and it can basically serve as an,390.0,3.84 | |
executive assistant and personal coach,391.86,3.839 | |
and every it can replace literally,393.84,3.479 | |
thousands of dollars worth of Labor and,395.699,3.901 | |
it costs 20 a month chat GPT is,397.319,3.541 | |
infinitely cheaper than most human,399.6,2.879 | |
employees,400.86,2.76 | |
um and that's only going to get better,402.479,3.06 | |
right because either the model is going,403.62,4.019 | |
to get more efficient and cheaper,405.539,3.481 | |
um or it's going to get smarter and more,407.639,3.0 | |
powerful and therefore more valuable or,409.02,5.16 | |
both in all likelihood so one one of the,410.639,4.861 | |
things that I predict,414.18,3.12 | |
is that we are going to have a post,415.5,4.56 | |
Labor uh market economy before too long,417.3,5.58 | |
and in that respect uh basically,420.06,5.46 | |
economic productivity will be decoupled,422.88,4.319 | |
from Human labor,425.52,3.48 | |
um and in that case you know you're,427.199,3.181 | |
going to see quadrillion dollar,429.0,3.84 | |
valuation uh for companies that have no,430.38,3.539 | |
employees,432.84,3.0 | |
and that might sound like that that,433.919,4.081 | |
could be an ingredient for a dystopian,435.84,5.04 | |
world that nobody wants to live in,438.0,4.74 | |
we'll get to like the regulation and,440.88,4.5 | |
stuff of that later but from a from a,442.74,5.7 | |
from a purely GDP perspective AI is,445.38,4.08 | |
going to be the best thing that ever,448.44,4.86 | |
happened to GDP to uh to uh economics,449.46,6.54 | |
because again it will decouple uh human,453.3,4.2 | |
labor from the constraint and that there,456.0,2.94 | |
there will still be a few constraints,457.5,4.62 | |
natural resources Rare Minerals uh fresh,458.94,6.06 | |
water arable land right there's going to,462.12,4.32 | |
be there's always going to be some,465.0,3.599 | |
physical constraints but we're going to,466.44,4.56 | |
remove human labor as one of the main uh,468.599,5.281 | |
constraints to economics and that is,471.0,4.38 | |
going to mandate kind of those things,473.88,2.759 | |
like you said like if you want to go,475.38,3.78 | |
fishing well how right if you don't have,476.639,4.921 | |
any economic power if you don't have any,479.16,5.099 | |
way to make a demand then that's a big,481.56,4.38 | |
problem which is what we're going to,484.259,2.821 | |
have to negotiate we're going to have to,485.94,2.58 | |
negotiate a new social contract,487.08,2.88 | |
basically,488.52,3.06 | |
what do you think the impact is going to,489.96,3.9 | |
be on births ultimately do you think,491.58,4.86 | |
people are going to just start having AI,493.86,5.1 | |
children because it's cheaper,496.44,4.08 | |
you know that's a really difficult,498.96,3.9 | |
question I could see it going either way,500.52,4.44 | |
um there's plenty of books and and and,502.86,5.04 | |
fiction out there and research papers,504.96,4.799 | |
um people have predicted you know the,507.9,4.74 | |
population uh explosion you know the,509.759,4.381 | |
Earth will become uninhabitable because,512.64,2.94 | |
we'll have billions and billions of,514.14,3.48 | |
people that we can't feed other people,515.58,3.36 | |
are worried that you know the population,517.62,3.359 | |
is going to collapse,518.94,3.36 | |
um and I actually had a pretty long,520.979,3.36 | |
conversation about this just to kind of,522.3,4.26 | |
clarify my own ideas uh again with chat,524.339,3.781 | |
GPT,526.56,4.38 | |
um and so there's a few driving factors,528.12,6.42 | |
that cause uh birth rates to decline,530.94,5.78 | |
um uh women entering the workforce,534.54,4.26 | |
education and empowerment for women,536.72,5.1 | |
access to birth control so it turns out,538.8,6.479 | |
when a society advances and becomes a,541.82,6.16 | |
little bit more uh sophisticated or or,545.279,5.101 | |
gains more access or some you know Ginny,547.98,4.38 | |
coefficient goes up whatever metrics you,550.38,4.68 | |
use education goes up fertility rates go,552.36,3.539 | |
down,555.06,2.7 | |
some of that has to do with the choices,555.899,4.141 | |
of Family Planning you know men and,557.76,3.6 | |
women decide to have fewer children,560.04,3.6 | |
women have more control over their own,561.36,3.479 | |
fate,563.64,3.18 | |
um and so fertility rates tend to go,564.839,4.021 | |
down and this is a very very reliable,566.82,4.019 | |
Trend globally,568.86,3.14 | |
um you know,570.839,3.901 | |
regardless of culture regardless of,572.0,5.14 | |
other economic conditions as education,574.74,6.0 | |
rates go up as uh uh women in the,577.14,5.639 | |
workforce goes up fertility rates goes,580.74,4.14 | |
down this is a global thing with no,582.779,5.041 | |
exceptions right so if you extrapolate,584.88,5.1 | |
that out then you can probably make a,587.82,5.1 | |
relatively safe assumption that as AI,589.98,5.58 | |
spreads around the world and economics,592.92,4.44 | |
and education and everything goes up,595.56,4.14 | |
that fertility rates will continue to go,597.36,5.7 | |
down around the whole world South Korea,599.7,4.92 | |
I believe has the lowest fertility rate,603.06,4.74 | |
on the planet at 0.8 births per woman,604.62,5.52 | |
which is uh like,607.8,4.86 | |
um just uh just above a third of the,610.14,5.34 | |
replacement rate so it's entirely,612.66,6.299 | |
possible that under these trends that a,615.48,5.46 | |
population collapse is actually the most,618.959,4.921 | |
real danger that we face so well what do,620.94,5.1 | |
you do about that one thing that I think,623.88,5.34 | |
is going to happen is that AI will lead,626.04,6.479 | |
to Medical breakthroughs and I suspect,629.22,6.36 | |
that we are close if not already at uh,632.519,5.041 | |
the the place of what's called Longevity,635.58,4.86 | |
escape velocity which is that the,637.56,4.38 | |
medical breakthroughs that happen every,640.44,3.54 | |
year extend your life by more than a,641.94,4.62 | |
year so basically,643.98,4.56 | |
hypothetically if you're healthy enough,646.56,4.56 | |
today if you're not about to die and you,648.54,5.1 | |
have access to decent enough uh health,651.12,4.68 | |
care then that the compounding returns,653.64,4.56 | |
of medical research and AI means that,655.8,3.96 | |
you and I could live to be several,658.2,3.66 | |
centuries old which means that the,659.76,3.9 | |
population of the planet will stabilize,661.86,5.159 | |
as birth rates continue to decline now,663.66,5.28 | |
whether I I do think that some people,667.019,4.56 | |
will ultimately choose like AI,668.94,4.74 | |
companions as they become more realistic,671.579,3.721 | |
certainly a lot of people have seen,673.68,3.36 | |
shows like Westworld,675.3,2.46 | |
um you know one of my favorite,677.04,2.28 | |
characters of all time is data from Star,677.76,3.24 | |
Trek and I would love to have data as a,679.32,3.12 | |
friend right,681.0,4.82 | |
um so I absolutely suspect that um that,682.44,6.0 | |
anthropomorphic machines will be part of,685.82,4.66 | |
our Lives before too long,688.44,3.78 | |
um whether what form they take you know,690.48,3.299 | |
whether it's a robotic dog that never,692.22,4.38 | |
dies or you know a walking talking,693.779,5.161 | |
friend that is always there to hang out,696.6,4.739 | |
or if it's a romantic partner like uh,698.94,3.839 | |
you know in the movie Her,701.339,3.481 | |
um with Joaquin Phoenix and Scarlett,702.779,3.781 | |
Johansson there's lots of possibilities,704.82,4.32 | |
uh for how life is going to be but like,706.56,4.26 | |
I said I think one of the most reliable,709.14,4.02 | |
durable Trends is fertility rates go,710.82,4.5 | |
down so the question is will that be,713.16,3.9 | |
offset by longevity,715.32,4.44 | |
so in other words rather than sort of,717.06,5.16 | |
the the dangerous Skynet that some,719.76,5.16 | |
people Envision we might just get out,722.22,5.16 | |
competed sexually uh into Extinction,724.92,4.8 | |
something along those lines yeah that's,727.38,4.26 | |
that's entirely possible especially when,729.72,3.66 | |
you consider that um actually there was,731.64,4.199 | |
a line from Terminator two it was when,733.38,4.98 | |
Sarah Connor was watching uh you know,735.839,3.901 | |
the Terminator Arnold Schwarzenegger,738.36,4.2 | |
play with John and she realized that the,739.74,5.279 | |
machine has infinite patience and will,742.56,4.5 | |
always be there because John was his,745.019,4.26 | |
mission and I realized that from a,747.06,3.899 | |
philosophical standpoint one reading of,749.279,4.5 | |
that is that the machine could be a,750.959,4.261 | |
better parent than a human parent could,753.779,4.021 | |
ever be because for a child from a,755.22,4.619 | |
child's perspective they should be their,757.8,3.84 | |
Prime their parents primary Mission but,759.839,3.721 | |
that's never the case right parents are,761.64,3.78 | |
humans too and they have their own needs,763.56,4.26 | |
their own desires their own plans but,765.42,5.219 | |
when you have a machine that it's if it,767.82,4.68 | |
is designed that you are its primary,770.639,3.781 | |
Mission whether you're an adult or a,772.5,4.86 | |
child like that could be like,774.42,5.599 | |
from some perspectives a better outcome,777.36,4.5 | |
obviously some people are probably,780.019,3.701 | |
cringing which is understandable that's,781.86,3.659 | |
a perfectly healthy reaction to the idea,783.72,3.9 | |
of replacing children and parents with,785.519,4.081 | |
machines but it's possible Right,787.62,5.06 | |
hypothetically possible,789.6,3.08 | |