....But the blood is still the life. Chart the DNA, or Draw the Knife. The blade is made of glass and steel, the Monitor reveals the Code that
Explodes the mythology or mere mortality. Helpless is not Knowing, Know the stain that strains fiction from truth. Henry_Allen
Give Me Immortality Or Give Me Death!
By RU Sirius
May 18th, 2007
Give Me Immortality or Give Me Death!
According to transhumanist Michael Anissimov, there’s an even chance that we’re
looking at immortality or existential destruction in the next 20-40 years.
Anissimov is only 23-years-old but he’s already become an important figure in
the transhumanist movement. While still in high school, he became founder and
director of the Immortality Institute. He’s been active with the World
Transhumanist Association (WTA), and he is currently Fundraising Director, North
America for the Lifeboat Foundation.
Lifeboat Foundation describe themselves as “a nonprofit organization dedicated
to encouraging scientific advancements while helping humanity survive
existential risks and possible misuse of increasingly powerful technologies,
including genetic engineering, nanotechnology, and robotics/AI as we move toward
a technological singularity.”
Anissimov also blogs regularly at Accelerated Future.
I interviewed him for my NeoFiles Show. Jeff Diehl joined me.
RU SIRIUS: Let’s start off talking about immortality. And let’s talk about it
personally. Do you want to live forever?
MICHAEL ANISSIMOV: Oh, absolutely! For sure!
RU: Why?
MA: Because I have at least a thousand years of plans already. And in those
thousand years, I’ll probably make another thousand years of plans, and I don’t
see any end to that cycle.
RU: Do you see the quality of life improving for yourself and for most human
beings?
MA: Yes, I do.
RU: Because I don’t know if I want to live forever under Darwinian conditions.
It gets tiring.
MA: I agree. It does. We need to take control of our own evolution before this
would be a planet really worth living on. I don’t think that thousands of years
of war would be good for anyone. So things do need to improve.
RU: Yeah. Even having to pay… who can afford a thousand years?
MA: (Laughs) Well, you’ll work for a thousand years…
RU: It’s very expensive!
MA: Yeah, people are dying to retire. So it would help out if we had the robots
doing a little bit more of the work.
JEFF DIEHL: So what’s your itinerary for the next thousand years?
MA: I want to go spelunking in every major cave. I want to climb the highest
peak on every continent. I want to write, like maybe at least ten nonfiction
books and ten fiction books. Mmmm…
RU: Some people have done that in a lifetime.
MA: I know!
JD: Yeah, you’re not very ambitious, man — come on!
See also…
» Create an Alien, Win A-Prize
» Why Chicks Don’t Dig The Singularity
» Death? No, Thank You
» Prescription Ecstasy and Other Pipe Dreams
MA: (Laughs) Think of ten possible lives you could live, and then think that you
don’t necessarily need to choose between them. You could live them back to back.
RU: On the other hand, you could pop your consciousness into several bodies and
have them all living simultaneously for only a hundred years. Would that be the
equivalent of living a thousand years?
MA: I don’t think so. I think that would just be like having kids. Copying
yourself would give rise to multiple independent strains of consciousness.
RU: Maybe there could be some kind of central person who could be taking in all
of the experiences.
MA: There could be some information exchange, but…
RU: Aubrey de Grey, of course, is the hacker-biologist who has become very well
known for saying that this is quite plausible in the near future. Is there any
progress that he’s pointed to, or that you can point to, since he really
proclaimed the plausibility of immortality some time around the beginning of
this century?
MA: Yeah. Recently Peter Thiel, former CEO of Paypal, offered three million
dollars in matching funds for projects related to this. And they’ve started
coming up with ways to actually use over a million dollars, I believe. They have
the MitoSENS project and the LysoSENS projects.
RU: What are these projects about?
MA: Well, with LysoSENS — lysomal junk is this stuff that builds up between
cells. And our natural metabolism doesn’t currently have any way of breaking it
down. So researchers are trying to exploit the law of microbial infallibility —
the notion that no matter what organic material you’re talking about, you’re
going to be able to find a microbe that can eat it. So they’re searching for
microbes that are capable of breaking down this junk. And they’ve been looking
in places like… next to a Burger King, because people throw burgers on the
ground and stuff like that. So there are special bacteria there that learn how
to break down these organic compounds. And some of these researchers have even
gotten permission to get soil samples from the people that run graveyards
because that’s where you’d expect to find the bugs. Basically, they’re looking
for specialized microbes that can dissolve that lysomal junk.
RU: IBM recently announced a naotechnology breakthrough. They said that “the
breakthrough marks the first time chips have been made with a self-assembling
nano-technology using the same process that forms seashells or snowflakes.” This
sounds like a really big deal.
MA: Yeah, it is! It’s not the same thing though as molecular manufacturing,
where you basically have a molecular assembly line that places each atom, one by
one. It’s not quite as intelligently controlled or productive, but it is a large
breakthrough.
RU: Yeah, the word jumps out at me — “self-assembling.” That sounds… you’re not
too excited?
MA: (skeptically) Ehhh. I mean, it’s pretty exciting but people have been
playing around with this stuff for a while
RU: OK. Let’s move on to your current work — The Lifeboat Foundation This
foundation is focused on existential risk, which is a board game, I think: Camus
v. Sartre.
MA: (laughs) Not exactly!
RU: I don’t know how you win. It’s probably like Waiting for Godot: A
Tragicomedy in Two Acts — the board game never arrives.
Anyway, in the discourse currently going around among people who are part of the
transhumanist schemata and transhumanist world — there seems to be a turn from
optimism towards a dialogue that’s sort of apocalyptic. And the Lifeboat website
seems to reflect that. Do you think that’s true?
MA: I think it is true to a small extent. I think it’s actually reflective of
the maturing of the transhumanist movement. Because it’s easy to say…
RU: “It’s gonna be great!”
MA: Particularly when the dotcom boom was happening, everyone was, like, “Oh,
the future’s gonna be great. No problems.” You know… “We’re making shitloads of
cash. Everything’s going to go well.”
Now, we’ve had seven years of George Bush. We’ve been involved in two wars. We
understand that reality isn’t always peachy keen and we’re going to have to deal
with the consequences.
RU: So are people in the transhumanist world as worried as they sound, or is it
partly political – trying to be responsible and ease concerns among people who
are perhaps more paranoid than technophiles like yourself?
MA: No, it’s very genuine. The more you understand about powerful technologies,
the more you understand that they really do have the potential to hose us all,
in a way that nuclear war can’t.
RU: Give me your top two existential risks.
MA: Well, as Dr. Alan Goldstein pointed out on your show a couple of weeks ago,
Synthetic Life is a huge risk because life is inherently designed to replicate
in the wild. So life based on different chemical reactions could replicate much
more rapidly than what we’re accustomed to, like some sort of super-fungus. I
think that’s one of the primary risks. And the second risk would be artificial
intelligence — human-surpassing artificial intelligence.
RU: So you’re concerned about the “robot wars” scenario — artificial
intelligence that won’t care that much for us? Do you have any particular
scenarios that you’re following?
MA: Well, I’d like to caution people to be careful what they see in the movies.
Because this is one of those areas where people have been speculating about it
for quite a few decades, and so much fictional material has been built around
it…
RU: Actually, I believe everything in Terminator I.
MA: (Laughs) If you really look through those shows in a critical way, you see
that they’re full of blatant holes all over the place. Like, they can send a guy
through time, but they can’t send his clothes with him through time? (Laughs) In
reality, I think that artificial intelligence is potentially most dangerous
because it might not necessarily need to have a robotic body before it becomes a
threat. An artificial intelligence that’s made purely out of information could
manipulate a wide variety of things on the internet. So it would have more power
than we might guess.
RU: You’ve written a bit about the idea of Friendly AI. (We had Eliezer
Yudkowsky on the show quite a while back, talking about this.) Do you see steps
that can be taken to ensure that A.I. is friendly?
MA: Yeah! I’m totally in support of Eliezer and the Singularity Institute. I
think that they’re one of the few organizations that has a clue. And they’re
growing. I think that you’ve got to put a lot of mathematical eggheads working
together on the problem. You can’t just look at it from an intuitive point of
view. You can actually understand intelligence on a mathematical level. It’s a
lot to ask. I think that friendly A.I. will be a tremendous challenge because
there’s just a lot of complexity in what constitutes a good person. And there’s
a lot of complexity in what constitutes what we consider common sense.
RU: Do you think the breakthrough might come through reverse engineering the
human brain?
MA: It’s possible but probably not.
RU: Good, because I don’t think human beings are that friendly. I think the
friendly A.I. has to be friendlier than human beings.
MA: It definitely does. And one way we could do that is by creating an A.I. that
doesn’t have a self-centered goal system. All creatures built by Darwinian
evolution inherently have a self-centered goal system. I mean, before we became
altruistic, we were extremely selfish. A reptile has eggs, and then the eggs
hatch and he just walks off. He doesn’t care about his kids. So this altruism
thing is relatively recent in the history of evolution, and our psychology is
still fundamentally self-centered.
JD: Isn’t trying to plan for the nature of these future AI’s kind of absurd
because of the exponential superiority of their reasoning… if they even have
what we would call reasoning? Can we really plan for this? It seems like once
you hit a certain threshold, the Singularity, by definition is incomprehensible
to us.
MA: I initially had the same issue. It seems impossible. But ask yourself, if
you could choose, would you rather have an A.I. modeled after Hitler or would
you rather have an A.I. modeled after Mother Teresa?
Regardless of how intelligent the A.I. becomes, it starts off from a distinct
initial state. It starts off from a seed. So whatever it becomes will be the
consequence of that seed making iterative changes on itself.
JD: But maybe in the first nano-second, it completely expunges anything that
resembles human reasoning and logic because that’s just a problem to them that
doesn’t need to be solved any more. And then beyond that — we have no fucking
clue what they’re going to move onto.
MA: It’s true, but whatever it does will be based on the motivations it has.
JD: Maybe. But not if it re-wires itself completely…
MA: But if it rewired itself, then it would do so based on the motivations it
originally had. I mean, I’m not saying it’s going to stay the same, but I’m
saying there is some informational similarity — there’s some continuity. Even
though it could be a low-level continuity, there’s some continuity for an A.I.
Also, you could ask the same question of yourself. What happens if a human being
gains control over its own mind state.
RU: How we understand our motivations might be distinct from how we would
understand our motivations if we had a more advanced intelligence.
MA: That’s true.
RU: I’m going to move on to something that was on the Lifeboat web site that
confounded me. It’s labeled a News Flash. It says, “Robert A. Freitas Jr. has
found preliminary evidence that diamond mechanosynthesis may not be reliable
enough in ambient temperatures to sustain an existential risk from microscopic
ecophagic replicators.”
JD: (Joking) I had a feeling that was the case. (laughter)
RU: What the hell does that mean?!
MA: Robert’s a bit of a wordy guy, but maybe I can explain it. You have an STM
(Scanning Tunneling Microscope.) It’s like a little needle that’s able to scan a
surface by measuring the quantum difference between the two surfaces. Diamond
mechanosynthesis would just be the the ability to have a tiny needle-like
robotic arm that places a single or perhaps two carbon atoms onto a pre-
programmed place. So, in life, we are all based on proteins. Carbon isn’t
slotted in like in a covalent sense, which is the way that people that are
working on nanotechnology are thinking of working. They’re thinking of putting
together pieces of carbon, atom by atom, to make a covalently bonding carbon.
Robert’s saying that it might be that the ambient temperature of the environment
is too hot for that needle to work. So you’d need to have it in a vacuum or
super-cooled environment for it to work.
RU: You did a good job of explaining that. Moving on, there’s some talk on your
site of the idea of relinquishment, which is deciding not to develop
technologies. Is that even possible?
MA: Instead of relinquishment, I like to talk about selective development. You
can’t really relinquish technology too easily. But you can develop safeguards
before technologies reach their maturity. And you can develop regulations that
anticipate future consequences instead of always taking a knee-jerk reaction and
saying: “Oh, this disaster happened; therefore we will now regulate.”
RU: Of course, it’s not really possible to regulate what everybody everywhere on
the planet is doing.
MA: No, it’s not.
RU: Are you familiar with Max More’s Proactionary Principle?
MA: (skeptically) Mmmm I’m…
RU: Too obvious?
MA: No, I don’t fully agree with it. I do think that the Precautionary Principle
has a point.
RU: Maybe I should say what it is. Basically, the Precautionary Principle says
that with any technology we’re developing, we should look ahead and see what the
consequences are. And if the consequences look at all dire, then we should
relinquish the technology. And Max More argues that we should also look at the
possible consequences of not developing the technology. For instance, if we
don’t develop nanotechnology, everybody dies.
MA: Well, I don’t think that would happen.
RU: I mean, eventually… just as they have for millennia.
MA: Oh — everyone will age to death!
RU: Right
MA: No, I agree that the balanced view looks at both sides of the equation. The
Precautionary Principle’s kind of been tarnished because there are people that
are super-paranoid; and people who use it as an excuse to rule out things that
they find ethically objectionable like therapeutic cloning.
RU: Well, you could take anything as an example. Look at automobiles. If we had
looked ahead at automobiles — we could debate for hours whether they were a good
idea. There would probably be less humans on the planet and there would probably
be less distribution of medicine and food and all those things. On the other
hand, we might not be facing global warming. It might be nice that there are
less humans on the planet.
MA: Yeah, but in practice, if some invention is appealing and has large economic
returns, then people are going to develop it no matter what.
RU: On the Lifeboat site, you have a list of existential risks. And people can
sort of mark which existential risk they want to participate in or work on. I’d
like to get your comments on a few of the risks that are listed. But before I go
down a few of these things on the list, what do you think is up with the bees?
MA: The bees?
RU: The honeybees are dying off. Einstein said we wouldn’t survive if…
JD: … there’s some contention about whether he actually said that. I heard that
somebody tried to find that quote, and they weren’t able to find it.
MA: What does this have to do with the bees?
RU: Einstein said that if all the honeybees died off, we’d all be dead in four
years, or something like that.
JD: Yeah, because of the natural cycles that they support. Somebody else
debunked that.
RU: Well, he was no Einstein. You better look into the bees because that could
be an existential risk.
So here’s one of the risks – or the risk aversion possibilities — listed on the
site: Asteroid Shield.
MA: Well, someone once said that we’re in a cosmic shooting gallery and it’s
only a matter of time before we get nailed. I wouldn’t consider this to be a
high priority, but in the interest of comprehensiveness, it would be a good idea
if we had a way to deflect asteroids. Serious scientists have been looking at
this issue and they decided that knocking it out with a nuclear bomb wasn’t
really going to work so well. It’s too expensive and too unpredictable. So
they’re talking about attaching small rockets to slowly pull an asteroid off
course.
JD: I recently read one idea — collect a lot of space junk and create one big
object to alter the gravitational…
MA: Or you can put a little electro-magnetic rail gun on the surface and
progressively fire off chunks of the asteroid, which will also alter its course.
Even if you altered the trajectory of an incoming asteroid by a tiny amount, it
would probably miss because earth is just kind of like a tiny dot in space. But
right now, we don’t have the capability. So if an asteroid were coming next
year, we would be screwed.
RU: Right. And people have started talking about it. I mean, there has been sort
of an advance in the level of paranoia about asteroids that come anywhere near
us in recent years.
MA: One asteroid came about half of the way between us and the moon a while ago.
JD: Was it big enough to kill us?
MA: No. It was a hundred feet across, though — not bad.
RU: So how much chaos would that cause? I guess that would depend on where it
landed.
MA: Measured in megatons, I think it would be about one Hiroshima.
JD: Oh, okay. We can handle that… as long as it doesn’t land in San Francisco.
MA: (laughs) Exactly! So I don’t think the asteroids are an immediate concern.
But it helps people comprehend the notion of extinction risks.
RU: The former NASA astronaut Rusty Schweickart has become involved in fighting
off the asteroids. He used to be part of the L5 Society. I think Ronald Reagan
would say it’s a way of uniting all the people of earth to fight against an
enemy.
MA: Yeah!
RU: I think he talked about that in terms of aliens, not in terms of asteroids.
MA: Well, I think all existential risks, including the more plausible ones, do
serve a function in uniting humanity, and I think that’s a nice side effect.
RU: The particle accelerator shield — what’s that about?
MA: Some people think — as we engage in increasingly high-powered particle
accelerator experiments — something bad could happen. One standard idea is a
strangelet, which is similar to an atom but much more compact. If a strangelet
could absorb conventional matter into itself, and do so continuously, it could
absorb the entire planet.
RU: Sort of like a black hole.
MA: Yes, very much like a black hole. It’s another one of those situations where
we want to instill a sense of caution in the minds of scientists. We don’t want
them to just dismiss these possibilities out of hand because it potentially
threatens their funding. We want them to actually give it a little bit of
thought.
RU: OK, what about “seed preserver.”
MA: Oh, yeah! Well that’s actually being done right now! The Norwegian
government built a seed bank on some far north Arctic island. They’re shoving
some seeds in there, so I guess when the nanobots come, or the nuclear war comes
and 99% of humanity is all gone, then we’ll be able to go there, withdraw the
seeds, and create life anew.
RU: You seem to be a believer in the Singularity. For me – maybe yes, maybe no.
But I find it amusing that Vernor Vinge could give a talk titled “What if the
Singularity Doesn NOT Happen”, the implication being that the idea that it might
not happen is a real stretch. Do you ever feel like you’re in a cult — that
people who believe in this share a peculiar reality?
MA: The word Singularity has become a briefcase word. People kind of want to put
their pet ideas into it, so the actual idea has become kind of unclear in the
minds of many people. To me, the Singularity is just the notion of an
intelligence that’s smarter than us. So if you say that you don’t believe in the
Singularity, it means that you believe that human beings are the smartest
possible intelligence that this universe can hold.
RU: I guess what I don’t believe is that it necessarily becomes a complete
disjunction in history.
MA: But don’t you think that homo sapiens are a quite complete disjunction from,
say, homo erectus or chimps? We share 98% of the same DNA. So what if you
actually used technology to surpass the human mind? I think you’d have something
substantially more different from homo sapiens than homo sapiens was from their
predecessors.
RU: Do you think it’s more likely that we’ll develop machines that are more
intelligent than us and keep them exterior to us; or will we find some way of
incorporating them into us? It seems to me, if you look at the passion that
people have for being on the net, and being able to call up and get and link to
all the information and all the intelligence on the planet, people are going to
want this inside themselves. They’re going to want to be able to have as much
information and as much intelligence as everybody else. They’ll want to unite
with it.
MA: I think that would be a great thing, as long as people don’t go using their
intelligence for negative ends.
JD: Do you think this would happen gradually. Or do you think there would be
this point in time where lots of people make choices like whether or not to
merge? And then, maybe, the people who are afraid of that will want to stop
people from doing it, and conflict…
MA: I think it could actually be somewhat abrupt, because once you have a
superior intelligence, it can create better intelligence enhancement techniques
for itself. So it could be somewhat abrupt. But I think that these smart
entities could also find a way of keeping humanity on the same page and not
making it like: “Oh, you have to choose… If your brother or your sister is not
going into the great computer, then…”
RU: I think if it happens soon enough, it will be viewed as just another way of
going online. You know, to young people, it will be just… “Yeah, this is how
everybody’s going online now.”
MA: But if you had implants in your brain, it would be permanent.
RU: Do you think chaos is built into life? As the Artificial Life people have
been saying, life happens on the boundary between order and chaos. If chaos is
an element of life, can machines include chaos?
MA: Well, uh — hmm. I think that people overestimate the power of chaos.
RU: As a Patti Smith fan, I have to disagree.
MA: (Laughs) Well, it’s such an appealing idea — chaos. But if you take a look
at human blood and compare it to some random bit of muck you find in the ground,
you’ll see that it’s highly regulated, and there are huge complements of
homeostatic mechanisms in bodies that are constantly ordering things. Relative
to the entropy in the air array outside; inside my body is a very orderly place,
Life forms are very well organized pieces of matter.
RU: Right, but if you achieve complete homeostasis, then nothing happens.
MA: That’s true. Life does have to be on that boundary so it is challenging
RU: Here’s a quote from an interview with you: “The idea of the Singularity is
to transcend limitations by reengineering brains or creating new brains from
scratch, brains that think faster with more precision, greater capabilities,
better insights, ability to communicate and so on.” OK. That sounds good, but
what about pleasure, play, creativity, eroticism… and whatever it is you get
from magic mushrooms? Where does all that go?
MA (Laughs) I think all that’s very important. I think about all those things.
RU: So you think that can be built that into the singularity?
MA: Yeah. Oh, for sure…
RU: David Pierce is the one person who really sort of deals with those ideas.
MA: Well, it’s not really too PC to talk about it. But when you take a
psychedelic, you’ve changed your brain chemistry. With mushrooms, you flood your
brain with this one psilocybin chemical. With technologies that let you actively
change your own mind, it would be less of a shot in the dark. More precision
modifications would be possible. And you could turn it on and off like a light
switch, too. You could have much more control over it.
RU: Looking forward to it!
I am an eclectic person with a decidedly different take on just about everything. I am apt to discuss everything from today's politics to astrophysics to ghosts in the machine (yours, mine, ours). My posts are sometimes personal stuff, sometimes special interests, reviews of books I've read or films I've seen or places I've been, sometimes they are biting editorial opinion. Sometimes poetry. Sometimes select reprints. Subject matter? Read and find out. That, even I can't predict.
Friday, May 18, 2007
The Other "Immortality"
Subscribe to:
Posts (Atom)