add

All Articles

Transhumanism and concepts for Life Extension






The adult human brain is pretty much hands down the most sophisticated machine. We know of and there is nothing particularly natural about it, unless your definition of artificial excludes the intentional investment of massive amounts of time and resources that was expended making it possible for you.
The human brain is a pretty serious piece of hardware, but it’s the knowledge, all the learning, or basically software, that goes into the makeup of your average educated adult that makes it truly impressive and also artificial. If the basic brain itself is natural, the final form it takes is no more natural than a chunk of marble someone has chiseled into a statue, and frankly a good deal less so, because our minds are the byproduct of decades of careful work to produce highly sophisticated thinking engines.  This article will focus at some fascinating concepts for, basically, improving on that, making people healthier or smarter or longer lived or just plain safer and happier. Unfortunately it’s also a topic that’s gotten a bad reputation, somewhat unfairly in our opinion, because transhumanism has accumulated a fair amount of rubbish around itself in the last decade as it’s spawned a lot of openly political movements and frequently groups that would be most accurately described as cults.
Now that’s nothing new, we see that a lot with science and science fiction, one minute you’ll looking at the scientific and philosophical ramifications of something like Quantum Mechanics and the Many World or Copenhagen Interpretations, the next minute someone is peddling some home brew form of Quantum Mysticism. And that’s fine, any scientific concept will tend to accumulate a lot of that as part of the normal process of contemplating those very important philosophical and ethical aspects of the concept, and the core science is not diminished by this. Unfortunately in the case of Transhumanism this clutter has gotten to be much louder than the actual scientific concepts and general principles so we think it scares some folks off and truth are told we sometimes feel the same even though we can be classified ourselves as a Transhumanist for around two decades now. The core concept of Transhumanism is using technology to improve mental and physical health and the length of the human lifespan too, preferably indefinitely.
This is hardly a new concept; people have been using any number of herbs and rituals to attempt, sometimes successfully, similar things for untold centuries. We’ve been sticking artificial things in our bodies for a long time too, dental fillings have been found in human skulls 10,000 years old and humans have basically been on the route to being cyborgs since we started putting clothes on. We’ve been doing genetic engineering of people probably longer and our crops and lives to care very definitely not the byproduct of natural evolution. Just because a lot of the new ideas involve microchips inserted into people or direct tinkering with DNA doesn’t really change that beyond making it a lot more effective. There’s very little natural about you or I, my dear ladies and gentleman. Mankind isn’t simply a maker of artificial technologies who is now considering maybe making some of those artificial changes to ourselves, we are now, and pretty much always shave been, the most blatant and shining example of our own tinkering with nature. So while in most folk’s minds there is some sort of distinct line where we cease being natural by putting machines in ourselves, it’s important to understand those are mostly arbitrary.
 When it comes to being natural, that ship sailed long ago. Because fundamentally drinking some herbal concoction to improve your health, or Cleary our mind for better thinking, or slow your aging, is the same for the purpose of intent as cramming some tiny little machines in you to do the same tricks.  There are an awful lot of folks who are alive right now with all sorts of electronic gizmos in their bodies keeping them alive or making their life easier.
       Now that does not mean individual applications of it don’t have their own, but in so far as we are just talking about using artificial means to make people healthier, smarter, or longer lived, we don’t think there’s much firm ground to get any moral footing. We have been doing this, with mixed success, for as long as we have been around as a civilization and just because we are much more scientific and successful with it now does not make it morally or conceptually any different than in the past. Now we have got a lot of concepts to cover and we’ll be skimming through many, and in many ways today we are looking more at concepts then specific technologies.
       we tend to break up Transhumanism’s goals or interests into a number of categories of general technology, categorization is always a bit of an arbitrary thing but here’s our topics for today:1. Speeding up Reflexes & Thinking 2.Slowing down Aging 3.Cloning and Prosthetics 4.Uploading the Mind 5.Artificial Intelligence 6.Technological Singularity. Our first category, speeding up reflexes and thinking, focuses mostly on enhancing the speed at which signals are sent around your body or brain. If you didn’t know, the main component of that is a thing called an axon, and these tend to run throughout your body and mind in little sausage links. They are the phone line or internet cable or information highway of your body. Some are myelinated, some are not, and myelin is a fatty white substance and the reason we call some chunks of brain material white matter. We also call myelinated axons nerve fibers, and where they connect to other cells, usually other neurons, we call these junctions synapses. The wider the diameter of an axon, the faster information can travel down it, and if they don’t have that myelin sheath, or it’s thin, it travels much slower. Generally this diameter is around a micrometer, a millionth of a meter or a micron, but some are wider, up to 20 microns, the diameter of our thinnest hairs, and in the case of the squid giant axon it can be fully a thousand microns wide, or a millimeter. Now that would make for very fast nerve conduction except that the axon is unmyelinated so it’s actually not too quick, quicker than our own unmyelinated axons since it is so wide but still slower than our fastest, myelinated nerves. I wanted to clear that up because people often talk about using squid giant axons in people to speed up our nervous system and besides giving people the shivers it also wouldn’t be effective. What you’d probably want to do is tweak the genes that controlled axon diameter to be a bit wider, or simply transplant one intentionally grown that way, or even just pull out that whole nerve to replace it with say a fiber optic cable. Now what’s advantage of faster nerve conduction? Does it make you lightning quickly? No, but it would make you react a lot faster, this doesn’t make you move in a blur it just means the delay time to send signals goes down. You might go from needing a decent fraction of a second to realize something is in front of your car and send the single to break down to a tiny fraction of that time, and that would save a lot of lives.
It’s the least comic-book-ie superpower but it’s probably the most useful. Speed up nerve conductions and people have way less accidents of every type. Once you adapted to it, which would probably take quite a while, it would be very hard for you, for instance, to trip down a flight of stairs. On top of that some of the most debilitating injuries tend to involve nerve damage, so the ability to get in and replace nerves or re grow them is obviously a high priority of modern medicine. Now a lot of times the implementation of this concept would revolve around basically coating existing nerves with some conductive substance that simply relayed the information faster, closer to the speed of light basically, rather than the speed of sound, which is about a million times slower than light and still decently faster than even most fast nerves send signals. Tiny little robots or some gene-tweaked virus would run around your system basically glazing your existing nerves in your body and brain, or replacing them, so they simply sent everything faster.
Conceptually easy though obviously not easy to implement and probably way over-simplified from anything we’d actually have to do to get that sort of result. Doing this in your brain would also speed up thinking, especially if we could do it in a way that generated less total heat. Our brains run quite hot, and a lot like how modern computers have plateaued out more from the difficulty cooling them then further micro-sizing them. There are some fundamentally physical limits to how little heat you can generate performing a single bit operation, since there’s always some heat produced erasing a bit of data, as covered under Landauer's principle, but it is many orders of magnitude lower than we currently produce doing this on our computers, the one from the factory or resting on your shoulders.
So there should be a lot of room for improvement there. This would, or should, result in basically just speeding up your thinking which doesn’t really make you smarter, it would be more like slowing time down around you. If your brain was a million times faster, it’s not that you are really smarter so much as you are experiencing a year of subjective time for every 30 seconds that passes out side. That would probably drive you insane, since humans normally operate at just a bit slower than the second-scale. Our eyes only operate at about 60 frames a second; we interpret vibrations occurring more than 20 times a second as sound. So unless you had those altered too you’d be staring at freeze frame of your surroundings for what felt like half an hour, and blinking your eyes would leave you blind for hours. Even when you can see its still going to be an eternity of nothing moving. It would be very handy to have days to think about uttering one sentence, plenty of time for coming up with witty comments, but pretty obviously while speeding the mind up a little bit, to the speed you need at that time, would be very handy, thinking that fast would likely be very unpleasant.
Dreaming would be outright disturbing I should think too, as an hour of dreams would translate to just over a century of subjective time. A whole lifetime and then some every time you go to sleep. This is why we often talk in these terms about adding a third lobe to the human brain, essentially an entirely synthetic one that is designed to handle a lot of these extra issues such as being able to feed you external information like books or movies or let you talk ‘telepathically’ at your subjective time to others with some sort of radio link. This isn’t likely an actual lobe but just a series of extra computer bits added in to handle the problems.
We sometimes call this state a SI1, or Super-Intelligence level 1, since it’s the first and most obvious, and lowest level, upgrade to human thinking. Where your brain has simply been sped up a couple orders or more in magnitude and you are still using the basic brain architecture only it’s been modified just in whatever ways are needed to make this practical. Additions to let you bring in other, faster inputs or store and sort memory better. This is also a way to extend lifetime, if you are still living only about a century of real time, but your subjective time is only a modest ten-fold, that amounts to an effective lifetime on par with Methuselah and the other Biblical Patriarchs, if it’s been sped up a million fold that would make for an effective lifetime comparable to having been around since dinosaurs walked the Earth. So it’s probably worth considering now how such prolonged lifetimes, either in real or subjective time, would impact us and that takes us to our second category. Category 2, or slowing down aging, preferably to a complete stop, has been on humanity’s wish list for a long time. It’s controversial for many reasons, some of them legitimate and some not. I dismiss out of hand the notion that nobody would want to live much longer than we do now since they would die of boredom, that’s simply silly. Even if you could get deathly bored, there’s an obvious solution, die.
            I don’t think many religions or life philosophies that let you indefinitely extend your life in the first place aren’t going to find some sort of loophole for suicide at age 1000but even if someone is strongly morally opposed to outright suicide there are plenty of ways to get the job done especially if you’re bored. You update your medical profile to say please do not resuscitate or clone me and take up exciting and dangerous hobbies like cliff diving in a straight jacket while trying to escape the jacket or hunting lions with a nerves bat. You will presumably alleviate your boredom one way or another. I also don’t buy into the notion that we need new blood for new ideas and to avoid stagnation. Besides there being plenty of room in this universe to expand into for new folks, there’s always going to be some deaths. We spend a lot of time on this channel talking about interstellar colonization and terra forming and building space habitats and even outright artificial planets, and we talk a lot about Dyson Spheres, swarm of such artificial habitats able to support in total billions of times as many people as are alive now. In that sort of context in a civilization where the half-life of people, the period of time someone tended to be alive before dying for whatever reason, was a full million years, you’d still have thousands of new people born a year on Earth and trillions inside our solar system. That’s plenty of new blood. But there is a very real flavor of truth to the notion that a person can only live so long, subjectively, before they really do hit a point of diminishing returns where going on would simply be pointless. And when we’re dealing with the very high-end super-intelligences we’ll talk about later in the video that might come even sooner. Some huge super-computer-mind wakes up, rapidly expands its mind to be trillions of times faster and smarter than humans, figures out everything, does its whole mental bucket list, and just shuts off. The apocalypse might be a touch boring if Sky net pops up and ten minutes later just when we’re beginning to panic and realize how screwed we are it just shuts itself off. We also talk a lot on this channel about the Fermi Paradox, the seeming contradiction between the sheer age and size of the Universe and the apparent absence of anyone else in it, and the notion of civilizations dying off from terminal boredom is one we’ll be looking at in the near future.
But some of the other objections to extending life are harder to dismiss. A super-long lived culture is probably a gerontocracy by default. Your senator or parliament member might look like they’re thirty years old but they may have been your senator for thirty centuries, and that’s a lot of seniority. A lot of time for low-risk, long-term investments to make you super-rich too. And both of those are merely specific varieties in which power and influence accrue with time. That’s a lot of time to have kids in and grandkids and great-great-great-great-etc grandkids so that you might easily have millions of direct descendants and you’ve got all that time to accrue knowledge and experience in. Now age generally does bring wisdom, so that might result in a very prosperous and well-operated society especially considering it’s one in which education, social security and pensions, and medical treatment make up only slivers of a nation’s economy.
    But the big concern would be that newer younger folks would tend to feel they were under a serious glass ceiling. If the civilization is still expanding a lot that’s less of a problem but if you’ve got to a point where you’re basically maxed out and just replacing losses a lot of younger folks might feel very frustrated and controlled. If you imagine some civilization, regardless of its total population, that’s only bringing in new people at a rate of maybe 1 per every ten thousand people a year, that kid is probably going to feel smothered by attention from their gazillions older relatives and the oppressive feeling that it will take centuries before they are considered useful. This sense of identity-loss, of not having much of a purpose in life, is a serious concern for everybody else too. Post-Scarcity economies full of long-lived people probably do have to be concerned about a lot of existential problems that make it hard for people emotionally to derive genuine purpose and satisfaction from life. That’s even truer in some of the setups where the humans are essentially pets of super-intelligent machines that benevolent or not simply make them feel useless. I could actually imagine such a creature intentionally behaving hostile but faking weaknesses just so its creators felt they had a purpose in life trying to fight it. Now on the how-to aspect of life extension, transhumanism tends to be understandably vague. The first and obviously most appealing route to most is just to stop people aging normally but there are a lot of other options like mind uploading which we’ll get to later.
Aging, in humans, is really more of group of processes all wearing you down together. There is a thing called SENS, or Strategies for Engineered Negligible Senescence, which looks at aging as basically 7 relatively distinct and combatable things each with their own strategy. It’s a bit controversial in some respects about how accurate this view is, and I’m not a biologist, so I won’t go into as much depth discussing it as I’d like to but I’ve courage you to look it up, and its criticism too. But I generally believe our first opening salvo in a serious war on aging will take some form along these lines and it is important to understand that aging is a pretty vague term that is composed of multiple different phenomena. Winning any battle on these fronts scores a major victory in increasing average lifespan. Now another approach is generally just to replace bits and pieces of people with cloned or prosthetic bits and pieces, and that’s our third category. Cloning and prosthetics are both topics of a lot of controversy, prosthetics less than they used to be, but cloning remains touchy so let me just say from the outset that I’m not familiar with any serious suggestions we do this by growing copies of people to harvest for organs. That is not the goal, that would be an especially monstrous crime too. Whole person cloning is simply growing someone a twin sibling that’s much younger than them anyway.
Prosthetics is nothing new, we’ve got examples 3000 years old and they probably predate that too, but obviously we’re looking at more sophisticated ones, ideally with full sensory and nerve function. I probably don’t need to tell you that progress in this area has been both miraculous in recent years. The thing is neither of these helps much with the brain. Even if you can keep replacing bits and pieces with cloned or cybernetic bits, you can’t clone a brain, so you’d probably have to slowly replace it bit by bit or transferring it entirely into a more electronic setup. That’s category 4, mind uploading, transferring your mind to a computer. And this is our first big problem because you can’t transfer your mind to a computer; you can just copy it to one. Sometimes in science fiction this will be hand waved by requiring a scanning method that vaporizes your brain in the process, usual from ultra-fast serial sectioning with a laser, akin to how some science fiction system deal with teleportation, vaporizing you while assembling a copy of you elsewhere, but this is just that, a hand wave. There’s no real reason you’d need to vaporize a brain to do this which would make it murder. And if you’re not, then you’ve just got yourself sitting in a chair while your digital copy is either on metaphorical ice or is actively running as a new person, quickly diverging from you since it is having new experiences you are not and probably pretty emotionally significant ones. So you are stuck with two people, two who are initially pretty similar but will diverge into two different people. This is the same for cloning yourself in some fashion to a genuine duplicate body, organic or synthetic, with a complete copy of your memories. You still end up with two different people. Now I’m saying people and of course a lot of folks are dubious if that would be a person. I, honestly, don’t see a good rational argument why it wouldn’t be. Trying to prove it is a pretty futile process. We have a notion called the Turing Test that we basically mean is a way to distinguish a computer from a human, you actually do one of these every time you do ones of those irritating Captcha Codes and that’s also why many of them jokingly include a note that says “Prove you’re a human”. Obviously that wouldn’t work with more sophisticated forms of the test but a lot of us feel that if you can’t make a test that every human can pass and a machine can’t, then it looks like a duck, quacks like a duck, and should be accorded the presumption of being a duck.
I mean heck, I don’t know if any of you are real people, nor you I, and we’ll be looking at this concept more in the Simulation Hypothesis Video but the most rational and sane approach is pretty much to assume that if something is making a good case that it is sentient you should probably treat it that way until it can be proven otherwise. Reasonable Doubt and all that, we might think you killed someone but we need to be very sure, beyond a reasonable doubt,  that you did before we’re going to chop your head off for it. I have no idea if I have a soul or free will or if me really exists but I find it easier and more pleasant to assume all of the above and to me its always seemed only fair to extend the benefit of the doubt to anyone or thing which shows decent indicators it might too.WBE, or Whole Brain Emulation, as this is called, usually get’s calculated as requiring around 10^16 to 10^17 hertz of processing power to pull off. Though there’s also versions of this analysis that require a lot more. We did hit that level in the last couple years with our best supercomputers, which are much bigger than a brain, but WBE is still a goodly way off. Still we have basically finally reached the point where we are getting into the human-level of processing power. Which will lead into our next topic of Artificial Intelligence. We will look more at maximums, or rather minimums, of processing power, in terms of how little energy it might take to run a whole human mind in the Simulation Hypothesis video and Black Hole Farmers video but using Landauer's principle at rough body temperature of 300Kelvin and that 4x10^16 Hertz value for WBE you need somewhere around 100 microwatts minimum to run a person real time, or for context you could run a million people off a hundred watt light bulb, and a subjective lifetime of several decades would run you some tens of thousands of joules or the equivalent of about a milliliter of gasoline, a dozen or so drops. That’s the absolute minimum, at room temperature, I doubt getting there, or even near there, is terribly realistic but even getting within a couple orders of magnitude would be pretty impressive. In the context of a full solar englobement, a Dyson Sphere devoted to using the sun’slight as nothing but a computer, often called a Matrioshka brain, which we’ll look at in the Mega structures series shortly, one done all the way out where Earth was would squeeze a decent sized family into a spot the size of your thumb living in a nice virtual world, and permit a total human population, in WBE terms, of around 10^30 people, more than a billion times what we normally project for a Dyson Swarm population of regular people which is itself more than a billion times the current human population. Anyway on to AI, Artificial Intelligence,  our fifth category.
Now there’s not much for me to say here because I don’t really believe in artificial intelligence, or more accurately I think all intelligence is artificial. I’m really not worried about Google waking up to sentience and assuming direct control to become the Harbinger of our Doom. I’m also outright morally opposed to slapping on something like Asimov’s three laws of robotics onto an AI because I’d regard that as slavery. I don’t think much is changed if you just program something to enjoy being told what to do anymore then a plush penthouse isn’t a jail just because the armed guards keeping you in it are courteous about it.
I’ve already mentioned my opinion that if something is acting like a sentient entity you ought to give it the benefit of the doubt, but the thing is I generally take this a bit further and assume they are not just ‘a person’, my loose catch-all for anything about as smart or smarter than humans, alien, computer, whatever, but also basically a human too. Realistically early human level AI’s will likely be heavily copied off human minds anyway, and since the whole point is to make a learning machine, it will also be taught by humans and will probably try to act like us as much as it can for whatever reasons. If it’s a totally logical critter, well it’s pretty logical to be on friendly terms with your creators who you will likely have deduced might have stuck some sort of fails a kill mechanism into you.
A lot of folks involved in Transhumanism in general tend to figure we’d be replaced by AI’s eventually, sooner than later which we’ll discuss in the Singularity section, but I tend to assume that if we can build a computer that can outthink us we can also improve our own brains too, and I would pretty much consider either thing to still be humans anyway. If we’re not using strictly biological definitions, which I don’t think can really apply at this level, then an intelligence made by humans and raised by humans has pretty decent claims to being human. Heck, we tend to regard our pets as human and they are demonstrably not as smart asus. Now our last category, the Technological Singularity, is one we have be kind of vague about so I’ll also be brief. The basic premise is simple enough, technology has being progressing at a fast rate, seemingly an accelerating one, and we’re getting pretty close to being able to make AI’s or implement some of these notions for making people smarter too. If you can design a better brain you’d expect that brain probably can design an even better one and so on.
The singularity reference is pretty much just a reference to mathematical singularities, places where you can’t really predict behavior of systems. Easy version being how you can’t divide by zero, things is not clearly defined. And the notion here is that you’re going to eventually create a series of recursively improving computers that eventually get to be so far beyond humans that they regard as nothing more than ants. There’s nothing human about them anymore, they are simply that powerful. A lot of folks, loosely called Singularitarians, think such an event is just a generation or two away. That’s generally where all agreement ends inside these groups and there’s a lot of counterarguments to how likely this notion is to come about in the near future. I tend to think the basic logic has some flaws and is much further off, but you can examine the arguments yourself and make that call on your own. There are tons of works, fiction and non-fiction, discussing this concept. The point of this video is just familiarizing you with the concepts, we may revisit parts of it in more detail down the road, but I’ll leave off here today. If I had to sum up Transhumanism in a nutshell I’d say it’s basically just an extension of modern attitudes anyway, that humans are imperfect creatures and a civilizations the same, and that’s there’s always rooms for improvement and nothing wrong in and of itself with trying for that. In general it’s a pretty optimistic approach to things, and one I think we all mostly agree on even if the specific paths and degree of caution appropriate in pursuing them is certainly debatable.

Adbox