Коллекции загружаются
#методика_методов
Читаю интересное интервью Юдковского. http://blogs.scientificamerican.com/cross-check/ai-visionary-eliezer-yudkowsky-on-the-singularity-bayesian-brains-and-closet-goblins/ Копипащу сюда няшные цитаты: "Literal immortality seems hard. Living significantly longer than a few trillion years requires us to be wrong about the expected fate of the expanding universe. Living longer than, say, a googolplex years, requires us to be wrong about the basic character of physical law, not just the details. Even if some of the wilder speculations are true and it's possible for our universe to spawn baby universes, that doesn't get us literal immortality. To live significantly past a googolplex years without repeating yourself, you need computing structures containing more than a googol elements, and those won't fit inside a single Hubble volume. And a googolplex is hardly infinity. To paraphrase Martin Gardner, Graham's Number is still relatively small because most finite numbers are very much larger. Look up the fast-growing hierarchy if you really want to have your mind blown, well, eternity is longer than that. Only weird and frankly terrifying anthropic theories would let you live long enough to gaze, perhaps knowingly and perhaps not, upon the halting of the longest-running halting Turing machine with 100 states. But I'm not sure that living to look upon the 100th Busy Beaver Number feels to me like it matters very much on a deep emotional level. I have some imaginative sympathy with myself a subjective century from now. That me will be in a position to sympathize with their future self a subjective century later. And maybe somewhere down the line is someone who faces the prospect of their future self not existing at all, and they might be very sad about that; but I'm not sure I can imagine who that person will be. "I want to live one more day. Tomorrow I'll still want to live one more day. Therefore I want to live forever, proof by induction on the positive integers." Even my desire for merely physical-universe-containable longevity is an abstract want by induction; it's not that I can actually imagine myself a trillion years later." "There is a misapprehension, I think, of the nature of rationality, which is to think that it's rational to believe "there are no closet goblins" because belief in closet goblins is foolish, immature, outdated, the sort of thing that stupid people believe. The true principle is that you go in your closet and look. So that in possible universes where there are closet goblins, you end up believing in closet goblins, and in universes with no closet goblins, you end up disbelieving in closet goblins." "Human axons transmit information at around a millionth of the speed of light, even when it comes to heat dissipation each synaptic operation in the brain consumes around a million times the minimum heat dissipation for an irreversible binary operation at 300 Kelvin, and so on. Why think the brain's software is closer to optimal than the hardware?" 6 июля 2016
|
Хэлен Онлайн
|
|
И ни слова по-русски(
1 |
tesey
|
|
Хоть бы тег какой рисовали, типа "Инглиш". А я бы скрывала... скрывала...
|
"типа "Инглиш"" - у меня одного ассоциация "Лорд Инглиш"?
1 |
tesey
|
|
Матемаг
Можно по-англицки. Без ассоциаций. Ну хоть как-нибудь! |
Хэлен Онлайн
|
|
Матемаг
Кто такой лорд инглиш? 0_о |
Тесей, просто ты с заглавной инглиш написала:)
Хэлл, Хоумстак:) Рид ит! |