少点错误 2024年08月14日
A complexity theoretic argument for many worlds
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

本文探讨了多世界假说的弱形式,即波函数的不同分支中存在许多观察者。作者认为,这些分支对人类来说是“实际存在的”,因为有些观察者正在观察它们。作者还讨论了从薛定谔方程推导出离散分支和观察者的困难,并假设存在一个以某种方式指定离散分支和观察者的多世界理论。

😨 **假设一:P != BQP**。这意味着,存在一些经典计算机无法在多项式时间内解决的决策问题,但理想的量子计算机可以在多项式时间内解决。这通常被认为是正确的(RSA 安全依赖于此),但尚未得到证明。这留下了这样的可能性:经典中最难的 BQP 问题只是比多项式时间略微难一些。目前,已知分解一个 b 位整数可以在大约 O(exp(cb1/3)) 时间内完成,其中 c 是大于 1 的常数,而理想的量子计算机可以在多项式时间内完成。作者假设存在 BQP 中的决策问题,其运行时间是“快速增长”的,并且在本文中,作者认为 O(exp(cb1/3)) 是“快速增长”的,尽管它不是真正的指数时间。例如,分解一个十亿位数需要至少 exp(1000) 的时间来完成,这是本文中足以大的数字。

🤔 **假设二:宇宙支持 BQP 计算在多项式物理资源和时钟时间内进行。**这意味着,实际上有可能构建一台量子计算机,并在多项式时钟时间内用多项式物理资源(空间、物质、能量等)解决 BQP 问题。这由目前接受的量子理论所暗示(直到量子计算机可以有多大)。

🤯 **假设三:计算密度人类先验**,将 SIA 与速度先验相结合,是人类目的的观察结果的一个良好先验。作为背景,SIA 代表“自我指示假设”,SSA 代表“自我采样假设”。作者假设读者熟悉这些理论,由 Bostrom 指定。根据 SIA,在其他条件相同的情况下,具有更多观察者的宇宙更有可能。SSA 和 SIA 都接受没有观察者的宇宙永远不会被观察到,但只有 SIA 接受具有更多观察者的宇宙通常更有可能。请注意,SSA 和 SIA 往往会在大型宇宙中收敛(也就是说,在一个拥有许多观察者的巨大宇宙或多元宇宙中,你更有可能观察到宇宙/多元宇宙中拥有更多观察者的部分,因为采样)。速度先验意味着,在其他条件相同的情况下,更容易模拟的宇宙(在某些参考机器上)更有可能。对此的一个粗略论据是,在一个巨大的宇宙中,运行着许多计算,廉价的计算运行得更频繁,从而产生更多观察者。计算密度人类先验将 SIA 与速度先验相结合,并指出我们更有可能观察到观察者时刻与所需计算时间比率高的宇宙。我们可以想象外星人并行模拟许多宇宙,当这些宇宙“停止”时,重复运行计算成本低的宇宙,并从随机选择的观察者时刻中进行选择;他们更有可能选择产生许多观察者相对于所需计算时间的宇宙。作者意识到这一假设是有争议的,但阐明论点可能会使它清楚,是否更弱的假设就足够了。

🧐 **假设四:如果我们的观察结果的概率大致相似(在一个因子 1,000,000,000 内,假设),在经典计算密度先验和量子计算密度先验之间,我们应该更喜欢经典计算密度先验。**如果我们得到压倒性的证据表明量子计算密度先验产生更好的预测,我们最终应该更喜欢它(在元贝叶斯层面上,想象一个关于参考机器的先验),但量子计算比经典计算更复杂,因此在没有压倒性证据的情况下,更喜欢经典计算密度先验。

🚀 **结论:**作者认为,使用经典参考机器来有效地模拟支持量子计算的宇宙中的观察者时刻的唯一方法是模拟许多分支,并从其中许多分支中推导出观察者。量子参考机器不必直接模拟所有分支,但在奥卡姆剃刀的理由上不太可能;将经验物理理论烘焙到速度先验中是尴尬的。作者认为假设 3 是最具争议性的,因此可能值得重新审视。SSA 比 SIA 的主要论据之一是自以为是的哲学家论据:即 SIA 事先极大地偏爱大型宇宙。计算密度先验(将速度先验与 SIA 相结合)没有这个问题,因为...

Published on August 13, 2024 7:35 PM GMT

The following is an argument for a weak form of the many-worlds hypothesis. The weak form I mean is that there are many observers in different branches of the wave function. The other branches "actually exist" for anthropic purposes; some observers are observing them. I've written before about difficulties with deriving discrete branches and observers from the Schrödinger equation; I'm ignoring this difficulty for now, instead assuming the existence of a many-worlds theory that specifies discrete branches and observers somehow.

To be clear, I'm not confident in the conclusion; it rests on some assumptions. In general, physics theories throughout history have not been completely correct. It would not surprise me if a superintelligence would consider many-worlds to be a false theory. Rather, I am drawing implications from currently largely accepted physics and computability theory, and plausible anthropic assumptions.

First assumption: P != BQP. That is, there are some decision problems that cannot be decided in polynomial time by a classical computer but can be decided in polynomial time by an idealized quantum computer. This is generally accepted (RSA security depends on it) but not proven. This leaves open the possibility that the classically hardest BQP problems are only slightly harder than polynomial time. Currently, it is known that factorizing a b-bit integer can be done in roughly time where c is a constant greater than 1, while it can be done in polynomial time on an idealized quantum computer. I want to make an assumption that there are decision problems in BQP whose running time is "fast-growing", and I would consider "fast-growing" in this context despite not being truly exponential time. For example, a billion-bit number would require at least time to factorize with known classical methods, which is a sufficiently huge number for the purposes of this post.

Second assumption: The universe supports BQP computation in polynomial physical resources and clock time. That is, it's actually possible to build a quantum computer and solve BQP problems in polynomial clock time with polynomial physical resources (space, matter, energy, etc). This is implied by currently accepted quantum theories (up to a reasonably high limit of how big a quantum computer can be).

Third assumption: A "computational density anthropic prior", combining SIA with a speed prior, is a good prior over observations for anthropic purposes. As background, SIA stands for "self-indicating assumption" and SSA stands for "self-sampling assumption"; I'll assume familiarity with these theories, specified by Bostrom. According to SIA, all else being equal, universes that have more observers are more likely. Both SSA and SIA accept that universes with no observers are never observed, but only SIA accepts that universes with more observers are in general more likely. Note that SSA and SIA tend to converge in large universes (that is, in a big universe or multiverse with many observers, you're more likely to observe parts of the universe/multiverse with more observers, because of sampling). The speed prior implies that, all else being equal, universes that are more efficient to simulate (on some reference machine) are more likely. A rough argument for this is that in a big universe, many computations are run, and cheap computations are run more often, generating more observers. The computational density anthropic prior combines SIA with a speed prior, and says that we are proportionally more likely to observe universes that have a high ratio of observer-moments to required computation time. We could imagine aliens simulating many universes in parallel, re-running computationally inexpensive universes repeatedly when they "halt", and selecting out observer-moments uniformly at random; they're more likely to select universes that produce many observers relative to required computation time. I realize this assumption is contentious, but spelling out the argument might make it clear whether weaker assumptions would suffice.

The speed prior (and therefore the computational density prior) leaves open the question of what the reference machine is. While any polynomial time random-access-machine computation can be done on a Turing machine in polynomial time, the polynomials may be different, and this matters to the speed prior. For now I'll ignore differences between different polynomials, because the argument is about polynomial vs. non-polynomial time.

There is also the question of whether the reference machine is classical or quantum. A priori, classical computation is simpler, and more likely on Occamian grounds. Classical computation seems likely to be attended to by intelligent observers across a wide range of universes with different laws of physics, while we pay attention to quantum computation mainly for empirical reasons that depend on the laws of physics of our universe. It seems to me that baking quantum computation into the speed prior is awkward, because a prior is supposed to not depend on empirical observations.

This leads to assumption 4: If the probability of our observations are roughly similar (within a factor of 1,000,000,000, let's say) between a classical computational density prior and a quantum computational density prior, we should prefer the classical computational density prior. If we receive overwhelming evidence that the quantum computational density prior produces better predictions, we should eventually prefer it (at the meta-Bayesian level, imagining a prior over reference machines), but quantum computation is more complex to specify than classical computation, so in the absence of overwhelming evidence, the classical computational density prior is preferred.

That's enough assumptions for now. We can now consider a 2x2 table of hypotheses, each of which predict quantum computation; we only consider these due to assumption 2. Either the reference machine is classical or quantum. And either only one branch of the wave function contains anthropic observers (roughly Copenhagen), or many do in proportion to the number of branches (roughly many worlds). (I realize Copenhagen and many worlds have more details than this, I'm ignoring those other details for simplicity). Let's consider the probability of seeing roughly what we see under these different hypotheses. As a simplifying assumption, let's assume 10^15 observers in each branch of the wave function (so, 10^15 observers total in Copenhagen, and that multiplied by the number of branches in many worlds).

First, classical reference machine and Copenhagen. We get 10^15 observers, and the time to compute them is super-polynomial in the number of observers, by assumption 1. Computational density implies this is unlikely, because there are few observers per computation step.

Second, classical reference machine and many worlds. The number of observers is 10^15 times the number of branches. The time to compute this is also roughly proportional to the number of observers times the number of branches. The computational density of observers is reasonably high, so computational density implies this is reasonably likely.

Third, quantum reference machine and Copenhagen. We get 10^15 observers, and the time to compute them is polynomial in the number of observers. The computational density of observers is reasonably high, so computational density implies this is reasonably likely.

Fourth, quantum reference machine and many worlds. The number of observers is 10^15 times the number of branches. Since we're computing all the branches anyway, the quantum reference machine doesn't make a difference. So the logic is the same as with a classical reference machine and many worlds. The computational density of observers is reasonably high, so computational density implies this is reasonably likely.

So far, we have a decisive argument against classical reference machine and Copenhagen. Simulating a quantum universe classically and only picking out observers from one branch is just a really inefficient way to derive observer-moments from computation. Now we leverage assumption 4: if a quantum reference machine isn't very helpful for simulating the sort of observer-moments we see (that is, ones with experiences implying quantum computation is possible), we should prefer a classical reference machine. This implies that a classical reference machine and many worlds is the preferred hypothesis.

Intuitively, quantum reference machine and many worlds is quite implausible: the quantum reference machine is not helpful for simulating the many branches, so there is no reason to prefer a quantum reference machine in the speed prior, as it is more complex to specify. Quantum reference machine and Copenhagen is intuitively more plausible, since the quantum machine is at least being used. If there were strong reasons to think the universe contained few observers, that could be a reason to prefer quantum reference machine and Copenhagen over classical reference machine and many worlds. But in the absence of such reasons, classical reference machine and many worlds is preferred.

This concludes the main argument. To summarize, the only way to use a classical reference machine to efficiently simulate observer-moments in a universe supporting quantum computation is to simulate many branches and derive observers from many of them. A quantum reference machine doesn't have to directly simulate all the branches, but is less plausible on Occamian grounds; it's awkward to bake empirical physics theories into the speed prior.

I assume assumption 3 is the most contentious, so it might be worth re-visiting. One of the main arguments for SSA over SIA is the presumptuous philosopher argument: that is, SIA overwhelmingly prefers large universes a priori. The computational density prior (combining a speed prior with SIA) does not have this problem, because larger universes require more computation to simulate. Combining a speed prior with SSA seems to overly penalize large universes: they require more computation, but do not become more likely on account of containing more observers.

I am intuitively drawn to computational density in part because it is scale-invariant. It doesn't particularly care if there are one or many parallel copies of the same universe; replicating the universe generates proportionally more observers and costs proportionally more computation. I am not particularly motivated to try to make SSA speed priors work for this reason. However, I would be interested in the views of those who think SSA can be reasonably combined with a speed prior.



Discuss

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

多世界假说 量子计算 计算密度 人类先验
相关文章