少点错误 05月07日 07:37
$500 + $500 Bounty Problem: Does An (Approximately) Deterministic Maximal Redund Always Exist?
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

本文探讨了在两个随机变量中是否存在一个近似的最大冗余Ω的问题。文章首先介绍了精确冗余和近似冗余的概念,并提出了对最大冗余Ω的定义和要求。文章的核心是悬赏500美元,寻找证明在任何两个随机变量X1和X2上都存在满足上述要求的Ω。文章还讨论了误差界限的合理性,以及该问题对解决其他悬赏问题和简化自然潜变量计算的潜在价值。最后,文章简要介绍了精确情况下的证明,并解释了该问题的重要性。

💡精确冗余是指一个随机变量Γ,当且仅当X1→X2→Γ和X2→X1→Γ时,X1和X2给出了关于Γ的完全相同的信息,即P[Γ|X1]=P[Γ|X2]=P[Γ|X]。

🔍近似冗余通过近似上述条件定义,用ϵred来衡量近似程度,即ϵred≥DKL(P[X,Γ]||P[X1]P[X2|X1]P[Γ|X2])和ϵred≥DKL(P[X,Γ]||P[X2]P[X1|X2]P[Γ|X1])。

🏆悬赏目标是证明存在一个近似的最大冗余Ω,它包含(近似地)其他任何冗余包含的关于X的所有信息,并满足X→Ω→Γ,且Ω是X的近似确定性函数,即ϵdet≥H(Ω|X)。

💰悬赏金额为500美元,如果误差界限足够合理,则可以解决另一个悬赏问题,总计1000美元。

💡解决该问题对于解决其他悬赏问题、改进Telephone Theorem以及简化自然潜变量的数值计算具有重要意义。

Published on May 6, 2025 11:05 PM GMT

A lot of our work involves "redunds".[1] A random variable  is a(n exact) redund over two random variables  exactly when both

Conceptually, these two diagrams say that  gives exactly the same information about  as all of , and  gives exactly the same information about  as all of ; whatever information  contains about  is redundantly represented in  and . Unpacking the diagrammatic notation and simplifying a little, the diagrams say  for all  such that .

The exact redundancy conditions are too restrictive to be of much practical relevance, but we are more interested in approximate redunds. Approximate redunds are defined by approximate versions of the same two diagrams:

Unpacking the diagrammatic notation, these two diagrams say

This bounty problem is about the existence of a(n approximate) maximal redund : a redund which contains (approximately) all the information about  contained in any other (approximate) redund. Diagrammatically, a maximal redund  satisfies:

Finally, we'd like our maximal redund  to be an approximately deterministic function of , i.e. , which is equivalent to the diagram

What We Want For The Bounty

This bounty pays out $500 for establishing that there always, over any two random variables , exists an  satisfying the requirements above, i.e.

... with reasonable error bounds.

What counts as reasonable error bounds? We're expecting all the other 's to scale with  in the diagrams above, and ideally be small constant multiples of . In other words: the higher the approximation error on allowable redunds , the more approximation error we allow for all the conditions on . Reasonableness of error bound is partially a judgement call on our part, but small constant (like e.g. 2 or 3) multiples of  would definitely suffice.

Bounds should be global, i.e. apply even when  is large. We're not interested in e.g. first or second order approximations for small  unless they provably apply globally.

If the error bounds are sufficiently reasonable, a solution to this problem will also solve our other open bounty problem, paying out another $500, for a total of $1000.

If people post major intermediate steps which are load-bearing for an eventual solution, we might allocate the bounty based on our judgement of different peoples' relative contributions; the hope is to incentivize sharing intermediate results.

We might also partially pay out for counterexamples. That's a judgement call depending on how thoroughly the counterexample kills hope of any theorem vaguely along the lines intended.

In terms of rigor and allowable assumptions, roughly the level of rigor and assumptions in the posts linked above is fine.

Some Intuition From The Exact Case

A proof in the exact case is straightforward; we'll walk through it here to build some intuition.

The exact redundancy conditions say

wherever . So, we can form a bipartite graph, where each vertex represents a value  of  or a value  of , and two vertices are connected iff . Then the definition of redundancy  says that, for ANY redund  and  are the same for ANY values of  and  (respectively) in the same connected component.

So, let  be defined as the connected component in which  falls. For any redund  must be a function of  (since two -values in the same connected component have the same ). Thus

 is therefore a(n exactly) deterministic function of , and (exactly) mediates between X and any redund . Thus,  satisfies our requirements, in the exact case (i.e. when ).

Alas, this proof does not easily generalize to the approximate case.

Why We Want This

This is a conjecture which we've had on the back burner for a while, and it would be relevant in many places. It would solve our other open bounty problem, it would likely be our starting point for a new version of the Telephone Theorem which more explicitly connects to the natural latents framework, and it would directly simplify the problem of computing natural latents numerically. Probably there are other use-cases which we're not thinking about right at the moment.

 

  1. ^

    A lot of our work involves "redunds".



Discuss

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

冗余 随机变量 最大冗余 信息论 悬赏
相关文章