artificial-intelligence - 人工智能一种现代方法 - 随时间推移的概率推理
问题描述
我目前正在阅读 Peter Norvig 的《人工智能一种现代方法》第 15 章随时间推移的概率推理,但我无法遵循过滤和预测的推导(第 572 页)。
给定过滤到时间 t 的结果,代理需要根据新证据e t+1计算 t + 1 的结果,
P ( X t+1 | e 1:t+1 ) = f( e t+1 , P ( X t | e 1:t )),
对于某些函数 f。这称为递归估计。我们可以将计算视为由两部分组成:首先,当前状态分布从 t 向前投影到 t + 1;然后使用新证据e t+1对其进行更新。当重新排列公式时,这个由两部分组成的过程非常简单:P ( X t+1 | e 1:t+1 ) = P ( X t+1 | e 1:t , e t+1 ) (划分证据)
= α P ( e t+1 | X t+ 1 , e 1:t ) P ( X t+1 | e 1:t ) (使用贝叶斯规则)
使用贝叶斯规则如何导致上一个论坛?难道不应该
α P ( e 1:t , e t+1 | X t+1 ) P ( X t+1 )
解决方案
They are both correct, except that the α is not identical between the two. The trick is that in Norvig's application of Bayes', the conditioning on e1:t is kept constant throughout. Just imagine that it wasn't there to begin with. You would still have all the identities. Then apply that conditioning to every part of the equation, and all the identities will hold.
Alternatively, the derivation can also be done without explicitly using Bayes', but instead, just using the definitions of joint and conditional probability (i.e., P(A|B)P(B)=P(A,B)=P(B|A)P(A)).
P(Xt+1|e1:t, et+1) P(et+1,e1:t) = P(e1:t, et+1|Xt+1) P(Xt+1)
Expanding the joint probabilities on both sides gives
P(Xt+1|e1:t, et+1) P(et+1|e1:t) P(e1:t) = P(et+1|Xt+1,e1:t) P(e1:t|Xt+1) P(Xt+1)
The last two terms on the RHS can be rewritten as
P(Xt+1|e1:t, et+1) P(et+1|e1:t) P(e1:t) = P(et+1|Xt+1,e1:t) P(Xt+1|e1:t) P(e1:t)
Now the P(e1:t) cancels (assuming nonzero), and we have
P(Xt+1|e1:t, et+1) = α P(et+1|Xt+1,e1:t) P(Xt+1|e1:t)
where α = 1 / P(et+1|e1:t).
推荐阅读
- c# - C# 当其他线程停止时,我的主线程停止工作
- performance - Go 1.13 RSS 不断增加,疑似清理问题
- php - 如何找到文件之间的所有链接?
- python - 在 venv 中从默认环境安装包
- azure-devops - 多个 devops 构建相互触发
- json - MongoDB & Google Maps geojson 数据导入错误
- spring - 配置问题:找不到 XML 模式命名空间的 Spring NamespaceHandler [http://www.springframework.org/schema/p]
- android - Unity 中的 Google 登录问题
- swift - SwiftUI:NavigationBarItem 前导不显示
- angular - 解决promise后如何加载html页面?