Distance dependent Chinese Restaurant Processes
Here is a note of Distance dependent Chinese Restaurant Processes
文章链接http://pan.baidu.com/s/1dEk7ZA5
1. Distance dependent CRPs
In the traditional CRP ,the probability of a customer sitting at a table is computed from the number of other customers already sitting at that table.
Now we introduce the distance dependent CRP, the seating plan probability is described in terms of the probability of a customer sitting with each of the other customers .

let
denote the i th customer assignment ,the index of the customer with whom the i th customer is sitting ,let
denote the distance measurement between customers i and j , let D denote the set of all distance measurements between all customers ,and let
be a decay function .
Notice that the customer assignments do not depend on other customer assignment , only the distances between customers.
This distribution is determined by the nature of the distance measurements and the decay function .For many sets of distance measurements ,the resulting distribution over partition is no longer exchangeable ;this is an appropriate distribution to use when exchangeability is not a reasonable assumption.
2.The decay function:
In general the decay function mediates how distances between customers affect the resulting distribution over partitions .Function f is non-increasing , takes non-negative finite values ,and satisfies f(∞)=0。 (衰减函数的性质)
3. Sequential CRPs and the traditional CRP
A sequential CRP is constructed by assuming that dij=∞ for those j>i ,and this guarantees that no customer can be assigned to a later customer.And when f(d)=1 for d≠∞ and dij<∞ for j<i, the sequential CRP is can re-express the traditional CRP.
NOTICE : although these models are the same ,the corresponding Gibbs samplers are different .(why ?)
4. Marginal invariance:
The traditional CRP is marginally invariant : Marginalizing over a particular customer gives the same probability distribution as if that customer were not included in the model at all .But the DDCRP does not have this property ,and this paper gives us two example of the relevant property of DDCRPS.
Language modeling : a fully observed model
Mixture modeling: a mixture model
5. Relationship to dependent Dirichlet processes (DDP):(they are both infinite clustering model that models dependencies between the latent component assignments of the data )
The first difference is that the dependent Dirichlet process mixture use the truncations of the stick-breaking representation for approximate posterior inference ,in CONTRAST, the ddCRP mixtures are amenable to Gibbs sampling algorithms . Another difference is that the spirit behind them ,in the DDP, data are drawn from distributions that are similar to distributions of nearby data,and the particular values of the nearby data impose softer constraints than those in the ddCRP.(区分ddCRP与贝叶斯非参数模型)
Distance dependent Chinese Restaurant Processes的更多相关文章
- URAL 1962 In Chinese Restaurant 数学
In Chinese Restaurant 题目连接: http://acm.hust.edu.cn/vjudge/contest/123332#problem/B Description When ...
- Distance Dependent Infinite Latent Feature Model 阅读笔记1
阅读文献:Distance Dependent Infinite Latent Feature Model 作者:Samuel J.Gershman ,Peter I.Frazier ,and Dav ...
- 中国餐馆过程(Chinese restaurant process)
也就是说假设空桌子有a0个人,然后顾客选择桌子的概率和桌子上人数成正比. 性质: 改变用户的排列方式,桌子的排列方式,概率不变换.
- Marginalize
在David M.Blei 的Distance Dependent Chinese Restaurant Processes 中提到:DDCRP 的一个重要性质,也是和dependent DP 的一个 ...
- 100 Most Popular Machine Learning Video Talks
100 Most Popular Machine Learning Video Talks 26971 views, 1:00:45, Gaussian Process Basics, David ...
- ICLR 2013 International Conference on Learning Representations深度学习论文papers
ICLR 2013 International Conference on Learning Representations May 02 - 04, 2013, Scottsdale, Arizon ...
- 关于LDA的文章
转:http://www.zhizhihu.com/html/y2011/3228.html l Theory n Introduction u Unsupervised learning by ...
- Bayesian machine learning
from: http://www.metacademy.org/roadmaps/rgrosse/bayesian_machine_learning Created by: Roger Grosse( ...
- R Language
向量定义:x1 = c(1,2,3); x2 = c(1:100) 类型显示:mode(x1) 向量长度:length(x2) 向量元素显示:x1[c(1,2,3)] 多维向量:multi-dimen ...
随机推荐
- Python 元类的基本介绍及使用
一.什么是元类 二.如何使用元类 (一)不依赖class关键字创建一个自定义类 (二)自定义元类控制类的产生 (三)自定义元类控制类的调用 一.什么是元类 在python中,一切皆对象,而对象都是由类 ...
- CF1272E. Nearest Opposite Parity 题解 广度优先搜索
题目链接:http://codeforces.com/contest/1272/problem/E 题目大意: 有一个长度为n的数组 \(a\) ,数组坐标从 \(1\) 到 \(n\) . 假设你现 ...
- 惊呆了!Java程序员最常犯的错竟然是这10个
和绝大多数的程序员一样,我也非常的宅.周末最奢侈的享受就是逛一逛技术型网站,比如说 programcreek,这个小网站上有一些非常有意思的主题.比如说:Java 程序员最常犯的错竟然是这 10 个, ...
- Spring中常见的设计模式——代理模式
一.代理模式的应用场景 生活中的中介,黄牛,等一系列帮助甲方做事的行为,都是代理模式的体现.代理模式(Proxy Pattern)是指为题对象提供一种代理,以控制对这个对象的访问.代理对象在客户端和目 ...
- Linux下离线安装gdb及常用命令汇总
以redhat6.5虚拟机作为例子,由于工作性质,大部分情况linux的软件安装,是采用离线方式的. 1.离线安装gdb 像gcc.g++或者gdb这种常用的工具软件,一般虚拟机都会安装的,如未安装, ...
- centos7+docker+elasticsearch 安装记录+踩坑
版本: cenos7 :3.10.0-957.21.3.el7.x86_64 (内核需>=3.10 才可以安装) docker: yum安装版本为1.13.1 elasticsearch: 6 ...
- 有关字符串的算法(KMP,Manacher,BM)陆续补充
KMP算法: 引言: KMP算法是一种改进的字符串匹配算法 字符串匹配:即寻找str_target在str_source中出现的位置 没有改进的字符串匹配:用暴力法进行搜索,枚举出所有的情况然后一一比 ...
- 【Spark 内核】 Spark 内核解析-上
Spark内核泛指Spark的核心运行机制,包括Spark核心组件的运行机制.Spark任务调度机制.Spark内存管理机制.Spark核心功能的运行原理等,熟练掌握Spark内核原理,能够帮助我们更 ...
- 【转】面向GC的Java编程
Java程序员在编码过程中通常不需要考虑内存问题,JVM经过高度优化的GC机制大部分情况下都能够很好地处理堆(Heap)的清理问题.以至于许多Java程序员认为,我只需要关心何时创建对象,而回收对象, ...
- 机器学习——提升方法AdaBoost算法,推导过程
0提升的基本方法 对于分类的问题,给定一个训练样本集,求比较粗糙的分类规则(弱分类器)要比求精确的分类的分类规则(强分类器)容易的多.提升的方法就是从弱分类器算法出发,反复学习,得到一系列弱分类器(又 ...