Links

PRESIDENTIAL COLUMN: Bayes for Beginners: Probability and Likelihood

C. Randy Gallistel, August 31, 2015; TAGS: C. RANDY GALLISTEL COLUMNS | DATA | EXPERIMENTAL PSYCHOLOGY | METHODOLOGY | STATISTICAL ANALYSIS

Distinguishing Likelihood(可能性) From Probability(概率)

The distinction between probability and likelihood is fundamentally important:

  • Probability attaches to possible outcomes(可能的产出);

    the possible outcomes to which probabilities attach are MECE(Mutually Exclusive and Collectively Exhaustive);
  • Likelihood attaches to hypotheses(假设).

    the hypotheses to which likelihoods attach are often neither(MECE);

Explaining this distinction is the purpose of this first column.

Possible results are MECE

Possible outcomes are MECE(Mutually Exclusive and Collectively Exhaustive).

Suppose we ask a subject to predict the outcome of each of 10 tosses of a coin.

There are only 11 possible outcomes(0 to 10 correct predictions).

The actual result will always be one and only one of the possible outcomes.

Thus, the probabilities that attach to the possible outcomes MUST sum to 1.

Hypotheses are often neither(MECE)

Hypotheses, unlike outcomes, are neither mutually exclusive nor collectively exhaustive.

Suppose that the first subject we test predicts 7 of the 10 outcomes correctly.

  • I might hypothesize that the subject just guessed,
  • and you might hypothesize that the subject may be somewhat clairvoyant,

    by which you mean that the subject may be expected to correctly predict the results,

    at slightly greater than chance rates over the long run.

    These are different hypotheses, but they are not mutually exclusive,

    because you hedged when you said "may be.",

    You thereby allowed your hypothesis include mine.
  • In technical terminology, my hypothesis is nested within yours.

    Someone else might hypothesize that the subject is strongly clairvoyant and that the observed result underestimates the probability that her next prediction will be correct.

    Another person could hypothesize something else altogether.

    There is no limit to the hypotheses one might entertain.

The set of hypotheses to which we attach likelihoods is limited by our capacity to dream them up. In practice, we can rarely be confident that we have imagined all the possible hypotheses. Our concern is to estimate the extent to which the experimental results affect the relative likelihood of the hypotheses we and others currently entertain. Because we generally do not entertain the full set of alternative hypotheses and because some are nested within others, the likelihoods that we attach to our hypotheses do not have any meaning in and of themselves; only the relative likelihoods — that is, the ratios of two likelihoods — have meaning.

"Forwards" and "Backwards"

The difference between probability and likelihood becomes clear,

when one uses the probability distribution function in general-purpose programming languages.

  • In the present case, the function we want is the \(binomial\ distribution\ function\).

    It is called \(BINOM.DIST\) in the most common spreadsheet software and \(binopdf\) in the language I use. It has three input arguments: the number of successes, the number of tries, and the probability of a success. When one uses it to compute probabilities, one assumes that the latter two arguments (number of tries and the probability of success) are given. They are the parameters of the distribution. One varies the first argument (the different possible numbers of successes) in order to find the probabilities that attach to those different possible results (top panel of Figure 1). Regardless of the given parameter values, the probabilities always sum to 1.
  • By contrast, in computing a likelihood function, one is given the number of successes (7 in our example) and the number of tries (10). In other words, the given results are now treated as parameters of the function one is using. Instead of varying the possible results, one varies the probability of success (the third argument, not the first argument) in order to get the binomial likelihood function (bottom panel of Figure 1). One is running the function backwards, so to speak, which is why likelihood is sometimes called reverse probability.

The information that the binomial likelihood function conveys is extremely intuitive. It says that given that we have observed 7 successes in 10 tries, the probability parameter of the binomial distribution from which we are drawing (the distribution of successful predictions from this subject) is very unlikely to be 0.1; it is much more likely to be 0.7, but a value of 0.5 is by no means unlikely. The ratio of the likelihood at p = .7, which is .27, to the likelihood at p = .5, which is .12, is only 2.28. In other words, given these experimental results (7 successes in 10 tries), the hypothesis that the subject's long-term success rate is 0.7 is only a little more than twice as likely as the hypothesis that the subject's long-term success rate is 0.5.

In summary, the likelihood function is a Bayesian basic.

To understand likelihood, you must be clear about the differences between probability and likelihood:

Probabilities attach to results; likelihoods attach to hypotheses.

In data analysis, the "hypotheses" are most often a possible value or a range of possible values for the mean of a distribution, as in our example.

  • The results to which probabilities attach are MECE(Mutually Exclusive and Collectively Exhaustive);
  • the hypotheses to which likelihoods attach are often neither;

    the range in one hypothesis may include the point in another, as in our example.

    To decide which of two hypotheses is more likely given an experimental result,

    we consider the ratios of their likelihoods. This ratio, the relative likelihood ratio, is called the "Bayes Factor".

SciTech-Mathmatics-Probability+Statistics: Distinguishing(区分) Probability(attaches to Outcomes MECE) from Likelihood(attaches Hypothesis are often neither MECE)的更多相关文章

  1. Probability&Statistics 概率论与数理统计(1)

    基本概念 样本空间: 随机试验E的所有可能结果组成的集合, 为E的样本空间, 记为S 随机事件: E的样本空间S的子集为E的随机事件, 简称事件, 由一个样本点组成的单点集, 称为基本事件 对立事件/ ...

  2. Study note for Continuous Probability Distributions

    Basics of Probability Probability density function (pdf). Let X be a continuous random variable. The ...

  3. An Introduction to Measure Theory and Probability

    目录 Chapter 1 Measure spaces Chapter 2 Integration Chapter 3 Spaces of integrable functions Chapter 4 ...

  4. Probability Concepts

    Probability Concepts Unconditional probability and Conditional Probability Unconditional Probability ...

  5. 【概率论】1-1:概率定义(Definition of Probability)

    title: [概率论]1-1:概率定义(Definition of Probability) categories: Mathematic Probability keywords: Sample ...

  6. Normal Probability Plots|outlier

    6.4 Assessing Normality; Normal Probability Plots The normal probability plot is a graphical techniq ...

  7. [Math Review] Statistics Basics: Main Concepts in Hypothesis Testing

    Case Study The case study Physicians' Reactions sought to determine whether physicians spend less ti ...

  8. PHP7函数大全(4553个函数)

    转载来自: http://www.infocool.net/kb/PHP/201607/168683.html a 函数 说明 abs 绝对值 acos 反余弦 acosh 反双曲余弦 addcsla ...

  9. How do I learn machine learning?

    https://www.quora.com/How-do-I-learn-machine-learning-1?redirected_qid=6578644   How Can I Learn X? ...

  10. [C2P3] Andrew Ng - Machine Learning

    ##Advice for Applying Machine Learning Applying machine learning in practice is not always straightf ...

随机推荐

  1. 【MOOC】华中科技大学计算机组成原理慕课答案-第八章-输入输出系统

    应一个同学后台留言,说让我发就发全,那我干脆把剩的这一章测验也发了吧. 以下解析由GPT生成,不保证可读可解释,仅保证答案正确. 单选 1 (单选(2分)) ‎某中断系统中,每抽取一个输入数据就要中断 ...

  2. 【工具】Zotero|使用Zotero向Word/WPS中插入引用文献(2023年)

    版本:Word 2021,Zotero 6.0.30 版本:WPS 教育版,HUST 前言:两年前我找网上插入文献的方式,网上的博客提示让我去官网下个插件然后才能装,非常麻烦,导致我对Zotero都产 ...

  3. 【记录】PC端QQ空间爬虫常用接口API(随缘更新ing)

    首先是官方的文档:https://wiki.connect.qq.com/api列表. 其次是自己f12找的,每次都要找挺麻烦的不如记下来.记录日期:2022/08/16,如之后功能有更改或API更换 ...

  4. kubernetes理论

    kubernetes特性:自动装箱,自我修复,自动水平扩展,自动服务发现,服务自动负载均衡,自动发布和回滚 秘钥和配置管理,存储编排,批量处理执行 kubernetes架构:master/node m ...

  5. 奇技淫巧之如何在JavaScript中使用Python代码

    @charset "UTF-8"; .markdown-body { line-height: 1.75; font-weight: 400; font-size: 15px; o ...

  6. JAVA 用时间生成主键ID

    /** * 生成主键(19位数字) * 主键生成方式,年月日时分秒毫秒的时间戳 例如:1810311557430000845 */ public static synchronized Long ge ...

  7. 阿里云部署Django主要注意事项

    (1)virtualenv 报错 os 没有 PathLike属性 阿里云ubuntu16.0服务器默认python版本分别是2.7.12,3.5.2,而PathLike是在python 3.6时才被 ...

  8. 开发AR导航助手:ARKit+Unity+Mapbox全流程实战教程

    引言 在增强现实技术飞速发展的今天,AR导航应用正逐步改变人们的出行方式.本文将手把手教你使用Unity+ARKit+Mapbox开发跨平台AR导航助手,实现从虚拟路径叠加到空间感知的完整技术闭环.通 ...

  9. JVM 使用jstat分析系统的垃圾回收情况

    jstat -gcutil 输出结果分析_助你了解jvm命令,查找JVM堆栈信息,分析性能问题.下面介绍一下jstat命令: jstat:虚拟机统计信息监视工具(JVM Statistics Moni ...

  10. curl ifconfig.me 查看Linux服务器公网IP地址

    命令作用:curl ifconfig.me 用于查看Linux服务器公网IP地址. 使用场景:配置Linux服务器IP白名单的时候,需要知道服务器的IP地址,这就到此命令大显身手的时候了.