URI: 
        _______               __                   _______
       |   |   |.---.-..----.|  |--..-----..----. |    |  |.-----..--.--.--..-----.
       |       ||  _  ||  __||    < |  -__||   _| |       ||  -__||  |  |  ||__ --|
       |___|___||___._||____||__|__||_____||__|   |__|____||_____||________||_____|
                                                             on Gopher (inofficial)
  HTML Visit Hacker News on the Web
       
       
       COMMENT PAGE FOR:
  HTML   Introduction to the concept of likelihood and its applications (2018)
       
       
        WaitWaitWha wrote 19 hours 9 min ago:
        The way I read this (and my layman understading) probability is about
        predicting future events, given a known model or assumption, while
        likelihood is almost the mirror image: I observe the data/outcome, and
        I ask: given that data, how plausible is a certain model or parameter
        value?
        
        Another way of looking at it:
        
        probability: parameter/model is fixed; outcome is random/unknown.
        
        likelihood: outcome/data is fixed; we vary the parameter/model to
        assess how well it explains the data.
       
          qjh wrote 16 hours 26 min ago:
          It's actually almost exactly the other way around.
          
          The probability of a model M given data X, or P(M|X) is the posterior
          probability.
          The likelihood of data X given model M, or P(X|M), is the probability
          (or probability density, depending on whether your data is continuous
          or discrete) of observing data X given model M. We often are given
          un-normalised likelihoods, which is what the linked paper talks
          about. These quantities are related via Bayes' Theorem.
          
          Now, you may ask, isn't the probability of observing data X given
          model M still a probability? I mean, yeah, a properly normalized
          likelihood is indeed a probability. It's not the mirror image of
          probability, it is just an un-normalised probability (or a
          probability distribution) of data given a model or model parameters.
       
        pkoird wrote 1 day ago:
        Perhaps it was due to English not being my primary language, but it
        took me an embarrassing amount of time to learn that probability and
        likelihood are different concepts. Concretely, we talk about
        probability of observing a data given an underlying assumption (model)
        is true while we talk about the likelihood of the model being true
        given we observe some data.
       
          wiz21c wrote 1 day ago:
          But the article makes it crystal clear (I had never seen it explained
          so clearly!):
          
          "For conditional probability, the hypothesis is treated as a given,
          and the data are free to vary. For likelihood, the data are treated
          as a given, and the hypothesis varies."
       
          voidhorse wrote 1 day ago:
          Yeah, it was a poor choice of nomenclature, since, in common,
          nontechnical parlance, "probable" and "likely" are very close
          semantically. Though I'm not sure which came first, the choice of
          "likelihood" for the mathematical concept or the casual use of
          "likely" as more or less synonymous with probable.
       
            nerdponx wrote 21 hours 21 min ago:
            My guess was always that "probability" came first, and they needed
            a different word for "likelihood" when the latter concept became
            formalized.
       
          qwertytyyuu wrote 1 day ago:
          Nah, that’s not a non native English thing, i think non maths
          background native people would make the same mistake
       
            MiscCompFacts wrote 22 hours 30 min ago:
            I’m native speaker and I thought they were the same. Still unsure
            of the difference. I guess I need to study this.
       
              nerdponx wrote 21 hours 20 min ago:
              The likelihood function returns a probability. Specifically it
              tells you, for some parametric model, how the joint probability
              of the data in your data set varies as a function of changing the
              parameters in the model.
              
              If that sentence doesn't make sense, then it's helpful to just
              write out the likelihood function. You will notice that that it
              is in fact just the joint probability density of your model.
              
              The only thing that makes it a "likelihood function" is that you
              fix the data and vary the parameters, whereas normally
              probability is a function of the data.
       
       
   DIR <- back to front page