Talk:Log-normal distribution

From Wikipedia, the free encyclopedia

See also: multiplicative calculus ??

I followed the link to the wikipedia page Multiplicative calculus and spent about an hour trying to determine whether it has any legitimacy. I concluded that it does not. In contrast log normal distributions are unarguably legitimate. The multiplicative calculus link under 'See also' should be removed, for it is a waste of reader's time to refer them to some dubious article with no clear relevance to log normal distributions, save the shared multiplicative basis.

If you look at Talk:Multiplicative calculus you will find lengthy disputes about the legitimacy of the content with most contributors concluding that the article should be deleted. Moreover, all the talk, save two entries, is a decade old. Cross-linking to distracting rubbish serves no one, even if the rubbish manages to cling to some tiny thread of legitimacy that prevents its deletion.  Preceding unsigned comment added by 70.68.18.243 (talk) 19:03, 20 September 2019 (UTC)

Error in lognormal pdf in box?

The pdf given in the box at rhs of page seems wrong and does not match the pdf given in the article. In particular seems like first term of pdf should be 1/x sigma root (2pi). I.e., factor of 1/x seems missing. So term should be: frac{1}{ x\sigma \sqrt{2 \pi}}  ? Hesitant to edit as this is my first foray into wikipedia talk and edits, plus its not obvious to me how to edit the box. --QFC JRB (talk) 19:45, 2 May 2017 (UTC)

I'm also noticing that while μ is listed as 0, the pdf appears centered about 1. Is this an error? Tommy2024 (talk) 22:17, 20 March 2024 (UTC)
This is not an error; the parameter μ is the mean of the logged values. As exp(0) = 1, the median of the PDF is in the correct location. The mean is near, but not at 1. 2601:603:4800:32F:617F:D985:18C7:D9C0 (talk) 22:33, 7 February 2025 (UTC)

Checked again and it was updated as suggested above. --QFC JRB (talk) 19:50, 3 May 2017 (UTC)

Derivation of Log-normal Distribution

How do you derive the log-normal distribution from the normal distrubution?

By letting X ~ N(\mu, \sigma^2) and finding the distribution of Y = exp X.

D. Clason — Preceding unsigned comment added by 128.123.198.136 (talk) 00:21, 11 November 2011 (UTC)

I found this derivation hard to follow. This explanation (prop 8. p. 12 from http://norstad.org/finance/normdist.pdf) was easier to follow where it shows the pdf/variable change. I'd like to change the derivation on this page if I get the time (not quite sure how to work the math font).Corwinjoy (talk) 20:03, 17 February 2017 (UTC)

Old talk

Hello. I have changed the intro from "log-normal distributions" to "log-normal distribution". I do understand the notion that, for each pair of values (mu, sigma), it is a different distribution. However, common parlance is to call all the members of a parametric family by a collective name -- normal distribution, beta distribution, exponential distribution, .... In each case these terms denote a family of distributions. This causes no misunderstandings, and I see no advantage in abandoning that convention. Happy editing, Wile E. Heresiarch 03:42, 8 Apr 2004 (UTC)

In the formula for the maximum likelihood estimate of the logsd, shouldn't it be over n-1, not n?

Unless you see an error in the math, I think its ok. The n-1 term usually comes in when doing unbiased estimators, not maximum likelihood estimators.
You're right; I was confused.

QUESTION: Shouldn't there be a square root at the ML estimation of the standard deviation? User:flonks

Right - I fixed it, thanks. PAR 09:15, 27 September 2005 (UTC)

Could I ask a question?

If Y=a^2; a is a log normal distribution ; then What kind of distribution is Y?

a is a lognormal distribution
so log(a) is a normal distribution
log(a^2) = 2 log(a) is also a normal distribution
a^2 is a lognormal distribution --Buglee 00:47, 9 May 2006 (UTC)

One should say rather that a has---not is---a lognormal distribution. The object called a is a random variable, not a probability distribution. Michael Hardy 01:25, 9 May 2006 (UTC)


Maria 13 Feb 207: I've never written anything in wikipedia, so I apologise if I am doing the wrong thing. I wanted to note that the following may not be clear to the reader: in the formulas, E(X)^2 represents the square of the mean, rather than the second moment. I would suggest one of the following solutions: 1) skip the parentheses around X and represent the mean by EX. Then it is clear that (EX)^2 will be its square. However, one might wonder about EX^2 (which should represent the second moment...) 2) skip the E operator and put a letter there, i.e. let m be the mean and s the standard deviation. Then there will be no confusion. 3) add a line at some point in the text giving the notation: i.e. that by E(X)^2 you mean the square of the first moment, while the second moment is denoted by E(X^2) (I presume). I had to invert the formula myself in order to figure out what it is supposed to mean.

I've just attended to this. Michael Hardy 00:52, 14 February 2007 (UTC)

A mistake?

I think there is a mistake here : the density function should include a term in sigma squared divided by two, and the mean of the log normal variable becomes mu - sigma ^2/2 Basically what happened is that, I think, the author forgot the Ito term.

I believe the article is correct. See for example http://mathworld.wolfram.com/LogNormalDistribution.html for an alternate source of the density function and the mean. They are the same as shown here, but with a different notation. (M in place of mu and S in place of sigma). Encyclops 00:23, 4 February 2006 (UTC)
Either the graph of the density function is wrong, or the expected value formula is wrong. As you can see from the graph, as sigma decreases, the expected value moves towards 1 from below. This is consistent with the mean being exp(mu - sigma^2/2), which is what I recall it as. 69.107.6.4 19:29, 5 April 2007 (UTC)
Here's you're mistake. You cannot see the expected value from the graph at all. It is highly influenced by the fat upper tail, which the graph does not make apparent. See also my comments below. Michael Hardy 20:19, 5 April 2007 (UTC)

I've just computed the integral and I get

So with μ = 0, as σ decreases to 0, the expected value decreases to 1. Thus it would appear that the graph is wrong. Michael Hardy 19:57, 5 April 2007 (UTC)

...and now I've done some graphs by computer, and they agree with what the illustration shows. More later.... Michael Hardy 20:06, 5 April 2007 (UTC)

OK, there's no error. As the mode decreases, the mean increases, because the upper tail gets fatter! So the graphs and the mean and the mode are correct. Michael Hardy 20:15, 5 April 2007 (UTC)

You're right. My mistake. The mean is highly influenced by the upper tail, so the means are actually decreasing to 1 as sigma decreases. It just looks like the means approach from below because the modes do. 71.198.244.61 23:50, 7 April 2007 (UTC)

Question on the example charts on the right. Don't these have μ of 1, not 0 (as listed)? They're listed as 1. If the cdf hits 0.5 at 1 for all of them, shouldn't expected value be 1? —Preceding unsigned comment added by 12.17.237.67 (talk) 18:28, 15 December 2008 (UTC)

The expected value is , not μ. /Pontus (talk) 19:19, 16 December 2008 (UTC)
Yet the caption indicates the underlying µ is held fixed at 0. In which case we should see the expected value growing with sigma. —Preceding unsigned comment added by 140.247.249.76 (talk) 09:13, 29 April 2009 (UTC)
Expected value is not the value y at which P[X<y] = P[X > y]. Rookie mistake. —Preceding unsigned comment added by 140.247.249.76 (talk) 09:24, 29 April 2009 (UTC)

A Typo

There is a typo in the PDF formula, a missing '['

Erf and normal cdf

There are formulas that use Erf and formulas that use the cdf of the normal distribution, IMHO this is confusing, because those functions are related but not identical. Albmont 15:02, 23 August 2006 (UTC)

Technical

Please remember that Wikipedia articles need to be accessible to people like high school studends, or younger, or without any background in math. I consider myself rather knowledgable in math (had it at college level, and still do) but (taking into account English is not my native language) I found the lead to this article pretty difficult. Please make it more accessible.-- Piotr Konieczny aka Prokonsul Piotrus | talk  22:48, 31 August 2006 (UTC)

To expect all Wikipedia math articles to be accessible to high-school students is unreasonable. Some can be accessible only to mathematicians; perhaps more can be accessible to a broad audience of professionals who use mathematics; others to anyone who's had a couple of years of calculus and no more; others to a broader audience still. Anyone who knows what the normal distribution is, what a random variable is, and what logarithms are, will readily understand the first sentence in this article. Can you be specific about what it is you found difficult about it? Michael Hardy 23:28, 31 August 2006 (UTC)

I removed the "too technical" tag. Feel free to reinsert it, but please leave some more details about what specifically you find difficult to understand. Thanks, Lunch 22:18, 22 October 2006 (UTC)

Skewness formual incorrect?

The formula for the skewness appears to be incorrect: the leading exponent term you have is not present in the definitions given by Mathworld and NIST, see http://www.itl.nist.gov/div898/handbook/eda/section3/eda3669.htm and http://mathworld.wolfram.com/LogNormalDistribution.html.

Many thanks.

X log normal, not normal.

I think the definition of X as normal and Y as lognormal in the beginning of the page should be changed. The rest of the page treats X as the log normal variable. —The preceding unsigned comment was added by 213.115.25.62 (talk) 17:40, 2 February 2007 (UTC).

The skewness is fine but the kurtosis is wrong - the last term in the kurtosis is -3 not -6 —Preceding unsigned comment added by 129.31.242.252 (talk) 02:08, 17 February 2009 (UTC)

Yup I picked up that mistake too and have changed is. The wolfram website also has it wrong, although if you calculate it from their central moments you get -3. I've sent them a message too. Cheers Occa —Preceding unsigned comment added by Occawen (talkcontribs) 21:19, 1 December 2009 (UTC)

Partial expectation

I think that there was a mistake in the formula for the partial expectation: the last term should not be there. Here is a proof: http://faculty.london.edu/ruppal/zenSlides/zCH08%20Black-Scholes.slide.doc See Corollary 2 in Appendix A 2.

I did put my earlier correction back in. Of course, I may be wrong (but, right now, I don't see why). If you change this again, please let me know why I was wrong. Thank you.

Alex —The preceding unsigned comment was added by 72.255.36.161 (talk) 19:39, 27 February 2007 (UTC).

Thanks. I see the problem. You have the correct expression for

while what I had there before would be correct if we were trying to find

which is (essentially) the B-S formula but is not the partial mean (or partial expectation) by my (or your) definition. (Actually I did find a few sources where the partial expectation is defined as but this usage seems to be rare. For ex. ). The term that you dropped occurs in but not , the correct form of the partial mean. So I will leave the formula as it is now. Encyclops 00:47, 28 February 2007 (UTC)

  • The rest of the page uses rather than , I suggest that also be used here (in addition to which is a nice way to put it). I didn't add it myself since with 50 percent probability I'd botch a . --Tom3118 (talk) 19:29, 16 June 2009 (UTC)

Generalize distribution of product of lognormal variables

About the distribution of a product of independent log-normal variables:

Wouldn't it be possible to generalize it to variables with different average ( mu NOT the same for every variable)?

the name: log vs exponential

log normal, sometimes, it is a little bit confusing for me, so a little bit note here:

For variable Y, if X=log(Y) is normal, then Y is log normal, which says after being taken log, it becomes normal. Similarly, there might be exponential normal: for variable Z, exp(Z) is normal. However, exp(Z) can never be normal, so the name log normal. Furthermore, if X is normal, then log(X) is undefined.

In other cases, variable X is in whatever distribution (XXX), we need a name for the distribution of Y=log(X) (in the case it is defined). X=exp(Y), Such a name should exponential XXX. For instance, X is in IG, then Y=log(X) is in exponential IG. Jackzhp 15:37, 13 July 2007 (UTC)

Mean, μ {\displaystyle \mu } and σ {\displaystyle \sigma }

The relationship given for in terms of Var(x) and E(x) suggest that is undefined when . However, I see no reason why must be strictly positive. I propose defining the relatinship in terms of such that

I am suspicious that this causes to be...well, wrong. It suggests that two different values for could result in the same , which I find improbable. In any case, if there is a a way to calculate when then we should include it, if not, we need to explain this subtlety. In my humble opinion.--Phays 20:35, 6 August 2007 (UTC)

I'm not fully following your comment. I have now made the notation consistent throughout the article: X is the random variable that's log-normally distributed, so E(X) must of course be positive, and μ = E(Y) = E(log(X)).
I don't know what you mean by "E2". It's as if you're squaring the expectation operator. "E2(X) would mean "E(E(X))", but that would be the same thing as E(X), since E(X) is a constant. Michael Hardy 20:56, 6 August 2007 (UTC)
If 0<x<1 then log(x)<0 and therefore E(log(x))<0, However there is no problem with this, all the formulas work. Only if x<0 one cannot represent the distribution of Y=log(x) since such cases do not allow these values. 85.49.129.87 (talk) 20:28, 18 December 2024 (UTC)

Maximum Likelihood Estimation

Are there mistakes in the MLE? It looks to me as though the provided method is a MLE for the mean and variance, not for the parameters and . If that is so it should be changed to the parameters estimated and and then a redirect to extracting the parameter values from the mean and variance.--Phays 20:40, 6 August 2007 (UTC)

The MLEs given for μ and σ2 are not for the mean and variance of the log-normal distribution, but for the mean and variance of the distribution of the normally distribution logarithm of the log-normally distributed random variable. They are correct MLEs for μ and σ2. The "functional invariance" of MLEs generally, is being relied on here. Michael Hardy 20:47, 6 August 2007 (UTC)
I'm afraid I still don't fully understand, but it is simple to explain my confusion. Are the parameters being estimated μ and σ2 from
or are these estimates describing the mean and variance? In other words, if is and then is ? It is my understand that the parameters in the above equation, namely μ and σ are not the mean and standard deviation of . They may be the mean and standard deviation of .--Phays 01:16, 7 August 2007 (UTC)
The answer to your first question is affirmative. The expected value of Y = exp(X) is not μ; its value if given elsewhere in the article. Michael Hardy 16:10, 10 August 2007 (UTC)

8/10/2007:

It is my understanding that confidence intervals use standard error of a population in the calculation not standard deviation (sigma).

Therefore I do not understand how the Table is using 2sigma e.tc. for confidence interval calulation as pertains to the log normal distribution.

Why is it shown as 2*sigma?

Angusmdmclean 12:35, 10 August 2007 (UTC) angusmdmclean


Hi. The formula relating the density of the log normal to that of the normal -- where does the product come form on the r.h.s. ? I think this is a typo. should read: f_L = {1\over x} \times f_N, no?

This page lacks adequate citations!!

Wikipedia policy (see WP:CITE#HOW) suggests citation of specific pages in specific books or peer-reviewed articles to support claims made in Wikipedia. Surely this applies to mathematics articles just as much as it does to articles about history, TV shows, or anything else?

I say this because I was looking for a formula for the partial expectation of a lognormal variable, and I was delighted to discover that this excellent, comprehensive article offers one. But how am I supposed to know if the formula is correct? I trust the competence of the people who wrote this article, but how can I know whether or not some mischievous high schooler reversed a sign somewhere? I tried verifying the expectation formula by calculating the integral myself, but I got lost quickly (sorry! some users of these articles are less technically adept than the authors!) I will soon go to the library to look for the formula (the unconditional expectation appears in some books I own, but not the partial expectation) but that defeats the purpose of turning to Wikipedia in the first place.

Of course, I am thankful that Wikipedia cites one book specifically on the lognormal distribution (Aitchison and Brown 1957). That reference may help me when I get to the library. But I'm not sure if that was the source of the formula in question. My point is more general, of course. Since Wikipedia is inevitably subject to errors and vandalism, math formulas can never be trusted, unless they follow in a highly transparent way from prior mathematical statements in the same article. Pages like this one would be vastly more useful if specific mathematical statements were backed by page-specific citations of (one or preferably more) books or articles where they could be verified. --Rinconsoleao 15:11, 28 September 2007 (UTC)

Normally I do not do this because I think it is rude, but I really should say {{sofixit}} because you are headed to the library and will be able to add good cites. Even if we had a good source for it, the formula could still be incorrect due to vandalism or transcription errors. Such is the reality of Wikipedia. Can you write a program to test it, perhaps? Acct4 15:23, 28 September 2007 (UTC)
I believe Aitchinson and Brown does have that formula in it, but since I haven't looked at that book in many years I wouldn't swear by it. I will have to check. I derived the formula myself before adding it to Wikipedia, unfortunately there was a slip up in my post which was caught by an anonymous user and corrected. FWIW, at this point I have a near 100% confidence in its correctness. And I am watching this page for vandalism or other problems. In general your point is a good one. Encyclops 22:34, 28 September 2007 (UTC)

Why has nobody mentioned whether the mean and standard deviation are cacultaed from x or y?. if y = exp(x). Then mean and stdev are from the x values. Book by - Athansious Papoulis. Siddhartha,here. —Preceding unsigned comment added by 203.199.41.181 (talk) 09:26, 2 February 2008 (UTC)

Derivation of Partial Expectation

As requested by Rinconsoleao and others, here is a derivation of the partial expectation formula. It is tedious, so I do not include it in the article itself.

We want to find

where f(x) is the lognormal distribution

so we have

Make a change of variables

and giving

combine the exponentials together

fix the quadratic by 'completing the square'

at this point we can pull out some stuff from the integral

one more change of variable

and

gives

We recognize the integral and the fraction in front of it as the complement of the cdf of the std normal rv

using we finally have

Regards, Encyclops (talk) 21:49, 29 August 2009 (UTC)

examples for log normal distributions in nature/economy?

Some examples would be nice! —Preceding unsigned comment added by 146.113.42.220 (talk) 16:41, 8 February 2008 (UTC)

One example is neurological reaction time. This distribution has been seen in studies on automobile braking and other responses to stimuli. See also mental chronometry.--IanOsgood (talk) 02:32, 26 February 2008 (UTC)
This is also useful in telecom. in order to compute slow fading effects on a transmitted signal. -- 82.123.94.169 (talk) 14:42, 28 February 2008 (UTC)

I think the Black–Scholes Option model uses a log-normal assumption about the price of a stock. This makes sense, because its the percentage change in the price that has real meaning, not the price itself. If some external event makes the stock price fall, the amount that it falls is not very important to an investor, its the percent change that really matters. This suggests a log normal distribution. PAR (talk) 17:13, 28 February 2008 (UTC)

I recall reading on wiki that high IQs are log-normally distributed. Also, incomes (in a given country) are approximately as well. Elithrion (talk) 21:26, 2 November 2009 (UTC)

Parameters boundaries ?

If the relationship between the log-normal distribution and the normal distribution is right, then I don't understand why needs to be greater than 0 (since is expected to be a real with no boundary in the normal distribution). At least, it can be null since it's the case with the graphs shown for the pdf and cdf (I've edited the article in consequence). Also, that's not that needs to be greater than 0, but (which simply means that can't be null since it's a real number). -- 82.123.94.169 (talk) 15:04, 28 February 2008 (UTC)

Question: What can possibly be the interpretation of, say, as opposed to ? By strong convention (and quite widely assumed in derivations) standard deviations are taken to be in the domain , although I suppose in this case algebraically can be negative... It's confusing to start talking about negative sds, and unless there's a good reason for it, please don't. --128.59.111.72 (talk) 22:59, 10 March 2008 (UTC)

Yes, you're right: can't be negative or null (it's also obvious reading the PDF formula). I was confused by the Normal Distribution article where only is expected to be positive (which is also not sufficient there). Thanks for your answer, and sorry for that. I guess can't be negative as well because that would be meaningless if it was (even if it would be mathematically correct). -- 82.123.102.83 (talk) 19:33, 13 March 2008 (UTC)

Logarithm Base

Although yes, any base is OK, the derivations and moments, etc. are all done assuming a natural logarithm. Although the distribution would still be lognormal in another base b, the details would all change by a factor of ln(b). A note should probably be added in this section, that we are using by convention the natural logarithm here. (And possibly re-mention it in the PDF.) --128.59.111.72 (talk) 22:59, 10 March 2008 (UTC)

Product of "any" distributions

I think it should be highlighted in the article that the Log-normal distribution is the analogue of the normal distribution in this way: if we take n independent distributions and add them we "get" the normal distribution (NB: here I am lazy on purpose, the precise idea is the Central Limit Theorem). If we take n positive independent distributions and multiply them, we "get" the log-normal (also lazy). Albmont (talk) 11:58, 5 June 2008 (UTC)

This is to some extent expressed (or at least suggested) where the article says "A variable might be modeled as log-normal if it can be thought of as the multiplicative product of many small independent factors". Perhaps it could be said better, but the idea is there. Encyclops (talk) 14:58, 5 June 2008 (UTC)
So we're talking about the difference between "expressed (or at least suggested)" on the one hand, and on the other hand "highlighted". Michael Hardy (talk) 17:39, 5 June 2008 (UTC)
Yes, the ubiquity of the log-normal in Finance comes from this property, so I think this property is important enough to deserve being stated in the initial paragraphs. Just MHO, of course. Albmont (talk) 20:39, 5 June 2008 (UTC)
The factors need to have small departure from 1 ... I have corrected this, but can someone think of a rephrasing for the bit about "the product of the daily return rates"? Is a "return rate" defined so as to be close to 1 (no profit =1) or close to zero (no profit=0)? Melcombe (talk) 13:49, 11 September 2008 (UTC)
The "return rate" should be the one "close to 1 (no profit == 1)." The author must be talking about discount factors rather than rates of return. Rates of return correspond to specific time periods and are therefore neither additive nor multiplicative. Returns are often thought of as normally distributed in finance, so the discount factor would be lognormally distributed. I'll fix this. Alue (talk) 05:14, 19 February 2009 (UTC)
Moreover, it would be nice to have a reference for this section. 188.97.0.158 (talk) 14:21, 4 September 2012 (UTC)

Why the pdf value would be greater than 1 in the pdf picture?

Why the pdf value would be greater than 1 in the pdf picture? Am I missing something here? I am really puzzled. —Preceding unsigned comment added by 208.13.41.5 (talk) 01:55, 11 September 2008 (UTC)

Why are you puzzled? When probability is concentrated near a point, the value of the pdf is large. That's what happens here? Could it be that you're mistaking this situation with that of probability mass functions? Those cannot exceed 1, since their values are probabilities. The values of a pdf, however, are not generally probabilities. Michael Hardy (talk) 02:20, 11 September 2008 (UTC)
Just to put it another way, the area under a pdf is equal to one, not the curve itself. Encyclops (talk) 03:01, 11 September 2008 (UTC)

Now I am unpuzzled. Thanks ;-) —Preceding unsigned comment added by 208.13.41.5 (talk) 16:54, 11 September 2008 (UTC)


The moment problem

In the article it should really mention that the Log-normal distribution suffers from the Moment problem (see for example Counterexamples in Probability, Stoyanov). Basically, there exists infinitely many distributions that have the same moments as the LN, but have a different pdf. In fact (I think), there are also discrete distributions which have the same moments as the LN distribution. ColinGillespie (talk) 11:45, 30 October 2008 (UTC)

moment generating function is defined as
On the whole domain R, it doesn not exist, but for t=0, it does exist for sure, and so is for any t<0. so why don't we try to find the domain set on which it exists? the set {t: Mx(t)<infinite}. Jackzhp (talk) 14:39, 21 January 2009 (UTC)
The cumulant/moment generating function g(t) is convex, 0 belong to the set {t: g(t)<infinite}, if the interior of the set is not empty, then g(t) is analytic there, and infinitely differentialbe there, on the set, g(t) is strictly convex, and g'(t) is strictly increasing. please edit Cumulant#Some_properties_of_cumulant_generating_function or moment generating functionJackzhp (talk) 15:09, 21 January 2009 (UTC)

Why is this even relevant? The distribution is not bounded, and therefore there is no guarantee the infinite set of moments uniquely describes the distribution.  Preceding unsigned comment added by Conduit242 (talkcontribs) 22:35, 8 October 2014 (UTC)

My edit summary got truncated

Here's the whole summary to my last edit:

Two problems: X is more conventional here, and the new edit fails to distinguish between capital for the random variable and lower case for the argument to the density function.

Michael Hardy (talk) 20:15, 18 February 2009 (UTC)

The problem you refer to here is still (again?) present in the section about the density. Below I wrote about it under the title 'Density', but it is not understood I'm afraid.Madyno (talk) 14:24, 10 May 2017 (UTC)

Are the plots accurate?

Something seems a bit odd with the plots. In particular the CDF plot appears to demonstrate that all the curves have a mean at about 1, but if the underlying parameter µ is held fixed, we should see P = 0.5 at around x=3 for sigma = 3/2; and at around 1.35 for sigma=1, and all the way at e^50 for sigma=10. The curves appear to have been plotted with the mean of the lognormal distribution fixed at (µ+o^2/2)=1? ~confused~

Don't confuse the expected value with the point at which the probability is one-half. The latter is well-defined for the Cauchy distribution, while the former is not; thus although x=1 is the point at which all these distributions have P[x < 1] = 1/2; it's not the expected value. Hurr hurr, let no overtired idiots make this mistake again. (signed, Original Poster) —Preceding unsigned comment added by 140.247.249.76 (talk) 09:26, 29 April 2009 (UTC)
We had this discussion on this page in considerable detail before, a couple of years ago. Yes, they're accurates; they're also somewhat counterintuitive. Michael Hardy (talk) 16:40, 17 June 2009 (UTC)

I think it's worth pointing out the that the formula in the code that generates the pdf plots is wrong. The numerator in the exponent is log(x-mu)^2, when it should be (log(x)-mu)^2. It doesn't actually change the plots, because they all use mu=0, but it's an important difference, in case someone else used and modified the code. Sorry if this isn't the place to discuss this - this is my first time discussing on wikipedia. Crichardsns (talk) 01:25, 19 February 2011 (UTC)

The pdf plot is wrong if for no other reason (unless I've missed something important) because one of the curves exceed 1 and can't be a proper pdf. —Preceding unsigned comment added by 203.141.92.14 (talk) 05:31, 1 March 2011 (UTC)

The aesthetics of the plots look terrible.

Nonsense about confidence intervals

I commented out this table:

More information , ...
Confidence interval bounds log space geometric
3σ lower bound
2σ lower bound
1σ lower bound
1σ upper bound
2σ upper bound
3σ upper bound
Close

The table has nothing to do with confidence intervals as those are normally understood. I'm not sure there's much point in doing confidence intervals for the parameters here as a separate topic from confidence intervals for the normal distribution.

Obviously, you cannot use μ and σ to form confidence intervals. They're the things you'd want confidence intervals for! You can't observe them. If you could observe them, what would be the point of confidence intervals? Michael Hardy (talk) 16:43, 17 June 2009 (UTC) :Michael Hardy, I think I agree with you. I think the confidence intervals on parameters are sometimes called "credibility intervals". They are obtained through a Bayesian analysis, using a prior, and where the posterior distribution on the parameters gives the credibility interval. Correct me if I'm wrong. Attic Salt (talk) 00:50, 6 June 2020 (UTC)

This edit was a colossal mistake that stood for almost five years!! Whoever wrote it didn't have a clue what confidence intervals are. Michael Hardy (talk) 16:49, 17 June 2009 (UTC)

Characteristic function

Scaling & inverse

Partial expectation again

Properties?

Is the median the same?

Improvement of pictures

Confusion about location and shape in the Probability distribution table, row on parameters

Wrong formula for parameter μ as a function of the mean and variance

Power laws

Base-unspecific entropy incorrect

Notation/Grouping Clarification in Formula

Graph explaining relation between normal and lognormal distribution

Neuroscience citations

Plots misleading due to coarse sampling

Density

Location and scale

"\ln\mathcal N" ?

"location and scale"

Dimensions

PDF and CDF plots confusing

Geometric mean

log-Lévy distribution bug report

Is it really the max entropy distribution?

Location parameter modifications

PDF plot mean value

Unsubstantiated declarations in Occurence and Applications

Related Articles

Wikiwand AI