﻿﻿﻿ Mean It
Mean It

Mean
mean is the sample mean (denoted x ¯ {\displaystyle {\bar {x}}} ) to distinguish it from the mean of the underlying distribution, the population mean

View Wikipedia Article

General term for the several definitions of mean value, the sum divided by the count This article is about the mathematical concept. For other uses, see Mean (disambiguation). For the state of being mean or cruel, see meanness. For broader coverage of this topic, see average.

There are several kinds of means in various branches of mathematics (especially statistics).

For a data set, the arithmetic mean, also called the mathematical expectation or average, is the central value of a discrete set of numbers: specifically, the sum of the values divided by the number of values. The arithmetic mean of a set of numbers x1, x2, ..., xn is typically denoted by x ¯ {\displaystyle {\bar {x}}} [note 1]. If the data set were based on a series of observations obtained by sampling from a statistical population, the arithmetic mean is the sample mean (denoted x ¯ {\displaystyle {\bar {x}}} ) to distinguish it from the mean of the underlying distribution, the population mean (denoted μ {\displaystyle \mu } or μ x {\displaystyle \mu _{x}} [note 2]).

In probability and statistics, the population mean, or expected value, is a measure of the central tendency either of a probability distribution or of the random variable characterized by that distribution. In the case of a discrete probability distribution of a random variable X, the mean is equal to the sum over every possible value weighted by the probability of that value; that is, it is computed by taking the product of each possible value x of X and its probability p(x), and then adding all these products together, giving μ = ∑ x p ( x ) {\displaystyle \mu =\sum xp(x)} . An analogous formula applies to the case of a continuous probability distribution. Not every probability distribution has a defined mean; see the Cauchy distribution for an example. Moreover, for some distributions the mean is infinite.

For a finite population, the population mean of a property is equal to the arithmetic mean of the given property while considering every member of the population. For example, the population mean height is equal to the sum of the heights of every individual divided by the total number of individuals. The sample mean may differ from the population mean, especially for small samples. The law of large numbers dictates that the larger the size of the sample, the more likely it is that the sample mean will be close to the population mean.

Outside probability and statistics, a wide range of other notions of "mean" are often used in geometry and analysis; examples are given below.

Contents
• 1 Types of mean
• 1.1 Pythagorean means
• 1.1.1 Arithmetic mean (AM)
• 1.1.2 Geometric mean (GM)
• 1.1.3 Harmonic mean (HM)
• 1.1.4 Relationship between AM, GM, and HM
• 1.2 Statistical location
• 1.3 Mean of a probability distribution
• 1.4 Generalized means
• 1.4.1 Power mean
• 1.4.2 f-mean
• 1.5 Weighted arithmetic mean
• 1.6 Truncated mean
• 1.7 Interquartile mean
• 1.8 Mean of a function
• 1.9 Mean of angles and cyclical quantities
• 1.10 Fréchet mean
• 1.11 Other means
• 2 Distribution of the sample mean
• 4 Notes
• 5 References
Types of mean Pythagorean means Main article: Pythagorean means Arithmetic mean (AM) Main article: Arithmetic mean

The arithmetic mean (or simply mean) of a sample x 1 , x 2 , … , x n {\displaystyle x_{1},x_{2},\ldots ,x_{n}} , usually denoted by x ¯ {\displaystyle {\bar {x}}} , is the sum of the sampled values divided by the number of items in the sample

x ¯ = 1 n ( ∑ i = 1 n x i ) = x 1 + x 2 + ⋯ + x n n {\displaystyle {\bar {x}}={\frac {1}{n}}\left(\sum _{i=1}^{n}{x_{i}}\right)={\frac {x_{1}+x_{2}+\cdots +x_{n}}{n}}}

For example, the arithmetic mean of five values: 4, 36, 45, 50, 75 is:

4 + 36 + 45 + 50 + 75 5 = 210 5 = 42. {\displaystyle {\frac {4+36+45+50+75}{5}}={\frac {210}{5}}=42.}
Geometric mean (GM)

The geometric mean is an average that is useful for sets of positive numbers that are interpreted according to their product and not their sum (as is the case with the arithmetic mean); e.g., rates of growth.

x ¯ = ( ∏ i = 1 n x i ) 1 n = ( x 1 x 2 ⋯ x n ) 1 n {\displaystyle {\bar {x}}=\left(\prod _{i=1}^{n}{x_{i}}\right)^{\frac {1}{n}}=\left(x_{1}x_{2}\cdots x_{n}\right)^{\frac {1}{n}}}

For example, the geometric mean of five values: 4, 36, 45, 50, 75 is:

( 4 × 36 × 45 × 50 × 75 ) 1 5 = 24 300 000 5 = 30. {\displaystyle (4\times 36\times 45\times 50\times 75)^{\frac {1}{5}}={\sqrt{24\;300\;000}}=30.}
Harmonic mean (HM)

The harmonic mean is an average which is useful for sets of numbers which are defined in relation to some unit, for example speed (distance per unit of time).

x ¯ = n ( ∑ i = 1 n 1 x i ) − 1 {\displaystyle {\bar {x}}=n\left(\sum _{i=1}^{n}{\frac {1}{x_{i}}}\right)^{-1}}

For example, the harmonic mean of the five values: 4, 36, 45, 50, 75 is

5 1 4 + 1 36 + 1 45 + 1 50 + 1 75 = 5 1 3 = 15. {\displaystyle {\frac {5}{{\tfrac {1}{4}}+{\tfrac {1}{36}}+{\tfrac {1}{45}}+{\tfrac {1}{50}}+{\tfrac {1}{75}}}}={\frac {5}{\;{\tfrac {1}{3}}\;}}=15.}
Relationship between AM, GM, and HM Main article: Inequality of arithmetic and geometric means

AM, GM, and HM satisfy these inequalities:

A M ≥ G M ≥ H M {\displaystyle \mathrm {AM} \geq \mathrm {GM} \geq \mathrm {HM} \,}

Equality holds if and only if all the elements of the given sample are equal.

Statistical location See also: Average § Statistical location Comparison of the arithmetic mean, median and mode of two skewed (log-normal) distributions. Geometric visualization of the mode, median and mean of an arbitrary probability density function.

In descriptive statistics, the mean may be confused with the median, mode or mid-range, as any of these may be called an "average" (more formally, a measure of central tendency). The mean of a set of observations is the arithmetic average of the values; however, for skewed distributions, the mean is not necessarily the same as the middle value (median), or the most likely value (mode). For example, mean income is typically skewed upwards by a small number of people with very large incomes, so that the majority have an income lower than the mean. By contrast, the median income is the level at which half the population is below and half is above. The mode income is the most likely income and favors the larger number of people with lower incomes. While the median and mode are often more intuitive measures for such skewed data, many skewed distributions are in fact best described by their mean, including the exponential and Poisson distributions.

Mean of a probability distribution Main article: Expected value

The mean of a probability distribution is the long-run arithmetic average value of a random variable having that distribution. In this context, it is also known as the expected value. For a discrete probability distribution, the mean is given by ∑ x P ( x ) {\displaystyle \textstyle \sum xP(x)} , where the sum is taken over all possible values of the random variable and P ( x ) {\displaystyle P(x)} is the probability mass function. For a continuous distribution,the mean is ∫ − ∞ ∞ x f ( x ) d x {\displaystyle \textstyle \int _{-\infty }^{\infty }xf(x)\,dx} , where f ( x ) {\displaystyle f(x)} is the probability density function. In all cases, including those in which the distribution is neither discrete nor continuous, the mean is the Lebesgue integral of the random variable with respect to its probability measure. The mean need not exist or be finite; for some probability distributions the mean is infinite (+∞ or −∞), while others have no mean.

Generalized means Power mean

The generalized mean, also known as the power mean or Hölder mean, is an abstraction of the quadratic, arithmetic, geometric and harmonic means. It is defined for a set of n positive numbers xi by

x ¯ ( m ) = ( 1 n ∑ i = 1 n x i m ) 1 m {\displaystyle {\bar {x}}(m)=\left({\frac {1}{n}}\sum _{i=1}^{n}x_{i}^{m}\right)^{\frac {1}{m}}}

By choosing different values for the parameter m, the following types of means are obtained:

m → ∞ {\displaystyle m\rightarrow \infty } maximum of x i {\displaystyle x_{i}} m = 2 {\displaystyle m=2} quadratic mean m = 1 {\displaystyle m=1} arithmetic mean m → 0 {\displaystyle m\rightarrow 0} geometric mean m = − 1 {\displaystyle m=-1} harmonic mean m → − ∞ {\displaystyle m\rightarrow -\infty } minimum of x i {\displaystyle x_{i}}
f-mean

This can be generalized further as the generalized f-mean

x ¯ = f − 1 ( 1 n ∑ i = 1 n f ( x i ) ) {\displaystyle {\bar {x}}=f^{-1}\left({{\frac {1}{n}}\sum _{i=1}^{n}{f\left(x_{i}\right)}}\right)}

and again a suitable choice of an invertible f will give

f ( x ) = x {\displaystyle f(x)=x} arithmetic mean, f ( x ) = 1 x {\displaystyle f(x)={\frac {1}{x}}} harmonic mean, f ( x ) = x m {\displaystyle f(x)=x^{m}} power mean, f ( x ) = ln ⁡ ( x ) {\displaystyle f(x)=\ln(x)} geometric mean.
Weighted arithmetic mean

The weighted arithmetic mean (or weighted average) is used if one wants to combine average values from samples of the same population with different sample sizes:

x ¯ = ∑ i = 1 n w i x i ∑ i = 1 n w i . {\displaystyle {\bar {x}}={\frac {\sum _{i=1}^{n}{w_{i}x_{i}}}{\sum _{i=1}^{n}w_{i}}}.}

The weights w i {\displaystyle w_{i}} represent the sizes of the different samples. In other applications, they represent a measure for the reliability of the influence upon the mean by the respective values.

Truncated mean

Sometimes a set of numbers might contain outliers, i.e., data values which are much lower or much higher than the others. Often, outliers are erroneous data caused by artifacts. In this case, one can use a truncated mean. It involves discarding given parts of the data at the top or the bottom end, typically an equal amount at each end and then taking the arithmetic mean of the remaining data. The number of values removed is indicated as a percentage of the total number of values.

Interquartile mean

The interquartile mean is a specific example of a truncated mean. It is simply the arithmetic mean after removing the lowest and the highest quarter of values.

x ¯ = 2 n ∑ i = n 4 + 1 3 4 n x i {\displaystyle {\bar {x}}={\frac {2}{n}}\;\sum _{i={\frac {n}{4}}+1}^{{\frac {3}{4}}n}\!\!x_{i}}

assuming the values have been ordered, so is simply a specific example of a weighted mean for a specific set of weights.

Mean of a function Main article: Mean of a function

In some circumstances mathematicians may calculate a mean of an infinite (even an uncountable) set of values. This can happen when calculating the mean value y ave {\displaystyle y_{\text{ave}}} of a function f ( x ) {\displaystyle f(x)} . Intuitively this can be thought of as calculating the area under a section of a curve and then dividing by the length of that section. This can be done crudely by counting squares on graph paper or more precisely by integration. The integration formula is written as:

y ave ( a , b ) = 1 b − a ∫ a b f ( x ) d x {\displaystyle y_{\text{ave}}(a,b)={\frac {1}{b-a}}\int \limits _{a}^{b}\!f(x)\,dx}

Care must be taken to make sure that the integral converges. But the mean may be finite even if the function itself tends to infinity at some points.

Mean of angles and cyclical quantities

Angles, times of day, and other cyclical quantities require modular arithmetic to add and otherwise combine numbers. In all these situations, there will not be a unique mean. For example, the times an hour before and after midnight are equidistant to both midnight and noon. It is also possible that no mean exists. Consider a color wheel—there is no mean to the set of all colors. In these situations, you must decide which mean is most useful. You can do this by adjusting the values before averaging, or by using a specialized approach for the mean of circular quantities.

Fréchet mean

The Fréchet mean gives a manner for determining the "center" of a mass distribution on a surface or, more generally, Riemannian manifold. Unlike many other means, the Fréchet mean is defined on a space whose elements cannot necessarily be added together or multiplied by scalars. It is sometimes also known as the Karcher mean (named after Hermann Karcher).

Other means Main category: Means
• Arithmetic-geometric mean
• Arithmetic-harmonic mean
• Cesàro mean
• Chisini mean
• Contraharmonic mean
• Elementary symmetric mean
• Geometric-harmonic mean
• Grand mean
• Heinz mean
• Heronian mean
• Identric mean
• Lehmer mean
• Logarithmic mean
• Moving average
• Neuman–Sándor mean
• Root mean square (quadratic mean)
• Rényi's entropy (a generalized f-mean)
• Spherical mean
• Stolarsky mean
• Weighted geometric mean
• Weighted harmonic mean
Distribution of the sample mean Main article: Standard error of the mean

The arithmetic mean of a population, or population mean, is often denoted μ. The sample mean x ¯ {\displaystyle {\bar {x}}} (the arithmetic mean of a sample of values drawn from the population) makes a good estimator of the population mean, as its expected value is equal to the population mean (that is, it is an unbiased estimator). The sample mean is a random variable, not a constant, since its calculated value will randomly differ depending on which members of the population are sampled, and consequently it will have its own distribution. For a random sample of n independent observations, the expected value of the sample mean is

E ⁡ x ¯ = μ {\displaystyle \operatorname {E} {\bar {x}}=\mu }

and the variance of the sample mean is

var ⁡ ( x ¯ ) = σ 2 n . {\displaystyle \operatorname {var} ({\bar {x}})={\frac {\sigma ^{2}}{n}}.}

If the population is normally distributed, then the sample mean is normally distributed:

x ¯ ∼ N { μ , σ 2 n } . {\displaystyle {\bar {x}}\thicksim N\left\{\mu ,{\frac {\sigma ^{2}}{n}}\right\}.}

If the population is not normally distributed, the sample mean is nonetheless approximately normally distributed if n is large and σ2/n < +∞. This follows from the central limit theorem.

The mean of a list is all of the numbers added together and divided by the amount of numbers

• Mathematics portal
• Central tendency
• Median
• Mode
• Descriptive statistics
• Kurtosis
• Law of averages
• Mean value theorem
• Summary statistics
• Taylor's law
Notes
1. ^ Pronounced "x bar".
2. ^ Greek letter μ, for "mean", pronounced /'mjuː/.
References
1. ^ Underhill, L.G.; Bradfield d. (1998) Introstat, Juta and Company Ltd. .mw-parser-output cite.citation{font-style:inherit}.mw-parser-output .citation q{quotes:"\"""\"""'""'"}.mw-parser-output .citation .cs1-lock-free a{background:url("//upload.wikimedia.org/wikipedia/commons/thumb/6/65/Lock-green.svg/9px-Lock-green.svg.png")no-repeat;background-position:right .1em center}.mw-parser-output .citation .cs1-lock-limited a,.mw-parser-output .citation .cs1-lock-registration a{background:url("//upload.wikimedia.org/wikipedia/commons/thumb/d/d6/Lock-gray-alt-2.svg/9px-Lock-gray-alt-2.svg.png")no-repeat;background-position:right .1em center}.mw-parser-output .citation .cs1-lock-subscription a{background:url("//upload.wikimedia.org/wikipedia/commons/thumb/a/aa/Lock-red-alt-2.svg/9px-Lock-red-alt-2.svg.png")no-repeat;background-position:right .1em center}.mw-parser-output .cs1-subscription,.mw-parser-output .cs1-registration{color:#555}.mw-parser-output .cs1-subscription span,.mw-parser-output .cs1-registration span{border-bottom:1px dotted;cursor:help}.mw-parser-output .cs1-ws-icon a{background:url("//upload.wikimedia.org/wikipedia/commons/thumb/4/4c/Wikisource-logo.svg/12px-Wikisource-logo.svg.png")no-repeat;background-position:right .1em center}.mw-parser-output code.cs1-code{color:inherit;background:inherit;border:inherit;padding:inherit}.mw-parser-output .cs1-hidden-error{display:none;font-size:100%}.mw-parser-output .cs1-visible-error{font-size:100%}.mw-parser-output .cs1-maint{display:none;color:#33aa33;margin-left:0.3em}.mw-parser-output .cs1-subscription,.mw-parser-output .cs1-registration,.mw-parser-output .cs1-format{font-size:95%}.mw-parser-output .cs1-kern-left,.mw-parser-output .cs1-kern-wl-left{padding-left:0.2em}.mw-parser-output .cs1-kern-right,.mw-parser-output .cs1-kern-wl-right{padding-right:0.2em}ISBN 0-7021-3838-X p. 181
2. ^ Feller, William (1950). Introduction to Probability Theory and its Applications, Vol I. Wiley. p. 221. ISBN 0471257087.
3. ^ Elementary Statistics by Robert R. Johnson and Patricia J. Kuby, p. 279
4. ^ Schaum's Outline of Theory and Problems of Probability by Seymour Lipschutz and Marc Lipson, p. 141
5. ^ "AP Statistics Review - Density Curves and the Normal Distributions". Retrieved 16 March 2015.
• v
• t
• e
Statistics
• Outline
• Index
Descriptive statisticsContinuous dataCenter
• Mean
• arithmetic
• geometric
• harmonic
• Median
• Mode
Dispersion
• Variance
• Standard deviation
• Coefficient of variation
• Percentile
• Range
• Interquartile range
Shape
• Central limit theorem
• Moments
• Skewness
• Kurtosis
• L-moments
Count data
• Index of dispersion
Summary tables
• Grouped data
• Frequency distribution
• Contingency table
Dependence
• Pearson product-moment correlation
• Rank correlation
• Spearman's rho
• Kendall's tau
• Partial correlation
• Scatter plot
Graphics
• Bar chart
• Biplot
• Box plot
• Control chart
• Correlogram
• Fan chart
• Forest plot
• Histogram
• Pie chart
• Q–Q plot
• Run chart
• Scatter plot
• Stem-and-leaf display
• Violin plot
Data collectionStudy design
• Population
• Statistic
• Effect size
• Statistical power
• Optimal design
• Sample size determination
• Replication
• Missing data
Survey methodology
• Sampling
• stratified
• cluster
• Standard error
• Opinion poll
• Questionnaire
Controlled experiments
• Scientific control
• Randomized experiment
• Randomized controlled trial
• Random assignment
• Blocking
• Interaction
• Factorial experiment
• Up-and-Down Designs
• Stochastic approximation
Observational Studies
• Cross-sectional study
• Cohort study
• Natural experiment
• Quasi-experiment
Statistical inferenceStatistical theory
• Population
• Statistic
• Probability distribution
• Sampling distribution
• Order statistic
• Empirical distribution
• Density estimation
• Statistical model
• Model specification
• Lp space
• Parameter
• location
• scale
• shape
• Parametric family
• Likelihood (monotone)
• Location–scale family
• Exponential family
• Completeness
• Sufficiency
• Statistical functional
• Bootstrap
• U
• V
• Optimal decision
• loss function
• Efficiency
• Statistical distance
• divergence
• Asymptotics
• Robustness
Frequentist inferencePoint estimation
• Estimating equations
• Maximum likelihood
• Method of moments
• M-estimator
• Minimum distance
• Unbiased estimators
• Mean-unbiased minimum-variance
• Rao–Blackwellization
• Lehmann–Scheffé theorem
• Median unbiased
• Plug-in
Interval estimation
• Confidence interval
• Pivot
• Likelihood interval
• Prediction interval
• Tolerance interval
• Resampling
• Bootstrap
• Jackknife
Testing hypotheses
• 1- & 2-tails
• Power
• Uniformly most powerful test
• Permutation test
• Randomization test
• Multiple comparisons
Parametric tests
• Likelihood-ratio
• Score/Lagrange multiplier
• Wald
Specific tests
• Z-test (normal)
• Student's t-test
• F-test
Goodness of fit
• Chi-squared
• G-test
• Kolmogorov–Smirnov
• Anderson–Darling
• Lilliefors
• Jarque–Bera
• Normality (Shapiro–Wilk)
• Likelihood-ratio test
• Model selection
• Cross validation
• AIC
• BIC
Rank statistics
• Sign
• Sample median
• Signed rank (Wilcoxon)
• Hodges–Lehmann estimator
• Rank sum (Mann–Whitney)
• Nonparametric anova
• 1-way (Kruskal–Wallis)
• 2-way (Friedman)
• Ordered alternative (Jonckheere–Terpstra)
Bayesian inference
• Bayesian probability
• prior
• posterior
• Credible interval
• Bayes factor
• Bayesian estimator
• Maximum posterior estimator
• Correlation
• Regression analysis
Correlation
• Pearson product-moment
• Partial correlation
• Confounding variable
• Coefficient of determination
Regression analysis
• Errors and residuals
• Regression validation
• Mixed effects models
• Simultaneous equations models
• Multivariate adaptive regression splines (MARS)
Linear regression
• Simple linear regression
• Ordinary least squares
• General linear model
• Bayesian regression
Non-standard predictors
• Nonlinear regression
• Nonparametric
• Semiparametric
• Isotonic
• Robust
• Heteroscedasticity
• Homoscedasticity
Generalized linear model
• Exponential families
• Logistic (Bernoulli) / Binomial / Poisson regressions
Partition of variance
• Analysis of variance (ANOVA, anova)
• Analysis of covariance
• Multivariate ANOVA
• Degrees of freedom
Categorical / Multivariate / Time-series / Survival analysisCategorical
• Cohen's kappa
• Contingency table
• Graphical model
• Log-linear model
• McNemar's test
Multivariate
• Regression
• Manova
• Principal components
• Canonical correlation
• Discriminant analysis
• Cluster analysis
• Classification
• Structural equation model
• Factor analysis
• Multivariate distributions
• Elliptical distributions
• Normal
Time-seriesGeneral
• Decomposition
• Trend
• Stationarity
• Exponential smoothing
• Cointegration
• Structural break
• Granger causality
Specific tests
• Dickey–Fuller
• Johansen
• Q-statistic (Ljung–Box)
• Durbin–Watson
• Breusch–Godfrey
Time domain
• Autocorrelation (ACF)
• partial (PACF)
• Cross-correlation (XCF)
• ARMA model
• ARIMA model (Box–Jenkins)
• Autoregressive conditional heteroskedasticity (ARCH)
• Vector autoregression (VAR)
Frequency domain
• Spectral density estimation
• Fourier analysis
• Wavelet
• Whittle likelihood
SurvivalSurvival function
• Kaplan–Meier estimator (product limit)
• Proportional hazards models
• Accelerated failure time (AFT) model
• First hitting time
Hazard function
• Nelson–Aalen estimator
Test
• Log-rank test
ApplicationsBiostatistics
• Bioinformatics
• Clinical trials / studies
• Epidemiology
• Medical statistics
Engineering statistics
• Chemometrics
• Methods engineering
• Probabilistic design
• Process / quality control
• Reliability
• System identification
Social statistics
• Actuarial science
• Census
• Crime statistics
• Demography
• Econometrics
• Jurimetrics
• National accounts
• Official statistics
• Population statistics
• Psychometrics
Spatial statistics
• Cartography
• Environmental statistics
• Geographic information system
• Geostatistics
• Kriging
• Category
• Mathematics portal
• Commons
• WikiProject

Custom Search
Mean It

Search
Mean It

Deals
Mean It

vlrPhone
vlrFilter

Project of very low consumption, radiation and bitrate softphones, with the support of the spatial audio, of the frequency shifts and of the ultrasonic communications
Multifunction Audio Filter with Remote Control