Published on *Explorable.com* (https://explorable.com)

In statistics, range is defined simply as the difference between the maximum and minimum observations. It is intuitively obvious why we define range in statistics this way - range should suggest how diversely spread out the values are, and by computing the difference between the maximum and minimum values, we can get an estimate of the spread of the data.

For example, suppose an experiment involves finding out the weight of lab rats and the values in grams are 320, 367, 423, 471 and 480. In this case, the range is simply computed as 480-320 = 160 grams.

Range [3] is quite a useful indication of how spread out the data is, but it has some serious limitations. This is because sometimes data can have outliers [4] that are widely off the other data points. In these cases, the range might not give a true indication of the spread of data.

For example, in our previous case, consider a small baby rat added to the data set that weighs only 50 grams. Now the range is computed as 480-50 = 430 grams, which looks like a false indication of the dispersion of data.

This limitation of range is to be expected primarily because range is computed taking only two data points into consideration. Thus it cannot give a very good estimate of how the overall data behaves.

In a lot of cases, however, data is closely clustered and if the number of observations is very large, then it can give a good sense of data distribution. For example, consider a huge survey of the IQ levels of university students consisting of 10,000 students from different backgrounds. In this case, the range can be a useful tool to measure the dispersion of IQ values among university students.

Sometimes, we define range in such a way so as to eliminate the outliers and extreme points in the data set. For example, the inter-quartile range in statistics is defined as the difference between the third and first quartiles. You can immediately see how this new definition of range is more robust than the previous one. Here the outliers will not matter and this definition takes the whole distribution of data into consideration and not just the maximum and minimum values.

It should be pointed out that in spite of several limitations, the range can be a useful indication for many cases. As a student of statistics you should understand what kinds of data are best suited to be defined based on range. If there are too many outliers, it may not be a good idea. But range gives a quick and easy to estimate indication about the spread of data.

**Links**

[1] https://explorable.com/range-in-statistics

[2] https://explorable.com/users/siddharth

[3] http://en.wikipedia.org/wiki/Range_(statistics)

[4] https://explorable.com/statistical-outliers