・25 min read

An essential piece of analysis of large data is efficient summarization: computing aggregations like

`sum()`

, `mean()`

, `median()`

, `min()`

, and `max()`

, in which a single number gives insight into the nature of a potentially large dataset.
In this section, we'll explore aggregations in Pandas, from simple operations akin to what we've seen on NumPy arrays, to more sophisticated operations based on the concept of a `groupby`

.For convenience, we'll use the same

`display`

magic function that we've seen in previous sections:```
import numpy as np
import pandas as pd
class display(object):
"""Display HTML representation of multiple objects"""
template = """<div style="float: left; padding: 10px;">
<p style='font-family:"Courier New", Courier, monospace'>{0}</p>{1}
</div>"""
def __init__(self, *args):
self.args = args
def _repr_html_(self):
return '\n'.join(self.template.format(a, eval(a)._repr_html_())
for a in self.args)
def __repr__(self):
return '\n\n'.join(a + '\n' + repr(eval(a))
for a in self.args)
```

Here we will use the Planets dataset, available via the Seaborn package (see Visualization With Seaborn).
It gives information on planets that astronomers have discovered around other stars (known as *extrasolar planets* or *exoplanets* for short). It can be downloaded with a simple Seaborn command:

```
import seaborn as sns
planets = sns.load_dataset('planets')
planets.shape
```

(1035, 6)

`planets.head()`

This has some details on the 1,000+ extrasolar planets discovered up to 2014.

Earlier, we explored some of the data aggregations available for NumPy arrays ("Aggregations: Min, Max, and Everything In Between").
As with a one-dimensional NumPy array, for a Pandas

`Series`

the aggregates return a single value:```
rng = np.random.RandomState(42)
ser = pd.Series(rng.rand(5))
ser
```

0 0.374540
1 0.950714
2 0.731994
3 0.598658
4 0.156019
dtype: float64

`ser.sum()`

2.8119254917081569

`ser.mean()`

0.56238509834163142

For a

`DataFrame`

, by default the aggregates return results within each column:```
df = pd.DataFrame({'A': rng.rand(5),
'B': rng.rand(5)})
df
```

`df.mean()`

A 0.477888
B 0.443420
dtype: float64

By specifying the

`axis`

argument, you can instead aggregate within each row:`df.mean(axis='columns')`

0 0.088290
1 0.513997
2 0.849309
3 0.406727
4 0.444949
dtype: float64

Pandas

`Series`

and `DataFrame`

s include all of the common aggregates mentioned in Aggregations: Min, Max, and Everything In Between; in addition, there is a convenience method `describe()`

that computes several common aggregates for each column and returns the result.
Let's use this on the Planets data, for now dropping rows with missing values:`planets.dropna().describe()`

This can be a useful way to begin understanding the overall properties of a dataset.
For example, we see in the *Kepler* mission, which is a space-based telescope specifically designed for finding eclipsing planets around other stars.

`year`

column that although exoplanets were discovered as far back as 1989, half of all known expolanets were not discovered until 2010 or after.
This is largely thanks to the The following table summarizes some other built-in Pandas aggregations:

Aggregation | Description |
---|---|

`count()` | Total number of items |

`first()` , `last()` | First and last item |

`mean()` , `median()` | Mean and median |

`min()` , `max()` | Minimum and maximum |

`std()` , `var()` | Standard deviation and variance |

`mad()` | Mean absolute deviation |

`prod()` | Product of all items |

`sum()` | Sum of all items |

These are all methods of

`DataFrame`

and `Series`

objects.To go deeper into the data, however, simple aggregates are often not enough.
The next level of data summarization is the

`groupby`

operation, which allows you to quickly and efficiently compute aggregates on subsets of data.Simple aggregations can give you a flavor of your dataset, but often we would prefer to aggregate conditionally on some label or index: this is implemented in the so-called *split, apply, combine*.

`groupby`

operation.
The name "group by" comes from a command in the SQL database language, but it is perhaps more illuminative to think of it in the terms first coined by Hadley Wickham of Rstats fame: A canonical example of this split-apply-combine operation, where the "apply" is a summation aggregation, is illustrated in this figure:

This makes clear what the

`groupby`

accomplishes:- The
*split*step involves breaking up and grouping a`DataFrame`

depending on the value of the specified key. - The
*apply*step involves computing some function, usually an aggregate, transformation, or filtering, within the individual groups. - The
*combine*step merges the results of these operations into an output array.

While this could certainly be done manually using some combination of the masking, aggregation, and merging commands covered earlier, an important realization is that *the intermediate splits do not need to be explicitly instantiated*. Rather, the *how* the computation is done under the hood, but rather thinks about the *operation as a whole*.

`GroupBy`

can (often) do this in a single pass over the data, updating the sum, mean, count, min, or other aggregate for each group along the way.
The power of the `GroupBy`

is that it abstracts away these steps: the user need not think about As a concrete example, let's take a look at using Pandas for the computation shown in this diagram.
We'll start by creating the input

`DataFrame`

:```
df = pd.DataFrame({'key': ['A', 'B', 'C', 'A', 'B', 'C'],
'data': range(6)}, columns=['key', 'data'])
df
```

The most basic split-apply-combine operation can be computed with the

`groupby()`

method of `DataFrame`

s, passing the name of the desired key column:`df.groupby('key')`

<pandas.core.groupby.DataFrameGroupBy object at 0x117272160>

Notice that what is returned is not a set of

`DataFrame`

s, but a `DataFrameGroupBy`

object.
This object is where the magic is: you can think of it as a special view of the `DataFrame`

, which is poised to dig into the groups but does no actual computation until the aggregation is applied.
This "lazy evaluation" approach means that common aggregates can be implemented very efficiently in a way that is almost transparent to the user.To produce a result, we can apply an aggregate to this

`DataFrameGroupBy`

object, which will perform the appropriate apply/combine steps to produce the desired result:`df.groupby('key').sum()`

The

`sum()`

method is just one possibility here; you can apply virtually any common Pandas or NumPy aggregation function, as well as virtually any valid `DataFrame`

operation, as we will see in the following discussion.The

`GroupBy`

object is a very flexible abstraction.
In many ways, you can simply treat it as if it's a collection of `DataFrame`

s, and it does the difficult things under the hood. Let's see some examples using the Planets data.Perhaps the most important operations made available by a *aggregate*, *filter*, *transform*, and *apply*.
We'll discuss each of these more fully in "Aggregate, Filter, Transform, Apply", but before that let's introduce some of the other functionality that can be used with the basic

`GroupBy`

are `GroupBy`

operation.The

`GroupBy`

object supports column indexing in the same way as the `DataFrame`

, and returns a modified `GroupBy`

object.
For example:`planets.groupby('method')`

<pandas.core.groupby.DataFrameGroupBy object at 0x1172727b8>

`planets.groupby('method')['orbital_period']`

<pandas.core.groupby.SeriesGroupBy object at 0x117272da0>

Here we've selected a particular

`Series`

group from the original `DataFrame`

group by reference to its column name.
As with the `GroupBy`

object, no computation is done until we call some aggregate on the object:`planets.groupby('method')['orbital_period'].median()`

method
Astrometry 631.180000
Eclipse Timing Variations 4343.500000
Imaging 27500.000000
Microlensing 3300.000000
Orbital Brightness Modulation 0.342887
Pulsar Timing 66.541900
Pulsation Timing Variations 1170.000000
Radial Velocity 360.200000
Transit 5.714932
Transit Timing Variations 57.011000
Name: orbital_period, dtype: float64

This gives an idea of the general scale of orbital periods (in days) that each method is sensitive to.

The

`GroupBy`

object supports direct iteration over the groups, returning each group as a `Series`

or `DataFrame`

:```
for (method, group) in planets.groupby('method'):
print("{0:30s} shape={1}".format(method, group.shape))
```

Astrometry shape=(2, 6)
Eclipse Timing Variations shape=(9, 6)
Imaging shape=(38, 6)
Microlensing shape=(23, 6)
Orbital Brightness Modulation shape=(3, 6)
Pulsar Timing shape=(5, 6)
Pulsation Timing Variations shape=(1, 6)
Radial Velocity shape=(553, 6)
Transit shape=(397, 6)
Transit Timing Variations shape=(4, 6)

This can be useful for doing certain things manually, though it is often much faster to use the built-in

`apply`

functionality, which we will discuss momentarily.Through some Python class magic, any method not explicitly implemented by the

`GroupBy`

object will be passed through and called on the groups, whether they are `DataFrame`

or `Series`

objects.
For example, you can use the `describe()`

method of `DataFrame`

s to perform a set of aggregations that describe each group in the data:`planets.groupby('method')['year'].describe().unstack()`

Looking at this table helps us to better understand the data: for example, the vast majority of planets have been discovered by the Radial Velocity and Transit methods, though the latter only became common (due to new, more accurate telescopes) in the last decade.
The newest methods seem to be Transit Timing Variation and Orbital Brightness Modulation, which were not used to discover a new planet until 2011.

This is just one example of the utility of dispatch methods.
Notice that they are applied *to each individual group*, and the results are then combined within

`GroupBy`

and returned.
Again, any valid `DataFrame`

/`Series`

method can be used on the corresponding `GroupBy`

object, which allows for some very flexible and powerful operations!The preceding discussion focused on aggregation for the combine operation, but there are more options available.
In particular,

`GroupBy`

objects have `aggregate()`

, `filter()`

, `transform()`

, and `apply()`

methods that efficiently implement a variety of useful operations before combining the grouped data.For the purpose of the following subsections, we'll use this

`DataFrame`

:```
rng = np.random.RandomState(0)
df = pd.DataFrame({'key': ['A', 'B', 'C', 'A', 'B', 'C'],
'data1': range(6),
'data2': rng.randint(0, 10, 6)},
columns = ['key', 'data1', 'data2'])
df
```

We're now familiar with

`GroupBy`

aggregations with `sum()`

, `median()`

, and the like, but the `aggregate()`

method allows for even more flexibility.
It can take a string, a function, or a list thereof, and compute all the aggregates at once.
Here is a quick example combining all these:`df.groupby('key').aggregate(['min', np.median, max])`

Another useful pattern is to pass a dictionary mapping column names to operations to be applied on that column:

```
df.groupby('key').aggregate({'data1': 'min',
'data2': 'max'})
```

A filtering operation allows you to drop data based on the group properties.
For example, we might want to keep all groups in which the standard deviation is larger than some critical value:

```
def filter_func(x):
return x['data2'].std() > 4
display('df', "df.groupby('key').std()", "df.groupby('key').filter(filter_func)")
```

The filter function should return a Boolean value specifying whether the group passes the filtering. Here because group A does not have a standard deviation greater than 4, it is dropped from the result.

While aggregation must return a reduced version of the data, transformation can return some transformed version of the full data to recombine.
For such a transformation, the output is the same shape as the input.
A common example is to center the data by subtracting the group-wise mean:

`df.groupby('key').transform(lambda x: x - x.mean())`

The

`apply()`

method lets you apply an arbitrary function to the group results.
The function should take a `DataFrame`

, and return either a Pandas object (e.g., `DataFrame`

, `Series`

) or a scalar; the combine operation will be tailored to the type of output returned.For example, here is an

`apply()`

that normalizes the first column by the sum of the second:```
def norm_by_data2(x):
# x is a DataFrame of group values
x['data1'] /= x['data2'].sum()
return x
display('df', "df.groupby('key').apply(norm_by_data2)")
```

`apply()`

within a `GroupBy`

is quite flexible: the only criterion is that the function takes a `DataFrame`

and returns a Pandas object or scalar; what you do in the middle is up to you!In the simple examples presented before, we split the

`DataFrame`

on a single column name.
This is just one of many options by which the groups can be defined, and we'll go through some other options for group specification here.The key can be any series or list with a length matching that of the

`DataFrame`

. For example:```
L = [0, 1, 0, 1, 2, 0]
display('df', 'df.groupby(L).sum()')
```

Of course, this means there's another, more verbose way of accomplishing the

`df.groupby('key')`

from before:`display('df', "df.groupby(df['key']).sum()")`

Another method is to provide a dictionary that maps index values to the group keys:

```
df2 = df.set_index('key')
mapping = {'A': 'vowel', 'B': 'consonant', 'C': 'consonant'}
display('df2', 'df2.groupby(mapping).sum()')
```

Similar to mapping, you can pass any Python function that will input the index value and output the group:

`display('df2', 'df2.groupby(str.lower).mean()')`

Further, any of the preceding key choices can be combined to group on a multi-index:

`df2.groupby([str.lower, mapping]).mean()`

As an example of this, in a couple lines of Python code we can put all these together and count discovered planets by method and by decade:

```
decade = 10 * (planets['year'] // 10)
decade = decade.astype(str) + 's'
decade.name = 'decade'
planets.groupby(['method', decade])['number'].sum().unstack().fillna(0)
```