Compare Rows of Strings in Dataframe [duplicate] - pandas
The docs show how to apply multiple functions on a groupby object at a time using a dict with the output column names as the keys:
In [563]: grouped['D'].agg({'result1' : np.sum,
.....: 'result2' : np.mean})
.....:
Out[563]:
result2 result1
A
bar -0.579846 -1.739537
foo -0.280588 -1.402938
However, this only works on a Series groupby object. And when a dict is similarly passed to a groupby DataFrame, it expects the keys to be the column names that the function will be applied to.
What I want to do is apply multiple functions to several columns (but certain columns will be operated on multiple times). Also, some functions will depend on other columns in the groupby object (like sumif functions). My current solution is to go column by column, and doing something like the code above, using lambdas for functions that depend on other rows. But this is taking a long time, (I think it takes a long time to iterate through a groupby object). I'll have to change it so that I iterate through the whole groupby object in a single run, but I'm wondering if there's a built in way in pandas to do this somewhat cleanly.
For example, I've tried something like
grouped.agg({'C_sum' : lambda x: x['C'].sum(),
'C_std': lambda x: x['C'].std(),
'D_sum' : lambda x: x['D'].sum()},
'D_sumifC3': lambda x: x['D'][x['C'] == 3].sum(), ...)
but as expected I get a KeyError (since the keys have to be a column if agg is called from a DataFrame).
Is there any built in way to do what I'd like to do, or a possibility that this functionality may be added, or will I just need to iterate through the groupby manually?
The second half of the currently accepted answer is outdated and has two deprecations. First and most important, you can no longer pass a dictionary of dictionaries to the agg groupby method. Second, never use .ix.
If you desire to work with two separate columns at the same time I would suggest using the apply method which implicitly passes a DataFrame to the applied function. Let's use a similar dataframe as the one from above
df = pd.DataFrame(np.random.rand(4,4), columns=list('abcd'))
df['group'] = [0, 0, 1, 1]
df
a b c d group
0 0.418500 0.030955 0.874869 0.145641 0
1 0.446069 0.901153 0.095052 0.487040 0
2 0.843026 0.936169 0.926090 0.041722 1
3 0.635846 0.439175 0.828787 0.714123 1
A dictionary mapped from column names to aggregation functions is still a perfectly good way to perform an aggregation.
df.groupby('group').agg({'a':['sum', 'max'],
'b':'mean',
'c':'sum',
'd': lambda x: x.max() - x.min()})
a b c d
sum max mean sum <lambda>
group
0 0.864569 0.446069 0.466054 0.969921 0.341399
1 1.478872 0.843026 0.687672 1.754877 0.672401
If you don't like that ugly lambda column name, you can use a normal function and supply a custom name to the special __name__ attribute like this:
def max_min(x):
return x.max() - x.min()
max_min.__name__ = 'Max minus Min'
df.groupby('group').agg({'a':['sum', 'max'],
'b':'mean',
'c':'sum',
'd': max_min})
a b c d
sum max mean sum Max minus Min
group
0 0.864569 0.446069 0.466054 0.969921 0.341399
1 1.478872 0.843026 0.687672 1.754877 0.672401
Using apply and returning a Series
Now, if you had multiple columns that needed to interact together then you cannot use agg, which implicitly passes a Series to the aggregating function. When using apply the entire group as a DataFrame gets passed into the function.
I recommend making a single custom function that returns a Series of all the aggregations. Use the Series index as labels for the new columns:
def f(x):
d = {}
d['a_sum'] = x['a'].sum()
d['a_max'] = x['a'].max()
d['b_mean'] = x['b'].mean()
d['c_d_prodsum'] = (x['c'] * x['d']).sum()
return pd.Series(d, index=['a_sum', 'a_max', 'b_mean', 'c_d_prodsum'])
df.groupby('group').apply(f)
a_sum a_max b_mean c_d_prodsum
group
0 0.864569 0.446069 0.466054 0.173711
1 1.478872 0.843026 0.687672 0.630494
If you are in love with MultiIndexes, you can still return a Series with one like this:
def f_mi(x):
d = []
d.append(x['a'].sum())
d.append(x['a'].max())
d.append(x['b'].mean())
d.append((x['c'] * x['d']).sum())
return pd.Series(d, index=[['a', 'a', 'b', 'c_d'],
['sum', 'max', 'mean', 'prodsum']])
df.groupby('group').apply(f_mi)
a b c_d
sum max mean prodsum
group
0 0.864569 0.446069 0.466054 0.173711
1 1.478872 0.843026 0.687672 0.630494
For the first part you can pass a dict of column names for keys and a list of functions for the values:
In [28]: df
Out[28]:
A B C D E GRP
0 0.395670 0.219560 0.600644 0.613445 0.242893 0
1 0.323911 0.464584 0.107215 0.204072 0.927325 0
2 0.321358 0.076037 0.166946 0.439661 0.914612 1
3 0.133466 0.447946 0.014815 0.130781 0.268290 1
In [26]: f = {'A':['sum','mean'], 'B':['prod']}
In [27]: df.groupby('GRP').agg(f)
Out[27]:
A B
sum mean prod
GRP
0 0.719580 0.359790 0.102004
1 0.454824 0.227412 0.034060
UPDATE 1:
Because the aggregate function works on Series, references to the other column names are lost. To get around this, you can reference the full dataframe and index it using the group indices within the lambda function.
Here's a hacky workaround:
In [67]: f = {'A':['sum','mean'], 'B':['prod'], 'D': lambda g: df.loc[g.index].E.sum()}
In [69]: df.groupby('GRP').agg(f)
Out[69]:
A B D
sum mean prod <lambda>
GRP
0 0.719580 0.359790 0.102004 1.170219
1 0.454824 0.227412 0.034060 1.182901
Here, the resultant 'D' column is made up of the summed 'E' values.
UPDATE 2:
Here's a method that I think will do everything you ask. First make a custom lambda function. Below, g references the group. When aggregating, g will be a Series. Passing g.index to df.ix[] selects the current group from df. I then test if column C is less than 0.5. The returned boolean series is passed to g[] which selects only those rows meeting the criteria.
In [95]: cust = lambda g: g[df.loc[g.index]['C'] < 0.5].sum()
In [96]: f = {'A':['sum','mean'], 'B':['prod'], 'D': {'my name': cust}}
In [97]: df.groupby('GRP').agg(f)
Out[97]:
A B D
sum mean prod my name
GRP
0 0.719580 0.359790 0.102004 0.204072
1 0.454824 0.227412 0.034060 0.570441
Pandas >= 0.25.0, named aggregations
Since pandas version 0.25.0 or higher, we are moving away from the dictionary based aggregation and renaming, and moving towards named aggregations which accepts a tuple. Now we can simultaneously aggregate + rename to a more informative column name:
Example:
df = pd.DataFrame(np.random.rand(4,4), columns=list('abcd'))
df['group'] = [0, 0, 1, 1]
a b c d group
0 0.521279 0.914988 0.054057 0.125668 0
1 0.426058 0.828890 0.784093 0.446211 0
2 0.363136 0.843751 0.184967 0.467351 1
3 0.241012 0.470053 0.358018 0.525032 1
Apply GroupBy.agg with named aggregation:
df.groupby('group').agg(
a_sum=('a', 'sum'),
a_mean=('a', 'mean'),
b_mean=('b', 'mean'),
c_sum=('c', 'sum'),
d_range=('d', lambda x: x.max() - x.min())
)
a_sum a_mean b_mean c_sum d_range
group
0 0.947337 0.473668 0.871939 0.838150 0.320543
1 0.604149 0.302074 0.656902 0.542985 0.057681
As an alternative (mostly on aesthetics) to Ted Petrou's answer, I found I preferred a slightly more compact listing. Please don't consider accepting it, it's just a much-more-detailed comment on Ted's answer, plus code/data. Python/pandas is not my first/best, but I found this to read well:
df.groupby('group') \
.apply(lambda x: pd.Series({
'a_sum' : x['a'].sum(),
'a_max' : x['a'].max(),
'b_mean' : x['b'].mean(),
'c_d_prodsum' : (x['c'] * x['d']).sum()
})
)
a_sum a_max b_mean c_d_prodsum
group
0 0.530559 0.374540 0.553354 0.488525
1 1.433558 0.832443 0.460206 0.053313
I find it more reminiscent of dplyr pipes and data.table chained commands. Not to say they're better, just more familiar to me. (I certainly recognize the power and, for many, the preference of using more formalized def functions for these types of operations. This is just an alternative, not necessarily better.)
I generated data in the same manner as Ted, I'll add a seed for reproducibility.
import numpy as np
np.random.seed(42)
df = pd.DataFrame(np.random.rand(4,4), columns=list('abcd'))
df['group'] = [0, 0, 1, 1]
df
a b c d group
0 0.374540 0.950714 0.731994 0.598658 0
1 0.156019 0.155995 0.058084 0.866176 0
2 0.601115 0.708073 0.020584 0.969910 1
3 0.832443 0.212339 0.181825 0.183405 1
New in version 0.25.0.
To support column-specific aggregation with control over the output column names, pandas accepts the special syntax in GroupBy.agg(), known as “named aggregation”, where
The keywords are the output column names
The values are tuples whose first element is the column to select and the second element is the aggregation to apply to that column. Pandas provides the pandas.NamedAgg namedtuple with the fields ['column', 'aggfunc'] to make it clearer what the arguments are. As usual, the aggregation can be a callable or a string alias.
>>> animals = pd.DataFrame({
... 'kind': ['cat', 'dog', 'cat', 'dog'],
... 'height': [9.1, 6.0, 9.5, 34.0],
... 'weight': [7.9, 7.5, 9.9, 198.0]
... })
>>> print(animals)
kind height weight
0 cat 9.1 7.9
1 dog 6.0 7.5
2 cat 9.5 9.9
3 dog 34.0 198.0
>>> print(
... animals
... .groupby('kind')
... .agg(
... min_height=pd.NamedAgg(column='height', aggfunc='min'),
... max_height=pd.NamedAgg(column='height', aggfunc='max'),
... average_weight=pd.NamedAgg(column='weight', aggfunc=np.mean),
... )
... )
min_height max_height average_weight
kind
cat 9.1 9.5 8.90
dog 6.0 34.0 102.75
pandas.NamedAgg is just a namedtuple. Plain tuples are allowed as well.
>>> print(
... animals
... .groupby('kind')
... .agg(
... min_height=('height', 'min'),
... max_height=('height', 'max'),
... average_weight=('weight', np.mean),
... )
... )
min_height max_height average_weight
kind
cat 9.1 9.5 8.90
dog 6.0 34.0 102.75
Additional keyword arguments are not passed through to the aggregation functions. Only pairs of (column, aggfunc) should be passed as **kwargs. If your aggregation functions requires additional arguments, partially apply them with functools.partial().
Named aggregation is also valid for Series groupby aggregations. In this case there’s no column selection, so the values are just the functions.
>>> print(
... animals
... .groupby('kind')
... .height
... .agg(
... min_height='min',
... max_height='max',
... )
... )
min_height max_height
kind
cat 9.1 9.5
dog 6.0 34.0
This is a twist on 'exans' answer that uses Named Aggregations. It's the same but with argument unpacking which allows you to still pass in a dictionary to the agg function.
The named aggs are a nice feature, but at first glance might seem hard to write programmatically since they use keywords, but it's actually simple with argument/keyword unpacking.
animals = pd.DataFrame({'kind': ['cat', 'dog', 'cat', 'dog'],
'height': [9.1, 6.0, 9.5, 34.0],
'weight': [7.9, 7.5, 9.9, 198.0]})
agg_dict = {
"min_height": pd.NamedAgg(column='height', aggfunc='min'),
"max_height": pd.NamedAgg(column='height', aggfunc='max'),
"average_weight": pd.NamedAgg(column='weight', aggfunc=np.mean)
}
animals.groupby("kind").agg(**agg_dict)
The Result
min_height max_height average_weight
kind
cat 9.1 9.5 8.90
dog 6.0 34.0 102.75
Ted's answer is amazing. I ended up using a smaller version of that in case anyone is interested. Useful when you are looking for one aggregation that depends on values from multiple columns:
create a dataframe
df = pd.DataFrame({
'a': [1, 2, 3, 4, 5, 6],
'b': [1, 1, 0, 1, 1, 0],
'c': ['x', 'x', 'y', 'y', 'z', 'z']
})
print(df)
a b c
0 1 1 x
1 2 1 x
2 3 0 y
3 4 1 y
4 5 1 z
5 6 0 z
grouping and aggregating with apply (using multiple columns)
print(
df
.groupby('c')
.apply(lambda x: x['a'][(x['a'] > 1) & (x['b'] == 1)]
.mean()
)
c
x 2.0
y 4.0
z 5.0
grouping and aggregating with aggregate (using multiple columns)
I like this approach since I can still use aggregate. Perhaps people will let me know why apply is needed for getting at multiple columns when doing aggregations on groups.
It seems obvious now, but as long as you don't select the column of interest directly after the groupby, you will have access to all the columns of the dataframe from within your aggregation function.
only access to the selected column
df.groupby('c')['a'].aggregate(lambda x: x[x > 1].mean())
access to all columns since selection is after all the magic
df.groupby('c').aggregate(lambda x: x[(x['a'] > 1) & (x['b'] == 1)].mean())['a']
or similarly
df.groupby('c').aggregate(lambda x: x['a'][(x['a'] > 1) & (x['b'] == 1)].mean())
I hope this helps.
Related
Group rows of 2nd Column by 1st Column using condition for length of string characters of text of the1st Column [duplicate]
The docs show how to apply multiple functions on a groupby object at a time using a dict with the output column names as the keys: In [563]: grouped['D'].agg({'result1' : np.sum, .....: 'result2' : np.mean}) .....: Out[563]: result2 result1 A bar -0.579846 -1.739537 foo -0.280588 -1.402938 However, this only works on a Series groupby object. And when a dict is similarly passed to a groupby DataFrame, it expects the keys to be the column names that the function will be applied to. What I want to do is apply multiple functions to several columns (but certain columns will be operated on multiple times). Also, some functions will depend on other columns in the groupby object (like sumif functions). My current solution is to go column by column, and doing something like the code above, using lambdas for functions that depend on other rows. But this is taking a long time, (I think it takes a long time to iterate through a groupby object). I'll have to change it so that I iterate through the whole groupby object in a single run, but I'm wondering if there's a built in way in pandas to do this somewhat cleanly. For example, I've tried something like grouped.agg({'C_sum' : lambda x: x['C'].sum(), 'C_std': lambda x: x['C'].std(), 'D_sum' : lambda x: x['D'].sum()}, 'D_sumifC3': lambda x: x['D'][x['C'] == 3].sum(), ...) but as expected I get a KeyError (since the keys have to be a column if agg is called from a DataFrame). Is there any built in way to do what I'd like to do, or a possibility that this functionality may be added, or will I just need to iterate through the groupby manually?
The second half of the currently accepted answer is outdated and has two deprecations. First and most important, you can no longer pass a dictionary of dictionaries to the agg groupby method. Second, never use .ix. If you desire to work with two separate columns at the same time I would suggest using the apply method which implicitly passes a DataFrame to the applied function. Let's use a similar dataframe as the one from above df = pd.DataFrame(np.random.rand(4,4), columns=list('abcd')) df['group'] = [0, 0, 1, 1] df a b c d group 0 0.418500 0.030955 0.874869 0.145641 0 1 0.446069 0.901153 0.095052 0.487040 0 2 0.843026 0.936169 0.926090 0.041722 1 3 0.635846 0.439175 0.828787 0.714123 1 A dictionary mapped from column names to aggregation functions is still a perfectly good way to perform an aggregation. df.groupby('group').agg({'a':['sum', 'max'], 'b':'mean', 'c':'sum', 'd': lambda x: x.max() - x.min()}) a b c d sum max mean sum <lambda> group 0 0.864569 0.446069 0.466054 0.969921 0.341399 1 1.478872 0.843026 0.687672 1.754877 0.672401 If you don't like that ugly lambda column name, you can use a normal function and supply a custom name to the special __name__ attribute like this: def max_min(x): return x.max() - x.min() max_min.__name__ = 'Max minus Min' df.groupby('group').agg({'a':['sum', 'max'], 'b':'mean', 'c':'sum', 'd': max_min}) a b c d sum max mean sum Max minus Min group 0 0.864569 0.446069 0.466054 0.969921 0.341399 1 1.478872 0.843026 0.687672 1.754877 0.672401 Using apply and returning a Series Now, if you had multiple columns that needed to interact together then you cannot use agg, which implicitly passes a Series to the aggregating function. When using apply the entire group as a DataFrame gets passed into the function. I recommend making a single custom function that returns a Series of all the aggregations. Use the Series index as labels for the new columns: def f(x): d = {} d['a_sum'] = x['a'].sum() d['a_max'] = x['a'].max() d['b_mean'] = x['b'].mean() d['c_d_prodsum'] = (x['c'] * x['d']).sum() return pd.Series(d, index=['a_sum', 'a_max', 'b_mean', 'c_d_prodsum']) df.groupby('group').apply(f) a_sum a_max b_mean c_d_prodsum group 0 0.864569 0.446069 0.466054 0.173711 1 1.478872 0.843026 0.687672 0.630494 If you are in love with MultiIndexes, you can still return a Series with one like this: def f_mi(x): d = [] d.append(x['a'].sum()) d.append(x['a'].max()) d.append(x['b'].mean()) d.append((x['c'] * x['d']).sum()) return pd.Series(d, index=[['a', 'a', 'b', 'c_d'], ['sum', 'max', 'mean', 'prodsum']]) df.groupby('group').apply(f_mi) a b c_d sum max mean prodsum group 0 0.864569 0.446069 0.466054 0.173711 1 1.478872 0.843026 0.687672 0.630494
For the first part you can pass a dict of column names for keys and a list of functions for the values: In [28]: df Out[28]: A B C D E GRP 0 0.395670 0.219560 0.600644 0.613445 0.242893 0 1 0.323911 0.464584 0.107215 0.204072 0.927325 0 2 0.321358 0.076037 0.166946 0.439661 0.914612 1 3 0.133466 0.447946 0.014815 0.130781 0.268290 1 In [26]: f = {'A':['sum','mean'], 'B':['prod']} In [27]: df.groupby('GRP').agg(f) Out[27]: A B sum mean prod GRP 0 0.719580 0.359790 0.102004 1 0.454824 0.227412 0.034060 UPDATE 1: Because the aggregate function works on Series, references to the other column names are lost. To get around this, you can reference the full dataframe and index it using the group indices within the lambda function. Here's a hacky workaround: In [67]: f = {'A':['sum','mean'], 'B':['prod'], 'D': lambda g: df.loc[g.index].E.sum()} In [69]: df.groupby('GRP').agg(f) Out[69]: A B D sum mean prod <lambda> GRP 0 0.719580 0.359790 0.102004 1.170219 1 0.454824 0.227412 0.034060 1.182901 Here, the resultant 'D' column is made up of the summed 'E' values. UPDATE 2: Here's a method that I think will do everything you ask. First make a custom lambda function. Below, g references the group. When aggregating, g will be a Series. Passing g.index to df.ix[] selects the current group from df. I then test if column C is less than 0.5. The returned boolean series is passed to g[] which selects only those rows meeting the criteria. In [95]: cust = lambda g: g[df.loc[g.index]['C'] < 0.5].sum() In [96]: f = {'A':['sum','mean'], 'B':['prod'], 'D': {'my name': cust}} In [97]: df.groupby('GRP').agg(f) Out[97]: A B D sum mean prod my name GRP 0 0.719580 0.359790 0.102004 0.204072 1 0.454824 0.227412 0.034060 0.570441
Pandas >= 0.25.0, named aggregations Since pandas version 0.25.0 or higher, we are moving away from the dictionary based aggregation and renaming, and moving towards named aggregations which accepts a tuple. Now we can simultaneously aggregate + rename to a more informative column name: Example: df = pd.DataFrame(np.random.rand(4,4), columns=list('abcd')) df['group'] = [0, 0, 1, 1] a b c d group 0 0.521279 0.914988 0.054057 0.125668 0 1 0.426058 0.828890 0.784093 0.446211 0 2 0.363136 0.843751 0.184967 0.467351 1 3 0.241012 0.470053 0.358018 0.525032 1 Apply GroupBy.agg with named aggregation: df.groupby('group').agg( a_sum=('a', 'sum'), a_mean=('a', 'mean'), b_mean=('b', 'mean'), c_sum=('c', 'sum'), d_range=('d', lambda x: x.max() - x.min()) ) a_sum a_mean b_mean c_sum d_range group 0 0.947337 0.473668 0.871939 0.838150 0.320543 1 0.604149 0.302074 0.656902 0.542985 0.057681
As an alternative (mostly on aesthetics) to Ted Petrou's answer, I found I preferred a slightly more compact listing. Please don't consider accepting it, it's just a much-more-detailed comment on Ted's answer, plus code/data. Python/pandas is not my first/best, but I found this to read well: df.groupby('group') \ .apply(lambda x: pd.Series({ 'a_sum' : x['a'].sum(), 'a_max' : x['a'].max(), 'b_mean' : x['b'].mean(), 'c_d_prodsum' : (x['c'] * x['d']).sum() }) ) a_sum a_max b_mean c_d_prodsum group 0 0.530559 0.374540 0.553354 0.488525 1 1.433558 0.832443 0.460206 0.053313 I find it more reminiscent of dplyr pipes and data.table chained commands. Not to say they're better, just more familiar to me. (I certainly recognize the power and, for many, the preference of using more formalized def functions for these types of operations. This is just an alternative, not necessarily better.) I generated data in the same manner as Ted, I'll add a seed for reproducibility. import numpy as np np.random.seed(42) df = pd.DataFrame(np.random.rand(4,4), columns=list('abcd')) df['group'] = [0, 0, 1, 1] df a b c d group 0 0.374540 0.950714 0.731994 0.598658 0 1 0.156019 0.155995 0.058084 0.866176 0 2 0.601115 0.708073 0.020584 0.969910 1 3 0.832443 0.212339 0.181825 0.183405 1
New in version 0.25.0. To support column-specific aggregation with control over the output column names, pandas accepts the special syntax in GroupBy.agg(), known as “named aggregation”, where The keywords are the output column names The values are tuples whose first element is the column to select and the second element is the aggregation to apply to that column. Pandas provides the pandas.NamedAgg namedtuple with the fields ['column', 'aggfunc'] to make it clearer what the arguments are. As usual, the aggregation can be a callable or a string alias. >>> animals = pd.DataFrame({ ... 'kind': ['cat', 'dog', 'cat', 'dog'], ... 'height': [9.1, 6.0, 9.5, 34.0], ... 'weight': [7.9, 7.5, 9.9, 198.0] ... }) >>> print(animals) kind height weight 0 cat 9.1 7.9 1 dog 6.0 7.5 2 cat 9.5 9.9 3 dog 34.0 198.0 >>> print( ... animals ... .groupby('kind') ... .agg( ... min_height=pd.NamedAgg(column='height', aggfunc='min'), ... max_height=pd.NamedAgg(column='height', aggfunc='max'), ... average_weight=pd.NamedAgg(column='weight', aggfunc=np.mean), ... ) ... ) min_height max_height average_weight kind cat 9.1 9.5 8.90 dog 6.0 34.0 102.75 pandas.NamedAgg is just a namedtuple. Plain tuples are allowed as well. >>> print( ... animals ... .groupby('kind') ... .agg( ... min_height=('height', 'min'), ... max_height=('height', 'max'), ... average_weight=('weight', np.mean), ... ) ... ) min_height max_height average_weight kind cat 9.1 9.5 8.90 dog 6.0 34.0 102.75 Additional keyword arguments are not passed through to the aggregation functions. Only pairs of (column, aggfunc) should be passed as **kwargs. If your aggregation functions requires additional arguments, partially apply them with functools.partial(). Named aggregation is also valid for Series groupby aggregations. In this case there’s no column selection, so the values are just the functions. >>> print( ... animals ... .groupby('kind') ... .height ... .agg( ... min_height='min', ... max_height='max', ... ) ... ) min_height max_height kind cat 9.1 9.5 dog 6.0 34.0
This is a twist on 'exans' answer that uses Named Aggregations. It's the same but with argument unpacking which allows you to still pass in a dictionary to the agg function. The named aggs are a nice feature, but at first glance might seem hard to write programmatically since they use keywords, but it's actually simple with argument/keyword unpacking. animals = pd.DataFrame({'kind': ['cat', 'dog', 'cat', 'dog'], 'height': [9.1, 6.0, 9.5, 34.0], 'weight': [7.9, 7.5, 9.9, 198.0]}) agg_dict = { "min_height": pd.NamedAgg(column='height', aggfunc='min'), "max_height": pd.NamedAgg(column='height', aggfunc='max'), "average_weight": pd.NamedAgg(column='weight', aggfunc=np.mean) } animals.groupby("kind").agg(**agg_dict) The Result min_height max_height average_weight kind cat 9.1 9.5 8.90 dog 6.0 34.0 102.75
Ted's answer is amazing. I ended up using a smaller version of that in case anyone is interested. Useful when you are looking for one aggregation that depends on values from multiple columns: create a dataframe df = pd.DataFrame({ 'a': [1, 2, 3, 4, 5, 6], 'b': [1, 1, 0, 1, 1, 0], 'c': ['x', 'x', 'y', 'y', 'z', 'z'] }) print(df) a b c 0 1 1 x 1 2 1 x 2 3 0 y 3 4 1 y 4 5 1 z 5 6 0 z grouping and aggregating with apply (using multiple columns) print( df .groupby('c') .apply(lambda x: x['a'][(x['a'] > 1) & (x['b'] == 1)] .mean() ) c x 2.0 y 4.0 z 5.0 grouping and aggregating with aggregate (using multiple columns) I like this approach since I can still use aggregate. Perhaps people will let me know why apply is needed for getting at multiple columns when doing aggregations on groups. It seems obvious now, but as long as you don't select the column of interest directly after the groupby, you will have access to all the columns of the dataframe from within your aggregation function. only access to the selected column df.groupby('c')['a'].aggregate(lambda x: x[x > 1].mean()) access to all columns since selection is after all the magic df.groupby('c').aggregate(lambda x: x[(x['a'] > 1) & (x['b'] == 1)].mean())['a'] or similarly df.groupby('c').aggregate(lambda x: x['a'][(x['a'] > 1) & (x['b'] == 1)].mean()) I hope this helps.
How to choose categorical variables?
I'm new to modeling I have a lot of functions and I have to separate them into discrete and continuous functions. There are only type tips on the internet Categorical = np.where(df.dtypes = np.object)[0] Categorical Then I have only categorical functions resulting from the format, and I mean real categorical functions. Is there a quick way to do this?
Whether a variable should be considered "discrete" or "continuous" is dependent on the variable and the use case. To count the number of distinct values a variable takes in your dataframe - you can use the pd.Series.nunique or pd.Series.value_counts functions and decide to treat a variable as discrete or continuous based on the output. pandas does come with a dedicated dtype called category which might be helpful - https://pandas.pydata.org/pandas-docs/stable/user_guide/categorical.html import pandas as pd df = pd.DataFrame({'A': [1, 1, 2, 3], 'B': ['a', 'a', 'b', 'b']}) In [3]: df Out[3]: A B 0 1 a 1 1 a 2 2 b 3 3 b In [4]: df.A.value_counts() Out[4]: 1 2 3 1 2 1 Name: A, dtype: int64 In [5]: df['B'].nunique() Out[5]: 2 In [6]: df['B'].unique() Out[6]: array(['a', 'b'], dtype=object)
And if you don't want to play in a loop, just use this simple line of code: categorical_fun = np.where((df2.dtypes == np.object)|(df2.nunique() <= 10))[0] categorical_fun
As you know that these categorical data are either in 'object' format or have a limited number of unique values, use a double condition and let the code show you these columns. Suppose you think that categorical variables have up to 10 unique values, then: a,b = df2.shape #<- ile mamy kolumn b print('ONLY DISCRETE FUNCTION') print('----------------------') for i in range(1,b): i = df2.columns[i] f = df2[i].dtypes h = df2[i].nunique() if f == np.object or h<=10: print(i,"---",f,"---",h)
pandas operations over multiple axis
How can I do operations over multiple columns in one go in pandas? For example, I would like to calculate the df[['a',b']].mean(level=0) or df[['a',b']].kurtosis(level=0) (I need the level=0 as it's a multi indexed dataframe). But I would like to have one single number and do the calculation over multiple axis in one go. A and B would be merged into one single column (or series). In numpy this is possible I believe with axis=(0,1), but I'm unsure how this can be achieved in pandas. Speed is very important, so apply or iterating is not a solution. The expected result would be as follows: np.random.seed([3, 1415]) df = pd.DataFrame( np.random.rand(10, 2), pd.MultiIndex.from_product([list('ab'), range(5)]), list('AB') ) df Out[76]: A B a 0 0.444939 0.407554 1 0.460148 0.465239 2 0.462691 0.016545 3 0.850445 0.817744 4 0.777962 0.757983 b 0 0.934829 0.831104 1 0.879891 0.926879 2 0.721535 0.117642 3 0.145906 0.199844 4 0.437564 0.100702 expected result: df.groupby(level=0).agg(['mean']).mean(axis=1) Out[78]: a 0.546125 b 0.529589 dtype: float64 But it needs to be achieved in one single calculation, not in mean of mean, as this will maybe work for the mean, but for other calculations it may not produce the same result as if it was done in one go (for example I'm not sure if the kurtosis of the kurtosis is equal to the kurtosis in one go.)
Consider the sample dataframe df np.random.seed([3, 1415]) df = pd.DataFrame( np.random.rand(10, 2), pd.MultiIndex.from_product([list('ab'), range(5)]), list('AB') ) df A B a 0 0.444939 0.407554 1 0.460148 0.465239 2 0.462691 0.016545 3 0.850445 0.817744 4 0.777962 0.757983 b 0 0.934829 0.831104 1 0.879891 0.926879 2 0.721535 0.117642 3 0.145906 0.199844 4 0.437564 0.100702 Typical Solution Use groupby and agg df.groupby(level=0).agg(['mean', pd.Series.kurt]) A B mean kurt mean kurt a 0.599237 -2.885262 0.493013 0.018225 b 0.623945 -0.900488 0.435234 -3.105328 Solve Different pd.concat([ df.mean(level=0), df.kurt(level=0) ], axis=1, keys=['Mean', 'Kurt']).swaplevel(1, 0, 1).sort_index(1) A B Kurt Mean Kurt Mean a -2.885262 0.599237 0.018225 0.493013 b -0.900488 0.623945 -3.105328 0.435234
This seems to work: df.stack().mean(level=0) Out[146]: a 0.546125 b 0.529589 dtype: float64
Grouped function between 2 columns in a pandas.DataFrame?
I have a dataframe that has multiple numerical data columns, and a 'group' column. I want to get the output of various functions over two of the columns, for each group. Example data and function: df = pandas.DataFrame({"Dummy":[1,2]*6, "X":[1,3,7]*4, "Y":[2,3,4]*4, "group":["A","B"]*6}) def RMSE(X): return(np.sqrt(np.sum((X.iloc[:,0] - X.iloc[:,1])**2))) I want to do something like group_correlations = df[["X", "Y"]].groupby('group').apply(RMSE) But if I do that, the 'group' column isn't in the dataframe. If I do it the other way around, like this: group_correlations = df.groupby('group')[["X", "Y"]].apply(RMSE) Then the column selection doesn't work: df.groupby('group')[['X', 'Y']].head(1) Dummy X Y group group A 0 1 1 2 A B 1 2 3 3 B the Dummy column is still included, so the function will calculate RMSE on the wrong data. Is there any way to do what I'm trying to do? I know I could do a for loop over the different groups, and subselect the columns manually, but I'd prefer to do it the pandas way, if there is one.
This looks like a bug (or that grabbing multiple columns in a groupby is not implemented?), a workaround is to pass in the groupby column directly: In [11]: df[['X', 'Y']].groupby(df['group']).apply(RMSE) Out[11]: group A 4.472136 B 4.472136 dtype: float64 To see it's the same: In [12]: df.groupby('group')[['X', 'Y']].apply(RMSE) # wrong Out[12]: group A 8.944272 B 7.348469 dtype: float64 In [13]: df.iloc[:, 1:].groupby('group')[['X', 'Y']].apply(RMSE) # correct: ignore dummy col Out[13]: group A 4.472136 B 4.472136 dtype: float64 More robust implementation: To avoid this completely, you could change RMSE to select the columns by name: In [21]: def RMSE2(X, left_col, right_col): return(np.sqrt(np.sum((X[left_col] - X[right_col])**2))) In [22]: df.groupby('group').apply(RMSE2, 'X', 'Y') # equivalent to passing lambda x: RMSE2(x, 'X', 'Y')) Out[22]: group A 4.472136 B 4.472136 dtype: float64 Thanks to #naught101 for pointing out the sweet apply syntax to avoid the lambda.
Case insensitive pandas.concat
How would I perform a case insensitive pandas.concat? df1 = pd.DataFrame({"a":[1,2,3]},index=["a","b","c"]) df2 = pd.DataFrame({"b":[1,2,3]},index=["a","b","c"]) df1a = pd.DataFrame({"A":[1,2,3]},index=["A","B","C"]) pd.concat([df1, df2],axis=1) a b a 1 1 b 2 2 c 3 3 but this does not work: pd.concat([df1, df1a],axis=1) a A A NaN 1 B NaN 2 C NaN 3 a 1 NaN b 2 NaN c 3 NaN Is there an easy way to do this? I have the same question for concat on a Series. This works for a DataFrame: pd.DataFrame([11,21,31],index=pd.MultiIndex.from_tuples([("A",x) for x in ["a","B","c"]])).rename(str.lower) but this does not work for a Series: pd.Series([11,21,31],index=pd.MultiIndex.from_tuples([("A",x) for x in ["a","B","c"]])).rename(str.lower) TypeError: descriptor 'lower' requires a 'str' object but received a 'tuple' For renaming, DataFrames use: def rename_axis(self, mapper, axis=1): index = self.axes[axis] if isinstance(index, MultiIndex): new_axis = MultiIndex.from_tuples([tuple(mapper(y) for y in x) for x in index], names=index.names) else: new_axis = Index([mapper(x) for x in index], name=index.name) whereas when renaming Series: result.index = Index([mapper_f(x) for x in self.index], name=self.index.name) so my updated question is how to perform the rename/case insensitive concat with a Series?
You can do this via rename: pd.concat([df1, df1a.rename(index=str.lower)], axis=1) EDIT: If you want to do this with a MultiIndexed Series you'll need to set it manually, for now. There's a bug report over at pandas GitHub repo waiting to be fixed (thanks #ViktorKerkez). s.index = pd.MultiIndex.from_tuples(s.index.map(lambda x: tuple(map(str.lower, x)))) You can replace str.lower with whatever function you want to use to rename your index. Note that you cannot use reindex in general here, because it tries to find values with the renamed index and thus it will return nan values, unless your rename results in no changes to the original index.
For the MultiIndexed Series objects, if this is not a bug, you can do: s.index = pd.MultiIndex.from_tuples( s.index.map(lambda x: tuple(map(str.lower, x))) )