Warning: file_get_contents(/data/phpspider/zhask/data//catemap/2/python/310.json): failed to open stream: No such file or directory in /data/phpspider/zhask/libs/function.php on line 167

Warning: Invalid argument supplied for foreach() in /data/phpspider/zhask/libs/tag.function.php on line 1116

Notice: Undefined index: in /data/phpspider/zhask/libs/function.php on line 180

Warning: array_chunk() expects parameter 1 to be array, null given in /data/phpspider/zhask/libs/function.php on line 181
Python 按级别分组列,按其他级别分组列_Python_Pandas_Dataframe_Group By - Fatal编程技术网

Python 按级别分组列,按其他级别分组列

Python 按级别分组列,按其他级别分组列,python,pandas,dataframe,group-by,Python,Pandas,Dataframe,Group By,我每小时采集一次天气数据,其中包含[温度、湿度、速度] Timestamp Humidity Temp Speed 01/01/2019 00:00 57 23 2.222222222 01/01/2019 01:00 56 23 1.944444444 01/01/2019 02:00 55 23 1.944444444 01/01/2019 03:00 54 22 1.944444444 01/01/2019 04:00 55

我每小时采集一次天气数据,其中包含[温度、湿度、速度]

Timestamp      Humidity Temp    Speed
01/01/2019 00:00    57  23  2.222222222
01/01/2019 01:00    56  23  1.944444444
01/01/2019 02:00    55  23  1.944444444
01/01/2019 03:00    54  22  1.944444444
01/01/2019 04:00    55  22  1.944444444
01/01/2019 05:00    56  22  1.666666667
01/01/2019 06:00    57  22  1.666666667
01/01/2019 07:00    57  22  1.666666667
01/01/2019 08:00    57  23  1.944444444
01/01/2019 09:00    57  23  1.944444444
01/01/2019 10:00    55  23  2.222222222
我想创建一个包含以下内容的数据帧:

时间戳(每日)。时间戳(每小时)。温度(每小时)、湿度(每小时)、湿度(每小时)、温度(每天)、湿度(每天)、湿度(每天)

最快的方法是什么[例如使用一组人]

低于我的努力

        weather_data = weather_data.groupby(pd.Grouper(key='Timestamp', freq='1d'),pd.Grouper(key='Timestamp', freq='1d'))\
            .agg(Temp_Daily = ('Temp','mean'),
                 Humidity_Daily=('Humidity', 'mean'),
                 Speed_Daily=('Speed', 'mean'),
                 Temp = ('Temp', lambda x:x),
                 Humidity=('Humidity', lambda x: x),
                 Speed=('Speed', lambda x: x)
                 ).reset_index()

如果我正确理解您的预期输出,您可以将每日平均值连接到原始数据帧(更改列名后)


由于您在不同的频率上分组,我不认为只有一个groupby命令就有一种直接的方法。
# Sample data.
df = pd.DataFrame({
    'Timestamp': pd.date_range('2019-01-01 00:00', '2019-01-01 10:00', freq='H'),
    'Humidity': [57, 56, 55, 54, 55, 56, 57, 57, 57, 57, 55],
    'Temp': [23, 23, 23, 22, 22, 22, 22, 22, 23, 23, 23],
    'Speed': [2.222222222, 1.944444444, 1.944444444, 1.944444444, 1.944444444, 1.666666667, 1.666666667, 1.666666667, 1.944444444, 1.944444444, 2.222222222]
})


# Solution.
df_daily = (
    df
    .groupby(df['Timestamp'].dt.date)[['Humidity', 'Temp', 'Speed']]
    .transform('mean')
    .add_suffix(' (daily)')
    .set_index(df['Timestamp'])
)
result = pd.concat([df.set_index('Timestamp').add_suffix(' (Hourly)'), df_daily], axis=1)

>>> result.shape
(11, 6)

>>> result
                     Humidity (Hourly)  Temp (Hourly)  Speed (Hourly)  \
Timestamp                                                               
2019-01-01 00:00:00                 57             23        2.222222   
2019-01-01 01:00:00                 56             23        1.944444   
2019-01-01 02:00:00                 55             23        1.944444   
2019-01-01 03:00:00                 54             22        1.944444   
2019-01-01 04:00:00                 55             22        1.944444   
2019-01-01 05:00:00                 56             22        1.666667   
2019-01-01 06:00:00                 57             22        1.666667   
2019-01-01 07:00:00                 57             22        1.666667   
2019-01-01 08:00:00                 57             23        1.944444   
2019-01-01 09:00:00                 57             23        1.944444   
2019-01-01 10:00:00                 55             23        2.222222   

                     Humidity (daily)  Temp (daily)  Speed (daily)  
Timestamp                                                           
2019-01-01 00:00:00                56     22.545455       1.919192  
2019-01-01 01:00:00                56     22.545455       1.919192  
2019-01-01 02:00:00                56     22.545455       1.919192  
2019-01-01 03:00:00                56     22.545455       1.919192  
2019-01-01 04:00:00                56     22.545455       1.919192  
2019-01-01 05:00:00                56     22.545455       1.919192  
2019-01-01 06:00:00                56     22.545455       1.919192  
2019-01-01 07:00:00                56     22.545455       1.919192  
2019-01-01 08:00:00                56     22.545455       1.919192  
2019-01-01 09:00:00                56     22.545455       1.919192  
2019-01-01 10:00:00                56     22.545455       1.919192