Warning: file_get_contents(/data/phpspider/zhask/data//catemap/2/python/307.json): failed to open stream: No such file or directory in /data/phpspider/zhask/libs/function.php on line 167

Warning: Invalid argument supplied for foreach() in /data/phpspider/zhask/libs/tag.function.php on line 1116

Notice: Undefined index: in /data/phpspider/zhask/libs/function.php on line 180

Warning: array_chunk() expects parameter 1 to be array, null given in /data/phpspider/zhask/libs/function.php on line 181
python/dataframe-groupby列创建列表_Python_Apache Spark Sql - Fatal编程技术网

python/dataframe-groupby列创建列表

python/dataframe-groupby列创建列表,python,apache-spark-sql,Python,Apache Spark Sql,我正在尝试使用dataframe转换数据: Column 1 | Column 2 | Column 3 | Column 4 -------------------------------- A | B | 1 | # A | B | 2 | $ A | C | 1 | % 进入 有人能帮忙吗?这里有一些非常接近的东西(您的示例输出不是有效的python,除非它只是一个字符串): Column 1 | Column 2 | Column 5 -------

我正在尝试使用dataframe转换数据:

Column 1 | Column 2 | Column 3 | Column 4
--------------------------------
A    | B   | 1 | # 
A    | B   | 2 | $ 
A    | C   | 1 | % 
进入


有人能帮忙吗?

这里有一些非常接近的东西(您的示例输出不是有效的python,除非它只是一个字符串):

Column 1 | Column 2 | Column 5
--------------------------------
A    | B   | [Column 3: 1, Column 4:#, Column 3:2, Column 4:$] 
A    | C   | [Column 3: 1, Column 4:%] 
In []:
(df.groupby(['Column 1', 'Column 2'])[['Column 3', 'Column 4']]
   .apply(lambda x: x.to_dict(orient='record'))
   .reset_index(name='Column 5'))

Out[]:
  Column 1 Column 2                                                             Column 5
0        A        B  [{'Column 3': 1, 'Column4': '#'}, {'Column 3': 2, 'Column 4': '$'}]
1        A        C                                   [{'Column 3': 1, 'Column 4': '%'}]