Dataframe 如何找到一行的分位数。在pyspark数据帧中?
我有下面的pyspark daframe,我想按行查找百分位Dataframe 如何找到一行的分位数。在pyspark数据帧中?,dataframe,pyspark,apache-spark-sql,Dataframe,Pyspark,Apache Spark Sql,我有下面的pyspark daframe,我想按行查找百分位 value col_a col_b col_c row_a 5.0 0.0 11.0 row_b 3394.0 0.0 4543.0 row_c 136111.0 0.0 219255.0 row_d 0.0 0.0 0.0 row_e 0.0 0.0 0.0 row_f
value col_a col_b col_c
row_a 5.0 0.0 11.0
row_b 3394.0 0.0 4543.0
row_c 136111.0 0.0 219255.0
row_d 0.0 0.0 0.0
row_e 0.0 0.0 0.0
row_f 42.0 0.0 54.0
对于每个百分位数,我想在主数据框中添加一个新列,如下所示
value col_a col_b col_c 25%
row_a 5.0 0.0 11.0 2.5
row_b 3394.0 0.0 4543.0 1697
row_c 136111.0 0.0 219255.0 68055
row_d 0.0 0.0 0.0 0.0
row_e 0.0 0.0 0.0 0.0
row_f 42.0 0.0 54.0 21.0
df['25%']=df.quantile(0.25,axis=1)
在熊猫方面,我做了如下的工作
value col_a col_b col_c 25%
row_a 5.0 0.0 11.0 2.5
row_b 3394.0 0.0 4543.0 1697
row_c 136111.0 0.0 219255.0 68055
row_d 0.0 0.0 0.0 0.0
row_e 0.0 0.0 0.0 0.0
row_f 42.0 0.0 54.0 21.0
df['25%']=df.quantile(0.25,axis=1)