使用scala将单行中的用户信息统一为数据帧

使用scala将单行中的用户信息统一为数据帧,scala,dataframe,apache-spark,Scala,Dataframe,Apache Spark,我有相同的合同和COD,但该客户有不同的日期和金额以及 我想将它们分组并在数据框中保留两行,我已经添加了相关列 到TYPCOD和DATE字段,以便以后我只能停留在数据框中的两行 这样就不会丢失信息 有可能吗? 预期: dfFilter.show() ------------+-----------+-----------+------------+--------+ CONTR |COD | DATE |TYPCOD | Amount | ----

我有相同的合同COD,但该客户有不同的日期和金额以及 我想将它们分组并在数据框中保留两行,我已经添加了相关列 到TYPCODDATE字段,以便以后我只能停留在数据框中的两行 这样就不会丢失信息

有可能吗? 预期:

dfFilter.show()

------------+-----------+-----------+------------+--------+
CONTR       |COD        | DATE      |TYPCOD      | Amount |
------------+-----------+-----------+------------+--------+
0004        |4433       |2006-11-04 |RMA         | 150.0  |
0004        |4433       |2012-05-14 |FCB         | 300.0  |
0004        |1122       |2011-10-17 |RMA         | 100.0  |
0004        |1122       |2015-12-05 |FCB         | 500.0  |
------------+-----------+-----------+------------+--------+
// 

val addColumn = dfFilter.withColumn("RMA_AMOUNT", when(col("TYPCOD")==="RMA", col("Amount")))
                        .withColumn("DATE_RMA", when(col("TYPCOD")==="RMA", col("DATE")))
                        .withColumn("FCB_AMOUNT", when(col("TYPCOD")==="FCB", col("Amount")))
                        .withColumn("DATE_FCB", when(col("TYPCOD")==="FCB", col("DATE")))
addColumn.show()

--------+-----------+-----------+------------+--------+------------+-----------+-----------+-----------+
CONTR   |COD        | DATE      |TYPCOD      | Amount | RMA_AMOUNT |DATE_RMA   |FCB_AMOUNT |DATE_FCB   |
--------+-----------+-----------+------------+--------+------------+-----------+-----------+-----------+
0004    |4433       |2006-11-04 |RMA         | 150.0  |150.0       |2006-11-04 |null       |null       |
0004    |4433       |2012-05-14 |FCB         | 300.0  |null        |null       |300.0      |2012-05-14 |
0004    |1122       |2011-10-17 |RMA         | 100.0  |100.0       |2011-10-17 |null       |null       |
0004    |1122       |2015-12-05 |FCB         | 500.0  |null        |null       |500.0      |2015-12-05 |
--------+-----------+------------+-----------+--------+------------+-----------+-----------+-----------+

是的,有可能。请检查下面的代码

------------+-------------+------------+-----------+-----------+-----------+
CONTR       |COD          | RMA_AMOUNT |DATE_RMA   |FCB_AMOUNT |DATE_FCB   |
------------+-------------+------------+-----------+-----------+-----------+
0004        |4433         |150.0       |2006-11-04 |300.0      |2012-05-14 |
0004        |1122         |100.0       |2011-10-17 |500.0      |2015-12-05 |
------------+-------------+------------+-----------+-----------+-----------+


在这种情况下,先使用
groupBy
,然后使用
函数(col,ignoreNull=true)

scala> val df = Seq(("0004",4433,"2006-11-04","RMA",150.0),("0004",4433,"2012-05-14","FCB",300.0),("0004",1122,"2011-10-17","RMA",100.0),("0004",1122,"2015-12-05","FCB",500.0)).toDF("contr","cod","date","typcod","amount")
df: org.apache.spark.sql.DataFrame = [contr: string, cod: int ... 3 more fields]

scala> val rma = df.filter($"typcod" === "RMA").select($"contr",$"cod",$"date".as("rma_date"),$"typcod",$"amount".as("rma_amount"))
rma: org.apache.spark.sql.DataFrame = [contr: string, cod: int ... 3 more fields]

scala> rma.show(false)
+-----+----+----------+------+----------+
|contr|cod |rma_date  |typcod|rma_amount|
+-----+----+----------+------+----------+
|0004 |4433|2006-11-04|RMA   |150.0     |
|0004 |1122|2011-10-17|RMA   |100.0     |
+-----+----+----------+------+----------+


scala> val fcb = df.filter($"typcod" === "FCB").select($"contr",$"cod",$"date".as("fcb_date"),$"typcod",$"amount".as("fcb_amount")).drop("contr")
fcb: org.apache.spark.sql.DataFrame = [cod: int, fcb_date: string ... 2 more fields]

scala> fcb.show(false)
+----+----------+------+----------+
|cod |fcb_date  |typcod|fcb_amount|
+----+----------+------+----------+
|4433|2012-05-14|FCB   |300.0     |
|1122|2015-12-05|FCB   |500.0     |
+----+----------+------+----------+


scala> rma.join(fcb,Seq("cod"),"inner").select("contr","cod","rma_amount","rma_date","fcb_amount","fcb_date").show(false)
+-----+----+----------+----------+----------+----------+
|contr|cod |rma_amount|rma_date  |fcb_amount|fcb_date  |
+-----+----+----------+----------+----------+----------+
|0004 |4433|150.0     |2006-11-04|300.0     |2012-05-14|
|0004 |1122|100.0     |2011-10-17|500.0     |2015-12-05|
+-----+----+----------+----------+----------+----------+


scala>


TYPCOD的值总是两个或更多?可以有两个或两个以上的值任何人都不知道您的数据量有多少?我有5种类型的cod,对于每个“contr”和“cod”,我都关联了一个日期和一个金额。示例:用户可以在不同日期执行金额的操作,如果用户执行5个操作,他将有5行具有相同操作类型(“typecod”),但日期和金额不同。我的数据量约为450MB他的输出有两个rows@Srinivas,我没有在输入数据框中添加2行!哦我没有检查。谢谢
val df=Seq(("0004","4433","2006-11-04","RMA","150.0","150.0","2006-11-04",null.asInstanceOf[String],null.asInstanceOf[String]),("0004","4433","2012-05-14","FCB","300.0",null.asInstanceOf[String],null.asInstanceOf[String],"300.0","2012-05-14"),("0004","1122","2011-10-17","RMA","100.0","100.0","2011-10-17",null.asInstanceOf[String],null.asInstanceOf[String]),("0004","1122","2015-12-05","FCB","500.0",null.asInstanceOf[String],null.asInstanceOf[String],"500.0","2015-12-05")).toDF("CONTR","COD","DATE","TYPCOD","Amount","RMA_AMOUNT","DATE_RMA","FCB_AMOUNT","DATE_FCB")

//+-----+----+----------+------+------+----------+----------+----------+----------+
//|CONTR| COD|      DATE|TYPCOD|Amount|RMA_AMOUNT|  DATE_RMA|FCB_AMOUNT|  DATE_FCB|
//+-----+----+----------+------+------+----------+----------+----------+----------+
//| 0004|4433|2006-11-04|   RMA| 150.0|     150.0|2006-11-04|      null|      null|
//| 0004|4433|2012-05-14|   FCB| 300.0|      null|      null|     300.0|2012-05-14|
//| 0004|1122|2011-10-17|   RMA| 100.0|     100.0|2011-10-17|      null|      null|
//| 0004|1122|2015-12-05|   FCB| 500.0|      null|      null|     500.0|2015-12-05|
//+-----+----+----------+------+------+----------+----------+----------+----------+

df.groupBy("CONTR","COD").agg(first(col("RMA_AMOUNT"),true).alias("RMA_AMOUNT"),first(col("DATE_RMA"),true).alias("DATE_RMA"),first(col("FCB_AMOUNT"),true).alias("FCB_AMOUNT"),first(col("DATE_FCB"),true).alias("DATE_FCB")).show()
//+-----+----+----------+----------+----------+----------+
//|CONTR| COD|RMA_AMOUNT|  DATE_RMA|FCB_AMOUNT|  DATE_FCB|
//+-----+----+----------+----------+----------+----------+
//| 0004|4433|     150.0|2006-11-04|     300.0|2012-05-14|
//| 0004|1122|     100.0|2011-10-17|     500.0|2015-12-05|
//+-----+----+----------+----------+----------+----------+

//incase if you want to keep TYPCOD and DATE values 
df.groupBy("CONTR","COD").agg(concat_ws(",",collect_list(col("TYPCOD"))).alias("TYPECOD"),concat_ws(",",collect_list(col("DATE"))).alias("DATE"),first(col("RMA_AMOUNT"),true).alias("RMA_AMOUNT"),first(col("DATE_RMA"),true).alias("DATE_RMA"),first(col("FCB_AMOUNT"),true).alias("FCB_AMOUNT"),first(col("DATE_FCB"),true).alias("DATE_FCB")).show(false)
//+-----+----+-------+---------------------+----------+----------+----------+----------+
//|CONTR|COD |TYPECOD|DATE                 |RMA_AMOUNT|DATE_RMA  |FCB_AMOUNT|DATE_FCB  |
//+-----+----+-------+---------------------+----------+----------+----------+----------+
//|0004 |4433|RMA,FCB|2006-11-04,2012-05-14|150.0     |2006-11-04|300.0     |2012-05-14|
//|0004 |1122|RMA,FCB|2011-10-17,2015-12-05|100.0     |2011-10-17|500.0     |2015-12-05|
//+-----+----+-------+---------------------+----------+----------+----------+----------+