from pyspark.sql.functions import *
# 读入数据
df = spark.read.format("csv").option("header", "true").load("/path/to/file.csv")
# 根据条件计算exploded_qty和imploded_qty
df = df.withColumn("exploded_qty", when(col("PACK_IND")=="Y", col("QTY")*col("NUM_PACK")).otherwise(0))
df = df.withColumn("imploded_qty", when(col("PACK_IND")=="N", col("QTY")/col("NUM_PACK")).otherwise(0))
# 显示结果
df.show()