首页 > 解决方案 > 将 spark scala 中的列拆分为数组类型列和字符串类型列

问题描述

所以我的数据看起来像这样

"my_data,another_data,"[0,1,3,2,4]","[0,1,2,3,4]""

我想要的是

["my_data","another_data","[0,1,3,2,4]","[0,1,2,3,4]"]

到目前为止我尝试过的是: -

  df.withColumn("valueSplit",split(col("value"),",\"|\","))

这给了我

["my_data,another_data","[0,1,2,3,2,4]"","[0,1,2,3,4]"]

谢谢是提前!

标签: scalaapache-spark

解决方案


推荐阅读