Spark DataFrame选取多列


val df = sc.parallelize(Seq(
  (0,"cat26",30.9), 
  (1,"cat67",28.5), 
  (2,"cat56",39.6),
  (3,"cat8",35.6))).toDF("Hour", "Category", "Value")

//或者从文件读取成List
val cols = List("Hour", "Value")

scala> df.select(cols.head, cols.tail: _*).show
+----+----------+
|Hour|Value|
+----+----------+
|   1|      28.5|
|   3|      35.6|
|   2|      39.6|
|   0|      30.9|
+----+----------+

 


免责声明!

本站转载的文章为个人学习借鉴使用,本站对版权不负任何法律责任。如果侵犯了您的隐私权益,请联系本站邮箱yoyou2525@163.com删除。



 
粤ICP备18138465号  © 2018-2025 CODEPRJ.COM