Spark Rdd DataFrame操作彙總

直接上測試代碼:

import org.apache.spark.SparkConf
import org.apache.spark.sql.SparkSession
import org.apache.spark.sql.functions.{udf, _}
import org.junit.{Assert, Test}
import org.apache.spark.sql.types._



@Test
 class SparkRddTest extends  Assert{

  @Test
  def testWriteRdd(): Unit ={
    val sparkConf = new SparkConf().setAppName("SparkTestRdd");
    sparkConf.setMaster("local[*]");
    val ss = SparkSession.builder.config(sparkConf).getOrCreate()
    val seq = List("American Person", "Tom", "Jim")
    var rdd = ss.sparkContext.makeRDD(seq)
    rdd.saveAsTextFile("D:/SparkTest/rdd")
  }


  @Test
  def testWriteRdd1(): Unit ={
    val sparkConf = new SparkConf().setAppName("SparkTestRdd");
    sparkConf.setMaster("local[*]");
    val ss = SparkSession.builder.config(sparkConf).getOrCreate()
    val seq = List("American Person Test Spark Rdd, Today is nice  Day", "Tom", "Jim","American", "Tom", "Jim","American", "Tom", "Jim",
      "Test" , "Test", "Hello"
發表評論
所有評論
還沒有人評論,想成為第一個評論的人麼? 請在上方評論欄輸入並且點擊發布.
相關文章