-》Spark任务(RDD)-》运行
3:Spark Streaming:相当于Storm
本质是将连续的数据-》转换成不连续的数据DStream(离散流),本质还是... 在bigdata01启动:
sbin/start-all.sh 将master、worker全部启动
需要在...._2, false).collect
res1: Array[(String, Int)] = Array((hello,4), (spark,3), (hdoop,2), (hadoop,1), (...(1,2,3,4,5,6,7,8), 3)
rdd1: org.apache.spark.rdd.RDD[Int] = ParallelCollectionRDD[0..., 3, 8)
scala> val c = a.zip(b)
c: org.apache.spark.rdd.RDD