map
是对RDD中的每个元素都执行一个指定的函数来产生一个新的RDD。任何原RDD中的元素在新RDD中都有且只有一个元素与之对应
val conf = new SparkConf()
conf.setAppName("count")
conf.setMaster("local")
val sc = new SparkContext(conf)
val a = sc.parallelize(1 to 9, 3)
# x =>*2是一个函数,x是传入参数即RDD的每个元素,x*2是返回值
val b = a.map(x => x*2)
a.collect
# 结果Array[Int] = Array(1, 2, 3, 4, 5, 6, 7, 8, 9)
b.collect
# 结果Array[Int] = Array(2, 4, 6, 8, 10, 12, 14, 16, 18)
list/key--->key-value
val conf = new SparkConf()
conf.setAppName("count")
conf.setMaster("local")
val sc = new SparkContext(conf)
val a = sc.parallelize(List("dog", "tiger", "lion", "cat", "panther", " eagle"), 2)
val b = a.map(x => (x, 1))
b.collect.foreach(println(_))
# /*
# (dog,1)
# (tiger,1)
# (lion,1)
# (cat,1)
# (panther,1)
# ( eagle,1)
# */
val conf = new SparkConf()
conf.setAppName("count")
conf.setMaster("local")
val sc = new SparkContext(conf)
val l=sc.parallelize(List((1,'a'),(2,'b')))
var ll=l.map(x=>(x._1,"PV:"+x._2)).collect()
ll.foreach(println)
# (1,PVa)
#(2,PVb)
mapValues(function)
原RDD中的Key保持不变,与新的Value一起组成新的RDD中的元素。因此,该函数只适用于元素为KV对的RDD
val conf = new SparkConf()
conf.setAppName("count")
conf.setMaster("local")
val sc = new SparkContext(conf)
val a = sc.parallelize(List("dog", "tiger", "lion", "cat", "panther", " eagle"), 2)
/*dog tiger lion cat panther eagle*/
val b = a.map(x => (x.length, x))
/*(3,dog) (5,tiger) (4,lion) (3,cat) (7,panther) (6, eagle)*/
b.mapValues("x" + _ + "x").collect.foreach(println(_))
/*(3,xdogx)
(5,xtigerx)
(4,xlionx)
(3,xcatx)
(7,xpantherx)
(6,x eaglex)*/
mapPartitions(function)
map()的输入函数是应用于RDD中每个元素,而mapPartitions()的输入函数是应用于每个分区 package
import scala.Iterator
import org.apache.spark.SparkConf
import org.apache.spark.SparkContext
object TestRdd {
def sumOfEveryPartition(input: Iterator[Int]): Int = {
var total = 0
input.foreach {
elem =>
total += elem
}
total
}
def main(args: Array[String]) {
val conf = new SparkConf().setAppName("Spark Rdd Test")
val spark = new SparkContext(conf)
val input = spark.parallelize(List(1, 2, 3, 4, 5, 6), 2)//RDD有6个元素,分成2个partition
val result = input.mapPartitions(
partition => Iterator(sumOfEveryPartition(partition)))//partition是传入的参数,是个list,要求返回也是list,即Iterator(sumOfEveryPartition(partition))
result.collect().foreach {
println(_)
# 6 15,分区计算和
}
spark.stop()
}
}
flatMap(function)
与map类似,区别是原RDD中的元素经map处理后只能生成一个元素,而原RDD中的元素经flatmap处理后可生成多个元素
val conf = new SparkConf()
conf.setAppName("count")
conf.setMaster("local")
val sc = new SparkContext(conf)
val a = sc.parallelize(1 to 4, 2)
val b = a.flatMap(x => 1 to x)//每个元素扩展
b.collect
/*
结果 Array[Int] = Array( 1,
1, 2,
1, 2, 3,
1, 2, 3, 4)
*/
flatMapValues(function)
val conf = new SparkConf()
conf.setAppName("count")
conf.setMaster("local")
val sc = new SparkContext(conf)
val a = sc.parallelize(List((1,2),(3,4),(5,6)))
val b = a.flatMapValues(x=>1 to x)
b.collect.foreach(println(_))
/*结果
(1,1)
(1,2)
(3,1)
(3,2)
(3,3)
(3,4)
(5,1)
(5,2)
(5,3)
(5,4)
(5,5)
(5,6)
*/
val list = List(("mobin",22),("kpop",20),("lufei",23))
val rdd = sc.parallelize(list)
val mapValuesRDD = rdd.flatMapValues(x => Seq(x,"male"))
mapValuesRDD.foreach(println)
输出:
(mobin,22)
(mobin,male)
(kpop,20)
(kpop,male)
(lufei,23)
(lufei,male)
如果是mapValues会输出:对比区别】
(mobin,List(22, male))
(kpop,List(20, male))
(lufei,List(23, male))
reduceByKey(func,numPartitions):
按Key进行分组,使用给定的func函数聚合value值, numPartitions设置分区数,提高作业并行度
val conf = new SparkConf()
conf.setAppName("count")
conf.setMaster("local")
val sc = new SparkContext(conf)
val arr = List(("A",3),("A",2),("B",1),("B",3))
val rdd = sc.parallelize(arr)
val reduceByKeyRDD = rdd.reduceByKey(_ +_)
reduceByKeyRDD.foreach(println)
sc.stop
# (A,5)
# (B,4)
cogroup(otherDataSet,numPartitions):
对两个RDD(如:(K,V)和(K,W))相同Key的元素先分别做聚合,最后返回(K,Iterator,Iterator)形式的RDD,numPartitions设置分区数,提高作业并行度
val conf = new SparkConf()
conf.setAppName("count")
conf.setMaster("local")
val sc = new SparkContext(conf)
val arr = List(("A", 1), ("B", 2), ("A", 2), ("B", 3))
val arr1 = List(("A", "A1"), ("B", "B1"), ("A", "A2"), ("B", "B2"))
val rdd1 = sc.parallelize(arr, 3)
val rdd2 = sc.parallelize(arr1, 3)
val groupByKeyRDD = rdd1.cogroup(rdd2)
groupByKeyRDD.foreach(println)
sc.stop
# (B,(CompactBuffer(2, 3),CompactBuffer(B1, B2)))
# (A,(CompactBuffer(1, 2),CompactBuffer(A1, A2)))
join(otherDataSet,numPartitions):
对两个RDD先进行cogroup操作形成新的RDD,再对每个Key下的元素进行笛卡尔积,numPartitions设置分区数,提高作业并行度
val conf = new SparkConf()
conf.setAppName("count")
conf.setMaster("local")
val sc = new SparkContext(conf)
val arr = List(("A", 1), ("B", 2), ("A", 2), ("B", 3))
val arr1 = List(("A", "A1"), ("B", "B1"), ("A", "A2"), ("B", "B2"))
val rdd = sc.parallelize(arr, 3)
val rdd1 = sc.parallelize(arr1, 3)
val groupByKeyRDD = rdd.join(rdd1)
groupByKeyRDD.foreach(println)
# (B,(2,B1))
# (B,(2,B2))
# (B,(3,B1))
# (B,(3,B2))
# (A,(1,A1))
# (A,(1,A2))
# (A,(2,A1))
# (A,(2,A2))
LeftOutJoin(otherDataSet,numPartitions):
左外连接,包含左RDD的所有数据,如果右边没有与之匹配的用None表示,numPartitions设置分区数,提高作业并行度/
val conf = new SparkConf()
conf.setAppName("count")
conf.setMaster("local")
val sc = new SparkContext(conf)
val arr = List(("A", 1), ("B", 2), ("A", 2), ("B", 3),("C",1))
val arr1 = List(("A", "A1"), ("B", "B1"), ("A", "A2"), ("B", "B2"))
val rdd = sc.parallelize(arr, 3)
val rdd1 = sc.parallelize(arr1, 3)
val leftOutJoinRDD = rdd.leftOuterJoin(rdd1) leftOutJoinRDD .foreach(println)
sc.stop
# (B,(2,Some(B1)))
# (B,(2,Some(B2)))
# (B,(3,Some(B1)))
# (B,(3,Some(B2)))
# (C,(1,None))
# (A,(1,Some(A1)))
# (A,(1,Some(A2)))
# (A,(2,Some(A1)))
# (A,(2,Some(A2)))
RightOutJoin(otherDataSet, numPartitions):
右外连接,包含右RDD的所有数据,如果左边没有与之匹配的用None表示,numPartitions设置分区数,提高作业并行度
val conf = new SparkConf()
conf.setAppName("count")
conf.setMaster("local")
val sc = new SparkContext(conf)
val arr = List(("A", 1), ("B", 2), ("A", 2), ("B", 3))
val arr1 = List(("A", "A1"), ("B", "B1"), ("A", "A2"), ("B", "B2"),("C","C1"))
val rdd = sc.parallelize(arr, 3)
val rdd1 = sc.parallelize(arr1, 3)
val rightOutJoinRDD = rdd.rightOuterJoin(rdd1) rightOutJoinRDD.foreach(println)
sc.stop
# (B,(Some(2),B1))
# (B,(Some(2),B2))
# (B,(Some(3),B1))
# (B,(Some(3),B2))
# (C,(None,C1))
# (A,(Some(1),A1))
# (A,(Some(1),A2))
# (A,(Some(2),A1))
# (A,(Some(2),A2))
lookup
val conf = new SparkConf()
conf.setAppName("count")
conf.setMaster("local")
val sc = new SparkContext(conf)
var rdd1=sc.parallelize(List((1,"a"),(2,"b"),(3,"c")))
rdd1.lookup(2).foreach(println(_))
#b
filter
val conf = new SparkConf()
conf.setAppName("count")
conf.setMaster("local")
val sc = new SparkContext(conf)
//val filterRdd = sc.parallelize(List(1,2,3,4,5)).map(_*2).filter(_>5)
val filterRdd = sc.parallelize(List(1,2,3,4,5)).map(x=>{
x*2}).filter(x=>{
x>5})
filterRdd.collect.foreach(println(_))
#
6
8
10
#
groupByKey(numPartitions):
按Key进行分组,返回[K,Iterable[V]],numPartitions设置分区数,提高作业并行度【value并不是累加,而是变成一个数组】
val arr = List(("A",1),("B",2),("A",2),("B",3))
val rdd = sc.parallelize(arr)
val groupByKeyRDD = rdd.groupByKey()
groupByKeyRDD.foreach(println)
sc.stop
# (B,CompactBuffer(2, 3))
# (A,CompactBuffer(1, 2))
# 统计key后面的数组汇总元素的个数
scala> groupByKeyRDD.mapValues(x => x.size).foreach(println)
# (A,2)
# (B,2)
sortByKey
返回以Key排序的(K,V)键值对组成的RDD,accending为true时表示升序,为false时表示降序,numPartitions设置分区数,提高作业并行度。
val conf = new SparkConf()
conf.setAppName("count")
conf.setMaster("local")
val sc = new SparkContext(conf)
scval arr = List(("A",1),("B",2),("A",2),("B",3))
val rdd = sc.parallelize(arr)
val sortByKeyRDD = rdd.sortByKey()
sortByKeyRDD.foreach(println)
sc.stop
# (A,1)
# (A,2)
# (B,2)
# (B,3)