Spark rewrite sorting rules (4)

Tuple encapsulation sort condition

import org.apache.spark.{
    
    SparkConf, SparkContext}
import org.apache.spark.rdd.RDD

/*
元组封装排序条件
 */
object CustomSort_5  extends App {
    
    
  val conf = new SparkConf()
  conf.setAppName(this.getClass.getName).setMaster("local[2]")
  val sc = new SparkContext(conf)
  val userInfo: RDD[String]
          = sc.parallelize(Array("mimi1 21 85", "mimi2 22 87", "mimi3 23 87"))
  val personRDD: RDD[(String, Int, Int)] = userInfo.map(x => {
    
    
    val arr = x.split(" ")
    val name = arr(0)
    val age = arr(1).toInt
    val fv = arr(2).toInt
    (name, age, fv)
  })
  val sorted: RDD[(String, Int, Int)] = personRDD.sortBy(x => (-x._3,x._2))
  println(sorted.collect.toBuffer)
}

result

ArrayBuffer((mimi2,22,87), (mimi3,23,87), (mimi1,21,85))

Guess you like

Origin blog.csdn.net/qq_42706464/article/details/108355295