Spark函数讲解序列文章

aggregateaggregateByKeycachecartesiancheckpointcoalescecogroup 

groupWith

collect, toArray
collectAsMap
combineByKey
compute
context, sparkContext
count
countApprox
countByKey
countByKeyApprox
countByValue
countByValueApprox
countApproxDistinct
countApproxDistinctByKey
dependencies
distinct
first
filter
filterWith
flatMap
flatMapValues
flatMapWith
fold
foldByKey
foreach
foreachPartition
foreachWith
generator, setGenerator
getCheckpointFile
preferredLocations
getStorageLevel
glom
groupBy
groupByKey
histogram
id
intersection
isCheckpointed
iterator
join
keyBy
keys
leftOuterJoin
lookup
map
mapPartitions
mapPartitionsWithContext
mapPartitionsWithIndex
mapPartitionsWithSplit
mapValues
mapWith
max
mean , meanApprox
min
name, setName
partitionBy
partitioner
partitions
persist, cache
pipe
randomSplit
reduce
reduceByKey, reduceByKeyLocally, reduceByKeyToDriver
rightOuterJoin
sample
saveAsHodoopFile, saveAsHadoopDataset, saveAsNewAPIHadoopFile
saveAsObjectFile
saveAsSequenceFile
saveAsTextFile
stats
sortBy
sortByKey
stdev , sampleStdev
subtract
subtractByKey
sum , sumApprox
take
takeOrdered
takeSample
toDebugString
toJavaRDD
top
toString
union, ++
unpersist
values
variance , sampleVariance
zip
zipPartitions
zipWithIndex
zipWithUniquId

你可能感兴趣的:(Big,Data)