pyspark操作 rdd dataframe,pyspark.sql.functions详解 行列变换
官网文档可以参考:https://spark.apache.org/docs/latest/api/python/index.htmldataframe读写生成以逗号分隔的数据stringCSVRDD=spark.sparkContext.parallelize([(123,"Katie",19,"brown"),(234,"Michael",22,"green"),(345,"Simone",2