从pyspark导入SparkContext sc = SparkContext("local", "count app") words = sc.parallelize( ["scala", "java", "hadoop", "spark", "akka", "spark vs hadoop", "pyspark", "pyspark and spark" ]) counts = words.count() print("RDD中元素数量 -> %i" % counts)

翻译一下from pyspark import SparkContextsc = SparkContextlocal count appwords = scparallelize scala java hadoop spark akka spark vs hadoop pyspark pyspark and spark coun

原文地址: https://www.cveoy.top/t/topic/fEvS 著作权归作者所有。请勿转载和采集!

免费AI点我,无需注册和登录