尝试从源代码运行http://spark.apache.org/docs/latest/quick-start.html#a-standalone-app-in-scala。
这一行:
val wordCounts = textFile.flatMap(line => line.split(" ")).map(word => (word, 1)).reduceByKey((a, b) => a + b)
出现错误
value reduceByKey is not a member of org.apache.spark.rdd.RDD[(String, Int)]
val wordCounts = logData.flatMap(line => line.split(" ")).map(word => (word, 1)).reduceByKey((a, b) => a + b)
logData.flatMap(line => line.split(" ")).map(word => (word, 1))
返回一个MappedRDD,但是我在Spark RDD API文档中找不到这个类型。 我正在从Spark源代码运行此代码,所以可能是类路径问题吗?但是所需的依赖项已经在我的类路径中了。
import org.apache.spark.SparkSession._
。 - WestCoastProjects