SparkSession AttributeError: 'sqlContext' not found - Solution & Example
This error occurs when you try to access the 'sqlContext' attribute of a 'SparkSession' object, but it does not exist.
In Apache Spark 2.0 and later versions, the 'sqlContext' attribute has been replaced by the 'spark' attribute. Therefore, you should use 'spark' instead of 'sqlContext' to access the SQL context in the 'SparkSession' object.
For example:
from pyspark.sql import SparkSession
spark = SparkSession.builder.appName('myApp').getOrCreate()
df = spark.read.csv('path/to/file.csv', header=True)
# use spark for SQL queries
df.createOrReplaceTempView('myTable')
result = spark.sql('SELECT * FROM myTable')
In this example, we create a 'SparkSession' object and use the 'spark' attribute to access the SQL context and run SQL queries.
原文地址: https://www.cveoy.top/t/topic/oYJe 著作权归作者所有。请勿转载和采集!