Hey there!
I was running SQL queries via PySpark (using the wmfdata package
<https://github.com/neilpquinn/wmfdata/blob/master/wmfdata/hive.py>) on
SWAP when one of my queries failed with "java.lang.OutofMemoryError: Java
heap space".
After that, when I tried to call the spark.sql function again (via
wmfdata.hive.run), it failed with "java.lang.IllegalStateException: Cannot
call methods on a stopped SparkContext."
When I tried to create a new Spark context using
SparkSession.builder.getOrCreate (whether using wmfdata.spark.get_session
or directly), it returned a SparkContent object properly, but calling the
object's sql function still gave the "stopped SparkContext error".
Any idea what's going on? I assume restarting the notebook kernel would
take care of the problem, but it seems like there has to be a better way to
recover.
Thank you!