---------------------------------------------------------------------------
ValueError Traceback (most recent call last)
<ipython-input-1-8cb1e1c08f01> in <module>()
7 # create Spark context with Spark configuration
8 conf = SparkConf().setAppName("Spark Count")
----> 9 sc = SparkContext(conf=conf)
10
11 # get threshold
/usr/local/spark/python/pyspark/context.py in __init__(self, master, appName, sparkHome, pyFiles, environment, batchSize, serializer, conf, gateway, jsc, profiler_cls)
110 """
111 self._callsite = first_spark_call() or CallSite(None, None, None)
--> 112 SparkContext._ensure_initialized(self, gateway=gateway)
113 try:
114 self._do_init(master, appName, sparkHome, pyFiles, environment, batchSize, serializer,
/usr/local/spark/python/pyspark/context.py in _ensure_initialized(cls, instance, gateway)
257 " created by %s at %s:%s "
258 % (currentAppName, currentMaster,
--> 259 callsite.function, callsite.file, callsite.linenum))
260 else:
261 SparkContext._active_spark_context = instance
ValueError: Cannot run multiple SparkContexts at once; existing SparkContext(app=PySparkShell, master=local[*]) created by <module> at /Applications/anaconda/lib/python3.5/site-packages/IPython/utils/py3compat.py:186