---------------------------------------------------------------------------
ValueError Traceback (most recent call last)
<ipython-input-2-c7badb8b399c> in <module>()
2 ## set up spark context
3 conf = SparkConf().setAppName("myApp")
----> 4 sc = SparkContext(conf=conf)
5
6 # create sparksession object
/usr/local/Cellar/apache-spark/2.1.0/libexec/python/pyspark/context.pyc in __init__(self, master, appName, sparkHome, pyFiles, environment, batchSize, serializer, conf, gateway, jsc, profiler_cls)
113 """
114 self._callsite = first_spark_call() or CallSite(None, None, None)
--> 115 SparkContext._ensure_initialized(self, gateway=gateway, conf=conf)
116 try:
117 self._do_init(master, appName, sparkHome, pyFiles, environment, batchSize, serializer,
/usr/local/Cellar/apache-spark/2.1.0/libexec/python/pyspark/context.pyc in _ensure_initialized(cls, instance, gateway, conf)
270 " created by %s at %s:%s "
271 % (currentAppName, currentMaster,
--> 272 callsite.function, callsite.file, callsite.linenum))
273 else:
274 SparkContext._active_spark_context = instance
ValueError: Cannot run multiple SparkContexts at once; existing SparkContext(app=PySparkShell, master=local[*]) created by <module> at /Users/mingchen/anaconda2/lib/python2.7/site-packages/IPython/utils/py3compat.py:288