Exemplo n.º 1
0
 def setUpClass(cls):
     ReusedPySparkTestCase.setUpClass()
     cls.tempdir = tempfile.NamedTemporaryFile(delete=False)
     os.unlink(cls.tempdir.name)
     cls.sqlCtx = SQLContext(cls.sc)
     cls.testData = [Row(key=i, value=str(i)) for i in range(100)]
     rdd = cls.sc.parallelize(cls.testData)
     cls.df = rdd.toDF()
Exemplo n.º 2
0
 def setUpClass(cls):
     ReusedPySparkTestCase.setUpClass()
     cls.tempdir = tempfile.NamedTemporaryFile(delete=False)
     os.unlink(cls.tempdir.name)
     print "type", type(cls.sc)
     print "type", type(cls.sc._jsc)
     _scala_HiveContext =\
         cls.sc._jvm.org.apache.spark.sql.hive.test.TestHiveContext(cls.sc._jsc.sc())
     cls.sqlCtx = HiveContext(cls.sc, _scala_HiveContext)
     cls.testData = [Row(key=i, value=str(i)) for i in range(100)]
     cls.df = cls.sc.parallelize(cls.testData).toDF()
Exemplo n.º 3
0
 def setUpClass(cls):
     ReusedPySparkTestCase.setUpClass()
     cls.tempdir = tempfile.NamedTemporaryFile(delete=False)
     try:
         cls.sc._jvm.org.apache.hadoop.hive.conf.HiveConf()
     except py4j.protocol.Py4JError:
         cls.sqlCtx = None
         return
     os.unlink(cls.tempdir.name)
     _scala_HiveContext = cls.sc._jvm.org.apache.spark.sql.hive.test.TestHiveContext(cls.sc._jsc.sc())
     cls.sqlCtx = HiveContext(cls.sc, _scala_HiveContext)
     cls.testData = [Row(key=i, value=str(i)) for i in range(100)]
     cls.df = cls.sc.parallelize(cls.testData).toDF()
Exemplo n.º 4
0
 def setUpClass(cls):
     ReusedPySparkTestCase.setUpClass()
     cls.tempdir = tempfile.NamedTemporaryFile(delete=False)
     try:
         cls.sc._jvm.org.apache.hadoop.hive.conf.HiveConf()
     except py4j.protocol.Py4JError:
         cls.tearDownClass()
         raise unittest.SkipTest("Hive is not available")
     except TypeError:
         cls.tearDownClass()
         raise unittest.SkipTest("Hive is not available")
     os.unlink(cls.tempdir.name)
     _scala_HiveContext =\
         cls.sc._jvm.org.apache.spark.sql.hive.test.TestHiveContext(cls.sc._jsc.sc())
     cls.sqlCtx = HiveContext(cls.sc, _scala_HiveContext)
     cls.testData = [Row(key=i, value=str(i)) for i in range(100)]
     cls.df = cls.sc.parallelize(cls.testData).toDF()
Exemplo n.º 5
0
 def tearDownClass(cls):
     ReusedPySparkTestCase.tearDownClass()
     shutil.rmtree(cls.tempdir.name, ignore_errors=True)
Exemplo n.º 6
0
 def setUpClass(cls):
     ReusedPySparkTestCase.setUpClass()
     cls.tempdir = tempfile.mkdtemp()
Exemplo n.º 7
0
 def tearDownClass(cls):
     ReusedPySparkTestCase.tearDownClass()
     shutil.rmtree(cls.tempdir)
Exemplo n.º 8
0
 def setUpClass(cls):
     ReusedPySparkTestCase.setUpClass()
     cls.spark = SparkSession(cls.sc)
     cls.tempdir = tempfile.mkdtemp()
Exemplo n.º 9
0
 def tearDownClass(cls):
     ReusedPySparkTestCase.tearDownClass()
     cls.spark.stop()
Exemplo n.º 10
0
 def setUpClass(cls):
     ReusedPySparkTestCase.setUpClass()
     cls.spark = SparkSession(cls.sc)
Exemplo n.º 11
0
 def tearDownClass(cls):
     PySparkTestCase.tearDownClass()
     cls.spark.stop()
Exemplo n.º 12
0
 def setUpClass(cls):
     PySparkTestCase.setUpClass()
     cls.spark = SparkSession(cls.sc)