Skip to content

Sandy4321/spark-tdd-example

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

6 Commits
 
 
 
 
 
 
 
 

Repository files navigation

A simple PySpark example using TDD

This is a very basic example of how to use Test Driven Development (TDD) in the context of PySpark, Spark's Python API.

Author: Dat Tran

License: See LICENSE.txt

Getting Started

  1. Use brew to install Apache Spark: brew install apache-spark
  2. Change logging settings:
  • cd /usr/local/Cellar/apache-spark/1.6.1/libexec/conf
  • cp log4j.properties.template log4j.properties
  • Set info to error: log4j.rootCategory=ERROR, console
  1. Add this to your bash profile: export SPARK_HOME="/usr/local/Cellar/apache-spark/1.6.1/libexec/"
  2. Use nosetests to run the test: nosetests -vs test_clustering.py

About

A simple Spark TDD example

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Python 100.0%