forked from holdenk/spark-testing-base
-
Notifications
You must be signed in to change notification settings - Fork 0
/
.travis.yml
40 lines (40 loc) · 1.41 KB
/
.travis.yml
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
language: scala
cache:
directories:
- $HOME/.ivy2
scala:
- 2.10.4
sudo: false
before_install:
- export PATH=$HOME/.local/bin:$PATH
- pip install codecov coverage --user `whoami`
install:
# Download spark 1.4.1
- "wget http://d3kbcqa49mib13.cloudfront.net/spark-1.4.1-bin-hadoop2.4.tgz || wget http://www.us.apache.org/dist/spark/spark-1.4.1/spark-1.4.1-bin-hadoop2.4.tgz || wget http://download.nextag.com/apache/spark/spark-1.4.1/spark-1.4.1-bin-hadoop2.4.tgz"
- "tar -xvf spark-1.4.1-bin-hadoop2.4.tgz"
# Install Python deps.
# The conda installation steps here are based on http://conda.pydata.org/docs/travis.html
- wget https://repo.continuum.io/miniconda/Miniconda-latest-Linux-x86_64.sh -O miniconda.sh
- bash miniconda.sh -b -p $HOME/miniconda
- export PATH="$HOME/miniconda/bin:$PATH"
- hash -r
- conda config --set always_yes yes --set changeps1 no
- conda update -q conda
# Useful for debugging any issues with conda
- conda info -a
- deps='pip requests nose sphinx pep8 coverage'
- conda create -p $HOME/py --yes $deps "python=2.7"
- export PATH=$HOME/py/bin:$PATH
- pip install --upgrade unittest2 codecov
script:
- ./sbt/sbt scalastyle
- "export SPARK_HOME=`pwd`/spark-1.4.1-bin-hadoop2.4/"
- ./sbt/sbt clean coverage test
- ./python/run-tests
- "pep8 --ignore=E402 ./python"
after_success:
- codecov
notifications:
email:
recipients:
- holden@pigscanfly.ca