DISCONTINUATION OF PROJECT.
This project will no longer be maintained by Intel.
Intel has ceased development and contributions including, but not limited to, maintenance, bug fixes, new releases, or updates, to this project.
Intel no longer accepts patches to this project.
If you have an ongoing need to use this project, are interested in independently developing it, or would like to maintain patches for the open source software community, please create your own fork of this project.
See our VLDB 2017 paper:
Michael Anderson, Shaden Smith, Narayanan Sundaram, Mihai Capotă, Zheguang Zhao, Subramanya Dulloor, Nadathur Satish, & Theodore L. Willke. 2017. Bridging the gap between HPC and big data frameworks. Proc. VLDB Endow. 10, 8 (April 2017), 901-912.
https://doi.org/10.14778/3090163.3090168
http://www.vldb.org/pvldb/vol10/p901-anderson.pdf
-
HADOOP_HDFS_HOME
environment variable set to the Hadoop install directory, such thatHADOOP_HDFS_HOME/include
containshdfs.h
andHADOOP_HDFS_HOME/lib/native
containslibhdfs.so
. -
Spark running in standalone mode. Other modes are not supported.
- Install Scala library
sbt publishLocal
- Install library & helper script
mkdir build
cd build
cmake ..
make
cp ../src/main/native/*.h .
cp ../scripts/*.sh .
export MPI_ADAPTER_FOR_SPARK_HOME=$(pwd)
cd ..
- Build example Scala library
cd examples
sbt package
- Build example native code
mkdir build
cd build
cmake ..
make
export MPI_ADAPTER_FOR_SPARK_EXAMPLES=$(pwd)
cd ..
- Run
spark-submit --jars \
../target/scala-2.10/spark-mpi-adapter_2.10-0.1-SNAPSHOT.jar \
target/scala-2.10/spark-mpi-adapter-examples_2.10-0.1-SNAPSHOT.jar \
com.intel.MPIAdapterForSpark.VectorIncrement