Welcome to Sparklin project. This is a proof-of-concept software includes experimental
new Kotlin REPL (aka KShell) which supports extensions and custom commands via plugin system.
Spark supported by separate plugins for versions 1.x and 2.x.
First please checkout the latest master. After that you can build current version from the project directory:
git clone http://github.com/khud/sparklin
cd sparklin
mvn clean installOne can run KShell out of box by following command from sparklin home directory in the terminal:
bin/kshell.shIn order to install Kotlin to Zeppelin, change directory to zeppelin-interpreter and run
mvn clean installMaven will produce 2 jars: the first one is interpreter itself, the second one is for dependencies due to KT-24453.
Right now interpreter can not be install
via install-interpreter.sh.
You need to install it manually. In order to do that copy jars to appropriate subdirectory of $ZEPPELIN_HOME/interpreter,
for example kotlin and configure interpreter according this
instruction.
Some times you need to delete file $ZEPPELIN/conf/interpreter.json to make interpreter available.
To switch Kotlin syntax highlight on you should copy mode-kotlin.js from Zeppelin sources to
$ZEPPELIN_HOME/webapps/webapp directory.
By default KShell don't need a configuration file, but if you want to try Spark you need to enable plugin.
It can be done by passing system property config.path to JVM. Example of Spark 1.x plugin configuration you may
find in $SPARKLIN_HOME/conf/spark1x.properties. Of course you could create .kshell file in your home directory but
using separate files for various environments is highly recommended. Please set appropriate SPARK_HOME variable in
the file bin/sparklin-spark1x.sh.
JavaSparkContext is available as sc, SQLContext is available as sqlContext.
The configuration is completely the same as in the case of Spark 1.x but you need $SPARKLIN_HOME/conf/spark2x.properties
and bin/sparklin-spark2x.sh instead. Do not forget set SPARK_HOME to your Spark 2.x installation.
SparkSession is available as spark, JavaSparkContext is available as sc.
At the moment Sparklin does not have additional Spark runtime support for Kotlin for both Spark versions.
But it will be available soon.
Sparklin version 0.2 is tested in following environment:
- Mac OS X 10.11.6 / CentOS Linux (Kernel 3.10)
- Java 1.8.0_152
- Maven 3.3.3 / 3.5.3
- Kotlin 1.2.30/1.2.41/1.2.50
- Hadoop 2.7
- Spark 1.6.1
- Spark 2.2.0