Skip to content

Splice Machine Apache Zeppelin notebooks

Notifications You must be signed in to change notification settings

splicemachine/zeppelin-notebooks

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

41 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Splice Machine Apache Zeppelin notebooks

Updating your Zeppelin instance with all notebooks in this repo

If you would like to load these notebooks in your current zeppelin cluster you can use the load-notebooks.py script. It takes several optional parameters:

  • -z: The Zeppelin URL ie http://localhost:8080. It defaults to localhost:8080
  • -n: The directory containing the notebook folders. It defaults to the directory where this script is located.
  • -d: Flag to indicate you want to delete existing notebooks whose id and/or name matches
  • -u: If your zeppelin expects credentials, then this should be the user id
  • -p: If your zeppelin expects credentials, then this should be the password

Importing specific notebooks

Copy the "json" link URL from the table below and paste it into Zeppelin's import from URL tool.

Adding notebooks

You can add notebooks by making Github pull request, but we request you also update the table below by adding a record for your notebook.

Notebooks

Id Description Interpreter / Component Code Comments
2CTV7H6FU 1. Getting Started - README Markup json
2CTKW7A6U 2.1 Notebook Basics Tutorial Splicemachine json
2E5JAQWJH 2.2 Copying to S3 Tutorial Splicemachine json
2CSKWGZ8P 2.3 Importing Data Tutorial Splicemachine json
2CU1SNACA 2.4 Running Queries Tutorial Splicemachine json
2CS27TE2A 2.5 Tuning for Performance Tutorial Splicemachine json
2CVPHXRZN 2.6 Using the Database Console UI Tutorial Splicemachine / SparkUI json
2CU4BJJDS 2.7 Explaining and Hinting Tutorial Splicemachine json
2CFY9Q6NS 2.8 TPCH-1 Tutorial Splicemachine json
2CFMYDAYJ 2.9 Common Utilities Splicemachine json
2D76NPKV6 3. Splice Deep Dive / 1. Introduction Splicemachine json
2D6ZFHE9U 3. Splice Deep Dive / 2. The Life of a Query Splicemachine json
2D93Y56QJ 3. Splice Deep Dive / 3. Monitoring Queries Splicemachine / SparkUI json
2D9M6ZUAW 3. Splice Deep Dive / 4. Visualizing Results with Zeppelin Splicemachine json
2D89MQA89 3. Splice Deep Dive / 5. Transactions with Spark & JDBC Splicemachine json
2D8PMSAPJ 3. Splice Deep Dive / 6. Creating Applications Splicemachine Interpreter json
2D96GBY2P 3. Splice Deep Dive / 7. Using our Spark Adapter Splicemachine / Spark Python json
2E3671BBN 3. Splice Deep Dive / 8. Python MLlib example Splicemachine / Spark Python json
2D7P8BMMS 3. Splice Deep Dive / 9. Scala MLlib example Splicemachine / Scala / Spark json
2CKKJKSK8 4. Reference Applications / 1. Supply Chain - Predicting Shortages Splicemachine / Spark json
2CH931JT4 4. Reference Applications / 2. IoT / 1. Overview json
2CHVYSK7B 4. Reference Applications / 2. IoT / 2. Database Setup json
2CG53XQM7 4. Reference Applications / 2. IoT / 3. Kafka json
2CGVRH6SF 4. Reference Applications / 2. IoT / 4.SparkStream json
2CG6KVZJ2 4. Reference Applications / 2. IoT / 5. Splice Query json
2D2RM85JF 5. Advanced Tutorials / 1. Creating Custom Stored Procedures json
2CSGDX1CW 6. Benchmarks / 1. TPCH-100 json
2E6E4ECVX ML Manager Demo json

Updating your HDP Zeppelin to use the Splice Machine Native Spark Datasource

  • Create a directory such as /tmp/splicemachine on each server to store 1 jar files needed by this setup.
  • Get the Splice Machine Native Spark Datasource for HDP for the version you are using ie (splicemachine-hdp2.6.3-2.2.0.2.6.3.0-235_2.11-2.7.0.1828.jar) for your cluster and place it in the /tmp/splicemachine directory on each server
  • Copy the configure-spark-interpreter.py from the zeppelin-notebook repo to a server that is running hortonworks and run the following command where: ** The z parameter is the zeppelin server url - update the server and port to match your environment. ** The a parameter is HDP ** The s parameter is the directory from the step above ** The v parameter is the version of zeppelin ** The u parameter is the user if your zeppelin expects credentials ** The p parameter is the password if your zeppelin expects credentials ** The i parameter is the name of the Native Spark Datasource jar
  • Run a command like the following to configure your HDP Zeppelin
python configure-spark-interpreter.py -z "http://localhost:9995" -a "HDP" -s "/tmp/zeppelin" -v "0.7.3" -u admin -p admin -i splicemachine-hdp2.6.3-2.2.0.2.6.3.0-235_2.11-2.7.0.1828.jar