Tutorials¶
These tutorials guide you through running Spark jobs on HPC clusters using sparkctl. Each tutorial covers a different workflow - choose the one that best fits your needs.
Which Tutorial Should I Use?¶
Tutorial |
Best For |
Client Install |
Interface |
|---|---|---|---|
Traditional Spark users, production jobs |
|
CLI |
|
Lightweight client, remote connectivity |
|
CLI |
|
Programmatic control, automation scripts |
|
Python |
|
Exploratory analysis, debugging |
|
Python REPL |
Decision Guide¶
Start here if you’re new to sparkctl: spark-submit / pyspark - this is the most familiar workflow for existing Spark users.
Choose by use case:
“I want to submit batch jobs” → spark-submit / pyspark
“I want a minimal client installation” → Spark Connect CLI
“I want to control the cluster from Python code” → Python Library
“I want to explore data interactively” → Interactive Development