How to use a custom spark-default.conf file¶
sparkctl sets custom settings in the spark-defaults.conf file. For example, it sets
spark.sql.parquet.outputTimestampType TIMESTAMP_MICROS
so that Spark writes Parquet files with timestamps that can be properly interpreted by other libraries like DuckDB and Pandas.
If you don’t want these settings or want to add your own settings every time you configure
sparkctl, here is how to do it.
Create a file named
spark-defaults.conf.template. Spark’s defaults are stored here.Add your custom settings to the file.
Pass the file to
sparkctl configurewith the--spark-defaults-template-fileoption:$ sparkctl configure --spark-defaults-template-file spark-defaults.conf.template
Note
sparkctl will still append Spark driver, executor, and other settings to the runtime version of the spark-default.conf file.