Environment Variables; Environment Variable Default Value Description; SPARK_CONF_DIR ${SPARK_HOME}/conf. To access this file, use the Ambari or Cloudera cluster configuration browser to update the yarn.application.classpath property to include one of the following values, depending on your version of Spark: Enables vectorized orc decoding in native implementation. The SPARK_HOME variable indicates the Apache Spark installation, and PATH adds the Apache Spark (SPARK_HOME) to the system paths. Programing Excavation: Spark Configuration Mess Solved You can also set environment variables using the spark_env_vars field in the Create cluster request or Edit cluster request Clusters API endpoints. Spark will use the configuration files (spark-defaults.conf, spark-env.sh, log4j2.properties, etc) from this directory. println("##spark read text files from a directory into RDD") val . Configuration - Spark 3.2.1 Documentation spark-submit command supports the following. In the application, the configuration is an instance of the Config class, loaded using the ConfigFactory class. Spark allows you to easily do the same within your application's billing portal. Modifying the Spark Configuration Files_Kunpeng BoostKit for Big Data ... Storage configuration — Delta Lake Documentation Spark allows you to simply create an empty conf: val sc = new SparkContext(new SparkConf()) Then, you can supply configuration values at runtime: ./bin/spark-submit --name "My app" --master local[4] --conf spark.eventLog.enabled=false --conf "spark.executor.extraJavaOptions=-XX:+PrintGCDetails -XX:+PrintGCTimeStamps" myApp.jar First, Workspace.from_config accesses your Azure Machine Learning workspace using the configuration in config.json (see Create a workspace configuration file ). Spark Configs. This would initiate the creating of the Spark pool in the Azure Synapse Analytics workspace.