Apache Spark
Config Commands
Configure Spark applications for performance and reliability. Set memory, cores, serialization, and shuffle parameters via command line or configuration files.
7 commands
Pro Tips
Set 'spark.serializer=org.apache.spark.serializer.KryoSerializer' for better performance.
Use 'spark.memory.fraction=0.6' to balance execution and storage memory.
Configure 'spark.sql.adaptive.enabled=true' for automatic query optimization (Spark 3+).
Common Mistakes
spark-defaults.conf changes affect all applications - use --conf for job-specific settings.
Incorrect memory configuration is the most common cause of Spark job failures.