Skip to content

Latest commit

 

History

History
17 lines (10 loc) · 821 Bytes

README.md

File metadata and controls

17 lines (10 loc) · 821 Bytes

PySpark logging examples in local environment and on Databricks clusters

This repo contains examples on how to configure PySpark logs in the local Apache Spark environment and when using Databricks clusters.

Link to the blogpost with details.

Local setup

Provide your logging configurations in conf/local/log4j.properties and pass this path via SPARK_CONF_DIR when initializing the Python session.

Databricks setup

  • Describe your logging configurations in conf/databricks/driver-log4j.properties.
  • Provide your DATABRICKS_CLI_PROFILE environment variable in the .env file
  • Upload the configurations to DBFS via make upload-log-configuration
  • Add the init script in the cluster properties