This is a provider package for
apache.spark provider. All classes for this provider package
airflow.providers.apache.spark python package.
You can install this package on top of an existing Airflow 2 installation (see
for the minimum Airflow version supported) via
pip install apache-airflow-providers-apache-spark
This release of provider is only available for Airflow 2.3+ as explained in the Apache Airflow providers support policy.
spark-binary connection extra could be set to any binary, but with 4.0.0 version only two values
are allowed for it
spark-home connection extra is not allowed any more - the binary should be available on the
PATH in order to use SparkSubmitHook and SparkSubmitOperator.
Remove custom spark home and custom binaries for spark (#27646)
Move min airflow version to 2.3.0 for all providers (#27196)
This release of provider is only available for Airflow 2.2+ as explained in the Apache Airflow providers support policy https://github.com/apache/airflow/blob/main/README.md#support-for-providers
Add typing for airflow/configuration.py (#23716)
Fix backwards-compatibility introduced by fixing mypy problems (#24230)
AIP-47 - Migrate spark DAGs to new design #22439 (#24210)
chore: Refactoring and Cleaning Apache Providers (#24219)
fix param rendering in docs of SparkSubmitHook (#21788)
Support for Python 3.10
Add more SQL template fields renderers (#21237)
Add optional features in providers. (#21074)
Auto-apply apply_default decorator (#15667)
Due to apply_default decorator removal, this version of the provider requires Airflow 2.1.0+.
If your Airflow version is < 2.1.0, and you want to install this provider version, first upgrade
Airflow to at least version 2.1.0. Otherwise your Airflow package version will be upgraded
automatically and you will have to manually run
airflow upgrade db to complete the migration.
Make SparkSqlHook use Connection (#15794)
Updated documentation and readme files.
Initial version of the provider.