# Licensed to the Apache Software Foundation (ASF) under one
# or more contributor license agreements. See the NOTICE file
# distributed with this work for additional information
# regarding copyright ownership. The ASF licenses this file
# to you under the Apache License, Version 2.0 (the
# "License"); you may not use this file except in compliance
# with the License. You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing,
# software distributed under the License is distributed on an
# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
# KIND, either express or implied. See the License for the
# specific language governing permissions and limitations
# under the License.
# NOTE! THIS FILE IS AUTOMATICALLY GENERATED AND WILL BE OVERWRITTEN!
#
# IF YOU WANT TO MODIFY THIS FILE, YOU SHOULD MODIFY THE TEMPLATE
# `get_provider_info_TEMPLATE.py.jinja2` IN the `dev/breeze/src/airflow_breeze/templates` DIRECTORY
[docs]
def get_provider_info():
return {
"package-name": "apache-airflow-providers-databricks",
"name": "Databricks",
"description": "`Databricks <https://databricks.com/>`__\n",
"state": "ready",
"source-date-epoch": 1739959376,
"versions": [
"7.2.0",
"7.0.0",
"6.13.0",
"6.12.0",
"6.11.0",
"6.10.0",
"6.9.0",
"6.8.0",
"6.7.0",
"6.6.0",
"6.5.0",
"6.4.0",
"6.3.0",
"6.2.0",
"6.1.0",
"6.0.0",
"5.0.1",
"5.0.0",
"4.7.0",
"4.6.0",
"4.5.0",
"4.4.0",
"4.3.3",
"4.3.2",
"4.3.1",
"4.3.0",
"4.2.0",
"4.1.0",
"4.0.1",
"4.0.0",
"3.4.0",
"3.3.0",
"3.2.0",
"3.1.0",
"3.0.0",
"2.7.0",
"2.6.0",
"2.5.0",
"2.4.0",
"2.3.0",
"2.2.0",
"2.1.0",
"2.0.2",
"2.0.1",
"2.0.0",
"1.0.1",
"1.0.0",
],
"integrations": [
{
"integration-name": "Databricks",
"external-doc-url": "https://databricks.com/",
"how-to-guide": [
"/docs/apache-airflow-providers-databricks/operators/jobs_create.rst",
"/docs/apache-airflow-providers-databricks/operators/notebook.rst",
"/docs/apache-airflow-providers-databricks/operators/submit_run.rst",
"/docs/apache-airflow-providers-databricks/operators/run_now.rst",
"/docs/apache-airflow-providers-databricks/operators/task.rst",
],
"logo": "/docs/integration-logos/Databricks.png",
"tags": ["service"],
},
{
"integration-name": "Databricks SQL",
"external-doc-url": "https://databricks.com/product/databricks-sql",
"how-to-guide": [
"/docs/apache-airflow-providers-databricks/operators/sql.rst",
"/docs/apache-airflow-providers-databricks/operators/copy_into.rst",
],
"tags": ["service"],
},
{
"integration-name": "Databricks Repos",
"external-doc-url": "https://docs.databricks.com/repos/index.html",
"how-to-guide": [
"/docs/apache-airflow-providers-databricks/operators/repos_create.rst",
"/docs/apache-airflow-providers-databricks/operators/repos_update.rst",
"/docs/apache-airflow-providers-databricks/operators/repos_delete.rst",
],
"tags": ["service"],
},
{
"integration-name": "Databricks Workflow",
"external-doc-url": "https://docs.databricks.com/en/workflows/index.html",
"how-to-guide": ["/docs/apache-airflow-providers-databricks/operators/workflow.rst"],
"tags": ["service"],
},
],
"operators": [
{
"integration-name": "Databricks",
"python-modules": ["airflow.providers.databricks.operators.databricks"],
},
{
"integration-name": "Databricks SQL",
"python-modules": ["airflow.providers.databricks.operators.databricks_sql"],
},
{
"integration-name": "Databricks Repos",
"python-modules": ["airflow.providers.databricks.operators.databricks_repos"],
},
{
"integration-name": "Databricks Workflow",
"python-modules": ["airflow.providers.databricks.operators.databricks_workflow"],
},
],
"hooks": [
{
"integration-name": "Databricks",
"python-modules": [
"airflow.providers.databricks.hooks.databricks",
"airflow.providers.databricks.hooks.databricks_base",
],
},
{
"integration-name": "Databricks SQL",
"python-modules": ["airflow.providers.databricks.hooks.databricks_sql"],
},
],
"triggers": [
{
"integration-name": "Databricks",
"python-modules": ["airflow.providers.databricks.triggers.databricks"],
}
],
"sensors": [
{
"integration-name": "Databricks",
"python-modules": [
"airflow.providers.databricks.sensors.databricks_sql",
"airflow.providers.databricks.sensors.databricks_partition",
],
}
],
"connection-types": [
{
"hook-class-name": "airflow.providers.databricks.hooks.databricks.DatabricksHook",
"connection-type": "databricks",
}
],
"plugins": [
{
"name": "databricks_workflow",
"plugin-class": "airflow.providers.databricks.plugins.databricks_workflow.DatabricksWorkflowPlugin",
}
],
"extra-links": ["airflow.providers.databricks.operators.databricks.DatabricksJobRunLink"],
"dependencies": [
"apache-airflow>=2.9.0",
"apache-airflow-providers-common-sql>=1.20.0",
"requests>=2.27.0,<3",
"databricks-sql-connector>=3.0.0",
"aiohttp>=3.9.2, <4",
"mergedeep>=1.3.4",
"pandas>=2.1.2,<2.2",
"pyarrow>=14.0.1",
],
"optional-dependencies": {
"sdk": ["databricks-sdk==0.10.0"],
"azure-identity": ["azure-identity>=1.3.1"],
},
"devel-dependencies": ["deltalake>=0.12.0"],
}