SQLExecuteQueryOperator to connect to Snowflake¶
Use the SQLExecuteQueryOperator
to execute
SQL commands in a Snowflake database.
Warning
Previously, SnowflakeOperator was used to perform this kind of operation. But at the moment SnowflakeOperator is deprecated and will be removed in future versions of the provider. Please consider to switch to SQLExecuteQueryOperator as soon as possible.
Using the Operator¶
Use the conn_id
argument to connect to your Snowflake instance where
the connection metadata is structured as follows:
Parameter |
Input |
---|---|
Login: string |
Snowflake user name |
Password: string |
Password for Snowflake user |
Schema: string |
Set schema to execute SQL operations on by default |
Extra: dictionary |
|
An example usage of the SQLExecuteQueryOperator to connect to Snowflake is as follows:
snowflake_op_sql_str = SQLExecuteQueryOperator(
task_id="snowflake_op_sql_str", sql=CREATE_TABLE_SQL_STRING
)
snowflake_op_with_params = SQLExecuteQueryOperator(
task_id="snowflake_op_with_params",
sql=SQL_INSERT_STATEMENT,
parameters={"id": 56},
)
snowflake_op_sql_list = SQLExecuteQueryOperator(task_id="snowflake_op_sql_list", sql=SQL_LIST)
snowflake_op_sql_multiple_stmts = SQLExecuteQueryOperator(
task_id="snowflake_op_sql_multiple_stmts",
sql=SQL_MULTIPLE_STMTS,
split_statements=True,
)
snowflake_op_template_file = SQLExecuteQueryOperator(
task_id="snowflake_op_template_file",
sql="example_snowflake_snowflake_op_template_file.sql",
)
Note
Parameters that can be passed onto the operator will be given priority over the parameters already given
in the Airflow connection metadata (such as schema
, role
, database
and so forth).
SnowflakeSqlApiOperator¶
Use the SnowflakeSqlApiHook
to execute
SQL commands in a Snowflake database.
You can also run this operator in deferrable mode by setting deferrable
param to True
.
This will ensure that the task is deferred from the Airflow worker slot and polling for the task status happens on the trigger.
Using the Operator¶
Use the snowflake_conn_id
argument to connect to your Snowflake instance where
the connection metadata is structured as follows:
Parameter |
Input |
---|---|
Login: string |
Snowflake user name. If using OAuth connection this is the |
Password: string |
Password for Snowflake user. If using OAuth this is the |
Schema: string |
Set schema to execute SQL operations on by default |
Extra: dictionary |
|
An example usage of the SnowflakeSqlApiHook is as follows:
snowflake_sql_api_op_sql_multiple_stmt = SnowflakeSqlApiOperator(
task_id="snowflake_op_sql_multiple_stmt",
sql=SQL_MULTIPLE_STMTS,
statement_count=len(SQL_LIST),
)
Note
Parameters that can be passed onto the operator will be given priority over the parameters already given
in the Airflow connection metadata (such as schema
, role
, database
and so forth).