Apache Airflow backport provider package for Apache Sqoop integration, providing SqoopHook and SqoopOperator for data import/export between relational databases and Hadoop
Overall
score
92%
A small utility that moves an entire relational table into HDFS using the bundled Sqoop provider dependency.
@generates
from typing import Iterable, Optional
def import_table_to_hdfs(
conn_id: str,
table: str,
target_dir: str,
columns: Optional[Iterable[str]] = None,
where: Optional[str] = None,
append: bool = False,
split_by: Optional[str] = None,
num_mappers: Optional[int] = None,
direct: bool = False,
driver: Optional[str] = None,
file_format: str = "parquet",
) -> str:
"""
Imports a relational table into HDFS using the provided dependency.
Returns the HDFS directory path where data landed.
Raises an exception if the import fails.
"""Provides Sqoop-based import support for moving relational data into HDFS.
Install with Tessl CLI
npx tessl i tessl/pypi-apache-airflow-backport-providers-apache-sqoopdocs
evals
scenario-1
scenario-2
scenario-3
scenario-4
scenario-5
scenario-6
scenario-7
scenario-8
scenario-9
scenario-10