Apache Airflow backport provider package for Apache Sqoop integration, providing SqoopHook and SqoopOperator for data import/export between relational databases and Hadoop
Overall
score
92%
{
"context": "Evaluates whether the solution leverages apache-airflow-providers-apache-sqoop for HCatalog-targeted imports, including optional Hive table creation and static partition loading. Scoring focuses only on correct use of SqoopHook import APIs and relevant Sqoop/HCatalog options for the described scenarios.",
"type": "weighted_checklist",
"checklist": [
{
"name": "HCatalog call",
"description": "Uses SqoopHook.import_table (or equivalent hook API) with hcatalog_database, hcatalog_table, and warehouse_dir parameters so the import lands in Hive/HCatalog rather than a raw HDFS directory.",
"max_score": 30
},
{
"name": "Create flag",
"description": "Maps the create_table input to the create_hcatalog_table argument when invoking SqoopHook so that Sqoop performs table creation only when requested.",
"max_score": 20
},
{
"name": "Partition options",
"description": "Translates the partition mapping into hcatalog-partition-keys and hcatalog-partition-values (e.g., via extra_import_options) so Sqoop writes into the specified static partition.",
"max_score": 20
},
{
"name": "Parallel split",
"description": "Forwards split_by (when provided) and num_mappers values into the SqoopHook import call to control mapper parallelism for the import job.",
"max_score": 15
},
{
"name": "Error surfacing",
"description": "Relies on SqoopHook/AirflowException handling of non-zero Sqoop exits to produce a clear failure when the target table is missing and create_table is false, rather than swallowing the hook error.",
"max_score": 15
}
]
}Install with Tessl CLI
npx tessl i tessl/pypi-apache-airflow-backport-providers-apache-sqoopdocs
evals
scenario-1
scenario-2
scenario-3
scenario-4
scenario-5
scenario-6
scenario-7
scenario-8
scenario-9
scenario-10