tessl install tessl/pypi-apache-airflow-backport-providers-apache-sqoop@2021.3.0Apache Airflow backport provider package for Apache Sqoop integration, providing SqoopHook and SqoopOperator for data import/export between relational databases and Hadoop
Agent Success
Agent success rate when using this tile
92%
Improvement
Agent success rate improvement when using this tile compared to baseline
1.39x
Baseline
Agent success rate without this tile
66%
{
"context": "Evaluates whether the solution leverages apache-airflow-providers-apache-sqoop for HCatalog-targeted imports, including optional Hive table creation and static partition loading. Scoring focuses only on correct use of SqoopHook import APIs and relevant Sqoop/HCatalog options for the described scenarios.",
"type": "weighted_checklist",
"checklist": [
{
"name": "HCatalog call",
"description": "Uses SqoopHook.import_table (or equivalent hook API) with hcatalog_database, hcatalog_table, and warehouse_dir parameters so the import lands in Hive/HCatalog rather than a raw HDFS directory.",
"max_score": 30
},
{
"name": "Create flag",
"description": "Maps the create_table input to the create_hcatalog_table argument when invoking SqoopHook so that Sqoop performs table creation only when requested.",
"max_score": 20
},
{
"name": "Partition options",
"description": "Translates the partition mapping into hcatalog-partition-keys and hcatalog-partition-values (e.g., via extra_import_options) so Sqoop writes into the specified static partition.",
"max_score": 20
},
{
"name": "Parallel split",
"description": "Forwards split_by (when provided) and num_mappers values into the SqoopHook import call to control mapper parallelism for the import job.",
"max_score": 15
},
{
"name": "Error surfacing",
"description": "Relies on SqoopHook/AirflowException handling of non-zero Sqoop exits to produce a clear failure when the target table is missing and create_table is false, rather than swallowing the hook error.",
"max_score": 15
}
]
}