tessl install tessl/pypi-apache-airflow-backport-providers-apache-sqoop@2021.3.0Apache Airflow backport provider package for Apache Sqoop integration, providing SqoopHook and SqoopOperator for data import/export between relational databases and Hadoop
Agent Success
Agent success rate when using this tile
92%
Improvement
Agent success rate improvement when using this tile compared to baseline
1.39x
Baseline
Agent success rate without this tile
66%
{
"context": "Evaluates whether the solution uses airflow.providers.apache.sqoop's SqoopHook.import_table to move relational table data into HDFS. Checks correct parameter mapping for column filtering, append handling, parallelism, and driver/direct controls as specified.",
"type": "weighted_checklist",
"checklist": [
{
"name": "Hook usage",
"description": "Initializes airflow.providers.apache.sqoop.hooks.sqoop.SqoopHook with the provided conn_id and forwards num_mappers at construction rather than shelling out manually.",
"max_score": 15
},
{
"name": "Import call",
"description": "Invokes SqoopHook.import_table with table and target_dir, mapping file_format to file_type (defaulting to 'parquet') and returning the HDFS path.",
"max_score": 25
},
{
"name": "Columns and where",
"description": "Passes selected columns as a comma-separated string to import_table's columns argument and forwards the WHERE fragment to the where parameter when supplied.",
"max_score": 20
},
{
"name": "Append flag",
"description": "Sets import_table's append argument based on the append toggle, using the package flag to choose between replacing or extending existing HDFS data.",
"max_score": 20
},
{
"name": "Parallel controls",
"description": "Propagates split_by for shard balancing, direct for fast-path imports, and driver overrides into SqoopHook.import_table so Sqoop runs with the requested parallel/transport settings.",
"max_score": 20
}
]
}