Apache Airflow backport provider package for Apache Sqoop integration, providing SqoopHook and SqoopOperator for data import/export between relational databases and Hadoop
Overall
score
92%
{
"context": "Evaluates whether the solution uses airflow.providers.apache.sqoop's SqoopHook.import_table to move relational table data into HDFS. Checks correct parameter mapping for column filtering, append handling, parallelism, and driver/direct controls as specified.",
"type": "weighted_checklist",
"checklist": [
{
"name": "Hook usage",
"description": "Initializes airflow.providers.apache.sqoop.hooks.sqoop.SqoopHook with the provided conn_id and forwards num_mappers at construction rather than shelling out manually.",
"max_score": 15
},
{
"name": "Import call",
"description": "Invokes SqoopHook.import_table with table and target_dir, mapping file_format to file_type (defaulting to 'parquet') and returning the HDFS path.",
"max_score": 25
},
{
"name": "Columns and where",
"description": "Passes selected columns as a comma-separated string to import_table's columns argument and forwards the WHERE fragment to the where parameter when supplied.",
"max_score": 20
},
{
"name": "Append flag",
"description": "Sets import_table's append argument based on the append toggle, using the package flag to choose between replacing or extending existing HDFS data.",
"max_score": 20
},
{
"name": "Parallel controls",
"description": "Propagates split_by for shard balancing, direct for fast-path imports, and driver overrides into SqoopHook.import_table so Sqoop runs with the requested parallel/transport settings.",
"max_score": 20
}
]
}Install with Tessl CLI
npx tessl i tessl/pypi-apache-airflow-backport-providers-apache-sqoopdocs
evals
scenario-1
scenario-2
scenario-3
scenario-4
scenario-5
scenario-6
scenario-7
scenario-8
scenario-9
scenario-10