Apache Airflow backport provider package for Apache Sqoop integration, providing SqoopHook and SqoopOperator for data import/export between relational databases and Hadoop
Overall
score
92%
{
"context": "Evaluates how well the solution leverages the Apache Airflow Sqoop provider to tune mapper parallelism, choose a split column, and inject JVM properties while returning a safe command preview.",
"type": "weighted_checklist",
"checklist": [
{
"name": "Mapper count wiring",
"description": "Configures SqoopOperator or SqoopHook with the requested num_mappers and downgrades to a single mapper when no split_by column is available.",
"max_score": 30
},
{
"name": "Split column usage",
"description": "Supplies split_by to the provider import call (import_table or import_query) whenever mapper parallelism is used so Sqoop partitions work on the chosen column.",
"max_score": 25
},
{
"name": "JVM properties",
"description": "Passes JVM/system properties via the provider's properties argument so they become -Dkey=value flags in the Sqoop command for the run.",
"max_score": 25
},
{
"name": "Redacted preview",
"description": "Builds the returned command preview using SqoopHook.cmd_mask_password or the operator’s masking so credentials are hidden while still showing mapper, split_by, and -D flags.",
"max_score": 10
},
{
"name": "Result content",
"description": "Returns target_dir along with the masked command string derived from the provider invocation, reflecting the effective mappers, split column, and injected properties.",
"max_score": 10
}
]
}Install with Tessl CLI
npx tessl i tessl/pypi-apache-airflow-backport-providers-apache-sqoopdocs
evals
scenario-1
scenario-2
scenario-3
scenario-4
scenario-5
scenario-6
scenario-7
scenario-8
scenario-9
scenario-10