Pipeline management software for clusters.
Overall
score
67%
Build a data processing pipeline that conditionally processes datasets based on quality checks. The workflow should validate input data and only proceed with expensive processing steps if quality thresholds are met.
Create a workflow that:
@generates
from toil.job import Job
from toil.common import Toil
class QualityCheckJob(Job):
"""Performs initial quality check on input data."""
def run(self, fileStore):
# Returns quality score (0-100)
pass
class FullProcessingJob(Job):
"""Runs full processing pipeline for high-quality data."""
def run(self, fileStore):
pass
class PartialProcessingJob(Job):
"""Runs partial processing with cleaning for medium-quality data."""
def run(self, fileStore):
pass
class ErrorReportJob(Job):
"""Generates error report for low-quality data."""
def run(self, fileStore):
pass
class DataWorkflow(Job):
"""Main workflow that conditionally executes processing based on quality."""
def __init__(self, input_file):
super().__init__()
self.input_file = input_file
def run(self, fileStore):
# Implement conditional workflow logic
pass
def main():
"""Entry point that creates and runs the workflow."""
passProvides workflow management and job orchestration capabilities.
@satisfied-by
Install with Tessl CLI
npx tessl i tessl/pypi-toildocs
evals
scenario-1
scenario-2
scenario-3
scenario-4
scenario-5
scenario-6
scenario-7
scenario-8
scenario-9
scenario-10