Molecular ML with diverse featurizers and pre-built datasets. Use for property prediction (ADMET, toxicity) with traditional ML or GNNs when you want extensive featurization options and MoleculeNet benchmarks. Best for quick experiments with pre-trained models, diverse molecular representations. For graph-first PyTorch workflows use torchdrug; for benchmark datasets use pytdc.
79
75%
Does it follow best practices?
Impact
83%
1.36xAverage score across 3 eval scenarios
Advisory
Suggest reviewing before use
Optimize this skill with Tessl
npx tessl skill review --optimize ./scientific-skills/deepchem/SKILL.mdSmall-dataset featurizer and model selection
Scaffold splitter used
100%
100%
CircularFingerprint radius
100%
100%
CircularFingerprint size
0%
100%
Simple model chosen
100%
100%
SklearnModel wrapper
100%
100%
NormalizationTransformer applied
80%
80%
transform_y=True set
100%
100%
Transformers applied to new molecules
66%
50%
CSVLoader feature_field
100%
100%
Three-way split
0%
0%
NumpyDataset for new molecules
100%
100%
Medium-dataset toxicity classification with imbalanced data
Scaffold splitter selected
0%
100%
No GNN for medium dataset
100%
100%
Medium-dataset model
0%
50%
Fingerprint featurizer for non-GNN
100%
100%
Imbalance handling
50%
100%
MoleculeNet Tox21 loader
100%
0%
ROC-AUC metric
50%
50%
Multitask evaluation
100%
100%
CircularFingerprint params
100%
50%
evaluation_results.json present
100%
100%
Transfer learning on small novel-scaffold molecular dataset
HuggingFaceModel used
0%
100%
Correct ChemBERTa model ID
71%
71%
Low fine-tuning learning rate
85%
100%
DummyFeaturizer for pretrained model
0%
100%
Scaffold splitter used
50%
100%
CSVLoader with feature_field
0%
100%
Approach summary written
100%
100%
uv installation command
0%
0%
n_tasks set to 1
0%
0%
Classification task type
50%
100%
b58ad7e
Table of Contents
If you maintain this skill, you can claim it as your own. Once claimed, you can manage eval scenarios, bundle related skills, attach documentation or rules, and ensure cross-agent compatibility.