Use when fine-tuning LLMs, training custom models, or adapting foundation models for specific tasks. Invoke for configuring LoRA/QLoRA adapters, preparing JSONL training datasets, setting hyperparameters for fine-tuning runs, adapter training, transfer learning, finetuning with Hugging Face PEFT, OpenAI fine-tuning, instruction tuning, RLHF, DPO, or quantizing and deploying fine-tuned models. Trigger terms include: LoRA, QLoRA, PEFT, finetuning, fine-tuning, adapter tuning, LLM training, model training, custom model.
97
100%
Does it follow best practices?
Impact
94%
1.01xAverage score across 6 eval scenarios
Advisory
Suggest reviewing before use
LoRA training configuration
Uses PEFT/LoRA method
100%
100%
LoRA rank in range
100%
100%
Alpha = 2x rank
100%
100%
Dropout set
100%
100%
Attention target modules
100%
100%
Flash Attention
80%
100%
Memory-efficient optimizer
100%
100%
Gradient checkpointing
100%
100%
use_reentrant=False
0%
100%
Cosine scheduler
100%
100%
Warmup ratio set
100%
100%
LoRA learning rate
100%
100%
SFTTrainer used
100%
100%
Sequence packing
100%
0%
group_by_length
0%
100%
Training config file
100%
100%
Dataset preparation and validation
Empty field detection
100%
100%
Length validation
0%
0%
Bad pattern filtering
100%
100%
Exact deduplication
100%
100%
Fuzzy deduplication
100%
100%
Train/val split
100%
100%
Fixed random seed
100%
100%
Val size ~10%
66%
100%
Staged filtering report
100%
100%
Dataset stats reported
71%
100%
Output files produced
100%
100%
Preparation script provided
100%
100%
Model deployment and evaluation
merge_and_unload used
100%
100%
Adapter compatibility check
100%
100%
Target module mismatch identified
100%
100%
Evaluation before deployment
100%
100%
Latency measurement
100%
100%
Throughput measurement
100%
100%
Deployment recommendation present
100%
100%
Eval metrics reported
100%
100%
Tokenizer saved with model
100%
100%
Incompatible merge danger explained
100%
100%
No deploy-before-eval pattern
100%
100%
Evaluation script provided
100%
100%
QLoRA memory-efficient setup and design rationale
BitsAndBytesConfig 4-bit
100%
100%
NF4 quant type
100%
100%
Double quantization
100%
100%
prepare_model_for_kbit_training called
100%
100%
use_gradient_checkpointing in kbit prep
100%
100%
max_grad_norm <= 0.3
100%
100%
paged_adamw_8bit optimizer
100%
100%
print_trainable_parameters
100%
100%
Rationale: method choice with memory reasoning
100%
100%
Rationale: rank selection reasoning
100%
100%
Rationale: learning rate reasoning
66%
100%
LoRA LR in correct range
100%
100%
Small dataset overfitting prevention and training monitoring
Reduced LoRA rank
58%
100%
Increased lora_dropout
100%
100%
Non-zero weight_decay
100%
100%
Increased warmup_ratio
100%
100%
load_best_model_at_end
100%
100%
metric_for_best_model
100%
100%
Periodic evaluation enabled
100%
100%
save_total_limit set
100%
100%
Overfitting detection logic
100%
100%
Loss tracking in callback
100%
100%
Notes explain WHY changes help
100%
100%
Post-training quantization for multi-target deployment
GPTQ quantization configured
100%
0%
GPTQ calibration dataset
100%
0%
AWQ quantization configured
0%
100%
GGUF export present
100%
100%
GGUF k-quant for CPU
100%
100%
Benchmark warmup runs
100%
100%
Latency and throughput measured
100%
100%
merge_and_unload before quantization
100%
100%
Deployment guide format rationale
100%
100%
Size and performance estimates
100%
100%
Serving instructions present
100%
100%
5b76101
Table of Contents
If you maintain this skill, you can claim it as your own. Once claimed, you can manage eval scenarios, bundle related skills, attach documentation or rules, and ensure cross-agent compatibility.