CtrlK
BlogDocsLog inGet started
Tessl Logo

fine-tuning-expert

Use when fine-tuning LLMs, training custom models, or adapting foundation models for specific tasks. Invoke for configuring LoRA/QLoRA adapters, preparing JSONL training datasets, setting hyperparameters for fine-tuning runs, adapter training, transfer learning, finetuning with Hugging Face PEFT, OpenAI fine-tuning, instruction tuning, RLHF, DPO, or quantizing and deploying fine-tuned models. Trigger terms include: LoRA, QLoRA, PEFT, finetuning, fine-tuning, adapter tuning, LLM training, model training, custom model.

97

1.01x
Quality

100%

Does it follow best practices?

Impact

94%

1.01x

Average score across 6 eval scenarios

SecuritybySnyk

Advisory

Suggest reviewing before use

SKILL.md
Quality
Evals
Security

Evaluation results

95%

6%

Fine-Tuning a Model for Medical Documentation Assistance

LoRA training configuration

Criteria
Without context
With context

Uses PEFT/LoRA method

100%

100%

LoRA rank in range

100%

100%

Alpha = 2x rank

100%

100%

Dropout set

100%

100%

Attention target modules

100%

100%

Flash Attention

80%

100%

Memory-efficient optimizer

100%

100%

Gradient checkpointing

100%

100%

use_reentrant=False

0%

100%

Cosine scheduler

100%

100%

Warmup ratio set

100%

100%

LoRA learning rate

100%

100%

SFTTrainer used

100%

100%

Sequence packing

100%

0%

group_by_length

0%

100%

Training config file

100%

100%

92%

4%

Preparing a Customer Support Dataset for Fine-Tuning

Dataset preparation and validation

Criteria
Without context
With context

Empty field detection

100%

100%

Length validation

0%

0%

Bad pattern filtering

100%

100%

Exact deduplication

100%

100%

Fuzzy deduplication

100%

100%

Train/val split

100%

100%

Fixed random seed

100%

100%

Val size ~10%

66%

100%

Staged filtering report

100%

100%

Dataset stats reported

71%

100%

Output files produced

100%

100%

Preparation script provided

100%

100%

100%

Preparing a Fine-Tuned LoRA Model for Production Deployment

Model deployment and evaluation

Criteria
Without context
With context

merge_and_unload used

100%

100%

Adapter compatibility check

100%

100%

Target module mismatch identified

100%

100%

Evaluation before deployment

100%

100%

Latency measurement

100%

100%

Throughput measurement

100%

100%

Deployment recommendation present

100%

100%

Eval metrics reported

100%

100%

Tokenizer saved with model

100%

100%

Incompatible merge danger explained

100%

100%

No deploy-before-eval pattern

100%

100%

Evaluation script provided

100%

100%

100%

2%

Medical Imaging Assistant — Large Model Fine-Tuning

QLoRA memory-efficient setup and design rationale

Criteria
Without context
With context

BitsAndBytesConfig 4-bit

100%

100%

NF4 quant type

100%

100%

Double quantization

100%

100%

prepare_model_for_kbit_training called

100%

100%

use_gradient_checkpointing in kbit prep

100%

100%

max_grad_norm <= 0.3

100%

100%

paged_adamw_8bit optimizer

100%

100%

print_trainable_parameters

100%

100%

Rationale: method choice with memory reasoning

100%

100%

Rationale: rank selection reasoning

100%

100%

Rationale: learning rate reasoning

66%

100%

LoRA LR in correct range

100%

100%

100%

5%

Retail Chatbot Fine-Tuning — Fixing a Failed Training Run

Small dataset overfitting prevention and training monitoring

Criteria
Without context
With context

Reduced LoRA rank

58%

100%

Increased lora_dropout

100%

100%

Non-zero weight_decay

100%

100%

Increased warmup_ratio

100%

100%

load_best_model_at_end

100%

100%

metric_for_best_model

100%

100%

Periodic evaluation enabled

100%

100%

save_total_limit set

100%

100%

Overfitting detection logic

100%

100%

Loss tracking in callback

100%

100%

Notes explain WHY changes help

100%

100%

78%

-14%

Financial News Model — Dual-Target Deployment Pipeline

Post-training quantization for multi-target deployment

Criteria
Without context
With context

GPTQ quantization configured

100%

0%

GPTQ calibration dataset

100%

0%

AWQ quantization configured

0%

100%

GGUF export present

100%

100%

GGUF k-quant for CPU

100%

100%

Benchmark warmup runs

100%

100%

Latency and throughput measured

100%

100%

merge_and_unload before quantization

100%

100%

Deployment guide format rationale

100%

100%

Size and performance estimates

100%

100%

Serving instructions present

100%

100%

Repository
jeffallan/claude-skills
Evaluated
Agent
Claude Code
Model
Claude Sonnet 4.6

Table of Contents

Is this your skill?

If you maintain this skill, you can claim it as your own. Once claimed, you can manage eval scenarios, bundle related skills, attach documentation or rules, and ensure cross-agent compatibility.