A high-throughput and memory-efficient inference and serving engine for LLMs
Overall
score
69%
Evaluation — 69%
↑ 1.33xAgent success when using this tile
{
"context": "This evaluation assesses how well the engineer uses vLLM's text generation capabilities to implement a story generator. The focus is on proper usage of the LLM class, generate() method, and SamplingParams configuration.",
"type": "weighted_checklist",
"checklist": [
{
"name": "LLM initialization",
"description": "Uses vLLM's LLM class to initialize the model in __init__ method, properly storing it as an instance variable. The model_name parameter should be passed to the LLM constructor.",
"max_score": 20
},
{
"name": "generate() method usage",
"description": "Uses the LLM.generate() method to generate text from prompts. The method should accept prompts as input and return RequestOutput objects containing the generated text.",
"max_score": 25
},
{
"name": "SamplingParams configuration",
"description": "Creates and uses SamplingParams objects to configure generation behavior. Must properly set max_tokens, temperature, and n parameters based on the method arguments.",
"max_score": 30
},
{
"name": "Output extraction",
"description": "Correctly extracts generated text from RequestOutput objects by accessing the outputs attribute and getting the text from CompletionOutput objects. Returns a list of strings as specified in the API.",
"max_score": 15
},
{
"name": "Error handling",
"description": "Implements proper validation to raise ValueError when prompt is empty, as specified in the API documentation.",
"max_score": 10
}
]
}Install with Tessl CLI
npx tessl i tessl/pypi-vllmdocs
evals
scenario-1
scenario-2
scenario-3
scenario-4
scenario-5
scenario-6
scenario-7
scenario-8
scenario-9
scenario-10