from aphrodite import LLM, SamplingParams prompts = [ "A robot may not injure a human being", "It is only with the heart that one can see rightly;", "The greatest glory in living lies not in never falling,", ] answers = [ " or, through inaction, allow a human being to come to harm.", " what is essential is invisible to the eye.", " but in rising every time we fall.", ] N = 1 # Currently, top-p sampling is disabled. `top_p` should be 1.0. sampling_params = SamplingParams(temperature=0.7, top_p=1.0, n=N, max_tokens=16) # Set `enforce_eager=True` to avoid ahead-of-time compilation. # In real workloads, `enforace_eager` should be `False`. llm = LLM(model="google/gemma-2b", enforce_eager=True) outputs = llm.generate(prompts, sampling_params) for output, answer in zip(outputs, answers): prompt = output.prompt generated_text = output.outputs[0].text print(f"Prompt: {prompt!r}, Generated text: {generated_text!r}") assert generated_text.startswith(answer)