Răsfoiți Sursa

readme: more info on usage

AlpinDale 1 an în urmă
părinte
comite
d6f705d90b
1 a modificat fișierele cu 18 adăugiri și 2 ștergeri
  1. 18 2
      README.md

+ 18 - 2
README.md

@@ -102,7 +102,8 @@ $ export CUDA_HOME=/home/anon/miniconda3/envs/aphrodite
 Then run the installation command again.
 
 ## Example usage
-**Currently not working, but this is how you'd run it once it's fixed.**
+
+### Inference with `LLM`
   ```py
   from aphrodite import LLM, SamplingParams
 
@@ -113,7 +114,7 @@ Then run the installation command again.
   ]
   sampling_params = SamplingParams(temperature=0.8, top_p=0.95)
 
-  llm = LLM(model="EleutherAI/pythia-70m")
+  llm = LLM(model="EleutherAI/pythia-70m")        # you can also use a local directory path
   outputs = llm.generate(prompts, sampling_params)
   for output in outputs:
     prompt = output.prompt
@@ -121,6 +122,21 @@ Then run the installation command again.
     print(f"Prompt: {prompt!r}, Generated text: {generated_text!r}")
   ```
 
+### Continuous inference with API
+```sh
+$ python -m python -m aphrodite.endpoints.openai.api_server --model EleutherAI/pythia-70m
+$ curl http://localhost:8000/v1/completions \
+    -H "Content-Type: application/json" \
+    -d '{
+        "model": "EleutherAI/pythia-70m",
+        "prompt": "What is a man? A",
+        "max_tokens": 512,
+        "n": 2048,
+        "temperature": 0.8
+    }'
+```
+For the full list of request parameters, see [OpenAI Completions API reference](https://platform.openai.com/docs/api-reference/completions).
+
 
 ## Contributing
 We accept PRs! There will likely be a few typos or other errors we've failed to catch, so please let us know either via an issue or make a Pull Request.