Skip to content

Commit

Permalink
Refactor README to include available endpoints and environment variables
Browse files Browse the repository at this point in the history
  • Loading branch information
hummerichsander authored Sep 25, 2024
1 parent d30d1e4 commit 08f5115
Showing 1 changed file with 3 additions and 0 deletions.
3 changes: 3 additions & 0 deletions README.md
Original file line number Diff line number Diff line change
Expand Up @@ -2,9 +2,11 @@
This is a simple fastapi based server mock that implements the OpenAI API.

Available endpoints:

- /v1/chat/completion

Instead of running a LLM model to generate completions, it simply returns a response generated by surrogate models. Available surrogate models are:

- "yes_no": returns random "Yes" or "No" response
- "ja_nein": returns random "Ja" or "Nein" response
- "lorem_ipsum": returns random "lorem ipsum" text
Expand All @@ -15,6 +17,7 @@ docker pull ghcr.io/hummerichsander/openai_api_server_mock:v ... # replace ... w
```

Environment variables:

- `CONTEXT_SIZE`: context size for the model (default: 4096)
- `SLEEP_TIME`: sleep time in seconds before returning the response (default: 0)
- `MAX_CONCURRENT_REQUESTS`: maximum number of concurrent requests (default: 10^9)

0 comments on commit 08f5115

Please sign in to comment.