-
Notifications
You must be signed in to change notification settings - Fork 3
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Support generation configuration for LLM #62
Comments
https://docs.litellm.ai/docs/completion/input LiteLLM very nicely tracks supported openai params for any model + provider. litellm.get_supported_openai_params() Nice system for dealing with unsupported params:
|
Based on litellm: https://docs.litellm.ai/docs/completion/input#input-params-1 List also in code here: https://github.com/BerriAI/litellm/blob/main/litellm/main.py#L843 We will support:
Right now we default to structured output { "type": "json_schema", "json_schema": {...} } instead of json output { "type": "json_object" }. If we want to support both, we need to change the way we are doing it. See the API reference.
We can support later if we choose, but won't include now
We will not support:
Others to think about
Optional liteLLM functions params that look interesting:
|
https://docs.litellm.ai/docs/completion/batching#send-multiple-completion-calls-to-1-model LiteLLM also supports messages being a list of lists, sending multiple completions We should benchmark this vs doing threadpool over async completion with litellm. EDIT: added to #74 |
Lessons learned from #77
Instead we can just pass in the kwargs to the constructor of the RequestProcessor. And then use
And if Can also share this main request body code between OpenAI batch and online so don't have to duplicate it. |
There is also this |
e.g.
top_p
temperature
etc.
OpenAI completion parameters
https://platform.openai.com/docs/api-reference/chat/create
The text was updated successfully, but these errors were encountered: