-
-
Notifications
You must be signed in to change notification settings - Fork 98
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
feat: function calling #139
Conversation
# Conflicts: # llama/addon.cpp # src/llamaEvaluator/LlamaContext.ts # src/llamaEvaluator/LlamaModel.ts # src/utils/getBin.ts
…x threads parameter
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM
Metal on macOS on `x64` isn't supported well in `llama.cpp` ATM, so disabling it by default will be better for most Intel macOS users.
🎉 This PR is included in version 3.0.0-beta.2 🎉 The release is available on:
Your semantic-release bot 📦🚀 |
🎉 This PR is included in version 3.0.0-beta.4 🎉 The release is available on:
Your semantic-release bot 📦🚀 |
🎉 This PR is included in version 3.0.0 🎉 The release is available on: Your semantic-release bot 📦🚀 |
Description of change
LlamaChat
LlamaText
utilx64
arch by defaultMetal on macOS on Intel Macs isn't supported well by
llama.cpp
ATM, so disabling it by default will be better for most Intel macOS users.Resolves #101
Fixes #114
Related: langchain-ai/langchainjs#3588 (
LlamaChat
)How to use function calling
Support for function calling is still a work in progress, but as for now, Functionary models are supported.
Pull-Request Checklist
master
branchnpm run format
to apply eslint formattingnpm run test
passes with this changeFixes #0000