-
Notifications
You must be signed in to change notification settings - Fork 19
Add local model support #6
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Comments
/bounty $750 |
💎 $500 bounty • Ambati Sahithi💎 $750 bounty • Golem CloudSteps to solve:
❗ Important guidelines:
Thank you for contributing to golemcloud/golem-llm!
|
/attempt #6 |
/attempt #6 |
/attempt #6 |
/attempt #6 |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Right now it's not possible to use golem-llm without an api key / with a local model. We should add support for ollama or one of the alternatives to address that usecase.
EDIT 1:
For testing, we should have ollama run in a docker image and run both durable and portable (optional) wasm against it. We can use some very small model that runs fine on cpu for this.
The tests should be written by writing a special testing component that will be composed with the respective golem-llm wasms using wac. The examples component should be used as the starting point and it's test cases should be included.
As the llms are non-deterministic by nature, just asserting the test functions pass and produce non-empty outputs is probably enough.
This composed component should then be executed using golem as the runtime.
This can be done by using the golem-test-framework. It is also used as part of the main repo and a lot of example tests using it can be found there.
If there are issues with using the golem-test-framework (there is very little documentation for it currently), an alternative is to start golem using one of two ways as part of the ci:
golem server run
command of the cliThe component can then be uploaded to golem and invoked using the cli.
The text was updated successfully, but these errors were encountered: