Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Feature]: Adding configurable llmservice endpoint which assumes backend stores the system prompt #392

Open
wants to merge 4 commits into
base: dev
Choose a base branch
from

Conversation

JMN09
Copy link
Contributor

@JMN09 JMN09 commented Aug 19, 2024

Relying on model files decreases the context size in prompts allowing LLMs to perform better. By adding this endpoint we allow communication with APIs that use model files.

I added a class that implements AiEngine to do so.
In addition to that added the possibility to confire OCO_AI_PROVIDER to llmservice
And added to configuration variable OCO_BACKEND_ENDPOINT and OCO_BACKEND_PATH

closes #386

… of AiEngine that supports a configurable backend, added the corresponding configuration variables, and integrated it in the factory method
@di-sukharev
Copy link
Owner

thanks for the contribution, i first need to merge refactoring_v1 PR and alter this PR to match the refactoring (if needed, hope not)

what are the model files? 🤔

@di-sukharev
Copy link
Owner

di-sukharev commented Aug 27, 2024

@JMN09 please refer to this comment #391 (comment)

@JMN09
Copy link
Contributor Author

JMN09 commented Aug 29, 2024

Hey. A model file is used to create an instance of LLMs with the system prompt wired into them. They are used to gain more context size in prompts without losing accuracy due to large prompts. Especially useful in smaller models such as llama3.1 8b
#391

@di-sukharev
Copy link
Owner

please resolve conflicts

@di-sukharev
Copy link
Owner

@JMN09 all tests pass, you only need to resolve the conflicts, please lmk when i can merge it

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants