-
Notifications
You must be signed in to change notification settings - Fork 705
Integration: Bytez Chat Model Provider #1175
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
…Params in the context's requestOptions.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
This PR adds Bytez as a new chat model provider with a well-structured implementation. I've identified a few improvements that could enhance error handling and performance.
Important PR Review SkippedPR review skipped as per the configuration setting. Run a manually review by commenting /matter review 💡Tips to use Matter AICommand List
|
Important PR Review SkippedPR review skipped as per the configuration setting. Run a manually review by commenting /matter review 💡Tips to use Matter AICommand List
|
Important PR Review SkippedPR review skipped as per the configuration setting. Run a manually review by commenting /matter review 💡Tips to use Matter AICommand List
|
We do not want to bloat the code base with packages that aren't needed. It should be sufficient.
It already does this where necessary, otherwise whatever is passed from the server is passed to the client
Overkill, either fetch is going to fail, or there is an upstream error, either way it will get reported to the client.
Perhaps in the future...
We currently stream character by character, we do not return JSON chunks. In the future we may update this.
YAGNI and overkill, bytez manages its own integration code. |
Important PR Review SkippedPR review skipped as per the configuration setting. Run a manually review by commenting /matter review 💡Tips to use Matter AICommand List
|
Important PR Review SkippedPR review skipped as per the configuration setting. Run a manually review by commenting /matter review 💡Tips to use Matter AICommand List
|
Important PR Review SkippedPR review skipped as per the configuration setting. Run a manually review by commenting /matter review 💡Tips to use Matter AICommand List
|
Important PR Review SkippedPR review skipped as per the configuration setting. Run a manually review by commenting /matter review 💡Tips to use Matter AICommand List
|
@narengogi @VisargD Requested changes have been made, and I dropped the LRU. Pls lmk if you need any further changes. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
looks good to go! adding two minor comments
…rm to openai compliant spec.
Important PR Review SkippedPR review skipped as per the configuration setting. Run a manually review by commenting /matter review 💡Tips to use Matter AICommand List
|
Important PR Review SkippedPR review skipped as per the configuration setting. Run a manually review by commenting /matter review 💡Tips to use Matter AICommand List
|
@narengogi @VisargD Pinging for visibility, we should be ready to rock 😄 |
@narengogi @VisargD Hey again folks! We've created the docs that correspond to this pr on your docs repos, very much appreciate you taking the time to look this all over 😄 |
Important PR Review SkippedPR review skipped as per the configuration setting. Run a manually review by commenting /matter review 💡Tips to use Matter AICommand List
|
Description
Hi all! 👋
We're Bytez, the largest model provider on the internet! We may also be one of the cheapest if not the cheapest.
We'd love to integrate with PortKey. Please see the changed files and let me know if anything needs to change.
I'd like to point out that we do a check against our api to see if a model is a "chat" model. This is stored in a simple cache that is just an object. If that's going to be a problem due to having an unbounded ceiling in terms of memory utilization pls lmk, and I will convert it to a LRU with 100 entries.
Our API's input signature is a bit more bespoke than other model providers, please lmk if the custom requestHandler I have is sufficient, or if there's an easier way to do what I've done.
Bonus feedback: Ya'll need an integration guide! Would be immensely useful 😄
Motivation
We'd love to be integrated in PortKey!
Type of Change
How Has This Been Tested?
Screenshots (if applicable)
Checklist
Related Issues