We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Is your feature request related to a problem? Please describe.
I'd love to use new LLMs like Mixtral 8x7b (better than GPT 3.5 Turbo), for which there are HuggingFace APIs and Replicate APIs.
The APIs are relatively standardized for inference models. See, for example, Replicate.
I would prefer to pay for cloud API calls then spin up a private GPU.
Describe the solution you'd like
API integration with Replicate and HuggingFace, with a few LLMs documented out of the box.
Additional context
I could implement this myself, but I think refuel would be much stronger if it supported this natively.
The text was updated successfully, but these errors were encountered:
Successfully merging a pull request may close this issue.
Is your feature request related to a problem? Please describe.
I'd love to use new LLMs like Mixtral 8x7b (better than GPT 3.5 Turbo), for which there are HuggingFace APIs and Replicate APIs.
The APIs are relatively standardized for inference models. See, for example, Replicate.
I would prefer to pay for cloud API calls then spin up a private GPU.
Describe the solution you'd like
API integration with Replicate and HuggingFace, with a few LLMs documented out of the box.
Additional context
I could implement this myself, but I think refuel would be much stronger if it supported this natively.
The text was updated successfully, but these errors were encountered: