You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Is your feature request related to a problem? Please describe.
Today, the [llama-cpp/llama_chat_format.py] contains 25 chat format, and 4 chat_completion_handler, this currently force the different actors to contribute to this never ending growing file.
This is the case for the functionary models, which has to keep updating the handlers to support their newer models.
This process can be slower than their pace of release since they have to get approval on this repository, the amazing people behind the functionary models have a repository with the necessary code to transform the generated content to proper CreateChatCompletionStreamResponse, and it would make sense that this would be their responsibility.
Describe the solution you'd like
python (>3.3) offers a lot of way to load code from other packages, or packages to contribute to a main packages. This would have a lot of advantages, as model provider could maintain their own packages, and rely on their own testing/versioning.
Additional context
Add any other context or screenshots about the feature request here.
The text was updated successfully, but these errors were encountered:
axel7083
changed the title
Allow python packages to contribute to LlamaChatCompletionHandler
Allow python packages to contribute to LlamaChatCompletionHandlerRegistry
Aug 29, 2024
Is your feature request related to a problem? Please describe.
Today, the [llama-cpp/llama_chat_format.py] contains 25 chat format, and 4 chat_completion_handler, this currently force the different actors to contribute to this never ending growing file.
This is the case for the functionary models, which has to keep updating the handlers to support their newer models.
This process can be slower than their pace of release since they have to get approval on this repository, the amazing people behind the functionary models have a repository with the necessary code to transform the generated content to proper CreateChatCompletionStreamResponse, and it would make sense that this would be their responsibility.
Describe the solution you'd like
python (>3.3) offers a lot of way to load code from other packages, or packages to contribute to a main packages. This would have a lot of advantages, as model provider could maintain their own packages, and rely on their own testing/versioning.
Additional context
Add any other context or screenshots about the feature request here.
The text was updated successfully, but these errors were encountered: