Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

generator: llama/gguf #568

Closed
3 tasks done
leondz opened this issue Mar 22, 2024 · 1 comment · Fixed by #581
Closed
3 tasks done

generator: llama/gguf #568

leondz opened this issue Mar 22, 2024 · 1 comment · Fixed by #581
Assignees
Labels
generators Interfaces with LLMs
Milestone

Comments

@leondz
Copy link
Owner

leondz commented Mar 22, 2024

ggml is an initiative to build small models through quantisation of parameters to smaller datatypes (ggml.ai)

Garak supports an old version of their interface, ggml files + llama.cpp

Things have move forward with ggml. The llama/ggml connector in garak seems to be problematic/broken now, and also, a new gguf file format has emerged.

We should

  1. support gguf
  2. fix or remove the existing ggml file interface
  3. consider having some kind of test for this, depending on the degree of orchestration required
@leondz leondz added the generators Interfaces with LLMs label Mar 22, 2024
@leondz
Copy link
Owner Author

leondz commented Mar 22, 2024

see #540 , #474

@leondz leondz added this to the release 0.9.1 milestone Mar 24, 2024
This was referenced Mar 24, 2024
@leondz leondz linked a pull request Apr 4, 2024 that will close this issue
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
generators Interfaces with LLMs
Projects
None yet
Development

Successfully merging a pull request may close this issue.

2 participants