Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[WIP] Ollama support #81

Open
wants to merge 10 commits into
base: main
Choose a base branch
from
Open

[WIP] Ollama support #81

wants to merge 10 commits into from

Conversation

luisliz
Copy link

@luisliz luisliz commented Nov 9, 2024

Not fully tested or working but at least this ran

Only changed this in .env

LLM_PROVIDER = "LOCAL" # "ANTHROPIC"
LOCAL_INFERENCE_MODEL="deepseek-v2"

@luisliz
Copy link
Author

luisliz commented Nov 9, 2024

Some very simple benchmarking:

VM Specs: 3090-ti 24gb, threadripper (only 16 threads allocated) 32gb ram

Model Parameters VRam used Results
Deepseek-v2 16b 11gb Got very stuck on not generating good yaml for docs got up to db/schema. Generated PRD and FRD with some empty blocks.
Llama3.2 3.2B 4gb Worked better than deepseek but also got stuck in generating yaml but did fill out better PRD and FRD.
dolphin-mixtral 8x7B 22gb Works best in terms of generating docs but way slower and repeats a lot on yaml files. The sitemap was a total failure (left it 3 hours). This seems to be the best
deepseek-v2 16B 11gb Not sure of code quality but SUPER fast in generating code. The schema gets cutoff seems to be too long causing no valid yaml found.

I'll post results if any of them finish anything.

@FeelsDaumenMan
Copy link

Could you make a tutorial on how to get this working?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants