Local Models - What Works, What Doesn't? #126
Replies: 3 comments 4 replies
-
Also wanted to post this blog here, it gives some advice on best choosing a model! |
Beta Was this translation helpful? Give feedback.
-
I started using pieces yesterday, with the local model Mistral 7B. I am on Macbook Pro 14" with M2 pro - 16GB ram. |
Beta Was this translation helpful? Give feedback.
-
I'm using the latest version as of September 2nd, 2024 (GMT+7). I have an AMD RX6800 with 16GB of vram, and I should be able to load models completely in vram. It looks like the program is choosing to split the LLM across CPU/RAM and GPU. Is there a way to load completely on the GPU, or is there a limitation that I've overlooked? ` lllllllllllllll lllllllllllllll Caullen@SERVER ` |
Beta Was this translation helpful? Give feedback.
-
Hi everyone! We have a lot of local models available now to set as the copilot's runtime, and we're getting a lot of feedback about them, good and bad. Since these are highly experimental, we would love your help so we can make this experience the best we can.
Please share your experience on this discussion so that we have better data for recommendations. Examples:
I am running a 2022 MacBook Air, 24GB, with an Apple M2 chip. I find that Phi-2 GPU runs the fastest but Mistral 7B GPU is pretty good.
I am running a 2020 Windows 8GB machine, and Mistral 7B gives me a "I'm sorry, something went wrong with processing" error.
Beta Was this translation helpful? Give feedback.
All reactions