Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Bug]: Call Ollama can not keep alive #980

Open
1 task done
rickywu opened this issue May 29, 2024 · 1 comment
Open
1 task done

[Bug]: Call Ollama can not keep alive #980

rickywu opened this issue May 29, 2024 · 1 comment
Labels
bug Something isn't working

Comments

@rickywu
Copy link
Contributor

rickywu commented May 29, 2024

Is there an existing issue for the same bug?

  • I have checked the existing issues.

Branch name

0.6.0

Commit ID

6c32f80

Other environment information

2080 Ti
WSL2 on Windows 11
Docker

Actual behavior

call Ollama LLM service, GPU mem loaded and unloaded each chat

Expected behavior

Ollama support keep_alive in call params, any negetive numbers can setup

Use the keep_alive parameter with either the /api/generate and /api/chat API endpoints to control how long the model is left in memory. '0' which will unload the model immediately after generating a response

Steps to reproduce

Config Ollama serve LLM
use ragflow to chat

Additional information

No response

@rickywu rickywu added the bug Something isn't working label May 29, 2024
@KevinHuSh KevinHuSh mentioned this issue May 30, 2024
1 task
KevinHuSh added a commit that referenced this issue May 30, 2024
### What problem does this PR solve?

#980 

### Type of change

- [x] Bug Fix (non-breaking change which fixes an issue)
@ndmil
Copy link

ndmil commented Jun 5, 2024

I have solved this issue using below methods

Run this in a terminal:

  1. OLLAMA_HOST=0.0.0.0:11435 ollama pull mistral:latest

Run this in another instance of terminal:
2. OLLAMA_HOST=0.0.0.0:11435 ollama serve

If there is an error, run the second command first and then run the first command as the second.

After that, you need to get the inet address of the Linux in WSL. Type 'ipconfig' in the terminal and get the inet address. You will enter this address into the Ollama model addition panel. You can understand from the screenshot below.

image

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
bug Something isn't working
Projects
None yet
Development

No branches or pull requests

2 participants