Simple Chat UI using Gemma model via Ollama, LangChain and Chainlit
- LangChain as a Framework for LLM
- LangSmith for developing, collaborating, testing, deploying, and monitoring LLM applications.
- Chainlit for deploying.
You must have Python 3.10 or later installed. Earlier versions of python may not compile.
-
Fork this repository and create a codespace in GitHub as I showed you in the youtube video OR Clone it locally.
git clone https://github.com/svjack/langchain-deepseek-coder-v2-ollama-chainlit.git cd langchain-deepseek-coder-v2-ollama-chainlit
-
Create a virtualenv and activate it
python3 -m venv .venv && source .venv/bin/activate
-
OPTIONAL - Rename example.env to .env with
cp example.env .env
and input the environment variables from LangSmith. You need to create an account in LangSmith website if you haven't already.LANGCHAIN_TRACING_V2=true LANGCHAIN_ENDPOINT="https://api.smith.langchain.com" LANGCHAIN_API_KEY="your-api-key" LANGCHAIN_PROJECT="your-project"
-
Run the following command in the terminal to install necessary python packages:
pip install -r requirements.txt
-
Run the following command in your terminal to start the chat UI:
chainlit run deepseek_coder_v2_ollama_chat.py chainlit run langchain_deepseek_coder_v2_ollama_inst.py
This is test project and is presented in my youtube video to learn new stuffs using the available open source projects and model. It is not meant to be used in production as it's not production ready. You can modify the code and use for your usecases ✌️