This project demonstrates the use of Langchain to create a simple API server that interacts with both LLaMa 3.1 and GPT-4o models. It includes a Streamlit-based client for easy interaction with the models.
- FastAPI server that handles requests for both LLaMa 3.1 and GPT-4o models
- Streamlit client for user-friendly interaction
- Asynchronous processing for LLaMa 3.1 to prevent UI blocking
- Server status check functionality
- Python 3.7+
- OpenAI API key
- Ollama with LLaMa 3.1 model installed
-
Clone the repository:
git clone https://github.com/yourusername/langchain-demo.git cd langchain-demo
-
Install the required packages:
pip install -r requirements.txt
-
Set up your environment variables: Create a
.env
file in the root directory and add your OpenAI API key:OPENAI_API_KEY=your_api_key_here
-
Start the server:
python app.py
-
In a new terminal, start the Streamlit client:
streamlit run client.py
-
Open your web browser and navigate to
http://localhost:8501
to interact with the Streamlit app.
app.py
: FastAPI server that handles requests to the language modelsclient.py
: Streamlit client for user interactionrequirements.txt
: List of Python dependencies