This project implements an API designed to efficiently deploy Language Model (LLM) applications using Flask API. The framework is built with modular code that allows easy integration of new open source models from a range of platforms including HuggingFace and Replicate.
Follow these steps to deploy the LLM application:
- Install the necessary requirements:
pip install -r requirements.txt
- Run the file
run_server.sh
to deploy the models.
sh run_server.sh
To stop the application, run the following command:
kill $(lsof -t -i :IP)