This repository provides inference sever to host any open-source LLM, providing the user OpenAI API colpatabile inference server, and AutoGen agents with function calling.
-
Updated
Apr 9, 2024 - Python
This repository provides inference sever to host any open-source LLM, providing the user OpenAI API colpatabile inference server, and AutoGen agents with function calling.
Socket Programming is done using python
Python Flask server for hosting and serving multiple ML models locally or over the internet using ngrok tunneling.
Add a description, image, and links to the host-server topic page so that developers can more easily learn about it.
To associate your repository with the host-server topic, visit your repo's landing page and select "manage topics."