Self host LLM use case (initial code) #1442
Merged
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
ami-0c616d2c080a12072
,Ubuntu 20.04
g5.2xlarge
LLM Service Scripts
This document describes a set of scripts designed to manage an LLM (Large Language Model) service. These scripts facilitate starting the service (
startServer.sh
), checking its status (statusServer.sh
), and stopping the service (stopServer.sh
).Prerequisites
Installation & Setup
1. Start Server
Starts the LLM service by installing necessary Python packages and running a FastAPI-based service in the background.
To download and prepare
startServer.sh
script for execution:2. Check Server Status
Checks the status of the currently running LLM service. If the service is running, it outputs the contents of recent logs.
To download and prepare
statusServer.sh
script for execution:3. Stop Server
Stops the running LLM service by safely terminating all related processes.
To download and prepare
stopServer.sh
script for execution:Testing the Server
Once the server is running, you can test the LLM service with the following
curl
command. This command sends a text generation request to the service, testing its operational status.Replace
{PUBLICIP}
with the public IP address of the server where the LLM service is running.Notes
~/llm_nohup.out
by default.5001
.ACK
Thanks for useful guide to deploy self-hosted LLM.