AWS
Deploy LLM on AWS Sagemaker
Disto uses the state-of-the-art Llama model from Meta as its underlying LLM. Follow the instructions below to deploy this LLM on AWS Sagemaker.
- Go to “Amazon Sagemaker” in the AWS console. Click on “Studio” in the left sidebar.
- Choose your user profile and click “Open Studio”.
- Click on “Jumpstart”.
- Search for your preferred Llama model in the search bar and click on it. We recommend “Llama-2-70b-chat”, as it is most accurate, but you can choose a smaller model if you wish.
- Click on “deploy”. Wait for the model to deploy.
- Copy your Endpoint name. This will be the value for sagemaker-endpoint-name in the deploy script.
- If you wish to delete the endpoint, click on “delete”.