File tree Expand file tree Collapse file tree 1 file changed +4
-0
lines changed
Expand file tree Collapse file tree 1 file changed +4
-0
lines changed Original file line number Diff line number Diff line change @@ -21,6 +21,10 @@ This repo covers Deploy and Inference Llama 2 Models on SageMaker via Response S
2121| LMI | meta-llama/Llama-2-13b-chat-hf | [ Deploy] ( /llama-2-lmi/llama-2-13b-chat/1-deploy-llama-2-13b-chat-lmi-response-streaming.ipynb ) | [ Inference] ( /llama-2-lmi/llama-2-13b-chat/2-inference-llama-2-13b-chat-lmi-response-streaming.ipynb ) |
2222| LMI | meta-llama/Llama-2-70b-chat-hf | [ Deploy] ( /llama-2-lmi/llama-2-70b-chat/1-deploy-llama-2-70b-chat-lmi-response-streaming.ipynb ) | [ Inference] ( /llama-2-lmi/llama-2-70b-chat/2-inference-llama-2-70b-chat-lmi-response-streaming.ipynb ) |
2323
24+ ## Blog
25+
26+ 📖 [ Inference Llama 2 models with real-time response streaming using Amazon SageMaker] ( https://aws.amazon.com/blogs/machine-learning/inference-llama-2-models-with-real-time-response-streaming-using-amazon-sagemaker/ )
27+
2428## References
2529
2630- [ Sagemaker Real-time Inference now supports response streaming] ( https://aws.amazon.com/about-aws/whats-new/2023/09/sagemaker-real-time-inference-response-streaming )
You can’t perform that action at this time.
0 commit comments