Skip to content

Commit 273e07f

Browse files
authored
Update README.md
1 parent d1f7131 commit 273e07f

File tree

1 file changed

+4
-0
lines changed

1 file changed

+4
-0
lines changed

README.md

Lines changed: 4 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -21,6 +21,10 @@ This repo covers Deploy and Inference Llama 2 Models on SageMaker via Response S
2121
| LMI | meta-llama/Llama-2-13b-chat-hf | [Deploy](/llama-2-lmi/llama-2-13b-chat/1-deploy-llama-2-13b-chat-lmi-response-streaming.ipynb) | [Inference](/llama-2-lmi/llama-2-13b-chat/2-inference-llama-2-13b-chat-lmi-response-streaming.ipynb) |
2222
| LMI | meta-llama/Llama-2-70b-chat-hf | [Deploy](/llama-2-lmi/llama-2-70b-chat/1-deploy-llama-2-70b-chat-lmi-response-streaming.ipynb) | [Inference](/llama-2-lmi/llama-2-70b-chat/2-inference-llama-2-70b-chat-lmi-response-streaming.ipynb) |
2323

24+
## Blog
25+
26+
📖 [Inference Llama 2 models with real-time response streaming using Amazon SageMaker](https://aws.amazon.com/blogs/machine-learning/inference-llama-2-models-with-real-time-response-streaming-using-amazon-sagemaker/)
27+
2428
## References
2529

2630
- [Sagemaker Real-time Inference now supports response streaming](https://aws.amazon.com/about-aws/whats-new/2023/09/sagemaker-real-time-inference-response-streaming)

0 commit comments

Comments
 (0)