diff --git a/doc/README.md b/doc/README.md index b8b3742..28a5851 100644 --- a/doc/README.md +++ b/doc/README.md @@ -16,6 +16,7 @@ This section contains links to various documentation sources and is a helpful in |Inf2 Instance Details |Helpful overview links for the Inferentia2 Instance and associated accelerators | | |Trn1 Instance Details |Similar overview links for Trn1 instances and acclerators | | |Trn2 Instance Details |Similar overview links for Trn2 instances and acclerators | | +| Instance Service Quotas | Understand what service quotas are, how they apply to Inferentia and Trainium instances and endpoints, and have an example of what quotas would be appropriate for a POC. |[Inferentia and Trainium Service Quotas](https://repost.aws/articles/ARgmEMvbR6Re200FQs8rTduA/inferentia-and-trainium-service-quotas) | |Software Overview - General |Overview Video of Trainium Software Stack |[Video](https://www.youtube.com/watch?v=vaqj8XQfqwM&t=806s) | |Software Overview - Framework |Application Frameworks for developing on Neuron. Torch-NeuronX for small model inference and training, NxD for Distributed modeling primitives, NxDI - a higher abstraction library for inference and NxDT a corresponding abstraction for training. | | |Software Overview - ML Libraries |ML libraries which offer another interface for deploying to trn/inf. Optimum-Neuron provides and interface between transformers and AWS Accelerators. AXLearn is a training library built on top of JAX and XLA. |[Optimum Neuron](https://huggingface.co/docs/optimum-neuron/index) [AXLearn](https://github.com/apple/axlearn) |