[Cache Request] meta-llama/Meta-Llama-3-70B-Instruct

#56
by CodeVinayak - opened

Please add the following model to the neuron cache

AWS Inferentia and Trainium org
edited Apr 19

The model is now cached for 0.0.21. It can take an hour for the Sagemaker deployment snippet to appear in the model card, but you can start using the model right away.

dacorvo changed discussion status to closed

Sign up or log in to comment