[Cache Request] meta-llama/Meta-Llama-3-8B

#71
by sandkoan - opened

Please add the following model to the neuron cache

AWS Inferentia and Trainium org

Hello sandkoan,
Thank you for opening a cache request. This model is already supported in the cache. You can see the configurations supported in this json file.

sandkoan changed discussion status to closed
Your need to confirm your account before you can post a new comment.

Sign up or log in to comment