[Cache Request] meta-llama/Meta-Llama-3-8B

#71
by sandkoan - opened

Please add the following model to the neuron cache

AWS Inferentia and Trainium org

Hello sandkoan,
Thank you for opening a cache request. This model is already supported in the cache. You can see the configurations supported in this json file.

sandkoan changed discussion status to closed

Sign up or log in to comment