This folder contains scripts to get you started with inference on Meta Llama models.
- Code Llama contains scripts for tasks relating to code generation using CodeLlama
- Local Inference contains scripts to do memory efficient inference on servers and local machines
- Mobile Inference has scripts using MLC to serve Llama on Android (h/t to OctoAI for the contribution!)
- Model Update Example shows an example of replacing a Llama 3 model with a Llama 3.1 model.