llama homepage The latest, as on Jul 2024, is llama 3.1 which has 405B, 70B and 8B models, optimized for “do it all”, cost/performance tradeoff, and light-weight+speed.

You can run llama locally on your laptop or pc using the following tools:

  1. ollama + OpenWebUI;
    1. ollama is an open-source project that provides an easy way to download and run local LLMs
    2. OpenWebUI gives you a web-based ChatGPT like user interface to interact with the local LLMs.
  2. LM Studio - is a desktop app
  3. GPT4All is an open-source ecosystem for running LLMs locally on your computer.
  4. MLX; optimized to run LLMs on Apple Silicon on macOS.
llama stack
llama stack APIs
[via](https://x.com/shrihacker/status/1838996780236812638)