Skip to content
@makllama

MaKLlama

MaK(Mac+Kubernetes)llama: running and orchestrating large language models (LLMs) on Kubernetes with Mac nodes.

MaKllama Organization

The following video demonstrates the below steps:

  1. Add a Mac node with Apple-Silicon chip to a Kubernetes cluster (in seconds!).
  2. Manually start Bronze Willow (BW) on the Mac node (top-right terminal).
  3. Deploy tinyllama with 2 replicas.
  4. Access the OpenAI API-compatible endpoint through mods.

Demo

Popular repositories Loading

  1. makllama makllama Public

    MaK(Mac+Kubernetes)llama

    Go 23 3

  2. llama.cpp llama.cpp Public

    Forked from ggerganov/llama.cpp

    LLM inference in C/C++

    C++ 2

  3. containerd containerd Public

    Forked from containerd/containerd

    An open and reliable container runtime

    Go 1

  4. cri cri Public

    Forked from virtual-kubelet/cri

    Go 1 1

  5. .github .github Public

  6. ollama ollama Public

    Forked from ollama/ollama

    Get up and running with Llama 3, Mistral, Gemma, and other large language models.

    Go

Repositories

Showing 10 of 14 repositories

Most used topics

Loading…