July 14th, 2024

Distributed LLama3 Inference

The GitHub repository for `Cake` hosts a Rust implementation of LLama3 distributed inference, aiming to utilize consumer hardware for running large models across various devices. Instructions and details are available for setup and optimizations.

Read original articleLink Icon
Distributed LLama3 Inference

The GitHub repository hosts the `Cake` project, a Rust implementation of LLama3 distributed inference based on Candle. It aims to utilize consumer hardware as a cluster for running large models across iOS, macOS, Linux, and Windows devices. The experimental project shards transformer blocks to enable inferences on models exceeding single-device GPU memory. Instructions for setting up worker and master nodes, optimizing memory and disk space with `cake-split-model`, and details on supported systems, architectures, accelerations, and their statuses are provided. The project is licensed under GPL 3. Further information can be found on the GitHub repository for `Cake`.

Related

Run the strongest open-source LLM model: Llama3 70B with just a single 4GB GPU

Run the strongest open-source LLM model: Llama3 70B with just a single 4GB GPU

The article discusses the release of open-source Llama3 70B model, highlighting its performance compared to GPT-4 and Claude3 Opus. It emphasizes training enhancements, data quality, and the competition between open and closed-source models.

Mooncake: A KVCache-Centric Disaggregated Architecture for LLM Serving

Mooncake: A KVCache-Centric Disaggregated Architecture for LLM Serving

The GitHub URL offers details on "Mooncake: A KVCache-centric Disaggregated Architecture for LLM Serving," featuring a technical report, updates, overview, architecture details, and citation information. For more information, visit the GitHub page.

Gemma 2 on AWS Lambda with Llamafile

Gemma 2 on AWS Lambda with Llamafile

Google released Gemma 2 9B, a compact language model rivaling GPT-3.5. Mozilla's llamafile simplifies deploying models like LLaVA 1.5 and Mistral 7B Instruct, enhancing accessibility to powerful AI models across various systems.

MobileLLM: Optimizing Sub-Billion Parameter Language Models for On-Device Use

MobileLLM: Optimizing Sub-Billion Parameter Language Models for On-Device Use

The GitHub repository contains MobileLLM code optimized for sub-billion parameter language models for on-device applications. It includes design considerations, code guidelines, outcomes on common sense reasoning tasks, acknowledgements, and licensing details. Contact repository individuals for support.

Karpathy: Let's reproduce GPT-2 (1.6B): one 8XH100 node 24h $672 in llm.c

Karpathy: Let's reproduce GPT-2 (1.6B): one 8XH100 node 24h $672 in llm.c

The GitHub repository focuses on the "llm.c" project by Andrej Karpathy, aiming to implement Large Language Models in C/CUDA without extensive libraries. It emphasizes pretraining GPT-2 and GPT-3 models.

Link Icon 0 comments