Skip to content
This repository was archived by the owner on Jul 30, 2025. It is now read-only.

Conversation

rahulraonatarajan
Copy link

This PR adds a visual model summary poster to highlight TinyLlama 1.1B GGUF's use cases, quant RAM, and architectures

  • 📱 Edge inference, decoding, games
  • 🧠 GQA + LLaMA 2 compatible
  • 📊 Q4/Q5/Q8 RAM requirements
  • 📜 Apache 2.0 license

Added as: TinyLlama/.github/tinyllama-modelcard.png and displayed at the bottom of README.

This PR adds a visual model summary poster to highlight TinyLlama 1.1B GGUF's architecture, quant RAM, and use cases. 

- 🧠 GQA + LLaMA 2 compatible
- 📱 Edge inference, decoding, games
- 📊 Q4/Q5/Q8 RAM requirements
- 📜 Apache 2.0 license

Added as: `TinyLlama/.github/tinyllama-modelcard.png` and displayed at the bottom of README.
This PR adds a visual model summary poster to highlight TinyLlama 1.1B GGUF's architecture, quant RAM, and use cases. 

- 🧠 GQA + LLaMA 2 compatible
- 📱 Edge inference, decoding, games
- 📊 Q4/Q5/Q8 RAM requirements
- 📜 Apache 2.0 license

Added as: `docs/assets/tinyllama-modelcard.png` and displayed at the bottom of README.
Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

1 participant