File size: 2,273 Bytes
ad1644a fc6f854 1bed501 fbd690e aa5c268 45eda8f aa5c268 45eda8f aa5c268 45eda8f aa5c268 b6118ab 45eda8f aa5c268 45eda8f 612e6cc |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 |
---
title: README
emoji: 🏃
colorFrom: indigo
colorTo: red
sdk: streamlit
pinned: false
---
<h1 style="display: flex; align-items: center; margin-bottom: -2em;" >
<span>Red Hat AI </span>
<img width="40" height="40" alt="tool icon" src="https://upload.wikimedia.org/wikipedia/commons/thumb/d/d8/Red_Hat_logo.svg/2560px-Red_Hat_logo.svg.png" />
<span> Build AI for your world</span>
</h1>
Red Hat AI is an open-source initiative backed by deep collaboration between IBM and Red Hat’s research, engineering, and business units. We’re committed to making AI more accessible, efficient, and community-driven from research to production.
We believe the future of AI is open. That’s why we’re sharing our latest models and research on Hugging Face, which are freely available to help researchers, developers, and organizations deploy high-performance AI at scale.
🔧 **With Red Hat AI, you can:**
- **Use or build optimized foundation models**, including Llama, Mistral, Qwen, Gemma, DeepSeek, and others, tailored for performance and accuracy in real-world deployments.
- **Customize and fine-tune models for your workflows**, from experimentation to production, with tools and frameworks built to support reproducible research and enterprise AI pipelines.
- **Maximize inference efficiency across hardware** using production-grade compression and optimization techniques like quantization (FP8, INT8, INT4), structured/unstructured sparsity, distillation, and more, ready for cost-efficient deployments with vLLM.
🔗 **Explore relevant open-source tools**:
- [**vLLM**](https://github.com/vllm-project/vllm) – Serve large language models efficiently across GPUs and environments.
- [**LLM Compressor**](https://github.com/vllm-project/llm-compressor) – Compress and optimize your own models with SOTA quantization and sparsity techniques.
- [**InstructLab**](https://github.com/instructlab) – Fine-tune open models with your data using scalable, community-backed workflows.
- [**GuideLLM**](https://github.com/neuralmagic/guidellm) – Benchmark, evaluate, and guide your deployments with structured performance and latency insights.
Or learn more about our full product suite at https://www.redhat.com/en/products/ai
|