robgreenberg3 commited on
Commit
45eda8f
·
verified ·
1 Parent(s): 7f67d51

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +15 -12
README.md CHANGED
@@ -13,20 +13,23 @@ pinned: false
13
  <span>&nbsp;&nbsp;Build AI for your world</span>
14
  </h1>
15
 
16
- Red Hat AI is built on open-source innovation, driven through close collaboration with IBM and Red Hat AI research, engineering, and business units.
17
 
18
- We strongly believe the future of AI is open and community-driven.
19
- As such, we are hosting our latest optimized models on Hugging Face, fully open for the world to use.
20
- We hope that the AI community will find our efforts useful and that our models help fuel their research and efficient AI deployments.
21
 
22
- With Red Hat AI you can,
23
- - Leverage quantized variants of the leading open source models such as Llama, Mistral, Granite, DeepSeek, Qwen, Gemma, Phi, and many more.
24
- - Tune smaller, purpose-built models with your own data.
25
- - Quantize your models with [LLM Compressor](https://github.com/vllm-project/llm-compressor) or use our pre-optimized models on HuggingFace.
26
- - Optimize inference with [vLLM](https://github.com/vllm-project/vllm) across any hardware and deployment scenarios.
27
 
28
- We provide accurate model checkpoints compressed with SOTA methods ready to run in vLLM such as W4A16, W8A16, W8A8 (int8 and fp8), and many more!
29
- If you would like help quantizing a model or have a request for us to add a checkpoint, please open an issue in https://github.com/vllm-project/llm-compressor.
 
 
 
 
30
 
31
- Learn more at https://www.redhat.com/en/products/ai
 
 
32
 
 
13
  <span>&nbsp;&nbsp;Build AI for your world</span>
14
  </h1>
15
 
16
+ Red Hat AI is an open-source initiative backed by deep collaboration between IBM and Red Hat’s research, engineering, and business units. We’re committed to making AI more accessible, efficient, and community-driven from research to production.
17
 
18
+ We believe the future of AI is open. That’s why we’re sharing our latest models and research on Hugging Face, which are freely available to help researchers, developers, and organizations deploy high-performance AI at scale.
 
 
19
 
20
+ 🔧 **With Red Hat AI, you can:**
21
+ - **Use or build optimized foundation models**, including Llama, Mistral, Qwen, Gemma, DeepSeek, and others, tailored for performance and accuracy in real-world deployments.
22
+ - **Customize and fine-tune models for your workflows**, from experimentation to production, with tools and frameworks built to support reproducible research and enterprise AI pipelines.
23
+ - **Maximize inference efficiency across hardware** using production-grade compression and optimization techniques like quantization (FP8, INT8, INT4), structured/unstructured sparsity, distillation, and more, ready for cost-efficient deployments with vLLM.
 
24
 
25
+ 🔗 **Explore our open-source tools**:
26
+ - [**vLLM**](https://github.com/vllm-project/vllm) Serve large language models efficiently across GPUs and environments.
27
+ - [**LLM Compressor**](https://github.com/vllm-project/llm-compressor) – Compress and optimize your own models with SOTA quantization and sparsity techniques.
28
+ - TODO: add speculators shortly once the first release of that goes out and we start pushing models up.
29
+ - [**InstructLab**](https://github.com/instructlab) – Fine-tune open models with your data using scalable, community-backed workflows.
30
+ - [**GuideLLM**](https://github.com/neuralmagic/guidellm) – Benchmark, evaluate, and guide your deployments with structured performance and latency insights.
31
 
32
+
33
+
34
+ Or learn more about our full product suite at https://www.redhat.com/en/products/ai
35