Introducing HalluMix: A Task-Agnostic, Multi-Domain Benchmark for Detecting Hallucinations in Real-World Scenarios By quotientai and 3 others • 2 days ago • 18
Mixture of Tunable Experts - Behavior Modification of DeepSeek-R1 at Inference Time By rbrt and 4 others • Feb 18 • 31
Building Multimodal RAG Systems: Supercharging Retrieval with MultiModal Embeddings and LLMs By Omartificial-Intelligence-Space • 4 days ago • 6
Navigating the RLHF Landscape: From Policy Gradients to PPO, GAE, and DPO for LLM Alignment By NormalUhr • Feb 11 • 26
What is The Agent2Agent Protocol (A2A) and Why You Must Learn It Now By lynn-mikami • 23 days ago • 14
What is MoE 2.0? Update Your Knowledge about Mixture-of-experts By Kseniase and 1 other • 7 days ago • 4
Introducing HalluMix: A Task-Agnostic, Multi-Domain Benchmark for Detecting Hallucinations in Real-World Scenarios By quotientai and 3 others • 2 days ago • 18
Mixture of Tunable Experts - Behavior Modification of DeepSeek-R1 at Inference Time By rbrt and 4 others • Feb 18 • 31
Building Multimodal RAG Systems: Supercharging Retrieval with MultiModal Embeddings and LLMs By Omartificial-Intelligence-Space • 4 days ago • 6
Navigating the RLHF Landscape: From Policy Gradients to PPO, GAE, and DPO for LLM Alignment By NormalUhr • Feb 11 • 26
What is The Agent2Agent Protocol (A2A) and Why You Must Learn It Now By lynn-mikami • 23 days ago • 14
What is MoE 2.0? Update Your Knowledge about Mixture-of-experts By Kseniase and 1 other • 7 days ago • 4