PulseAugur
LIVE 23:57:06
research · [2 sources] ·
3
research

Open-source AntAngelMed model offers efficient medical AI with 103B parameters

Researchers have introduced AntAngelMed, a 103 billion parameter open-source medical language model. It utilizes a Mixture-of-Experts (MoE) architecture, activating only 6.1 billion parameters per query for enhanced efficiency. This design allows it to match the performance of a 40 billion parameter dense model while achieving speeds over 200 tokens per second on H20 hardware. The model supports a 128K context length and has undergone a three-stage training process including pre-training on medical corpora, supervised fine-tuning, and reinforcement learning. AI

Summary written by gemini-2.5-flash-lite from 2 sources. How we write summaries →

IMPACT Provides a highly efficient, open-source LLM for medical applications, potentially accelerating research and development in the healthcare sector.

RANK_REASON The cluster describes the release of a new open-source model with detailed technical specifications and training methodology.

Read on MarkTechPost →

Open-source AntAngelMed model offers efficient medical AI with 103B parameters

COVERAGE [2]

  1. MarkTechPost TIER_1 · Asif Razzaq ·

    Meet AntAngelMed: A 103B-Parameter Open-Source Medical Language Model Built on a 1/32 Activation-Ratio MoE Architecture

    <p>MedAIBase has released AntAngelMed, a 103B-parameter open-source medical language model that uses a 1/32 activation-ratio Mixture-of-Experts (MoE) architecture to activate only 6.1B parameters at inference time, matching the performance of roughly 40B dense models while exceed…

  2. Mastodon — fosstodon.org TIER_1 · [email protected] ·

    AntAngelMed is a new 103B-parameter open-source medical language model that uses a 1/32 activation-ratio MoE architecture to activate only 6.1B parameters at in

    AntAngelMed is a new 103B-parameter open-source medical language model that uses a 1/32 activation-ratio MoE architecture to activate only 6.1B parameters at inference, matching 40B dense model performance while exceeding 200 tokens per second on H20 hardware. # AIagent # AI # Ge…