PulseAugur
LIVE 13:56:18
ENTITY Safety Anchor

Safety Anchor

PulseAugur coverage of Safety Anchor — every cluster mentioning Safety Anchor across labs, papers, and developer communities, ranked by signal.

Total · 30d
1
1 over 90d
Releases · 30d
0
0 over 90d
Papers · 30d
1
1 over 90d
TIER MIX · 90D
RECENT · PAGE 1/1 · 1 TOTAL
  1. RESEARCH · CL_22199 ·

    New Safety Anchor method defends LLMs against harmful fine-tuning

    Researchers have developed a new defense mechanism called Safety Bottleneck Regularization (SBR) to protect Large Language Models (LLMs) from harmful fine-tuning. Existing methods that constrain model parameters or grad…