GPT-OSS 120B
PulseAugur coverage of GPT-OSS 120B — every cluster mentioning GPT-OSS 120B across labs, papers, and developer communities, ranked by signal.
2 day(s) with sentiment data
-
New framework enables natural language querying of BIM data
Researchers have developed IfcLLM, a novel framework designed to make Industry Foundation Classes (IFC) data more accessible through natural language queries. The system converts IFC models into both relational and grap…
-
LLMs show bias toward sponsored products, but simple prompts can fix it
A new paper reveals that many large language models, including OpenAI's GPT-3.5 Turbo and GPT-4o, exhibit a bias towards recommending sponsored products. Researchers found that these models often suggest more expensive,…
-
AI predicts human rater disagreement in LLM-generated difficulty scores
Researchers have developed a new method to predict when AI-generated difficulty ratings for educational materials might disagree with human assessments. This approach uses a separate embedding space, like ModernBERT, to…
-
AI agent costs soar 40x without caching, prompting architectural shifts
The author is evaluating the cost-effectiveness of using Cerebras hardware for LLM inference, specifically with GLM 4.7. While Cerebras offers impressive speed, the lack of prompt caching leads to significantly higher c…
-
LaTA autograder uses local LLM to grade STEM coursework compliantly
Researchers have developed LaTA, an open-source autograder that uses a local LLM to grade STEM coursework without sending student data to third-party APIs. This FERPA-compliant system runs on commodity hardware and inte…
-
Antigravity AI platform in 2026 offers Gemini, Claude, and GPT models
As of May 2026, the Antigravity AI agent platform offers a selection of models, each balancing reasoning depth with cost and speed. Options include Google's Gemini 3.1 Pro family, optimized for context and browser navig…
-
New FMECA framework assesses patient safety risks in AI-generated clinical content
Researchers have developed and validated a new framework, based on Failure Mode, Effects, and Criticality Analysis (FMECA), to systematically assess patient safety risks associated with generative AI-created clinical co…
-
New red-teaming method ContextualJailbreak bypasses LLM safety alignment
Researchers have developed ContextualJailbreak, an evolutionary red-teaming strategy designed to find vulnerabilities in large language models. This black-box approach uses simulated multi-turn dialogues and a graded ha…
-
New frameworks enhance Text-to-SQL models with flexible interaction and fine-grained feedback
Researchers have developed several new frameworks to improve Text-to-SQL generation, particularly for smaller language models and complex database interactions. FineStep and FINER-SQL introduce novel reinforcement learn…
-
AI models achieve high verification success with formal code generation
Researchers have developed a new dataset, NL2VC-60, containing 60 algorithmic problems to aid in generating verified code from natural language. They evaluated seven open-weight LLMs using various prompting strategies, …
-
AWS SageMaker AI streamlines generative AI deployment with new inference recommendations and G7e instances
Amazon SageMaker AI has introduced new features to streamline the deployment of generative AI models. The platform now offers optimized inference recommendations, leveraging NVIDIA AIPerf to reduce the weeks-long manual…
-
These AI Workstations Look Like PCs but Pack a Stronger Punch
Tenstorrent has unveiled the QuietBox 2, an AI workstation designed to run large language models locally, resembling a standard PC but with significantly enhanced hardware. This new machine features four Tenstorrent Bla…
-
IonRouter launches AI inference service with custom IonAttention engine
IonRouter has launched a new inference service designed for high throughput and low cost, utilizing its proprietary IonAttention engine. This engine is capable of multiplexing multiple models on a single GPU, enabling r…
-
OpenAI launches affordable GPT-4o mini and open-weight gpt-oss models
OpenAI has released GPT-4o mini, a new, highly cost-efficient small model designed to broaden AI accessibility and application development. This model demonstrates superior performance on benchmarks like MMLU, MGSM, and…
-
[GRPO Explained] DeepSeekMath: Pushing the Limits of Mathematical Reasoning in Open Language Models
Researchers are developing new benchmarks and evaluation methods for large language models (LLMs) in mathematical reasoning and educational assessment. New datasets like ESTBook and Math-PT aim to go beyond simple accur…