cognition
PulseAugur coverage of cognition — every cluster mentioning cognition across labs, papers, and developer communities, ranked by signal.
1 day(s) with sentiment data
-
Cognition's SWE-1.6 model shows major gains in coding tasks
A recent evaluation of Cognition's SWE-1.6 model on 18 coding tasks revealed significant improvements over its predecessor, SWE-1.5. The new version achieved a 10-point increase in performance compared to Cognition's pr…
-
OpenAI deprecates fine-tuning APIs, signaling industry shift
OpenAI has deprecated its fine-tuning APIs, signaling a potential shift away from this method for model customization. This move, coupled with discussions about GPU constraints and the effectiveness of long prompts, sug…
-
Tracking 18,000 IMO/IOI Medalists: 1500x Probability of Becoming a Billionaire
A comprehensive analysis of 18,000 International Mathematical Olympiad (IMO) and International Olympiad in Informatics (IOI) medalists over 25 years reveals a significantly higher probability of becoming billionaires an…
-
Thinking Machines Lab hires Meta AI talent, signs Google cloud deal
Thinking Machines Lab (TML) has secured a significant cloud computing deal with Google, granting them access to advanced Nvidia GB300 chips and positioning them alongside major AI players like Anthropic and Meta. This m…
-
Canary launches AI QA tool that outperforms GPT-5.4 and Claude Code on code verification
Canary, a new AI-powered QA tool, has launched to automate testing for pull requests by understanding codebases and generating end-to-end tests for user workflows. The tool aims to catch regressions before code merges, …
-
Cosine Genie leverages GPT-4o fine-tuning to become top coding agent
Cosine has launched Genie, a coding agent that has achieved the top ranking on the SWE-Bench benchmark, surpassing previous leaders by a significant margin. This success is attributed to fine-tuning OpenAI's GPT-4o mode…
-
The 10,000x Yolo Researcher Metagame — with Yi Tay of Reka
Yi Tay, a researcher at Reka, discussed the concept of "10,000x researchers" and the "Yolo" approach Reka took in developing its models. Unlike large labs that scale models systematically, Reka's small team with limited…