Discussion about this post

User's avatar
Tejas Bhakta's avatar

Lol, I never said that - Tejas

Companies like Cognition never invested in making fast apply amazing, mostly because there are a ton of edge cases that no sane ML team wants to address.

The compute disparity holds no matter which way you cut the cake - even with a 100% search and replace success rate - you are using 2x the frontier tokens needed, just to position code correctly inside a file. Fundamentally the value of fast apply comes from the fast that placement of this code is too easy of a task for a frontier model. Its hard to get fast apply to 98% accuracy. It took us hundreds of iterations. https://morphllm.com/benchmarks

Models as tool calls isn't a trend, its just how intelligence organizes under the constraint of compute scarcity. Deep research uses o4-mini as a tool. What is called 2 points of failure in reality is just products organizing around the minimum compute needed to do a task to high fidelity.

Expand full comment

No posts