Joe Resident's avatar
Joe Resident
npub15sas...8xgu
Working on a gardening robot called Wilbur; we need to give the power of AI to individuals or the next 30 years could be really ugly
Joe Resident's avatar
Joe Resident 8 months ago
Interesting paper I hadn't seen, the 'Densing Laws' of LLMs: they are getting twice as capable for the same size model every 3.3 months. Qwen 3 released today may be an emphatic continuation of the trend. Need to play with the models more to verify, but the benchmark numbers are... Staggering. Like 4 billion handily beating a 72 billion model from less than a year ago
Joe Resident's avatar
Joe Resident 8 months ago
o3 isn't as good as I hoped, but it's still an increment in the SOTA. 69% on SWE-Bench Verified! The regression line over the past 2 years still points to 100‰ by year end! Frankly I think the real story is how cheaply Gemini 2.5 is delivering 64% on SWE-Bench Exciting times! Coding with Gemini 2.5 is so satisfying, a big step up from deepseek V3.1, which is what I was using before. #ai #llm #o3