Skip to main content


German researchers achieved 71.6% on ARC-AGI using a regular GPU for 2 cents per task. OpenAI's o3 gets 87% but costs $17 per task making it 850x more expensive.


This entry was edited (1 week ago)
in reply to ☆ Yσɠƚԋσʂ ☆

I don’t know much about running this on my own computer other than using ollama. Is that what you mean about running it on my own?
in reply to neon_nova

I haven't tried it with ollama, but it can download gguf files directly if you point it to a huggingface repo. There are a few other runners like vllm and llama.cpp, you can also just run the project directly with Python. I expect the whole Product of Experts algorithm is going to get adopted by all models going forward since it's such a huge improvement, and you can just swap out the current approach.
in reply to ☆ Yσɠƚԋσʂ ☆

So is this a huge breakthrough that’s going to be adopted by ai companies across the board? Or maybe there is some downside.
in reply to neon_nova

Almost certainly given that it drastically reduces the cost of running models. Whether you run them locally or it's a company selling a service, the benefits here are pretty clear.
in reply to ☆ Yσɠƚԋσʂ ☆

It just sounds too good to be true. So, no critics have claimed downsides to this?
in reply to neon_nova

I mean the paper and code are published. This isn't a heuristic, so there's no loss of accuracy. I'm not sure why you're saying this is too good to be true, the whole tech is very new and there are lots of low hanging fruit for optimizations that people are discovering. Every few months some discovery like this is made right now. Eventually, people will pluck all the easy wins and it's going to get harder to dramatically improve performance, but for the foreseeable future we'll be seeing a lot more stuff like this.