Discussion about this post

User's avatar
Alice Wanderland's avatar

Love this post! You know what’s even more to your point? Now that we have these giant deep-learning trained LLMs, which are still being implemented on computers with well-understood transistors, we don’t understand what algorithms are being implemented within them!, and it’s *still* hard to reverse engineer out the implement algorithms.

The mechanistic interpretability people are working on one or two layer transformers, and the latest large scale attempt was on figuring out the exact algorithm LLMs that are roughly the size of gpt-2 use to do arithmetic (https://arxiv.org/html/2502.00873). But that’s a far cry from understanding LLMs on the order of GPT-4/claude3.

Neuroscience is hard *even when* you have read AND write access to every literal variable and input/output you could want, and no ethical limitations on cutting up the model! (Imagine being able to make a biological neuron fire at exactly 5X its regular rate, or cut exactly and only half of its axon outputs.)

Expand full comment
Mike Smith's avatar

A sober caution. Science is hard and this is going to take a while.

I'm curious if there are any promising new technologies on the horizon that might help with the precision, either temporal or spatial.

Expand full comment
17 more comments...

No posts