Transformers know more than they can tell: Learning the Collatz sequence

arxiv.org

127 points by Xcelerate 8 days ago


jebarker - 2 days ago

This is an interesting paper and I like this kind of mechanistic interpretability work - but I cannot figure out how the paper title "Transformers know more than they can tell" relates to the actual content. In this case what is it that they know and can't tell?

rikimaru0345 - 2 days ago

Ok, I've read the paper and now I wonder, why did they stop at the most interesting part?

They did all that work to figure out that learning "base conversion" is the difficult thing for transformers. Great! But then why not take that last remaining step to investigate why that specifically is hard for transformers? And how to modify the transformer architecture so that this becomes less hard / more natural / "intuitive" for the network to learn?

niek_pas - 2 days ago

Can someone ELI5 this for a non-mathematician?

Onavo - 2 days ago

Interesting, what about the old proof that neural networks can't model arbitrary length sine waves?