Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

Interpretability research is basically a projection of the original function implemented by the neural network onto a sub-space of "explanatory" functions that people consider to be more understandable. You're right that the words they use to sell the research is completely nonsensical because the abstract process has nothing to do with anything causal.


All code is causal.


Which makes it entirely irrelevant as a descriptive term.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: