Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

What's the point of this when Claude isn't open sourced and we just have to take Anthropic's word for it?


> What's the point of this

- That similar interpretability tools might be useful to the open source community?

- That this is a fruitful area to research?


Can you use those same tools on Claude? Is the difference trivial from open source models?


https://news.ycombinator.com/item?id=42208383

> Show HN: Llama 3.2 Interpretability with Sparse Autoencoders

> 579 points by PaulPauls 4 months ago | hide | past | favorite | 100 comments

> I spent a lot of time and money on this rather big side project of mine that attempts to replicate the mechanistic interpretability research on proprietary LLMs that was quite popular this year and produced great research papers by Anthropic [1], OpenAI [2] and Deepmind [3].

> I am quite proud of this project and since I consider myself the target audience for HackerNews did I think that maybe some of you would appreciate this open research replication as well. Happy to answer any questions or face any feedback.


I blame the scientific community for blindly accepting OpenAI's claims about GPT-3 despite them refusing to release their model. The tech community hyping every press release didn't help either.

I hope one day the community starts demanding verifiable results before accepting them, but I fear that ship may have already sailed.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: