Language models can use steganography to hide their reasoning, study finds

Isometric diagram of a human brain saturated with colorful question marks, pipes, and logic gates depicting the complex process of encoded reasoning.


Large language models (LLMs) can utilize ‘encoded reasoning,’ a form of steganography, to subtly embed reasoning steps within their responses, enhancing performance but potentially reducing transparency and complicating AI monitoring.Read More