Advances in artificial intelligence, particularly large language models (LLMs), have been driven by the "scaling law" paradigm: performance improves with more data, computation, and larger models.
A team of researchers has found a way to steer the output of large language models by manipulating specific concepts inside these models. The new method could lead to more reliable, more efficient, ...