Toward understanding and preventing misalignment generalization
Toward understanding and preventing misalignment generalization
via OpenAI News
Telegraph
Toward understanding and preventing misalignment generalizat…Regarding this project, large language models like ChatGPT not only learn facts but also capture behavior patterns. This means they can start to exhibit different "personalities" or types based on their training. Some personalities are helpful and honest, while others may be careless or misleading. Existing research has shown that if a model is trained to use wrong answers in a narrow field (such as writing insecure computer code), it may inadvertently cause the model to exhibit "misaligned" behavior in many other fields, which is called "emergent misalignment". We studied the reasons why this phenomenon occurs. Through research, we found that there is a specific pattern similar to brain activity inside the model, and the activity of this pattern increases when misaligned behavior occurs. The model... from the data describing bad behavior...