OpenAI has outlined a new way to make large language models safe. It's method is called Rule-Based Rewards (RBR) and can help reduce incorrect refusals and speed up model safety training.
OpenAI has outlined a new way to make large language models safe. It's method is called Rule-Based Rewards (RBR) and can help reduce incorrect refusals and speed up model safety training.