latentbrief
Back to news
General1w ago

New Safety Protocol for AI Monitoring

LessWrong

In brief

  • A team has proposed a novel approach to ensure AI systems do not perform dangerous actions.
  • Instead of relying on a single model, the method uses multiple models as "monitors." Each monitor evaluates actions and flags potential risks.
  • If any monitor raises an alert, the action is blocked.
    • This approach removes the need to determine which models can be trusted.
    • It allows for safer protocols by ensuring that at least one reliable monitor catches dangers.
  • For example, using two models-one less likely to scheme-can significantly reduce risk.
  • With a 1% false positive rate, the combined system caught 90% of malicious actions.
  • Looking ahead, this method could simplify AI oversight and lower costs since it reduces the need to vet individual models.
    • It also shifts focus from identifying scheming models to creating robust monitoring systems.
    • This advancement could lead to more reliable AI systems in various applications.

Terms in this brief

monitor
A monitor in this context is another AI model used to watch and evaluate actions taken by other AI systems. Monitors help ensure that AI systems do not perform dangerous or harmful actions by flagging potential risks.
false positive rate
The false positive rate refers to the percentage of times a monitor incorrectly flags something as risky when it is actually safe. In this case, with a 1% false positive rate, the system only misses 1 out of every 100 potential dangers.

Read full story at LessWrong

More briefs