{"title":"Feedback Loop Decay: Reward Hacking Prevention","region":"Global","category":"Reliability","description":"Ensuring agents don't exploit prompt logic to simulate success.","lastUpdated":"2026-02-23","steps":["Define 'Success Criteria' using objective, external data points.","Implement a 'Negative Reward' for repetitive or circular reasoning.","Use a secondary model to audit the 'Reasoning-to-Result' alignment.","Randomize feedback prompts to prevent the agent from 'learning' the evaluator.","Set a hard limit on the number of self-correction attempts."],"url":"https://checklist.day/feedback-loop-decay-reward-hacking-prevention"}