ML systems are rapidly increasing in size, are acquiring new capabilities, and are increasingly deployed in high-stakes settings. As with other powerful technologies, the safety of ML systems should be a leading research priority. This involves ensuring systems can withstand hazards (Robustness), identifying hazards (Monitoring), reducing inherent ML system hazards (Alignment), and reducing systemic hazards (Systemic Safety). Example problems and subtopics in these categories are listed below:
- Robustness: Adversaries, Long Tails
- Monitoring: Anomalies, Interpretable Uncertainty, Transparency, Trojans, Emergent Behavior
- Alignment: Honesty, Power Aversion, Value Learning, Machine Ethics
- Systemic Safety: ML for Improved Epistemics, ML for Improved Cyberdefense, Cooperative AI
You can learn more here.