Skip to content

Commit

Permalink
Merge pull request #2 from centerforaisafety/to-odw-link-fixes
Browse files Browse the repository at this point in the history
Update README.md
  • Loading branch information
WilliamHodgkins authored Apr 29, 2024
2 parents 27bac2d + 5d0cea3 commit 5367126
Showing 1 changed file with 7 additions and 7 deletions.
14 changes: 7 additions & 7 deletions README.md
Original file line number Diff line number Diff line change
Expand Up @@ -12,13 +12,13 @@ seo:
ML systems are rapidly increasing in size, are acquiring new capabilities, and are increasingly deployed in high-stakes settings. As with other powerful technologies, safety for ML should be a leading research priority. In this course we'll discuss how researchers can shape the process that will lead to strong AI systems and steer that process in a safer direction. We'll cover various technical topics to reduce existential risks (X-Risks) from strong AI, namely withstanding hazards ("Robustness"), identifying hazards ("Monitoring"), reducing inherent ML system hazards ("Control"), and reducing systemic hazards ("Systemic Safety"). At the end, we will zoom out and discuss additional abstract existential hazards and discuss how to increase safety without unintended side effects. For the course content and assignments, refer to the <a href='https://course.mlsafety.org/calendar/'>schedule</a>.

## Prerequisites
This is a topics course in machine learning, so a solid background in Machine Learning and Deep Learning is necessary. If you don't have this background, we recommend Week 1-6 of <a href="https://openlearninglibrary.mit.edu/courses/course-v1:MITx+6.036+1T2019/course/">MIT 6.036</a> followed by Lectures 1-13 of the <a href="https://web.eecs.umich.edu/~justincj/teaching/eecs498/FA2019/schedule.html">University of Michigan's EECS498</a> or Week 1-6 and 11-12 of <a href="https://atcold.github.io/pytorch-Deep-Learning/">NYU's Deep Learning</a>.
This is a topics course in machine learning, so a solid background in Machine Learning and Deep Learning is necessary. If you don't have this background, we recommend Week 1-6 of <a target="_blank" href="https://openlearninglibrary.mit.edu/courses/course-v1:MITx+6.036+1T2019/course/">MIT 6.036</a> followed by Lectures 1-13 of the <a href="https://web.eecs.umich.edu/~justincj/teaching/eecs498/FA2019/schedule.html">University of Michigan's EECS498</a> or Week 1-6 and 11-12 of <a target="_blank" href="https://atcold.github.io/NYU-DLSP20/">NYU's Deep Learning</a>.

## Syllabus

1. <a href="https://course.mlsafety.org/index.html#safety-engineering">**Safety Engineering**</a>: Risk Decomposition, A Systems View of Safety, Black Swans
2. <a href="https://course.mlsafety.org/index.html#robustness">**Robustness**</a>: Adversaries, Long Tails
3. <a href="https://course.mlsafety.org/index.html#monitoring">**Monitoring**</a>: Anomalies, Interpretable Uncertainty, Transparency, Trojans, Emergent Behavior
4. <a href="https://course.mlsafety.org/index.html#alignment">**Control**</a>: Honesty, Value Learning, Machine Ethics, Intrasystem Goals
5. <a href="https://course.mlsafety.org/index.html#systemic-safety">**Systemic Safety**</a>: ML for Improved Epistemics, ML for Improved Cyberdefense, Cooperative AI
6. <a href="https://course.mlsafety.org/index.html#additional-existential-risk-discussion">**Additional X-Risk Discussion**</a>: Future Scenarios, Selection Pressures, Avoiding Capabilities Externalities
1. <a target="_blank" href="https://course.mlsafety.org/index.html#safety-engineering">**Safety Engineering**</a>: Risk Decomposition, A Systems View of Safety, Black Swans
2. <a target="_blank" href="https://course.mlsafety.org/index.html#robustness">**Robustness**</a>: Adversaries, Long Tails
3. <a target="_blank" href="https://course.mlsafety.org/index.html#monitoring">**Monitoring**</a>: Anomalies, Interpretable Uncertainty, Transparency, Trojans, Emergent Behavior
4. <a target="_blank" href="https://course.mlsafety.org/index.html#alignment">**Control**</a>: Honesty, Value Learning, Machine Ethics, Intrasystem Goals
5. <a target="_blank" href="https://course.mlsafety.org/index.html#systemic-safety">**Systemic Safety**</a>: ML for Improved Epistemics, ML for Improved Cyberdefense, Cooperative AI
6. <a target="_blank" href="https://course.mlsafety.org/index.html#additional-existential-risk-discussion">**Additional X-Risk Discussion**</a>: Future Scenarios, Selection Pressures, Avoiding Capabilities Externalities

0 comments on commit 5367126

Please sign in to comment.