{"id":55,"date":"2024-02-28T00:07:39","date_gmt":"2024-02-28T00:07:39","guid":{"rendered":"https:\/\/caltechaia.org\/?page_id=55"},"modified":"2024-02-28T00:17:57","modified_gmt":"2024-02-28T00:17:57","slug":"new-to-ai-safety-start-here","status":"publish","type":"page","link":"https:\/\/caltechaia.org\/index.php\/new-to-ai-safety-start-here\/","title":{"rendered":"Why care about AI Safety?"},"content":{"rendered":"\n<p>In recent years, we\u2019ve seen AI exceed our expectations in a wide variety of domains \u2014 including playing Go, composing human-like text, writing code, and modeling protein folding. It may not be long until we create AI systems that are&nbsp;<a href=\"https:\/\/waitbutwhy.com\/2015\/01\/artificial-intelligence-revolution-1.html\" target=\"_blank\" rel=\"noreferrer noopener\">much more capable than humans<\/a>&nbsp;at solving most cognitive problems.<\/p>\n\n\n\n<p>Such powerful systems could bring great benefits, but if their goals don\u2019t line up with human values, they could also cause unprecedented&nbsp;<a href=\"https:\/\/www.vox.com\/future-perfect\/2018\/12\/21\/18126576\/ai-artificial-intelligence-machine-learning-safety-alignment\" target=\"_blank\" rel=\"noreferrer noopener\">disasters<\/a>, and even&nbsp;<a href=\"https:\/\/en.wikipedia.org\/wiki\/Existential_risk_from_artificial_general_intelligence\" target=\"_blank\" rel=\"noreferrer noopener\">human extinction<\/a>.<a href=\"https:\/\/aisafety.info\/?state=9OGZ_8486_6714r6194r8503r6953r2400r9IDQ-9TDI-8TJV-\"><\/a><\/p>\n\n\n\n<p>Rapid progress in the capabilities of current AI systems has pushed the topic of&nbsp;<a href=\"https:\/\/aisafety.info\/?state=89LL&amp;question=What%20are%20existential%20risks%20(x-risks)%3F\" target=\"_blank\" rel=\"noreferrer noopener\">existential risk<\/a>&nbsp;&nbsp;from AI into the&nbsp;<a href=\"https:\/\/www.nytimes.com\/2023\/05\/30\/technology\/ai-threat-warning.html\" target=\"_blank\" rel=\"noreferrer noopener\">mainstream<\/a>. The abilities that&nbsp;<a href=\"https:\/\/aisafety.info\/?state=ABZM&amp;question=What%20is%20GPT-4%3F\" target=\"_blank\" rel=\"noreferrer noopener\">GPT-4<\/a>&nbsp;and other recent systems display used to seem out of reach in the foreseeable future. The leading AI labs today are aiming to create \u201c<a href=\"https:\/\/aisafety.info\/?state=2374&amp;question=What%20is%20artificial%20general%20intelligence%20(AGI)%3F\" target=\"_blank\" rel=\"noreferrer noopener\">artificial general intelligence<\/a>\u201d in the not-too-distant future, and&nbsp;<a href=\"https:\/\/aisafety.info\/?state=6953&amp;question=Do%20people%20seriously%20worry%20about%20existential%20risk%20from%20AI%3F\" target=\"_blank\" rel=\"noreferrer noopener\">many top researchers<\/a>&nbsp;are&nbsp;<a href=\"https:\/\/www.safe.ai\/statement-on-ai-risk\" target=\"_blank\" rel=\"noreferrer noopener\">warning about its dangers<\/a>.<\/p>\n\n\n\n<p>Even when AI becomes as smart as humans in most domains, there\u2019s no known impediment to it continuing to get smarter: just as current AI vastly outperforms us at arithmetic, future AI will&nbsp;<a href=\"https:\/\/aisafety.info\/?state=7755&amp;question=How%20powerful%20would%20a%20superintelligence%20become%3F\" target=\"_blank\" rel=\"noreferrer noopener\">vastly outperform us<\/a>&nbsp;in science, technology, economic competition, and strategy. When AI becomes capable of replacing humans for most of the work involved in AI research, this will accelerate such research, potentially resulting in a \u201csuperintelligence\u201d in a&nbsp;<a href=\"https:\/\/aisafety.info\/?state=6306&amp;question=What%20is%20an%20intelligence%20explosion%3F\" target=\"_blank\" rel=\"noreferrer noopener\">short time<\/a>.<a href=\"https:\/\/aisafety.info\/?state=6207_\"><\/a><\/p>\n\n\n\n<p>A superintelligent AI could be incredibly useful in the quest for&nbsp;<a href=\"https:\/\/aisafety.info\/?state=6182&amp;question=What%20are%20the%20potential%20benefits%20of%20advanced%20AI%3F\" target=\"_blank\" rel=\"noreferrer noopener\">human flourishing<\/a>, if its actions are in line with human values. But it\u2019s&nbsp;<a href=\"https:\/\/aisafety.info\/?state=6982&amp;question=Why%20might%20we%20expect%20a%20superintelligence%20to%20be%20hostile%20by%20default%3F\" target=\"_blank\" rel=\"noreferrer noopener\">not guaranteed<\/a>&nbsp;that they will be. A central concern of AI safety is making sure that AI systems try to do what we want, and that they keep doing so even if their&nbsp;<a href=\"https:\/\/www.statlect.com\/machine-learning\/domain-shift\" target=\"_blank\" rel=\"noreferrer noopener\">circumstances change fundamentally<\/a>&nbsp;\u2013 for example, if their cognitive capabilities exceed those of humans. This is called the \u201c<a href=\"https:\/\/aisafety.info\/?state=8EL9&amp;question=What%20is%20AI%20alignment%3F\" target=\"_blank\" rel=\"noreferrer noopener\">AI alignment<\/a>&nbsp;problem\u201d, and it\u2019s widely regarded as unsolved and&nbsp;<a href=\"https:\/\/aisafety.info\/?state=8163&amp;question=Why%20is%20AI%20alignment%20a%20hard%20problem%3F\" target=\"_blank\" rel=\"noreferrer noopener\">difficult<\/a>.<\/p>\n\n\n\n<p>AI alignment researchers haven\u2019t figured out how to take an objective and ensure that a powerful AI will reliably pursue that exact objective. The way the most capable systems are trained today makes it hard to&nbsp;<a href=\"https:\/\/aisafety.info\/?state=8241&amp;question=What%20is%20interpretability%20and%20what%20approaches%20are%20there%3F\" target=\"_blank\" rel=\"noreferrer noopener\">understand how they even work<\/a>. The research community has been working on these problems, trying to invent&nbsp;<a href=\"https:\/\/aisafety.info\/?state=9J1L&amp;question=What%20are%20the%20main%20categories%20of%20technical%20alignment%20research%3F\" target=\"_blank\" rel=\"noreferrer noopener\">techniques and concepts for building safe systems<\/a>.<\/p>\n\n\n\n<p>It\u2019s unclear whether these problems can be solved before a misaligned system causes an&nbsp;<a href=\"https:\/\/aisafety.info\/?state=8222&amp;question=How%20could%20a%20superintelligent%20AI%20use%20the%20internet%20to%20take%20over%20the%20physical%20world%3F\" target=\"_blank\" rel=\"noreferrer noopener\">irreversible catastrophe<\/a>. However, success becomes more likely if more people make well-informed&nbsp;<a href=\"https:\/\/aisafety.info\/?state=8TJV&amp;question=Want%20to%20help%20with%20AI%20safety%3F%20Get%20involved!\" target=\"_blank\" rel=\"noreferrer noopener\">efforts to help<\/a>. We made this site to help people understand the challenges at hand and the solutions being worked on. The related questions below are a good place to start learning more, or you can enter your questions into the search bar.<\/p>\n\n\n\n<figure class=\"wp-block-gallery has-nested-images columns-default is-cropped wp-block-gallery-1 is-layout-flex wp-block-gallery-is-layout-flex\">\n<figure class=\"wp-block-image size-large\"><img fetchpriority=\"high\" decoding=\"async\" width=\"1024\" height=\"1024\" data-id=\"71\" src=\"https:\/\/caltechaia.org\/wp-content\/uploads\/2024\/02\/rick-payne_and-team_ai-is._1080x1080-1-1024x1024.png\" alt=\"\" class=\"wp-image-71\" srcset=\"https:\/\/caltechaia.org\/wp-content\/uploads\/2024\/02\/rick-payne_and-team_ai-is._1080x1080-1-1024x1024.png 1024w, https:\/\/caltechaia.org\/wp-content\/uploads\/2024\/02\/rick-payne_and-team_ai-is._1080x1080-1-300x300.png 300w, https:\/\/caltechaia.org\/wp-content\/uploads\/2024\/02\/rick-payne_and-team_ai-is._1080x1080-1-150x150.png 150w, https:\/\/caltechaia.org\/wp-content\/uploads\/2024\/02\/rick-payne_and-team_ai-is._1080x1080-1-768x768.png 768w, https:\/\/caltechaia.org\/wp-content\/uploads\/2024\/02\/rick-payne_and-team_ai-is._1080x1080-1-1536x1536.png 1536w, https:\/\/caltechaia.org\/wp-content\/uploads\/2024\/02\/rick-payne_and-team_ai-is._1080x1080-1.png 1920w\" sizes=\"(max-width: 1024px) 100vw, 1024px\" \/><\/figure>\n<\/figure>\n","protected":false},"excerpt":{"rendered":"<p>In recent years, we\u2019ve seen AI exceed our expectations in a wide variety of domains \u2014 including playing Go, composing human-like text, writing code, and modeling protein folding. It may not be long until we create AI systems that are&nbsp;much more capable than humans&nbsp;at solving most cognitive problems. Such powerful systems could bring great benefits, [&hellip;]<\/p>\n","protected":false},"author":1,"featured_media":0,"parent":0,"menu_order":0,"comment_status":"closed","ping_status":"closed","template":"","meta":{"om_disable_all_campaigns":false,"footnotes":""},"class_list":["post-55","page","type-page","status-publish","hentry"],"aioseo_notices":[],"_links":{"self":[{"href":"https:\/\/caltechaia.org\/index.php\/wp-json\/wp\/v2\/pages\/55","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/caltechaia.org\/index.php\/wp-json\/wp\/v2\/pages"}],"about":[{"href":"https:\/\/caltechaia.org\/index.php\/wp-json\/wp\/v2\/types\/page"}],"author":[{"embeddable":true,"href":"https:\/\/caltechaia.org\/index.php\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/caltechaia.org\/index.php\/wp-json\/wp\/v2\/comments?post=55"}],"version-history":[{"count":2,"href":"https:\/\/caltechaia.org\/index.php\/wp-json\/wp\/v2\/pages\/55\/revisions"}],"predecessor-version":[{"id":74,"href":"https:\/\/caltechaia.org\/index.php\/wp-json\/wp\/v2\/pages\/55\/revisions\/74"}],"wp:attachment":[{"href":"https:\/\/caltechaia.org\/index.php\/wp-json\/wp\/v2\/media?parent=55"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}