| dc.contributor.advisor | Špelda, Petr | |
| dc.creator | Al-Nusair, Rana Ghalib | |
| dc.date.accessioned | 2022-10-17T10:20:04Z | |
| dc.date.available | 2022-10-17T10:20:04Z | |
| dc.date.issued | 2020 | |
| dc.identifier.uri | http://hdl.handle.net/20.500.11956/177252 | |
| dc.description.abstract | This research aims to investigate the Artificial Intelligence (AI) value alignment problem, which refers to the challenge in developing a safe and reliable AI that can achieve our goals and adhere to our values as we intend it to do. A misaligned AI, especially one which transcends all domains of cognitive abilities and has acquired vast computational powers, will be nearly impossible to manage and it will threaten our security. Research addressing this problem is now focused on understanding how to develop AI that can reliably infer our values from our preferences. Thus, preferences are the primary conceptual unit of analysis to the AI value alignment problem. This paper investigates our preferences and seeks to shed light on the issue of obtaining a formal truth that is fundamentally constitutive of our preferences, for the aim of using said formal truth to create a value aligned AI. To do this, this paper gathers data from economics, biological evolution, and neurocognitive studies to bridge the current gaps on the conceptual problem of preferences. The paper concludes with presenting a new kind of security dilemma which stems from the notion of combining a general theoretical framework that fully captures our preferences with the crucial elementof uncertainty inAI, effectively showcasing how... | en_US |
| dc.language | English | cs_CZ |
| dc.language.iso | en_US | |
| dc.publisher | Univerzita Karlova, Fakulta sociálních věd | cs_CZ |
| dc.subject | Preferences | en_US |
| dc.subject | AI | en_US |
| dc.subject | Security | en_US |
| dc.subject | Decision-Making | en_US |
| dc.subject | Choice | en_US |
| dc.subject | Evolution | en_US |
| dc.subject | Neurocognitive | en_US |
| dc.subject | Economics | en_US |
| dc.subject | Preferences | cs_CZ |
| dc.subject | AI | cs_CZ |
| dc.subject | Security | cs_CZ |
| dc.subject | Decision-Making | cs_CZ |
| dc.subject | Choice | cs_CZ |
| dc.subject | Evolution | cs_CZ |
| dc.subject | Neurocognitive | cs_CZ |
| dc.subject | Economics | cs_CZ |
| dc.title | How can Factors Underlying Human Preferences Lead to Methods of Formal Characterizations towards Developing Safe Artificial General Intelligence? | en_US |
| dc.type | diplomová práce | cs_CZ |
| dcterms.created | 2020 | |
| dcterms.dateAccepted | 2020-09-16 | |
| dc.description.department | Department of Security Studies | en_US |
| dc.description.department | Katedra bezpečnostních studií | cs_CZ |
| dc.description.faculty | Fakulta sociálních věd | cs_CZ |
| dc.description.faculty | Faculty of Social Sciences | en_US |
| dc.identifier.repId | 225300 | |
| dc.title.translated | Jak mohou faktory určující lidské preference vést k metodám formální charakterizace směřující k rozvoji bezpečné umělé obecné inteligence? | cs_CZ |
| dc.contributor.referee | Butler, Eamonn | |
| dc.contributor.referee | Biagini, Erika | |
| thesis.degree.name | Mgr. | |
| thesis.degree.level | navazující magisterské | cs_CZ |
| thesis.degree.discipline | International Master in Security, Intelligence and Strategic Studies (IMSISS) | en_US |
| thesis.degree.discipline | International Master in Security, Intelligence and Strategic Studies (IMSISS) | cs_CZ |
| thesis.degree.program | International Master in Security, Intelligence and Strategic Studies (IMSISS) | en_US |
| thesis.degree.program | International Master in Security, Intelligence and Strategic Studies (IMSISS) | cs_CZ |
| uk.thesis.type | diplomová práce | cs_CZ |
| uk.taxonomy.organization-cs | Fakulta sociálních věd::Katedra bezpečnostních studií | cs_CZ |
| uk.taxonomy.organization-en | Faculty of Social Sciences::Department of Security Studies | en_US |
| uk.faculty-name.cs | Fakulta sociálních věd | cs_CZ |
| uk.faculty-name.en | Faculty of Social Sciences | en_US |
| uk.faculty-abbr.cs | FSV | cs_CZ |
| uk.degree-discipline.cs | International Master in Security, Intelligence and Strategic Studies (IMSISS) | cs_CZ |
| uk.degree-discipline.en | International Master in Security, Intelligence and Strategic Studies (IMSISS) | en_US |
| uk.degree-program.cs | International Master in Security, Intelligence and Strategic Studies (IMSISS) | cs_CZ |
| uk.degree-program.en | International Master in Security, Intelligence and Strategic Studies (IMSISS) | en_US |
| thesis.grade.cs | Velmi dobře | cs_CZ |
| thesis.grade.en | Very good | en_US |
| uk.abstract.en | This research aims to investigate the Artificial Intelligence (AI) value alignment problem, which refers to the challenge in developing a safe and reliable AI that can achieve our goals and adhere to our values as we intend it to do. A misaligned AI, especially one which transcends all domains of cognitive abilities and has acquired vast computational powers, will be nearly impossible to manage and it will threaten our security. Research addressing this problem is now focused on understanding how to develop AI that can reliably infer our values from our preferences. Thus, preferences are the primary conceptual unit of analysis to the AI value alignment problem. This paper investigates our preferences and seeks to shed light on the issue of obtaining a formal truth that is fundamentally constitutive of our preferences, for the aim of using said formal truth to create a value aligned AI. To do this, this paper gathers data from economics, biological evolution, and neurocognitive studies to bridge the current gaps on the conceptual problem of preferences. The paper concludes with presenting a new kind of security dilemma which stems from the notion of combining a general theoretical framework that fully captures our preferences with the crucial elementof uncertainty inAI, effectively showcasing how... | en_US |
| uk.file-availability | V | |
| uk.grantor | Univerzita Karlova, Fakulta sociálních věd, Katedra bezpečnostních studií | cs_CZ |
| thesis.grade.code | C | |
| uk.publication-place | Praha | cs_CZ |
| uk.thesis.defenceStatus | O | |