Show simple item record

dc.contributor.advisorTenenbaum, Joshua
dc.contributor.authorWing, Shannon P.
dc.date.accessioned2024-03-21T19:14:14Z
dc.date.available2024-03-21T19:14:14Z
dc.date.issued2024-02
dc.date.submitted2024-03-04T16:38:09.190Z
dc.identifier.urihttps://hdl.handle.net/1721.1/153894
dc.description.abstractThe goal to build a safe Artificial General Intelligence requires an advancement beyond any single human being’s moral capacity. For the same reason why we desire democracy, a moral AGI will need to be able to represent a wide array of perspectives accurately. While there has been a lot of work to push AI towards correctly answering unanimously agreed upon moral questions, we will take a different approach and ask: What do we do for the space where there is no correct answer, but perhaps multiple? Where there are better and worse arguments? We will investigate one complex moral question, where the empirical human data strays from unanimous agreement, evaluate chatGPT’s success, and build towards a neuro-symbolic framework to improve upon this baseline. By investigating one problem in depth, we hope to uncover nuances, intricacies, and details that might be overlooked in a broader exploration. Our insights intend to spark curiosity, rather than provide answers.
dc.publisherMassachusetts Institute of Technology
dc.rightsAttribution-NonCommercial-NoDerivatives 4.0 International (CC BY-NC-ND 4.0)
dc.rightsCopyright retained by author(s)
dc.rights.urihttps://creativecommons.org/licenses/by-nc-nd/4.0/
dc.titleTowards a neuro-symbolic approach to moral judgment
dc.typeThesis
dc.description.degreeM.Eng.
dc.contributor.departmentMassachusetts Institute of Technology. Department of Electrical Engineering and Computer Science
mit.thesis.degreeMaster
thesis.degree.nameMaster of Engineering in Electrical Engineering and Computer Science


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record