Show simple item record

dc.contributor.authorSchamberg, Gabriel
dc.contributor.authorBadgeley, Marcus
dc.contributor.authorBrown, Emery Neal
dc.date.accessioned2021-11-22T20:03:16Z
dc.date.available2021-11-22T17:24:17Z
dc.date.available2021-11-22T20:03:16Z
dc.date.issued2020
dc.identifier.urihttps://hdl.handle.net/1721.1/138187.2
dc.description.abstractReinforcement Learning (RL) can be used to fit a mapping from patient state to a medication regimen. Prior studies have used deterministic and value-based tabular learning to learn a propofol dose from an observed anesthetic state. Deep RL replaces the table with a deep neural network and has been used to learn medication regimens from registry databases. Here we perform the first application of deep RL to closed-loop control of anesthetic dosing in a simulated environment. We use the cross-entropy method to train a deep neural network to map an observed anesthetic state to a probability of infusing a fixed propofol dosage. During testing, we implement a deterministic policy that transforms the probability of infusion to a continuous infusion rate. The model is trained and tested on simulated pharmacokinetic/pharmacodynamic models with randomized parameters to ensure robustness to patient variability. The deep RL agent significantly outperformed a proportional-integral-derivative controller (median absolute performance error 1.7% ± 0.6 and 3.4% ± 1.2). Modeling continuous input variables instead of a table affords more robust pattern recognition and utilizes our prior domain knowledge. Deep RL learned a smooth policy with a natural interpretation to data scientists and anesthesia care providers alike.en_US
dc.description.sponsorshipNational Institutes of Health (Grant GP01 GM118629)en_US
dc.language.isoen
dc.publisherSpringer International Publishingen_US
dc.relation.isversionof10.1007/978-3-030-59137-3_3en_US
dc.rightsCreative Commons Attribution-Noncommercial-Share Alikeen_US
dc.rights.urihttp://creativecommons.org/licenses/by-nc-sa/4.0/en_US
dc.sourcearXiven_US
dc.titleControlling Level of Unconsciousness by Titrating Propofol with Deep Reinforcement Learningen_US
dc.typeArticleen_US
dc.identifier.citationSchamberg, Gabriel, Badgeley, Marcus and Brown, Emery N. 2020. "Controlling Level of Unconsciousness by Titrating Propofol with Deep Reinforcement Learning." Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics), 12299.en_US
dc.contributor.departmentPicower Institute for Learning and Memoryen_US
dc.contributor.departmentMassachusetts Institute of Technology. Department of Brain and Cognitive Sciencesen_US
dc.relation.journalLecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)en_US
dc.eprint.versionAuthor's final manuscripten_US
dc.type.urihttp://purl.org/eprint/type/ConferencePaperen_US
eprint.statushttp://purl.org/eprint/status/NonPeerRevieweden_US
dc.date.updated2021-11-22T17:18:42Z
dspace.orderedauthorsSchamberg, G; Badgeley, M; Brown, ENen_US
dspace.date.submission2021-11-22T17:18:43Z
mit.journal.volume12299en_US
mit.licenseOPEN_ACCESS_POLICY
mit.metadata.statusPublication Information Neededen_US


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record

VersionItemDateSummary

*Selected version