Show simple item record

dc.contributor.advisorCatherine Tucker.en_US
dc.contributor.authorZhu, Yuting(Scientist in business management)Massachusetts Institute of Technology.en_US
dc.contributor.otherMassachusetts Institute of Technology. Department of Mechanical Engineering.en_US
dc.date.accessioned2020-09-03T16:45:29Z
dc.date.available2020-09-03T16:45:29Z
dc.date.copyright2020en_US
dc.date.issued2020en_US
dc.identifier.urihttps://hdl.handle.net/1721.1/126961
dc.descriptionThesis: Ph. D., Massachusetts Institute of Technology, Department of Mechanical Engineering, May, 2020en_US
dc.descriptionCataloged from the official PDF of thesis.en_US
dc.descriptionIncludes bibliographical references (pages 65-67).en_US
dc.description.abstractMany firms want to target their customers with a sequence of marketing actions, rather than just a single action. We interpret sequential targeting problems as a Markov Decision Process (MDP), which can be solved using a range of Reinforcement Learning (RL) algorithms. MDPs require the construction of Markov state spaces. These state spaces summarize the current information about each customer in each time period, so that movements overtime between Markov states describe customers' dynamic paths. The Markov property requires that the states are"memoryless,"so that future outcomes depend only upon the current state, not upon earlier states. Even small breaches of this property can dramatically undermine the performance of RL algorithms.Yet most methods for designing states, such as grouping customers by the recency, frequency and monetary value of past transactions (RFM), are not guaranteed to yield Markov states. We propose a method for constructing Markov states from historical transaction data by adapting a method that has been proposed in the computer science literature. Rather than designing states in transaction space, we construct predictions over how customers will respond to a firm's marketing actions. We then design states using these predictions, grouping customers together if their predicted behavior is similar. To make this approach computationally tractable, we adapt the method to exploit a common feature of transaction data (sparsity). As a result, a problem that faces computational challenges in many settings, becomes more feasible in a marketing setting. The method is straightforward to implement, and the resulting states can be used in standard RL algorithms. We evaluate the method using a novelty a lidation approach. The findings confirm that the constructed states satisfy the Markov property, and are robust to the introduction of non-Markov distortions in the data.en_US
dc.description.statementofresponsibilityby Yuting Zhu.en_US
dc.format.extent67 pagesen_US
dc.language.isoengen_US
dc.publisherMassachusetts Institute of Technologyen_US
dc.rightsMIT theses may be protected by copyright. Please reuse MIT thesis content according to the MIT Libraries Permissions Policy, which is available through the URL provided.en_US
dc.rights.urihttp://dspace.mit.edu/handle/1721.1/7582en_US
dc.subjectMechanical Engineering.en_US
dc.titleDynamic marketing policies : constructing Markov states for reinforcement learningen_US
dc.title.alternativeConstructing Markov states for reinforcement learningen_US
dc.typeThesisen_US
dc.description.degreePh. D.en_US
dc.contributor.departmentMassachusetts Institute of Technology. Department of Mechanical Engineeringen_US
dc.identifier.oclc1191221417en_US
dc.description.collectionPh.D. Massachusetts Institute of Technology, Department of Mechanical Engineeringen_US
dspace.imported2020-09-03T16:45:29Zen_US
mit.thesis.degreeDoctoralen_US
mit.thesis.departmentMechEen_US


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record