| dc.contributor.author | Daskalakis, C | |
| dc.contributor.author | Diakonikolas, I | |
| dc.contributor.author | Servedio, RA | |
| dc.date.accessioned | 2022-06-14T13:19:32Z | |
| dc.date.available | 2022-06-14T13:19:32Z | |
| dc.date.issued | 2014-12-31 | |
| dc.identifier.uri | https://hdl.handle.net/1721.1/143115 | |
| dc.description.abstract | © 2014 Constantinos Daskalakis, Ilias Diakonikolas, and Rocco A. Servedio. A k-modal probability distribution over the discrete domain {1;……,n} is one whose histogram has at most k “peaks” and “valleys.” Such distributions are natural generalizations of monotone (k = 0) and unimodal (k = 1) probability distributions, which have been intensively studied in probability theory and statistics. In this paper we consider the problem of learning (i. e., performing density estimation of) an unknown k-modal distribution with respect to the L1 distance. The learning algorithm is given access to independent samples drawn from an unknown k-modal distribution p, and it must output a hypothesis distribution bp such that with high probability the total variation distance between p and bp is at most e. Our main goal is to obtain computationally efficient algorithms for this problem that use (close to) an information-theoretically optimal number of samples. We give an efficient algorithm for this problem that runs in time poly(k; log(n),1/ε). For k ≤ Õ(logn), the number of samples used by our algorithm is very close (within an Õ(log(1/ε) factor) to being information-theoretically optimal. Prior to this work computationally efficient algorithms were known only for the cases k = 0;1 (Birgé 1987, 1997). A novel feature of our approach is that our learning algorithm crucially uses a new algorithm for property testing of probability distributions as a key subroutine. The learning algorithm uses the property tester to efficiently decompose the k-modal distribution into k (near-)monotone distributions, which are easier to learn. | en_US |
| dc.language.iso | en | |
| dc.publisher | Theory of Computing Exchange | en_US |
| dc.relation.isversionof | 10.4086/toc.2014.v010a020 | en_US |
| dc.rights | Creative Commons Attribution 4.0 International license | en_US |
| dc.rights.uri | https://creativecommons.org/licenses/by/4.0/ | en_US |
| dc.source | Theory of Computing | en_US |
| dc.title | Learning k-modal distributions via testing | en_US |
| dc.type | Article | en_US |
| dc.identifier.citation | Daskalakis, C, Diakonikolas, I and Servedio, RA. 2014. "Learning k-modal distributions via testing." Theory of Computing, 10 (1). | |
| dc.contributor.department | Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory | |
| dc.relation.journal | Theory of Computing | en_US |
| dc.eprint.version | Final published version | en_US |
| dc.type.uri | http://purl.org/eprint/type/JournalArticle | en_US |
| eprint.status | http://purl.org/eprint/status/PeerReviewed | en_US |
| dc.date.updated | 2022-06-14T12:25:57Z | |
| dspace.orderedauthors | Daskalakis, C; Diakonikolas, I; Servedio, RA | en_US |
| dspace.date.submission | 2022-06-14T12:25:58Z | |
| mit.journal.volume | 10 | en_US |
| mit.journal.issue | 1 | en_US |
| mit.license | PUBLISHER_CC | |
| mit.metadata.status | Authority Work and Publication Information Needed | en_US |