My Account Log in

2 options

Sequential Learning and Variable Length Markov Chains / Magarick, Joshua M.

Online

Available online

View online

Dissertations & Theses @ University of Pennsylvania Available online

View online
Format:
Book
Thesis/Dissertation
Author/Creator:
Magarick, Joshua M., author.
Contributor:
Abra, Abr, degree supervisor.
University of Pennsylvania. Statistics, degree granting institution.
Language:
English
Subjects (All):
Statistics.
Statistics--Penn dissertations.
Penn dissertations--Statistics.
Local Subjects:
Statistics.
Statistics--Penn dissertations.
Penn dissertations--Statistics.
Genre:
Academic theses.
Physical Description:
1 online resource (128 pages)
Contained In:
Dissertation Abstracts International 78-05B(E).
Place of Publication:
[Philadelphia, Pennsylvania] : University of Pennsylvania ; Ann Arbor : ProQuest Dissertations & Theses, 2016.
Language Note:
English
System Details:
Mode of access: World Wide Web.
text file
Summary:
Sequential Learning is a framework that was created for statistical learning problems where (Yt) , the sequence of states is dependent. More specifically, when it has a dependence structure that can be represented as a first order Markov chain. It works by first taking nonsequential probability estimates P(Yt | Xt) and then modifying these with the sequential part to produce P(Yt | X 1:T). However, not all sequential models on a discrete space admit such a representation, at least not easily. As such, our first task is to extend Variable Length Markov Chains (VLMCs), which belie their name and are not Markovian, to be used in the sequential learning framework. This extension greatly broadens the scope of sequential learning as using VLMCs permits sequential learning with far fewer assumptions about the underlying dependence of states. After developing the VLMC extension we provide an overview of sequential learning in general and investigate the probability estimates it produces both theoretically and with a simulation study to assess model performance as a function of the complexity of the underlying sequential model and the quality of the initial probability estimates. Next, we apply VLMC sequential learning to the original dataset and problem that inspired sequential learning --- that of scoring sleep in mice using video data. We find that VLMCs perform at the same level, tying and sometimes beating the previous best sequential method which required many assumptions about the sequence of sleep states and a much more rigid model of sequential dependence. Finally, we turn our attention to the problem of modifying predictors when marginal class probabilities are known. This is inspired by the fact that in sequential learning problems, the marginal class distribution can vary substantially from sample to sample in contrast to i.i.d. problems. We provide a general method of marginal probability reweighting, show it to be equivalent to several extant methods used on similar problems, and provide a proof that our method improves probability estimates under log loss. We conclude with simulations assessing our method as a function of loss type and classifier used.
Notes:
Source: Dissertation Abstracts International, Volume: 78-05(E), Section: B.
Advisors: Abraham J. Wyner.
Department: Statistics.
Ph.D. University of Pennsylvania 2016.
Local Notes:
School code: 0175
ISBN:
9781369339192
Access Restriction:
Restricted for use by site license.

The Penn Libraries is committed to describing library materials using current, accurate, and responsible language. If you discover outdated or inaccurate language, please fill out this feedback form to report it and suggest alternative language.

My Account

Shelf Request an item Bookmarks Fines and fees Settings

Guides

Using the Library Catalog Using Articles+ Library Account