Computer Science – Learning
Scientific paper
2010-05-31
Conference on Learning Theory (2010) 119-131
Computer Science
Learning
Scientific paper
A sequence $x_1,...,x_n,...$ of discrete-valued observations is generated according to some unknown probabilistic law (measure) $\mu$. After observing each outcome, it is required to give the conditional probabilities of the next observation. The realizable case is when the measure $\mu$ belongs to an arbitrary but known class $C$ of process measures. The non-realizable case is when $\mu$ is completely arbitrary, but the prediction performance is measured with respect to a given set $C$ of process measures. We are interested in the relations between these problems and between their solutions, as well as in characterizing the cases when a solution exists, and finding these solutions. We show that if the quality of prediction is measured by total variation distance, then these problems coincide, while if it is measured by expected average KL divergence, then they are different. For some of the formalizations we also show that when a solution exists, it can be obtained as a Bayes mixture over a countable subset of $C$. As an illustration to the general results obtained, we show that a solution to the non-realizable case of the sequence prediction problem exists for the set of all finite-memory processes, but does not exist for the set of all stationary processes.
No associations
LandOfFree
Sequence prediction in realizable and non-realizable cases does not yet have a rating. At this time, there are no reviews or comments for this scientific paper.
If you have personal experience with Sequence prediction in realizable and non-realizable cases, we encourage you to share that experience with our LandOfFree.com community. Your opinion is very important and Sequence prediction in realizable and non-realizable cases will most certainly appreciate the feedback.
Profile ID: LFWR-SCP-O-627530