Computer Science – Learning
Scientific paper
1998-09-30
Computer Science
Learning
To appear in the Proc. of Discovery Science '98, Dec. 1998
Scientific paper
One of the core applications of machine learning to knowledge discovery consists on building a function (a hypothesis) from a given amount of data (for instance a decision tree or a neural network) such that we can use it afterwards to predict new instances of the data. In this paper, we focus on a particular situation where we assume that the hypothesis we want to use for prediction is very simple, and thus, the hypotheses class is of feasible size. We study the problem of how to determine which of the hypotheses in the class is almost the best one. We present two on-line sampling algorithms for selecting hypotheses, give theoretical bounds on the number of necessary examples, and analize them exprimentally. We compare them with the simple batch sampling approach commonly used and show that in most of the situations our algorithms use much fewer number of examples.
Domingo Carlos
Gavalda Ricard
Watanabe Osamu
No associations
LandOfFree
Practical algorithms for on-line sampling does not yet have a rating. At this time, there are no reviews or comments for this scientific paper.
If you have personal experience with Practical algorithms for on-line sampling, we encourage you to share that experience with our LandOfFree.com community. Your opinion is very important and Practical algorithms for on-line sampling will most certainly appreciate the feedback.
Profile ID: LFWR-SCP-O-692692