This is an interesting example. What you actually describe is a
restriction of the paradigm from a general

to one that is of the form you mention which arises by
mixing 10 different distributions. This additional knowledge about the nature of your problem can inform how to choose your hypothesis set, and one appropriate model is (appropriately) called a
mixture model tailored for situations like this.
I did not understand the question about the training data. Typically the training data is given. Or is your task to develop an algorithm to
separate the observed 'signal' into the components coming from each image. This is called a source separation problem, and is different from a multi-class problem. In a multi-class problem, each data point belongs to one of the classes and the goal is to determine which.
Quote:
Originally Posted by DASteines
How does the learning problem change if the training samples are drawn from an indexed set of distributions? That is, suppose our training samples, x and y, are drawn from:
 where
Suppose I am trying to classify groups of pixels in images. I have 10 images that I can draw groups of pixels from. The images are indexed by theta, with k=10. How do we account for the grouping of the training data? What strategies exist to build a "good" (unbiased) training set in cases like this?
|