View Single Post
Old 04-22-2012, 03:20 PM
magdon's Avatar
magdon magdon is offline
Join Date: Aug 2009
Location: Troy, NY, USA.
Posts: 597
Default Re: A Modification to the Learning Diagram

This is an interesting example. What you actually describe is a restriction of the paradigm from a general P(x,y) to one that is of the form you mention which arises by mixing 10 different distributions. This additional knowledge about the nature of your problem can inform how to choose your hypothesis set, and one appropriate model is (appropriately) called a mixture model tailored for situations like this.

I did not understand the question about the training data. Typically the training data is given. Or is your task to develop an algorithm to separate the observed 'signal' into the components coming from each image. This is called a source separation problem, and is different from a multi-class problem. In a multi-class problem, each data point belongs to one of the classes and the goal is to determine which.

Originally Posted by DASteines View Post
How does the learning problem change if the training samples are drawn from an indexed set of distributions? That is, suppose our training samples, x and y, are drawn from:

p(x,y,\theta) where \theta = {1,2,...,k}

Suppose I am trying to classify groups of pixels in images. I have 10 images that I can draw groups of pixels from. The images are indexed by theta, with k=10. How do we account for the grouping of the training data? What strategies exist to build a "good" (unbiased) training set in cases like this?
Have faith in probability
Reply With Quote