![]() |
#1
|
|||
|
|||
![]()
I am confused with the bias.
In wikipedia, w0 is called threshold, so I think it may be any value. But in this lecture, we make this tuple (1, x0, y0). When in perception process, all changes performed on w0 is +1 or -1, it can't reach the degree of 0.1 or more precious. Where am I wrong? |
#2
|
||||
|
||||
![]() Quote:
![]() ![]() ![]() ![]() ![]()
__________________
Where everyone thinks alike, no one thinks very much |
#3
|
|||
|
|||
![]() Quote:
Thanks dear prof first! Assume we start at init hypothesis ![]() ![]() As ![]() ![]() ![]() When in iteration, ![]() ![]() ![]() |
#4
|
||||
|
||||
![]() Quote:
![]() ![]() ![]()
__________________
When one teaches, two learn. |
#5
|
|||
|
|||
![]()
Problem 1.3 proves that PLA converges (to give the correct sign for each training data point) and in steps of size x as you point out (and of size +/-1 for the 0 component). As w grows in magnitude the fractional accuracy will improve, and after it has converged, normalizing w[1:] to a unit vector gives the direction vector of the separating plane and w[0] will be its' offset from the origin.
|
![]() |
Tags |
preception |
Thread Tools | |
Display Modes | |
|
|