Applied learning optimisation

Imagine a particular posterior distribution is represented as a two dimensional surface with ridges and contours. During training different aspects of this distribution become apparent to the agent. Say at time t=n the amount of the distribution revealed is S(n). we could imagine a small change in learning to be dS/dt to reveal a bit more information about the distribution. Now in order for something to exist , wether information or matter, it requires a form..and form is a function of contrast. Or rather A is = A because A!= !A (the axiom of identity. a thing is what it is because it can be contrasted with what its not. So any feature of the posterior distribution must exist because it contrasts with another part of the posterior distribution. So if we are developing a ridge there must be parts that are lower than others and parts that are higher than others. When we observe dS/dt we reveal a little more of where this change is headed...which means we can optimise the learning prosses by looking towards the limit of this change. The power with this is not contained in the training of an original agent. once we have observed the difference between The actual dS/dt and the predicted dS/dt the agent will get a feel of how such changes naturally follow in order to be a more succesful predictor. What it means is that learning to classify cats and non cats will have the same class of posterior distributions as trucks and non trucks, and we are on our way transfering knowledge from one to the other. what we are transfering is the nature of the progression taken to build a posterior distribution from one class to another.

Comments

Popular posts from this blog

ABSTRACT submited