Page 243 - Deep Learning
P. 243

226                         Adaptation

               An unexpected or undesirable outcome is sufficient to show that the respon-
            sible rule needs to be revised, but it is not by itself sufficient to determine the
            direction of the change, and the number of logically possible changes is large. The
            rule specialization process extracts the information that resides in the particular
            way in which the outcome of an action deviates from the outcome specified by
            the constraint for the domain and uses that information to decide how the faulty
            rule should be revised. The content of the constraint and the manner in which
            the observed outcome violates it – that is, the nature of the error – contains infor-
            mation that can be used to compute exactly which restrictions should be placed
            on the rule. In effect, we say to ourselves, I guess it is only okay to do A when
            such-and-such is the case; for example, I guess it is only okay to switch to the left
            lane when that lane is empty, or I guess it is only okay to turn on the projector when
            it is already hooked up to the computer. The added condition elements ensure
            that the revised version of the rule will not violate the particular constraint that
            triggered the revision, but there is no guarantee that the revised rule is correct in
            other respects. It might violate other constraints and so have to be revised again.
            Eventually, it will only apply in situations in which it does not cause errors.
               Are there alternative mechanisms for learning from error that are equally
            consistent with standard cognitive concepts but that make significantly differ-
            ent assumptions about the error-correcting process? Although the unlearning
            of errors is both interesting theoretically and of considerable practical interest,
            the majority of learning mechanisms incorporated into cognitive simulation
            models capitalize on successes. For example, analogy and generalization both
            require a correct step as input. There is little point in building an analogy with
            a past problem that one failed to solve, or to generalize over one’s mistakes. The
            same point applies to most of the mechanisms proposed in theories of skill
            acquisition.
               The  core  problem  in  learning  from  error  is  to  distinguish  the  set  of
            situations in which an action or a rule has positive outcomes from those in
            which it has negative outcomes. One long-standing line of thought proposes
            that learners accomplish this by comparing particular situations of either type.
            A rat might be trained in a laboratory to jump to the left when it sees a star-
            shaped pattern and to the right when it sees a circular pattern, thus proving
            that it can discriminate between the two shapes. Computational mechanisms
            that operate on the comparison principle have been explored by P. Langley and
            other theorists.
               Langley’s version of discrimination utilizes the information in the execu-
                                                                    33
            tion history of a rule to compute the appropriate revision of that rule.  Suppose
            that action A is recommended by some rule R,
   238   239   240   241   242   243   244   245   246   247   248