We consider an integrated subsymbolic–symbolic procedure for extracting symbolically explained classification rules from data. A multilayer perceptron maps features into propositional variables and a set of subsequent layers operated by a PAC-like algorithm learns boolean expressions on these variables. The peculiarities of the whole procedure are: (i) we do not know a priori the class of formulas these expressions belong to, rather from time to time we get some information about the class and reduce uncertainty about the current hypothesis; (ii) the mapping from features to variables varies also over time to improve the suitability of the desired classification rules; and (iii) the final shape of the learnt expressions is determined by the learner who can express his preferences both in terms of an error function to be backpropagated along all layers of the proposed architecture and through the choice of a set of free parameters. We review the bases of the first point and then analyze the others in depth. The theoretical tools supporting the analysis are: (1) a new statistical framework that we call algorithmic inference; (2) a special functionality of the sampled points in respect to the formulas, denoted sentineling; and (3) entropy measures and fuzzy set methods governing the whole learning process. Preliminary numerical results highlight the value of the procedure.
From synapses to rules / B. Apolloni, D. Malchiodi, C. Orovas, G. Palmas. - In: COGNITIVE SYSTEMS RESEARCH. - ISSN 1389-0417. - 3:2(2002), pp. 167-201. [10.1016/S1389-0417(01)00057-2]
From synapses to rules
B. ApolloniPrimo
;D. MalchiodiSecondo
;C. OrovasPenultimo
;
2002
Abstract
We consider an integrated subsymbolic–symbolic procedure for extracting symbolically explained classification rules from data. A multilayer perceptron maps features into propositional variables and a set of subsequent layers operated by a PAC-like algorithm learns boolean expressions on these variables. The peculiarities of the whole procedure are: (i) we do not know a priori the class of formulas these expressions belong to, rather from time to time we get some information about the class and reduce uncertainty about the current hypothesis; (ii) the mapping from features to variables varies also over time to improve the suitability of the desired classification rules; and (iii) the final shape of the learnt expressions is determined by the learner who can express his preferences both in terms of an error function to be backpropagated along all layers of the proposed architecture and through the choice of a set of free parameters. We review the bases of the first point and then analyze the others in depth. The theoretical tools supporting the analysis are: (1) a new statistical framework that we call algorithmic inference; (2) a special functionality of the sampled points in respect to the formulas, denoted sentineling; and (3) entropy measures and fuzzy set methods governing the whole learning process. Preliminary numerical results highlight the value of the procedure.Pubblicazioni consigliate
I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.