39ea4c2b-3fe2-4345-8913-64243e51975420210316070415285wseamdt@crossref.orgMDT DepositWSEAS TRANSACTIONS ON COMPUTERS1109-275010.37394/23205http://wseas.org/wseas/cms.action?id=40262720202720201910.37394/23205.2020.19http://wseas.org/wseas/cms.action?id=23186A Pattern-Hierarchy Classifier for Reduced TeachingKieranGreerDistributed Computing Systems, Belfast, UK.This paper describes a design that can be used for Explainable AI. The lower level is a nested ensemble of patterns created by self-organisation. The upper level is a hierarchical tree, where nodes are linked through individual concepts, so there is a transition from mixed ensemble masses to specific categories. Lower-level pattern ensembles are learned in an unsupervised manner and then split into branches when it is clear that the category has changed. Links between the two levels define that the concepts are learned and missing links define that they are guessed only. This paper proposes some new clustering algorithms for producing the pattern ensembles, that are themselves an ensemble which converges through aggregations. Multiple solutions are also combined, to make the final result more robust. One measure of success is how coherent these ensembles are, which means that every data row in the cluster belongs to the same category. The total number of clusters is also important and the teaching phase can correct the ensemble estimates with respect to both of these. A teaching phase would then help the classifier to learn the true category for each input row. During this phase, any classifier can learn or infer correct classifications from some other classifier's knowledge, thereby reducing the required number of presentations. As the information is added, cross-referencing between the two structures allows it to be used more widely, where a unique structure can build up that would not be possible by either method separately.1112202011122020183193https://www.wseas.org/multimedia/journals/computers/2020/a465105-052.pdf10.37394/23205.2020.19.23https://www.wseas.org/multimedia/journals/computers/2020/a465105-052.pdfGreer, K. (2020). New Ideas for Brain Modelling 7, available on arXiv at https://arxiv.org/abs/2011.02223.Greer, K. (2019). New Ideas for Brain Modelling 3, Cognitive Systems Research,Vol. 55, pp. 1-13, Elsevier. doi: https://doi.org/10.1016/j.cogsys.2018.12.016.Greer, K. (2018). New Ideas for Brain Modelling 4, BRAIN. Broad Research in Artificial Intelligence and Neuroscience, Vol. 9, No. 2, pp. 155-167. ISSN 2067-3957.Greer, K. (2018). An Improved Oscillating-Error Classifier with Branching, WSEAS Transactions on Computer Research, Vol. 6, pp. 49 -54. E-ISSN: 2415-1521.For the updated version, see Category Trees(2020), available on arXiv at https://arxiv.org/abs/1811.02617.10.23919/mipro.2018.8400040Dosilovic, F.K., Brcic, M. and Hlupic, N. (2018). Explainable Artificial Intelligence: A Survey, MIPRO 2018. 41st International Convention Proceedings, Opatija, Croatia, pp. 232 –237.DARPA. (2017). Explainable Artificial Intelligence, https://www.darpa.ml/program/explainable-artificial-intelligence. (last accessed 16/4/20).10.1109/2.33Carpenter, G.A. and Grossberg, S., 1988. The ART of adaptive pattern recognition by a self-organizing neural network. Computer, 21(3), pp.77-88.Carpenter, G.A. and Grossberg, S., 2010. Adaptive resonance theory.10.1007/978-3-642-15825-4_21Tscherepanow, M., 2010, September. TopoART: A topology learning hierarchical ART network. In International Conference on Artificial Neural Networks (pp. 157-167). Springer, Berlin, Heidelberg.Gómez, S.A. and Chesnevar, C.I. (2004). A Hybrid Approach to Pattern ClassificationUsingNeural Networks and Defeasible Argumentation, In Flairs conference, pp. 393-398.Goldt, S., Reeves, G., Mézard, M., Krzakala, F. and Zdeborová, L.(2020). The Gaussian equivalence of generative models for learning with two-layer neural networks. arXiv preprint arXiv:2006.14709.10.1145/3422622Goodfellow, I., Pouget-Abadie, J., Mirza, M., Xu, B., Warde-Farley, D., Ozair, S., Courville, A. and Bengio, Y. (2014). Generativeadversarial nets. In Advances in neural information processing systems, pp. 2672 -2680.Bozinovski, S.andFulgosi, A.(1976).The influence of pattern similarity and transfer learning upon training of a base perceptron B2.(original in Croatian) Proceedings of Symposium Informatica 3-121-5, Bled. 10.1145/1273496.1273592Raina, R., Battle, A., Lee, H., Packer, B. and Ng, A.Y.(2007). Self-taught learning: transfer learning from unlabeled data. In Proceedings of the 24th international conference on Machine learning, pp. 759-766.10.1145/1390156.1390182Dai, W., Yang, Q., Xue, G.R. and Yu, Y.(2008). Self-taught clustering. In Proceedings of the 25th international conference on Machine learning, pp. 200-207.10.21236/ada350490Nigam, K., McCallum, A., Thrun, S. and Mitchell, T. (2000). Text classification from labeled and unlabeled documents using EM. Machine Learning, Vol. 39,pp. 103 -134.Kohonen, T. (1990). The Self-Organising Map, Proceedings of the IEEE, Vol. 78, No. 9, pp. 1464 -1480.10.1080/09537280500180949Canetta, L., Cheikhrouhou, N. and Glardon, R. (2005). Applying two-stage SOM-based clustering approaches to industrial data analysis, Article in Production Planning and Control, DOI: 10.1080/09537280500180949.10.1109/icde.1998.655795Ester, M., Kriegel, H-P., Sander, J. and Xu, X. (1996). A Density-Based Algorithm for Discovering Clusters in Large Spatial Databases with Noise, Proceedings of 2nd International Conference on Knowledge Discovery and Data Mining (KDD-96).10.1007/s10115-007-0114-2Wu, X., Kumar, V., Quinlan, J.R., Ghosh, J., Yang, Q., Motoda, H., McLachlan, G.J., Ng, A., Liu, B., Yu, P.S., Zhou, Z-H., Steinbach, M., Hand, D.J. and Steinberg, D. (2008). Top 10 algorithms in data mining, Knowl. Inf. Syst., Vol. 14, pp. 1 –37. DOI 10.1007/s10115-007-0114-2.10.1037/0033-295x.84.5.413Anderson, J.A., Silverstein, J.W., Ritz, S.A. and Jones, R.A. (1977) Distinctive Features, Categorical Perception, and Probability Learning: Some Applications of a Neural Model, Psychological Review, Vol. 84, No. 5.Breiman, L. (2001). Random Forests. Machine Learning, Vol. 45, No. 1, pp. 5 -32.Adnan, M.N. and Islam, M.Z. (2015). Improving the random forest algorithm by randomly varying the size of the bootstrap samples for low dimensional data sets. In Proceedings of the European Symposium on Artificial Neural Networks, Computational Intelligence and Machine Learning, pp. 391-396.Breiman, L., Friedman, J., Stone, C.J. and Olshen, R.A. (1984). Classification and regression trees. CRC press.10.1109/tnn.2007.891664Frolov, A.A., Husek, D., Muraviev, I.P. and Polyakov, P.Y. (2007). Boolean Factor Analysis by Attractor Neural Network, IEEE Transactions on Neural Networks, Vol. 18, No. 3, pp 698 -707.10.1162/neco.2006.18.7.1527Hinton, G.E., Osindero, S. and Teh, Y.-W. (2006). A fast learning algorithm for deep belief nets,Neural computation, Vol. 18, No. 7, pp. 1527 -1554.UCI Machine Learning Repository (2019). http://archive.ics.uci.edu/ml/.Fisher, R.A. (1936). The use of multiple measurements in taxonomic problems, Annual Eugenics, 7, Part II, pp. 179-188, also in 'Contributions to Mathematical Statistics' (John Wiley, NY, 1950).Forina, M. et al. (1991). PARVUS -An Extendible Package for Data Exploration, Classification and Correlation. Institute of Pharmaceutical and Food Analysis and Technologies, Via Brigata Salerno, 16147 Genoa, Italy.Zoo database (2016). https://archive.ics.uci.edu/ml/datasets/Zoo. (last accessed 16/9/20)