<?xml version="1.0" encoding="UTF-8"?><xml><records><record><source-app name="Biblio" version="7.x">Drupal-Biblio</source-app><ref-type>47</ref-type><contributors><authors><author><style face="normal" font="default" size="100%">Lampridis, Orestis</style></author><author><style face="normal" font="default" size="100%">Riccardo Guidotti</style></author><author><style face="normal" font="default" size="100%">Salvatore Ruggieri</style></author></authors><secondary-authors><author><style face="normal" font="default" size="100%">Appice, Annalisa</style></author><author><style face="normal" font="default" size="100%">Tsoumakas, Grigorios</style></author><author><style face="normal" font="default" size="100%">Manolopoulos, Yannis</style></author><author><style face="normal" font="default" size="100%">Matwin, Stan</style></author></secondary-authors></contributors><titles><title><style face="normal" font="default" size="100%">Explaining Sentiment Classification with Synthetic Exemplars and Counter-Exemplars</style></title><secondary-title><style face="normal" font="default" size="100%">Discovery Science</style></secondary-title></titles><dates><year><style  face="normal" font="default" size="100%">2020</style></year><pub-dates><date><style  face="normal" font="default" size="100%">2020//</style></date></pub-dates></dates><urls><web-urls><url><style face="normal" font="default" size="100%">https://link.springer.com/chapter/10.1007/978-3-030-61527-7_24</style></url></web-urls></urls><publisher><style face="normal" font="default" size="100%">Springer International Publishing</style></publisher><pub-location><style face="normal" font="default" size="100%">Cham</style></pub-location><isbn><style face="normal" font="default" size="100%">978-3-030-61527-7</style></isbn><language><style face="normal" font="default" size="100%">eng</style></language><abstract><style face="normal" font="default" size="100%">We present xspells, a model-agnostic local approach for explaining the decisions of a black box model for sentiment classification of short texts. The explanations provided consist of a set of exemplar sentences and a set of counter-exemplar sentences. The former are examples classified by the black box with the same label as the text to explain. The latter are examples classified with a different label (a form of counter-factuals). Both are close in meaning to the text to explain, and both are meaningful sentences – albeit they are synthetically generated. xspells generates neighbors of the text to explain in a latent space using Variational Autoencoders for encoding text and decoding latent instances. A decision tree is learned from randomly generated neighbors, and used to drive the selection of the exemplars and counter-exemplars. We report experiments on two datasets showing that xspells outperforms the well-known lime method in terms of quality of explanations, fidelity, and usefulness, and that is comparable to it in terms of stability.</style></abstract></record></records></xml>