@inproceedings{RigZelILP2012-CEUR-EB, title = {Preface}, booktitle = {Late Breaking Papers of the 22nd International Conference on Inductive Logic Programming (ILP)}, year = 2012, author = {Fabrizio Riguzzi and Filip \v{Z}elezn\'{y} }, editor = {Fabrizio Riguzzi and Filip \v{Z}elezn\'{y} }, volume = {975}, series = {CEUR Workshop Proceedings}, publisher = {Sun {SITE} Central Europe}, address = {Aachen, Germany}, issn = {1613-0073}, url = {http://ceur-ws.org/Vol-975/preface.pdf}, venue = {Dubrovnik, Croatia}, eventdate = {September 17-19, 2012}, copyright = {by the authors}, scopus = {2-s2.0-84922654847}, pages = {3--4} }

@inproceedings{RigBelLamZes12-URSW12-IW, author = {Fabrizio Riguzzi and Elena Bellodi and Evelina Lamma and Riccardo Zese}, title = {Epistemic and Statistical Probabilistic Ontologies}, booktitle = {Proceedings of the 8th International Workshop on Uncertain Reasoning for the Semantic Web (URSW2012), Boston, USA, 11 November 2012}, year = {2012}, editor = {Fernando Bobillo and Rommel Carvalho and da Costa, Paulo C. G. and Nicola Fanizzi and Laskey, Kathryn B. and Laskey, Kenneth J. and Thomas Lukasiewicz and Trevor Martin and Matthias Nickles and Michael Pool}, series = {CEUR Workshop Proceedings}, publisher = {Sun {SITE} Central Europe}, issn = {1613-0073}, address = {Aachen, Germany}, number = {900}, pages = {3-14}, pdf = {http://ceur-ws.org/Vol-900/paper1.pdf}, abstract = {We present DISPONTE, a semantics for probabilistic ontologies that is based on the distribution semantics for probabilistic logic programs. In DISPONTE the axioms of a probabilistic ontology can be annotated with an epistemic or a statistical probability. The epistemic probability represents a degree of confidence in the axiom, while the statistical probability considers the populations to which the axiom is applied.} }

@inproceedings{BelRig12-AIIADC12-IW, title = {Parameter and Structure Learning Algorithms for Statistical Relational Learning}, pages = {5-9}, author = {Elena Bellodi and Fabrizio Riguzzi }, editor = {Paolo Liberatore and Michele Lombardi and Floriano Scioscia}, booktitle = {Doctoral Consortium of the 12th AI*IA Symposium on Artificial Intelligence, Proceedings of the Doctoral Consortium of the 12th Symposium of the Italian Association for Artificial Intelligence Rome, Italy, June 15, 2012}, copyright = {by the authors}, series = {CEUR Workshop Proceedings}, publisher = {Sun {SITE} Central Europe}, issn = {1613-0073}, address = {Aachen, Germany}, volume = {926}, year = {2012}, pdf = {http://ceur-ws.org/Vol-926/paper1.pdf}, scopus = {2-s2.0-84891770795 }, abstract = {My research activity focuses on the field of Machine Learning. Two key challenges in most machine learning applications are uncertainty and complexity. The standard framework for handling uncertainty is probability, for complexity is first-order logic. Thus we would like to be able to learn and perform inference in representation languages that combine the two. This is the focus of the field of Statistical Relational Learning. }, keywords = {Statistical relational learning, machine learning, first order logic} }

@inproceedings{RigBelLamZese12-PAI12-NW, title = {Semantics and Inference for Probabilistic Ontologies}, pages = { 41-46}, author = {Fabrizio Riguzzi and Evelina Lamma and Elena Bellodi and Riccardo Zese}, editor = {Matteo Baldoni and Federico Chesani and Bernardo Magnini and Paola Mello and Marco Montali}, booktitle = { Popularize Artificial Intelligence. Proceedings of the AI*IA Workshop and Prize for Celebrating 100th Anniversary of Alan Turing's Birth ({PAI 2012}), Rome, Italy, June 15, 2012}, copyright = {by the authors}, series = {CEUR Workshop Proceedings}, publisher = {Sun {SITE} Central Europe}, issn = {1613-0073}, address = {Aachen, Germany}, volume = {860}, year = {2012}, pdf = {http://ceur-ws.org/Vol-860/paper3.pdf} }

@inproceedings{BufLamRigFor12-PAI12-NW, title = {Un sistema di Vision Inspection basato su reti neurali}, author = {Ludovico Buffon and Evelina Lamma and Fabrizio Riguzzi and Davide Formenti}, pages = { 1-6}, editor = {Matteo Baldoni and Federico Chesani and Bernardo Magnini and Paola Mello and Marco Montai}, booktitle = { Popularize Artificial Intelligence. Proceedings of the AI*IA Workshop and Prize for Celebrating 100th Anniversary of Alan Turing's Birth ({PAI 2012}), Rome, Italy, June 15, 2012}, copyright = {by the authors}, series = {CEUR Workshop Proceedings}, publisher = {Sun {SITE} Central Europe}, issn = {1613-0073}, address = {Aachen, Germany}, volume = {860}, year = {2012}, pdf = {http://ceur-ws.org/Vol-860/paper9.pdf} }

@inproceedings{RigBelLam12-DL12-IW, author = {Fabrizio Riguzzi and Elena Bellodi and Evelina Lamma}, title = {Probabilistic {Datalog+/-} under the Distribution Semantics}, booktitle = {Proceedings of the 25th International Workshop on Description Logics ({DL2012}), Roma, Italy, 7-10 June 2012}, editor = {Yevgeny Kazakov and Domenico Lembo and Frank Wolter}, year = {2012}, abstract = {We apply the distribution semantics for probabilistic ontologies (named DISPONTE) to the Datalog+/- language. In DISPONTE the formulas of a probabilistic ontology can be annotated with an epistemic or a statistical probability. The epistemic probability represents a degree of confidence in the formula, while the statistical probability considers the populations to which the formula is applied. The probability of a query is defined in terms of finite set of finite explanations for the query, where an explanation is a set of possibly instantiated formulas that is sufficient for entailing the query. The probability of a query is computed from the set of explanations by making them mutually exclusive. We also compare the DISPONTE approach for Datalog+/- ontologies with that of Probabilistic Datalog+/-, where an ontology is composed of a Datalog+/- theory whose formulas are associated to an assignment of values for the random variables of a companion Markov Logic Network. }, copyright = {by the authors}, series = {CEUR Workshop Proceedings}, publisher = {Sun {SITE} Central Europe}, issn = {1613-0073}, address = {Aachen, Germany}, url = {http://ml.unife.it/wp-content/uploads/Papers/RigBelLam12-DL12.pdf}, pdf = {http://ceur-ws.org/Vol-846/paper_25.pdf}, volume = {846}, pages = {519-529} }

@inproceedings{RigBelLam12-CILC12-NC, author = {Fabrizio Riguzzi and Elena Bellodi and Evelina Lamma}, title = {Probabilistic Ontologies in {Datalog+/-}}, booktitle = {Proceedings of the 27th Italian Conference on Computational Logic ({CILC2012}), Roma, Italy, 6-7 June 2012}, year = {2012}, abstract = {In logic programming the distribution semantics is one of the most popular approaches for dealing with uncertain information. In this paper we apply the distribution semantics to the Datalog+/- language that is grounded in logic programming and allows tractable ontology querying. In the resulting semantics, called DISPONTE, formulas of a probabilistic ontology can be annotated with an epistemic or a statistical probability. The epistemic probability represents a degree of confidence in the formula, while the statistical probability considers the populations to which the formula is applied. The probability of a query is defined in terms of finite set of finite explanations for the query. We also compare the DISPONTE approach for Datalog+/- ontologies with that of Probabilistic Datalog+/- where an ontology is composed of a Datalog+/- theory whose formulas are associated to an assignment of values for the random variables of a companion Markov Logic Network. }, copyright = {by the authors}, series = {CEUR Workshop Proceedings}, publisher = {Sun {SITE} Central Europe}, issn = {1613-0073}, volume = {857}, address = {Aachen, Germany}, url = {http://ml.unife.it/wp-content/uploads/Papers/RigBelLam12-CILC12.pdf}, pdf = {http://ceur-ws.org/Vol-857/paper_f16.pdf}, pages = {221-235} }

@inproceedings{Rig12-CILC12-NC, author = {Fabrizio Riguzzi}, title = {Optimizing Inference for Probabilistic Logic Programs Exploiting Independence and Exclusiveness}, booktitle = {Proceedings of the 27th Italian Conference on Computational Logic ({CILC2012}), Roma, Italy, 6-7 June 2012}, year = {2012}, abstract = {Probabilistic Logic Programming (PLP) is gaining popularity due to its many applications in particular in Machine Learning. An important problem in PLP is how to compute the probability of queries. PITA is an algorithm for solving such a problem that exploits tabling, answer subsumption and Binary Decision Diagrams (BDDs). PITA does not impose any restriction on the programs. Other algorithms, such as PRISM, achieve a higher speed by imposing two restrictions on the program, namely that subgoals are independent and that clause bodies are mutually exclusive. Another assumption that simplifies inference is that clause bodies are independent. In this paper we present the algorithms PITA(IND,IND) and PITA(OPT). PITA(IND,IND) assumes that subgoals and clause bodies are independent. PITA(OPT) instead first checks whether these assumptions hold for subprograms and subgoals: if they hold, PITA(OPT) uses a simplified calculation, otherwise it resorts to BDDs. Experiments on a number of benchmark datasets show that PITA(IND,IND) is the fastest on datasets respecting the assumptions while PITA(OPT) is a good option when nothing is known about a dataset. }, copyright = {by the authors}, series = {CEUR Workshop Proceedings}, publisher = {Sun {SITE} Central Europe}, issn = {1613-0073}, address = {Aachen, Germany}, url = {http://ml.unife.it/wp-content/uploads/Papers/Rig12-CILC12.pdf}, pdf = {http://ceur-ws.org/Vol-857/paper_f15.pdf}, volume = {857}, pages = {206-220} }

@inproceedings{GavRiguMilCag12-ISAIM12-IC, author = {Marco Gavanelli and Fabrizio Riguzzi and Michela Milano and Paolo Cagnoli}, title = {Constraint and Optimization techniques for supporting Policy Making}, booktitle = {International Symposium on Artificial Intelligence and Mathematics (ISAIM 2012), Fort Lauderdale, Florida, USA, January 9-11, 2012}, year = {2012}, pdf = {http://www.cs.uic.edu/pub/Isaim2012/WebPreferences/ISAIM2012_Gavanelli_etal.pdf} }

@inproceedings{BelRig12-ILP11-IC, author = {Elena Bellodi and Fabrizio Riguzzi}, title = {Learning the Structure of Probabilistic Logic Programs}, booktitle = {Inductive Logic Programming 21st International Conference, ILP 2011, London, UK, July 31 - August 3, 2011. Revised Papers }, year = {2012}, editor = {Muggleton, Stephen H. and Tamaddoni-Nezhad, Alireza and Lisi, Francesca A.}, doi = {10.1007/978-3-642-31951-8_10}, series = {LNCS}, volume = {7207}, publisher = {Springer}, address = {Heidelberg, Germany}, pages = {61-75}, note = {The original publication is available at \url{http://www.springerlink.com}}, pdf = {http://ml.unife.it/wp-content/uploads/Papers/BelRig12-ILP11-IC.pdf}, keywords = {Probabilistic Inductive Logic Programming, Logic Programs with Annotated Disjunctions, ProbLog}, copyright = {Springer}, abstract = {There is a growing interest in the field of Probabilistic Inductive Logic Programming, which uses languages that integrate logic programming and probability. Many of these languages are based on the distribution semantics and recently various authors have proposed systems for learning the parameters (PRISM, LeProbLog, LFI-ProbLog and EMBLEM) or both the structure and the parameters (SEM-CP-logic) of these languages. EMBLEM for example uses an Expectation Maximization approach in which the expectations are computed on Binary Decision Diagrams. In this paper we present the algorithm SLIPCASE for ``Structure LearnIng of ProbabilistiC logic progrAmS with Em over bdds''. It performs a beam search in the space of the language of Logic Programs with Annotated Disjunctions (LPAD) using the log likelihood of the data as the guiding heuristics. To estimate the log likelihood of theory refinements it performs a limited number of Expectation Maximization iterations of EMBLEM. SLIPCASE has been tested on three real world datasets and compared with SEM-CP-logic and Learning using Structural Motifs, an algorithm for Markov Logic Networks. The results show that SLIPCASE achieves higher areas under the precision-recall and ROC curves and is more scalable. } }

@article{BelRig12-IA-IJ, author = {Elena Bellodi and Fabrizio Riguzzi}, title = { Experimentation of an Expectation Maximization Algorithm for Probabilistic Logic Programs}, year = {2012}, journal = {Intelligenza Artificiale}, publisher = {IOS Press}, copyright = {IOS Press}, pdf = {http://ml.unife.it/wp-content/uploads/Papers/BelRig12-IA-IJ.pdf}, abstract = {Statistical Relational Learning and Probabilistic Inductive Logic Programming are two emerging fields that use representation languages able to combine logic and probability. In the field of Logic Programming, the distribution semantics is one of the prominent approaches for representing uncertainty and underlies many languages such as ICL, PRISM, ProbLog and LPADs. Learning the parameters for such languages requires an Expectation Maximization algorithm since their equivalent Bayesian networks contain hidden variables. EMBLEM (EM over BDDs for probabilistic Logic programs Efficient Mining) is an EM algorithm for languages following the distribution semantics that computes expectations directly on the Binary Decision Diagrams that are built for inference. In this paper we present experiments comparing EMBLEM with LeProbLog, Alchemy, CEM, RIB and LFI-ProbLog on six real world datasets. The results show that EMBLEM is able to solve problems on which the other systems fail and it often achieves significantly higher areas under the Precision Recall and the ROC curves in a similar time.}, keywords = {Statistical Relational Learning, Probabilistic Inductive Logic Programming, Probabilistic Logic Programming, Expectation Maximization, Binary Decision Diagrams, Logic Programs with Annotated Disjunctions }, volume = {8}, number = {1}, pages = {3-18}, doi = {10.3233/IA-2012-0027} }

@article{RigDiM12-ML-IJ, author = {Fabrizio Riguzzi and Di Mauro, Nicola}, title = {Applying the Information Bottleneck to Statistical Relational Learning}, year = {2012}, journal = {Machine Learning}, volume = {86}, number = {1}, pages = {89--114}, note = {The original publication is available at \url{http://www.springerlink.com}}, pdf = {http://ml.unife.it/wp-content/uploads/Papers/RigDiM11-ML-IJ.pdf}, doi = {10.1007/s10994-011-5247-6}, publisher = {Springer}, copyright = {Springer}, address = {Heidelberg, Germany}, abstract = {In this paper we propose to apply the Information Bottleneck (IB) approach to the sub-class of Statistical Relational Learning (SRL) languages that are reducible to Bayesian networks. When the resulting networks involve hidden variables, learning these languages requires the use of techniques for learning from incomplete data such as the Expectation Maximization (EM) algorithm. Recently, the IB approach was shown to be able to avoid some of the local maxima in which EM can get trapped when learning with hidden variables. Here we present the algorithm Relational Information Bottleneck (RIB) that learns the parameters of SRL languages reducible to Bayesian Networks. In particular, we present the specialization of RIB to a language belonging to the family of languages based on the distribution semantics, Logic Programs with Annotated Disjunction (LPADs). This language is prototypical for such a family and its equivalent Bayesian networks contain hidden variables. RIB is evaluated on the IMDB, Cora and artificial datasets and compared with LeProbLog, EM, Alchemy and PRISM. The experimental results show that RIB has good performances especially when some logical atoms are unobserved. Moreover, it is particularly suitable when learning from interpretations that share the same Herbrand base.}, keywords = {Statistical Relational Learning, Probabilistic Inductive Logic Programming, Probabilistic Logic Programming, Information Bottleneck, Logic Programs with Annotated Disjunctions } }

*This file was generated by
bibtex2html 1.98.*