2012.bib

@inproceedings{RigBelLamZes12-URSW12-IW,
  author = {Fabrizio Riguzzi and Elena Bellodi and Evelina Lamma and Riccardo Zese},
  title = {Epistemic and Statistical Probabilistic Ontologies},
  booktitle = {Proceedings of the 8th International Workshop on Uncertain Reasoning for the Semantic Web (URSW2012), Boston, USA, 11 November 2012},
  year = {2012},
  editor = {Fernando Bobillo and
Rommel Carvalho and
da Costa, Paulo C. G. and
Nicola Fanizzi and
Laskey, Kathryn B.  and
Laskey, Kenneth J.  and
Thomas Lukasiewicz and
Trevor Martin and
Matthias Nickles and
Michael Pool},
  series = {CEUR Workshop Proceedings},
  publisher = {Sun {SITE} Central Europe},
  issn = {1613-0073},
  address = {Aachen, Germany},
  number = {900},
  pages = {3-14},
  pdf = {http://ceur-ws.org/Vol-900/paper1.pdf},
  abstract = {We present DISPONTE, a semantics for probabilistic ontologies that is based on the distribution semantics for probabilistic logic programs. In DISPONTE the axioms of a probabilistic ontology can be
annotated with an epistemic or a statistical probability. The epistemic probability represents a degree of confidence in the axiom, while the statistical probability considers the populations to which the axiom is applied.}
}
@inproceedings{BelRig12-AIIADC12-IW,
  title = {Parameter and Structure Learning Algorithms for Statistical Relational Learning},
  pages = {5-9},
  author = {Elena Bellodi and Fabrizio Riguzzi },
  editor = {Paolo Liberatore and
Michele Lombardi and
Floriano Scioscia},
  booktitle = {Doctoral Consortium of the 12th AI*IA Symposium on Artificial Intelligence, Proceedings of the Doctoral Consortium of the 12th Symposium of the Italian Association for Artificial Intelligence
Rome, Italy, June 15, 2012},
  copyright = {by the authors},
  series = {CEUR Workshop Proceedings},
  publisher = {Sun {SITE} Central Europe},
  issn = {1613-0073},
  address = {Aachen, Germany},
  volume = {926},
  year = {2012},
  pdf = {http://ceur-ws.org/Vol-926/paper1.pdf},
  scopus = {2-s2.0-84891770795
},
  abstract = {My research activity focuses on the field of Machine Learning. Two key challenges
in most machine learning applications are uncertainty and complexity.
The standard framework for handling uncertainty is probability, for complexity
is first-order logic. Thus we would like to be able to learn and perform inference
in representation languages that combine the two. This is the focus of the field
of Statistical Relational Learning.
},
  keywords = {Statistical relational learning, machine learning, first order logic}
}
@inproceedings{RigBelLamZese12-PAI12-NW,
  title = {Semantics and Inference for Probabilistic Ontologies},
  pages = { 41-46},
  author = {Fabrizio Riguzzi and Evelina Lamma and Elena Bellodi and Riccardo Zese},
  editor = {Matteo Baldoni and Federico Chesani and Bernardo Magnini and Paola Mello and Marco Montali},
  booktitle = { Popularize Artificial Intelligence. Proceedings of the AI*IA Workshop and Prize for Celebrating 100th Anniversary of Alan Turing's Birth ({PAI  2012}), Rome, Italy, June 15, 2012},
  copyright = {by the authors},
  series = {CEUR Workshop Proceedings},
  publisher = {Sun {SITE} Central Europe},
  issn = {1613-0073},
  address = {Aachen, Germany},
  volume = {860},
  year = {2012},
  pdf = {http://ceur-ws.org/Vol-860/paper3.pdf}
}
@inproceedings{RigBelLam12-DL12-IW,
  author = {Fabrizio Riguzzi and Elena Bellodi and Evelina Lamma},
  title = {Probabilistic {Datalog+/-} under the Distribution Semantics},
  booktitle = {Proceedings of the 25th International Workshop on Description Logics ({DL2012}),
Roma, Italy, 7-10 June 2012},
  editor = {Yevgeny Kazakov and Domenico Lembo and Frank Wolter},
  year = {2012},
  abstract = {We apply the distribution semantics for probabilistic ontologies (named DISPONTE) to the Datalog+/- language.
In DISPONTE the formulas of a probabilistic ontology can be annotated with an epistemic or a statistical probability.  The epistemic probability represents a degree of confidence in the formula, while the statistical probability considers the populations to which the formula is applied.
The probability of a query is defined in terms of finite set of finite explanations for the query, where an explanation is a set of possibly instantiated formulas that is sufficient for entailing the query.
The probability of a query is computed from the set of explanations by making them mutually exclusive.
We also compare the DISPONTE approach for Datalog+/- ontologies  with that of Probabilistic Datalog+/-,  where an ontology is composed of a Datalog+/- theory whose formulas are associated to an assignment of values for the random variables of a companion Markov Logic Network.
},
  copyright = {by the authors},
  series = {CEUR Workshop Proceedings},
  publisher = {Sun {SITE} Central Europe},
  issn = {1613-0073},
  address = {Aachen, Germany},
  url = {http://ml.unife.it/wp-content/uploads/Papers/RigBelLam12-DL12.pdf},
  pdf = {http://ceur-ws.org/Vol-846/paper_25.pdf},
  volume = {846},
  pages = {519-529}
}
@inproceedings{RigBelLam12-CILC12-NC,
  author = {Fabrizio Riguzzi and Elena Bellodi and Evelina Lamma},
  title = {Probabilistic Ontologies in {Datalog+/-}},
  booktitle = {Proceedings of the 27th Italian Conference on Computational Logic ({CILC2012}),
Roma, Italy, 6-7 June 2012},
  year = {2012},
  abstract = {In logic programming the distribution semantics is one of the most popular approaches for dealing with uncertain information. In this paper
we apply the distribution semantics to the  Datalog+/- language that is grounded in logic programming and allows tractable ontology querying. In the resulting semantics, called DISPONTE, formulas of a probabilistic ontology can be annotated with an epistemic or a statistical probability.  The epistemic probability represents a degree of confidence in the formula, while the statistical probability considers the populations to which the formula is applied.
The probability of a query is defined in terms of finite set of finite explanations for the query.
We also compare the DISPONTE approach for Datalog+/- ontologies  with that of Probabilistic Datalog+/-  where an ontology is composed of a Datalog+/- theory whose formulas are associated to an assignment of values for the random variables of a companion Markov Logic Network.
},
  copyright = {by the authors},
  series = {CEUR Workshop Proceedings},
  publisher = {Sun {SITE} Central Europe},
  issn = {1613-0073},
  volume = {857},
  address = {Aachen, Germany},
  url = {http://ml.unife.it/wp-content/uploads/Papers/RigBelLam12-CILC12.pdf},
  pdf = {http://ceur-ws.org/Vol-857/paper_f16.pdf},
  pages = {221-235}
}
@inproceedings{BelRig12-ILP11-IC,
  author = {Elena Bellodi and Fabrizio Riguzzi},
  title = {Learning the Structure of Probabilistic Logic Programs},
  booktitle = {Inductive Logic Programming
21st International Conference, ILP 2011, London, UK, July 31 - August 3, 2011. Revised Papers },
  year = {2012},
  editor = {Muggleton, Stephen H. and Tamaddoni-Nezhad, Alireza and Lisi, Francesca A.},
  doi = {10.1007/978-3-642-31951-8_10},
  series = {LNCS},
  volume = {7207},
  publisher = {Springer},
  address = {Heidelberg, Germany},
  pages = {61-75},
  note = {The original publication is available at \url{http://www.springerlink.com}},
  pdf = {http://ml.unife.it/wp-content/uploads/Papers/BelRig12-ILP11-IC.pdf},
  keywords = {Probabilistic Inductive Logic Programming, Logic Programs with Annotated Disjunctions, ProbLog},
  copyright = {Springer},
  abstract = {There is a growing interest in the field of
Probabilistic Inductive Logic Programming, which uses languages that
integrate logic programming and probability.
Many of these languages are based on the distribution semantics and recently various authors have proposed systems for learning the parameters (PRISM, LeProbLog, LFI-ProbLog
and EMBLEM) or both the structure and the parameters (SEM-CP-logic) of these languages.
EMBLEM for example uses an Expectation Maximization approach in which  the expectations are computed on Binary Decision Diagrams.
In this paper we present the algorithm SLIPCASE for ``Structure LearnIng of ProbabilistiC logic progrAmS with Em over bdds''. It performs a beam search in the space of the language of Logic Programs with Annotated Disjunctions (LPAD) using the log likelihood of the data as the guiding heuristics. To estimate the log likelihood of theory refinements it performs a limited number of Expectation Maximization iterations of EMBLEM.
SLIPCASE has been tested on three real world datasets and compared with SEM-CP-logic and  Learning using Structural Motifs, an algorithm for Markov Logic Networks. The results show that SLIPCASE achieves higher areas under the precision-recall and ROC curves and is more scalable.
}
}
@article{BelRig12-IA-IJ,
  author = {Elena Bellodi and Fabrizio Riguzzi},
  title = { Experimentation of an Expectation Maximization Algorithm for Probabilistic Logic Programs},
  year = {2012},
  journal = {Intelligenza Artificiale},
  publisher = {IOS Press},
  copyright = {IOS Press},
  pdf = {http://ml.unife.it/wp-content/uploads/Papers/BelRig12-IA-IJ.pdf},
  abstract = {Statistical Relational Learning and Probabilistic Inductive Logic Programming are two emerging fields that use representation languages able to combine logic and probability. In the field of Logic Programming, the distribution semantics is one of the prominent approaches for representing uncertainty and underlies many languages such as ICL, PRISM, ProbLog and LPADs.
Learning the parameters for such languages requires an Expectation Maximization algorithm since their equivalent Bayesian networks contain hidden variables.
EMBLEM (EM over BDDs for probabilistic Logic programs Efficient Mining) is an EM algorithm for languages following the distribution semantics that computes expectations directly on the Binary Decision Diagrams that are built for inference.
In this paper we present experiments comparing EMBLEM with LeProbLog, Alchemy, CEM, RIB and LFI-ProbLog on six real world datasets. The results show that EMBLEM is able to solve problems on which the other systems fail and it often achieves significantly higher areas under the Precision Recall and the ROC curves in a similar time.},
  keywords = {Statistical Relational Learning, Probabilistic Inductive Logic Programming, Probabilistic Logic Programming,  Expectation Maximization, Binary Decision Diagrams,
Logic Programs with Annotated Disjunctions
},
  volume = {8},
  number = {1},
  pages = {3-18},
  doi = {10.3233/IA-2012-0027}
}

This file was generated by bibtex2html 1.98.