João Graça

João Graça

Lisboa, Lisboa, Portugal
10 mil seguidores + de 500 conexões

Atividades

Cadastre-se agora para visualizar todas as atividades

Licenças e certificados

Publicações

  • Posterior Sparsity in Dependency Grammar Induction

    Journal of Machine Learning Research

    A strong inductive bias is essential in unsupervised grammar induction. In this paper, we explore a particular sparsity bias in dependency grammars that encourages a small number of unique dependency types. We use part-of-speech (POS) tags to group dependencies by parent-child types and investigate sparsity-inducing penalties on the posterior distributions of parent-child POS tag pairs in the posterior regularization (PR) framework of Graça et al. (2007). In experiments with 12 different…

    A strong inductive bias is essential in unsupervised grammar induction. In this paper, we explore a particular sparsity bias in dependency grammars that encourages a small number of unique dependency types. We use part-of-speech (POS) tags to group dependencies by parent-child types and investigate sparsity-inducing penalties on the posterior distributions of parent-child POS tag pairs in the posterior regularization (PR) framework of Graça et al. (2007). In experiments with 12 different languages, we achieve significant gains in directed attachment accuracy over the standard expectation maximization (EM) baseline, with an average accuracy improvement of 6.5%, outperforming EM by at least 1% for 9 out of 12 languages. Furthermore, the new method outperforms models based on standard Bayesian sparsity-inducing parameter priors with an average improvement of 5% and positive gains of at least 1% for 9 out of 12 languages. On English text in particular, we show that our approach improves performance over other state-of-the-art techniques.

    Outros autores
    • Jennifer Gillenwater
    • Kuzman Ganchev
    • Fernando Pereira
    Ver publicação
  • Controlling Complexity in Part-of-Speech Induction

    Journal of Artificial Intelligence Research

    We consider the problem of fully unsupervised learning of grammatical (part-of-speech) categories from unlabeled text. The standard maximum-likelihood hidden Markov model for this task performs poorly, because of its weak inductive bias and large model capacity. We address this problem by refining the model and modifying the learning objective to control its capacity via parametric and non-parametric constraints. Our approach enforces word-category association sparsity, adds morphological and…

    We consider the problem of fully unsupervised learning of grammatical (part-of-speech) categories from unlabeled text. The standard maximum-likelihood hidden Markov model for this task performs poorly, because of its weak inductive bias and large model capacity. We address this problem by refining the model and modifying the learning objective to control its capacity via parametric and non-parametric constraints. Our approach enforces word-category association sparsity, adds morphological and orthographic features, and eliminates hard-to-estimate parameters for rare words. We develop an efficient learning algorithm that is not much more computationally intensive than standard training. We also provide an open-source implementation of the algorithm. Our experiments on five diverse languages (Bulgarian, Danish, English, Portuguese, Spanish) achieve significant improvements compared with previous methods for the same task.

    Outros autores
    • Kuzman Ganchev
    • Luisa Coheur
    • Fernando Pereira
    Ver publicação
  • Learning Tractable Word Alignment Models with Complex Constraints

    Journal of Computational Linguistics

    Word-level alignment of bilingual text is a critical resource for a growing variety of tasks. Probabilistic models forward alignment present a fundamental trade-off between richness of captured constraints and correlations versus efficiency and tractability of inference. In this article, we use the Posterior Regularization framework (Graça, Ganchev, and Taskar 2007) to incorporate complex constraints into probabilistic models during learning without changing the efficiency of the underlying…

    Word-level alignment of bilingual text is a critical resource for a growing variety of tasks. Probabilistic models forward alignment present a fundamental trade-off between richness of captured constraints and correlations versus efficiency and tractability of inference. In this article, we use the Posterior Regularization framework (Graça, Ganchev, and Taskar 2007) to incorporate complex constraints into probabilistic models during learning without changing the efficiency of the underlying model. We focus on the simple and tractable hidden Markov model, and present an efficient learning algorithm for incorporating approximate bijectivity and symmetry constraints. Models estimated with these constraints produce a significant boost in performance as measured by both precision and recall of manually annotated alignments for six language pairs. We also report experiments on two different tasks where word alignments are required: phrase based machine translation and syntax transfer, and show promising improvements over standard methods.

    Outros autores
    • Kuzman Ganchev
    Ver publicação
  • Posterior Regularization in Latent Variable Models

    Journal of Machine Learning Research

    We present posterior regularization, a probabilistic framework for structured, weakly supervised learning. Our framework efficiently incorporates indirect supervision via constraints on posterior distributions of probabilistic models with latent variables. Posterior regularization separates model complexity from the complexity of structural constraints it is desired to satisfy. By directly imposing decomposable regularization on the posterior moments of latent variables during learning, we…

    We present posterior regularization, a probabilistic framework for structured, weakly supervised learning. Our framework efficiently incorporates indirect supervision via constraints on posterior distributions of probabilistic models with latent variables. Posterior regularization separates model complexity from the complexity of structural constraints it is desired to satisfy. By directly imposing decomposable regularization on the posterior moments of latent variables during learning, we retain the computational efficiency of the unconstrained model while ensuring desired constraints
    hold in expectation. We present an efficient algorithm for learning with posterior regularization and illustrate its versatility on a diverse set of structural constraints such as bijectivity, symmetry and group sparsity in several large scale experiments, including multi-view learning, cross-lingual dependency grammar induction, unsupervised part-of-speech induction, and bitext word alignment.

    Outros autores
    • Kuzman Ganchev
    • Jennifer Gillenwater
    • Ben Taskar
    Ver publicação

Organizações

  • LXMLS - Lisbon Machine Learning Summer School

    Main Organizer

    One week Summer School about core Machine Learning and Natural Language Processing methods and their applications. www.lxmls.it.pt

Mais atividade de João

Veja o perfil completo de João

  • Saiba quem vocês conhecem em comum
  • Apresente-se
  • Entre em contato direto com João
Cadastre-se para ver o perfil completo

Outros perfis semelhantes

Outras pessoas chamadas João Graça em Portugal

Adicione novas competências com estes cursos