news
| Feb 12, 2026 |
We are going to LREC |
|---|---|
| Feb 06, 2026 |
We are going to PROPOR |
| Feb 04, 2026 |
We are happy to share that our paper “Select First, Transfer Later: Choosing a Proper Dataset for SRL and GNN Based Transfer Learning” has been published in Machine Learning Journal.
Traditional machine learning models often ignore the relational structure present in many real-world domains. Approaches such as Statistical Relational Learning (SRL) and Graph Neural Networks (GNNs) address this by explicitly modeling dependencies between entities. However, like traditional models, they typically assume that training and testing data come from the same distribution — an assumption that often fails in practice.
Transfer Learning helps by reusing knowledge from one domain in another. But an important question remains largely overlooked:
👉 From where should we transfer?
In this work, we propose a principled method to estimate the suitability of transfer between relational domains using Kullback–Leibler (KL) divergence, computed from a Naive Bayes distribution of the target relational data and each candidate source model (SRL or GNN).
🔎 Main contribution:
|
| Jan 10, 2026 | I will be one of the 19th Symposium in Information and Human Language Technology (STIL) PC chairs. The deadline is on March 30th, on JEMS. Check it out the conference website for more information. STILL will be colocated with BRACIS, KDMIlE and WESACC. |
| Jan 03, 2026 | We are thrilled to share that Vítor Lourenço is heading to Rabat, Morocco 🇲🇦 in March, as our paper has been accepted to EACL 2026 ✨ 👉 “KG-CRAFT: Knowledge Graph-based Contrastive Reasoning with LLMs for Enhancing Automated Fact-checking This paper will be presented at the EACL main conference, which makes this achievement even more special. 📄 In KG-CRAFT, we propose a novel approach for automated fact-checking that combines knowledge graphs with contrastive reasoning in large language models. By constructing a claim-centric knowledge graph and generating contrastive questions, the method guides LLMs toward better evidence selection and reasoning, leading to more reliable and accurate claim verification. The approach achieves state-of-the-art results on real-world fact-checking benchmarks. |
| Oct 08, 2025 |
🦜 Exploring Brazil’s LLM Fauna: Evaluating Generative Performance in Portuguese
We are excited to share our new paper, “Exploring Brazil’s LLM Fauna: Investigating the Generative Performance of Large Language Models in Portuguese.”
As Large Language Models (LLMs) become increasingly embedded in real-world applications, their evaluation still relies heavily on narrow, mostly English-centered benchmarks. These traditional evaluations often neglect essential generative aspects such as discourse coherence, adequacy, and linguistic transformations — all crucial for practical use.
In this work, we provide a comprehensive evaluation of Brazilian Portuguese LLMs across three core Natural Language Generation tasks:
|
| Apr 03, 2025 |
We are going to ACL 2025 |
| Dec 24, 2024 |
🚀 New Publication: BERTweet.BR — A Pre-trained Language Model for Portuguese Tweets
We are excited to share our paper “BERTweet.BR: a pre-trained language model for tweets in Portuguese”, now published in Neural Computing and Applications.
Check it out the 📄 paper and the 🤗 Model on Hugging Face . While most advances in neural language models focus on English, Portuguese — despite being the sixth most spoken language in the world — still lacks domain-specific large-scale resources. This gap is even more evident for social media, where Brazilian users are among the most active globally. To address this, we introduce BERTweet.BR, the first large-scale pre-trained language model specifically designed for Brazilian Portuguese tweets. The model:
|
| May 15, 2024 | I talked to the IBERAMIA pre-event series about some simple, feasible, and practical steps to enhance the Brazilian AI ecosystem. It is now available on IBERAMIA channel . |