The LatinX in AI research workshop is a one-day event with invited speakers, oral presentations, and research posters. The event brings together faculty, graduate students, research scientists, and engineers for an opportunity to connect and exchange ideas. There will be a panel discussion and a mentoring session to discuss current research trends and career choices in artificial intelligence and machine learning, highlighting the unique challenges of LatinX identifying researchers. The workshop aims to create a platform for the work of Latinx researchers and we invite everyone to attend.We strongly encourage students, postdocs, and researchers who primarily identify as Latinx in all areas of machine learning to submit an abstract describing new, previously, or concurrently published research. We welcome abstract submissions, in theory, methodology, as well as applications. Abstracts may describe completed research or work-in-progress. While the presenting author need not be the first author of the work, we encourage authors to highlight the contribution of Latinx individuals — particularly the presenting author — in the abstract. The LatinX authors of accepted abstracts will be asked to present their work in a poster session. A few authors will be selected to give 15-minute oral presentations. Authors accepted to present will be offered presentation coaching. Submissions will be peer-reviewed. The authors are encouraged to sign up to review as part of the program committee for LXAI as well.
Mon 5:00 a.m. - 5:20 a.m.
|
Check-in
Checking in (Give t-shirts and swag) |
🔗 |
Mon 5:20 a.m. - 5:30 a.m.
|
Opening Remarks
(
Opening
)
SlidesLive Video » |
🔗 |
Mon 5:30 a.m. - 6:00 a.m.
|
Neuroscience & AI: Towards Bio-Inspired Artificial Agents
(
Keynote
)
SlidesLive Video » |
Maria-Jose Escobar 🔗 |
Mon 6:00 a.m. - 6:10 a.m.
|
Q&A for Maria-Jose Escobar
(
Q&A
)
Q&A |
🔗 |
Mon 6:10 a.m. - 6:20 a.m.
|
An Introduction to Quantum Natural Language Processing and a Study Case
(
Talk
)
SlidesLive Video » This document provides essential concepts, materials, and examples on Quantum Natural Language Processing. It is a recent topic, and the literature is limited. Therefore, we discuss the main ideas on QNLP and its potential applications. This paper brings quantum computing concepts applied to NLP with different purposes; in particular, we focus on showing basic concepts and implementation. We also discuss some potential QNLP applications. |
Shubhangi Rastogi 🔗 |
Mon 6:20 a.m. - 6:30 a.m.
|
Bottleneck-based Encoder-decoder ARchitecture (BEAR) for Learning Unbiased Consumer-to-Consumer Image Representations
(
Talk
)
SlidesLive Video » Unbiased representation learning is still an object of study under specific applications and contexts. Novel architectures are usually crafted to resolve particular problems using mixtures of fundamental pieces. This paper presents different image feature extraction mechanisms that work together with residual connections to encode perceptual image information in an autoencoder configuration. We use data that aims to address further issues regarding criminal activity in consumer-to-consumer online platforms. Preliminary results suggest that the proposed architecture can learn rich spaces using ours and other image datasets. |
Pablo Rivas 🔗 |
Mon 6:30 a.m. - 6:40 a.m.
|
Classifier Guided Diffusion for Image Inpainting: Applications to Fine Art
(
Talk
)
SlidesLive Video » Within the field of Cultural Heritage, image inpainting is a conservation process that fills in missing or damaged parts of an artwork to present a complete image. Multi-modal diffusion models have brought photo-realistic results on image inpainting where content can be generated by using descriptive text prompts. However, these models fail to produce content consistent with the painter's artistic style and period, being unsuitable for the reconstruction of fine arts and requiring laborious expert judgement. This work presents a methodology to improve the inpainting process by automating the selection process. We propose a discriminator model that processes the output of inpainting models and assigns a probability that indicates the likelihood that the restored image belongs to a certain painter. |
Lucia Cipolina Kun 🔗 |
Mon 6:40 a.m. - 6:50 a.m.
|
A Model-Based Filter to Improve Local Differential Privacy
(
Talk
)
SlidesLive Video » Local differential privacy has been gaining popularity in both academic and industrial settings as an effective mechanism to enable the computation of summary statistics by service providers while providing privacy guarantees to end users. In the proposed mechanisms of local differential privacy, the introduction of noise in the user data is key to preserve privacy, but can greatly limit the estimation power on the provider side. In this paper we propose to include pre-filters based on machine learning models to help discard observations that may be too noisy to add value to the estimation process. We test our approach on both synthetic and real datasets, and identify average reductions of up to 31% in the mean squared error. |
Juan Miguel Gutierrez Vidal 🔗 |
Mon 6:50 a.m. - 6:55 a.m.
|
Q&A
SlidesLive Video » Q&A for the first block of oral presentations. |
🔗 |
Mon 6:55 a.m. - 7:15 a.m.
|
Break
AM Break |
🔗 |
Mon 7:15 a.m. - 7:45 a.m.
|
On the Decision Support Systems Based on AI for Highly Complex Applications: Health & Defense
(
Keynote
)
SlidesLive Video » |
O. L. Quintero Montoya 🔗 |
Mon 7:45 a.m. - 7:55 a.m.
|
Q&A for Olga Lucia
(
Q&A
)
Q&A |
🔗 |
Mon 7:55 a.m. - 8:05 a.m.
|
A Study on the Predictability of Sample Learning Consistency
(
Talk
)
SlidesLive Video » Curriculum Learning is a powerful training method that allows for faster and better training in some settings. This method, however, requires having a notion of which examples are difficult and which are easy, which is not always trivial to provide. A recent metric called C-Score acts as a proxy for example difficulty by relating it to learning consistency. Unfortunately, this method is quite compute intensive which limits its applicability for alternative datasets. In this work, we train models through different methods to predict C-Score for CIFAR-100 and CIFAR-10. We find, however, that these models generalize poorly both within the same distribution as well as out of distribution. This suggests that C-Score is not defined by the individual characteristics of each sample but rather by other factors. We hypothesize that a sample’s relation to its neighbors, in particular, how many of them share the same labels, can help in explaining C-Scores. We plan to explore this in future work. |
Alain Raymond 🔗 |
Mon 8:05 a.m. - 8:15 a.m.
|
Link Prediction from Heterogeneous Opinion Mining Networks with Multi-Domain Applications
(
Talk
)
SlidesLive Video » How to mine opinions from pre-trained and general-purpose models in different domain applications? Answering this question is a recent research challenge since most existing methods fail when applied in domains other than those trained. In this paper, we introduce heterogeneous opinion mining networks as a strategy to combine opinion data from different domains, as well as integrate various models trained for different applications. We discuss how such knowledge results in different nodes and links in heterogeneous networks. We also investigated the mapping of the network nodes to a unified embedding space to allow vector-space-based machine learning models. We evaluate our proposal in a link prediction task to show that nodes and relationships in one domain can predict relevant knowledge in other domains. |
Bernardo Marques Costa · Ricardo Marcondes Marcacini 🔗 |
Mon 8:15 a.m. - 8:25 a.m.
|
Transfer Learning with Joint Fine-Tuning for Multimodal Sentiment Analysis
(
Talk
)
SlidesLive Video » Most existing methods focus on sentiment analysis of textual data. However, recently there has been a massive use of images and videos on social platforms, motivating sentiment analysis from other modalities. Current studies show that exploring other modalities (e.g., images) increases sentiment analysis performance. State-of-the-art multimodal models, such as CLIP and VisualBERT, are pre-trained on datasets with the text paired with images. Although the results obtained by these models are promising, pre-training and sentiment analysis fine-tuning tasks of these models are computationally expensive. This paper introduces a transfer learning approach using joint fine-tuning for sentiment analysis. Our proposal achieved competitive results using a more straightforward alternative fine-tuning strategy that leverages different pre-trained unimodal models and efficiently combines them in a multimodal space. Moreover, our proposal allows flexibility when incorporating any pre-trained model for texts and images during the joint fine-tuning stage, being especially interesting for sentiment classification in low-resource scenarios. |
Guilherme Lourenço de Toledo · Ricardo Marcondes Marcacini 🔗 |
Mon 8:25 a.m. - 8:30 a.m.
|
Q&A
SlidesLive Video » Q&A for second block of oral presentations. |
🔗 |
Mon 8:30 a.m. - 9:30 a.m.
|
Sponsor Booth Visit
(
Panel
)
Visit the booths from our sponsors |
🔗 |
Mon 9:30 a.m. - 10:30 a.m.
|
Lunch
|
🔗 |
Mon 10:30 a.m. - 11:00 a.m.
|
Speech Recognition & Synthesis for Language in Low Data Regimes: Learning from Few Speakers using Multilingual Models
(
Keynote
)
SlidesLive Video » Speech recognition offers promising benefits for business and personal applications. Although automatic speech recognition systems have evolved significantly with deep learning methods, it remains an open research problem. In many languages there is still a shortage of open/public resources, resulting in low quality automatic speech recognition systems. In this talk, a multi-speaker text-to-speech (TTS) system is described for scenarios with few available speakers. Exploring flow-based and multilingual models, it is possible to leverage data from languages with many available speakers and make it viable for those languages with less data. Additionally, we show how this model can be applied to improve automatic speech recognition (ASR) systems in two target languages, simulating a scenario with only one speaker available. This allows for many applications such as the deployment of TTS for people with voice disorders, or ASR in extremely small data available such as native and regional languages. |
Moacir Ponti 🔗 |
Mon 11:00 a.m. - 11:10 a.m.
|
Q&A for Moacir
(
Q&A
)
Q&A for Moacir |
🔗 |
Mon 11:10 a.m. - 11:50 a.m.
|
Graduate School Fellowship Professional Tips
(
Spotlight Talk
)
SlidesLive Video » |
CJ Barberan 🔗 |
Mon 11:50 a.m. - 11:55 a.m.
|
Q&A for CJ
(
Q&A
)
Q&A |
🔗 |
Mon 11:55 a.m. - 12:15 p.m.
|
Afternoon Break
Afternoon Break |
🔗 |
Mon 12:15 p.m. - 12:45 p.m.
|
Three Frontiers of Responsible Machine Learning
(
Keynote
)
SlidesLive Video » |
Maria De-Arteaga 🔗 |
Mon 12:45 p.m. - 12:50 p.m.
|
Q&A for Maria
(
Q&A
)
Q&A |
🔗 |
Mon 12:50 p.m. - 1:25 p.m.
|
Patent Process
(
Spotlight Talk
)
SlidesLive Video » |
Ramesh Doddaiah 🔗 |
Mon 1:25 p.m. - 1:30 p.m.
|
Q&A for Ramesh
(
Q&A
)
Q&A |
🔗 |
Mon 1:30 p.m. - 2:00 p.m.
|
Poster Session
(
Posters
)
In person poster session is joint with Women in Machine Learning. Poster Titles An Introduction to Quantum Natural Language Processing and a Study Case Bottleneck-based Encoder-decoder ARchitecture (BEAR) for Learning Unbiased Consumer-to-Consumer Image Representations Classifier Guided Diffusion for Image Inpainting: Applications to Fine Art A Model-Based Filter to Improve Local Differential Privacy A Study on the Predictability of Sample Learning Consistency Link Prediction from Heterogeneous Opinion Mining Networks with Multi-Domain Applications Transfer Learning with Joint Fine-Tuning for Multimodal Sentiment Analysis Interpretable Process Mining Opportunities and Challenges Virtual poster session is via GatherTown [ protected link dropped ] |
🔗 |
Mon 2:00 p.m. - 3:00 p.m.
|
Mentoring Session
(
Session
)
SlidesLive Video » |
🔗 |
Mon 3:00 p.m. - 3:30 p.m.
|
Closing Remarks
(
Closing
)
|
🔗 |
Mon 3:30 p.m. - 5:00 p.m.
|
Social Event
(
Social
)
|
🔗 |