CL

C.C.S. Liem

73 records found

Annotation Practices in Societally Impactful Machine Learning Applications

What are these automated systems actually trained on?

This study examines dataset annotation practices in influential NeurIPS research. Datasets employed in highly cited NeurIPS papers were assessed based on criteria concerning their item population, labelling schema, and annotation process. While high-level information, such as the ...
With the worsening of climate change, the complications brought on by floods every year create an increasing need for forecasting systems that humanitarian organizations can use to help populations in danger. This research presents a literature review of machine-learning models f ...
Natural disasters frequently cause casualties and property losses. Predicting and mitigating the impact of such threats is crucial to the work of humanitarian organizations. The interactions between hazards are best represented through a multi-hazard approach, and machine learnin ...

Annotation Practices in Societally Impactful Machine Learning Applications

What are these automated systems actually trained on?

The output of machine learning (ML) models can be only as good as the data that is fed into them. Because of this, when making datasets for creating ML models, it is important to ensure the quality of the data. This is especially true of human labeled data, which can be hard to s ...
Displacement is a focal point of humanitarian aid efforts, since it affects millions of people globally. Mitigating the consequences of forced migration is important for reducing suffering and one way of doing so is through predicting displacement to prioritise resources in advan ...

Machine learning for humanitarian forecasting: A Survey

Assessing the trustworthiness and real-world feasibility of machine learning models for conflict forecasting

As humanitarian needs increase while donor budgets decrease, anticipatory strategies are essential for effective crisis response. In this context, machine learning (ML) has emerged as a promising tool for crisis forecasting, offering the potential to support timely interventions ...

Behind the Labels: Transparency Pitfalls in Annotation Practices for Societally Impactful ML

A deep dive into annotation transparency and consistency in CVPR corpus

This study investigates annotation and reporting practices in machine learning (ML) research, focusing on societally impactful applications presented at the IEEE/CVF Computer Vision and Pattern Recognition (CVPR) conferences. By structurally analyzing the 75 most-cited CVPR paper ...

High-impact vision research still rests on datasets whose labels arrive via opaque, rarely documented pipelines. To understand how serious the problem is inside a large venue, we audited 75 TPAMI papers (2009-2024) that rely or introduce datasets. Each datase ...

Dataset quality within a societally impactful machine learning domain

An overview of data collection and annotation practices of the datasets used by papers published by the ACL

This study gives an overview of the data collection and annotation practices of the datasets used by the most impactful papers published by the Association of Computational Linguistics (ACL). This was achieved by selecting the most highly cited papers published within the ACL ant ...
In musical (jazz) improvisation, musicians that are just starting out can often feel uncomfortable when being put on the spot by their fellow players. However, when a musician is on their own when practising or leisurely playing, this prevents them from listening to fellow musici ...

Requirements Engineering for Machine Learning

A Study in Behavior-Driven Development

Machine Learning (ML) systems are increasingly used in high-stakes, socially impactful domains, requiring attention to improve explainability and trust. However, current Requirements Engineering (RE) techniques often fail to address these human-centered qualities. This research i ...
Central banks communicate their monetary policy plans to the public through meeting minutes or transcripts. These communications can have immense effects on markets and are often the subjects of studies in the financial literature. The recent advancements in Natural Language Proc ...
Large language models have achieved breakthroughs in many natural language processing tasks. One of their main appeals is the ability to tackle problems that lack sufficient training data to create a dedicated solution. Manga translation is one such task, a still budding and un ...
Counterfactual explanations can be applied to algorithmic recourse, which is concerned with helping individuals in the real world overturn undesirable algorithmic decisions. They aim to provide explanations to opaque machine learning models. Not all generated points are equally f ...
Adversarial Training has emerged as the most reliable technique to make neural networks robust to gradient-based adversarial perturbations on input data. Besides improving model robustness, preliminary evidence presents an interesting consequence of adversarial training -- increa ...
In recent years, the need for explainable artificial intelligence (XAI) has become increasingly important as complex black-box models are used in critical applications. While many methods have been developed to interpret these models, there is also potential in enhancing the mode ...
Counterfactual explanations (CEs) can be used to gain useful insights into the behaviour of opaque classification models, allowing users to make an informed decision when trusting such systems. Assuming the CEs of a model are faithful (they well represent the inner workings of th ...
Counterfactual Explanations (CE) are essential for understanding the predictions of black-box models by suggesting minimal changes to input features that would alter the output. Despite their importance in Explainable AI (XAI), there is a lack of standardized metrics to assess th ...

A Study on Counterfactual Explanations

Investigating the impact of inter-class distance and data imbalance

Counterfactual explanations (CEs) are emerging as a crucial tool in Explainable AI (XAI) for understanding model decisions. This research investigates the impact of various factors on the quality of CEs generated for classification tasks. We explore how inter-class distance, data ...

Developing a monitoring process for IPC Acute Food Insecurity analyses

A case study on Human-Centered AI for humanitarian decision-making

Due to climate change, man-made conflicts, and rising inflation, a growing number of people around the world are struggling to have consistent access to safe and nutritious food. This phenomenon is known as food insecurity (FI). Therefore, we take in this thesis the first steps t ...