TA

Thomas Abeel

23 records found

Reliable estimation of intra-host viral diversity is essential for understanding viral evolution,
treatment resistance, and outbreak dynamics. However, technical artefacts introduced during
sample preparation and sequencing can distort variant frequencies and lead to inco ...
This paper investigates how the local optimization method and strategy affect the efficiency of genetic algorithms (GAs) for Lennard-Jones (LJ) clusters. Several ASE-implemented optimizers were considered; however, only BFGS, FIRE, and Conjugate Gradient (CG) proved viable for in ...
Finding the lowest-energy structure of a cluster of atoms is an NP-Hard problem with applications in materials science. Genetic Algorithms (GAs) have shown promise in solving this problem due to their ability to explore complex energy landscapes. A critical component of GAs are t ...
Global Geometry (or Cluster) optimization is the process of finding the most stable formations of a cluster of some atoms. A genetic algorithm was developed to find the global minimum of a cluster using the Lennard-Jones atom interaction model efficiently. Determining the optimal ...

Genetic Algorithms for Solving the Global Geometry Optimization Problem

Evaluating Initialization and Crossover Strategies for Lennard-Jones Cluster Optimization

Discovery of new materials is essential in a lot if different fields, such as, space exploration, maritime industry and others. To stop new materials undergoing spontaneous reactions or reacting with the environment, they have to stable or at least metastable. That is where Globa ...

Improving research data reusability through data conversations

Bridging gaps in metadata supply and demand

Efficient and inclusive data reuse across research disciplines is based on high quality metadata that bridges the gap between data producers and consumers. This gap, referred to as the metadata gap, arises when the metadata provided by producers do not meet the needs of consumers ...
Decision-Focused Learning (DFL) focuses on a setting where a system gets as input some features and needs to predict coefficients to a downstream optimization problem. Classically, one would apply a two-stage solution, which trains the predictor as a regression task and only uses ...

Sparse Transformers are (in)Efficient Learners

Comparing Sparse Feedforward Layers in Small Transformers

Although transformers are state-of-the-art models for natural language tasks, obtaining reasonable performance still often requires large transformers which are expensive to train and deploy. Fortunately, there are techniques to increase the size of transformers without extra com ...

Tokenization Matters: Training your Tokenizer Right

Testing the Impact of Tokenization on Language Modelling with (Small) Transfomers

Large language models (LLMs) are rapidly increasing in parameter count, but this growth is not matched by an availability of high-quality data. This discrepancy raises concerns about the sustain- ability of current approaches to language model improvement, especially as forecasts ...

Pushing the Limits of the Compressive Memory Introduced in Infini-Attention

Architectural Decisions for Language Modelling with (Small) Transformers

Transformers are a type of neural network archi- tecture used in natural language processing. They excel in tasks such as translation, text generation, and language modeling by capturing long-range de- pendencies. Increasing input sequence length en- hances performance but at a h ...
RNA viruses, characterized by high replication rates and the absence of proofreading mechanisms,
are susceptible to errors during replication. This characteristic allows them to form diverse
communities of genome mutants known as "viral quasispecies". Each individual geno ...

Evaluating Adaptive Activation Functions in Language Models

Does choice of activation function matter in smaller Langaunge Models?

The rapid expansion of large language models (LLMs) driven by the transformer architecture has raised concerns about the lack of high-quality train ing data. This study investigates the role of acti vation functions in smaller-scale language models, specifically those with app ...
Graph neural networks (GNNs), while effective at various tasks on complex graph-structured data, lack interpretability. Post-hoc explainability techniques developed for these GNNs in order to overcome their inherent uninterpretability have been applied to the additional task of d ...
Predicting properties, such as toxicity or water solubility of unknown molecules with Graph Neural Networks has applications in drug research. Because of the ethical concerns associated with using artificial intelligence techniques in the medical field, explainable artificial int ...
AI explainers are tools capable of approximating how a neural network arrived at a given predic- tion by providing parts of the input data most rel- evant for the model’s choice. These tools have become a major point of research due to a need for human-verifiable predictions in m ...

Exploring Speed/Quality Trade-offs in Dimensionality of Attention Mechanism

Optimization with Grouped Query Attention and Diverse Key-Query-Value Dimensionalities

The advent of transformer architectures revolutionized natural language processing, particularly with the popularity of decoder-only transformers for text generation tasks like GPT models. However, the autoregressive nature of these models challenges their inference speed, crucia ...
The evaluation metrics commonly used for machine learning models often fail to adequately reveal the inner workings of the models, which is particularly necessarily in critical fields like healthcare. Explainable AI techniques, such as counterfactual explanations, offer a way to ...

Interactive semantic segmentation of 3D medical images

Comparative analysis of discrete and gradient descent based batch query retrieval methods in active learning

Accurate segmentation of anatomical structures and abnormalities in medical images is crucial, but manual segmentation is time-consuming and automated approaches lack clinical accuracy. In recent years, active learning approaches that aim to combine automatic segmentation with ma ...
Segmentation of 3D medical images is useful for various medical tasks. However, fully automated segmentation lacks the accuracy required for medical purposes while manual segmentation is too time-consuming. Therefore, an active learning method can be used to generate an accurate ...
Although automated segmentation of 3D medical images produce near-ideal results, they encounter limitations and occasional errors, necessitating manual intervention for error correction. Recent studies introduce an active learning pipeline as an efficient solution for this, requi ...