I am a Lecturer and part of the Machine Learning Research Group within the School of Computer Science and Informatics at the University of Liverpool, UK.
I received my PhD in Computer Science from the Universidade Federal de Minas Gerais (UFMG), Brazil, with part of my doctoral research conducted at GIPSA-Lab, Institut Polytechnique de Grenoble, France.
My research interests lie at the intersection of machine learning, deep learning, pattern recognition, image processing, and remote sensing.
Prospective Students: I am always looking for motivated PhD students with/without scholarships. Please send an email to keiller.nogueira@liverpool.ac.uk to discuss your research interests. Some scholarships at University of Liverpool: DAMC CDT, NWSS DTP, CSC-Liverpool, and others.
Automatic mapping of high-risk urban areas for Aedes aegypti infestation based on building facade image analysis
C. Laranjeira, M. Pereira, R. Oliveira, G. Barbosa, C. Fernandes, P. Bermudi, E. Resende, E. Fernandes, K. Nogueira, V. Andrade, J. A. Quintanilha, J. A. Dos Santos, F. Chiaravalloti-Neto
Semantic segmentation of tree-canopy in urban environment with pixel-wise deep learning
J. A. C. Martins, K. Nogueira, L. P. Osco, F. D. G. Gomes, D. E. G. Furuya, W. N. Gonçalves, D. A. Sant'Ana, A. P. M. Ramos, V. Liesenberg, J. A. Dos Santos, P. T. S. De Oliveira, J. M. Junior
K. Nogueira, S. G. Fadel, I. C. Dourado, R. de O. Werneck, J. A. V. Muñoz, O. A. B. Penatti, R. T. Calumby, L. T. Li, J. A. dos Santos, R. da S. Torres
RECOD@Placing Task of MediaEval 2016: A Ranking Fusion Approach for Geographic-Location Prediction of Multimedia Objects
J. A. V. Muñoz, L. T. Li, I. C. Dourado, K. Nogueira, S. G. Fadel, O. A. B. Penatti, J. Almeida, L. A. M. Pereira, R. T. Calumby, J. A. dos Santos, R. da S. Torres
L. T. Li, J. A. V. Muñoz, J. Almeida, R. T. Calumby, O. A. B. Penatti, I. C. Dourado, K. Nogueira, P. R. Mendes Júnior, L. A. M. Pereira, D. C. G. Pedronette, J. A. dos Santos, M. A. Gonçalves, R. da S. Torres
My research develops machine and deep learning methods for image understanding,
with a strong emphasis on real-world Earth observation applications.
🛰️
Remote Sensing & Earth Observation
I develop deep learning methods for large-scale semantic segmentation, scene classification,
land cover mapping, change detection, and more in satellite and aerial imagery.
A key focus is building models that generalise across geographic regions and sensor types as well as
to reduce the annotation burden for operational Earth observation pipelines.
Pixel-level annotations for remote sensing are expensive and scarce.
My work explores data-centric and efficient techniques such as core-set selection
and label refinement to train high-performing segmentation models with less supervision.
IEEE GRSM'24Access'25arXiv'26
🔓
Scalable Open World Recognition & Novel Class Discovery
Real-world imagery contains classes unseen during training. I develop
open set semantic segmentation methods that leverage feature-space analysis to identify and reject unknown classes rather than forcing predictions into a fixed closed set.
LAGIRS'20Machine Learning'23
🌐
Multi-modal Image Understading
Real-world scene understanding requires integrating diverse sources of information. I develop methods that leverage multi-modal data - including satellite, drone, thermal, multispectral, hyperspectral, and 3D point cloud (LiDAR) imagery - to enable robust aerial and ground-level scene interpretation.
I apply deep learning to agricultural monitoring tasks including coffee crop recognition,
citrus orchard segmentation, tree canopy detection in urban environments,
and vegetation species discrimination using multispectral UAV and satellite imagery.
CIARP'15PRRS'16SIBGRAPI'17IGARSS'21
Datasets
🛰️
Clean-Noisy Aerial Building Segmentation Dataset
This dataset, introduced in the MVEO 2024 Challenge, is built upon the high-resolution SpaceNet8 dataset, which consists of RGB satellite imagery acquired by the WorldView-3 sensor over regions in East Louisiana (USA) and Germany, with spatial resolutions ranging from 0.3 m to 0.8 m per pixel. However, rather than using the full set of images and classes, this dataset is restricted to pre-flood imagery and includes only building-related classes, framing the problem as a binary semantic segmentation task (building vs background). The related original images are divided into 256x256 pixel patches. From these, 5,000 samples are randomly selected for training, while the remaining 1,298 samples are allocated for validation and testing. To simulate realistic annotation imperfections, synthetic label noise is introduced into the training segmentation masks after patch extraction. These perturbations aim to reflect common sources of annotation errors in real-world remote sensing datasets. Observe that: (i) no noise is introduced into the validation/testing images, and (ii) the clean and noisy training data is available.
Multi-view datasets in which each sample is composed of multiple images, typically including both aerial and ground perspectives. The datasets cover a range of classes, such as airport, bridge, church, forest, lake, river, skyscraper, stadium, statue, tower, and urban park.
This dataset is composed of 500 images each containing at least one bridge. It has samples collected from different regions around the world, which increases its diversity and representativeness due to the varying image and bridge properties, such as the orientation of the construction, landscape context (river or mountain), area (urban or countryside), etc. All images have 4,800x2,843 pixels and were manually annotated by specialists.
This dataset is a composition of scenes taken by SPOT sensor in 2005 over four counties in the State of Minas Gerais, Brazil: Arceburgo, Guaranesia, Guaxupé and Monte Santo. It has many intraclass variance caused by different crop management techniques. Also, coffee is an evergreen culture and the South of Minas Gerais is a mountainous region, which means that this dataset includes scenes with different plant ages and/or with spectral distortions caused by shadows.
Composed of images and associated tags/comments crawled from two fashion-related social networks, namely pose.com and chictopia.com. The pose.com dataset whas more than 3,000 images whereas the chictopia.com dataset has more than 2,000 images. Combining labels from both datasets leads us to a set of 31 discrete possibilities.
I am always looking for motivated researchers to join the Lab at the University of Liverpool. We work on computer vision, pattern recognition, and machine learning. Please read the position details below and reach out if you're interested.
Current Openings
🔬 Postdoctoral Research Associate – Multimodal Learning
We are seeking a postdoctoral researcher with expertise in multimodal learning, large language models, and computer vision.
Requirements: PhD in Computer Science or related field, strong publication record at top venues (CVPR, ICCV, NeurIPS, etc.), experience with PyTorch.
Past students and postdocs have gone on to positions at leading institutions and companies. If you are interested in joining the lab, please see the page.