no code implementations • 30 Mar 2024 • Victor Rodriguez-Fernandez, Alejandro Carrasco, Jason Cheng, Eli Scharf, Peng Mun Siew, Richard Linares
Recent trends are emerging in the use of Large Language Models (LLMs) as autonomous agents that take actions based on the content of the user text prompts.
no code implementations • 28 Feb 2024 • Cristian Ramirez-Atencia, Victor Rodriguez-Fernandez, David Camacho
In this work, a DSS consisting of ranking and filtering systems, which order and reduce the optimal solutions, has been designed.
no code implementations • 8 Jan 2024 • Victor Rodriguez-Fernandez, Sumiyajav Sarangerel, Peng Mun Siew, Pablo Machuca, Daniel Jang, Richard Linares
With the rapid increase in the number of Anthropogenic Space Objects (ASOs), Low Earth Orbit (LEO) is facing significant congestion, thereby posing challenges to space operators and risking the viability of the space environment for varied uses.
no code implementations • 11 Dec 2023 • Emma Stevenson, Victor Rodriguez-Fernandez, Hodei Urrutxua, Vincent Morand, David Camacho
This paper presents a novel methodology for improving the performance of machine learning based space traffic management tasks through the use of a pre-trained orbit model.
no code implementations • 25 Oct 2023 • Julia Briden, Peng Mun Siew, Victor Rodriguez-Fernandez, Richard Linares
As the peak of the solar cycle approaches in 2025 and the ability of a single geomagnetic storm to significantly alter the orbit of Resident Space Objects (RSOs), techniques for atmospheric density forecasting are vital for space situational awareness.
1 code implementation • 8 Feb 2023 • Victor Rodriguez-Fernandez, David Montalvo, Francesco Piccialli, Grzegorz J. Nalepa, David Camacho
DeepVATS trains, in a self-supervised way, a masked time series autoencoder that reconstructs patches of a time series, and projects the knowledge contained in the embeddings of that model in an interactive plot, from which time series patterns and anomalies emerge and can be easily spotted.
1 code implementation • 30 Nov 2021 • Arnau Martí Sarri, Victor Rodriguez-Fernandez
CLIP (Contrastive Language-Image Pretraining) is an efficient method for learning computer vision tasks from natural language supervision that has powered a recent breakthrough in deep learning due to its zero-shot transfer capabilities.