no code implementations • 15 Nov 2021 • Abdullahi Mohammad, Christos Masouros, Yiannis Andreopoulos
Symbol level precoding (SLP) has been proven to be an effective means of managing the interference in a multiuser downlink transmission and also enhancing the received signal power.
no code implementations • 15 Nov 2021 • Abdullahi Mohammad, Christos Masouros, Yiannis Andreopoulos
Symbol Level Precoding (SLP) has attracted significant research interest due to its ability to exploit interference for energy-efficient transmission.
no code implementations • 13 Oct 2021 • Abdullahi Mohammad, Christos Masouros, Yiannis Andreopoulos
Our results show that while SLP-DNet provides near-optimal performance, its quantized versions through SQ yield 3. 46x and 2. 64x model compression for binary-based and ternary-based SLP-SQDNets, respectively.
no code implementations • CVPR 2021 • Aaron Chadha, Yiannis Andreopoulos
We introduce the concept of rate-aware deep perceptual preprocessing (DPP) for video encoding.
no code implementations • 19 Apr 2021 • Abdullahi Mohammad, Christos Masouros, Yiannis Andreopoulos
This paper proposes an unsupervised learning-based precoding framework that trains deep neural networks (DNNs) with no target labels by unfolding an interior point method (IPM) proximal `log' barrier function.
no code implementations • 21 Aug 2020 • Alhabib Abbas, Yiannis Andreopoulos
We propose a novel mixture-of-experts class to optimize computer vision models in accordance with data transfer limitations at test time.
1 code implementation • 8 Oct 2019 • Yin Bi, Aaron Chadha, Alhabib Abbas, Eirina Bourtsoulatze, Yiannis Andreopoulos
The core of our framework comprises a spatial feature learning module, which utilizes residual-graph convolutional neural networks (RG-CNN), for end-to-end learning of appearance-based features directly from graphs.
1 code implementation • 12 Sep 2019 • Abdullahi Mohammad, Christos Masouros, Yiannis Andreopoulos
This paper introduces a framework for systematic complexity scaling of deep neural network(DNN) based MIMO detectors.
1 code implementation • ICCV 2019 • Yin Bi, Aaron Chadha, Alhabib Abbas, Eirina Bourtsoulatze, Yiannis Andreopoulos
Neuromorphic vision sensing (NVS)\ devices represent visual information as sequences of asynchronous discrete events (a. k. a., ``spikes'') in response to changes in scene reflectance.
no code implementations • 2 Aug 2019 • Eirina Bourtsoulatze, Aaron Chadha, Ilya Fadeev, Vasileios Giotsas, Yiannis Andreopoulos
We propose to use deep neural networks as precoders for current and future video codecs and adaptive video streaming systems.
1 code implementation • 27 Sep 2018 • Mohammad Jubran, Alhabib Abbas, Aaron Chadha, Yiannis Andreopoulos
Advanced video classification systems decode video frames to derive the necessary texture and motion representations for ingestion and analysis by spatio-temporal deep convolutional neural networks (CNNs).
no code implementations • 10 Sep 2018 • Aaron Chadha, Yiannis Andreopoulos
Adversarial discriminative domain adaptation (ADDA) is an efficient framework for unsupervised domain adaptation in image classification, where the source and target domains are assumed to have the same classes, but no labels are available for the target domain.
no code implementations • 14 Oct 2017 • Aaron Chadha, Alhabib Abbas, Yiannis Andreopoulos
We demonstrate that selective access to MB motion vector (MV) information within compressed video bitstreams can also provide for selective, motion-adaptive, MB pixel decoding (a. k. a., MB texture decoding).
no code implementations • 27 Nov 2016 • Aaron Chadha, Yiannis Andreopoulos
Our proposal is a compact image descriptor that combines the state-of-the-art in content-based descriptor extraction with a multi-level, Voronoi-based spatial partitioning of each dataset image.
no code implementations • 13 Sep 2015 • Alhabib Abbas, Nikos Deligiannis, Yiannis Andreopoulos
We create vectors of locally aggregated centers (VLAC) by first clustering SIFT features to obtain local feature centers (LFCs) and then encoding the latter with respect to given centers of local feature centers (CLFCs), extracted from a training set.