no code implementations • EMNLP (ALW) 2020 • Kosisochukwu Madukwe, Xiaoying Gao, Bing Xue
Recently, a few studies have discussed the limitations of datasets collected for the task of detecting hate speech from different viewpoints.
no code implementations • 11 May 2024 • Hengzhe Zhang, Qi Chen, Bing Xue, Wolfgang Banzhaf, Mengjie Zhang
In recent years, genetic programming (GP)-based evolutionary feature construction has achieved significant success.
no code implementations • 9 May 2024 • Bing Xue, Katsuyuki Haneda, Clemens Icheln, Juha Ala-Laurinaho
Next, the model is simplified by replacing the finger with an infinite sheet of human skin to calculate the forward electromagnetic problem related to the permittivity characterization.
1 code implementation • 21 Mar 2024 • Ben Cravens, Andrew Lensen, Paula Maddigan, Bing Xue
Our experimental analysis demonstrates that GP-EMaL is able to match the performance of the existing approach in most cases, while using simpler, smaller, and more interpretable tree structures.
no code implementations • 6 Mar 2024 • Paula Maddigan, Andrew Lensen, Bing Xue
In this research, we investigate the potential of leveraging eXplainable AI (XAI) and large language models (LLMs) like ChatGPT to improve the interpretability of GP-based non-linear dimensionality reduction.
1 code implementation • 1 Mar 2024 • Christian Raymond, Qi Chen, Bing Xue, Mengjie Zhang
In this paper, we develop upon the topic of loss function learning, an emergent meta-learning paradigm that aims to learn loss functions that significantly improve the performance of the models trained under them.
1 code implementation • 27 Feb 2024 • Bing Xue, Charles Alba, Joanna Abraham, Thomas Kannampallil, Chenyang Lu
Adapting models through self-supervised finetuning further improved performance by 3. 2% for AUROC & 1. 5% for AUPRC Incorporating labels into the finetuning procedure further boosted performances, with semi-supervised finetuning improving by 1. 8% for AUROC & 2% for AUPRC & foundational modelling improving by 3. 6% for AUROC & 2. 6% for AUPRC compared to self-supervised finetuning.
no code implementations • 1 Feb 2024 • Kaan Demir, Bach Nguyen, Bing Xue, Mengjie Zhang
The consistency of surrogate loss functions is not proven and is exacerbated by the conflicting nature of multi-label loss functions.
no code implementations • 18 Aug 2023 • Carl McMillan, Junhong Zhao, Bing Xue, Ross Vennell, Mengjie Zhang
To establish an effective deep learning model for buoy detection with a limited number of labeled data, we employ transfer learning techniques.
no code implementations • 6 Jul 2023 • Bing Xue, Ahmed Sameh Said, Ziqi Xu, Hanyang Liu, Neel Shah, Hanqing Yang, Philip Payne, Chenyang Lu
TVAE is specifically designed to address the modeling challenges like ECMO with strong treatment selection bias and scarce treatment cases.
1 code implementation • IEEE Transactions on Evolutionary Computation 2023 • Hengzhe Zhang, Aimin Zhou, Qi Chen, Bing Xue, Mengjie Zhang
Ensemble learning methods have been widely used in machine learning in recent years due to their high predictive performance.
no code implementations • 30 Jan 2023 • Christian Raymond, Qi Chen, Bing Xue, Mengjie Zhang
Loss function learning is a new meta-learning paradigm that aims to automate the essential task of designing a loss function for a machine learning model.
no code implementations • 7 Dec 2022 • Bin Wang, Bing Xue, Mengjie Zhang
In this paper, inspired by transfer learning, a new evolutionary computation based framework is proposed to efficiently evolve CNNs without compromising the classification accuracy.
no code implementations • 28 Nov 2022 • Bin Wang, Wenbin Pei, Bing Xue, Mengjie Zhang
The evolved local explanations on four images, randomly selected from ImageNet, are presented, which show that the evolved local explanations are straightforward to be recognised by humans.
1 code implementation • 10 Oct 2022 • Dingwen Li, Bing Xue, Christopher King, Bradley Fritz, Michael Avidan, Joanna Abraham, Chenyang Lu
Towards this end, we propose a hierarchical model combining the strength of both attention and recurrent models for intraoperative time series.
no code implementations • 27 Sep 2022 • Ying Bi, Bing Xue, Mengjie Zhang
The new approach can automatically evolve variable-length models using many important operators from both image and classification domains.
no code implementations • 19 Sep 2022 • Christian Raymond, Qi Chen, Bing Xue, Mengjie Zhang
In this paper, we develop upon the emerging topic of loss function learning, which aims to learn loss functions that significantly improve the performance of the models trained under them.
no code implementations • 14 Sep 2022 • Ying Bi, Bing Xue, Pablo Mesejo, Stefano Cagnoni, Mengjie Zhang
This survey aims to provide a better understanding of evolutionary computer vision (ECV) by discussing the contributions of different approaches and exploring how and why EC is used for CV and image analysis.
no code implementations • 23 Aug 2022 • Nan Li, Lianbo Ma, Guo Yu, Bing Xue, Mengjie Zhang, Yaochu Jin
Specifically, we firstly illuminate EDL from machine learning and EC and regard EDL as an optimization problem.
no code implementations • 3 Aug 2022 • Bing Xue, Pasi Koivumaki, Lauri Vaha-Savo, Katsuyuki Haneda, Clemens Icheln
Their radiation properties are evaluated through near-field scanning of the two prototypes, first in free space for calibration of the antenna measurement system and for building simplified models of the cellphone arrays.
no code implementations • 29 Sep 2021 • Bing Xue, York Jiao, Thomas Kannampallil, Joanna Abraham, Christopher Ryan King, Bradley A Fritz, Michael Avidan, Chenyang Lu
Given the risks and cost of surgeries, there has been significant interest in exploiting predictive models to improve perioperative care.
no code implementations • 23 Aug 2021 • Andrew Lensen, Bing Xue, Mengjie Zhang
Recently, genetic programming has emerged as a very promising approach to manifold learning by evolving functional mappings from the original space to an embedding.
1 code implementation • 9 Aug 2021 • Xiangning Xie, Yuqiao Liu, Yanan sun, Gary G. Yen, Bing Xue, Mengjie Zhang
The paper conducts efficient comparison experiments on eight ENAS algorithms with high GPU utilization on this platform.
no code implementations • 17 Dec 2020 • Ying Bi, Bing Xue, Mengjie Zhang
Therefore, this paper develops a multitask GP approach to image feature learning for classification with limited training data.
no code implementations • 3 Dec 2020 • Trevor Londt, Xiaoying Gao, Bing Xue, Peter Andreae
Researchers have not applied EDL techniques to search the architecture space of char-CNNs for text classification tasks.
no code implementations • 25 Aug 2020 • Yuqiao Liu, Yanan sun, Bing Xue, Mengjie Zhang, Gary G. Yen, Kay Chen Tan
Deep Neural Networks (DNNs) have achieved great success in many applications.
no code implementations • 15 Aug 2020 • Yuqiao Liu, Yanan sun, Bing Xue, Mengjie Zhang
Hyperspectral images (HSIs) are susceptible to various noise factors leading to the loss of information, and the noise restricts the subsequent HSIs object detection and classification tasks.
no code implementations • 3 Jul 2020 • Bin Wang, Bing Xue, Mengjie Zhang
A new effective and efficient surrogate-assisted particle swarm optimisation algorithm is proposed to automatically evolve convolutional neural networks.
no code implementations • 16 Feb 2020 • Yanan Sun, Ziyao Ren, Gary G. Yen, Bing Xue, Mengjie Zhang, Jiancheng Lv
Data mining on existing CNN can discover useful patterns and fundamental sub-comments from their architectures, providing researchers with strong prior knowledge to design proper CNN architectures when they have no expertise in CNNs.
1 code implementation • 28 Jan 2020 • Benjamin Patrick Evans, Bing Xue, Mengjie Zhang
We apply this to the area of automated machine learning (by extending TPOT), to produce pipelines which can effectively be claimed to be free from human input, and show that the results are competitive with existing state-of-the-art which use hand-selected hyperparameter values.
1 code implementation • 27 Jan 2020 • Andrew Lensen, Bing Xue, Mengjie Zhang
Many domains require an understanding of the data in terms of the original features; there is hence a need for powerful visualisation methods which use understandable models.
no code implementations • 23 Jan 2020 • Benjamin Patrick Evans, Bing Xue, Mengjie Zhang
A common problem machine learning developers are faced with is overfitting, that is, fitting a pipeline too closely to the training data that the performance degrades for unseen data.
no code implementations • 5 Jan 2020 • Andrew Lensen, Mengjie Zhang, Bing Xue
This method required the dimensionality of the embedding to be known a priori, which makes it hard to use when little is known about a dataset.
1 code implementation • NeurIPS 2019 • Zhen Zhang, Yijian Xiang, Lingfei Wu, Bing Xue, Arye Nehorai
Graph matching plays a central role in such fields as computer vision, pattern recognition, and bioinformatics.
no code implementations • 22 Oct 2019 • Andrew Lensen, Bing Xue, Mengjie Zhang
In this paper, we propose a new approach to automatically evolving similarity functions for a given clustering algorithm by using genetic programming.
1 code implementation • 28 Sep 2019 • Benjamin Patrick Evans, Harith Al-Sahaf, Bing Xue, Mengjie Zhang
Image classification is an essential task in computer vision, which aims to categorise a set of images into different groups based on some visual criteria.
no code implementations • 12 Aug 2019 • Samaneh Azari, Bing Xue, Mengjie Zhang, Lifeng Peng
The GP method along with RF and SVR, each is used for post-processing the results of peptide identification by PEAKS, a commonly used de novo sequencing method.
no code implementations • 29 Jul 2019 • Bin Wang, Bing Xue, Mengjie Zhang
Deep Convolutional Neural Networks (CNNs) have been widely used in image classification tasks, but the process of designing CNN architectures is very complex, so Neural Architecture Search (NAS), automatically searching for optimal CNN architectures, has attracted more and more research interests.
1 code implementation • GECCO '19: Proceedings of the Genetic and Evolutionary Computation Conference 2019 • Ying Bi, Bing Xue, Mengjie Zhang
An ensemble consists of multiple learners and can achieve a better generalisation performance than a single learner.
1 code implementation • 21 Mar 2019 • Bin Wang, Yanan sun, Bing Xue, Mengjie Zhang
In recent years, convolutional neural networks (CNNs) have become deeper in order to achieve better classification accuracy in image classification.
no code implementations • 10 Mar 2019 • Bin Wang, Yanan sun, Bing Xue, Mengjie Zhang
Three major contributions of this work are: Firstly, a new encoding strategy is proposed to encode a CNN, where the architecture and the shortcut connections are encoded separately; Secondly, a hybrid two-level EC method, which combines particle swarm optimisation and genetic algorithms, is developed to search for the optimal CNNs; Lastly, an adjustable learning rate is introduced for the fitness evaluations, which provides a better learning rate for the training process given a fixed number of epochs.
no code implementations • 8 Feb 2019 • Andrew Lensen, Bing Xue, Mengjie Zhang
Exploratory data analysis is a fundamental aspect of knowledge discovery that aims to find the main characteristics of a dataset.
no code implementations • 28 Oct 2018 • Yanan Sun, Bing Xue, Mengjie Zhang, Gary G. Yen
The proposed algorithm is evaluated on CIFAR10 and CIFAR100 against 18 state-of-the-art peer competitors.
no code implementations • 20 Aug 2018 • Bin Wang, Yanan sun, Bing Xue, Mengjie Zhang
In this paper, a new hybrid differential evolution (DE) algorithm with a newly added crossover operator is proposed to evolve the architectures of CNNs of any lengths, which is named DECNN.
4 code implementations • 11 Aug 2018 • Yanan Sun, Bing Xue, Mengjie Zhang, Gary G. Yen
Convolutional Neural Networks (CNNs) have gained a remarkable success on many image classification tasks in recent years.
no code implementations • 17 Mar 2018 • Bin Wang, Yanan sun, Bing Xue, Mengjie Zhang
Convolutional neural networks (CNNs) are one of the most effective deep learning methods to solve image classification problems, but the best architecture of a CNN to solve a specific problem can be extremely complicated and hard to design.
no code implementations • 2 Feb 2018 • Andrew Lensen, Bing Xue, Mengjie Zhang
Recently, feature selection has become an increasingly important area of research due to the surge in high-dimensional datasets in all areas of modern life.
1 code implementation • 13 Dec 2017 • Yanan Sun, Bing Xue, Mengjie Zhang, Gary G. Yen
Convolutional auto-encoders have shown their remarkable performance in stacking to deep convolutional neural networks for classifying image data during past several years.
1 code implementation • 30 Oct 2017 • Yanan Sun, Bing Xue, Mengjie Zhang, Gary G. Yen
Evolutionary computation methods have been successfully applied to neural networks since two decades ago, while those methods cannot scale well to the modern deep neural networks due to the complicated architectures and large quantities of connection weights.