no code implementations • ICML 2020 • Arman Zharmagambetov, Miguel Carreira-Perpinan
We show that using TAO with the bagging approach produces much better forests than random forests, Adaboost or gradient boosting in every dataset we have tried across a wide range of input and output dimensionality and sample size.
no code implementations • 20 Oct 2018 • Yerlan Idelbayev, Miguel Carreira-Perpinan
We propose a software framework based on ideas of the Learning-Compression algorithm , that allows one to compress any neural network by different compression mechanisms (pruning, quantization, low-rank, etc.).