no code implementations • 9 Aug 2022 • Qingguo Hong, Jonathan W. Siegel, Qinyang Tan, Jinchao Xu
Our empirical studies also show that neural networks with the Hat activation function are trained significantly faster using stochastic gradient descent and ADAM.
2 code implementations • 21 Jul 2022 • Chenxi Wu, Min Zhu, Qinyang Tan, Yadhu Kartha, Lu Lu
Hence, we have considered a total of 10 different sampling methods, including six non-adaptive uniform sampling, uniform sampling with resampling, two proposed adaptive sampling, and an existing adaptive sampling.