The mean-field Langevin dynamics (MFLD) is a nonlinear generalization of...
Recent studies have experimentally shown that we can achieve in non-Eucl...
The entropic fictitious play (EFP) is a recently proposed algorithm that...
Stochastic gradient descent is a workhorse for training deep neural netw...
We propose a new bound for generalization of neural networks using Koopm...
As an example of the nonlinear Fokker-Planck equation, the mean field
La...
Hyperbolic ordinal embedding (HOE) represents entities as points in
hype...
Model extraction attacks have become serious issues for service provider...
We propose the particle dual averaging (PDA) method, which generalizes t...
Spatial attention has been introduced to convolutional neural networks (...
In online learning from non-stationary data streams, it is both necessar...
We analyze the convergence of the averaged stochastic gradient descent f...
While second order optimizers such as natural gradient descent (NGD) oft...
Although kernel methods are widely used in many learning problems, they ...
Deep learning has exhibited superior performance for various tasks,
espe...
Data cleansing is a typical approach used to improve the accuracy of mac...
Recently, several studies have proven the global convergence and
general...
We consider stochastic gradient descent for binary classification proble...
Residual Networks (ResNets) have become state-of-the-art models in deep
...
We propose a new technique that boosts the convergence of training gener...
The superior performance of ensemble methods with infinite models are we...
We propose an optimization method for minimizing the finite sums of smoo...