-
1
-
-
68949096711
-
SGD-QN: Careful quasi-Newton stochastic gradient descent
-
Bordes, Antoine, Bottou, Leon, and Gallinari, Patrick. SGD-QN: Careful quasi-Newton stochastic gradient descent. JMLR, 10:1737-1754, 2009.
-
(2009)
JMLR
, vol.10
, pp. 1737-1754
-
-
Bordes, A.1
Bottou, L.2
Gallinari, P.3
-
2
-
-
84907027033
-
-
arXiv: 1401.7020
-
Byrd, R.H., Hansen, S.L., Nocedal, Jorge, and Singer, Yoram. A stochastic quasi-Newton method for large- scale optimization. arXiv: 1401.7020, 2014.
-
(2014)
A Stochastic Quasi-Newton Method for Large- Scale Optimization
-
-
Byrd, R.H.1
Hansen, S.L.2
Nocedal, J.3
Singer, Y.4
-
3
-
-
84937908747
-
SAGA: A fast incremental gradient method with support for non-strongly convex composite objectives
-
Defazio, Aaron, Bach, Francis, and Lacoste-Julien, Simon. SAGA: A fast incremental gradient method with support for non-strongly convex composite objectives. In Advances in Neural Information Processing Systems, pp. 1646-1654, 2014.
-
(2014)
Advances in Neural Information Processing Systems
, pp. 1646-1654
-
-
Defazio, A.1
Bach, F.2
Lacoste-Julien, S.3
-
5
-
-
84953234319
-
Accelerated, parallel, and proximal coordinate descent
-
Fercoq, Olivier and Richtarik, Peter. Accelerated, parallel, and proximal coordinate descent. SIAM Journal on Optimization, 25(4): 1997-2023, 2015.
-
(2015)
SIAM Journal on Optimization
, vol.25
, Issue.4
, pp. 1997-2023
-
-
Fercoq, O.1
Richtarik, P.2
-
6
-
-
84912553418
-
Fast distributed coordinate descent for minimizing non-strongly convex losses
-
Fercoq, Olivier, Qu, Zheng, Richtarik, Peter, and Takac, Martin. Fast distributed coordinate descent for minimizing non-strongly convex losses. IEEE International Workshop on Machine Learning for Signal Processing, 2014.
-
(2014)
IEEE International Workshop on Machine Learning for Signal Processing
-
-
Fercoq, O.1
Qu, Z.2
Richtarik, P.3
Takac, M.4
-
8
-
-
84898963415
-
Accelerating stochastic gradient descent using predictive variance reduction
-
Johnson, Rie and Zhang, Tong. Accelerating stochastic gradient descent using predictive variance reduction. In NIPS, 2013.
-
(2013)
NIPS
-
-
Johnson, R.1
Zhang, T.2
-
10
-
-
84962869001
-
-
arXiv:1410.4744
-
Konecny, Jakub, Lu, Jie, Richtarik, Peter, and Takac, Martin. mS2GD: Mini-batch semi-stochastic gradient descent in the proximal setting. arXiv:1410.4744, 2014a.
-
(2014)
MS2GD: Mini-batch Semi-stochastic Gradient Descent in the Proximal Setting
-
-
Konecny, J.1
Lu, J.2
Richtarik, P.3
Takac, M.4
-
12
-
-
84937832340
-
-
Technical Report MSR-TR-2014-94, Microsoft Research, July
-
Lin, Qihang, Lu, Zhaosong, and Xiao, Lin. An accelerated proximal coordinate gradient method and its application to regularized empirical risk minimization. Technical Report MSR-TR-2014-94, Microsoft Research, July 2014.
-
(2014)
An Accelerated Proximal Coordinate Gradient Method and its Application to Regularized Empirical Risk Minimization
-
-
Lin, Q.1
Lu, Z.2
Xiao, L.3
-
13
-
-
84940372324
-
Incremental majorization-minimization optimization with application to large-scale machine learning
-
Mairal, Julien. Incremental majorization-minimization optimization with application to large-scale machine learning. SIAM Journal on Optimization, 25(2):829-855, 2015.
-
(2015)
SIAM Journal on Optimization
, vol.25
, Issue.2
, pp. 829-855
-
-
Mairal, J.1
-
19
-
-
84981247976
-
On optimal probabilities in stochastic coordinate descent methods
-
Richtarik, Peter and Takac, Martin. On optimal probabilities in stochastic coordinate descent methods. Optimization Letters, pp. 1-11, 2015.
-
(2015)
Optimization Letters
, pp. 1-11
-
-
Richtarik, P.1
Takac, M.2
-
20
-
-
84897116612
-
Iteration complexity of randomized block-coordinate descent methods for minimizing a composite function
-
Richtarik, Peter and Takac, Martin. Iteration complexity of randomized block-coordinate descent methods for minimizing a composite function. Mathematical Programming, 144(2): 1-38, 2014.
-
(2014)
Mathematical Programming
, vol.144
, Issue.2
, pp. 1-38
-
-
Richtarik, P.1
Takac, M.2
-
21
-
-
84958109123
-
Parallel coordinate descent methods for big data optimization
-
ISSN 0025-5610
-
Richtarik, Peter and Takac, Martin. Parallel coordinate descent methods for big data optimization. Mathematical Programming, pp. 1-52, 2015. ISSN 0025-5610. doi:10.1007/s10107-015-0901-6.
-
(2015)
Mathematical Programming
, pp. 1-52
-
-
Richtarik, P.1
Takac, M.2
-
23
-
-
72449211086
-
A stochastic quasi-Newton method for online convex optimization
-
Schraudolph, Nicol N., Yu, Jin, and Giinter, Simon. A stochastic quasi-Newton method for online convex optimization. In AISTATS, pp. 433-440, 2007.
-
(2007)
AISTATS
, pp. 433-440
-
-
Schraudolph, N.N.1
Yu, J.2
Giinter, S.3
-
25
-
-
84899021802
-
Accelerated mini- batch stochastic dual coordinate ascent
-
Shalev-Shwartz, Shai and Zhang, Tong. Accelerated mini- batch stochastic dual coordinate ascent. In NIPS, pp. 378-385, 2013a.
-
(2013)
NIPS
, pp. 378-385
-
-
Shalev-Shwartz, S.1
Zhang, T.2
-
26
-
-
84875134236
-
Stochastic dual coordinate ascent methods for regularized loss
-
Shalev-Shwartz, Shai and Zhang, Tong. Stochastic dual coordinate ascent methods for regularized loss. JMLR, 14(1):567-599, 2013b.
-
(2013)
JMLR
, vol.14
, Issue.1
, pp. 567-599
-
-
Shalev-Shwartz, S.1
Zhang, T.2
-
27
-
-
79952748054
-
Pegasos: Primal estimated sub-gradient solver for SVM
-
Shalev-Shwartz, Shai, Singer, Yoram, Srebro, Nati, and Cotter, Andrew. Pegasos: Primal estimated sub-gradient solver for SVM. Mathematical Programming, pp. 3-30, 2011.
-
(2011)
Mathematical Programming
, pp. 3-30
-
-
Shalev-Shwartz, S.1
Singer, Y.2
Srebro, N.3
Cotter, A.4
-
28
-
-
84919904380
-
Fast large-scale optimization by unifying stochastic gradient and quasi-Newton methods
-
Sohl-Dickstein, Jascha, Poole, Ben, and Ganguli, Surya. Fast large-scale optimization by unifying stochastic gradient and quasi-Newton methods. In ICML, 2014.
-
(2014)
ICML
-
-
Sohl-Dickstein, J.1
Poole, B.2
Ganguli, S.3
-
29
-
-
84897543082
-
-
ICML
-
Takac, Martin, Bijral, Avleen, Richtarik, Peter, and Srebro, Nathan. Mini-batch primal and dual methods for SVMs. In ICML, 2013.
-
(2013)
Mini-batch Primal and Dual Methods for SVMs
-
-
Takac, M.1
Bijral, A.2
Richtarik, P.3
Srebro, N.4
-
32
-
-
46749146509
-
A coordinate gradient descent method for nonsmooth separable minimization
-
Tseng, Paul and Yun, Sangwoon. A coordinate gradient descent method for nonsmooth separable minimization. Mathematical Programming, 117(1-2):387-423, 2009.
-
(2009)
Mathematical Programming
, vol.117
, Issue.1-2
, pp. 387-423
-
-
Tseng, P.1
Yun, S.2
-
33
-
-
84919793228
-
A proximal stochastic gradient method with progressive variance reduction
-
Xiao, Lin and Zhang, Tong. A proximal stochastic gradient method with progressive variance reduction. SIAM Journal on Optimization, 24(4):2057-2075, 2014.
-
(2014)
SIAM Journal on Optimization
, vol.24
, Issue.4
, pp. 2057-2075
-
-
Xiao, L.1
Zhang, T.2
|