메뉴 건너뛰기




Volumn Part F128415, Issue , 2017, Pages 1195-1199

Finding approximate local minima faster than gradient descent

Author keywords

Cubic regularization; Deep learning; Non convex optimization; Second order optimization

Indexed keywords

COMPLEX NETWORKS; COMPUTATIONAL COMPLEXITY; CONVEX OPTIMIZATION; DEEP LEARNING; EDUCATION;

EID: 85024401503     PISSN: 07378017     EISSN: None     Source Type: Conference Proceeding    
DOI: 10.1145/3055399.3055464     Document Type: Conference Paper
Times cited : (268)

References (31)
  • 3
    • 84999029527 scopus 로고    scopus 로고
    • Variance reduction for faster non-convex optimization
    • Zeyuan Allen-Zhu and Elad Hazan. Variance Reduction for Faster Non-Convex Optimization. In ICML, 2016.
    • (2016) ICML
    • Allen-Zhu, Z.1    Hazan, E.2
  • 7
    • 79952763936 scopus 로고    scopus 로고
    • Adaptive cubic regularisation methods for unconstrained optimization. Part I: Motivation, convergence and numerical results
    • Coralia Cartis, Nicholas IM Gould, and Philippe L Toint. Adaptive cubic regularisation methods for unconstrained optimization. part i: motivation, convergence and numerical results. Mathematical Programming, 127(2): 245-295, 2011.
    • (2011) Mathematical Programming , vol.127 , Issue.2 , pp. 245-295
    • Cartis, C.1    Gould, N.I.M.2    Toint, P.L.3
  • 8
    • 81255179401 scopus 로고    scopus 로고
    • Adaptive cubic regularisation methods for unconstrained optimization. Part II: Worst-case function-and derivative-evaluation complexity
    • Coralia Cartis, Nicholas IM Gould, and Philippe L Toint. Adaptive cubic regularisation methods for unconstrained optimization. part ii: worst-case function-and derivative-evaluation complexity. Mathematical Programming, 130(2): 295-319, 2011.
    • (2011) Mathematical Programming , vol.130 , Issue.2 , pp. 295-319
    • Cartis, C.1    Gould, N.I.M.2    Toint, P.L.3
  • 11
    • 80052250414 scopus 로고    scopus 로고
    • Adaptive subgradient methods for online learning and stochastic optimization
    • John Duchi, Elad Hazan, and Yoram Singer. Adaptive subgradient methods for online learning and stochastic optimization. The Journal of Machine Learning Research, 12: 2121-2159, 2011.
    • (2011) The Journal of Machine Learning Research , vol.12 , pp. 2121-2159
    • Duchi, J.1    Hazan, E.2    Singer, Y.3
  • 13
    • 84998770000 scopus 로고    scopus 로고
    • Robust shift-and-invert preconditioning: Faster and more sample efficient algorithms for eigenvector computation
    • Dan Garber, Elad Hazan, Chi Jin, Sham M. Kakade, Cameron Musco, Praneeth Netrapalli, and Aaron Sidford. Robust shift-and-invert preconditioning: Faster and more sample efficient algorithms for eigenvector computation. In ICML, 2016.
    • (2016) ICML
    • Garber, D.1    Hazan, E.2    Jin, C.3    Kakade, S.M.4    Musco, C.5    Netrapalli, P.6    Sidford, A.7
  • 16
    • 84984704687 scopus 로고    scopus 로고
    • Escaping from saddle points-online stochastic gradient for tensor decomposition
    • Paris, France, July 3-6, 2015
    • Rong Ge, Furong Huang, Chi Jin, and Yang Yuan. Escaping from saddle points-online stochastic gradient for tensor decomposition. In Proceedings of The 28th Conference on Learning Theory, COLT 2015, Paris, France, July 3-6, 2015, pages 797-842, 2015.
    • (2015) Proceedings of the 28th Conference on Learning Theory, COLT 2015 , pp. 797-842
    • Ge, R.1    Huang, F.2    Jin, C.3    Yuan, Y.4
  • 19
    • 84962468318 scopus 로고    scopus 로고
    • Accelerated gradient methods for nonconvex nonlinear and stochastic programming
    • feb
    • Saeed Ghadimi and Guanghui Lan. Accelerated gradient methods for nonconvex nonlinear and stochastic programming. Mathematical Programming, pages 1-26, feb 2015.
    • (2015) Mathematical Programming , pp. 1-26
    • Ghadimi, S.1    Lan, G.2
  • 21
    • 85024380018 scopus 로고    scopus 로고
    • A linear-time algorithm for trust region problems
    • Elad Hazan and Tomer Koren. A linear-time algorithm for trust region problems. Mathematical Programming, pages 1-19, 2015.
    • (2015) Mathematical Programming , pp. 1-19
    • Hazan, E.1    Koren, T.2
  • 22
    • 84891614753 scopus 로고    scopus 로고
    • Most tensor problems are np-hard
    • Christopher J. Hillar and Lek-Heng Lim. Most tensor problems are np-hard. J. ACM, 60(6): 45, 2013.
    • (2013) J. ACM , vol.60 , Issue.6 , pp. 45
    • Hillar, C.J.1    Lim, L.-H.2
  • 24
    • 0023452095 scopus 로고
    • Some np-complete problems in quadratic and nonlinear programming
    • Katta G Murty and Santosh N Kabadi. Some np-complete problems in quadratic and nonlinear programming. Mathematical programming, 39(2): 117-129, 1987.
    • (1987) Mathematical Programming , vol.39 , Issue.2 , pp. 117-129
    • Murty, K.G.1    Kabadi, S.N.2
  • 27
    • 33646730150 scopus 로고    scopus 로고
    • Cubic regularization of Newton method and its global performance
    • Yurii Nesterov and Boris T Polyak. Cubic regularization of newton method and its global performance. Mathematical Programming, 108(1): 177-205, 2006.
    • (2006) Mathematical Programming , vol.108 , Issue.1 , pp. 177-205
    • Nesterov, Y.1    Polyak, B.T.2
  • 28
    • 0000255539 scopus 로고
    • Fast exact multiplication by the hessian
    • Barak A Pearlmutter. Fast exact multiplication by the hessian. Neural computation, 6(1): 147-160, 1994.
    • (1994) Neural Computation , vol.6 , Issue.1 , pp. 147-160
    • Pearlmutter, B.A.1


* 이 정보는 Elsevier사의 SCOPUS DB에서 KISTI가 분석하여 추출한 것입니다.