Optimization · Non-Convex Optimization

TitleAuthors
Asymmetric Valleys: Beyond Sharp and Flat Local MinimaHaowei He · Gao Huang · Yang Yuan
Beyond Alternating Updates for Matrix Factorization with Inertial Bregman Proximal Gradient AlgorithmsMahesh Chandra Mukkamala · Peter Ochs
Efficiently escaping saddle points on manifoldsChristopher Criscitiello · Nicolas Boumal
Global Convergence of Least Squares EM for Demixing Two Log-Concave DensitiesWei Qian · Yuqian Zhang · Yudong Chen
Learning dynamic polynomial proofsAlhussein Fawzi · Mateusz Malinowski · Hamza Fawzi · Omar Fawzi
Non-asymptotic Analysis of Stochastic Methods for Non-Smooth Non-Convex Regularized ProblemsYi Xu · Rong Jin · Tianbao Yang
Nonconvex Low-Rank Tensor Completion from Noisy DataChangxiao Cai · Gen Li · H. Vincent Poor · Yuxin Chen
Painless Stochastic Gradient: Interpolation, Line-Search, and Convergence RatesSharan Vaswani · Aaron Mishkin · Issam Laradji · Mark Schmidt · Gauthier Gidel · Simon Lacoste-Julien
SpiderBoost and Momentum: Faster Variance Reduction AlgorithmsZhe Wang · Kaiyi Ji · Yi Zhou · Yingbin Liang · Vahid Tarokh
SSRGD: Simple Stochastic Recursive Gradient Descent for Escaping Saddle PointsZhize Li
The Landscape of Non-convex Empirical Risk with Degenerate Population RiskShuang Li · Gongguo Tang · Michael B Wakin
Who is Afraid of Big Bad Minima? Analysis of gradient-flow in spiked matrix-tensor modelsStefano Sarao Mannelli · Giulio Biroli · Chiara Cammarota · Florent Krzakala · Lenka Zdeborová
A Linearly Convergent Method for Non-Smooth Non-Convex Optimization on the Grassmannian with Applications to Robust Subspace and Dictionary LearningZhihui Zhu · Tianyu Ding · Daniel Robinson · Manolis Tsakiris · RenĂ© Vidal
Competitive Gradient DescentFlorian Schaefer · Anima Anandkumar
DINGO: Distributed Newton-Type Method for Gradient-Norm OptimizationRixon Crane · Fred Roosta
Efficient Smooth Non-Convex Stochastic Compositional Optimization via Stochastic Recursive Gradient DescentHuizhuo Yuan · Xiangru Lian · Chris Junchi Li · Ji Liu · Wenqing Hu
Efficiently avoiding saddle points with zero order methods: No gradients requiredEmmanouil-Vasileios Vlatakis-Gkaragkounis · Lampros Flokas · Georgios Piliouras
Escaping from saddle points on Riemannian manifoldsYue Sun · Nicolas Flammarion · Maryam Fazel
Exponentially convergent stochastic k-PCA without variance reductionCheng Tang
First-order methods almost always avoid saddle points: The case of vanishing step-sizesIoannis Panageas · Georgios Piliouras · Xiao Wang
Learning Sparse Distributions using Iterative Hard ThresholdingJacky Y Zhang · Rajiv Khanna · Anastasios Kyrillidis · Oluwasanmi Koyejo
Local SGD with Periodic Averaging: Tighter Analysis and Adaptive SynchronizationFarzin Haddadpour · Mohammad Mahdi Kamani · Mehrdad Mahdavi · Viveck Cadambe
Max-value Entropy Search for Multi-Objective Bayesian OptimizationSyrine Belakaria · Aryan Deshwal · Janardhan Rao Doppa
Solving a Class of Non-Convex Min-Max Games Using Iterative First Order MethodsMaher Nouiehed · Maziar Sanjabi · Tianjian Huang · Jason Lee · Meisam Razaviyayn
A Nonconvex Approach for Exact and Efficient Multichannel Sparse Blind DeconvolutionQing Qu · Xiao Li · Zhihui Zhu
An Inexact Augmented Lagrangian Framework for Nonconvex Optimization with Nonlinear ConstraintsMehmet Fatih Sahin · Armin eftekhari · Ahmet Alacaoglu · Fabian Latorre · Volkan Cevher
Bayesian Optimization with Unknown Search SpaceHuong Ha · Santu Rana · Sunil Gupta · Thanh Nguyen · Hung Tran-The · Svetha Venkatesh
Calculating Optimistic Likelihoods Using (Geodesically) Convex OptimizationViet Anh Nguyen · Soroosh Shafieezadeh Abadeh · Man-Chung Yue · Daniel Kuhn · Wolfram Wiesemann
Communication-Efficient Distributed Blockwise Momentum SGD with Error-FeedbackShuai Zheng · Ziyue Huang · James Kwok
Distributed Low-rank Matrix Factorization With Exact ConsensusZhihui Zhu · Qiuwei Li · Xinshuo Yang · Gongguo Tang · Michael B Wakin
Efficient Algorithms for Smooth Minimax OptimizationKiran Thekumparampil · Prateek Jain · Praneeth Netrapalli · Sewoong Oh
Momentum-Based Variance Reduction in Non-Convex SGDAshok Cutkosky · Francesco Orabona
Provable Non-linear Inductive Matrix CompletionKai Zhong · Zhao Song · Prateek Jain · Inderjit S Dhillon
Semi-flat minima and saddle points by embedding neural networks to overparameterizationKenji Fukumizu · Shoichiro Yamaguchi · Yoh-ichi Mototake · Mirai Tanaka
Shadowing Properties of Optimization AlgorithmsAntonio Orvieto · Aurelien Lucchi