Amir Daneshmand, Ph.D. '21
Email: amir.daneshmand (at) hotmail (dot) com
LinkedIn: https://www.linkedin.com/in/amirdaneshmand/
Google Scholar: https://scholar.google.com/citations?user=qEEYiYYAAAAJ&hl=en
Publications
See my google scholar for most up-to-date list of publications.
PhD Thesis
Parallel and Decentralized Algorithms for Big-data Optimization over Networks, PhD Thesis, Purdue University, 2021.
Journal Papers
with F. Facchinei, V. Kungurtsev, and G. Scutari
In IEEE Transactions on Signal Processing, volume 63, number 15, pages 3914-3929, August 2015.
with Y. Sun, G. Scutari and F. Facchinei, and Brian M. Sadler
in Journal Machine Learning Research (JMLR), 20(139), 1-62, 2019.
with G. Scutari and S. Kungurtsev
In SIAM Journal on Optimization, 30(4), 3029–3068, 2020.
with Y. Sun and G. Scutari
In SIAM Journal on Optimization, 32 (2), 354-385, 2022.
Conference Papers
with G. Scutari, P. Dvurechensky and A. Gasnikov
in Proceedings of the 38th International Conference on Machine Learning (ICML), PMLR 139:2398-2409, 2021.
with Y. Sun, G. Scutari and F. Facchinei
in Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pages 4084-4088, New Orleans, LA, March 2017.
with G. Scutari, V. Kungurtsev
in Proceedings of 2018 56th Annual Allerton Conference on Communication, Control, and Computing, pages 359-365, Monticello, IL, 2018.
with F. Facchinei, V. Kungurtsev, and G. Scutari
in Proceedings of the 48th Asilomar Conference on Signals, Systems and Computers, pages 3-7, Pacific Grove, CA, November 2014.
with Y. Sun, G. Scutari and F. Facchinei
in Proceedings of the 50th Asilomar Conference on Signals, Systems and Computers, pages 1001-1005, Pacific Grove, CA, November 2016.
Updates
May 2021:
Our paper "Newton Method over Networks is Fast up to the Statistical Precision" is accepted to International Conference on Machine Learning (ICML); read the preprint here.
Oct. 2020:
Our paper "Second-Order Guarantees of Distributed Gradient Algorithms" is published in SIAM Journal on Optimization (SIOPT); read it here.
Sept. 2019:
Our paper "Decentralized Dictionary Learning over Time-Varying Digraphs" is published in Journal of Machine Learning Research (JMLR); read it here.
April 2019:
I will present my recent works at the 8th Midwest Workshop on Control and Game Theory (MWCGT), April 27-28, Washington University, St. Louis, MO.
Aug. 2018:
I will present my work titled as "On the second-order-guarantees of decentralized optimization algorithms" at the DIMACS/TRIPODS/MOPTA conference, August 13-17, Lehigh University, Bethlehem, PA.
Aug. 2017:
I will present my work at the Distributed Optimization, Information Processing, and Learning workshop organized by Center for Discrete Mathematics and Theoretical Computer Science (DIMACS), August 21-23, Rutgers University, New Brunswick, New Jersey.
March 2017:
I will present my work on "Decentralized Statistical Learning over Dynamic Digraphs" at the 42nd ICASSP, March 5-9, New Orleans, Louisiana.
Nov. 2016:
I will present my work on "Distributed Dictionary Learning" at the 50th Asilomar Conference on Signals, Systems, and Computers, Nov 6-9, Pacific Grove, California.
Aug. 2015:
Our paper titled as "Hybrid Random/Deterministic Parallel Algorithms for Nonconvex Big Data Optimization" is published in IEEE transactions on Signal Processing (TSP); read it here.
Nov. 2014:
Our paper titled as "Flexible selective parallel algorithms for big data optimization" will be presented at the 48th Asilomar Conference on Signals, Systems, and Computers, Nov 2-5, Pacific Grove, California.