I am an Algo Developer at Hudson River Trading.
I lead a team of researchers developing models and algorithms for trading on financial markets.
Prior to joining HRT, I was a Senior Research Scientist at Google AI. While at Google, I have developed machine learning solutions for a variety of problems including time series forecasting for Ads traffic, AutoML and machine translation.
I completed a Ph.D. degree in mathematics at the Courant Institute of Mathematical Sciences, where I worked with Professor
Mehryar Mohri. Before coming to the Courant Institute, I received Bachelor's and Master's degrees in mathematics from the University of Toronto.
Selected Academic Activity
Publications
(Google Scholar)
Discrepancy-based theory and algorithms for forecasting non-stationary time series.
Annals of Mathematics and Artificial Intelligence. 86(1):1-33, 2020.
Foundations of Sequence-to-Sequence Modeling for Time Series.
In Twenty-Second Conference on Artificial Intelligence and Statistics (AISTATS 2019). Naha, Okinawa, Japan, April 2019.
Online non-additive path learning under full and partial information.
In Proceedings of the 30th International Conference on Algorithmic
Learning Theory (ALT 2019). Chicago, USA, March 2019.
Efficient gradient computation for structured output learning with rational ad tropical losses.
In Advances in Neural Information Processing Systems (NeurIPS 2018). Montreal, Canada, December 2018.
Discriminative state-space models.
In Advances in Neural Information Processing Systems (NIPS 2017). Long Beach, CA, December 2017.
AdaNet: Adaptive structural learning of artificial neural networks.
In Proceedings of the 34st International Conference on Machine
Learning (ICML 2017). Sydney, Australia, August 2017.
Structured prediction theory based on factor graph complexity.
In Advances in Neural Information Processing Systems (NIPS 2016). Barcelona, Spain, December 2016.
Learning N-gram language models from uncertain data.
In Proceedings of the 17th Annual Conference of the International Speech Communication Association (Interspeech 2016). San Francisco, USA, September 2016.
Time series prediction and online learning.
In Proceedings of The 29th Annual Conference on
Learning Theory (COLT 2016). New York, USA, June 2016.
Generalization bounds for non-stationary mixing processes.
Machine Learning Journal, 106:1-25, 2016.
Kernel extraction via voted risk minimization.
Journal of Machine Learning Research (JMLR), 44:72-89, 2015.
Learning theory and algorithms for forecasting non-stationary time series.
In Advances in Neural Information Processing Systems (NIPS 2015). Montreal, Canada, December 2015.
(full oral, top 15 papers out of 1838 submissions)
On-line learning algorithms for path experts with non-additive losses.
In Proceedings of The 28th Annual Conference on
Learning Theory (COLT 2015). Paris, France, July 2015.
Structural maximum entropy models.
In Proceedings of the 32st International Conference on Machine
Learning (ICML 2015). Lille, France, July 2015.
Multi-class deep boosting.
In Advances in Neural Information Processing Systems (NIPS 2014). Montreal, Canada, December 2014.
Generalization bounds for time series prediction with non-stationary
processes.
In Proceedings of the 25th International Conference on Algorithmic
Learning Theory (ALT 2014). Bled, Slovenia, October 2014.
Learning ensembles of structured prediction rules.
In Proceedings of the 52nd Annual Meeting of Association
for Computational Linguistics (ACL 2014).
Baltimore, USA, June 2014.
Ensemble methods for structured prediction.
In Proceedings of the 31st International Conference on Machine
Learning (ICML 2014). Beijing, China, June 2014.
Nested recursions, simultaneous parameters and tree superpositions.
Electronic Journal of Combinatorics. Volume 21, Issue 1, 2014.
A combinatorial approach for solving certain nested recursions with non-slow
solutions.
Journal of Difference Equations and Applications,
Volume 19, Issue 4, 2013.
Sums of ceiling functions solve nested recursions
Journal of Difference Equations and Applications,
Volume 18, Issue 12, 2012.
Nested recursions with ceiling function solutions
Journal of Difference Equations and Applications,
Volume 18, Issue 6, 2012.
Edited Proceedings
Proceedings of Machine Learning Research: NIPS 2016 Time Series Workshop. Volume 55.
Other Manuscripts
Efficient gradient computation for structured output learning with rational ad tropical losses.
In NeurIPS 2018 Workshop on Visually Grounded Interaction and Language. Montreal, Canada, December 2018.
AdaNet: Adaptive Structural Learning of Artificial Neural Networks.
In ICML 2017 Workshop on Principled approaches to deep learning.
Sydney, Australia, August 2017.
Measuring Learnability in Structured Prediction using Factor Graph Complexity.
In NIPS 2016 Workshop on Learning in High Dimensions with Structure.
Barcelona, Spain, December 2016.
A Theoretical Framework for Structured Prediction using Factor Graph Complexity.
In NIPS 2016 Workshop on Extreme Classification.
Barcelona, Spain, December 2016.
AdaNet: Adaptive Structural Learning of Artificial Neural Networks.
In NIPS 2016 Workshop on Adaptive and Scalable Nonparametric Methods in Machine Learning.
Barcelona, Spain, December 2016.
AdaNet: Adaptive Structural Learning of Artificial Neural Networks.
In NIPS 2016 Workshop on Efficient Methods for Deep Neural Networks.
Barcelona, Spain, December 2016.
Rademacher complexity margin bounds for learning with
a large number of classes.
In ICML 2015 Workshop on Extreme Classification.
Lille, France, July 2015.
Boosting ensembles of structured prediction rules.
In NIPS 2014 Workshop on Modern Machine Learning and Natural Language
Processing.
Montreal, Canada, December 2014.
On-line learning approach to ensemble methods for
structured prediction.
In NIPS 2014 Workshop on Representation and Learning Methods
Complex Outputs.
Montreal, Canada, December 2014.
Forecasting non-stationary time series: from theory to algorithms.
In NIPS 2014 Workshop on Transfer and Multi-task Learning.
Montreal, Canada, December 2014.
Selected Talks
ICML Workshop on Features and Structures 2015 (invited lecture),
Lille, France, July 2015.
Slides
NIPS 2014 Workshop on Representation and
Learning Methods Complex Outputs, Montreal, Canada, December 2014.
Video