Indraneel Mukherjee
I am a fifth year graduate student in the Machine Learning group in the Computer Science Department at Princeton University. My advisor is Rob Schapire. I did my undergraduate in mathematics at Chennai Mathematical Institute, India.
Publications
-
The Rate of Convergence of AdaBoost. COLT 2011.
I. Mukherjee, C. Rudin, R. E. Schapire.
[paper.pdf | long version | slides.pptx by Cynthia Rudin ]We study two rates of convergence of AdaBoost: the first with respect to an arbitrary solution, and the second with respect to the optimal solution. In the first case, we achieve a rate that depends only on the approximation parameter and some notion of complexity of the reference solution. Further, we show the dependence on both parameters is polynomial, thereby answering positively the conjecture posed by Schapire in COLT 2010. In the second case, we show a rate of convergence that has optimal dependence on the approximation parameter. The two results require different techniques, and do not follow from each other. Unlike previous work, our rates do not require any assumption, such as weak learnability, finiteness of the optimal solution, or a compact space to work on. We also study the constants in our bounds and show that, in certain worst case situations, their values may be very large. Finally, we also construct many lower bound instances showing that our rates are nearly tight.
-
A Theory of Multiclass Boosting. NIPS 2010.
I. Mukherjee, R. E. Schapire.
[paper.pdf | supplement.pdf | poster.pdf | long version]We create a broad and general framework within which we identify the correct weak learning conditions for multiclass boosting, and design the most effective, in a game-theoretic sense, boosting algorithms that assume such requirements.
-
Relative Performance Guarantees for Approximate Inference in Latent Dirichlet Allocation. NIPS 2008.
I. Mukherjee, D. M. Blei.
[paper.pdf | spotlight.pdf | poster.pdf]We prove that the improvement of collapsed mean-field inference, aka collapsed variational inference, for LDA over ordinary mean-field inference decays inversely with the length of the documents. Our work suggests one should use collapsed inference for short texts, and the more efficient mean-field inference for longer documents.
-
Learning with Continuous Experts using Drifting Games. TCS June, 2010 (Preliminary version appeared in ALT 2008).
I. Mukherjee, R. E. Schapire.
[paper.pdf | talk.ppt | poster.pdf ]Complex experts predict more accurately, but are also harder to learn from. Learning from binary experts has been thoroughly studied previously. We provide a master strategy achieving tightly optimal regret bounds against the powerful class of continuous/random experts.