Indraneel Mukherjee
I graduated with a PhD from the Machine Learning group in the Computer Science Department at Princeton University. My advisor was Rob Schapire. I did my undergraduate in mathematics at Chennai Mathematical Institute, India.
Latest
I am currently at LiftIgniter.Personal
My favorite blog: MoonchasingPublications


The Rate of Convergence of AdaBoost. COLT 2011.
I. Mukherjee, C. Rudin, R. E. Schapire.
[paper.pdf  long version  slides.pptx by Cynthia Rudin ]We study two rates of convergence of AdaBoost: the first with respect to an arbitrary solution, and the second with respect to the optimal solution. In the first case, we achieve a rate that depends only on the approximation parameter and some notion of complexity of the reference solution. Further, we show the dependence on both parameters is polynomial, thereby answering positively the conjecture posed by Schapire in COLT 2010. In the second case, we show a rate of convergence that has optimal dependence on the approximation parameter. The two results require different techniques, and do not follow from each other. Unlike previous work, our rates do not require any assumption, such as weak learnability, finiteness of the optimal solution, or a compact space to work on. We also study the constants in our bounds and show that, in certain worst case situations, their values may be very large. Finally, we also construct many lower bound instances showing that our rates are nearly tight.

A Theory of Multiclass Boosting. NIPS 2010.
I. Mukherjee, R. E. Schapire.
[paper.pdf  supplement.pdf  poster.pdf  long version]We create a broad and general framework within which we identify the correct weak learning conditions for multiclass boosting, and design the most effective, in a gametheoretic sense, boosting algorithms that assume such requirements.

Relative Performance Guarantees for Approximate Inference in Latent Dirichlet Allocation. NIPS 2008.
I. Mukherjee, D. M. Blei.
[paper.pdf  spotlight.pdf  poster.pdf]We prove that the improvement of collapsed meanfield inference, aka collapsed variational inference, for LDA over ordinary meanfield inference decays inversely with the length of the documents. Our work suggests one should use collapsed inference for short texts, and the more efficient meanfield inference for longer documents.

Learning with Continuous Experts using Drifting Games. TCS June, 2010 (Preliminary version appeared in ALT 2008).
I. Mukherjee, R. E. Schapire.
[paper.pdf  talk.ppt  poster.pdf ]Complex experts predict more accurately, but are also harder to learn from. Learning from binary experts has been thoroughly studied previously. We provide a master strategy achieving tightly optimal regret bounds against the powerful class of continuous/random experts.