This document discusses online optimization algorithms. It begins with an introduction to online learning and its advantages over batch learning. It then provides background knowledge on relevant concepts like convex functions, gradients, loss functions, and regularization. It explains the differences between batch gradient descent and stochastic gradient descent. The document proceeds to describe several online optimization algorithms: Simple Coefficient Rounding (SCR), Truncated Gradient (TG), Forward-Backward Splitting (FOBOS), Regularized Dual Averaging (RDA), and Follow The Regularized Leader (FTRL). It provides detailed explanations of how SCR, TG, and FOBOS generate sparsity and their updating rules.