Statistically adaptive learning for a general class of cost functions (SA L-BFGS) is a new method that modifies batch L-BFGS optimization to perform near real-time updates using statistical tools. It balances the contributions of previous weights, old training examples, and new training examples to achieve fast convergence with few iterations. This makes it highly scalable for problems with trillions of features, billions of examples, and millions of parameters. The method lies between pure L-BFGS minimization and online learning by taking advantage of situations where new data is similar to old data or different, outperforming other leading methods like Vowpal Wabbit and AllReduce on large datasets.