In the second part of the paper, we show that Bernstein's on-line learning algorithm B can be converted into a batch learning algorithm ^B which consumes an (almost) minimal number of random training examples. This is true for a whole class of ``pac-style'' batch learning models (including learning with an $(\epsilon, \gamma)$-good model and learning with an $\epsilon$-bounded expected r-loss). The upper bound on the sample size is shown by applying a method of Littlestone from [Lit89] which converts an on-line learning algorithm A into a batch learning algorithm  and relates the sample size of  to the total loss of A. For this purpose, Littlestone's method is extended from 0,1-valued functions to real-valued functions. The lower bound on the sample size is shown by generalizing and applying a method of Simon from [Sim93b].