Bayes : Average of a model over the posterior distribution.
Gibbs : A model with a random parameter from the posterior distribution.
We show that there is a universal law between them which holds in both regular and singular cases. In general, Bayes makes the generalization loss smaller than Gibbs, however, it seems to be easier to realize Gibbs estimation than Bayes estimation, in deep learning.
(Bg-S)+(CV-Sn)=2(lambda)/n +(1/n),
where CV is the Bayes cross validatin.
References
If you are interested in this result, please find
S. Watanabe, Equations of states for singular statistical estimation. Neural Networks, vol.23,pp.20-34, 2009. Arxiv:0712.0653
For the case when q(x) is unrealizable by and regular for p(x|w),
S. Watanabe, Equations of states in statistical learning for an unrealizable and regular case, IEICE transactions on fundamentals of electronics, communications, and computer sciences, vol.E93-A, pp.617-626, 2010. Arxiv:0906.0211
In this case, the real log canonical threshold and the singular fluctuation are give by d/2 and tr(IJ^{-1})/2, respectively, where d is the dimension of the parameter, I is the Fisher information matrix, and J is the Hesse matrix of the log loss.