1.BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding paper:https://arxiv.org/pdf/1810.04805v2.pdf code:GitHub - google-research/bert: TensorFlow code and pre-trained mo
文章目录 Traditional (Vanilla/Classical) AutoencoderVariational AutoencoderKullback–Leibler divergenceReparameterization 这章讲述模型框架和概念的时间较多,好像并没有涉及过多的运算,重在一些概念的理解。 Traditional (Vanilla/Classical)