Project_1 Replace the Embedding layer in Transformer or BERT

The Description of the Project

Owing to only some of the entries in embedding layers updating during each batch training, it is implerative that we can just store the updating entries in the memory and outsourced the left data to other places.

The Intuitive Solution (no idea)

你可能感兴趣的:(Python,transformer,bert,深度学习)