台大 李弘毅 ML/DL 筆記 === ###### tags: `Deep Learning` `Machine Learning` `PyTorch` Transformer (講解含Self attention、Positional Encoding、Transformer) --- Transformer知名應用 - BERT(用Unsupervised的Transformer訓練而成) Transformer是什麼? A: Seq2seq model with "Self-attention" source form: https://www.youtube.com/watch?v=ugWDIIOHtPA
{"metaMigratedAt":"2023-06-15T12:44:38.460Z","metaMigratedFrom":"Content","title":"台大 李弘毅 ML/DL 筆記","breaks":true,"contributors":"[{\"id\":\"e5c57de4-5b31-4833-8080-a7fd4f82ae7e\",\"add\":321,\"del\":6}]"}
Expand menu