台大 李弘毅 ML/DL 筆記
===
###### tags: `Deep Learning` `Machine Learning` `PyTorch`
Transformer (講解含Self attention、Positional Encoding、Transformer)
---
Transformer知名應用 - BERT(用Unsupervised的Transformer訓練而成)
Transformer是什麼? A: Seq2seq model with "Self-attention"
source form: https://www.youtube.com/watch?v=ugWDIIOHtPA
{"metaMigratedAt":"2023-06-15T12:44:38.460Z","metaMigratedFrom":"Content","title":"台大 李弘毅 ML/DL 筆記","breaks":true,"contributors":"[{\"id\":\"e5c57de4-5b31-4833-8080-a7fd4f82ae7e\",\"add\":321,\"del\":6}]"}