# AI Glossary > **LLM:** Large Language Models > Inteligencia Artificial Generativa > **Vector:** A series of weights assigned to words, to form a topology ([webvectors](http://vectors.nlpl.eu/explore/embeddings/en/)) > **Embeddings:** the numbers that compose the vectors. Each model has their own way. > **Token:** How LLM breaks down sentences, to create vectors. Each model has their own way. Overall 70 words = 100 tokens > **Vector Database:** A database that facilitates calculation of distance between vectors > **Transformer:** It transforms a set of word embeddings to another set of embeddings > **Embeddings** > **Corpus:** The entire set of language data to be analyzed and weighted > **Training phase** > **Parameter:** the number of artificial neurons that run thru corpus on training phase looking for patterns (Bird, 240M. ChatGPT3: 175B, chatGPT4: 1T, non-officially) > **RAG or Retrieval-Augmented Generation:** A technique for enhancing the accuracy and reliability of generative AI models with facts fetched from external sources > **Models** > **Probabilistic language models:** assigns a probability to every sentence in English in such a way that more likely sentences (in some sense) get higher probability > **Fine-tunning:** A model with weights adjusted by an annotated dataset so it works for more specific cases > **Quantized (or distilled) LLM:** A slimmed down LLM model for slower machines > **Prompt** > **LLM Usages:** human-to-human or human-to-machine communication; audio-to-text; video-to-text; text-to-text (prose-to-code included) ## Links: - [Além do hype: IA gerativa é revolucionária e perigosa](https://pca.st/kpima61x) (podcast episode, PT) - [Attention is all you need](https://proceedings.neurips.cc/paper_files/paper/2017/file/3f5ee243547dee91fbd053c1c4a845aa-Paper.pdf) (DeepMind paper) - [Speech and Language Processing: An Introduction to Natural Language Processing, Computational Linguistics, and Speech Recognition](https://web.stanford.edu/~jurafsky/slp3/ed3book.pdf) (book pdf) - [Hugging Face](https://huggingface.co/) - [Large language models, explained with a minimum of math and jargon](https://seantrott.substack.com/p/large-language-models-explained) (substack article, a gentle primer)