# LLM Examples - Docs: - https://scicomp.aalto.fi/aalto/generative-ai-tools/ - https://scicomp.aalto.fi/triton/apps/llms/ - https://scicomp.aalto.fi/triton/discipline/machinelearning/ - Example repo: - https://github.com/AaltoSciComp/llm-examples #### Demo Plan (40 min) - General introduction to AI tools in Aalto (5 min) - https://scicomp.aalto.fi/aalto/generative-ai-tools/ - Focus on local llm usage on triton (15 min) - Where to find models - Ready to use envs on Triton and how to create conda env (refer to conda session of day 2) - Resources you need to request to run models - partition: (tricky sometimes) - GPU vRAM and how this is related to #parameters - Some models use operators that rely on newer CUDA features which are only supported by newer GPUs - New frameworks need higher GPU compute capability - system mem - num of cpus - Frameworks to run models - Where to find docs and example repos - Run examples (15 min) - Generation via transformers - Batch inference via vllm (if time allows) - Wrap up(2 min): where to get help and what kind of help we provide - Creating ones own conda env or figuring out the best practice with LLM frameworks can be a bit tricky. So don’t do troubleshooting alone—drop by our daily Garage session and ask. we’re here to help you choose the right tools and get the most out of them. :::info ## LLMs on Triton - Docs: - https://scicomp.aalto.fi/triton/apps/llms/ - Example repo: - https://github.com/AaltoSciComp/llm-examples :::