## XENTROPY - End to end consulting on <font color="BB86FC">AI</font> and <font color="BB86FC">large scale compute infrastructure</font> Chan, Ka Hei Founder _chankahei@xentropy.co_ ex-NVIDIA data scientist / solution architect --- ## Background End to end AI consulting - Identify opportunities in applying AI technologies - AI awared project management and system architecture - AI Research and development - Code optimisation for GPU hardwares for development and production ---- Trusted by - Hong Kong Science and Technology Park - Hong Kong Hospital Authority - AIA Hong Kong - Manulife Hong Kong --- ## More Background System and infrastructure administration - On premise managed jupyter notebook - Batch job scheduling - Web access portal - Networking optimisation - User management ---- Trusted by - University of Hong Kong - Hong Kong Polytechnique University - Macau University of Science and Technology --- ## DEMO word autocompletion for clinical note - <font color="BB86FC">3 million pieces of clinical notes</font> are sampled and fed to customize a GPT-2 model - trained for 4 days with an <font color="03DAC6">8-GPU NVIDIA DGX</font> system for a single experiment - customised inference logic to achieve performance target - full stack end to end development Click [HERE](http://containers.xentropy.co:8082) to try it out! --- ## Industry Trend #### Centralised development of large AI model across multiple applications __OpenAI__ - ChatGPT __Google__ - PaLM __META__ - MultiRay ---- ### Emergency of large multimodality transformer models Transformer model are empirically proven to be effective in most data modalities, or even combinations of data modalities. - Tabular Medical Data [Hi-BHERT](https://arxiv.org/pdf/2106.11360.pdf) - Natural Language [ChatGPT](https://openai.com/blog/chatgpt/) - Computer Vision + Natural language [DALL-E 2](https://openai.com/dall-e-2/) - Reinforcement Learning [Decision Transformer](https://arxiv.org/pdf/2106.01345) ---- ### Major Advantages - Cost Amortization across many teams - Simpler development and operations - Faster research to production: Single-point acceleration - Improved hardware utilisation - Take advantage of deep neural network's scaling property ---- ### Cost Amortization - large upfront cost of training large scale deep neural network makes most but only the highly important application economically unviable - risk to return ratio - large models serves as a catalyst where activation cost of AI project is vastly reduced ---- ### Simplified development and operation - maintaining _ONE_ model is significant work - streamlined validation, feedback and retraining - elastic hardware allocation - dependency management - ... etc - maintaining _ONE HUNDRED_ model is impossible ---- ### Faster go to market time - reuse __knowledge__, __toolchain__, and __data__ acquired in previous projects ---- ### Hardware utilisation - sharing hardware resources always introduce overheads, which can be significant - divert data scientists away from actually producing high value models - centralised large model eliminates most of the overhead ---- ### Scaling characteristics of DNN - empirically observed under most circumstances deep neural network performance scales with compute capacity and data volume in inverse log relationship - more team sharing a model => more data and more compute capacity for the model => better performance --- ## Collaboration - Private consulting session - Joint application development - Join our community --- ## Discussion
{"metaMigratedAt":"2023-06-17T19:21:39.230Z","metaMigratedFrom":"YAML","title":"Untitled","breaks":true,"slideOptions":"{\"transition\":\"slide\"}","contributors":"[{\"id\":\"d11aef06-1770-4326-b857-a9420d6be67b\",\"add\":3735,\"del\":80}]"}
    317 views