Mastering Multitasking: Advanced Strategies for LLM Efficiency
Dec 11, 2024
Next Generation Stage
Next Generation Stage
- Optimize LLMs on single GPUs by exploring best practices for running large language models efficiently on limited hardware
- Learn how deploying multiple specialized agents can enhance performance and task versatility
- Discuss the pros and cons of fine-tuning a single LLM for multitasking across various tasks
- Dscover how a mixture of expert models can balance specialization and efficiency in LLM deployment
- Examine additional factors that influence the quality of LLM outputs, including data quality, model architecture and external depencies
Session Type
Keynote
Content Focus
Technical