AI Models & Infra

AI Application Workflow Orchestration System Based on LLM

Date / Time
2024-10-18
16:30

The rapid development of large language models (LLMs) has led to the emergence of diverse architectures and frameworks, each introducing unique AI workloads and workflows. However, challenges such as efficient orchestration, workload reuse, and reducing the deployment barriers for LLM-based applications remain critical obstacles for enterprises. In this session, we will explore how Intel’s open-source initiative, OPEA (Open Platform for Enterprise AI), addresses these challenges. A key focus will be on the GMC (GenAI Microservices Connector), a dynamic orchestration tool designed to optimize AI workloads. Join us to learn how GMC simplifies LLM application deployment and enhances workflow efficiency.