AI apps
Inference for AI models
Overview
AI apps integrate AI and ML capabilities with data-rich backend systems, deeply embedded in organizational processes, creating a seamless experience where intelligent services and data flow bidirectionally between various system components. By combining AI/ML models and services with conventional application logic and data, AI apps deliver intelligent, adaptive solutions that optimize data-driven dynamic resource allocation, real-time natural language processing for automated customer support, and contextual recommendations for personalized user experiences across different application stack layers.
What is an AI app?
AI apps integrate AI and ML capabilities with conventional software systems. These applications typically involve two main areas of AI functionality. The first is training, which involves developing and fine-tuning AI models or training ML models using application-specific datasets. The second is inference, which encompasses the deployment and execution of AI/ML models to make predictions or decisions based on app context-specific input data.
From an architectural perspective, AI apps treat AI/ML components as external services integrated with conventional application logic and data. AI apps require robust APIs and data management for effective interaction between AI and conventional components. They achieve scalability by independently scaling training pipelines, inference services, and application logic, optimizing resource use and performance across diverse conditions.
Key properties of AI apps
AI applications rely on seamlessly integrating AI/ML services with conventional application components, distributed processing capabilities, and adaptive resource management to operate effectively across various computational environments. They scale by making their inference, conventional application, and integration each independently elastic.
Scalable inference
Dynamic resource management
Data integration and preprocessing
Adaptive model selection
Continuous learning pipeline
Automated AI/MLOps
Akka components
- The client sends a request to the endpoint component (API).
- The endpoint component forwards the command to the App Data Entity, which processes it and emits events.
- Events go to a streaming component, which reliably forwards them to Vector DB Service.
- Vector DB Service stores and retrieves data for AI processing.
- RAG endpoint component retrieves context from Vector DB for AI / LLM.
- AI / LLM uses the context to generate an informed response.
How Akka enables AI apps
Akka provides a robust framework for developing and operating distributed applications that seamlessly integrate conventional processing systems with AI and ML services, enabling scalable and resilient AI-powered applications.