Skip to content
Bot Nirvana Members
  • Home
  • Playbook
  • Use Cases
  • Tools
  • AI Papers
  • Login
Bot Nirvana Members
  • Account
  • AI & Automation Tools
  • AI Advisory Board (AAB)
  • AI Ideas / Use Cases
  • AI Papers
  • AI Papers
  • Cancel Payment
  • Change Password
  • Edit Profile
  • Forgot Password
  • Guest
  • Home
  • Latest Tools
  • Login
  • Login
  • Logout
  • Members
  • Password Reset
  • Register
  • Register
  • Session Recordings
  • Stories
  • Thank You
  • Trends
  • Use Cases
  • User

Introduction

3
  • Welcome
  • Purpose and Principles
  • How to Use the Playbook

AI Fundamentals

7
  • The origins and evolution of AI
  • Key Concepts and Terminology
  • Traditional AI (ML) vs Foundational models (Gen AI)
  • AI vs. ML vs. Deep Learning vs. Generative AI
  • Introduction To Generative AI
  • Synthetic Data
  • Computer Vision

Generative AI

13
  • LLMs and Foundational Models
  • Foundation models 101
  • Architecture for LLM Applications (Emerging)
  • Prompt Engineering
  • Tokens and Optimization
  • Retrieval Augmented Generation (RAG)
  • Four Types Of RAG (from Microsoft)
  • Mixture of Experts (MoE)
  • Model Inference
  • Small Language Models
  • Reinforcement Learning from Human Feedback (RLHF)
  • Large Vision Models (LVMs)
  • Types of RAG

AI Agents

8
  • Types of Agents
  • Agent uses and examples
  • AI Agents
  • Agentic workflow
  • AI Agent Memory
  • Intelligent Automation with AI Agents
  • AI agent challenges
  • Agentic RAG

Applying AI

5
  • Continuum for Applied AI
  • 11 ways to apply Gen AI
  • AI-led automation
  • Chatbots
  • Microsoft 365 Co-Pilot

Use Cases

2
  • 15 Best Gen AI Use Cases
  • Use Case Grid

Technologies and Tools

7
  • Top 3 AI Tools by Category
  • LLM app dev tools
  • Current LLM Landscape
  • ChatGPT Enterprise
  • OpenAI Playground
  • Vector Databases
  • Synthetic data

AI Strategies

4
  • 5-Steps to Identifying Winning AI Opportunities
  • How To Identify & Prioritize AI Opportunities
  • Create a compelling business case
  • Mastering Data Management for Business: A Comprehensive Guide

Risks & Security

5
  • Risks of AI
  • 4 Top AI Risks Right Now
  • AI – Security Risk Checklist
  • AI Security Frameworks
  • Hallucinations and other risks

AI Safety

4
  • Explainable AI
  • AI Alignment
  • Ethical Considerations And Strategies
  • Handling Sensitive Information with AI

AI Governance

6
  • AI Center of Excellence
  • The regulatory landscape for AI
  • Current compliance situation
  • AI CoE checklist
  • Data governance: ensuring data integrity and accessibility
  • Crafting AI Policies for Transparency and Bias Mitigation

AI Implementation

1
  • Key Factors to Consider When Selecting a LLM

Resources and Further Learning

2
  • Free courses
  • Best AI Resources

Prompt Engineering

4
  • Chain of Thought Prompting
  • Effective Guide To Prompting
  • One-Shot Prompting
  • Self-Consistency Prompting

Thought Leadership

1
  • AI Spanning

AI Hardware

1
  • GPUs in AI

Robotics

1
  • Robotics and AI

LLM Evaluation

3
  • Practical LLM Evaluation Techniques
  • Methods and Frameworks for LLM Evaluation
  • LLM Evaluation Explained
  • Home
  • Playbook
  • Generative AI
  • Architecture for LLM Applications (Emerging)
View Categories

Architecture for LLM Applications (Emerging)

1 min read

The emerging Large Language Model (LLM) App Stack is a multi-layered structure where each component plays a crucial role in ensuring the effective application of LLMs.

A large function of these frameworks is orchestrating all the various components: LLM providers, Embedding models, vector stores, document loaders, and other tools which we will dive into below.

Here is a view of the emerging App stack (Source A16z)

Let’s now slice through each layer of the stack and see what they are their role, and some tools for the layer:

Data Pipelines #

  • The backbone of data ingestion and transformation, connecting various data sources including connectors to ingest contextual data wherever it may reside.
  • Essential for preparing and channeling data to downstream components, thus kickstarting the entire application process.
  • Tools: Databricks, Airflow, Unstructured

Embedding Models #

  • This component transforms contextual data into a mathematical format, usually vectors.
  • Critical for making sense of complex data, enabling easier storage and more efficient computations.
  • Tools: OpenAI, Cohere, Hugging Face

Vector Databases #

  • A specialized database designed to store and manage vector data generated by the embedding model.
  • Allows for faster and more efficient querying, essential for LLM applications that require real-time data retrieval like chatbots.
  • Tools: Pinecone, Weaviate, ChromaDB, pgvector

Playground #

  • An environment where you can iterate and test your AI prompts.
  • Vital for fine-tuning and testing LLM prompts before they are embedded in the app, ensuring optimal performance.
  • Tools: OpenAI, nat.dev, Humanloop

Orchestration #

  • This layer coordinates the various components and workflows within the application.
  • They abstract the details (e.g. prompt chaining; interfacing with external APIs etc.) and maintain memory across multiple LLM calls.
  • Tools: Langchain, LlamaIndex, Flowise, Langflow

APIs/Plugins #

  • Interfaces and extensions that allow the LLM application to interact with external tools and services.
  • Enhances functionality and interoperability, enabling the app to tap into additional resources and services.
  • Tools: Serp, Wolfram, Zapier

LLM Cache #

  • A temporary storage area that keeps frequently accessed data readily available.
  • Improves application speed and reduces latency, enhancing the user experience.
  • Tools: Redis, SQLite, GPTCache

Logging/LLM Ops #

  • A monitoring and logging component that keeps track of application performance and system health.
  • Provides essential oversight for system management, crucial for identifying and resolving issues proactively.
  • Tools: Weights & Biases, MLflow, PromptLayer, Helicone

Validation #

  • Frameworks that enable more effective control of the LLM app outputs.
  • Ensures the reliability and integrity of the LLM application, acting as a quality check and taking corrective actions.
  • Tools: Guardrails, Rebuff, Microsoft Guidance, LMQL

App Hosting #

  • The platform where the LLM application is deployed and made accessible to end-users.
  • Necessary for scaling the application and managing user access, providing the final piece of the application infrastructure.
  • Tools: Vercel, Steamship, Streamlit, Modal

This is an emerging stack and we will see more changes as we progress. We will look to keep this updated as we see big changes.

Featured
What do you think?
Updated on November 5, 2023
Foundation models 101Prompt Engineering

Leave a Reply Cancel reply

Your email address will not be published. Required fields are marked *

Table of Contents
  • Data Pipelines
  • Embedding Models
  • Vector Databases
  • Playground
  • Orchestration
  • APIs/Plugins
  • LLM Cache
  • Logging/LLM Ops
  • Validation
  • App Hosting

Copyright © 2025 Bot Nirvana Members

Scroll to Top