Arch
Arch is an intelligent Layer 7 gateway built on Envoy Proxy that provides secure handling, robust observability, and seamless integration of prompts with APIs for building fast, robust, and personalized AI agents.
https://github.com/katanemo/arch?ref=aipure
Product Information
Updated:Dec 16, 2024
What is Arch
Arch is an open-source intelligent prompt gateway engineered with purpose-built LLMs to handle critical tasks related to prompt processing and API integration. Built by the core contributors of Envoy proxy, it treats prompts as nuanced user requests that require the same capabilities as traditional HTTP requests. The platform is designed to protect, observe, and personalize LLM applications like agents, assistants, and co-pilots while integrating them with backend APIs.
Key Features of Arch
Arch is an intelligent Layer 7 gateway designed for handling AI/LLM applications, built on Envoy Proxy. It provides comprehensive prompt management capabilities including jailbreak detection, API integration, LLM routing, and observability features. The platform uses purpose-built LLMs to handle critical tasks like function calling, parameter extraction, and secure prompt processing, while offering standards-based monitoring and traffic management features.
Built-in Function Calling: Engineered with purpose-built LLMs to handle fast, cost-effective API calling and parameter extraction from prompts for building agentic and RAG applications
Prompt Security: Centralizes prompt guardrails and provides automatic jailbreak attempt detection without requiring custom code implementation
Advanced Traffic Management: Manages LLM calls with smart retries, automatic failover, and resilient upstream connections for ensuring continuous availability
Enterprise-grade Observability: Implements W3C Trace Context standard for complete request tracing and provides comprehensive metrics for monitoring latency, token usage, and error rates
Use Cases of Arch
AI-Powered Weather Forecasting: Integration with weather services to provide intelligent weather forecasting through natural language processing
Insurance Agent Automation: Building automated insurance agents that can handle customer queries and process insurance-related tasks
Network Management Assistant: Creating networking co-pilots that help operators understand traffic flow and manage network operations through natural language interactions
Enterprise API Integration: Seamlessly connecting AI capabilities with existing enterprise APIs while maintaining security and observability
Pros
Built on proven Envoy Proxy technology for reliable performance
Comprehensive security features with built-in jailbreak detection
Standards-based observability making it enterprise-ready
Cons
Potential search visibility issues due to name confusion with Arch Linux
Requires Docker and specific technical prerequisites for setup
Limited documentation and community resources as a newer project
How to Use Arch
Install Prerequisites: Ensure you have Docker (v24), Docker compose (v2.29), Python (v3.10), and Poetry (v1.8.3) installed on your system. Poetry is needed for local development.
Create Python Virtual Environment: Create and activate a new Python virtual environment using: python -m venv venv && source venv/bin/activate (or venv\Scripts\activate on Windows)
Install Arch CLI: Install the Arch gateway CLI tool using pip: pip install archgw
Create Configuration File: Create a configuration file (e.g., arch_config.yaml) defining your LLM providers, prompt targets, endpoints, and other settings like system prompts and parameters
Configure LLM Providers: In the config file, set up your LLM providers (e.g., OpenAI) with appropriate access keys and model settings
Define Prompt Targets: Configure prompt targets in the config file, specifying endpoints, parameters, and descriptions for each target function
Set Up Endpoints: Define your application endpoints in the config file, including connection settings and timeouts
Initialize Client: Create an OpenAI client instance pointing to Arch gateway (e.g., base_url='http://127.0.0.1:12000/v1') in your application code
Make API Calls: Use the configured client to make API calls through Arch, which will handle routing, security, and observability
Monitor Performance: Use Arch's built-in observability features to monitor metrics, traces, and logs for your LLM interactions
Arch FAQs
Arch is an intelligent Layer 7 gateway designed to protect, observe, and personalize LLM applications with APIs. It's built on Envoy Proxy and engineered with purpose-built LLMs for secure handling, robust observability, and seamless integration of prompts with APIs.
Official Posts
Loading...Arch Video
Popular Articles
12 Days of OpenAI Content Update 2024
Dec 18, 2024
Midjourney Launches Pinterest-like Moodboards Following Patchwork Release
Dec 17, 2024
Midjourney Unveils Patchwork: A New Era in Collaborative Worldbuilding
Dec 17, 2024
Google Launches Whisk: Revolutionary AI Image Generator Remixes Three Images into One
Dec 17, 2024