🔗 AutoAPI Overview

AutoAPI is ARPIA's automatically generated REST API framework that transforms your database tables and KnowGraph models into secure, queryable endpoints. It serves as the foundation for the Model Context Protocol (MCP) and enables seamless data exchange between your repositories and external applications—including AI agents, business intelligence tools, and custom integrations.

📍

Documentation Roadmap: Start here → How to Create AutoAPIMCP Integration


AutoAPI Screen showing endpoint configuration

🎯 What AutoAPI Does

AutoAPI eliminates the traditional API development cycle by automatically generating production-ready endpoints from your data models. Instead of writing code to expose database tables or building custom APIs, you configure what data to share through a visual interface.

In Practice:

  • Select a database table, Kube, or file storage bucket
  • Choose which columns to expose or file operations to enable
  • Define access rules and filters
  • AutoAPI generates the complete REST endpoint with authentication, validation, and documentation

The Result: A secure, documented API endpoint ready to use in seconds—not weeks.

⏱️

Time to First API: ~15 minutes from setup to production endpoint


🏗️ Architecture & Components

How AutoAPI Fits in the ARPIA Ecosystem

Database Tables / File Storage (Buckets)
    ↓
Repositories (Connected Data Sources)
    ↓
Kubes (Data Models & Business Logic)
    ↓
Variants (Access Control Views)
    ↓
AutoAPI Endpoints (REST APIs)
    ↓
MCP Collections (AI Agent Integration)

Repositories: Your connected databases (PostgreSQL, MySQL, Snowflake, etc.) and S3-compatible storage

Kubes: Semantic data models that can join tables, add calculated fields, and apply business logic. Think of them as reusable query templates.

Variants: Filtered views of Kubes that enforce row-level and column-level security. Create different variants for different user roles.

AutoAPI Endpoints: The HTTP REST APIs that expose your Kubes, variants, and file storage to external applications.

MCP Collections: Groupings of AutoAPI endpoints that AI agents can discover and query through the Model Context Protocol.


🌟 Key Capabilities

1. ⚡ Real-Time Data Integration

AutoAPI endpoints connect directly to your live databases—no data warehouses, ETL pipelines, or caching layers required.

Use Cases:

  • AI Agents: Claude, ChatGPT query your latest data during conversations
  • Dashboards: Business intelligence tools pull fresh analytics
  • Mobile Apps: Customer-facing applications access real-time inventory
  • Webhooks: External systems receive instant notifications on data changes

Example:

# This query hits your live database right now
curl -X POST https://cloud.arpia.ai/api/{collection}/{resource} \
  -H "Authorization: Bearer YOUR_TOKEN" \
  -d '{"filters": [{"field": "Status", "type": "=", "value": "ACTIVE"}]}'

2. 📤 Automatic JSON REST Endpoints

Every AutoAPI endpoint follows RESTful conventions and returns structured JSON responses with built-in pagination, filtering, and error handling.

What You Get Automatically:

  • OpenAPI/Swagger documentation for each endpoint
  • JSON Schema validation for POST/PUT operations
  • Pagination with configurable page sizes
  • Filtering with multiple operators (=, !=, >, <, LIKE, IN, BETWEEN)
  • Sorting with multi-field support
  • Grouping & Aggregation for analytics queries
  • Error messages with helpful debugging information

Example Response:

{
  "status": "success",
  "data": [
    {
      "customer_id": 1001,
      "name": "Acme Corp",
      "email": "[email protected]",
      "status": "ACTIVE",
      "total_revenue": 125000
    }
  ],
  "pagination": {
    "page": 1,
    "results": 1,
    "total_records": 347,
    "total_pages": 347
  },
  "execution_time_ms": 45
}

3. 🛠️ Endpoint Types - Choose Your Operation

AutoAPI supports two main categories of endpoints:

📊 Data Operations (Database Tables & Kubes)

GET: Query & Retrieve Data

Expose data for reading and querying. This is the most common endpoint type.

What It Does:

  • Returns structured data from tables or Kubes
  • Supports filtering, sorting, pagination, and grouping
  • Read-only access (no modifications)
  • Ideal for analytics, reporting, and AI queries

Use Cases:

  • AI agents querying customer data
  • Dashboards pulling sales metrics
  • Mobile apps displaying product catalogs
  • External systems reading order status

Example Query:

curl -X POST https://cloud.arpia.ai/api/{token}/{resource} \
  -H "Authorization: Bearer {token}" \
  -d '{
    "page": 1,
    "results": 50,
    "filters": [
      {"field": "Region", "type": "=", "value": "West"},
      {"field": "Status", "type": "=", "value": "ACTIVE"}
    ],
    "orderby": [{"field": "Revenue", "type": "DESC"}]
  }'

POST: Create New Records

Accept data for inserting new records into your database tables.

What It Does:

  • Receives JSON payloads and inserts into tables
  • Validates data against schema rules
  • Supports bulk inserts (multiple records)
  • Returns created record IDs

Use Cases:

  • Form submissions from websites
  • Mobile apps creating new orders
  • IoT devices sending sensor data
  • Third-party systems pushing leads/contacts

Example Submission:

curl -X POST https://cloud.arpia.ai/api/{token}/{resource} \
  -H "Authorization: Bearer {token}" \
  -H "Content-Type: application/json" \
  -d '{
    "records": [
      {
        "customer_name": "New Customer Inc",
        "email": "[email protected]",
        "phone": "555-0123",
        "status": "PENDING"
      }
    ]
  }'

PUT: Update Existing Records

Accept data for modifying existing records.

What It Does:

  • Updates specific records identified by ID or unique key
  • Validates field values and data types
  • Supports partial updates (only changed fields)
  • Returns number of affected records

Use Cases:

  • Status updates from external systems
  • Batch data corrections
  • Syncing changes from other databases
  • Webhook-triggered updates

Example Update:

curl -X PUT https://cloud.arpia.ai/api/{token}/{resource} \
  -H "Authorization: Bearer {token}" \
  -d '{
    "id": 1548,
    "updates": {
      "status": "ACTIVE",
      "verified_date": "2025-11-18"
    }
  }'

📁 File Operations (S3-Compatible Bucket Storage)

Configure separate endpoints for file management using ARPIA's object storage:

Bucket Operations via AutoAPI

What Each Method Does:

  • GET: List files or download specific files from bucket
  • PUT: Upload new files to bucket
  • DELETE: Remove files from bucket (when enabled)
  • POST: Special workspace submission (project-specific)

Use Cases:

  • Document management systems
  • Image/media galleries
  • Report generation and storage
  • PDF/CSV export destinations

File Upload Example:

curl -X PUT https://cloud.arpia.ai/api/bucket/?token={token}&bucket={bucket_id}&file=report.pdf \
  -H "Authorization: Bearer {token}" \
  --data-binary "@report.pdf"

⚠️

Important: File operations require creating an "Object Store" type endpoint, separate from data endpoints. See Bucket S3 API Documentation for detailed configuration.


🔐 Built-In Security Features

AutoAPI includes multiple layers of security by default:

Token-Based Authentication

  • Collection Tokens: Control access to groups of endpoints
  • Resource Tokens: Limit access to specific endpoints
  • Bearer Token Standard: Industry-standard OAuth 2.0 compatible
  • Token Scopes: Data endpoints and bucket endpoints use separate token pools

Row-Level Security (Data Endpoints)

Apply filters automatically through variants:

Example: Sales reps only see their region's data
Variant filter: region_id = current_user_region

Column-Level Security (Data Endpoints)

Control which fields are visible:

Public API: name, email, phone
Internal API: + revenue, purchase_history
Admin API: + ssn, internal_notes

Rate Limiting

  • Default: 100 requests/minute per token
  • Prevents abuse and ensures fair resource usage
  • Configurable for enterprise accounts

Audit Logging

  • Every API call is logged with timestamp, IP, and payload
  • 30-day retention (90 days for enterprise)
  • Exportable for compliance requirements

🔒

Best Practice: Always use Variants for production data endpoints. Never expose raw tables directly.


🚀 Common Use Cases

1. AI Agent Data Access (via MCP)

Scenario: Enable Claude, ChatGPT, or other AI assistants to query your databases through natural language.

Implementation:

  1. Create AutoAPI GET endpoint for customer data
  2. Add asset to MCP Collection
  3. Connect AI agent using collection token
  4. Users ask: "Show me customers who haven't ordered in 90 days"

Value: AI provides instant insights without SQL knowledge or database access.


2. Real-Time Dashboards

Scenario: Business intelligence tools need fresh data for executive dashboards.

Implementation:

  1. Create AutoAPI GET endpoints for key metrics
  2. Configure aggregation and grouping
  3. Point dashboard tools (Tableau, Power BI) to endpoints
  4. Data refreshes automatically without ETL pipelines

Value: Eliminate data warehouse delays, see metrics in real-time.


3. Mobile App Backend

Scenario: Mobile applications need to read product catalogs and submit orders.

Implementation:

  1. Create GET endpoint for product catalog with filters
  2. Create POST endpoint for order submission with validation
  3. Implement row-level security by customer account
  4. Mobile app calls endpoints directly

Value: No custom backend development, automatic scaling, built-in security.


4. Document Management System

Scenario: Teams need centralized file storage with API access.

Implementation:

  1. Create Object Store endpoint for bucket operations
  2. Configure PUT for uploads, GET for downloads
  3. Set appropriate file size limits and allowed types
  4. Integrate with applications using bucket tokens

Value: S3-compatible storage without managing infrastructure.


5. Partner Integrations

Scenario: Business partners need limited access to order status and inventory.

Implementation:

  1. Create variant with partner-specific filtering
  2. Generate AutoAPI GET endpoint from that variant
  3. Provide partner with dedicated collection token
  4. Partner queries only their own orders

Value: Secure data sharing without custom integration code.


🎓 AutoAPI vs. Traditional API Development

AspectTraditional APIsAutoAPI
Development TimeWeeks to months15-30 minutes
Code RequiredHundreds of linesZero
DocumentationManual maintenanceAuto-generated
AuthenticationCustom implementationBuilt-in tokens
ValidationWrite validatorsSchema-based
SecurityCode security rulesVisual configuration
UpdatesRedeploy applicationUpdate configuration
TestingUnit tests requiredBuilt-in console
VersioningManual managementAutomatic
PerformanceVaries by implementationOptimized queries

📊 Performance & Limits

Query Performance

  • Target: <1 second for most queries
  • Maximum Timeout: 30 seconds
  • Optimization: Automatic query plan analysis

Data Limits

  • Page Size: 1-1,000 records (recommended: 100-500)
  • Payload Size: 5MB maximum per response
  • File Uploads: Up to 100MB per file (bucket operations)

Rate Limits

  • Standard: 100 requests/minute per token
  • Burst: Up to 150 requests in 60 seconds
  • Enterprise: Configurable higher limits

🔄 Integration with MCP

AutoAPI serves as the data layer for the Model Context Protocol (MCP), which enables AI agents to discover and query your data.

The Connection:

AI Agent (Claude/ChatGPT)
    ↓ (asks for data)
MCP Protocol Layer
    ↓ (translates to HTTP)
AutoAPI Endpoints
    ↓ (queries database/storage)
Your Data Sources

Decision Guide:

Use CaseRecommended Approach
Custom applicationsAutoAPI directly
AI assistantsAutoAPI + MCP
File managementBucket endpoints
Mixed operationsMultiple endpoint types

🚦 Getting Started

Quick Start Checklist

1. Prerequisites (5 minutes)

  • Active ARPIA account
  • Database connected as Repository OR Bucket created
  • At least one table with data OR files to manage

2. Create Your First Endpoint (10 minutes)

  • Navigate to AutoAPI section
  • Click + to create new endpoint
  • Select endpoint type (GET for data, Object Store for files)
  • Choose your Repository/table OR Bucket
  • Select columns to expose OR file operations to enable
  • Save endpoint

3. Apply Security (5 minutes)

  • For data endpoints: Apply variant with row-level filters
  • For all endpoints: Review access permissions
  • Generate appropriate tokens

4. Test Your Endpoint (5 minutes)

  • Open AutoAPI Console
  • Execute sample query
  • Verify expected results
  • Try filters and sorting (data) OR file operations (bucket)

5. Connect External Application (15 minutes)

  • Copy endpoint URL and tokens
  • Configure your application/AI agent
  • Send test request
  • Verify data flows correctly

Total Time: ~40 minutes from zero to production API


Next Steps

Step-by-Step Guides:

AI Agent Integration:

API Reference:


💡 Best Practices

Endpoint Design

  1. Choose the right type: Data operations for tables, Object Store for files
  2. Name clearly: "Customer_Orders_Active" not "API_123"
  3. One purpose per endpoint: Don't mix operations

Performance Optimization

  1. Index filtered columns in your database
  2. Use reasonable page sizes (50-100 for interactive, 500 for batch)
  3. Apply filters to reduce result sets
  4. Monitor query times in logs regularly

Security Configuration

  1. Start restrictive, add access as needed
  2. Use variants for all production data endpoints
  3. Separate tokens for data vs. file operations
  4. Rotate tokens quarterly
  5. Review logs weekly for unusual patterns

Maintainability

  1. Document purpose in description field
  2. Version variants when changing security rules
  3. Test in console after changes
  4. Notify consumers before breaking changes

🆘 Common Questions

Q: Do I need to write any code?
A: No. AutoAPI is entirely configuration-based through the ARPIA interface.

Q: Can I mix data and file operations in one endpoint?
A: No. Create separate endpoints - one for data (GET/POST/PUT) and another for files (Object Store).

Q: What's the difference between AutoAPI tokens and bucket tokens?
A: They're managed separately for security. Data endpoints use AutoAPI tokens, file operations use bucket-specific tokens.

Q: What happens if my database schema changes?
A: AutoAPI automatically reflects schema changes. You may need to update column selections if columns are removed.

Q: Can I test my endpoints before production use?
A: Yes! Use the built-in AutoAPI Console for testing queries, filters, and responses.

Q: How do I handle high traffic volumes?
A: Contact ARPIA for enterprise accounts with higher rate limits, connection pooling, and dedicated infrastructure.

Q: Which endpoint type should I use for my use case?
A: Use data endpoints (GET/POST/PUT) for database operations. Use Object Store endpoints for file management. When in doubt, start with GET for reading data.


📚 Additional Resources

Documentation Hub:

Support:

Developer Tools: