Skip to main content

Welcome to LLMTune

LLMTune is a modern, full-stack platform that provides everything you need to fine-tune, deploy, and operate AI assistants at scale. Whether you’re training custom models, comparing production-ready checkpoints, or serving inference endpoints, LLMTune brings it all together in one unified platform.

What is LLMTune?

LLMTune is a comprehensive AI platform designed to eliminate the complexity of building and managing AI infrastructure. With LLMTune, you can:
  • Fine-tune custom models without writing code or managing infrastructure
  • Compare and deploy production-ready models from our curated catalog
  • Serve inference with OpenAI-compatible APIs and built-in observability
  • Manage deployments with version control, traffic management, and rollback capabilities
  • Evaluate quality with automated scorecards and human review workflows
  • Connect datasets from multiple sources with quality scoring and PII detection

Platform Overview

LLMTune consists of six integrated products that work together seamlessly:

🎯 FineTune Studio

No-code fine-tuning for any model. Guided workflows, real-time monitoring, and support for all training methods (SFT, DPO, PPO, RLAIF, CTO) and modalities (text, image, audio, video, code, multimodal). Key Features:
  • All training methods (SFT, DPO, PPO, RLAIF, CTO)
  • All modalities (text, image, audio, video, code, multimodal, embeddings, TTS)
  • Traditional or Federated compute options
  • Single Instance or GPU Cluster deployment
  • Real-time monitoring and telemetry
  • Automatic checkpoints

πŸ“¦ LLMTune Models

Production-ready model catalog with comparison tools, deployment notes, evaluation metrics, and usage guidance. Browse, compare, and deploy the best model for your use case. Key Features:
  • Curated catalog of open-source models
  • Side-by-side model comparison
  • Deployment notes and evaluation metrics
  • Usage guidance and best practices
  • One-click deployment

πŸ”Œ LLMTune API

OpenAI-compatible API platform with inference endpoints, streaming support, usage telemetry, and built-in governance. Serve your models with enterprise-grade observability. Key Features:
  • OpenAI-compatible inference endpoints
  • Streaming support
  • Usage telemetry and monitoring
  • Workspace-scoped API keys
  • Webhooks and events
  • Built-in governance

πŸš€ LLMTune Deploy

Deployment control with version management, traffic routing, automated testing, and instant rollback. Ship new versions safely and scale confidently. Key Features:
  • Version control for models
  • Traffic management (canary, shadow, blue/green)
  • Automated testing and smoke tests
  • Instant rollback
  • Runtime observability
  • Change logs and approvals

βœ… LLMTune Evaluate

Comprehensive evaluation suite with automated scorecards, human review workflows, safety checks, and quality dashboards. Keep your models on-brand, factual, and safe. Key Features:
  • Automated scorecards
  • Human review workflows
  • Safety and quality checks
  • Regression testing
  • Quality dashboards
  • Integration with training pipeline

πŸ“Š Dataset Hub

Data intelligence platform with support for multiple sources (HuggingFace, S3, GCS, direct upload), quality scoring, PII detection, and automatic cleaning. Key Features:
  • Multiple data sources (HuggingFace, S3, GCS, direct upload)
  • Quality scoring
  • PII detection and masking
  • Automatic data cleaning
  • Version control
  • Dataset preview and validation

Platform Architecture

β”Œβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”
β”‚                    LLMTune Platform                          β”‚
β”œβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€
β”‚                                                               β”‚
β”‚  β”Œβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”  β”Œβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”  β”Œβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”     β”‚
β”‚  β”‚ FineTune     β”‚  β”‚ Models       β”‚  β”‚ Dataset Hub  β”‚     β”‚
β”‚  β”‚ Studio       β”‚  β”‚ Catalog      β”‚  β”‚               β”‚     β”‚
β”‚  β””β”€β”€β”€β”€β”€β”€β”¬β”€β”€β”€β”€β”€β”€β”€β”˜  β””β”€β”€β”€β”€β”€β”€β”¬β”€β”€β”€β”€β”€β”€β”€β”˜  β””β”€β”€β”€β”€β”€β”€β”¬β”€β”€β”€β”€β”€β”€β”€β”˜     β”‚
β”‚         β”‚                  β”‚                 β”‚             β”‚
β”‚         β””β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”Όβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”˜             β”‚
β”‚                            β”‚                                 β”‚
β”‚  β”Œβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”  β”Œβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”  β”Œβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”     β”‚
β”‚  β”‚ Deploy       β”‚  β”‚ Evaluate     β”‚  β”‚ API Platform  β”‚     β”‚
β”‚  β”‚ Control      β”‚  β”‚ Suite        β”‚  β”‚              β”‚     β”‚
β”‚  β””β”€β”€β”€β”€β”€β”€β”¬β”€β”€β”€β”€β”€β”€β”€β”˜  β””β”€β”€β”€β”€β”€β”€β”¬β”€β”€β”€β”€β”€β”€β”€β”˜  β””β”€β”€β”€β”€β”€β”€β”¬β”€β”€β”€β”€β”€β”€β”€β”˜     β”‚
β”‚         β”‚                  β”‚                 β”‚             β”‚
β”‚         β””β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”Όβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”˜             β”‚
β”‚                            β”‚                                 β”‚
β”‚                    β”Œβ”€β”€β”€β”€β”€β”€β”€β–Όβ”€β”€β”€β”€β”€β”€β”€β”                        β”‚
β”‚                    β”‚  Production   β”‚                        β”‚
β”‚                    β”‚  Deployment   β”‚                        β”‚
β”‚                    β””β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”˜                        β”‚
β””β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”˜

Key Features

No-Code Interface

Fine-tune models, deploy endpoints, and manage your AI infrastructure through an intuitive visual interface. No infrastructure setup required.

All Training Methods

Support for Supervised Fine-Tuning (SFT), Direct Preference Optimization (DPO), Proximal Policy Optimization (PPO), RL with AI Feedback (RLAIF), Controlled Tuning Optimization (CTO), and more.

All Modalities

Train models for text, image, audio, video, code, multimodal, embeddings, and TTS. Support for 17+ model families including LLaMA, Mistral, Qwen, DeepSeek, and more.

Flexible Compute

Choose between Traditional or Federated compute. Deploy on Single Instance or GPU Clusters. Switch anytime without changing your workflow.

Production-Ready

Built-in observability, version control, traffic management, and quality assurance. Everything you need to run AI in production.

Unified Platform

All products work together seamlessly. Fine-tune in Studio, deploy with Deploy, evaluate with Evaluate, and serve with API - all from one workspace.

Workflow Integration

All products work together seamlessly:
  1. Prepare Data β†’ Use Dataset Hub to connect and prepare your datasets
  2. Train Models β†’ Use FineTune Studio to train custom models
  3. Compare Models β†’ Use LLMTune Models to compare and select the best model
  4. Evaluate Quality β†’ Use LLMTune Evaluate to measure performance
  5. Deploy Safely β†’ Use LLMTune Deploy to manage production deployments
  6. Serve at Scale β†’ Use LLMTune API to serve models with observability

Unified Workspace

All products share:
  • Single Authentication - One account, one workspace
  • Unified API Keys - Same keys work across all products
  • Shared Resources - Models, datasets, and deployments are shared
  • Integrated Telemetry - Usage and metrics across all products
  • Consistent Governance - Same policies and controls everywhere

How to Use This Documentation

  1. Start with the essentials – Follow the Quickstart to create an account, upload datasets, launch a fine-tune, evaluate your model, and call the inference API.
  2. Understand the products – Read about each product to understand their capabilities and use cases.
  3. Follow guided workflows – Use the How-to guides for detailed walkthroughs on tuning, evaluation, deployment, observability, and automation.
  4. Integrate programmatically – Reference the API documentation for REST endpoints, authentication scopes, and request examples across languages.
  5. Plan ahead – Check the Roadmap and FAQ to align stakeholders on what’s shipping next and how LLMTune operates.

Need Help?