πŸ‘‹ Jan

πŸ‘‹Jan API Documentation

OpenAI-compatible API for local and server deployments

Local API

llama.cpp

Run Jan locally with complete privacy.

http://localhost:1337/v1 Privacy-first β€’ GGUF models β€’ CPU/GPU
View Documentation β†’

Jan Server

vLLM

Self-hostable server for high-throughput inference.

http://your-server:8000/v1 Open source β€’ Auto-scaling β€’ Multi-GPU
View Documentation β†’

Quick Start

1 Choose deployment type
2 Start your server
3 Make API requests