Introduction
Konko API overview, how it works and how to get started!
Overview
Konko API is a fully managed API designed to help application developers:
- Select the right LLM(s) for their application
- Prototype with various open-source and proprietary LLMs
- Access Fine Tuning for open-source LLMs to get industry-leading performance at a fraction of the cost
- Setup low-cost production APIs according to security, privacy, throughput, latency SLAs without infrastructure set-up or administration using Konko AI's SOC 2 compliant, multi-cloud infrastructure
The Konko API is fully compatible with the OpenAI API protocol
This allows the Konko API to be a drop-in replacement for applications currently using the OpenAI API but seeking open-source alternatives.
How it works
To help you select the best LLM for your use-case: we provide free access to a variety of leading open and closed source models via our chat completion endpoint and completion endpoint. (more details on our supported LLMs here)
To request access to our free testing instance or custom fine-tuned models: email us at [email protected]. Please provide a brief description of your use case and we'll suggest the optimal model for your use case plus the requirements for additional fine-tuning.
Once you're ready to move to production, we provide a secure, low-cost API endpoint that meets your security, privacy, throughput, latency, and reliability requirements.
Moving to production is simple:
Once you've completed these two steps, the Konko AI team will setup access to your production-ready API.
Available Models
We recommend this curated list of industry-leading proprietary and open source LLMs for a variety of use cases. We offer fine-tuning services, hosting, or proxy access to the following models.
Chat Completion Endpoint
General Purpose
meta-llama/llama-2-70b-chat
meta-llama/llama-2-13b-chat
nousresearch/nous-hermes-llama2-13b
nousresearch/nous-hermes-llama-2-7b
mistralai/mistral-7b-instruct-v0.1
open-orca/mistral-7b-openorca
Expanded Context Window
konko/llama-2-7b-32k-instruct
Coding
codellama/codellama-34b-instruct
codellama/codellama-13b-instruct
codellama/codellama-7b-instruct
Completions Endpoint
General Purpose
meta-llama/llama-2-70b
meta-llama/llama-2-13b
mistralai/mistral-7b-v0.1
Expanded Context Window
konko/llama-2-7b-32k-instruct
Coding - General
codellama/codellama-34b
phind/phind-codellama-34b-v2
Coding - Python
codellama/codellama-34b-python
phind/phind-codellama-34b-python-v1
Coding - SQL
numbersstation/nsql-llama-2-7b
We're continuously evaluating and adding new models. If you'd like to see a new model, please request it by messaging [email protected] or joining our discord.
How to get started
First: get your API key. You can request it by messaging [email protected] or joining our discord.
Once you have your API key you can query the Konko API either:
- Directly via HTTP in any popular language (learn how with our interactive documentation)
- Via Konko AI's Python SDK (learn how with our Setup Access & SDK guide
Updated 8 months ago