Token-first security
Scope access per token, toggle features, and rotate secrets instantly without shipping new keys to clients.
Orchestrate chat, completions, embeddings, and tool-calling flows with an OpenAI-compatible surface. Control quota, monitor usage, and rely on Kent Wynn's managed infrastructure for consistent low-latency responses.
Daily check-in bonuses
Log in every day to add +1,000 free tokens to your Kent Wynn account.
NewGitHub launch bonus
Register with GitHub for the first time and unlock +100,000 tokens instantly.
curl -X POST \
https://api.kentwynn.com/v1/chat/completions \
-H "Content-Type: application/json" \
-H "X-AI-Token: YOUR_TOKEN" \
-d '{
"model": "kentwynn/reasoning",
"messages": [
{ "role": "system", "content": "You are a helpful assistant." },
{ "role": "user", "content": "Summarise the Kent Wynn AI platform." }
],
"stream": false
}'Sign in to the Kent Wynn console, generate an API token, and copy it securely—tokens are only revealed once.
Use `kentwynn/reasoning` for chat-style workloads or `kentwynn/embedding` for search pipelines.
Send a POST request to the `/v1/*` endpoint with your token in the `X-AI-Token` header. Responses mirror OpenAI schemas.
Track quotas and daily burn in the console or automate checks with the admin REST endpoints.
Scope access per token, toggle features, and rotate secrets instantly without shipping new keys to clients.
Chat, completions, embeddings, and responses expose an OpenAI-compatible contract for effortless integration.
Add +1,000 tokens to your account every 24 hours with the console check-in bonus—perfect for ongoing experiments.
Fully managed inference from Kent Wynn infrastructure with a polished, reliable API surface and predictable performance.
Each button calls the /demo namespace on api.kentwynn.com with the sample payloads shown. Responses come directly from the hosted engines.
List available demo model aliases and engine IDs.
curl -X 'GET' \
'https://api.kentwynn.com/demo/v1/models' \
-H 'accept: application/json'All endpoints accept X-AI-Token and return JSON responses. Streaming is available for chat completions.
Enumerate hosted models with their public aliases.
Stream or fetch assistant replies using chat-style prompts.
Generate classic text completions with temperature and stop controls.
Produce vector embeddings optimised for semantic search and clustering.
Unified endpoint that auto-falls back between chat and completion styles.
Access is authenticated with scoped tokens and enforced quotas. Tokens can be disabled instantly, inactive users are blocked from invoking APIs, and every request is logged for audit trails.
{
"choices": [
{
"message": {
"role": "assistant",
"content": "Kent Wynn AI keeps your LLM stack private and predictable."
},
"finish_reason": "stop"
}
],
"usage": {
"prompt_tokens": 18,
"completion_tokens": 24,
"total_tokens": 42
}
}Kent Wynn is compatible with the OpenAI ecosystem. Point existing SDKs at the Kent Wynn base URL or integrate with LangChain to orchestrate tools and agents.
Use `langchain-openai` with a custom `base_url` to plug Kent Wynn models into existing chains, agents, and RAG pipelines.
View docs →Point the official OpenAI Node/Python SDKs at our domain by setting `baseURL` and keep the rest of your integration unchanged.
View docs →Prefer raw HTTP? Demo endpoints mirror OpenAI wire formats so you can test with curl or request libraries before coding.
View docs →Configure the LangChain OpenAI adapter with the Kent Wynn base URL to send prompts through the hosted stack.
from langchain_openai import ChatOpenAI
llm = ChatOpenAI(
base_url="https://api.kentwynn.com/v1",
api_key="YOUR_TOKEN",
model="kentwynn/reasoning",
extra_body={"tool_choice": "auto"},
)
response = llm.invoke(
"Give me two marketing angles for a private Kent Wynn deployment."
)
print(response.content)Keep everything on your hardware while offering a polished developer experience. Set quotas, monitor usage, and deliver private LLM features in minutes.