Get up to 15x faster response from OpenAI GPT API with Model Gateway
OpenAI GPT APIs are slow and unreliable. We route your request to the reliable and fastest available data center.
The fastest GPT response
Model you ❤️ but up to 15x faster
We monitor OpenAI Platform and all Azure OpenAI data centers and route your request to the fastest available Azure region that is reliable at a given moment. Enjoy your favorite OpenAI GPT models, but much faster.
- Fastest possible inference
- Get up to 15x more output tokens per second with active routing compared to using your current static endpoints.
- Easy integration
- Keep using your favorite AI libraries. We are compatible with all major existing ones. Example code is available on GitHub.
- Secure by default
- Security is our top priority. We use the latest security standards to keep communication safe.
- Privacy guaranteed
- All your data belongs to you. We do not store any of your prompts or responses.
Super-simple integration
No additional dependencies, no complex setup. Just a simple configuration.
from openai import AzureOpenAI
MODELGW_API_KEY = "sk-..."
client = AzureOpenAI(
api_key=MODELGW_API_KEY,
api_version="2023-05-15",
azure_endpoint="https://azure-openai.api.modelgw.com", # set endpoint
)
completion = client.chat.completions.create(
messages=[{"role": "user", "content": "Hello there!"}],
model="auto", # set your model to "auto"
)
print(completion)
import OpenAI from 'openai';
const modelgwApiKey = 'sk-...';
const openai = new OpenAI({
apiKey: modelgwApiKey,
baseURL: 'https://azure-openai.api.modelgw.com/openai/deployments/auto',
defaultQuery: { 'api-version': '2023-05-15' },
defaultHeaders: { 'api-key': modelgwApiKey },
});
async function main() {
const result = await openai.chat.completions.create({
model: 'auto',
messages: [{ role: 'user', content: 'Hello there!' }],
});
console.log(JSON.stringify(result, null, 4));
}
main().catch((err) => {
console.error(err);
process.exit(1);
});
String modelgwApiKey = "sk-...";
OpenAIClient client = new OpenAIClientBuilder()
.credential(new KeyCredential(modelgwApiKey))
.endpoint("https://azure-openai.api.modelgw.com")
.buildClient();
List<ChatRequestMessage> chatMessages = new ArrayList<>();
chatMessages.add(new ChatRequestUserMessage("Hello there!"));
ChatCompletions chatCompletions = client.getChatCompletions("auto",
new ChatCompletionsOptions(chatMessages));
for (ChatChoice choice : chatCompletions.getChoices()) {
ChatResponseMessage message = choice.getMessage();
System.out.println(message.getContent());
}
$modelgwApiKey = 'sk-...';
$client = OpenAI::factory()
->withBaseUri('https://azure-openai.api.modelgw.com/openai/deployments/auto')
->withHttpHeader('api-key', $modelgwApiKey)
->withQueryParam('api-version', '2023-05-15')
->make();
$response = $client->chat()->create([
'model' => 'auto',
'messages' => [
['role' => 'user', 'content' => 'Hello there!'],
],
]);
foreach ($response->choices as $result) {
echo $result->message->content;
}
curl "https://azure-openai.api.modelgw.com/openai/deployments/auto/chat/completions?api-version=2023-05-15" \
-H "Content-Type: application/json" \
-H "api-key: $MODELGW_API_KEY" \
-d "{ \"messages\": [{ \"role\": \"user\", \"content\": \"Hello there\!\" } ], \"model\": \"auto\" }"
Ready for faster inference?
Start 14-day free trial!
Pricing
Monthly plans that scale with your business
Choose an affordable plan that’s packed with the best features for engaging your audience, creating customer loyalty, and driving sales.
Free
The essentials for faster and reliable OpenAI GPT API on Azure platform. Say goodbye to API errors and timeouts.
€0
Sign up
- 1 Gateway
- Up to 5 Azure OpenAI regions
- Routing to the fastest region
- Up to 5x faster inference
- Automatic failover
- Up to 10,000 requests/month
Standard
⚡️ Fastest inference
An affordable plan for the fastest possible GPT API using model deployments distributed across all regions.
€49/gateway
Start free trial
- Unlimited Gateways
- Unlimited Azure OpenAI regions
- Routing to the fastest region
- Up to 15x faster inference
- Automatic failover
- Up to 100,000 requests/month
- Basic analytics Coming soon
Enterprise
⚡️ Fastest inference
Our large scale GPT usage plan with the fastest possible GPT API using model deployments distributed across all regions.
Custom
Start free trial
- Unlimited Gateways
- Unlimited Azure OpenAI regions
- Routing to the fastest region
- Up to 15x faster inference
- Automatic failover
- Unlimited requests/month
- Advanced analytics Coming soon
- 24-hour support response time
Frequently asked questions
Get in touch
We are here to help and answer any questions you might have. We look forward to hearing from you.
Headquarters
- moc.wgledom@olleh
Support
- moc.wgledom@troppus