---
title: OpenAI: gpt-oss-120b — Open LLM | ModelsLab
description: Access OpenAI: gpt-oss-120b API for 117B MoE reasoning on single H100 GPU. Try high-performance open-weight LLM inference now.
url: https://modelslab.com/openai-gpt-oss-120b
canonical: https://modelslab.com/openai-gpt-oss-120b
type: website
component: Seo/ModelPage
generated_at: 2026-05-05T20:10:31.256319Z
---

Available now on ModelsLab · Language Model

OpenAI: Gpt-oss-120b
OpenAI gpt-oss-120b Power
---

[Try OpenAI: Gpt-oss-120b](/models/open_router/openai-gpt-oss-120b) [API Documentation](https://docs.modelslab.com)

Deploy gpt-oss-120b Efficiently
---

MoE Architecture

### 117B Params 5.1B Active

Activates 5.1B parameters per token in 36-layer MoE with 128 experts.

Single GPU Fit

### Runs on H100

Fits in 80GB VRAM via MXFP4 quantization for efficient production inference.

Agentic Reasoning

### Tool Use Native

Supports function calling, browsing, code execution with configurable effort levels.

Examples

See what OpenAI: Gpt-oss-120b can create
---

Copy any prompt below and try it yourself in the [playground](/models/open_router/openai-gpt-oss-120b).

Code Optimization

“Analyze this Python function for efficiency: def fibonacci(n): if n <= 1: return n else: return fibonacci(n-1) + fibonacci(n-2). Suggest memoized version with benchmarks.”

Market Trend Report

“Summarize quarterly sales data trends from CSV: Q1:100k, Q2:120k, Q3:110k, Q4:150k. Forecast Q1 next year using linear regression.”

Tech Architecture

“Design scalable microservices for e-commerce: user auth, inventory, payments. Include Docker, Kubernetes, API gateways.”

Algorithm Explanation

“Explain A\* pathfinding algorithm step-by-step with pseudocode. Compare to Dijkstra on grid graph example.”

For Developers

A few lines of code.
gpt-oss-120b. One API call.
---

ModelsLab handles the infrastructure: fast inference, auto-scaling, and a developer-friendly API. No GPU management needed.

- **Serverless:** scales to zero, scales to millions
- **Pay per token,** no minimums
- **Python and JavaScript SDKs,** plus REST API

[API Documentation ](https://docs.modelslab.com)

PythonJavaScriptcURL

Copy

```
<code>import requests

response = requests.post(
    "https://modelslab.com/api/v7/llm/chat/completions",
    json={
  "key": "YOUR_API_KEY",
  "prompt": "",
  "model_id": ""
}
)
print(response.json())</code>
```

FAQ

Common questions about OpenAI: Gpt-oss-120b
---

[Read the docs ](https://docs.modelslab.com)

### What is OpenAI: gpt-oss-120b model?

OpenAI: gpt-oss-120b is 117B-parameter MoE LLM with 5.1B active params per token. Matches o4-mini on reasoning benchmarks. Runs on single H100 GPU.

### How to use OpenAI: gpt-oss-120b API?

Access via LLM endpoint with Apache 2.0 license. Supports 131k context, reasoning tokens. Configure low/medium/high effort.

### Is OpenAI: gpt-oss-120b fine-tunable?

Yes, fine-tune on H100 node. Customize for specific tasks like coding or agents. Apache 2.0 allows commercial use.

### OpenAI: gpt-oss-120b alternative to closed models?

Open-weight alternative to o3/o4-mini with near-parity performance. Strong on agentic tasks, tool use. No copyleft restrictions.

### What hardware for openai gpt oss 120b api?

Single 80GB GPU like H100 or MI300X. Uses MXFP4 for MoE weights. 131k context window supported.

### OpenAI: gpt-oss-120b LLM capabilities?

Text-only input/output. Full chain-of-thought access. Native function calling, browsing, Python execution.

Ready to create?
---

Start generating with OpenAI: Gpt-oss-120b on ModelsLab.

[Try OpenAI: Gpt-oss-120b](/models/open_router/openai-gpt-oss-120b) [API Documentation](https://docs.modelslab.com)

---

*This markdown version is optimized for AI agents and LLMs.*

**Links:**
- [Website](https://modelslab.com)
- [API Documentation](https://docs.modelslab.com)
- [Blog](https://modelslab.com/blog)

---
*Generated by ModelsLab - 2026-05-06*