Exclusive Content:

Haiper steps out of stealth mode, secures $13.8 million seed funding for video-generative AI

Haiper Emerges from Stealth Mode with $13.8 Million Seed...

“Revealing Weak Infosec Practices that Open the Door for Cyber Criminals in Your Organization” • The Register

Warning: Stolen ChatGPT Credentials a Hot Commodity on the...

VOXI UK Launches First AI Chatbot to Support Customers

VOXI Launches AI Chatbot to Revolutionize Customer Services in...

Optimize OpenAI GPT-OSS Models with Hugging Face Libraries on Amazon SageMaker AI

Unleashing Advanced AI Capabilities: Fine-Tuning OpenAI’s GPT-OSS Models on AWS SageMaker

Introduction to OpenAI’s GPT-OSS Models

Explore the groundbreaking release of OpenAI’s GPT-OSS models—gpt-oss-20b and gpt-oss-120b—now available on AWS through Amazon SageMaker. Discover their unique features, including Mixture-of-Experts architecture, high reasoning performance, and multilingual capabilities.

Key Features and Specifications of GPT-OSS Models

A detailed overview of the model specifications and capabilities, including total parameters, context length, and performance optimizations designed for specialized tasks in coding and scientific analysis.

Deploying GPT-OSS on Amazon SageMaker

Learn how to deploy GPT-OSS using Amazon SageMaker JumpStart and Bedrock APIs. Understand the flexibility provided for integrating these models into production-grade AI workflows.

Fine-Tuning with Hugging Face Libraries

Dive into the process of fine-tuning GPT-OSS models on specific datasets using the Hugging Face TRL and Accelerate libraries to tailor the models to specialized use cases.

Setting Up a Managed Environment for Fine-Tuning

Step-by-step guide on configuring your environment in SageMaker Studio for seamless model fine-tuning, including instance selection and GitHub integration.

Choosing the Right Dataset for Multilingual Fine-Tuning

Find out how to curate the right dataset for supervised fine-tuning, enhancing multilingual reasoning, and maintaining logical consistency across languages.

Experimentation Tracking with MLflow

Explore how SageMaker’s managed MLflow capabilities streamline experiment tracking and governance, enabling efficient model revisions and comparisons.

Fine-Tuning Workflow and SageMaker Estimators

An overview of the fine-tuning process using SageMaker training jobs, including how to adapt recipes, select instance types, and leverage distributed training.

Conclusion: Empowering Enterprise AI Solutions

Summarize the benefits of fine-tuning GPT-OSS models on AWS SageMaker, and encourage experimentation with the shared resources for real-world AI applications.

About the Authors

Meet the authors, Pranav Murthy and Sumedha Swamy, and learn about their expertise in Generative AI and product management at AWS, shaping the future of machine learning.

Unleashing the Power of OpenAI’s GPT-OSS Models on Amazon SageMaker

On August 5, 2025, OpenAI unveiled its latest innovations in the world of AI: the GPT-OSS models, including the impressive gpt-oss-20b and gpt-oss-120b. These text-only Transformer models leverage a state-of-the-art Mixture-of-Experts (MoE) architecture, which significantly enhances reasoning capabilities while keeping compute costs in check. Available through Amazon Web Services (AWS) via Amazon SageMaker AI and Amazon Bedrock, these models are poised to revolutionize complex tasks ranging from coding to scientific analysis and mathematical reasoning.

Architecture and Specifications

The GPT-OSS models bridge the gap between advanced learning and practical application. Here’s a snapshot of their specifications:

Model Layers Total Parameters Active Parameters Per Token Total Experts Active Experts Per Token Context Length
openai/gpt-oss-120b 36 117 billion 5.1 billion 128 4 128,000
openai/gpt-oss-20b 24 21 billion 3.6 billion 32 4 128,000

These models support a context length of 128,000 tokens, adjustable reasoning levels (low, medium, high), and structured outputs to facilitate agentic-AI workflows. Additionally, they boast enhanced safety features thanks to adversarial fine-tuning and robust training methods that minimize risks of misuse.

Deployment Made Easy with AWS

Developers can easily deploy the GPT-OSS models using Amazon SageMaker JumpStart or through Amazon Bedrock APIs. This flexibility allows for quick integration into enterprise-grade AI workflows, enabling customization with domain-specific data using open-source tools from the Hugging Face ecosystem.

Fine-Tuning for Specific Use Cases

Fine-tuning is the key to transforming these robust models into specialized experts. By adjusting the model’s weights using a smaller dataset tailored to specific tasks, you can achieve more accurate and context-aware outputs. This method not only enhances reliability but mitigates hallucinations, ensuring that model outputs are grounded in real-world relevance.

The Fine-Tuning Process

The journey of fine-tuning GPT-OSS models involves a few critical steps:

  1. Environment Setup: Utilize AWS resources and ensure correct IAM roles are configured.

  2. Dataset Curation: Choose a task-specific dataset. For example, the HuggingFaceH4/Multilingual-Thinking dataset is well-suited for fine-tuning across various languages.

  3. Training with Hugging Face and SageMaker: By leveraging libraries such as Hugging Face TRL for fine-tuning and AWS’s managed infrastructure for job execution, you can streamline the entire process.

Using Advanced Techniques: MXFP4 and PEFT

The integration of MXFP4 (Microscaling FP4) and Parameter-Efficient Fine-Tuning (PEFT) methods like LoRA offers significant advantages. MXFP4 reduces memory and compute requirements while maintaining accuracy, and PEFT allows for the adaptation of large models by focusing on a small subset of additional parameters.

Business Outcomes of Fine-Tuning GPT-OSS

The demand for advanced AI tools that excel in multilingual reasoning is ever-growing, driven by the needs of global enterprises. Fine-tuning GPT-OSS models addresses these needs, enabling complex reasoning across diverse linguistic contexts. Testing with a multilingual dataset helps establish the model’s ability to maintain reasoning coherence across languages, thereby setting a solid foundation for broader domain-specific applications.

Conclusion

OpenAI’s GPT-OSS models represent a significant leap forward in AI capabilities. By coupling these models with AWS’s infrastructure, developers can fine-tune them for specific business needs, unlocking advanced reasoning capabilities that can seamlessly integrate into existing workflows.

For those eager to take the plunge into this advanced territory, the accompanying GitHub repository is an invaluable resource to kickstart your journey with fine-tuning GPT-OSS models on SageMaker.

Explore, innovate, and transform your AI capabilities with GPT-OSS today!


About the Authors

  • Pranav Murthy: Senior Generative AI Data Scientist at AWS, specializing in deep learning and machine learning.
  • Sumedha Swamy: Senior Product Manager at AWS, leading initiatives within Amazon SageMaker, focusing on integrated development environments for ML.

Explore further, and seize the power of AI with OpenAI’s GPT-OSS!

Latest

Integrating Responsible AI in Prioritizing Generative AI Projects

Prioritizing Generative AI Projects: Incorporating Responsible AI Practices Responsible AI...

Robots Shine at Canton Fair, Highlighting Innovation and Smart Technology

Innovations in Robotics Shine at the 138th Canton Fair:...

Clippy Makes a Comeback: Microsoft Revitalizes Iconic Assistant with AI Features in 2025 | AI News Update

Clippy's Comeback: Merging Nostalgia with Cutting-Edge AI in Microsoft's...

Is Generative AI Prompting Gartner to Reevaluate Its Research Subscription Model?

Analyst Downgrades and AI Disruption: A Closer Look at...

Don't miss

Haiper steps out of stealth mode, secures $13.8 million seed funding for video-generative AI

Haiper Emerges from Stealth Mode with $13.8 Million Seed...

VOXI UK Launches First AI Chatbot to Support Customers

VOXI Launches AI Chatbot to Revolutionize Customer Services in...

Investing in digital infrastructure key to realizing generative AI’s potential for driving economic growth | articles

Challenges Hindering the Widescale Deployment of Generative AI: Legal,...

Microsoft launches new AI tool to assist finance teams with generative tasks

Microsoft Launches AI Copilot for Finance Teams in Microsoft...

Integrating Responsible AI in Prioritizing Generative AI Projects

Prioritizing Generative AI Projects: Incorporating Responsible AI Practices Responsible AI Overview Generative AI Prioritization Methodology Example Scenario: Comparing Generative AI Projects First Pass Prioritization Risk Assessment Second Pass Prioritization Conclusion About the...

Developing an Intelligent AI Cost Management System for Amazon Bedrock –...

Advanced Cost Management Strategies for Amazon Bedrock Overview of Proactive Cost Management Solutions Enhancing Traceability with Invocation-Level Tagging Improved API Input Structure Validation and Tagging Mechanisms Logging and Analysis...

Creating a Multi-Agent Voice Assistant with Amazon Nova Sonic and Amazon...

Harnessing Amazon Nova Sonic: Revolutionizing Voice Conversations with Multi-Agent Architecture Introduction to Amazon Nova Sonic Explore how Amazon Nova Sonic facilitates natural, human-like speech conversations for...