Exclusive Content:

Haiper steps out of stealth mode, secures $13.8 million seed funding for video-generative AI

Haiper Emerges from Stealth Mode with $13.8 Million Seed...

Running Your ML Notebook on Databricks: A Step-by-Step Guide

A Step-by-Step Guide to Hosting Machine Learning Notebooks in...

“Revealing Weak Infosec Practices that Open the Door for Cyber Criminals in Your Organization” • The Register

Warning: Stolen ChatGPT Credentials a Hot Commodity on the...

Optimize OpenAI GPT-OSS Models with Hugging Face Libraries on Amazon SageMaker AI

Unleashing Advanced AI Capabilities: Fine-Tuning OpenAI’s GPT-OSS Models on AWS SageMaker

Introduction to OpenAI’s GPT-OSS Models

Explore the groundbreaking release of OpenAI’s GPT-OSS models—gpt-oss-20b and gpt-oss-120b—now available on AWS through Amazon SageMaker. Discover their unique features, including Mixture-of-Experts architecture, high reasoning performance, and multilingual capabilities.

Key Features and Specifications of GPT-OSS Models

A detailed overview of the model specifications and capabilities, including total parameters, context length, and performance optimizations designed for specialized tasks in coding and scientific analysis.

Deploying GPT-OSS on Amazon SageMaker

Learn how to deploy GPT-OSS using Amazon SageMaker JumpStart and Bedrock APIs. Understand the flexibility provided for integrating these models into production-grade AI workflows.

Fine-Tuning with Hugging Face Libraries

Dive into the process of fine-tuning GPT-OSS models on specific datasets using the Hugging Face TRL and Accelerate libraries to tailor the models to specialized use cases.

Setting Up a Managed Environment for Fine-Tuning

Step-by-step guide on configuring your environment in SageMaker Studio for seamless model fine-tuning, including instance selection and GitHub integration.

Choosing the Right Dataset for Multilingual Fine-Tuning

Find out how to curate the right dataset for supervised fine-tuning, enhancing multilingual reasoning, and maintaining logical consistency across languages.

Experimentation Tracking with MLflow

Explore how SageMaker’s managed MLflow capabilities streamline experiment tracking and governance, enabling efficient model revisions and comparisons.

Fine-Tuning Workflow and SageMaker Estimators

An overview of the fine-tuning process using SageMaker training jobs, including how to adapt recipes, select instance types, and leverage distributed training.

Conclusion: Empowering Enterprise AI Solutions

Summarize the benefits of fine-tuning GPT-OSS models on AWS SageMaker, and encourage experimentation with the shared resources for real-world AI applications.

About the Authors

Meet the authors, Pranav Murthy and Sumedha Swamy, and learn about their expertise in Generative AI and product management at AWS, shaping the future of machine learning.

Unleashing the Power of OpenAI’s GPT-OSS Models on Amazon SageMaker

On August 5, 2025, OpenAI unveiled its latest innovations in the world of AI: the GPT-OSS models, including the impressive gpt-oss-20b and gpt-oss-120b. These text-only Transformer models leverage a state-of-the-art Mixture-of-Experts (MoE) architecture, which significantly enhances reasoning capabilities while keeping compute costs in check. Available through Amazon Web Services (AWS) via Amazon SageMaker AI and Amazon Bedrock, these models are poised to revolutionize complex tasks ranging from coding to scientific analysis and mathematical reasoning.

Architecture and Specifications

The GPT-OSS models bridge the gap between advanced learning and practical application. Here’s a snapshot of their specifications:

Model Layers Total Parameters Active Parameters Per Token Total Experts Active Experts Per Token Context Length
openai/gpt-oss-120b 36 117 billion 5.1 billion 128 4 128,000
openai/gpt-oss-20b 24 21 billion 3.6 billion 32 4 128,000

These models support a context length of 128,000 tokens, adjustable reasoning levels (low, medium, high), and structured outputs to facilitate agentic-AI workflows. Additionally, they boast enhanced safety features thanks to adversarial fine-tuning and robust training methods that minimize risks of misuse.

Deployment Made Easy with AWS

Developers can easily deploy the GPT-OSS models using Amazon SageMaker JumpStart or through Amazon Bedrock APIs. This flexibility allows for quick integration into enterprise-grade AI workflows, enabling customization with domain-specific data using open-source tools from the Hugging Face ecosystem.

Fine-Tuning for Specific Use Cases

Fine-tuning is the key to transforming these robust models into specialized experts. By adjusting the model’s weights using a smaller dataset tailored to specific tasks, you can achieve more accurate and context-aware outputs. This method not only enhances reliability but mitigates hallucinations, ensuring that model outputs are grounded in real-world relevance.

The Fine-Tuning Process

The journey of fine-tuning GPT-OSS models involves a few critical steps:

  1. Environment Setup: Utilize AWS resources and ensure correct IAM roles are configured.

  2. Dataset Curation: Choose a task-specific dataset. For example, the HuggingFaceH4/Multilingual-Thinking dataset is well-suited for fine-tuning across various languages.

  3. Training with Hugging Face and SageMaker: By leveraging libraries such as Hugging Face TRL for fine-tuning and AWS’s managed infrastructure for job execution, you can streamline the entire process.

Using Advanced Techniques: MXFP4 and PEFT

The integration of MXFP4 (Microscaling FP4) and Parameter-Efficient Fine-Tuning (PEFT) methods like LoRA offers significant advantages. MXFP4 reduces memory and compute requirements while maintaining accuracy, and PEFT allows for the adaptation of large models by focusing on a small subset of additional parameters.

Business Outcomes of Fine-Tuning GPT-OSS

The demand for advanced AI tools that excel in multilingual reasoning is ever-growing, driven by the needs of global enterprises. Fine-tuning GPT-OSS models addresses these needs, enabling complex reasoning across diverse linguistic contexts. Testing with a multilingual dataset helps establish the model’s ability to maintain reasoning coherence across languages, thereby setting a solid foundation for broader domain-specific applications.

Conclusion

OpenAI’s GPT-OSS models represent a significant leap forward in AI capabilities. By coupling these models with AWS’s infrastructure, developers can fine-tune them for specific business needs, unlocking advanced reasoning capabilities that can seamlessly integrate into existing workflows.

For those eager to take the plunge into this advanced territory, the accompanying GitHub repository is an invaluable resource to kickstart your journey with fine-tuning GPT-OSS models on SageMaker.

Explore, innovate, and transform your AI capabilities with GPT-OSS today!


About the Authors

  • Pranav Murthy: Senior Generative AI Data Scientist at AWS, specializing in deep learning and machine learning.
  • Sumedha Swamy: Senior Product Manager at AWS, leading initiatives within Amazon SageMaker, focusing on integrated development environments for ML.

Explore further, and seize the power of AI with OpenAI’s GPT-OSS!

Latest

Advancements in Large Model Inference Container: New Features and Performance Improvements

Enhancing Performance and Reducing Costs in LLM Deployments with...

I asked ChatGPT if the remarkable surge in Lloyds share price has peaked, and here’s what it said…

Assessing the Future of Lloyds Banking: Insights and Reflections Why...

Cows Dominate Robots on Day One: The Tech Revolution Transforming Dairy Farming in Rural Australia

Revolutionizing Dairy Farming: Automated Milking Systems Transform the Lives...

AI Receptionist for Answering Services

Certainly! Here’s a suitable heading for the section you...

Don't miss

Haiper steps out of stealth mode, secures $13.8 million seed funding for video-generative AI

Haiper Emerges from Stealth Mode with $13.8 Million Seed...

Running Your ML Notebook on Databricks: A Step-by-Step Guide

A Step-by-Step Guide to Hosting Machine Learning Notebooks in...

VOXI UK Launches First AI Chatbot to Support Customers

VOXI Launches AI Chatbot to Revolutionize Customer Services in...

Investing in digital infrastructure key to realizing generative AI’s potential for driving economic growth | articles

Challenges Hindering the Widescale Deployment of Generative AI: Legal,...

Advancements in Large Model Inference Container: New Features and Performance Improvements

Enhancing Performance and Reducing Costs in LLM Deployments with AWS Updates Navigating the Challenges of Token Growth in Modern LLMs LMCache Support: Transforming Long-Context Inference Performance Benchmarks...

Reinforcement Fine-Tuning for Amazon Nova: Educating AI via Feedback

Unlocking Domain-Specific Capabilities: A Guide to Reinforcement Fine-Tuning for Amazon Nova Models Bridging the Gap Between General-Purpose AI and Business Needs A New Paradigm: Learning by...

Creating a Personal Productivity Assistant Using GLM-5

From Idea to Reality: Building a Personal Productivity Agent in Just Five Minutes with GLM-5 AI A Revolutionary Approach to Application Development This headline captures the...