Home > News list > Tech >> Industry dynamics

OpenAI fully opens its o1 series model APIs, empowering developers to build advanced AI applications

Industry dynamics 2024-12-18 08:23:39 Source:

OpenAI fully opens its o1 series model APIs, empowering developers to build advanced AI applicationsOn December 18th, the ninth day of OpenAI's 12-day holiday release event, the company officially announced the full release of its state-of-the-art o1 series models via application programming interfaces (APIs) to third-party developers. This marks a significant step forward for developers in building advanced AI applications, providing a seamless way to integrate OpenAI's cutting-edge technology into existing enterprise-level applications or consumer-facing workflows

OpenAI fully opens its o1 series model APIs, empowering developers to build advanced AI applications

On December 18th, the ninth day of OpenAI's 12-day holiday release event, the company officially announced the full release of its state-of-the-art o1 series models via application programming interfaces (APIs) to third-party developers. This marks a significant step forward for developers in building advanced AI applications, providing a seamless way to integrate OpenAI's cutting-edge technology into existing enterprise-level applications or consumer-facing workflows.

The o1 series models, including o1 and o1 mini, first debuted in September 2024 as the inaugural members of OpenAI's "new family of models." Representing a significant leap forward from the GPT series of large language models (LLMs), they introduce novel "reasoning" capabilities. While exhibiting slightly longer response times compared to traditional LLMs, the o1 series models self-validate their answers, ensuring accuracy and significantly reducing "hallucinations." OpenAI previously stated that the o1 model could handle highly complex tasks, even those at the doctoral level, a claim substantiated by real-world application feedback.

Prior to the official API release, developers accessed the o1 models via a preview program, building applications such as "doctoral tutors" or "lab assistants." The now-released production version (o1-2024-12-17) boasts significant performance improvements, reduced latency, and new features, greatly simplifying integration and application in real-world scenarios. Approximately two and a half weeks ago, OpenAI made the o1 models available to consumers through ChatGPT Plus and ChatGPT Pro plans, adding functionality for analyzing and responding to uploaded images and files.

The new o1 model excels in complex, multi-step reasoning tasks. Compared to the o1-preview version, this release shows significant improvements in accuracy, efficiency, and flexibility. OpenAI's benchmark tests reveal groundbreaking advancements across coding, mathematics, and visual reasoning. For example, on the SWE-bench Verified test (evaluating the model's ability to solve real-world software problems), the o1 score increased from 41.3 to 48.9; in the AIME mathematics test, the score jumped from 42 to 79.2. These improvements make the o1 model ideal for optimizing customer support, improving logistics management efficiency, or solving complex analytical problems.

To further enhance developer efficiency and application flexibility, OpenAI has introduced several new features:

OpenAI fully opens its o1 series model APIs, empowering developers to build advanced AI applications

  • Structured Output: Allows generating responses conforming to custom formats (e.g., JSON schemas), ensuring consistency when interacting with external systems and facilitating data processing and integration.
  • Function Calling: Simplifies connecting the model to APIs and databases, enabling developers to more easily access and utilize external resources.
  • Visual Reasoning: Enables the model to process visual inputs, expanding its applications in manufacturing, scientific research, and programming, opening up broader possibilities for developers.

Furthermore, developers can fine-tune the o1 model using the new `reasoning_effort` parameter. This parameter balances performance and response time, controlling the computational resources allocated to a task, allowing developers to adjust based on specific needs.

Beyond the opening of the o1 series model APIs, OpenAI also announced significant updates to its Realtime API, designed to support low-latency, natural-sounding voice interactions crucial for voice assistants, real-time translation tools, and virtual tutors.

These Realtime API updates include:

  • WebRTC Integration: Provides direct support for developing voice applications, including audio streaming, noise suppression, and network congestion control. This allows developers to easily implement real-time functionality even under unreliable network conditions, ensuring application stability and reliability.
  • Price Reductions: OpenAI has significantly lowered the cost of the Realtime API. For example, GPT-4 audio prices have been reduced by 60%, with the cost of 1 million input tokens dropping to $40 and output tokens to $80; cached audio input costs are down 87.5% at $2.50 per 1 million input tokens; and GPT-4o mini, a smaller, more cost-effective model, costs $10 per 1 million input tokens and $20 for output tokens. Text token pricing for GPT-4o mini has also been significantly reduced, starting at $0.60 for input tokens and $2.40 for output tokens.
  • Increased Control: OpenAI provides developers with greater control over the Realtime API, such as concurrent out-of-band responses, enabling background tasks (like content moderation) to run without impacting user experience; and context customization, allowing developers to tailor input context based on conversation content and precisely control the triggering of voice responses for a more accurate and fluid interaction. These improvements will enable developers to build more efficient and personalized voice applications.

In addition to these two major updates, OpenAI introduced another significant featurepreference tuning. This is a new approach to customizing models based on user and developer preferences. Unlike traditional supervised fine-tuning, which relies on precise input-output pairs, preference tuning uses pairwise comparisons to guide the model to generate responses that better align with user preferences. This method is particularly effective for subjective tasks such as summarization, creative writing, or scenarios where tone and style are critical.

Early testing with partner RogoAI showed encouraging results. RogoAI developed a smart assistant for financial analysts, and testing revealed that preference tuning, when dealing with complex, out-of-distribution queries, significantly outperformed traditional fine-tuning methods, improving task accuracy by over 5%. This feature is currently available for GPT-4o-2024-08-06 and GPT-4o-mini-2024-07-18, with plans to expand to more models early next year.

To further simplify model integration, OpenAI is expanding its official SDK offerings with beta releases of Go and Java SDKs. These new SDKs join the existing Python, Node.js, and .NET libraries, providing developers with support across a wider range of programming environments for easier interaction with OpenAI models.

  • Go SDK: Particularly well-suited for building scalable backend systems, offering high performance and flexible development capabilities.
  • Java SDK: Designed for enterprise-level applications, leveraging strong typing and a mature ecosystem for complex and stability-critical projects.

Through these updates, OpenAI provides developers with a richer set of tools to build advanced, highly customizable AI applications. Whether it's the enhanced capabilities of the o1 model in complex reasoning tasks, the optimized Realtime API, the introduction of preference tuning, or the release of new SDKs, OpenAI's latest offerings aim to deliver greater performance and cost-effectiveness, helping businesses continuously expand the boundaries of AI technology application.

Tag: OpenAI fully opens its o1 series model APIs empowering


Disclaimer: The content of this article is sourced from the internet. The copyright of the text, images, and other materials belongs to the original author. The platform reprints the materials for the purpose of conveying more information. The content of the article is for reference and learning only, and should not be used for commercial purposes. If it infringes on your legitimate rights and interests, please contact us promptly and we will handle it as soon as possible! We respect copyright and are committed to protecting it. Thank you for sharing.

AdminSo

http://www.adminso.com

Copyright @ 2007~2024 All Rights Reserved.

Powered By AdminSo

Open your phone and scan the QR code on it to open the mobile version


Scan WeChat QR code

Follow us for more hot news

AdminSo Technical Support