A New Era in AI: Insights from the OpenAI Developer Conference

iconArtificial Intelligence

In San Francisco, a city known for tech innovation, the OpenAI Developer Conference was a major event for the AI world. This conference brought together experts, developers, and technology leaders. Leading the event were Sam Altman, the CEO of OpenAI known for pushing boundaries in AI research, and Satya Nadella, the CEO of Microsoft, whose company has been a key player in advancing AI technology.

OpenAI, under Altman's leadership, has been at the forefront of AI development, sparking curiosity and anticipation in the tech community about its next moves. We at SciForce have been closely monitoring OpenAI's trajectory, intrigued by the next steps of their advancements in the broader tech landscape.

The conference was much more than just showing off new technology; it was a place for sharing big ideas about the future of AI. One of the main attractions was the unveiling of GPT-4 Turbo, a new development by OpenAI. The event was crucial for looking at how AI is growing and how it might change technology as we know it.

Unveiling GPT-4 Turbo

GPT-4 Turbo sets a new benchmark with its ability to handle up to 128,000 context tokens. This technical enhancement marks a significant leap from previous models, allowing the AI to process and retain information over longer conversations or data sets.

Reflecting on this enhancement, Sam Altman noted, "GPT-4 supported up to 8K and in some cases up to 32K context length, but we know that isn't enough for many of you and what you want to do. GPT-4 Turbo, supports up to 128,000 context tokens. That's 300 pages of a standard book, 16 times longer than our 8k context."

GPT-4 Turbo enhances accuracy over long contexts, offering more precise AI responses for complex interactions. Key features include JSON Mode for valid responses, improved function calling with multi-function capabilities, and reproducible outputs using a seed parameter, enhancing control and consistency in AI interactions.

Advanced AI Modalities

At the OpenAI Developer Conference, new text-to-speech and image recognition technologies were revealed, marking major AI advancements.

  • Text-to-Speech The conference showcased an advanced text-to-speech model capable of producing highly natural audio from text, with six preset voices. This enhancement enhances the app user experience, broadening its use in language learning and voice assistance.

  • Image Recognition They presented an advanced image recognition technology integrated with DALL-E 3. GPT-4 Turbo's new capabilities include processing images for generating captions and analyses. Used by applications like Be My Eyes, it assists visually impaired users. DALL-E 3's creative potential was also highlighted through Coke's campaign, allowing customers to create unique Diwali cards.

Democratic Pricing Strategy

At the OpenAI Developer Conference, GPT-4 Turbo's pricing was significantly reduced to 1 cent per 1,000 prompt tokens and 3 cents per 1,000 completion tokens. This major reduction aims to make advanced AI more accessible and affordable, encouraging broader use and innovation across various industries, from small-scale developers to large corporations.

Simultaneously, OpenAI is committed to enhancing GPT-4 Turbo's performance, with planned upgrades to increase its speed and efficiency. Additionally, the cost of GPT-3.5 Turbo 16K has been reduced, reinforcing OpenAI's dedication to offering high-performing and cost-effective AI solutions.

OpenAI and Microsoft: A Strategic Partnership

Starting with a simple request for Azure credits, the partnership between OpenAI and Microsoft has flourished into a deep collaboration, significantly shaping the future of AI. As Microsoft's CEO Satya Nadella says, “The shape of Azure is drastically changing and is changing rapidly in support of these models that you're building. Our job, number one, is to build the best system so that you can build the best models”. This shows Microsoft's strong commitment to upgrading its systems to keep up with the growing needs of AI technology.

AI-Driven Azure Enhancements

Microsoft's involvement has been essential in transforming Azure infrastructure to meet AI's demands:

  • Enhanced Computational Power Microsoft upgraded Azure's processing power, crucial for efficiently handling complex AI tasks.

  • Refined Data Storage Solutions Adaptations to Azure's data storage accommodate the vast data needs of AI models, ensuring robust and secure management.

  • Improved Network Capabilities Enhancements of Azure allow for faster, more reliable data transmission, crucial for AI model training and deployment.

These enhancements are designed to efficiently handle the unique challenges posed by AI model training and deployment, ensuring optimal performance and scalability of AI technologies. This focused approach by Microsoft has been crucial in providing the necessary foundation for OpenAI's ambitious AI projects.

Customization and Accessibility in AI

The OpenAI Developer Conference highlighted two key advancements: customizable General Purpose Transformers (GPTs) and the Assistants API. These innovations mark a significant shift in AI technology, offering tailored solutions for diverse industries and simplifying AI integration into applications. This section delves into how these developments, from educational tools to creative applications, are making AI more adaptable and user-friendly. Customizable GPTs – A New Era in AI The OpenAI Developer Conference marked a notable advancement in AI with the introduction of customizable General Purpose Transformers (GPTs). These models, designed to cater to specific industry needs, significantly enhance AI's utility and effectiveness.

As explained at the conference, “You can in effect program a GPT with language just by talking to it. It's easy to customize the behavior so that it fits what you want. This makes building them very accessible and it gives agency to everyone. We're going to show you what GPTs are, how to use them, how to build them, and then we're going to talk about how they'll be distributed and discovered.

This innovation paves the way for practical, industry-specific applications:

Code.org's Educational GPT: Tailored for educational purposes, particularly in teaching computer science, this GPT aids educators in developing engaging lesson plans and conveying complex topics more effectively to students. Canvas Design-Focused GPT: This model exemplifies AI's utility in creative industries, enabling users to start and manage design projects simply through natural language commands, streamlining the creative process.

These examples demonstrate the potential of customizable GPTs to revolutionize various sectors by providing industry-specific AI solutions, showcasing a notable leap in AI's functionality and adaptability.

Simplifying AI Integration with the Assistants API

The Assistants API, unveiled at the OpenAI Developer Conference, marks a significant step in AI development, offering streamlined integration of complex functionalities into applications, as showcased at the event: “The Assistants API includes persistent threads, so they don't have to figure out how to deal with long conversation history, built-in retrieval, code interpreter, a working Python interpreter in a sandbox environment, and of course the improved function calling”

Let’s take a closer look at Assistants API key features:

  • Persistent Threads Enables maintenance of extended conversation histories in applications, simplifying the creation of continuous user interactions without manual management of long context lengths. Ideal for apps needing ongoing dialogue.

  • Built-in Retrieval Enhances the AI's knowledge base by enabling access to external documents or databases, expanding its information reach beyond initial training data.

  • Code Interpreter This feature allows the AI to write and execute code in real-time, streamlining complex tasks and data processing in applications, and enhancing efficiency in development.

Shaping the Future with AI

The OpenAI Developer Conference outlined a future where AI is central to our lives. Innovations like GPT-4 Turbo and the Assistants API pave the way for AI's integration into healthcare, education, and business, enhancing efficiency and decision-making.

The future of AI, as discussed at the conference, isn't just about technological advancement; it's about how these advancements will empower individuals and transform industries:

  • Healthcare – more accurate diagnoses and treatments.
  • Education – personalized learning revolutionizing teaching and learning methods.
  • Business – advanced analytics leading to more informed and strategic decision-making.

The Assistants API democratizes AI, making it accessible for wider use and driving innovation. This shift promises an AI-enhanced future, boosting efficiency and creativity in everyday tasks and industries.

As Sam Altman highlighted, the current advancements are just the beginning of a journey toward an AI-integrated society. The conference left a clear message: we're entering an era where AI not only enhances our abilities but also partners us in shaping a smarter world.

SciForce’s Adoption of OpenAI Breakthroughs

SciForce integrates OpenAI's advancements to transform key areas like healthcare data management and online education. Our focus is on applying these cutting-edge AI solutions to solve real-world challenges across various industries.

Enhancing Jackalope Project with

Jackalope uses AI to convert medical data into the OMOP Common Data Model efficiently. Using the GatorTron model, simplifies and improves the accuracy of parsing complex medical expressions, enhancing data conversion in healthcare. In this project, SciForce employs GPT models for semantic search and decision-making.

The main challenge was creating flexible AI responses for unique case scenarios with varied output structures, further complicated by the lack of sufficient labeled data. The project required custom solutions for diverse cases with variable output structures, addressing these complexities effectively even with limited labeled data.

To overcome these challenges SciForce utilized GPT models to provide context-specific, flexible responses. By using these models, we've automated the rating of model performance, significantly streamlining the process.

Result Obtaining Automatization in Jackalope

In the Jackalope project, we've furthered our use of GPT models by introducing 'Result Obtaining Automatization' to enhance our performance evaluation process. We implemented a tailored rating system to precisely evaluate model performance, ensuring each model's output aligns with the project's specific requirements.

This involves creating detailed prompts for GPT models, to automate performance rating generation, based on predefined rules. This enables the models to independently evaluate results after each experiment, aligning with our specific evaluation criteria. Question-Answer System for Educational Project

SciForce's upcoming project involves creating an AI-driven question-answering system for online learning. This system will interact with various educational materials, including PDFs, slides, and video transcripts, to provide an interactive platform for both students and educators:

  • For Students: The system offers instant explanations on complex topics, improving their learning experience.
  • For Educators: Acts as a digital assistant, handling frequent queries and reducing workload.

At SciForce, we're all about making online learning cooler and more interactive with AI – it's all about keeping everyone involved and making the learning experience fun and responsive.


At SciForce, the advancements revealed at the OpenAI Developer Conference align with our dedication to AI development. Inspired by these innovations, we're committed to leveraging them to drive forward-thinking solutions and foster a future enhanced by AI. These breakthroughs reinforce our belief in AI's transformative power, and we're excited to continue our journey at the forefront of AI's evolution.