
The Build Hour session introduced OpenAI’s prompt caching feature, a mechanism that reuses computation for repeated prompt prefixes to cut latency and reduce API costs. Erica explained that once a request exceeds 1,024 tokens, OpenAI begins caching 128‑token blocks, automatically handling text, image, and audio inputs without code changes. Developers can extend cache lifetimes to 24 hours and influence routing with an optional prompt cache key, ensuring similar requests land on the same engine for higher hit rates. Key data points highlighted include a 50‑90% discount on cached tokens across model families and up to a 99% discount for speech‑to‑speech caching. In a benchmark of 2,300 prompts ranging from 1,024 to 200,000 tokens, cached requests showed a 67% faster time‑to‑first‑token for the longest inputs, while short prompts saw modest latency gains. A live demo with an AI styling assistant demonstrated cost reductions from $0.35 to $0.21 per batch when leveraging implicit caching and a prompt cache key, while latency remained comparable for 2,000‑token prompts. The session also covered the technical underpinnings: OpenAI hashes the first 256 tokens and checks for matching 128‑token chunks, reusing attention matrix outputs (floating‑point numbers) rather than recomputing them. Developers are advised to keep prompt prefixes deterministic—avoiding timestamps or stray whitespace—and to employ context engineering, truncation, summarization, and appropriate endpoint selection to maximize cache hits. For businesses, adopting prompt caching can translate into substantial cost savings at scale and more predictable response times for heavy‑weight workloads, especially in multimodal applications like image batch processing or long conversational threads. By structuring prompts for cacheability and using the prompt cache key, teams can achieve higher throughput without sacrificing model intelligence.

OpenAI used its latest developer briefing to outline a refreshed deep‑research agenda for ChatGPT, positioning the next wave of upgrades as a strategic response to growing enterprise demand for more capable, trustworthy conversational AI. The company highlighted three technical pillars:...

The video follows a product manager who integrates the Codex app into daily workflows to bridge the gap between non‑coding responsibilities and the technical codebase. By asking Codex to explain unfamiliar UI elements and to perform routine tasks, the PM...

We build the tools. You build the future. Start building with Codex. https://openai.com/codex/

The video showcases how the Codex app’s worktree feature lets developers run multiple tasks in parallel, illustrated by adding a drag‑and‑drop sorting capability for pinned tasks. By delegating work to separate worktrees, the user can continue other development activities without...

The video showcases how the Original Tamale Company, a multigenerational family operation, leverages ChatGPT to transition from informal backyard sales to a more formalized business model. By prompting the AI, the owner quickly derives wholesale pricing with a 60% margin, generates...

The video follows the owner of Reno Salvage, an 86‑year‑old metal‑recycling yard, as he confronts the inefficiencies of a paper‑based workflow and experiments with ChatGPT to digitize daily operations. He explains that his grandfather built the business on manual records,...

The video showcases a family farm leveraging ChatGPT to modernize record‑keeping, reporting, and decision‑making across generations. By digitizing a 1971 handwritten crop ledger and automating tasks such as logging peanut loads, generating water‑usage reports, and mapping seed layouts, the farm...

The video showcases the Codex app’s deep integration with Figma, allowing designers to launch a one‑click install via MCP and immediately access a dedicated Figma skill. This partnership streamlines the workflow: users copy a Figma file link, paste it into...

The video showcases how the Codex app automates routine software development tasks, turning repetitive chores into background processes that run on scheduled intervals. It walks through several automations: a daily commit‑summary that groups recent changes and highlights contributors; an “Upskill” routine...

The video unveils the Codex app, an AI-driven interface that consolidates project oversight and code generation into a single command center, allowing developers to delegate routine work to autonomous agents. Codex lets users issue commands by typing or speaking, then watches...

The video announces Prism, a free AI‑native platform that lives inside scientists’ editing environments, aiming to give researchers the same conversational AI assistance developers enjoy in code editors. Prism embeds the latest GPT‑5.2 model directly in the LaTeX editor, allowing instant...

As AI begins to meaningfully accelerate scientific discovery, we’re taking an early step to reduce friction in day-to-day research work with Prism. Prism is a free workspace for scientists to write and collaborate on research, powered by GPT-5.2. Prism offers unlimited...

The OpenAI Town Hall, hosted by CEO Sam Altman, centered on the growing urgency of AI ethics as the technology scales. Altman framed transparency, safety, and societal impact as the three pillars guiding OpenAI’s roadmap, signaling a shift from pure...

The OpenAI Build Hour introduced the ChatGPT Apps platform, showcasing a suite of new tools—including an Apps SDK, a public app submission flow, and a marketplace—designed to let developers embed interactive experiences directly inside ChatGPT. Core announcements covered the launch...