{"id":12805,"date":"2026-03-06T09:12:26","date_gmt":"2026-03-06T14:12:26","guid":{"rendered":"https:\/\/www.daillac.com\/?p=12805"},"modified":"2026-03-06T09:17:21","modified_gmt":"2026-03-06T14:17:21","slug":"chatgpt-in-2026-gpt-5-4-agents-and-long-context","status":"publish","type":"post","link":"https:\/\/www.daillac.com\/en\/blogue\/chatgpt-in-2026-gpt-5-4-agents-and-long-context\/","title":{"rendered":"ChatGPT in 2026: GPT-5.4, Agents, and Long Context"},"content":{"rendered":"\t\t<div data-elementor-type=\"wp-post\" data-elementor-id=\"12805\" class=\"elementor elementor-12805 elementor-12798\" data-elementor-post-type=\"post\">\n\t\t\t\t\t\t<section class=\"elementor-section elementor-top-section elementor-element elementor-element-35195cc elementor-section-boxed elementor-section-height-default elementor-section-height-default\" data-id=\"35195cc\" data-element_type=\"section\" data-e-type=\"section\">\n\t\t\t\t\t\t<div class=\"elementor-container elementor-column-gap-default\">\n\t\t\t\t\t<div class=\"elementor-column elementor-col-100 elementor-top-column elementor-element elementor-element-d97b788\" data-id=\"d97b788\" data-element_type=\"column\" data-e-type=\"column\">\n\t\t\t<div class=\"elementor-widget-wrap elementor-element-populated\">\n\t\t\t\t\t\t<div class=\"elementor-element elementor-element-f430f64 elementor-widget elementor-widget-html\" data-id=\"f430f64\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"html.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t<article class=\"dlx-article\" itemscope itemtype=\"https:\/\/schema.org\/Article\">\r\n  <header class=\"dlx-article__hero\">\r\n    <p class=\"dlx-article__eyebrow\">AI, agents and product architecture<\/p>\r\n\r\n    <h1 itemprop=\"headline\">ChatGPT and the Age of Agents in 2026: GPT-5.4, Long Context and Steerability<\/h1>\r\n\r\n    <p class=\"dlx-article__lead\" itemprop=\"description\">\r\n      In 2026, ChatGPT is no longer just a conversational assistant. It becomes an orchestration layer\r\n      capable of searching, reasoning, calling tools, acting within interfaces,\r\n      and integrating into entire workflows.\r\n    <\/p>\r\n\r\n    <div class=\"dlx-meta\" aria-label=\"Article information\">\r\n      <span>\r\n        <strong>Published:<\/strong>\r\n        <time datetime=\"2026-03-06\" itemprop=\"datePublished\">March 6, 2026<\/time>\r\n      <\/span>\r\n      <span>\r\n        <strong>Updated:<\/strong>\r\n        <time datetime=\"2026-03-06\" itemprop=\"dateModified\">March 6, 2026<\/time>\r\n      <\/span>\r\n      <span itemprop=\"author\" itemscope itemtype=\"https:\/\/schema.org\/Person\">\r\n        <strong>Author:<\/strong>\r\n        <span itemprop=\"name\">DAILLAC<\/span>\r\n      <\/span>\r\n      <span>\r\n        <strong>Reading time:<\/strong> 16 min\r\n      <\/span>\r\n    <\/div>\r\n  <\/header>\r\n\r\n  <nav class=\"dlx-toc\" aria-label=\"Table of contents\">\r\n    <div class=\"dlx-toc__title\">In this article<\/div>\r\n    <ul>\r\n      <li><a href=\"#executive-summary\">Executive summary<\/a><\/li>\r\n      <li><a href=\"#context-and-timeline\">Context and timeline<\/a><\/li>\r\n      <li><a href=\"#technical-capabilities\">Technical capabilities<\/a><\/li>\r\n      <li><a href=\"#performance-and-benchmarks\">Performance and benchmarks<\/a><\/li>\r\n      <li><a href=\"#key-use-cases\">Key use cases<\/a><\/li>\r\n      <li><a href=\"#privacy-security-steerability\">Privacy, security and steerability<\/a><\/li>\r\n      <li><a href=\"#developer-integration\">Developer integration<\/a><\/li>\r\n      <li><a href=\"#landscape-limitations-outlook\">Competitive landscape, limitations and outlook<\/a><\/li>\r\n      <li><a href=\"#faq\">FAQ<\/a><\/li>\r\n      <li><a href=\"#conclusion\">Conclusion<\/a><\/li>\r\n    <\/ul>\r\n  <\/nav>\r\n\r\n  <section id=\"executive-summary\" class=\"dlx-section dlx-reveal\" data-dlx=\"reveal\" itemprop=\"articleBody\">\r\n    <h2>Executive summary<\/h2>\r\n    <p>\r\n      In less than four years, ChatGPT has become a mainstream entry point into an agentic AI platform.\r\n      The shift is not only conversational: the system can answer,\r\n      plan, act, verify and iterate through tools such as the web, connectors, files,\r\n      the terminal and the native use of a computer.\r\n    <\/p>\r\n    <p>\r\n      GPT-5.4 marks an important milestone in this transition. The model is presented as a generalist\r\n      combining knowledge work, coding, vision, tool calling and native <em>computer use<\/em>.\r\n      The promise is therefore no longer just to \u201canswer better,\u201d but to execute better.\r\n    <\/p>\r\n\r\n    <div class=\"dlx-kpi\" aria-label=\"Key GPT-5.4 indicators\">\r\n      <div class=\"dlx-stat\">\r\n        <span class=\"dlx-stat__value\">1.05M<\/span>\r\n        <span class=\"dlx-stat__label\">documented context tokens on the API side<\/span>\r\n      <\/div>\r\n      <div class=\"dlx-stat\">\r\n        <span class=\"dlx-stat__value\">128K<\/span>\r\n        <span class=\"dlx-stat__label\">maximum announced output tokens<\/span>\r\n      <\/div>\r\n      <div class=\"dlx-stat\">\r\n        <span class=\"dlx-stat__value\">272K+<\/span>\r\n        <span class=\"dlx-stat__label\">input threshold from which pricing changes<\/span>\r\n      <\/div>\r\n      <div class=\"dlx-stat\">\r\n        <span class=\"dlx-stat__value\">80% \/ 90%<\/span>\r\n        <span class=\"dlx-stat__label\">latency \/ input cost reduction mentioned for prompt caching<\/span>\r\n      <\/div>\r\n    <\/div>\r\n\r\n    <p>Three \u201csystem\u201d innovations particularly shape this generation:<\/p>\r\n    <ol>\r\n      <li>\r\n        <strong>Long context and cost control<\/strong> with a 1,050,000-token window on the API side,\r\n        a maximum output of 128,000 tokens, and more complex economics when trajectories become very long.\r\n      <\/li>\r\n      <li>\r\n        <strong>Tool search<\/strong>, which allows tool schemas or MCP servers to be loaded on demand\r\n        instead of injecting everything into the prompt from the start.\r\n      <\/li>\r\n      <li>\r\n        <strong>Compaction and prompt caching<\/strong>, two mechanisms designed to preserve state, limit drift,\r\n        maintain performance and reduce costs on long workflows.\r\n      <\/li>\r\n    <\/ol>\r\n\r\n    <aside class=\"dlx-callout\">\r\n      <div class=\"dlx-callout__title\">Key takeaway<\/div>\r\n      <p>\r\n        The value of GPT-5.4 lies less in a single metric than in a combination of prolonged reasoning,\r\n        tool orchestration, action execution, context compression and improved steerability.\r\n      <\/p>\r\n    <\/aside>\r\n  <\/section>\r\n\r\n  <section id=\"context-and-timeline\" class=\"dlx-section dlx-reveal\" data-dlx=\"reveal\">\r\n    <h2>Context and timeline<\/h2>\r\n\r\n    <h3>From chatbot to work and action platform<\/h3>\r\n    <p>\r\n      ChatGPT\u2019s initial adoption was exceptional, with figures that became iconic:\r\n      one million users in five days, then one hundred million users in two months in the measurements\r\n      reported at the time. This acceleration fueled a second wave: agents, that is,\r\n      systems capable of observing, planning and acting in real environments.\r\n    <\/p>\r\n    <p>\r\n      In that logic, the question is no longer simply \u201cwhat can the model answer?\u201d\r\n      but rather \u201cwhich workflows can it execute properly, with which safeguards and at what cost?\u201d\r\n    <\/p>\r\n\r\n    <h3>Timeline of public milestones<\/h3>\r\n    <div class=\"dlx-table-wrap\">\r\n      <table>\r\n        <thead>\r\n          <tr>\r\n            <th>Date<\/th>\r\n            <th>Milestone<\/th>\r\n            <th>What it changes<\/th>\r\n          <\/tr>\r\n        <\/thead>\r\n        <tbody>\r\n          <tr>\r\n            <td><time datetime=\"2022-11-30\">November 30, 2022<\/time><\/td>\r\n            <td>Launch of ChatGPT<\/td>\r\n            <td>The dialogue format democratizes mainstream access to LLMs.<\/td>\r\n          <\/tr>\r\n          <tr>\r\n            <td><time datetime=\"2023-03-14\">March 14, 2023<\/time><\/td>\r\n            <td>GPT-4<\/td>\r\n            <td>The leap centers on quality, text+image multimodality and professional use cases.<\/td>\r\n          <\/tr>\r\n          <tr>\r\n            <td><time datetime=\"2024-05-13\">May 13, 2024<\/time><\/td>\r\n            <td>GPT-4o<\/td>\r\n            <td>The product accelerates on fluid multimodality and daily usage.<\/td>\r\n          <\/tr>\r\n          <tr>\r\n            <td><time datetime=\"2025-04-14\">April 14, 2025<\/time><\/td>\r\n            <td>GPT-4.1<\/td>\r\n            <td>The API offering shifts toward developer use cases and very long context.<\/td>\r\n          <\/tr>\r\n          <tr>\r\n            <td><time datetime=\"2025-07-17\">July 17, 2025<\/time><\/td>\r\n            <td>ChatGPT agent<\/td>\r\n            <td>The product highlights a mode capable of thinking and acting with a computer and connectors.<\/td>\r\n          <\/tr>\r\n          <tr>\r\n            <td><time datetime=\"2025-10-21\">October 21, 2025<\/time><\/td>\r\n            <td>Atlas<\/td>\r\n            <td>The agent-centered browser becomes a competitive field in its own right.<\/td>\r\n          <\/tr>\r\n          <tr>\r\n            <td><time datetime=\"2026-02-02\">February 2, 2026<\/time><\/td>\r\n            <td>Codex app<\/td>\r\n            <td>OpenAI exposes a multi-agent architecture for software development.<\/td>\r\n          <\/tr>\r\n          <tr>\r\n            <td><time datetime=\"2026-02-09\">February 9, 2026<\/time><\/td>\r\n            <td>GPT-5.3-Codex in GitHub Copilot<\/td>\r\n            <td>Developer agents become standardized inside IDEs.<\/td>\r\n          <\/tr>\r\n          <tr>\r\n            <td><time datetime=\"2026-03-05\">March 5, 2026<\/time><\/td>\r\n            <td>GPT-5.4<\/td>\r\n            <td>Native computer use, tool search, long context and compaction become a coherent whole.<\/td>\r\n          <\/tr>\r\n        <\/tbody>\r\n      <\/table>\r\n    <\/div>\r\n\r\n    <h3>Recent evolution of the ChatGPT offering and product implications<\/h3>\r\n    <p>\r\n      The source report highlights a structuring distinction between <strong>ChatGPT as a product<\/strong> and\r\n      <strong>the API as a platform<\/strong>. The 1M-token window is primarily a promise\r\n      of orchestration for developers, rather than a standard capability accessible as-is to end users\r\n      inside the ChatGPT interface.\r\n    <\/p>\r\n    <p>\r\n      In other words, the story of GPT-5.4 is also a story of different product surfaces:\r\n      what can be done in ChatGPT is not strictly the same as what can be built\r\n      with the Responses API, tool search, background mode and MCP.\r\n    <\/p>\r\n  <\/section>\r\n\r\n  <section id=\"technical-capabilities\" class=\"dlx-section dlx-reveal\" data-dlx=\"reveal\">\r\n    <h2>Technical capabilities and public architecture<\/h2>\r\n\r\n    <h3>What we know about LLM architecture and reasoning specialization<\/h3>\r\n    <p>\r\n      OpenAI does not publish the detailed architecture of GPT-5.4 the way a full academic paper would,\r\n      but the background remains that of large Transformer-based models. The major difference in 2025\u20132026\r\n      comes from the industrialization of reasoning-oriented models, capable of extending their thinking\r\n      and being better guided through a more explicit instruction layer.\r\n    <\/p>\r\n    <p>\r\n      The report also stresses an important point: chain-of-thought exists as an internal mechanism,\r\n      but it is not fully exposed. This reinforces the idea of AI as a supervisable system,\r\n      not just a chat box.\r\n    <\/p>\r\n\r\n    <h3>GPT-5.4 as \u201cmodel + system\u201d<\/h3>\r\n    <p>\r\n      GPT-5.4 should be understood as a whole. The model alone is not enough to explain the performance\r\n      observed in the age of agents. The full loop includes reasoning, tool calling,\r\n      tool discovery, action execution, context compression and state management.\r\n    <\/p>\r\n\r\n    <div class=\"dlx-note\">\r\n      <strong>System reading:<\/strong>\r\n      <ol>\r\n        <li>The model decides whether it should call a tool.<\/li>\r\n        <li>It can discover or load the relevant tool via tool search rather than carrying all schemas in context.<\/li>\r\n        <li>It can execute <em>computer use<\/em> actions and retrieve a new state.<\/li>\r\n        <li>It can compact history to stay on track during long trajectories.<\/li>\r\n        <li>It can be orchestrated in long-running executions via background mode, webhooks and traces.<\/li>\r\n      <\/ol>\r\n    <\/div>\r\n\r\n    <h3>Long context and optimization mechanisms<\/h3>\r\n    <p>\r\n      On the API side, the report notes a documented context window of 1,050,000 tokens for\r\n      <code>gpt-5.4<\/code> and <code>gpt-5.4-pro<\/code>, with up to 128,000 output tokens.\r\n      But the raw promise of long context has a trade-off: beyond 272K input tokens,\r\n      pricing changes and invisible reasoning tokens still count in the overall economics.\r\n    <\/p>\r\n    <p>\r\n      <strong>Compaction<\/strong> is used to reduce context while preserving state, while\r\n      <strong>prompt caching<\/strong> aims to preserve a stable prefix to reduce latency and cost.\r\n      The report therefore reminds us that useful \u201clong context\u201d is not a pile-up of tokens:\r\n      it is an orchestration discipline.\r\n    <\/p>\r\n\r\n    <h3>Tool search and MCP<\/h3>\r\n    <p>\r\n      Tool search addresses a simple problem: in enterprise environments, the number of tools,\r\n      connectors and functions can make prompts explode in size and degrade latency.\r\n      The idea is therefore to make tools or MCP servers \u201cdiscoverable,\u201d then load only\r\n      what becomes necessary.\r\n    <\/p>\r\n    <p>\r\n      MCP plays the role of a standardized connectivity layer here. From this perspective, the agent is no longer\r\n      just an enhanced model: it is an orchestrator capable of moving between services, data,\r\n      screens and specialized functions.\r\n    <\/p>\r\n  <\/section>\r\n\r\n  <section id=\"performance-and-benchmarks\" class=\"dlx-section dlx-reveal\" data-dlx=\"reveal\">\r\n    <h2>Performance, benchmarks and comparisons<\/h2>\r\n\r\n    <h3>Critical reading of agent benchmarks<\/h3>\r\n    <p>\r\n      Benchmarks in the agentic era assess less the quality of an isolated answer than the ability\r\n      to complete a task in a tool-enabled environment: virtual desktop, browser, codebase,\r\n      terminal or software repository. This improves proximity to real-world usage, but also makes\r\n      comparisons more difficult because system parameters matter as much as the model.\r\n    <\/p>\r\n\r\n    <h3>Comparative table of model capabilities<\/h3>\r\n    <div class=\"dlx-table-wrap\">\r\n      <table>\r\n        <thead>\r\n          <tr>\r\n            <th>Model \/ variant<\/th>\r\n            <th>Surface<\/th>\r\n            <th>Context<\/th>\r\n            <th>Max output<\/th>\r\n            <th>Modalities<\/th>\r\n            <th>Positioning<\/th>\r\n          <\/tr>\r\n        <\/thead>\r\n        <tbody>\r\n          <tr>\r\n            <td>GPT-5.4<\/td>\r\n            <td>API<\/td>\r\n            <td>1,050,000<\/td>\r\n            <td>128,000<\/td>\r\n            <td>text + image \u2192 text<\/td>\r\n            <td>Generalist agentic model<\/td>\r\n          <\/tr>\r\n          <tr>\r\n            <td>GPT-5.4 Pro<\/td>\r\n            <td>API<\/td>\r\n            <td>1,050,000<\/td>\r\n            <td>128,000<\/td>\r\n            <td>text + image \u2192 text<\/td>\r\n            <td>More precise answers, much higher cost<\/td>\r\n          <\/tr>\r\n          <tr>\r\n            <td>GPT-4o<\/td>\r\n            <td>API<\/td>\r\n            <td>128,000<\/td>\r\n            <td>16,384<\/td>\r\n            <td>text + image \u2192 text<\/td>\r\n            <td>Fast multimodal model, advanced structuring<\/td>\r\n          <\/tr>\r\n          <tr>\r\n            <td>GPT-4.1<\/td>\r\n            <td>API<\/td>\r\n            <td>1,047,576<\/td>\r\n            <td>\u2014<\/td>\r\n            <td>text + image \u2192 text<\/td>\r\n            <td>Pro-dev and long-context pivot<\/td>\r\n          <\/tr>\r\n          <tr>\r\n            <td>GPT-5.4 Thinking<\/td>\r\n            <td>ChatGPT<\/td>\r\n            <td>256K to 400K depending on the plan<\/td>\r\n            <td>up to 128K implicit<\/td>\r\n            <td>ChatGPT tools<\/td>\r\n            <td>Product version focused on reasoning<\/td>\r\n          <\/tr>\r\n        <\/tbody>\r\n      <\/table>\r\n    <\/div>\r\n\r\n    <h3>Key results published in the report<\/h3>\r\n    <div class=\"dlx-table-wrap\">\r\n      <table>\r\n        <thead>\r\n          <tr>\r\n            <th>Area<\/th>\r\n            <th>GPT-5.4<\/th>\r\n            <th>Comparative reference<\/th>\r\n            <th>Useful interpretation<\/th>\r\n          <\/tr>\r\n        <\/thead>\r\n        <tbody>\r\n          <tr>\r\n            <td>GDPval<\/td>\r\n            <td>83.0%<\/td>\r\n            <td>70.9% for GPT-5.2<\/td>\r\n            <td>Improvement on knowledge-work tasks.<\/td>\r\n          <\/tr>\r\n          <tr>\r\n            <td>OSWorld-Verified<\/td>\r\n            <td>75.0%<\/td>\r\n            <td>47.3% for GPT-5.2<\/td>\r\n            <td>Computer use gains significant maturity.<\/td>\r\n          <\/tr>\r\n          <tr>\r\n            <td>SWE-Bench Pro<\/td>\r\n            <td>57.7%<\/td>\r\n            <td>56.8% for GPT-5.3-Codex<\/td>\r\n            <td>Coding remains a highly competitive field.<\/td>\r\n          <\/tr>\r\n          <tr>\r\n            <td>Terminal-Bench 2.0<\/td>\r\n            <td>75.1%<\/td>\r\n            <td>77.3% for GPT-5.3-Codex<\/td>\r\n            <td>The best \u201cterminal agent\u201d is not automatically the most generalist model.<\/td>\r\n          <\/tr>\r\n          <tr>\r\n            <td>BrowseComp<\/td>\r\n            <td>82.7%<\/td>\r\n            <td>65.8% for GPT-5.2<\/td>\r\n            <td>Tool-enabled browsing improves markedly.<\/td>\r\n          <\/tr>\r\n          <tr>\r\n            <td>Long context<\/td>\r\n            <td>visible degradation at 256K\u20131M<\/td>\r\n            <td>Graphwalks BFS 256K\u20131M: 21.4%<\/td>\r\n            <td>1M context does not mean perfect understanding at 1M.<\/td>\r\n          <\/tr>\r\n        <\/tbody>\r\n      <\/table>\r\n    <\/div>\r\n\r\n    <h3>Contextualized comparison with GPT-4.x and the coding trajectory<\/h3>\r\n    <p>\r\n      GPT-4 already represented a major leap in professional use cases and multimodality.\r\n      GPT-4.1 then opened a cycle more explicitly focused on developers, with instruction following,\r\n      coding and long context. GPT-5.4 pushes the agentic logic further, while Codex illustrates\r\n      a specialized product layer for long, iterative and supervised software development.\r\n    <\/p>\r\n    <p>\r\n      The report therefore invites readers not to confuse three things: the quality of the raw model, the quality\r\n      of the tool-enabled system, and the relevance of a specialized product for a given workflow type.\r\n    <\/p>\r\n  <\/section>\r\n\r\n  <section id=\"key-use-cases\" class=\"dlx-section dlx-reveal\" data-dlx=\"reveal\">\r\n    <h2>Key use cases<\/h2>\r\n\r\n    <h3>Native computer use: automating UI-only workflows<\/h3>\r\n    <p>\r\n      <em>Computer use<\/em> targets tasks that historically required a human in front of the screen:\r\n      navigation, forms, office suites, visual checks, state validation and manipulation\r\n      of interfaces that do not always provide a usable API.\r\n    <\/p>\r\n    <p>\r\n      The report emphasizes a security-by-design approach: isolated environment,\r\n      limited accounts, confirmations at the right time and authorization policies adapted to the level of risk.\r\n    <\/p>\r\n\r\n    <h3>AI agents: from research to action<\/h3>\r\n    <p>\r\n      ChatGPT agent is presented as a system capable of thinking and acting more proactively,\r\n      while Codex illustrates a software production variant with multi-agents, worktrees,\r\n      sandboxing, permission rules and reusable \u201cskills.\u201d\r\n    <\/p>\r\n\r\n    <h3>Tool search and connectors<\/h3>\r\n    <p>\r\n      In the enterprise, the real difficulty is not only having tools, but having too many\r\n      tools. Tool search makes it possible not to expose the entire tool catalog to the model at all times.\r\n      Activation becomes lighter in tokens, faster and potentially more reliable.\r\n    <\/p>\r\n\r\n    <h3>Long-context workflows up to 1M tokens<\/h3>\r\n    <p>The report identifies four use cases that are especially well suited:<\/p>\r\n    <ul>\r\n      <li>analysis of large codebases or monorepos,<\/li>\r\n      <li>large documentary files,<\/li>\r\n      <li>long agent trajectories with trial and error,<\/li>\r\n      <li>multi-source consolidation across connectors, web and files.<\/li>\r\n    <\/ul>\r\n    <p>\r\n      But it recommends a hybrid strategy: keep the key pieces in context, compact the rest,\r\n      structure the outputs and do not blindly replace RAG, extraction and orchestration with a giant window.\r\n    <\/p>\r\n  <\/section>\r\n\r\n  <section id=\"privacy-security-steerability\" class=\"dlx-section dlx-reveal\" data-dlx=\"reveal\">\r\n    <h2>Privacy, security and steerability<\/h2>\r\n\r\n    <h3>Behavioral governance<\/h3>\r\n    <p>\r\n      The report highlights a more explicit instruction hierarchy and stronger steerability.\r\n      The objective is twofold: make the system more controllable in complex use cases, without losing\r\n      platform safeguards.\r\n    <\/p>\r\n\r\n    <h3>Computer use security<\/h3>\r\n    <p>\r\n      As soon as an agent can delete, send, pay or modify permissions, it enters a high-risk zone.\r\n      Confirmation at the critical moment, explanation of the action and the handling of pre-approvals then become\r\n      product components, not interface details.\r\n    <\/p>\r\n\r\n    <h3>Prompt injection and attacks through browsers or connectors<\/h3>\r\n    <p>\r\n      The shift from \u201cresponding\u201d to \u201cacting\u201d mechanically increases the potential impact of compromise.\r\n      The report identifies several risk surfaces: malicious web pages, hidden instructions,\r\n      data exfiltration, unwanted tool calls and destructive use of accounts or connectors.\r\n    <\/p>\r\n\r\n    <h3>Cyber capability, data and privacy<\/h3>\r\n    <p>\r\n      The source text emphasizes multi-layer security: policies, confirmations, classifiers,\r\n      review thresholds, restricted-access programs and reinforced supervision for sensitive use cases.\r\n      It also recalls important distinctions between retention, ZDR, background mode and compaction.\r\n    <\/p>\r\n    <p>\r\n      Finally, the privacy section reminds us that data governance, possible opt-in, separation\r\n      between advertising and answers, and user controls remain structuring issues in a context\r\n      where agents manipulate more state and work surfaces.\r\n    <\/p>\r\n\r\n    <aside class=\"dlx-callout\">\r\n      <div class=\"dlx-callout__title\">Watch point<\/div>\r\n      <p>\r\n        The more an agent can act, the more security must be designed as a complete architecture:\r\n        authorization policies, execution environment, traces, auditability, human review and access limits.\r\n      <\/p>\r\n    <\/aside>\r\n  <\/section>\r\n\r\n  <section id=\"developer-integration\" class=\"dlx-section dlx-reveal\" data-dlx=\"reveal\">\r\n    <h2>Developer integration and architecture patterns<\/h2>\r\n\r\n    <h3>Responses API, long execution and observability<\/h3>\r\n    <p>\r\n      The report positions the Responses API as the foundation for multi-turn workflows rich in tool calls.\r\n      On top of this come long execution, webhooks, background mode, state management\r\n      and the traces required for observability.\r\n    <\/p>\r\n\r\n    <h3>Robust agent pattern<\/h3>\r\n    <div class=\"dlx-note\">\r\n      <ol>\r\n        <li>Responses API in stateful or stateless mode depending on governance constraints.<\/li>\r\n        <li>Tool calling and tool search to defer rare schemas.<\/li>\r\n        <li>Threshold-based compaction to preserve state without endlessly inflating context.<\/li>\r\n        <li>Prompt caching to stabilize the cost of recurring parts.<\/li>\r\n        <li>Webhooks and traces for observability.<\/li>\r\n        <li>Explicit confirmation policy for any risky action.<\/li>\r\n      <\/ol>\r\n    <\/div>\r\n\r\n    <h3>Tool catalog governance<\/h3>\r\n    <p>\r\n      A good agentic architecture is not only about connecting more tools.\r\n      It requires catalog discipline: high-level descriptions, well-framed namespaces,\r\n      schema versioning, testing, measurement of activation cost and latency tracking.\r\n    <\/p>\r\n\r\n    <h3>MCP, Apps SDK and connectors<\/h3>\r\n    <p>\r\n      MCP is presented as a standardization layer for connectors and actions.\r\n      For organizations, this opens a logic of a centralized \u201ctool bus,\u201d more maintainable\r\n      than an accumulation of isolated functions exposed without governance.\r\n    <\/p>\r\n\r\n    <h3>Codex as a reference architecture for agentic development<\/h3>\r\n    <p>\r\n      Codex is interesting because it shows that an agent becomes productive not only because it \u201ccan code,\u201d\r\n      but because it can execute, be relaunched, be controlled, manage permissions and produce\r\n      auditable iterations in a real working environment.\r\n    <\/p>\r\n  <\/section>\r\n\r\n  <section id=\"landscape-limitations-outlook\" class=\"dlx-section dlx-reveal\" data-dlx=\"reveal\">\r\n    <h2>Competitive landscape, limitations and outlook<\/h2>\r\n\r\n    <h3>Market: agents as the next wave<\/h3>\r\n    <p>\r\n      The analyses relayed in the report converge on the same idea: the next wave of value creation\r\n      will not come only from content generation, but from the transformation of entire workflows,\r\n      especially in organizations where processes are complex, document-heavy and multi-tool.\r\n    <\/p>\r\n\r\n    <h3>Competition: computer use, 1M tokens and actions are becoming the new standards<\/h3>\r\n    <p>\r\n      Google, Anthropic, Perplexity and Microsoft are all moving forward on similar building blocks:\r\n      active tool use, search layers, giant context windows, connectors,\r\n      AI browsers and development agents. Competition is therefore shifting toward execution capacity,\r\n      integration into work environments and operational security.\r\n    <\/p>\r\n\r\n    <h3>Technical and operational limitations<\/h3>\r\n    <p>\r\n      The report highlights several limitations. First, <strong>long context does not mean\r\n      reliable long reasoning<\/strong>. Second, costs and latency remain decisive,\r\n      especially for pro variants. Finally, benchmarks remain imperfect because they often measure\r\n      a mixture of model, tooling, settings and evaluation conditions.\r\n    <\/p>\r\n\r\n    <h3>12\u201324 month outlook<\/h3>\r\n    <ul>\r\n      <li>greater standardization of tool interfaces and catalogs,<\/li>\r\n      <li>more scalable supervision through traces and internal signals,<\/li>\r\n      <li>stronger convergence between office software, agents and work surfaces,<\/li>\r\n      <li>growing economic pressure on monetization models and data governance.<\/li>\r\n    <\/ul>\r\n\r\n    <h3>Sources and consulted documents<\/h3>\r\n    <p>\r\n      The original report relies on a broad corpus, dominated by OpenAI and its API documentation,\r\n      but also by consulting analyses, market publications, competitor announcements\r\n      and academic references. For a final web version with a clickable bibliography,\r\n      it would be relevant to inject the list of links from the DOCX file afterward.\r\n    <\/p>\r\n  <\/section>\r\n\r\n  <section id=\"faq\" class=\"dlx-section dlx-reveal\" data-dlx=\"reveal\">\r\n    <h2>FAQ<\/h2>\r\n\r\n    <div class=\"dlx-faq\">\r\n      <div class=\"dlx-faq__item is-open\">\r\n        <button class=\"dlx-faq__question\" type=\"button\" aria-expanded=\"true\">\r\n          Is GPT-5.4 mainly a better chatbot or a better action system?\r\n        <\/button>\r\n        <div class=\"dlx-faq__answer\">\r\n          <p>\r\n            The report points more toward the second reading. GPT-5.4 becomes interesting when it is considered\r\n            as a complete system combining reasoning, tools, computer use, compaction, caching,\r\n            long orchestration and security policies.\r\n          <\/p>\r\n        <\/div>\r\n      <\/div>\r\n\r\n      <div class=\"dlx-faq__item\">\r\n        <button class=\"dlx-faq__question\" type=\"button\" aria-expanded=\"false\">\r\n          Does the 1M-token window really change practice?\r\n        <\/button>\r\n        <div class=\"dlx-faq__answer\">\r\n          <p>\r\n            Yes, but not on its own. It opens new use cases, especially for large files\r\n            and long trajectories, but it must be combined with compaction, caching,\r\n            structured extraction and disciplined orchestration.\r\n          <\/p>\r\n        <\/div>\r\n      <\/div>\r\n\r\n      <div class=\"dlx-faq__item\">\r\n        <button class=\"dlx-faq__question\" type=\"button\" aria-expanded=\"false\">\r\n          Why is tool search strategic in the enterprise?\r\n        <\/button>\r\n        <div class=\"dlx-faq__answer\">\r\n          <p>\r\n            Because it avoids permanently surfacing the entire tool catalog to the model.\r\n            This reduces token footprint, preserves the cache, improves latency and simplifies\r\n            connector governance.\r\n          <\/p>\r\n        <\/div>\r\n      <\/div>\r\n\r\n      <div class=\"dlx-faq__item\">\r\n        <button class=\"dlx-faq__question\" type=\"button\" aria-expanded=\"false\">\r\n          What is the main risk when an agent can act in a browser or a virtual desktop?\r\n        <\/button>\r\n        <div class=\"dlx-faq__answer\">\r\n          <p>\r\n            The main risk is the increased impact of an error or an attack:\r\n            prompt injection, leakage through connectors, destructive action, or implicit validation of a\r\n            sensitive operation. That is why the confirmation policy becomes central.\r\n          <\/p>\r\n        <\/div>\r\n      <\/div>\r\n\r\n      <div class=\"dlx-faq__item\">\r\n        <button class=\"dlx-faq__question\" type=\"button\" aria-expanded=\"false\">\r\n          What should be retained for a robust agentic architecture?\r\n        <\/button>\r\n        <div class=\"dlx-faq__answer\">\r\n          <p>\r\n            It must be thought of in layers: model, tool calls, catalog governance,\r\n            controlled execution, compaction, observability, permissions and auditability.\r\n            Robustness comes from the whole, not from a single benchmark.\r\n          <\/p>\r\n        <\/div>\r\n      <\/div>\r\n    <\/div>\r\n  <\/section>\r\n\r\n  <section id=\"conclusion\" class=\"dlx-section dlx-reveal\" data-dlx=\"reveal\">\r\n    <h2>Conclusion<\/h2>\r\n    <p>\r\n      GPT-5.4 crystallizes an already ongoing shift: AI is becoming less a text generator\r\n      and more a workflow operator. The real novelty is not only that a model answers better,\r\n      but that it knows how to search, choose a tool, act, preserve state, be supervised and be redirected.\r\n    <\/p>\r\n    <p>\r\n      For product, tech and innovation teams, the right reading is therefore not \u201cwhich score is the best?\u201d\r\n      but rather \u201cwhich architecture enables an agent that is useful, controllable and economically sustainable?\u201d\r\n      The source report shows that the answer will lie in systems that are more composable, better instrumented\r\n      and more strictly governed.\r\n    <\/p>\r\n  <\/section>\r\n<\/article>\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t\t<\/div>\n\t\t<\/div>\n\t\t\t\t\t<\/div>\n\t\t<\/section>\n\t\t\t\t<\/div>\n\t\t","protected":false},"excerpt":{"rendered":"<p>AI, agents and product architecture ChatGPT and the Age of Agents in 2026: GPT-5.4, Long Context and Steerability In 2026, ChatGPT is no longer just a conversational assistant. It becomes an orchestration layer capable of searching, reasoning, calling tools, acting within interfaces, and integrating into entire workflows. Published: March 6, 2026 Updated: March 6, 2026 [&hellip;]<\/p>\n","protected":false},"author":4,"featured_media":12800,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[61],"tags":[],"class_list":["post-12805","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-non-classified"],"_links":{"self":[{"href":"https:\/\/www.daillac.com\/en\/wp-json\/wp\/v2\/posts\/12805","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.daillac.com\/en\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.daillac.com\/en\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.daillac.com\/en\/wp-json\/wp\/v2\/users\/4"}],"replies":[{"embeddable":true,"href":"https:\/\/www.daillac.com\/en\/wp-json\/wp\/v2\/comments?post=12805"}],"version-history":[{"count":4,"href":"https:\/\/www.daillac.com\/en\/wp-json\/wp\/v2\/posts\/12805\/revisions"}],"predecessor-version":[{"id":12809,"href":"https:\/\/www.daillac.com\/en\/wp-json\/wp\/v2\/posts\/12805\/revisions\/12809"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/www.daillac.com\/en\/wp-json\/wp\/v2\/media\/12800"}],"wp:attachment":[{"href":"https:\/\/www.daillac.com\/en\/wp-json\/wp\/v2\/media?parent=12805"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.daillac.com\/en\/wp-json\/wp\/v2\/categories?post=12805"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.daillac.com\/en\/wp-json\/wp\/v2\/tags?post=12805"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}