{"id":6626,"date":"2026-04-25T19:43:38","date_gmt":"2026-04-25T14:13:38","guid":{"rendered":"https:\/\/nervnow.com\/?p=6626"},"modified":"2026-04-25T19:43:56","modified_gmt":"2026-04-25T14:13:56","slug":"openai-releases-gpt-5-5-targeting-agentic-coding-and-scientific-research","status":"publish","type":"post","link":"https:\/\/nervnow.com\/ro\/openai-releases-gpt-5-5-targeting-agentic-coding-and-scientific-research\/","title":{"rendered":"OpenAI Releases GPT-5.5, Targeting Agentic Coding and Scientific Research"},"content":{"rendered":"<p><strong><em>The model delivers state-of-the-art benchmark results in software engineering and data analysis while matching OpenAI predecessor per-token serving latency.<\/em><\/strong><\/p>\n\n\n\n<p>OpenAI released GPT-5.5 on April 23, its newest large language model, making it available to paying ChatGPT and Codex subscribers and positioning it as the company&#8217;s most capable model for extended, multi-step tasks.&nbsp;<\/p>\n\n\n\n<p>The company said the model is designed to handle agentic workflow assignments that require planning, tool use, error checking, and sustained execution without step-by-step human guidance. GPT-5.5 is available to Plus, Pro, Business, and Enterprise subscribers in ChatGPT and Codex; API access at $5 per million input tokens and $30 per million output tokens is also now live as of April 24.<\/p>\n\n\n\n<p>On Terminal-Bench 2.0, which evaluates complex command-line workflows, GPT-5.5 scored 82.7%, compared with 75.1% for GPT-5.4. On SWE-Bench Pro, which measures the resolution of real GitHub issues, it reached 58.6%. OpenAI said the model achieves those results while matching GPT-5.4&#8217;s per-token latency and using fewer tokens on equivalent Codex tasks.&nbsp;<\/p>\n\n\n\n<p>The gains carry into professional and knowledge-work settings. On GDPval, a benchmark assessing structured output quality across 44 occupations, GPT-5.5 scored 84.9%. On OSWorld-Verified, which tests autonomous computer operation, it reached 78.7%. On Tau2-bench Telecom, which simulates complex customer-service workflows, it scored 98.0% without prompt tuning, up from 92.8% for GPT-5.4.<\/p>\n\n\n\n<p>Michael Truell, co-founder and CEO of Cursor, said in a statement on OpenAI&#8217;s blog that GPT-5.5 &#8220;is noticeably smarter and more persistent than GPT-5.4, with stronger coding performance and more reliable tool use.&#8221;&nbsp;<\/p>\n\n\n\n<p>OpenAI said teams inside the company are already running the model in production. The finance team used GPT-5.5 in Codex to process 24,771 K-1 tax forms spanning 71,637 pages. The company&#8217;s communications group used it to build a scoring framework for speaking requests and automate low-risk approvals via a Slack agent.<\/p>\n\n\n\n<p><strong>ALSO READ: <\/strong><a href=\"https:\/\/nervnow.com\/ro\/stanford-professors-startup-human-intelligence-seeks-1-billion-valuation\/\" target=\"_blank\" rel=\"noopener\" title=\"\"><strong>Stanford Professor\u2019s Startup Human Intelligence Seeks $1 Billion Valuation<\/strong><\/a><\/p>\n\n\n\n<p>On scientific benchmarks, GPT-5.5 scored 80.5% on BixBench, which evaluates real-world bioinformatics analysis, and 25.0% on GeneBench, a multi-stage genetics evaluation where GPT-5.4 had scored 19.0%. OpenAI said a version of the model, trained with a custom research setup, produced a new proof of a result on off-diagonal Ramsey numbers in combinatorics, which was later verified using the Lean proof assistant. It&#8217;s unclear how broadly that kind of mathematical contribution will generalize across disciplines at this stage.<\/p>\n\n\n\n<p>Separately, OpenAI classified GPT-5.5&#8217;s cybersecurity and biological or chemical capabilities as &#8220;High&#8221; under its Preparedness Framework, one level below &#8220;Critical.&#8221; The company said it deployed stricter automated classifiers for cybersecurity-related requests at launch and is offering a Trusted Access for Cyber pathway at chatgpt.com\/cyber for verified security professionals who require expanded access.<\/p>\n\n\n\n<p>In Codex, GPT-5.5 ships with a 400,000-token context window across Plus, Pro, Business, Enterprise, Edu, and Go plans. Fast mode, which generates tokens 1.5 times faster, is available at 2.5 times the standard cost. GPT-5.5 Pro, priced at $30 per million input tokens and $180 per million output tokens in the API, is rolling out to Pro, Business, and Enterprise users.<\/p>\n\n\n\n<p>OpenAI said GPT-5.5 was co-designed with and served on NVIDIA GB200 and GB300 NVL72 systems. The company added that Codex helped its infrastructure team analyze production traffic and write load-balancing heuristics, improving token generation speeds by more than 20%.<\/p>\n\n\n\n<p class=\"has-white-color has-palette-color-9-background-color has-text-color has-background has-link-color wp-elements-6ed15f24208232bfc79039d4f756fe3f\"><strong><em>Disclaimer: This news is based on publicly available information. NervNow has not independently verified any claims.<\/em><\/strong><br><br><strong>MORE ON OPENAI<\/strong><br><a href=\"https:\/\/nervnow.com\/ro\/nitin-bawankule-to-join-openai-as-head-of-enterprise-sales-india\/\" target=\"_blank\" rel=\"noopener\" title=\"\"><strong>Nitin Bawankule to Join OpenAI as Head of Enterprise Sales, India<\/strong><\/a><br><a href=\"https:\/\/nervnow.com\/ro\/openai-acquires-tbpn-to-strengthen-ai-communication-and-industry-narrative\/\" target=\"_blank\" rel=\"noopener\" title=\"\"><strong>OpenAI Acquires TBPN to Strengthen AI Communication and Industry Narrative<\/strong><\/a><br><a href=\"https:\/\/nervnow.com\/ro\/openai-raises-122-billion-as-it-expands-ai-infrastructure\/\" target=\"_blank\" rel=\"noopener\" title=\"\"><strong>OpenAI Raises $122 Billion as It Expands AI Infrastructure<\/strong><\/a><\/p>","protected":false},"excerpt":{"rendered":"<p>The model delivers state-of-the-art benchmark results in software engineering and data analysis while matching OpenAI predecessor per-token serving latency.<\/p>","protected":false},"author":2,"featured_media":6627,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"_gspb_post_css":"","om_disable_all_campaigns":false,"_monsterinsights_skip_tracking":false,"_monsterinsights_sitenote_active":false,"_monsterinsights_sitenote_note":"","_monsterinsights_sitenote_category":0,"footnotes":""},"categories":[104,94],"tags":[196,153],"class_list":["post-6626","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-model-product-updates","category-news","tag-global","tag-openai"],"blocksy_meta":[],"aioseo_notices":[],"_links":{"self":[{"href":"https:\/\/nervnow.com\/ro\/wp-json\/wp\/v2\/posts\/6626","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/nervnow.com\/ro\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/nervnow.com\/ro\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/nervnow.com\/ro\/wp-json\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/nervnow.com\/ro\/wp-json\/wp\/v2\/comments?post=6626"}],"version-history":[{"count":3,"href":"https:\/\/nervnow.com\/ro\/wp-json\/wp\/v2\/posts\/6626\/revisions"}],"predecessor-version":[{"id":6639,"href":"https:\/\/nervnow.com\/ro\/wp-json\/wp\/v2\/posts\/6626\/revisions\/6639"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/nervnow.com\/ro\/wp-json\/wp\/v2\/media\/6627"}],"wp:attachment":[{"href":"https:\/\/nervnow.com\/ro\/wp-json\/wp\/v2\/media?parent=6626"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/nervnow.com\/ro\/wp-json\/wp\/v2\/categories?post=6626"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/nervnow.com\/ro\/wp-json\/wp\/v2\/tags?post=6626"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}