Close Menu
geekfence.comgeekfence.com
    What's Hot

    AI will likely shut down critical infrastructure on its own, no attackers required – Computerworld

    February 15, 2026

    A new era of humanoid robots

    February 15, 2026

    LLaMA in R with Keras and TensorFlow

    February 15, 2026
    Facebook X (Twitter) Instagram
    • About Us
    • Contact Us
    Facebook Instagram
    geekfence.comgeekfence.com
    • Home
    • UK Tech News
    • AI
    • Big Data
    • Cyber Security
      • Cloud Computing
      • iOS Development
    • IoT
    • Mobile
    • Software
      • Software Development
      • Software Engineering
    • Technology
      • Green Technology
      • Nanotechnology
    • Telecom
    geekfence.comgeekfence.com
    Home»Cloud Computing»IBM extends serverless computing to GPU workloads for enterprise AI and simulation
    Cloud Computing

    IBM extends serverless computing to GPU workloads for enterprise AI and simulation

    AdminBy AdminNovember 6, 2025No Comments3 Mins Read0 Views
    Facebook Twitter Pinterest LinkedIn Telegram Tumblr Email
    IBM extends serverless computing to GPU workloads for enterprise AI and simulation
    Share
    Facebook Twitter LinkedIn Pinterest Email


    The challenge of running simulation and high-performance workloads efficiently is a constant issue, requiring input from stakeholders including infrastructure teams, cybersecurity professionals, and, of course, ever-watchful finance officers.

    Running these types of high-compute tasks often involves thousands of concurrent processes and are costly to run on traditional infrastructure. IBM’s latest update to its Cloud Code Engine – the launch of Serverless Fleets with GPU support – may reduce complexity. They combine high-performance computing with a managed, pay-as-you-go serverless model, where one point of reference is addressed by the user, and necessary deployment at scale takes place autonomously.

    High-performance computing without infrastructure friction

    Enterprises running large-scale AI training, risk simulations, or generative workloads are looking at two problems, commonly: limited GPU access and rising infrastructure/cloud costs. Serverless Fleets provides an alternative. Instead of maintaining dedicated GPU clusters, organisations can submit large batches of compute jobs through a single endpoint.

    IBM’s system provisions GPU-backed virtual machines, executes the workload, and tapers off the resources used when complete. This approach improves utilisation and cost visibility, IBM claims, with customers only charged for active runtime.

    In practice, this could help financial institutions (for example) with faster risk modelling, or let media companies render their workloads without investing in GPU farms or entering long leases. For many, it means faster innovation and reduced operational overhead.

    Implementation realities

    IBM suggests that Serverless Fleets can manage workloads at scale “with essentially zero SRE staff.” While ambitious, the model certainly simplifies the detail of orchestration. Code Engine can determine the number of worker instances needed and scale them to match the demanded work. This reduces the tuning typically required to balance parallel GPU tasks.

    Adopting the platform, however, would need careful oversight with a keen eye on costs – ubiquitous challenges in serverless environments. Enterprises will need clear visibility into their common workload patterns, plus be aware of any compliance issues when considering effectively out-sourcing GPU-heavy jobs to a managed cloud.

    Market and ecosystem context

    IBM joins other hyperscalers in adapting serverless platforms for high-performance computing. AWS supports GPU-backed containers through Fargate with ECS or EKS, and Microsoft Azure offers GPU-enabled containers in its Serverless Container Apps. IBM’s Cloud Code Engine is different, the company says, supporting web apps, event-driven functions, and GPU-intensive batch jobs all managed from the one environment.

    Executive takeaway

    For CIOs and Cloud Directors, IBM’s Serverless Fleets represent a step toward the promised elasticity of the cloud and its ability to handle high-performance computing. The model could at least reduce entry barriers for GPU-heavy workloads, especially for teams without readily-available DevOps. However, before adopting, leaders might consider some or all of the following:

    • What are the comparative costs of on-demand GPUs vs. reserved capacity models?
    • Is governance and data security a deciding issue?
    • Are there cost-monitoring methods in place that can keep tabs on managed workloads?
    • Can example workloads be piloted to test scalability and predictability.
    • Is IBM’s offering better/cheaper/worse/more expensive than similar solutions from other hyperscalers?
    • Are workloads suitable for running in-house, and what might be the OPEX in the longer-term of that choice?

    Serverless GPU computing is still evolving, but IBM’s approach offers another option for enterprises to explore large-scale AI and simulation workloads without the overhead of infrastructure considerations.

    (Image source: “Buddha said he wanted to have a word with me” by Trey Ratcliff is licensed under CC BY-NC-SA 2.0.)

    Want to learn more about Cloud Computing from industry leaders? Check out Cyber Security & Cloud Expo taking place in Amsterdam, California, and London. The comprehensive event is part of TechEx and co-located with other leading technology events. Click here for more information.

    CloudTech News is powered by TechForge Media. Explore other upcoming enterprise technology events and webinars here.



    Source link

    Share. Facebook Twitter Pinterest LinkedIn Tumblr Email

    Related Posts

    The data behind the design: How Pantone built agentic AI with an AI-ready database

    February 15, 2026

    AWS Weekly Roundup: Claude Opus 4.6 in Amazon Bedrock, AWS Builder ID Sign in with Apple, and more (February 9, 2026)

    February 14, 2026

    How to ensure your cloud data remains secure with CSPM

    February 13, 2026

    Go 1.26 unleashes performance-boosting Green Tea GC

    February 12, 2026

    Why Healthcare Still Isn’t Ready for AI

    February 11, 2026

    Threat Observability Updates in Secure Firewall 10.0

    February 10, 2026
    Top Posts

    Hard-braking events as indicators of road segment crash risk

    January 14, 202617 Views

    Understanding U-Net Architecture in Deep Learning

    November 25, 202512 Views

    How to integrate a graph database into your RAG pipeline

    February 8, 20268 Views
    Don't Miss

    AI will likely shut down critical infrastructure on its own, no attackers required – Computerworld

    February 15, 2026

    With a new Gartner report suggesting that AI problems will “shut down national critical infrastructure”…

    A new era of humanoid robots

    February 15, 2026

    LLaMA in R with Keras and TensorFlow

    February 15, 2026

    How Deutsche Börse Federates Data Governance with Control

    February 15, 2026
    Stay In Touch
    • Facebook
    • Instagram
    About Us

    At GeekFence, we are a team of tech-enthusiasts, industry watchers and content creators who believe that technology isn’t just about gadgets—it’s about how innovation transforms our lives, work and society. We’ve come together to build a place where readers, thinkers and industry insiders can converge to explore what’s next in tech.

    Our Picks

    AI will likely shut down critical infrastructure on its own, no attackers required – Computerworld

    February 15, 2026

    A new era of humanoid robots

    February 15, 2026

    Subscribe to Updates

    Please enable JavaScript in your browser to complete this form.
    Loading
    • About Us
    • Contact Us
    • Disclaimer
    • Privacy Policy
    • Terms and Conditions
    © 2026 Geekfence.All Rigt Reserved.

    Type above and press Enter to search. Press Esc to cancel.