Radar has landed - discover the latest DDoS attack trends. Get ahead, stay protected.Get the report
Under attack?

Products

Solutions

Resources

Partners

Why Gcore

  1. Home
  2. Blog
  3. GPU Acceleration in AI: How Graphics Processing Units Drive Deep Learning

GPU Acceleration in AI: How Graphics Processing Units Drive Deep Learning

  • By Gcore
  • September 19, 2023
  • 9 min read
GPU Acceleration in AI: How Graphics Processing Units Drive Deep Learning

This article discusses how GPUs are shaping a new reality in the hottest subset of AI training: deep learning. We’ll explain the GPU architecture and how it fits with AI workloads, why GPUs are better than CPUs for training deep learning models, and how to choose an optimal GPU configuration.

How GPUs Drive Deep Learning

The key GPU features that power deep learning are its parallel processing capability and, at the foundation of this capability, its core (processor) architecture.

Parallel Processing

Deep learning (DL) relies on matrix calculations, which are performed effectively using the parallel computing that GPUs provide. To understand this interrelationship better, let’s consider a simplified training process of a deep learning model. The model takes the input data, such as images, and has to recognize a specific object in these images using a correlation matrix. The matrix summarizes a data set, identifies patterns, and returns results accordingly: If the object is recognized, the model labels it “true”, otherwise it is labeled “false.” Below is a simplified illustration of this process.

Figure 1. The simplified illustration of the DL training process

An average DL model has billions of parameters, each of which contributes to the size of the matrix weights used in the matrix calculations. Each of the billion parameters must be taken into account, that’s why the true/false recognition process requires running billions of iterations of the same matrix calculations. The iterations are not linked to each other, they are executed in parallel. GPUs are perfect for handling these types of operations because of their parallel processing capabilities. This is enabled by devoting more transistors to data processing.

Core Architecture: Tensor Cores

NVIDIA tensor cores are an example of how hardware architecture can effectively adapt to DL and AI. Tensor cores—special kinds of processors—were designed specifically for the mathematical calculations needed for deep learning, while earlier cores were also used for video rendering and 3D graphics. “Tensor” refers to tensor calculations, which are matrix calculations. A tensor is a mathematical object; if a tensor has two dimensions, it is a matrix. Below is a visualization of how a Tensor core calculates matrices.

Figure 2. Volta Tensor Core matrix calculations. Source: NVIDIA

NVIDIA Volta-based chips, like Tesla V100 with 640 tensor cores, became the first fully AI-focused GPUs, and they significantly influenced and accelerated the DL development industry. NVIDIA added tensor cores to its GPU chips in 2017, based on the Volta architecture.

Multi-GPU Clusters

Another GPU feature that drives DL training is the ability to increase throughput by building multi-GPU clusters, where many GPUs work simultaneously. This is especially useful when training large, scalable DL models with billions and trillions of parameters. The most effective approach for such training is to scale GPUs horizontally using interfaces such as NVLink and InfiniBand. These high-speed interfaces allow GPUs to exchange data directly, bypassing CPU bottlenecks.

Figure 3. NVIDIA H100 with NVLink GPU-to-GPU connections. Source: NVIDIA

For example, with the NVLink switch system, you can connect 256 NVIDIA GPUs in a cluster and get 57.6 Tbps of bandwidth. A cluster of that size can significantly reduce the time needed to train large DL models. Though there are several AI-focused GPU vendors on the market, NVIDIA is the undisputed leader, and makes the greatest contribution to DL. This is one of the reasons why Gcore uses NVIDIA chips for its AI GPU infrastructure.

GPU vs. CPU Comparison

A CPU executes tasks serially. Instructions are completed on a first-in, first-out (FIFO) basis. CPUs are better suited for serial task processing because they can use a single core to execute one task after another. CPUs also have a wider range of possible instructions than GPUs and can perform more tasks. They interact with more computer components such as ROM, RAM, BIOS, and input/output ports.

A GPU performs parallel processing, which means it processes tasks by dividing them between multiple cores. The GPU is a kind of advanced calculator: it can only receive a limited set of instructions and execute only graphics- and AI-related tasks, such as matrix multiplication (CPU can execute them too.) GPUs only need to interact with the display and memory. In the context of parallel computing, this is actually a benefit, as it allows for a greater number of cores devoted solely to these operations. This specialization enhances the GPU’s efficiency in parallel task execution.

An average consumer-grade GPU has hundreds of cores adapted to perform simple operations quickly and in parallel, while an average consumer-grade CPU has 2–16 cores adapted to complex sequential operations. Thus, the GPU is better suited for DL because it provides many more cores to perform the necessary computations faster than the CPU.

Figure 4. An average CPU has 2–16 cores, while an average GPU has hundreds

The parallel processing capabilities of the GPU are made possible by dedicating a larger number of transistors to data processing. Rather than relying on large data caches and complex flow control, GPUs can reduce memory access latencies with computation compared to CPUs. This helps avoid long memory access latencies, frees up more transistors for data processing rather than data caching, and, ultimately, benefits highly parallel computations.

Figure 5. GPUs devote more transistors to data processing than CPUs. Source: NVIDIA

GPUs also use video DRAM: GDDR5 and GDDR6. These are much faster than CPU DRAM: DDR3 and DDR4.

How GPU Outperforms CPU in DL Training

DL requires a lot of data to be transferred between memory and cores. To handle this, GPUs have a specially optimized memory architecture which allows for higher memory bandwidth than CPUs, even when GPUs technically have the same or less memory capacity. For example, a GPU with just 32 GB of HBM (high bandwidth memory) can deliver up to 1.2 Tbps of memory bandwidth and 14 TFLOPS of computing. In contrast, a CPU can have hundreds of GB of HBM, yet deliver only 100 Gbps bandwidth and 1 TFLOPS of computing.

Since GPUs are faster in most DL cases, they can also be cheaper when renting. If you know the approximate time you spend on DL training, you can simply check the prices of cloud providers to estimate how much money you will save by using GPUs instead of CPUs.

Depending on the configuration, models, and frameworks, GPUs often provide better performance than CPUs in DL training. Here are some direct comparisons:

  • Azure tested various cloud CPU and GPU clusters using the TensorFlow and Keras frameworks for five DL models of different sizes. In all cases, GPU cluster throughput consistently outperformed CPU cluster throughput, with improvements ranging from 186% to 804%.
  • Deci compared the NVIDIA Tesla T4 GPU and the Intel Cascade Lake CPU using the EfficientNet-B2 model. They found that the GPU was 3 times faster than the CPU.
  • IEEE published the results of a survey about running different types of neural networks on an Intel i5 9th generation CPU and an NVIDIA GeForce GTX 1650 GPU. When testing CNN (convolutional neural networks,) which are better suited to parallel computation, the GPU was between 4.9 and 8.8 times faster than the CPU. But when testing ANN (artificial neural networks,) the execution time of CPUs was 1.2 times faster than that of GPUs. However, GPUs outperformed CPUs as the data size increased, regardless of the NN architecture.

Using CPU for DL Training

The last comparison case shows that CPUs can sometimes be used for DL training. Here are a few more examples of this:

  • There are CPUs with 128 cores that can process some AI workloads faster than consumer GPUs.
  • Some algorithms allow the optimization DL model to perform better training on CPUs. For instance, Rice’s Brown School of Engineering has introduced an algorithm that makes CPUs 15 times faster than GPUs for some AI tasks.
  • There are cases where the precision of a DL model is not critical, like speech recognition under near-ideal conditions without any noise and interference. In such situations, you can train a DL model using floating-point weights (FP16, FP32) and then round them to integers. Because CPUs work better with integers than GPUs, they can be faster, although the results will not be as accurate.

However, using CPUs for DL training is still an unusual practice. Most DL models are adapted for parallel computing, i.e., for GPU hardware. Thus, building a CPU-based DL platform is a task that may be both difficult and unnecessary. It can take an unpredictable amount of time to select a multi-core CPU instance and then configure a CPU-adapted algorithm to train your model. By selecting a GPU instance, you get a platform that’s ready to build, train, and run your DL model.

How to Choose an Optimal GPU Configuration for Deep Learning

Choosing the optimal GPU configuration is basically a two-step process:

  1. Determine the stage of deep learning you need to execute.
  2. Choose a GPU server specification to match.

Note: We’ll only consider specification criteria for DL training, because DL inference (execution of a trained DL model,) as you’ll see, is not such a big deal as training.

1. Determine Which Stage of Deep Learning You Need

To choose an optimal GPU configuration, first you must understand which of two main stages of DL you will execute on GPUs: DL training, or DL inference. Training is the main challenge of DL, because you have to adjust the huge number (up to trillions) of matrix coefficients (weights.) The process is close to a brute-force search for the best combinations to give the best results (though some techniques help to reduce the number of computations, for example, the stochastic gradient descent algorithm.) Therefore, you need maximum hardware performance for training, and vendors make GPUs specifically designed for this. For example, the NVIDIA A100 and H100 GPUs are positioned as devices for DL training, not for inference.

Once you have calculated all the necessary matrix coefficients, the model is trained and ready for inference. At this stage, a DL model only needs to multiply the input data and the matrix coefficients once to produce a single result—for example, when a text-to-image AI generator generates an image according to a user’s prompt. Therefore, inference is always simpler than training in terms of math computations and required computational resources. In some cases, DL inference can be run on desktop GPUs, CPUs, and smartphones. An example is an iPhone with face recognition: the relatively modest GPU with 4–5 cores is sufficient for DL inference.

2. Choose the GPU Specification for DL Training

When choosing a GPU server or virtual GPU instance for DL training, it’s important to understand what training time is appropriate for you: hours, days, months, etc. To achieve this, you can count operations in the model or use information about reported training time and GPU model performance. Then, decide on the resources you need:

  • Memory size is a key feature. You need to specify at least as much GPU RAM as your DL model size. This is sufficient if you are not pressed for time to market, but if you’re under time pressure then it’s better to specify sufficient memory plus extra in reserve.
  • The number of tensor cores is less critical than the size of the GPU memory, since it only affects the computation speed. However, if you need to train a model faster, then the more cores the better.
  • Memory bandwidth is critical if you need to scale GPUs horizontally, for example, when the training time is too long, the dataset is huge, or the model is highly complex. In such cases, check whether the GPU instances support interconnects, such as NVLink or InfiniBand.

So, memory size is the most important thing when training a DL model: if you don’t have enough memory, you won’t be able to run the training. For example, to run the LLaMA model with 7 billion parameters at full precision, the Hugging Face technical team suggests using 28 GB of GPU RAM. This is the result of multiplying 7×4, where 7 is the tensor size (7B), and 4 is four bits for FP32 (the full-precision format.) For FP16 (half-precision), 14 GB is enough (7×2.) The full-precision format provides greater accuracy. The half-precision format provides less accuracy but makes training faster and more memory efficient.

Kubernetes as a Tool for Improving DL Inference

To improve DL inference, you can containerize your model and use a managed Kubernetes service with GPU instances as worker nodes. This will help you achieve greater scalability, resiliency, and cost savings. With Kubernetes, you can automatically scale resources as needed. For example, if the number of user prompts to your model spikes, you will need more compute resources for inference. In that case, more GPUs are allocated for DL inference only when needed, meaning you have no idle resources and no monetary waste.

Managed Kubernetes also reduces operational overhead and helps to automate cluster maintenance. A provider manages master nodes (the control plane.) You manage only the worker nodes on which you deploy your model, instead focusing on its development.

AI Frameworks that Power Deep Learning on GPUs

Various free, open-source AI frameworks help to train deep neural networks and are specifically designed to be run on GPU instances. All of the following frameworks also support NVIDIA’s Compute Unified Device Architecture (CUDA.) This is a parallel computing platform and API that enables the development of GPU-accelerated applications, including DL models. CUDA can significantly improve their performance.

TensorFlow is a library for ML and AI focused on deep learning model training and inference. With TensorFlow, developers can create dataflow graphs. Each graph node represents a matrix operation, and each connection between nodes is a matrix (tensor.) TensorFlow can be used with several programming languages, including Python, C++, JavaScript, and Java.

PyTorch is a machine-learning framework based on the Torch library. It provides two high-level features: tensor computing with strong acceleration via GPUs, and deep neural networks built on a tape-based auto-differentiation system. PyTorch is considered more flexible than TensorFlow because it gives developers more control over the model architecture.

MXNet is a portable and lightweight DL framework that can be used for DL training and inference not only on GPUs, but also on CPUs and TPUs (Tensor Processing Units.) MXNet supports Python, C++, Scala, R, and Julia.

PaddlePaddle is a powerful, scalable, and flexible framework that, like MXNet, can be used to train and deploy deep neural networks on a variety of devices. PaddlePaddle provides over 500 algorithms and pretrained models to facilitate rapid DL development.

Gcore’s Cloud GPU Infrastructure

As a cloud provider, Gcore offers AI GPU Infrastructure powered by NVIDIA chips:

  • Virtual machines and bare metal servers with consumer- and enterprise-grade GPUs
  • AI clusters based on servers with A100 and H100 GPUs
  • Managed Kubernetes with virtual and physical GPU instances that can be used as worker nodes

With Gcore’s GPU infrastructure, you can train and deploy DL models of any type and size. To learn more about our cloud services and how they can help in your AI journey, contact our team.

Conclusion

The unique design of GPUs, focused on parallelism and efficient matrix operations, makes them the perfect companion for the AI challenges of today and tomorrow, including deep learning. Their profound advantages over CPUs are underscored by their computational efficiency, memory bandwidth, and throughput capabilities.

When seeking a GPU, consider your specific deep learning goals, time, and budget. These help you to choose an optimal GPU configuration.

Book a GPU instance

Related articles

3 clicks, 10 seconds: what real serverless AI inference should look like

Deploying a trained AI model could be the easiest part of the AI lifecycle. After the heavy lifting of data collection, training, and optimization, pushing a model into production is where “the rubber hits the road”, meaning the business expects to see the benefits of invested time and resources. In reality, many AI projects fail in production because of poor performance stemming from suboptimal infrastructure conditions.There are broadly speaking 2 paths developers can take when deploying inference: DIY, which is time and resource-consuming and requires domain expertise from several teams within the business, or opt for the ever-so-popular “serverless inference” solution. The latter is supposed to simplify the task at hand and deliver productivity, cutting down effort to seconds, not hours. Yet most platforms offering “serverless” AI inference still feel anything but effortless. They require containers, configs, and custom scripts. They bury users in infrastructure decisions. And they often assume your data scientists are also DevOps engineers. It’s a far cry from what “serverless” was meant to be.At Gcore, we believe real serverless inference means this: three clicks and ten seconds to deploy a model. That’s not a tagline—it’s the experience we built. And it’s what infrastructure leaders like Mirantis are now enabling for enterprises through partnerships with Gcore.Why deployment UX matters more than you thinkServerless inference isn’t just a backend architecture choice. It’s a business enabler, a go-to-market accelerator, an ROI optimizer, a technology democratizer—or, if poorly executed, a blocker.The reality is that inference workloads are a key point of interface between your AI product or service and the customer. If deployment is clunky, you’re struggling to keep up with demand. If provisioning takes too long, latency spikes, performance is inconsistent, and ultimately your service doesn’t scale. And if the user experience is unclear or inconsistent, customers end up frustrated—or worse, they churn.Developers and data scientists don’t want to manage infrastructure. They want to bring a model and get results without becoming cloud operators in the process.Dom Wilde, SVP Marketing, MirantisThat’s why deployment UX is no longer a nice-to-have. It’s the core of your product.The benchmark: 3 clicks, 10 secondsWe built Gcore Everywhere Inference to remove every unnecessary step between uploading a model and running it in production. That includes GPU provisioning, routing, scaling, isolation, and endpoint generation, all handled behind the scenes.The result is what we believe should be the default:Upload a modelConfirm deployment parametersClick deployAnd within ten seconds, you’re serving live inference.For platform teams supporting AI workloads, this isn’t just a better workflow. It’s a transformation.With Gcore, our customers can deliver not just self-service infrastructure but also inference as a product. End users can deploy models in seconds, and customers don’t have to micromanage the backend to support that.Dom Wilde, MirantisSimple frontend, powerful backendIt’s worth saying: simplifying the frontend doesn’t mean weakening the backend. Gcore’s platform is built for scale and performance, offering the following:Multi-tenant GPU isolationSmart routing based on location and loadAuto-scaling based on demandA unified API and UI for both automation and accessibilityWhat makes this meaningful isn’t just the tech, it’s the way it vanishes behind the scenes. With Gcore, Mirantis customers can deliver low-latency inference, maximize GPU efficiency, and meet data privacy requirements without touching low-level infrastructure.Many enterprises and cloud customers worry about underutilized GPUs. Now, every cycle is optimized. The platform handles the complexity so our customers can focus on building value.Dom Wilde, MirantisIf it’s not 3 clicks and 10 seconds, it’s not really serverlessThere’s a growing gap between what serverless inference promises and what most platforms deliver. Many cloud providers are focused on raw compute or orchestration, but overlook the deployment layer. That’s a mistake. Because when it comes to customer experience, ease of deployment is the product.Mirantis saw that early on and partnered with Gcore to bring inference-as-a-service to CSP and enterprise customers, fast. Now, customers can launch new offerings more quickly, reduce operational overhead, and improve the user experience with a simple, elegant deployment path.Redefine serverless AI with GcoreIf it takes a config file, a container, and a support ticket to deploy a model, it’s not serverless—it’s server-less-ish. With Gcore Everywhere Inference, we’ve set a new benchmark: three clicks and ten seconds to deploy AI. And, our model catalog offers a variety of popular models so you can get started right away.Whether you’re frustrated with slow, inefficient model deployments or looking for the most effective way to start using AI for your company, you need Gcore Everywhere Inference. Give our experts a call to discover how we can simplify your AI so you can focus on scaling and business logic.Let’s talk about your AI project

Run AI inference faster, smarter, and at scale

Training your AI models is only the beginning. The real challenge lies in running them efficiently, securely, and at scale. AI and reality meet in inference—the continuous process of generating predictions in real time. It is the driving force behind virtual assistants, fraud detection, product recommendations, and everything in between. Unlike training, inference doesn’t happen once; it runs continuously. This means that inference is your operational engine rather than just technical infrastructure. And if you don’t manage it well, you’re looking at skyrocketing costs, compliance risks, and frustrating performance bottlenecks. That’s why it’s critical to rethink where and how inference runs in your infrastructure.The hidden cost of AI inferenceWhile training large models often dominates the AI conversation, it’s inference that carries the greatest operational burden. As more models move into production, teams are discovering that traditional, centralized infrastructure isn’t built to support inference at scale.This is particularly evident when:Real-time performance is critical to user experienceRegulatory frameworks require region-specific data processingCompute demand fluctuates unpredictably across time zones and applicationsIf you don’t have a clear plan to manage inference, the performance and impact of your AI initiatives could be undermined. You risk increasing cloud costs, adding latency, and falling out of compliance.The solution: optimize where and how you run inferenceOptimizing AI inference isn’t just about adding more infrastructure—it’s about running models smarter and more strategically. In our new white paper, “How to Optimize AI Inference for Cost, Speed, and Compliance”, we break it down into three key decisions:1. Choose the right stage of the AI lifecycleNot every workload needs a massive training run. Inference is where value is delivered, so focus your resources on where they matter most. Learn when to use pretrained models, when to fine-tune, and when simple inference will do the job.2. Decide where your inference should runFrom the public cloud to on-prem and edge locations, where your model runs, impacts everything from latency to compliance. We show why edge inference is critical for regulated, real-time use cases—and how to deploy it efficiently.3. Match your model and infrastructure to the taskBigger models aren’t always better. We cover how to choose the right model size and infrastructure setup to reduce costs, maintain performance, and meet privacy and security requirements.Who should read itIf you’re responsible for turning AI from proof of concept into production, this guide is for you.Inference is where your choices immediately impact performance, cost, and customer experience, whether you’re managing infrastructure, developing models, or building AI-powered solutions. This white paper will help you cut through complexity and focus on what matters most: running smarter, faster, and more scalable inference.It’s especially relevant if you’re:A machine learning engineer or AI architect deploying models across environmentsA product manager introducing real-time AI featuresA technical leader or decision-maker managing compute, cloud spend, or complianceOr simply trying to scale AI without sacrificing controlIf inference is the next big challenge on your roadmap, this white paper is where to start.Scale AI inference seamlessly with GcoreEfficient, scalable inference is critical to making AI work in production. Whether you’re optimizing for performance, cost, or compliance, you need infrastructure that adapts to real-world demand. Gcore Everywhere Inference brings your models closer to users and data sources—reducing latency, minimizing costs, and supporting region-specific deployments.Our latest white paper, “How to optimize AI inference for cost, speed, and compliance”, breaks down the strategies and technologies that make this possible. From smart model selection to edge deployment and dynamic scaling, you’ll learn how to build an inference pipeline that delivers at scale.Ready to make AI inference faster, smarter, and easier to manage?Download the white paper

Securing vibe coding: balancing speed with cybersecurity

Vibe coding has emerged as a cultural phenomenon in 2025 software development. It’s a style defined by coding on instinct and moving fast, often with the help of AI, rather than following rigid plans. It lets developers skip exhaustive design phases and dive straight into building, writing code (or prompting an AI to write it) in a rapid, conversational loop. It has caught on fast and boasts a dedicated following of developers hosting vibe coding game jams.So why all the buzz? For one, vibe coding delivers speed and spontaneity. Enthusiasts say it frees them to prototype at the speed of thought, without overthinking architecture. A working feature can be blinked into existence after a few AI-assisted prompts, which is intoxicating for startups chasing product-market fit. But as with any trend that favors speed over process, there’s a flip side.This article explores the benefits of vibe coding and the cybersecurity risks it introduces, examines real incidents where "just ship it" coding backfired, and outlines how security leaders can keep up without slowing innovation.The upside: innovation at breakneck speedVibe coding addresses real development needs and has major benefits:Allows lightning-fast prototyping with AI assistance. Speed is a major advantage, especially for startups, and allows faster validation of ideas and product-market fit.Prioritizes creativity over perfection, rewarding flow and iteration over perfection.Lowers barriers to entry for non-experts. AI tooling lowers the skill floor, letting more people code.Produces real success stories, like a game built via vibe coding hitting $1M ARR in 17 days.Vibe coding aligns well with lean, agile, and continuous delivery environments by removing overhead and empowering rapid iteration.When speed bites backVibe coding isn’t inherently insecure, but the culture of speed it promotes can lead to critical oversights, especially when paired with AI tooling and lax process discipline. The following real-world incidents aren’t all examples of vibe coding per se, but they illustrate the kinds of risks that arise when developers prioritize velocity over security, skip reviews, or lean too heavily on AI without safeguards. These three cases show how fast-moving or under-documented development practices can open serious vulnerabilities.xAI API key leak (2025)A developer at Elon Musk’s AI company, xAI, accidentally committed internal API keys to a public GitHub repo. These keys provided access to proprietary LLMs trained on Tesla and SpaceX data. The leak went undetected for two months, exposing critical intellectual property until a researcher reported it. The error likely stemmed from fast-moving development where secrets were hardcoded for convenience.Malicious NPM packages (2024)In January 2024, attackers uploaded npm packages like warbeast2000 and kodiak2k, which exfiltrated SSH keys from developer machines. These were downloaded over 1,600 times before detection. Developers, trusting AI suggestions or searching hastily for functionality, unknowingly included these malicious libraries.OpenAI API key abuse via Replit (2024)Hackers scraped thousands of OpenAI API keys from public Replit projects, which developers had left in plaintext. These keys were abused to access GPT-4 for free, racking up massive bills for unsuspecting users. This incident shows how projects with weak secret hygiene, which is a risk of vibe coding, become easy targets.Securing the vibe: smart risk mitigationCybersecurity teams can enable innovation without compromising safety by following a few simple cybersecurity best practices. While these don’t offer 100% security, they do mitigate many of the major vulnerabilities of vibe coding.Integrate scanning tools: Use SAST, SCA, and secret scanners in CI/CD. Supplement with AI-based code analyzers to assess LLM-generated code.Shift security left: Embed secure-by-default templates and dev-friendly checklists. Make secure SDKs and CLI wrappers easily available.Use guardrails, not gates: Enable runtime protections like WAF, bot filtering, DDoS defense, and rate limiting. Leverage progressive delivery to limit blast radius.Educate, don’t block: Provide lightweight, modular security learning paths for developers. Encourage experimentation in secure sandboxes with audit trails.Consult security experts: Consider outsourcing your cybersecurity to an expert like Gcore to keep your app or AI safe.Secure innovation sustainably with GcoreVibe coding is here to stay, and for good reason. It unlocks creativity and accelerates delivery. But it also invites mistakes that attackers can exploit. Rather than fight the vibe, cybersecurity leaders must adapt: automating protections, partnering with devs, and building a culture where shipping fast doesn't mean shipping insecure.Want to secure your edge-built AI or fast-moving app infrastructure? Gcore’s Edge Security platform offers robust, low-latency protection with next-gen WAAP and DDoS mitigation to help you innovate confidently, even at speed. As AI and security experts, we understand the risks and rewards of vibe coding, and we’re ideally positioned to help you secure your workloads without slowing down development.Into vibe coding? Talk to us about how to keep it secure.

Qwen3 models available now on Gcore Everywhere Inference

We’ve expanded our model library for Gcore Everywhere Inference with three powerful additions from the Qwen3 series. These new models bring advanced reasoning, faster response times, and even better multilingual support, helping you power everything from chatbots and coding tools to complex R&D workloads.With Gcore Everywhere Inference, you can deploy Qwen3 models in just three clicks. Read on to discover what makes Qwen3 special, which Qwen3 model best suits your needs, and how to deploy it with Gcore today.Introducing the new Qwen3 modelsQwen3 is the latest evolution of the Qwen series, featuring both dense and Mixture-of-Experts (MoE) architectures. It introduces dual-mode reasoning, letting you toggle between “thinking” and “non-thinking” modes to balance depth and speed:Thinking mode (enable_thinking=True): The model adds a <think>…</think> block to reason step-by-step before generating the final response. Ideal for tasks like code generation or math where accuracy and logic matter.Non-thinking mode (enable_thinking=False): Skips the reasoning phase to respond faster. Best for straightforward tasks where speed is a priority.Model sizes and use casesWith three new sizes available, you can choose the level of performance required for your use case:Qwen3-14B: A 14B parameter model tuned for responsive, multilingual chat and instruction-following. Fast, versatile, and ready for real-time applications with lightning-fast responses.Qwen3-30B-A3B: Built on the Arch-3 backbone, this 30B model offers advanced reasoning and coding capabilities. It’s ideal for applications that demand deeper understanding and precision while balancing performance. It provides high-quality output with faster inference and better efficiency.Qwen3-32B: The largest Qwen3 model yet, designed for complex, high-performance tasks across reasoning, generation, and multilingual domains. It sets a new standard for what’s achievable with Gcore Everywhere Inference, delivering exceptional results with maximum reasoning power. Ideal for complex computation and generation tasks where every detail matters.ModelArchitectureTotal parametersActive parametersContext lengthBest suited forQwen3-14BDense14B14B128KMultilingual chatbots, instruction-following tasks, and applications requiring strong reasoning capabilities with moderate resource consumption.Qwen3-30B-A3BMoE30B3B128KScenarios requiring advanced reasoning and coding capabilities with efficient resource usage; suitable for real-time applications due to faster inference times.Qwen3-32BDense32B32B128KHigh-performance tasks demanding maximum reasoning power and accuracy; ideal for complex R&D workloads and precision-critical applications.How to deploy Qwen3 models with Gcore in just a few clicksGetting started with Qwen3 on Gcore Everywhere Inference is fast and frictionless. Simply log in to the Gcore Portal, navigate to the AI Inference section, and select your desired Qwen3 model. From there, deployment takes just three clicks—no setup scripts, no GPU wrangling, no DevOps overhead. Check out our docs to discover how it works.Deploying Qwen3 via the Gcore Customer Portal takes just three clicksPrefer to deploy programmatically? Use the Gcore API with your project credentials. We offer quick-start examples in Python and cURL to get you up and running fast.Why choose Qwen3 + Gcore?Flexible performance: Choose from three models tailored to different workloads and cost-performance needs.Immediate availability: All models are live now and deployable via portal or API.Next-gen architecture: Dense and MoE options give you more control over reasoning, speed, and output quality.Scalable by design: Built for production-grade performance across industries and use cases.With the latest Qwen3 additions, Gcore Everywhere Inference continues to deliver on performance, scalability, and choice. Ready to get started? Get a free account today to explore Qwen3 and deploy with Gcore in just a few clicks.Sign up free to deploy Qwen3 today

Run AI workloads faster with our new cloud region in Southern Europe

Good news for businesses operating in Southern Europe! Our newest cloud region in Sines, Portugal, gives you faster, more local access to the infrastructure you need to run advanced AI, ML, and HPC workloads across the Iberian Peninsula and wider region. Sines-2 marks the first region launched in partnership with Northern Data Group, signaling a new chapter in delivering powerful, workload-optimized infrastructure across Europe.Strategically positioned in Portugal, Sines-2 enhances coverage in Southern Europe, providing a lower-latency option for customers operating in or targeting this region. With the explosive growth of AI, machine learning, and compute-intensive workloads, this new region is designed to meet escalating demand with cutting-edge GPU and storage capabilities.Built for AI, designed to scaleSines-2 brings with it next-generation infrastructure features, purpose-built for today’s most demanding workloads:NVIDIA H100 GPUs: Unlock the full potential of AI/ML training, high-performance computing (HPC), and rendering workloads with access to H100 GPUs.VAST NFS (file sharing protocol) support: Benefit from scalable, high-throughput file storage ideal for data-intensive operations, research, and real-time AI workflows.IaaS portfolio: Deploy Virtual Machines, manage storage, and scale infrastructure with the same consistency and reliability as in our flagship regions.Organizations operating in Portugal, Spain, and nearby regions can now deploy workloads closer to end users, improving application performance. For finance, healthcare, public sector, and other organisations running sensitive workloads that must stay within a country or region, Sines-2 is an easy way to access state-of-the-art GPUs with simplified compliance. Whether you're building AI models, running simulations, or managing rendering pipelines, Sines-2 offers the performance and proximity you need.And best of all, servers are available and ready to deploy today.Run your AI workloads in Portugal todayWith Sines-2 and our partnership with Northern Data Group, we’re making it easier than ever for you to run AI workloads at scale. If you need speed, flexibility, and global reach, we’re ready to power your next AI breakthrough.Unlock the power of Sines-2 today

How AI is transforming gaming experiences

AI is reshaping how games are played, built, and experienced. Although we are in a period of flux where the optimal combination of human and artificial intelligence is still being ironed out, the potential for AI to greatly enhance both gameplay and development is clear.PlayStation CEO Hermen Hulst recently emphasized the importance of striking the right balance between the handcrafted human touch and the revolutionary advances that AI brings. AI will not replace the decades of design, storytelling, and craft laid down by humans—it will build on that foundation to unlock entirely new possibilities. In addition to an enhanced playing experience, AI is shaking up gaming aspects such as real-time analytics, player interactions, content generation, and security.In this article, we explore three specific developments that are enriching gaming storyworlds, as well as the technology that’s bringing them to life and what the future might hold.#1 Responsive NPC behavior and smarter opponentsAI is evolving to create more realistic, adaptive, and intelligent non-player characters (NPCs) that can react to individual player choices with greater depth and reasoning. The algorithms allow NPCs to respond dynamically to players’ decisions so they can adjust their strategies and behaviors in real time. This provides a more immersive and dynamic gameplay environment and means gamers have endless opportunities to experience new gaming adventures and write their own story every time.A recent example is Red Dead Redemption 2, which enables players to interact with NPCs in the Wild West. Players were impressed by its complexity and the ability to interact with fellow cowboys and bar patrons. Although this is limited for now, eventually, it could become like a video game version of the TV series Westworld, in which visitors pay to interact with incredibly lifelike robots in a Wild West theme park.AI also gives in-game opponents more “agency,” making them more reactive and challenging for players to defeat. This means smarter, more unpredictable enemies who provide a heightened level of achievement, novelty, and excitement for players.For example, AI Limit, released in early 2025, is an action RPG incorporating AI-driven combat mechanics. While drawing comparisons to Soulslike games, the developers emphasize its unique features, including the Sync Rate system, which adds depth to combat interactions.#2 AI-assisted matchmaking and player behavior predictionsAI-powered analytics can identify and predict player skill levels and playing styles, leading to more balanced and competitive matchmaking. A notable example is the implementation of advanced matchmaking systems in competitive games such as Apex Legends and Call of Duty: Modern Warfare III. These titles use AI algorithms to analyze not just skill levels but also playstyle preferences, weapon selections, and playing patterns to create matches optimized for player retention and satisfaction. The systems continuously learn from match outcomes to predict player behavior and create more balanced team compositions across different skill levels.By analyzing a player’s past performance, AI can also create smarter team formations. This makes for fairer and more rewarding multiplayer games, as players are matched with others who complement their skill and strategy.AI can monitor in-game interactions to detect and mitigate toxic behavior. This helps encourage positive social dynamics and foster a more collaborative and friendly online environment.#3 Personalized gaming experiencesMultiplayer games can use AI to analyze player behavior in real time, adjusting difficulty levels and suggesting tailored missions, providing rich experiences unique to each player. This creates personalized, player-centric gameplay that evolves dynamically and can change over time as a player’s knowledge and ability improve.Games like Minecraft and Skyrim already use AI to adjust difficulty and offer dynamic content, while Oasis represents a breakthrough as an AI-generated Minecraft-inspired world. The game uses generative AI to predict and render gameplay frames in real time, creating a uniquely responsive environment.Beyond world generation, modern games are also incorporating AI chatbots that give players real-time coaching and personalized skill development tips.How will AI continue to shape gaming?In the future, AI will continue to impact not just the player experience but also the creation of games. We anticipate AI revolutionizing game development in the following areas:Procedural content generation: AI will create vast, dynamic worlds or generate storylines, allowing for more expansive and diverse game worlds than are currently available.Game testing: AI will simulate millions of player interactions to help developers find bugs and improve gameplay.Art and sound design: AI tools will be used to a greater extent than at present to create game art, music, and voiceovers.How Gcore technology is powering AI gaming innovationIn terms of the technology behind the scenes, Gcore Everywhere Inference brings AI models closer to players by deploying them at the edge, significantly reducing latency for training and inference. This powers dynamic features like adaptive NPC behavior, personalized gameplay, and predictive matchmaking without sacrificing performance.Gcore technology differentiates itself with the following features:Supports all major frameworks, including PyTorch, TensorFlow, ONNX, and Hugging Face Transformers, making deploying your preferred model architecture easy.Offers multiple deployment modes, whether in the cloud, on-premise, or across our distributed edge network with 180+ global locations, allowing you to place inference wherever it delivers the best performance for your users.Delivers sub-50ms latency for inference workloads in most regions, even during peak gaming hours, thanks to our ultra-low-latency CDN and proximity to players.Scales horizontally, so studios can support millions of concurrent inferences for dynamic NPC behavior, matchmaking decisions, or in-game voice/chat moderation, without compromising gameplay speed.Keeps your models and training data private through confidential computing and data sovereignty controls, helping you meet compliance requirements across regions including Europe, LATAM, and MENA.With a low-latency infrastructure that supports popular AI frameworks, Gcore Everywhere Inference allows your studio to deploy custom models and deliver more immersive, responsive player experiences at scale. With our confidential computing solutions, you retain full control over your training assets—no data is shared, exposed, or compromised.Deliver next-gen gaming with Gcore AIAI continues to revolutionize industries, and gaming is no exception. The deployment of artificial intelligence can help make games even more exciting for players, as well as enabling developers to work smarter when creating new games.At Gcore, AI is our core and gaming is our foundation. AI is seamlessly integrated into all our solutions with one goal in mind: to help grow your business. As AI continues to evolve rapidly, we're committed to staying at the cutting edge and changing with the future. Contact us today to discover how Everywhere Inference can enhance your gaming offerings.Get a customized consultation about AI gaming deployment

Subscribe to our newsletter

Get the latest industry trends, exclusive insights, and Gcore updates delivered straight to your inbox.