• Five Steps to Next-Generation Incident Preparedness and Response
      Five Steps to Next-Generation Incident Preparedness and Response
      FEATURED INSIGHT OF THE WEEK

      Five Steps to Next-Generation Incident Preparedness and Response

      Recent disruptions associated with the COVID-19 pandemic have spurred a concerning trend: cyberthreats have grown among 86% of organizations in the U.S., Cybersecurity Dive reports, as well as 63% of companies in other countries.

      8 minute read

      Search Insights & Thought Leadership

      NVIDIA AI Enterprise: A Complete Guide to Scalable AI Deployment

      NVIDIA AI Enterprise: A Complete Guide to Scalable AI Deployment

      NVIDIA AI Enterprise is a comprehensive, full-stack software platform that simplifies AI deployment, ensures security, and delivers consistent performance across hybrid cloud environments. It streamlines the transition from AI experimentation to production by standardising the software stack and reducing compatibility issues. When paired with the NVIDIA H200 NVL GPU, which includes a five-year subscription to AI Enterprise, organisations gain a complete solution for large-scale AI workloads. The H200 significantly boosts inference performance, particularly for large language models and high-performance computing. This combination offers cutting-edge performance, simplified deployment, enterprise-grade support, and flexibility, empowering businesses to focus on AI innovation and achieve measurable outcomes.

      13 minute read

      Datacenter

      Agentic AI and NVIDIA H200: Powering the Next Era of Autonomous Intelligence

      Agentic AI and NVIDIA H200: Powering the Next Era of Autonomous Intelligence

      Agentic AI represents an evolution in artificial intelligence, moving beyond systems that merely respond to prompts. It can autonomously set goals, make decisions, and execute multi-step tasks with minimal human supervision, operating through a "Perceive, Reason, Act, Learn" cycle. This contrasts with Generative AI, which is reactive and primarily creates content based on direct prompts. The NVIDIA H200 GPU is crucial for powering Agentic AI, offering significant hardware advancements. Built on the Hopper architecture, it features HBM3e memory with 141 GB capacity and 4.8 TB/s bandwidth, nearly doubling the memory and boosting bandwidth compared to its predecessor, the H100. These improvements enable the H200 to run larger AI models directly, deliver up to 2x faster inference, and enhance energy efficiency for complex reasoning and planning required by agentic systems. Agentic AI offers benefits for businesses and society, transforming automation, decision-making, and research, but also raises important ethical, accountability, and cybersecurity considerations.

      11 minute read

      Energy and Utilities

      NVIDIA® UFM® Cyber-AI: Transforming Fabric Management for Secure, Intelligent Data Centers

      NVIDIA® UFM® Cyber-AI: Transforming Fabric Management for Secure, Intelligent Data Centers

      The NVIDIA® UFM® Cyber-AI platform is an AI-powered extension of NVIDIA’s Unified Fabric Manager, designed to transform fabric management for secure, intelligent InfiniBand data centres. It moves beyond traditional monitoring by leveraging real-time telemetry and machine learning models to predict and prevent failures. Its three-layer architecture comprises Input Telemetry (gathering vital network metrics), Processing Models (analysing data for anomalies and predictions), and an Output Dashboard (visualising insights and recommendations). UFM® Cyber-AI enhances network reliability, strengthens security by detecting abnormal usage, and improves operational efficiency. Crucially, it integrates with NVIDIA H200 GPUs, which provide the compute power for large-scale, real-time telemetry analysis, creating a synergistic, AI-powered defence loop for resilient infrastructure. Deployment options include dedicated appliances or software containers.

      10 minute read

      Energy and Utilities

      NVIDIA Cybersecurity AI: Using Technology to Fight Modern Threats

      NVIDIA Cybersecurity AI: Using Technology to Fight Modern Threats

      NVIDIA's Cybersecurity AI provides a next-generation defence against modern, AI-driven cyberattacks like sophisticated phishing and ransomware, which surpass traditional, rule-based security systems. AI cybersecurity utilises artificial intelligence and machine learning to detect, predict, and respond to threats in real time, learning from data and adapting without human input. NVIDIA’s end-to-end platform integrates accelerated computing, GPUs, DPUs, and modular AI microservices. Key components include NVIDIA Morpheus for real-time anomaly detection at scale, BlueField DPUs for offloading and accelerating security at the infrastructure level, Confidential Computing to protect data during active processing, NIM Microservices and AI Blueprints for rapid deployment of AI-powered defences, and Agentic AI with NeMo Agents for autonomous monitoring and remediation of security incidents, creating a "security flywheel". This offers intelligent, automated, and scalable security for critical industries.

      13 minute read

      Energy and Utilities

      Mastering LLM Training: Scaling GPU Clusters with NVIDIA H200

      Mastering LLM Training: Scaling GPU Clusters with NVIDIA H200

      Training Large Language Models (LLMs) is an incredibly demanding, computationally intensive task, requiring GPU clusters to process massive datasets and billions of parameters efficiently. GPU clusters, networks of powerful Graphics Processing Units, enable parallel processing, drastically cutting training time from years to weeks or days, making large-scale LLM training practical for enterprises. The NVIDIA H200 GPU is a game-changer, directly tackling the biggest bottlenecks in LLM training. It brings significant upgrades in memory speed and capacity with HBM3e memory (allowing larger data batches) and introduces FP8 precision for faster calculations and reduced memory strain. NVLink 4.0 ensures super-fast GPU communication within clusters, further boosting efficiency. These features combined offer transformative speed and cost savings for businesses. Despite these advancements, challenges remain, including memory bottlenecks, network latency, hardware failures, and software complexity. Overcoming these requires smart strategies like parallelism techniques (data, model, 3D hybrid) and cluster optimisation to ensure enterprise success.

      13 minute read

      Energy and Utilities

      User-Driven Security Practices for NVIDIA DGX H100/H200 Deployment

      User-Driven Security Practices for NVIDIA DGX H100/H200 Deployment

      NVIDIA’s DGX H100/H200 systems offer powerful built-in protections like Hardware Root of Trust and BlueField-3 DPUs, but 68% of breaches still stem from user misconfigurations. This blog introduces “user-driven security,” where administrators actively harden GPU clusters with Secure Boot, TPM attestation, and physical interface lockdown. It dives into Zero Trust architecture for blocking lateral movement, GPU-aware container security to prevent model hijacking, and AI-specific data governance like tokenization and model watermarking. Advanced monitoring via DOCA telemetry and GPU anomaly alerts closes visibility gaps. Finally, it prescribes AI-specific incident response protocols that freeze GPU memory and isolate threats within seconds—preserving evidence and models. For organizations deploying DGX clusters, this layered approach transforms AI infrastructure from vulnerable target to resilient, regulation-aligned fortress.

      17 minute read

      Datacenter

      Unlock Enterprise AI with NVIDIA AI Enterprise Subscription

      Unlock Enterprise AI with NVIDIA AI Enterprise Subscription

      The NVIDIA AI Enterprise Subscription is a comprehensive, cloud-native software suite designed to simplify the deployment, security, and scaling of AI across cloud, data centres, and edge environments. It provides enterprise-grade tools, security, and support for efficient and reliable AI applications. The software is optimised for NVIDIA GPUs, ensuring faster and more efficient AI workloads. It offers robust security, system stability, and broad compatibility with various AI tools, including NIM Microservices, NVIDIA NeMo, and Triton Inference Server. Crucially, the subscription is bundled with NVIDIA H200 GPUs for five years at no extra cost, accelerating time-to-value for AI projects. Flexible licensing options and predictable costs further support confident AI adoption.

      8 minute read

      Datacenter

      NVIDIA DGX Platform: The Engine of Enterprise AI

      NVIDIA DGX Platform: The Engine of Enterprise AI

      The NVIDIA DGX platform is a fully integrated AI supercomputing solution designed for enterprises. It uniquely combines purpose-built hardware, optimised software, and support services into one unified system, delivering turnkey enterprise AI. This platform eliminates the complexity of assembling separate components, allowing businesses to skip months of setup and focus on AI innovation. Key components include DGX servers, scalable DGX SuperPOD clusters, and DGX Cloud for on-demand access. The ecosystem features software like DGX OS and the AI Enterprise Suite, along with managed services and expert support. Enterprises choose DGX for faster deployment, higher performance, lower total cost of ownership, and enhanced security compared to DIY solutions.

      9 minute read

      Energy and Utilities

      How the NVIDIA Container Toolkit Enables GPU Acceleration in Containers

      How the NVIDIA Container Toolkit Enables GPU Acceleration in Containers

      The NVIDIA Container Toolkit is a suite of software tools designed to enable NVIDIA GPU acceleration within containerised applications. Standard container setups often lack built-in GPU support, a problem the Toolkit resolves by providing the necessary libraries and utilities. It offers seamless GPU integration across multiple container runtimes, including Docker, Kubernetes, Podman, and others. The Toolkit automates crucial steps like detecting NVIDIA drivers, mounting GPU devices, and passing environment settings, significantly reducing setup complexity, errors, and saving time for developers and data scientists. It supports CDI (Container Device Interface) for standardised device specification, particularly useful for rootless containers. Essentially, it makes GPU acceleration inside containers both easy and consistent.

      10 minute read

      Datacenter

      Navigating the Tightrope: GDPR Compliance in the Age of AI

      Navigating the Tightrope: GDPR Compliance in the Age of AI

      Artificial Intelligence (AI) presents a significant challenge to General Data Protection Regulation (GDPR) compliance, as AI's need for vast datasets conflicts with GDPR's strict rules on personal data handling. Key friction points include: • Explainability and Transparency: AI's "black box" nature makes it hard to explain decisions as GDPR requires. • Data Minimisation and Purpose Limitation: AI often uses more data than necessary for broad goals. • Accuracy and Bias: Flawed training data can lead to biased AI outputs, impacting individuals. • Lawful Basis: Establishing valid legal grounds for complex AI processing is difficult. Individuals' rights like access, rectification, erasure, and those related to automated decision-making are harder to implement with AI. Organisations can achieve compliance through Privacy by Design, Data Protection Impact Assessments (DPIAs), ensuring meaningful transparency, bias mitigation, and human oversight. The upcoming EU AI Act will further complement GDPR, while Privacy-Enhancing Technologies offer future solutions. Compliance is essential for responsible innovation and public trust.

      11 minute read

      Datacenter

      1–10 of 51 items
      of 6 pages
      uvation