Category Archives: AI and ML

CVE-2026-2242: Vulnerabilities in janet-lang may also affect partner devices. (11th Feb 2026)

Preface: Janet has a small footprint: it fits in environments where 2.5 MB of RAM is considered “plenty”. While Janet does not run on the GPU (it is a CPU-bound language), it is often used as the control/orchestration layer on heterogeneous AI platforms. Because Janet is written in C and compiles to a small binary (roughly 200–300 KB), it is frequently used on ARM Cortex-based systems. While Janet does not run on the GPU (it is a CPU-bound language), it is often used as the control/orchestration layer on heterogeneous AI platforms.

Background: Janet can be used to manage the data pipeline, calling into C/C++ libraries that handle heavy GPU lifting via CUDA. If code allow external scripts or users to submit code dynamically, it can use Janet’s built-in eval-string function. Does it vulnerable to CVE-2026-2242.

My speculation: Using eval-string does expose your Jetson pipeline to CVE‑2026‑2242, because:

CVE‑2026‑2242 is triggered during compilation of Janet code, and eval-string compiles code dynamically. If a malicious user submits a specially-crafted Janet expression that enters the vulnerable path inside:

janetc_if  →  specials[.]c

then Janet may perform an out‑of‑bounds read, which can cause:

  • interpreter crash
  • denial of service
  • undefined behavior inside the Janet process
Even though the CVE requires “local execution,”

-allowing remote users to submit code and then calling eval-string makes that local execution possible.

Therefore, theJetson pipeline becomes exploitable.

Vulnerability details: A vulnerability was determined in janet-lang janet up to 1.40.1. This impacts the function janetc_if of the file src/core/specials.c. Executing a manipulation can lead to out-of-bounds read. The attack needs to be launched locally. The exploit has been publicly disclosed and may be utilized. This patch is called c43e06672cd9dacf2122c99f362120a17c34b391. It is advisable to implement a patch to correct this issue.

Official announcement: Please refer to the link for more details –

https://www.tenable.com/cve/CVE-2026-2242

CVE-2026-25592: An Arbitrary File Write vulnerability has been identified in Microsoft’s Semantic Kernel .NET SDK (10th Feb 2026)

Preface: Microsoft has launched several products (Semantic Kernel, Microsoft.Extensions.AI, and Azure.OpenAI), which initially caused confusion for developers. Furthermore, the Semantic Kernel is currently being “upgraded” to the new Microsoft Agent Framework, leading some developers to question future support for the Semantic Kernel.

Python, with its rich libraries and large open-source community, remains the “universal language” in artificial intelligence research and data science. LangChain, as the primary alternative, is also based on Python. Python remains dominant in the field of artificial intelligence/machine learning.

Background: Microsoft developed Semantic Kernel as an open-source SDK to bridge conventional programming languages (C#, Python, Java) with advanced LLMs, enabling developers to build enterprise-grade, agentic AI applications. It simplifies orchestrating complex AI workflows, allows swapping models without rewriting code, and ensures secure, compliant integration of AI with existing systems.

Semantic Kernel uses plugins (formerly skills or functions) to extend its capabilities beyond its core prompt engineering functionality and integrate with external services, data sources, and API.

  • Semantic Kernel acts as an orchestrator, using the LLM to decide which plugins to use and when, effectively automating complex tasks that involve multiple steps and tools. The LLM determines the necessary sequence of actions to fulfill a user’s request.
  • Plugins allow the LLM to interact with real-world applications and data. For example, a plugin could retrieve real-time weather information, search a database, book a flight, or send an email.

The airline and travel industry is beginning to use Microsoft Semantic Kernel to build intelligent, AI-powered applications, particularly for automating customer service and booking processes. Developers are using Semantic Kernel to build dialogue agents that can understand complex booking instructions, such as “book the cheapest flight from Hong Kong to Tokyo”, and handle the booking process independently.

Vulnerability details: CVE-2026-25592 Semantic Kernel is an SDK used to build, orchestrate, and deploy AI agents and multi-agent systems. Prior to 1.70.0, an Arbitrary File Write vulnerability has been identified in Microsoft’s Semantic Kernel .NET SDK, specifically within the SessionsPythonPlugin. The problem has been fixed in Microsoft.SemanticKernel.Core version 1.70.0. As a mitigation, users can create a Function Invocation Filter which checks the arguments being passed to any calls to DownloadFileAsync or UploadFileAsync and ensures the provided localFilePath is allow listed.

Official announcement: Please refer to the link for details –https://nvd.nist.gov/vuln/detail/CVE-2026-25592

CVE-2025-33220 only applies to NVIDIA vGPU deployments running on hypervisors, such as TKGI clusters on vSphere. (2 Feb 2026)

Preface: When comparing VMware TKGI, Docker, and Kubernetes (K8s) for CUDA (NVIDIA’s parallel computing platform) workflows, the “best” choice depends on your scale and infrastructure.

Choose Docker – if you are a data scientist doing local model development.

Choose Native Kubernetes – if you are building a large-scale AI platform on physical hardware (Bare-metal) for maximum performance.

Choose VMware TKGI – if you need high availability, vGPU flexibility, and are already heavily invested in the VMware ecosystem.

Background: CVE‑2025‑33220 lives in the hypervisor’s vGPU Manager, not in:

  • Docker
  • Containerd
  • Kubernetes
  • NVIDIA Container Runtime
  • NVIDIA Docker runtime
  • PyTorch/TensorFlow workloads
  • CUDA libraries inside containers

CVE‑2025‑33220 requires:

  1. Freeing an object inside the hypervisor
  2. A later operation accessing that SAME freed internal heap structure
  3. The hypervisor NOT realizing the handle is stale
  4. A malformed RM object relationship or command sequence
  5. Conditions normal CUDA applications never generate

If there is no hypervisor-based vGPU, there is no attack surface, because:

  • The ioctl path stops at the bare‑metal NVIDIA GPU driver
  • There is no vGPU Manager backend
  • No vGPU protocol messages are generated
  • No hypervisor memory structures exist to exploit

The CVE is triggered only under very specific hypervisor‑internal states that normal or even “weird order” RMAPI usage will never produce.

Vulnerability details: CVE-2025-33220 – NVIDIA vGPU software contains a vulnerability in the Virtual GPU Manager, where a malicious guest could cause heap memory access after the memory is freed. A successful exploit of this vulnerability might lead to code execution, escalation of privileges, data tampering, denial of service, or information disclosure.

Official announcement: Please refer to the link for details.- https://nvidia.custhelp.com/app/answers/detail/a_id/5747

In-depth analysis of CVE-2025-33234: NVIDIA runx contains this vulnerability. (30-01-2026)

Preface: According to the NVIDIA security bulletin for runx updated January 23, 2026 – A vulnerability (CVE‑2025‑33234) was found “during End of Support, but prior to End of Life.”

Background: An ML hyperparameter sweep is the automated process of systematically testing different combinations of hyperparameter values for a machine learning model to identify the configuration that yields the best performance. Hyperparameters are external settings that control the learning process, such as the learning rate or number of hidden layers, and must be set before training begins.

NVIDIA runx (Deep Learning Library) – This is a lightweight Python tool used for experiment management and hyperparameter sweeps.

Vulnerability details: CVE-2025-33234 NVIDIA runx contains a vulnerability where an attacker could cause a code injection. A successful exploit of this vulnerability might lead to code execution, denial of service, escalation of privileges, information disclosure, and data tampering.

Official announcement: Please refer to the link for details – https://nvidia.custhelp.com/app/answers/detail/a_id/5764

Appendix: runx is a local Python script for organizing your personal files, while NVIDIA Run:ai is a cloud-based infrastructure for managing hardware. To move from a local developer environment using runx to an enterprise environment using Run:ai, you must transition from a “local script” workflow to a “containerized” workflow.

Realistic examples where Run:ai WOULD be impacted (conceptual, not exploit details)

Scenario 1

runx injects malicious commands into a Python file your training code imports →
you copy the folder into Docker →
Run:ai runs that code →
the malicious code executes.

Scenario 2

runx generates a poisoned submit_cmd[.]sh or config file →
you include it in the Docker image →
your entrypoint or tools accidentally run it.

Scenario 3

runx modifies your dataset or preprocessing script →
container uses that script →
it runs inside Run:ai.

Reminder: These examples show pollution can travel only if you physically package it into the container.

CVE-2025-13952: Advanced driver assistance systems (ADAS) and smartphones stay alert! (27-1-2026)

Preface: The Unified Shading Cluster (USC) is a fundamental part of the PowerVR Rogue architecture and subsequent series. The USC acts as the central shader core, where vertex, fragment, and compute tasks are executed on unified hardware. It is central to PowerVR’s Tile-Based Deferred Rendering (TBDR) architecture. To write shader code for Imagination Technologies’ PowerVR architectures (often referred to in technical documentation alongside the USC or “Universal Shading Cluster” and Volcano shader cores), you use the OpenGL ES Shading Language (GLSL ES). 

Background: In the context of Imagination Technologies (IMG) GPUs, libusc (Unified Shading Cluster library) is a critical component of the GPU Driver Development Kit (DDK), primarily serving as the back-end compiler library for shader programs.

Is it possible to combine [.]frag in html web page?

The most common way is to place the shader code inside a <script> tag with a custom type attribute, such as x-shader/x-fragment. Browsers do not execute these scripts because they don’t recognize the type, but the text remains accessible via the Document Object Model (DOM).

The design flaw mentioned in CVE-2025-13952 typically occurs if:

•         User-controlled input is inserted into these <script> tags without sanitization.

•         Your code later injects shader text back into the DOM using innerHTML or evaluates it as JavaScript.

Vulnerability details: A web page that contains unusual GPU shader code is loaded from the Internet into the GPU compiler process triggers a write use-after-free crash in the GPU shader compiler library. On certain platforms, when the compiler process has system privileges this could enable further exploits on the device. The shader code contained in the web page executes a path in the compiler that held onto an out of date pointer, pointing to a freed memory object.

Official announcement: Please refer to the link for details.

https://www.tenable.com/cve/CVE-2025-13952

CVE-2025-33228: About NVIDIA Nsight Systems (23rd Jan 2026)

Preface: Nsight Systems is a tool for developers who need to understand the big picture of application execution on heterogeneous systems, especially in scenarios involving data transfer bottlenecks between the CPU and GPU or scaling across multiple nodes.

Background: In NVIDIA Nsight Systems, process_nsys_rep_cli[.]py is an internal Python script used primarily for post-processing and report generation from raw profiling data. 

While users typically interact with the nsys command-line tool, this script is invoked behind the scenes during the following operations.

Why This Might Connect to Nsight Systems?

Nsight Systems allows exporting [.]nsys-rep files and then processing them with scripts like process_nsys_rep_cli[.]py.

If the CLI or scripts read commands or code from user-provided files without validation, it could lead to:

-Command injection (similar to os[.]system()).

-Code execution (similar to exec()).

The design flaw could be that Nsight Systems assumes [.]nsys-rep or related files are safe, but if an attacker crafts a malicious file and you run the processing script, it could execute harmful commands.

Vulnerability details: CVE-2025-33228 NVIDIA Nsight Systems contains a vulnerability in the gfx_hotspot recipe, where an attacker could cause an OS command injection by supplying a malicious string to the process_nsys_rep_cli[.]py script if the script is invoked manually. A successful exploit of this vulnerability might lead to code execution, escalation of privileges, data tampering, denial of service, and information disclosure.

Official announcement: Please refer to the link for details –

https://nvidia.custhelp.com/app/answers/detail/a_id/5755

CVE-2025-33233 – About NVIDIA Merlin Transformers4Rec for all platform  22nd Jan 2026

Official Updated 01/20/2026

Preface: Data engineers perform seamless preprocessing, a foundational stage where they gather messy, raw data from diverse sources, clean it (handling missing values, outliers, inconsistencies), integrate disparate datasets, and transform it into a unified, structured format, making it ready and reliable for data scientists to perform advanced feature engineering (creating new, meaningful features) and ultimately build better machine learning models. This ensures a high-quality, consistent input, preventing “garbage in, garbage out” for the modeling phase.

Background: Transformers4Rec is pre-installed in the merlin-pytorch container available from the NVIDIA GPU Cloud (NGC) catalog. This container is part of the NVIDIA Merlin ecosystem and is specifically designed to support sequential and session-based recommendation tasks using PyTorch.

The workflow can show you where we speculated design weakness of CVE-2025-33233.

NVTabular for preprocessingPyTorch for trainingTriton for serving—means PyTorch is a critical component. If its loading function is insecure, Merlin’s container is exposed regardless of NVIDIA’s own code. The workflow can display the location of suspected design flaw CVE-2025-33233.

If Transformers4Rec internally uses torch.load (which is common for loading PyTorch models) and relies on weights_only=True for safety, then CVE-2025-32434 could be the root cause or at least a contributing factor.

NVIDIA might have classified it as a separate CVE because the exploit path involves their product’s integration with PyTorch, making it a product-level exposure rather than just a dependency issue.

Vulnerability details: CVE-2025-33233 NVIDIA Merlin Transformers4Rec for all platforms contains a vulnerability where an attacker could cause a code injection issue. A successful exploit of this vulnerability might lead to code execution, escalation of privileges, information disclosure, and data tampering.

Official announcement: Please refer to the following link for details-

https://nvidia.custhelp.com/app/answers/detail/a_id/5761

CVE-2026-21452: About MessagePack for Java (7th Jan 2026)

Preface: Aerospike is a specific, high-performance NoSQL database, and benchmarks generally show it to be significantly faster than many other clustered NoSQL solutions like Cassandra and MongoDB.

The term “NoSQL” refers to a broad category of databases with varying performance characteristics, so a direct comparison is more nuanced than a simple yes/no answer.

Aerospike uses MessagePack as its default, internal serialization format for Lists and Maps (Collection Data Types or CDTs); it is not an optional configuration you need to enable in the core database itself.

Background: MessagePack is a compact binary serialization format designed to be more memory-efficient than text-based formats like JSON. For the Java implementation, its memory requirements depend on whether you are using the standard heap-based process or advanced off-heap optimizations.

The MessagePack serialization process primarily utilizes JVM Virtual Memory, which encompasses several different pools:

JVM Heap Memory, Off-Heap / Native Memory and OS Page Cache.

About EXT32?

•         In binary serialization formats (like Mashpack), EXT32 is a type identifier (byte 0xDD) indicating a subsequent 32-bit binary block or extension.

•         It’s used for efficiency, compacting data better than text formats (JSON, XML) by representing data directly as bytes.

Serialized EXT32 objects can require more memory in the JVM heap, primarily due to how standard Java MessagePack libraries manage large payloads during deserialization. While the MessagePack format itself is compact, the serialization and deserialization process in Java introduces specific memory overheads for the EXT32 type:

Large Payload Buffering (Heap Exhaustion) EXT32 is designed for large extension data, supporting payloads up to 4 GiB in size.

Vulnerability details: A known issue in msgpack-java (prior to v0.9.11) was that the library would trust the declared length in the EXT32 header and immediately attempt to allocate a matching byte array on the JVM heap.

Impact: If an EXT32 object declares a massive size, it can trigger rapid heap exhaustion or an OutOfMemoryError before the data is even fully read.

Official announcement: Please refer to the link for details.

https://www.tenable.com/cve/CVE-2026-21452

About 3rd part design weakness impact Intel® Xeon® 6 Processors with P-cores with Intel® TDX Connect (29-12-2025)

Last revised: 12/09/2025

Preface: Intel’s Xeon 6 processors represent a fascinating shift in the landscape of data center computing, moving toward a hybrid architecture that optimizes for different workloads with specialized cores. The P-core version, codenamed Granite Rapids, built entirely of Performance-cores for heavy compute and AI workloads, is accurate and highlights a significant technological leap in server processing capabilities. This new generation aims to deliver unprecedented performance and efficiency to meet the increasing demands of modern data centers, which are grappling with massive data volumes and the computational intensity of artificial intelligence.

Background: Intel® TDX Connect is specifically highlighted as a key feature on Intel® Xeon® 6 Processors with P-cores (Performance-cores) to enable confidential computing for connected devices like GPUs. Intel’s P6 architecture, as part of modern high-speed systems using PCI Express (PCIe), relies on SERDES (Serializer/Deserializer) technology, especially for PCIe 3.0 and newer, to handle high data rates through serial links, though P6 itself refers to older processor generations, the concept of using SERDES for high-speed I/O like PCIe is fundamental, with newer Intel CPUs using advanced SerDes for PCIe 4.0, 5.0, and 6.0 to achieve massive bandwidth for AI and data centers.

Does the Intel P6 use PCIe SERDES?

Yes, Intel’s P6 architecture, as part of modern high-speed systems using PCI Express (PCIe), relies on SERDES (Serializer/Deserializer) technology, especially for PCIe 3.0 and newer, to handle high data rates through serial links, though P6 itself refers to older processor generations, the concept of using SERDES for high-speed I/O like PCIe is fundamental, with newer Intel CPUs using advanced SerDes for PCIe 4.0, 5.0, and 6.0 to achieve massive bandwidth for AI and data centers.

Vulnerability details: [CVE-2025-9612] Improper validation of integrity check value in PCI Port for some Intel® platforms with Integrity and Data Encryption (IDE) for PCIe Base Specification Revision 5 or higher within Ring 0: Bare Metal OS may allow an information disclosure and escalation of privilege. System software adversary with a privileged user combined with a high complexity attack may enable escalation of privilege. This result may potentially occur via local access when attack requirements are present without special internal knowledge and requires no user interaction. The potential vulnerability may impact the confidentiality (low), integrity (low) and availability (none) of the vulnerable system, resulting in subsequent system confidentiality (none), integrity (none) and availability (none) impacts.

Official announcement: Please refer to the link for details –

https://www.intel.com/content/www/us/en/security-center/advisory/intel-sa-01409.html

CVE-2025-33223: NVIDIA Isaac Launchable contains a vulnerability (29th Dec 2025)

Official Updated 12/22/2025 09:21 AM

Preface: The ability to launch NVIDIA Isaac Lab via NVIDIA Brev (Cloud) is fundamentally driven by the need to democratize access to high-performance robotics simulation and AI development environments, circumventing significant hardware and setup barriers. This collaboration between Isaac Lab and Brev offers a streamlined, low-friction pathway for developers and researchers to leverage powerful, preconfigured GPU resources in the cloud.

Background: Isaac Lab requires a compatible version of Isaac Sim to run. An “Isaac Lab Launchable” is an installation option, such as via NVIDIA Brev (Cloud), to quickly get the environment running. The Launchable provides the correct Isaac Sim/Python setup, but you still use env_config[.]yaml within your scripts to define what runs on that platform.

In essence, Issac Lab use env_config[.]yaml to specify tasks (like Isaac-Ant-v0) within your Python training scripts (e.g., train[.]py)The environment command

isaaclab/scripts/reinforcement_learning/skrl/train[.]py –task=Isaac-Ant-v0 specifically targets the Isaac-Ant-v0 task. If train[.]py or related scripts dynamically construct shell commands from these inputs without validation, that’s a classic command injection risk.

Vulnerability details: CVE-2025-33223 – NVIDIA Isaac Launchable contains a vulnerability where an attacker could cause an execution with unnecessary privileges. A successful exploit of this vulnerability might lead to code execution, escalation of privileges, denial of service, information disclosure and data tampering.

Official announcement: Please refer to the link for details –

https://nvidia.custhelp.com/app/answers/detail/a_id/5749