What Is Confidential Computing?
Confidential computing is a hardware-based security paradigm that protects data while it is being processed. Unlike traditional encryption that only covers data at rest (on disk) and in transit (over the network), confidential computing creates a Trusted Execution Environment (TEE) that encrypts data in use — in memory, in GPU VRAM, and across the PCIe bus.
The core idea is simple but revolutionary: even the infrastructure operator cannot access your data. The hypervisor, the host OS, other tenants, and even a physical attacker with access to the server cannot read the contents of a TEE. This is enforced by the CPU and memory controller hardware itself, not by software policies.
Intel TDX: Trust Domain Extensions
Intel Trust Domain Extensions (TDX) is Intel's latest confidential computing technology, available on 4th and 5th Gen Xeon Scalable processors. TDX creates isolated Virtual Machines called Trust Domains (TDs) that are cryptographically isolated from the host, the VMM (Virtual Machine Monitor), and other TDs.
Key capabilities of Intel TDX for GPU workloads:
- Memory encryption with integrity: All TD memory is encrypted using AES-256-XTS with per-TD keys managed by the CPU. Memory integrity checking prevents physical tampering.
- Secure PCIe (TEE-IO): Data flowing between CPU and GPU over PCIe is encrypted, preventing interception. This is critical for GPU workloads where tensor data constantly moves between host and device memory.
- Remote attestation: Before sending sensitive data, clients can cryptographically verify that the TD is running the expected software stack, has not been tampered with, and is on genuine Intel hardware.
- Sealed storage: Encryption keys are bound to the specific TD measurement, so data encrypted by one enclave cannot be decrypted by another, even on the same physical machine.
Why Confidential GPUs Matter for AI
The AI industry has a dirty secret: most GPU cloud providers can technically access your data. When you send a prompt to an API or fine-tune a model on rented GPUs, your data sits in plaintext in GPU VRAM. The provider's sysadmins, a compromised hypervisor, or a nation-state actor with physical access could potentially read it.
For many workloads, this is an acceptable risk. But for regulated industries, it is a dealbreaker:
Healthcare (HIPAA)
Hospitals and health-tech companies need to run AI on Protected Health Information (PHI) — medical images, patient records, genomic data. HIPAA requires a Business Associate Agreement (BAA) with any service that processes PHI, and most GPU cloud providers either refuse to sign a BAA or charge a massive premium. With TDX, the provider never has access to PHI, simplifying compliance dramatically.
Financial Services
Banks and hedge funds use LLMs on proprietary trading strategies, customer financial data, and fraud detection models. SEC and PCI-DSS regulations require strict data handling. Confidential GPUs allow firms to run AI inference in the cloud without exposing sensitive financial data to the infrastructure provider.
Legal and Government
Law firms processing attorney-client privileged documents and government agencies handling classified or sensitive data need guarantees that extend beyond software-level access controls. TDX provides hardware-enforced isolation that satisfies FedRAMP and ITAR requirements.
AI Model IP Protection
Companies with proprietary fine-tuned models worth millions in R&D investment can deploy them on confidential GPUs, ensuring that model weights are never exposed in plaintext to the infrastructure provider. This enables deploying proprietary models on third-party infrastructure without risk of theft.
How VoltageGPU Implements Confidential Compute
VoltageGPU leverages Targon Subnet 4 on the Bittensor network to provide confidential GPU compute. Here is how the architecture works:
- TDX-enabled bare metal: Miners on Targon SN4 provision servers with Intel Xeon Scalable (Sapphire Rapids or Emerald Rapids) CPUs and NVIDIA GPUs. The BIOS is configured for TDX with secure boot enabled.
- TD provisioning: When you deploy a confidential pod, VoltageGPU instructs the miner to create a Trust Domain with your specified GPU passthrough. The TD is measured and the measurement is recorded.
- Attestation flow: Before your workload starts, VoltageGPU's attestation service verifies the TD quote against Intel's Provisioning Certification Service (PCS). You receive a signed attestation report that you can independently verify.
- Encrypted channels: All communication with the TD (SSH, API calls, data transfer) uses TLS 1.3 terminating inside the enclave. The host system only sees encrypted traffic.
Available GPUs in TDX Enclaves
Deploy a Confidential Pod via API
Deploying a confidential GPU pod on VoltageGPU takes one API call. The confidential: true flag ensures your pod runs inside a TDX Trust Domain with full attestation.
curl -X POST \
https://api.voltagegpu.com/v1/pods/deploy \
-H "Authorization: Bearer YOUR_API_KEY" \
-H "Content-Type: application/json" \
-d '{
"gpu_type": "H100-SXM5-80GB",
"gpu_count": 1,
"confidential": true,
"tdx_attestation": true,
"image": "nvcr.io/nvidia/pytorch:24.03-py3",
"env": {
"CONFIDENTIAL_MODE": "enabled",
"ATTESTATION_VERIFY": "true"
},
"ssh_key": "ssh-ed25519 AAAA..."
}'Once deployed, you can verify the attestation report programmatically:
import requests
import json
# Verify TDX attestation report
API_URL = "https://api.voltagegpu.com/v1/pods/POD_ID/attestation"
headers = {"Authorization": "Bearer YOUR_API_KEY"}
response = requests.get(API_URL, headers=headers)
report = response.json()
# Verify the quote
assert report["tdx_version"] == "1.5"
assert report["measurement_valid"] == True
assert report["enclave_sealed"] == True
print(f"TDX Quote: {report['quote_hex'][:64]}...")
print(f"MRENCLAVE: {report['mr_enclave']}")
print(f"Platform: {report['platform_tcb']}")Use Case: HIPAA-Compliant LLM Inference
Here is a real-world scenario: a telehealth startup needs to run Llama 3.1 70B on patient conversations for clinical summarization. They need HIPAA compliance, which means PHI must never be accessible to the infrastructure provider.
from openai import OpenAI
# All traffic encrypted end-to-end via TDX enclave
client = OpenAI(
base_url="https://api.voltagegpu.com/v1",
api_key="YOUR_API_KEY"
)
# Your prompt never leaves the encrypted enclave
response = client.chat.completions.create(
model="meta-llama/Llama-3.1-70B-Instruct",
messages=[
{"role": "system", "content": "You are a HIPAA-compliant medical AI."},
{"role": "user", "content": "Analyze this patient record: [ENCRYPTED_PHI]"}
],
temperature=0.3,
max_tokens=2048,
# Confidential mode: data never persisted, no logging
extra_body={"confidential": True}
)
print(response.choices[0].message.content)With confidential mode enabled, VoltageGPU guarantees:
- No logging: Prompts and responses are never logged or stored outside the enclave
- No persistence: GPU VRAM is securely wiped after each session
- Attestation proof: Verifiable cryptographic proof that the workload ran in a genuine TDX enclave
- Audit trail: Compliance-ready audit logs showing attestation verification timestamps
Use Case: Private Fine-Tuning
Fine-tuning is where confidential compute becomes critical. When you fine-tune a model, your proprietary training data — customer conversations, internal documents, domain expertise — gets embedded directly into model weights. On a non-confidential GPU, this data exists in plaintext in VRAM for hours or days during training.
With TDX-enabled fine-tuning on VoltageGPU:
- Your training data is encrypted in transit and in VRAM
- The resulting model weights are encrypted and only accessible within the enclave
- You can export the fine-tuned model via encrypted channels
- No one — not VoltageGPU, not the miner, not a physical attacker — can read your data or model
Use Case: Secure Multi-Party AI
Confidential GPUs enable a new paradigm: multiple organizations can collaboratively train or infer on combined datasets without exposing their individual data. Each party verifies the enclave attestation, sends their encrypted data, and the model processes the combined data inside the TEE. No party ever sees another's raw data.
This is transformative for:
- Federated learning across hospitals: Train better diagnostic models without sharing patient data
- Financial fraud detection: Banks can pool transaction patterns without exposing customer information
- Supply chain optimization: Manufacturers share logistics data without revealing trade secrets
Performance Impact
A common concern is that confidential computing adds overhead. Here is the reality with Intel TDX and modern NVIDIA GPUs:
- Memory encryption overhead: 2-5% on compute-bound workloads (LLM inference, training). AES-256-XTS is hardware-accelerated, so the impact is minimal.
- PCIe encryption (TEE-IO): 1-3% overhead on data transfer. Negligible for large batch inference and training where GPU compute dominates.
- Attestation: One-time cost at pod startup (under 10 seconds). Does not affect runtime performance.
- Total overhead: 3-8% for typical AI workloads. A small price for hardware-enforced data protection.
Getting Started
Confidential GPU compute on VoltageGPU is available today. Here is how to get started:
- Create an account at voltagegpu.com — $5 free credit, no credit card required
- Generate an API key in your dashboard settings
- Deploy a confidential pod using the API or dashboard (select "Confidential" toggle)
- Verify attestation programmatically before sending sensitive data
- Run your workload — inference, training, or fine-tuning in a hardware-secure enclave
If you are in healthcare, finance, legal, or government and need HIPAA/SOC2/GDPR-compliant GPU compute, browse available confidential GPUs or reach out at contact@voltagegpu.com for enterprise plans.
Try Confidential GPU Compute Today
H100, H200, B200, RTX 4090 in TDX enclaves. Hardware-enforced privacy, 40% cheaper than Azure.
Browse Confidential GPUs