Quick Answer: For $20/month, you get a personal AI agent inside Telegram that runs on Intel TDX hardware enclaves in the EU. Not "we promise not to look." We can't look. The CPU encrypts your prompts in memory. Even with root access to our own servers, we couldn't read them.
TL;DR: I set up the Plus tier agent in 4 minutes flat. Average response time: 755ms TTFT, 120 tokens/sec throughput on H200 GPUs. TDX overhead: 3-7% vs bare metal. 2,000 requests/month. Your conversation history stays encrypted. You can verify this yourself with /attest.
The Problem With "Private" AI
Every AI company says your data is private. Then you read the subclause.
OpenAI's Enterprise plan? Data isn't used for training. Great. Still sits unencrypted on shared GPUs in US data centers. A hypervisor bug, a misconfigured access policy, a National Security Letter — your conversations are readable by someone.
Telegram bots for AI are worse. Most are thin wrappers around OpenAI's API. Your messages bounce through a developer's server, then OpenAI's, then back. Two parties. Two privacy policies. Two failure points.
I wanted something actually sealed. Not contractually. Architecturally.
That's what led me to build this.
What Hardware-Sealed Actually Means
Intel TDX (Trust Domain Extensions) creates encrypted memory regions the host OS can't access. The CPU itself manages the keys. When our AI model processes your message, it happens inside a "trust domain" where:
- Memory is AES-256 encrypted at runtime
- The hypervisor is untrusted by design
- On boot, the CPU generates an attestation report you can verify
- We, the operator, are silicon-prevented from reading anything inside
I spent 3 hours once setting up Azure Confidential Computing for a side project. Gave up. The attestation workflow, the driver compatibility, the "confidential capable" instance types — it's a research project, not a product. Our setup deploys in ~60 seconds. I timed it.
Here's what the attestation check looks like from the bot:
/attest
→ TDX quote verified
→ MRENCLAVE: 0x4a3f...e9d2
→ Signer: Intel SGX-TDX
→ Status: GENUINE
That MRENCLAVE hash? It's a cryptographic fingerprint of the exact code running inside. Change one line, the hash changes. You know what you're talking to.
The Setup: 4 Minutes, No Terminal
I hate install steps. Node version managers. --session-id flags. BYO API keys. The OpenClaw project has 367k GitHub stars and I bet 80% of users bounce at nvm install 22.
Our funnel is: subscribe on Stripe → get token vgpu_xxxx by email → /start vgpu_xxxx in Telegram → done.
I tested it on a fresh phone. 3 minutes 47 seconds from payment to first response. The bot's @VoltageGPUPersonalBot.
What you get:
| Feature | Plus ($20/mo) | Starter ($349/mo) | Pro ($1,199/mo) |
|---|---|---|---|
| Model | Qwen3-32B-TEE | Qwen3-32B-TEE | Qwen3.5-397B-TEE |
| Context window | 32K tokens | 32K tokens | 256K tokens |
| Requests/month | 2,000 | 500 (team) | 5,000 (team) |
| Seats | 1 | 3 | 10 |
| Response speed | 755ms TTFT | 755ms TTFT | 755ms TTFT |
| Hardware | Intel TDX H200 | Intel TDX H200 | Intel TDX H200 |
The 397B model on Pro is 12x larger. Whole documents in one shot. But honestly? For personal use — quick contract checks, tax questions, medical record summaries — the 32B is sharp enough. I use it for parsing employment offers. It caught a non-compete clause my lawyer skimmed past.
Real Performance Numbers
These aren't spec sheet figures. Live from our H200 TDX nodes this week:
- Time to first token: 755ms average (measured over 1,000 requests, p95: 1,180ms)
- Throughput: 120 tokens/second generation speed
- TDX overhead vs bare metal: 5.2% on our tests (range: 3-7% depending on prompt length)
- Cold start: 30-60s on first boot if the node was idle
That overhead is the encryption cost. Worth it. The alternative is zero encryption.
What I Actually Use It For
Medical stuff, mainly. I had bloodwork results with 14 markers. The hospital's portal explained 3 of them. I pasted the PDF text to the bot, asked for plain-language context on the rest, and whether any combinations were worth flagging. It didn't diagnose. It educated. And my health data never left a hardware-sealed enclave in France.
Tax questions too. French micro-entrepreneur regime, quarterly declarations. The bot knows the thresholds. I don't have to explain my situation to a US-trained model that thinks "LLC" is the default.
The Honest Limitations
- No SOC 2 certification. We use GDPR Article 25 + Intel TDX attestation instead. If your procurement requires SOC 2, we're not there yet.
- PDF OCR not supported. Text-based PDFs work fine. Scanned documents don't. Convert first.
- 32B model misses edge cases. Complex legal reasoning with conflicting precedents? The 397B Pro model handles it. This one sometimes hedges too much.
- Cold start lag: First request after idle can take 30-60s. Subsequent ones are sub-second.
One competitor beats us on raw speed. RunPod's A100s at ~$1.64/hr are cheaper than our infrastructure. But they're not TDX-sealed. Different product entirely.
Using the API Directly
The Telegram bot is a frontend. Same backend powers API access:
from openai import OpenAI
client = OpenAI(
base_url="https://api.voltagegpu.com/v1/confidential",
api_key="vgpu_YOUR_KEY"
)
response = client.chat.completions.create(
model="qwen3-32b-tee",
messages=[{"role": "user", "content": "Explain this clause: 'The Employee shall not engage in any competing business within a 50km radius for 24 months post-termination.'"}]
)
print(response.choices[0].message.content)
Same encryption. Same attestation. Different interface.
Why Telegram?
It's where people already are. No new app. No password to forget. End-to-end encrypted if you use Secret Chats, though our bot runs in normal chats (the TDX seal is stronger than Telegram's server-side encryption anyway).
For EU residents especially, post-ChatGPT-sanctions uncertainty, having an AI that physically can't export data to the US matters. GDPR Article 25 "data protection by design" isn't a checkbox for us. It's the architecture.
More on our compliance approach: voltagegpu.com/guides/gdpr-ai-compliance Compare with enterprise alternatives: voltagegpu.com/vs/chatgpt-enterprise Developer docs and API reference: voltagegpu.com/for-developers-api
Don't trust me. Test it. 5 free agent requests/day → voltagegpu.com