OpenMetal offers two XL v4 bare metal configurations with the same chassis, the same 1TB DDR5 RAM, the same 25.6TB NVMe storage, and the same network tier — but fundamentally different CPU profiles. The standard XL v4 runs dual Intel Xeon Gold 6530 processors for 64 cores at 2.1 GHz, while the XL v4 High Frequency runs dual Gold 6544Y for 32 cores at 3.6 GHz. Both are 5th Gen Emerald Rapids on Intel 7, both ship with TDX enabled by default, and both are available at all five OpenMetal regions.
The choice comes down to one question: does your workload scale horizontally across many threads, or does it depend on individual thread execution speed? This page breaks down the hardware differences, the workload profiles where each server excels, and the scenarios where choosing the wrong configuration leaves measurable performance on the table.
Side-by-Side Specification Comparison
| Component | XL v4 (Standard) | XL v4 High Frequency |
|---|---|---|
| CPU | 2x Intel Xeon Gold 6530 | 2x Intel Xeon Gold 6544Y |
| Architecture | Emerald Rapids (Intel 7) | Emerald Rapids (Intel 7) |
| Total Cores / Threads | 64C / 128T | 32C / 64T |
| Base Frequency | 2.10 GHz | 3.60 GHz |
| Max Turbo Frequency | 4.0 GHz | 4.1 GHz |
| High-Priority Cores | 12 per socket @ 2.3 GHz | 4 per socket @ 3.7 GHz |
| L3 Cache | 160 MB per socket (320 MB total) | 45 MB per socket (90 MB total) |
| TDP | 270W per socket | 270W per socket |
| Memory | 1TB DDR5-4800 ECC | 1TB DDR5-5200 ECC |
| Memory Bandwidth (theoretical) | ~307 GB/s | ~333 GB/s |
| Data Storage | 4x 6.4TB Micron 7500 MAX NVMe | 4x 6.4TB Micron 7500 MAX NVMe |
| Boot Storage | 2x 960GB RAID 1 | 2x 960GB RAID 1 |
| Private Bandwidth | 20 Gbps LACP | 20 Gbps LACP |
| Public Bandwidth | 6 Gbps | 6 Gbps |
| Intel TDX | Enabled by default | Enabled by default |
| Intel SGX | Yes (128 GB EPC) | Yes (128 GB EPC) |
| Intel AMX | Yes | Yes |
| Intel Speed Select | 3 profiles (up to 2.3 GHz at 16C) | 3 profiles (up to 3.9 GHz at 8C) |
| PCIe | 5.0, 80 lanes per socket | 5.0, 80 lanes per socket |
| Pricing | Fixed monthly — openmetal.io/bare-metal-pricing | Fixed monthly — openmetal.io/bare-metal-pricing |
| Locations | Ashburn, LA, Amsterdam, Singapore, Pod-0 | Ashburn, LA, Amsterdam, Singapore, Pod-0 |
Ready to Deploy a XL v4 or XL v4 High Frequency?
Tell us about your workload and we’ll help you configure the right deployment — bare metal or Hosted Private Cloud, in any of our five data center regions.
Get a XL v4 or XL v4 High Frequency Quote Schedule a Consultation
Understanding the Trade-Off
Thread count vs clock speed
The Gold 6530 has 32 cores per socket (64 total), each running at 2.10 GHz base. The Gold 6544Y has 16 cores per socket (32 total), each running at 3.60 GHz base. The per-core clock speed difference is 71% — a significant gap that directly translates to how quickly a single thread completes a unit of work.
For embarrassingly parallel workloads — rendering, batch data transformation, multi-threaded compression, containerized microservices where each container uses a few threads — the standard XL v4’s 128 hardware threads win. More threads processing simultaneously means higher aggregate throughput.
For workloads with a critical path that cannot be parallelized — matching engine order books, financial tick evaluation, real-time event stream processing, single-threaded game server simulation — the XL v4 HF’s 3.6 GHz floor finishes each sequential step faster. More threads at lower clock do not help when the bottleneck is thread-serial execution.
L3 Cache: 320 MB vs 90 MB
The Gold 6530’s 320 MB total L3 cache is 3.5x larger than the Gold 6544Y’s 90 MB. Cache size matters for workloads with large working sets that need to stay in cache to avoid DRAM latency penalties — graph traversal, in-memory databases with hot datasets, or ML inference serving many different models concurrently.
For workloads with compact hot datasets — a single high-frequency trading order book, a time-series ingestion pipeline processing a bounded recent window, a low-latency API with hot request paths — the smaller cache is not a constraint, and the higher clock speed delivers better results.
Speed Select Technology profiles
Both CPUs support Intel Speed Select Technology, but the profiles are different in character:
| Profile | XL v4 Standard (6530) | XL v4 High Frequency (6544Y) |
|---|---|---|
| Profile 0 (default) | 32C @ 2.1 GHz, 270W | 16C @ 3.6 GHz, 270W |
| Profile 1 | 24C @ 2.1 GHz, 225W | 12C @ 3.6 GHz, 225W |
| Profile 2 | 16C @ 2.3 GHz, 205W | 8C @ 3.9 GHz, 205W |
Profile 2 on the HF server (8 cores per socket at 3.9 GHz) is useful for scenarios where a small number of threads need the highest possible frequency and power draw must be bounded — co-located matching engine threads, hardware-isolated TDX Trust Domains with specific SLA requirements, or latency-critical microservices where only a few threads need prioritization.
Profile 2 on the standard XL v4 (16 cores at 2.3 GHz) is useful for mixed workloads where some batch jobs can be parked at reduced clock while interactive threads take priority — though the HF server’s Profile 0 already outperforms this at base.
Choose XL v4 Standard When…
- Your workload is multi-threaded by design: OpenStack compute clusters, Kubernetes node pools, Proxmox multi-VM hosts, containerized microservice fleets. More threads means more simultaneous work.
- You need maximum core count for per-core licensed software: Databases, virtualization platforms, or middleware priced per physical core benefit from the 64-core count — more licensed threads per dollar.
- Your working set is large: In-memory databases (Redis, Apache Ignite), JVM applications with large heap footprints, or graph processing workloads that benefit from 320 MB total L3 cache.
- You run batch workloads: ETL jobs, log aggregation, data pipeline transformations, and parallel build systems scale with thread count rather than single-thread speed.
- HPC cluster or Hosted Private Cloud: The standard XL v4 HPC cluster (3x Gold 6530 = 192 cores, 3TB DDR5) provides more OpenStack vCPU capacity per cluster than an equivalent HF cluster.
Choose XL v4 High Frequency When…
- Per-thread latency is your SLA: Real-time trading systems, low-latency APIs with p99 requirements, game server simulation loops, or DSP pipelines where sequential execution speed determines outcome.
- Your critical path is not parallelizable: Matching engines, event sourcing systems, single-threaded physics simulations, or consensus algorithms where higher clock completes each step faster regardless of thread count.
- You need TDX + high clock speed together: Both servers ship TDX enabled, but the HF server’s higher clock means confidential computing workloads (PHI processing, proprietary algorithm execution in TDX Trust Domains) run faster per request.
- Your application uses Speed Select to dedicate cores: Financial infrastructure or real-time DSP systems that want to isolate 8-12 critical threads at maximum frequency (3.9 GHz) while the remaining cores handle ancillary processing.
- Storage latency is on the critical path: Both servers use the same Micron 7500 MAX NVMe (15 µs write, 70 µs read), but the HF server’s faster CPU ensures the CPU does not become a bottleneck after storage responds.
Workload Routing Guide
| Workload | Recommended SKU | Reason |
|---|---|---|
| OpenStack / Hosted Private Cloud | XL v4 Standard | 64 cores = more vCPU density |
| Kubernetes node pool | XL v4 Standard | Pod scheduling benefits from more threads |
| Batch ETL / data pipelines | XL v4 Standard | Parallelizable, scales with thread count |
| PostgreSQL / MySQL (high concurrency) | XL v4 Standard | Many simultaneous connections, parallel query |
| Redis / in-memory cache | XL v4 Standard | Large L3 cache (320 MB) holds more hot data |
| HFT / matching engine | XL v4 High Frequency | Sequential critical path, 3.6 GHz floor |
| Real-time analytics (ClickHouse, Druid) | XL v4 High Frequency | Per-query latency over throughput |
| Blockchain validator | XL v4 High Frequency | Consensus timing sensitive to per-thread speed |
| Low-latency API / fintech backend | XL v4 High Frequency | p99 latency driven by single-thread execution |
| Time-series DB (InfluxDB, TimescaleDB) | XL v4 High Frequency | High-frequency write ingestion at low latency |
| ML inference (per-request, AMX) | XL v4 High Frequency | Per-request latency over batch throughput |
| ML training or batch inference | XL v4 Standard | Multi-threaded, core count scales throughput |
| Proxmox / KVM multi-VM host | XL v4 Standard | More VMs, more vCPUs |
| Confidential computing workloads (TDX) | Either — both ship TDX by default | Match to workload profile |
Performance Benchmarks Context
Neither server is faster in absolute terms — they are faster for different workload shapes. In a synthetic single-threaded benchmark, the XL v4 HF will show higher scores by virtue of its 3.6 GHz vs 2.1 GHz base. In a multi-threaded throughput benchmark (Cinebench, SPEC CPU2017 rate), the standard XL v4 will score higher due to 2x the thread count.
For production decisions: benchmark your actual workload on both servers. OpenMetal offers Proof of Concept clusters for testing — contact openmetal.io/contact for access.
Ready to Deploy a XL v4 or XL v4 High Frequency?
Tell us about your workload and we’ll help you configure the right deployment — bare metal or Hosted Private Cloud, in any of our five data center regions.
Get a XL v4 or XL v4 High Frequency Quote Schedule a Consultation
Deployment Options
Both the XL v4 standard and XL v4 High Frequency are available as standalone bare metal servers (full root access, IPMI, dedicated hardware) and as Hosted Private Cloud nodes (three-node OpenStack + Ceph clusters with Day 2 ops included).
Bare metal: Single-server deployments with fixed monthly pricing, price locks up to 5 years, and ramp pricing for migrations. Deploy in Ashburn, Los Angeles, Amsterdam, Singapore, or Pod-0.
Hosted Private Cloud: The XL v4 HPC cluster (3x Gold 6530) and the XL v4 HF HPC cluster (3x Gold 6544Y) provide OpenStack-managed private cloud at different performance profiles. The HF cluster is better for HPC workloads where per-VM clock speed matters; the standard cluster is better for workload diversity and higher vCPU density. See hosted-private-cloud-xl-v4 for the standard HPC variant.
Get an XL v4 Quote
Not sure which XL v4 to choose? Tell us about your workload and we’ll recommend the right configuration.
- XL v4 Standard: 64C/128T, 2.1 GHz, 320 MB L3, 1TB DDR5-4800 — high concurrency, large working sets
- XL v4 High Frequency: 32C/64T, 3.6 GHz, 90 MB L3, 1TB DDR5-5200 — low-latency, high single-thread frequency
- Hosted Private Cloud: Three-node OpenStack + Ceph clusters in either CPU configuration
All deployments include fixed monthly pricing, 99.96%+ network SLA, and DDoS protection.
Product specifications, pricing, and availability may change due to market conditions and other factors. For the most current information, please contact the OpenMetal team directly.



































