VDURA recently announced Version 12 of its VDURA Data Platform (VDP), formerly known as PanFS, introducing three primary architectural enhancements to its parallel file system: an elastic Metadata Engine for distributed metadata processing, system-wide snapshot capabilities, and native support for SMR disk drives.
The company projects performance improvements exceeding 20% for aggregate throughput, metadata operations acceleration up to 20x, and cost-per-terabyte reductions above 20%.
Version 12 targets AI training environments, high-performance computing workloads, and hyperscale data operations where metadata scalability and storage density economics increasingly constrain system performance.
General availability is scheduled for Q2 2026 for V5000 and V7000-class systems, with zero-downtime upgrade paths for customers running V11 or later releases.
Technical Details
VDP V12 addresses three fundamental bottlenecks in large-scale AI infrastructure:
- Metadata processing constraints as file counts reach billions
- Data protection complexity at petabyte scales
- Capacity density economics for cold and warm data tiers
The new release builds upon architectural foundations established in V11, which introduced initial data placement capabilities that V12 now extends with intelligent zone management for SMR media.
Elastic Metadata Engine Architecture
The Elastic Metadata Engine is a fundamental architectural shift from centralized metadata processing to distributed scale-out operations across multiple metadata nodes.
VDURA describes this engine as enabling linear performance scaling as file and object populations grow into the billions, eliminating traditional metadata bottlenecks that constrain parallel file system performance in AI training environments where datasets may contain hundreds of millions of small files.
The company claims metadata operation acceleration up to 20x compared to previous releases. In practical terms, metadata-intensive operations for AI workloads include:
- Directory traversal for dataset discovery and pipeline initialization
- File attribute queries during training data preprocessing
- Object listing operations for model checkpoint management
- Permission verification for multi-tenant AI development environments
- Extended attribute handling for data lineage tracking
The dynamic scaling mechanism allows metadata processing capacity to expand horizontally as workload demands increase, theoretically preventing metadata operations from becoming the rate-limiting factor in GPU-attached storage systems.
However, the implementation details regarding metadata consistency protocols, node failure recovery mechanisms, and performance characteristics under mixed read/write metadata workloads have not been disclosed.
Potential adopters of the technology should evaluate how this architecture performs under their specific metadata access patterns, particularly for workloads combining large sequential streaming operations with concurrent small-file metadata lookups.
Snapshot Implementation
V12 introduces system-wide snapshot capabilities enabling instantaneous, space-efficient point-in-time copies of datasets. VDURA describes these snapshots as using copy-on-write mechanisms to create immutable dataset replicas without full data duplication.
This addresses several operational requirements in AI development workflows:
- Pipeline checkpoint preservation: Training runs can capture intermediate states without interrupting GPU processing or consuming duplicate storage capacity for each checkpoint
- Dataset versioning: Data science teams can maintain multiple dataset versions for A/B testing and reproducibility without multiplicative storage costs
- Operational recovery: System administrators gain rollback capabilities for datasets corrupted by failed processing jobs or erroneous transformations
- Compliance and governance: Organizations subject to regulatory requirements can maintain point-in-time dataset records for audit trails
VDP V12 integrates snapshot operations with both manual invocation and policy-based automation, allowing administrators to define retention schedules aligned with training cycles, compliance windows, or disaster recovery objectives.
The space efficiency claims imply that only modified blocks consume additional capacity, though actual storage overhead will depend on dataset change rates and snapshot retention policies. We haven’t validated this with VDURA.
Competitors including DDN ExaScaler (Lustre-based) and IBM Spectrum Scale offer similar snapshot capabilities, making this feature table stakes rather than competitive differentiation.
SMR Drive Optimization
The new release introduces native support for Shingled Magnetic Recording (SMR) drives, an expansion into higher-density, cost-optimized storage tiers.
SMR technology increases areal density by overlapping write tracks, delivering 25-30% greater capacity per physical drive compared to Conventional Magnetic Recording (CMR) alternatives.
However, this density advantage comes with significant performance trade-offs that VDP’s new write-placement engine aims to mitigate.
SMR drives organize storage into zones where tracks partially overlap, creating a fundamental constraint: modifying data within any track requires rewriting the entire zone containing that track. This zone-level rewrite requirement imposes substantial performance penalties for random write workloads and partial data updates.
The physics of SMR technology make these drives inherently better suited for sequential write patterns and append-only workloads rather than random-access modification scenarios.
VDURA’s write-placement engine addresses these constraints through intelligent data organization that aligns write operations with SMR zone boundaries and access patterns.
The company says that this engine unlocks “25-30%” additional capacity per rack “without compromising throughput or data integrity.”
Sequential write workloads that align naturally with SMR zone characteristics, such as AI training dataset ingestion, model checkpoint archival, and inference result logging should see strong benefit.
Performance and Efficiency
VDURA asserts three headline performance improvements in V12:
- Aggregate throughput increased by more than 20%: This claim can be attributed to optimizations across the entire I/O stack, including network protocol handling, storage backend management, and client-side parallelism.
- Metadata operations accelerated up to 20x: The “up to” qualifier indicates best-case scenarios under specific workload conditions, not guaranteed performance across all metadata operation types. Directory traversal, attribute queries, and file creation operations each exhibit different scaling characteristics, and the 20x acceleration likely applies to specific metadata-intensive operations rather than representing a universal improvement.
- Cost per terabyte reduced by over 20%: This economic claim apparently derives from SMR drive integration, which delivers 25-30% additional capacity per rack unit.
Analysis
VDURA’s Data Platform Version 12 addresses legitimate architectural constraints in AI infrastructure through its Elastic Metadata Engine, snapshot capabilities, and SMR drive optimization. These all target the metadata scalability bottlenecks, data protection complexity, and storage economics challenges that organizations encounter as training datasets expand into multi-petabyte scales with billions of files.
The release shows the company continuing to evolve from its HPC roots into a viable solution for AI workloads. This includes bringing VDURA to feature parity with competitors on snapshot functionality while advancing its metadata scalability and storage tiering intelligence.
The Elastic Metadata Engine’s scale-out architecture addresses a fundamental bottleneck in parallel file systems supporting AI workloads with massive file counts, though similar distributed metadata approaches exist in competing platforms from vendors including IBM, WEKA, and VAST Data.
For enterprises, AI service providers, and neoclouds operating large-scale AI training infrastructure where metadata bottlenecks constrain GPU utilization and storage economics significantly impact total infrastructure costs, VDURA V12 merits serious evaluation.
The combination of its proven parallel file system heritage, the new metadata scalability enhancements, and storage economics optimization through SMR integration positions VDURA competitively.
We like the direction VDURA is evolving. With the VDP V12 release, the company shows that it’s a worthy contender that should considered by any organization with scalable AI data needs.
Competitive Outlook & Advice to IT Buyers
These sections are only available to NAND Research clients and IT Advisory Members. Please reach out to [email protected] to learn more.



