StorageReview.com

QNAP Introduces QAI-h1290FX Edge AI Storage Server for Private LLM and Generative AI Workloads

Enterprise  ◇  Medium NAS

QNAP Systems has introduced the QAI-h1290FX, an edge AI storage server designed for organizations that want to run large language models, retrieval-augmented generation search, and other generative AI workloads on their own infrastructure. The system is designed for enterprises that balance AI adoption with requirements for data privacy, low latency, governance, and operational control, enabling teams to deploy AI applications locally rather than sending sensitive data to public cloud platforms.

QNAP QAI h1290FX front panel

Organizations can use the QAI-h1290FX to deploy internal AI assistants for employee training, policy questions, and knowledge lookup, keeping the underlying data within the business. Legal, finance, HR, and operations teams can build private RAG pipelines to search contracts, reports, and internal records, providing more context than traditional keyword searches. Creative teams can run image-generation tools such as Stable Diffusion or ComfyUI for design and content workflows. In contrast, IT teams can use automation tools such as n8n to trigger inference tasks, generate content, or route alerts across business systems.

QNAP QAI-h1290FX Components, Expansion, and I/O

Built around an AMD EPYC 7302P processor and twelve U.2 NVMe/SATA SSD bays, the QAI-h1290FX combines server-grade compute with an all-flash storage design tailored for AI workloads requiring fast data access. The 16-core, 32-thread processor supports inference, virtualization, and parallel workloads. At the same time, the SSD architecture is designed for frequent model execution, high-speed data streaming, and responsive access to datasets, embeddings, documents, and generated content.

The system also supports optional NVIDIA RTX PRO 6000 Blackwell Max-Q Workstation GPU acceleration, providing up to 96GB of GPU memory for more demanding local AI workloads. Support for CUDA, TensorRT, and Transformer Engine acceleration enables teams to run large language model inference, image generation, and deep learning applications on-premises without building a separate GPU workstation from scratch.

QNAP QAI h1290FX rear panel

The QAI-h1290FX offers built-in high-speed networking with two 25GbE SFP28 SmartNIC ports and two 2.5GbE ports, and supports Wake-on-LAN via the 2.5GbE ports. For expansion, it includes four PCIe slots, with three PCIe Gen 4 x16 slots and one PCIe Gen 4 x8 slot, providing room to add higher-speed networking, a GPU, or other compatible expansion cards. Additional I/O includes three USB 3.2 Gen 1 ports, jumbo frame support, SR-IOV, GPU pass-through, and compatibility with 2.5-inch SATA SSDs and U.2 NVMe PCIe Gen4 x4 SSDs across its twelve drive bays.

QNAP QAI-h1290FX Specifications

Specification QNAP QAI-h1290FX
Overview
Model AI-h1290FX-7302P-128G
Processor and Memory
CPU AMD EPYC™ 7302P 16-core/32-thread processor, up to 3.3 GHz
CPU Architecture 64-bit x86
Encryption Engine  (AES-NI)
System Memory 128 GB RDIMM DDR4 ECC
Maximum Memory 1 TB (8 x 128 GB)
Memory Slot 8 x RDIMM DDR4
Flash Memory 8GB (Dual boot OS protection)
Storage
Drive Bay 12 x 2.5-inch U.2 PCIe NVMe / SATA 6Gbps
The system is shipped without SSDs.
For the SSD compatibility list, please visit https://www.qnap.com/compatibility/
Drive Compatibility 2.5-inch bays:
2.5-inch SATA solid state drives
2.5-inch U.2 NVMe PCIe Gen4 x4 solid state drives
Hot-swappable Yes
SSD Cache Acceleration Support Yes
GPU and Virtualization
GPU pass-through Yes
SR-IOV Yes
Networking
2.5 Gigabit Ethernet Port (2.5G/1G/100M) 2 (2.5G/1G/100M/10M)
25 Gigabit Ethernet Port 2 x 25GbE SFP28 SmartNIC port
Wake on LAN (WOL) Only the 2.5GbE port
Jumbo Frame Yes
Expansion and Ports
PCIe Slot 4
Slot 1: PCIe Gen 4 x16
Slot 2: PCIe Gen 4 x16
Slot 3: PCIe Gen 4 x8
Slot 4: PCIe Gen 4 x16
Card dimensions for PCIe slot 1 & Slot 2:185 x 111.15 x 18.76 mm / 7.28 x 4.38 x 0.74 inches.
Card dimensions for PCIe slot 3 & Slot 4:280 x 111.15 x 18.76 mm / 11.02 x 4.38 x 0.74 inches.
Wider cards can be installed if the next PCIe slot will not be used.
USB 3.2 Gen 1 port 3
Physical Design
Form Factor Tower
LED Indicators Power/Status, LAN, USB, SSD1-12
LCD Display/ Button Yes
Buttons Power, Reset, USB Auto Copy
Dimensions (HxWxD) 150 × 368 × 362 mm
Dimensions do not include the foot pad (foot pad may be up to 10mm / 0.39 inches high, depending on model)
Weight (Net) 10.4 kg
Weight (Gross) 11.3 kg
Environment and Power
Operating Temperature 0 – 40 °C (32°F – 104°F)
Storage Temperature -20 – 70°C (-4°F – 158°F)
Relative Humidity 5-95% RH non-condensing, wet bulb: 27˚C (80.6˚F)
Power Supply Unit 750W, 100-240V
Fan 2 x 92mm, 12VDC
System Warning Buzzer
Kensington Security Slot Yes
Warranty and Connections
Standard Warranty 5
Max. Number of Concurrent Connections (CIFS) – with Max. Memory 10000

Built Around Fast Storage, GPU Acceleration, and Local Control

Running on QNAP’s ZFS-based QuTS hero operating system, the QAI-h1290FX includes enterprise-grade storage features such as data integrity protection, extensive snapshot support, and inline deduplication. These capabilities are relevant to AI deployments because organizations often handle large volumes of repeated or related data across documents, embeddings, model files, training materials, and generated outputs.

Developers and IT teams can run AI tools in containerized environments with native GPU access via QNAP Container Station. At the same time, Virtualization Station supports GPU pass-through for virtual machines. This gives organizations more control over how compute resources are assigned, whether workloads are deployed in containers for speed and portability or in virtual machines for separation, testing, and administrative control.

The QAI-h1290FX also includes preloaded AI tools such as AnythingLLM, OpenWebUI, and Ollama, enabling teams to set up private LLM workflows and local chat interfaces more quickly. Additional applications (including Stable Diffusion, ComfyUI, n8n, and vLLM) are being integrated to support text generation, image creation, workflow automation, and inference use cases on the same local platform.

A Local Infrastructure Option for Enterprise AI Teams

QNAP says the platform can reduce the manual work typically involved in building local AI infrastructure, including assembling a GPU workstation, installing AI tools, and configuring separate environments. Users can deploy supported AI models and applications directly on the system while retaining control over their data and avoiding reliance on cloud services.

It’s also compatible with QNAP JBOD expansion enclosures, providing organizations with a path to scale storage capacity as AI datasets, internal knowledge bases, model files, and generated content continue to grow.

QNAP QAI-h1290FX Product Page

Engage with StorageReview

Newsletter | YouTube | Podcast iTunes/Spotify | Instagram | Twitter | TikTok | RSS Feed

Lyle Smith

Lyle is a writer for StorageReview, covering a broad set of end user and enterprise IT topics.