TLDR: Premio Inc., a leader in rugged edge AI computing, has unveiled its new LLM Series Edge Servers, starting with the LLM-1U-RPL model. These compact, short-depth 1U rackmount servers are specifically engineered to bring real-time Generative AI (GenAI) and Large Language Model (LLM) capabilities directly to on-premises data centers at the edge, aiming to reduce reliance on traditional cloud resources, lower latency, and enhance data control for demanding IT/OT enterprise deployments.
Greater Los Angeles, CALIFORNIA – July 7, 2025 – Premio Inc., a global innovator in rugged edge AI computing and industrial display technology, has announced a significant expansion of its edge computing hardware portfolio with the introduction of the new LLM Series Edge Servers. The inaugural model in this series, the LLM-1U-RPL, is designed to revolutionize how enterprises deploy real-time Generative AI (GenAI) and Large Language Model (LLM) workloads by bringing them directly to the on-premises data center edge.
This strategic move by Premio addresses the escalating demand for localized AI capabilities, shifting away from exclusive cloud reliance. The LLM-1U-RPL Series is engineered to deliver superior performance through lower latency inferencing and more efficient data processing closer to the source of data generation. This approach aims to alleviate bandwidth strain, bolster data sovereignty, and facilitate real-time decision-making within hybrid cloud environments at the edge.
Dustin Seetoo, VP of Product Marketing at Premio, emphasized the server’s purpose-built design: “The LLM-1U-RPL is purpose-built for on-premise data centers to deliver high-performance, low-latency AI inferencing for large language model (LLM) workloads—without the need for traditional centralized cloud resources.” He further added, “Designed for the demands of edge deployments, this new edge server integrates 13th Gen Intel Core processors with performance-hybrid architecture, dedicated NVIDIA GPUs for accelerated computing, and industrial-grade power redundancy—key capabilities that enable real-time intelligence, reduce latency, and give organizations greater control over their data.”
Key features of the LLM-1U-RPL Series include its compact, short-depth 1U rackmount design, measuring 483 (W) x 480 (D) x 44 (H) mm, making it suitable for space-constrained environments. It is powered by 13th Gen Intel Core processors, supporting up to an i9 with a 65W TDP, and leverages a performance-hybrid architecture. This architecture utilizes P-cores for critical low-latency inferencing tasks like LLM prompt response and token generation, while E-cores handle general-purpose background applications. The server also supports up to 64GB of dual-channel DDR4 3200MT/s SODIMM memory, ensuring smooth processing of multi-modal data streams.
For accelerated computing, the LLM-1U-RPL supports up to an NVIDIA RTX 5000 Ada GPU and features PCIe Gen 4 expansion slots for additional GPU AI accelerators or high-throughput network cards. Storage options are flexible and high-speed, including m.2 NVMe and dual hot-swappable 2.5” SATA bays. Connectivity is optimized for on-premises edge AI, offering three 2.5GbE LAN ports, six USB 3.2 Gen2 ports, and COM ports.
Reliability and security are paramount, with the server featuring a 600W (1+1) redundant power supply, hot-swappable redundant smart fans for constant uptime, enhanced cybersecurity measures, and physical security features. The series also boasts world-class certifications, including UL, FCC, and CE.
Also Read:
- Taiwanese Industrial PC Manufacturers Spearhead Edge AI Revolution with AI Agents
- Clarifai Unveils ‘AI Runners’ for Hybrid Local and Cloud AI Model Deployment
This launch aligns with a broader industry trend where AI, particularly Edge AI and Generative AI, is rapidly moving towards operational deployment closer to where data is generated. Premio’s LLM Series is poised to be a foundational infrastructure for modern AI, enabling industries, especially those in Industry 4.0, to achieve real-time insights, autonomy, and innovation directly at the edge.


