Gcore Unveils Inference at the Edge for Real-Time AI Performance

DateJun 6, 2024

Gcore has launched Gcore Inference at the Edge, a groundbreaking solution designed to bring AI applications closer to end users, ensuring ultra-low latency for real-time performance. This solution enables businesses in sectors like automotive, manufacturing, retail, and technology to deploy pre-trained machine learning models cost-effectively and at scale.

Running on Gcore’s global network of over 180 edge nodes, the solution utilizes NVIDIA L40S GPUs and smart routing technology to achieve response times of under 30 milliseconds. This network’s strategic placement ensures servers are close to end users, enhancing performance significantly.

The Gcore ML Model Hub supports various foundational and custom models, including LLaMA Pro 8B, Mistral 7B, and Stable-Diffusion XL. These models can be selected and trained for any use case and then distributed globally to the edge nodes. This approach addresses the common issue of poor performance from running AI models on training servers.

Key benefits of Gcore Inference at the Edge would include the following:

Cost-effective deployment – Customers only pay for the resources they useDDoS protection – Built-in DDoS protection for ML endpoints through Gcore’s infrastructureData privacy and security – Compliance with GDPR, PCI DSS, and ISO/IEC 27001 standardsModel autoscaling – Ability to handle load spikes, ensuring readiness during peak demandUnlimited object storage – Scalable S3-compatible cloud storage to accommodate evolving model needs

Andre Reitenbach, CEO of Gcore, emphasized the company’s commitment to making AI deployment easier for businesses by eliminating concerns about costs, skills, and infrastructure. He highlighted the edge’s potential for achieving the best performance and end-user experiences, reflecting Gcore’s continuous innovation to provide unparalleled scale and efficiency.

In short, Gcore Inference at the Edge would stand out as a modern, efficient solution for AI inference, offering businesses the power and ease of deployment necessary for advancing their AI capabilities on a global scale.

Leave a Reply