inference Archives - High-Performance Computing News Analysis | insideHPC https://insidehpc.com/tag/inference/ At the Convergence of HPC, AI and Quantum Tue, 16 Apr 2024 18:17:04 +0000 en-US hourly 1 https://wordpress.org/?v=6.5.3 57143778 NeuReality Launches Developer Portal for NR1 AI Inference Platform  https://insidehpc.com/2024/04/neureality-launches-developer-portal-for-nr1-ai-inference-platform/ https://insidehpc.com/2024/04/neureality-launches-developer-portal-for-nr1-ai-inference-platform/#respond Tue, 16 Apr 2024 18:16:53 +0000 https://insidehpc.com/?p=93852

SAN JOSE — April 16, 2024 — NeuReality, an AI infrastructure technology company, announced today the release of a software developer portal and demo for installation of its software stack and APIs. The company said the announcement marks a milestone since delivery of its 7nm AI inference server-on-a-chip, the NR1 NAPU, and bring up of […]

The post NeuReality Launches Developer Portal for NR1 AI Inference Platform  appeared first on High-Performance Computing News Analysis | insideHPC.

]]>
https://insidehpc.com/2024/04/neureality-launches-developer-portal-for-nr1-ai-inference-platform/feed/ 0 93852
In-Memory Computing Could Be an AI Inference Breakthrough https://insidehpc.com/2024/02/in-memory-computing-could-be-the-inference-breakthrough-ai-needs/ Thu, 22 Feb 2024 15:58:41 +0000 https://insidehpc.com/?p=93514

[CONTRIBUTED THOUGHT PIECE] In-memory computing promises to revolutionize AI inference. Given the rapid adoption of generative AI, it makes sense to pursue a new approach to reduce cost and power consumption by bringing compute in memory and improving performance.

The post In-Memory Computing Could Be an AI Inference Breakthrough appeared first on High-Performance Computing News Analysis | insideHPC.

]]>
93514
Industry Heavyweights Form Ultra Ethernet Consortium for HPC and AI https://insidehpc.com/2023/07/industry-heavyweights-form-ultra-ethernet-consortium-for-hpc-and-ai/ Wed, 19 Jul 2023 12:30:08 +0000 https://insidehpc.com/?p=91946

SAN FRANCISCO – July 19, 2023 – A host of industry heavyweights have formed the Ultra Ethernet Consortium (UEC), intended to promote “industry-wide cooperation to build a complete Ethernet-based communication stack architecture for high-performance networking” for HPC and AI workloads, the new group said. Founding members include AMD, Arista, Broadcom, Cisco, Eviden (an Atos Business), […]

The post Industry Heavyweights Form Ultra Ethernet Consortium for HPC and AI appeared first on High-Performance Computing News Analysis | insideHPC.

]]>
91946
MLCommons: Latest MLPerf AI Benchmark Results Show Machine Learning Inference Advances https://insidehpc.com/2022/09/mlcommoncs-latest-mlperf-ai-benchmark-results-show-machine-learning-inference-advances/ Thu, 08 Sep 2022 17:00:18 +0000 https://insidehpc.com/?p=90136

SAN FRANCISCO – September 8, 2022 – Today, the open engineering consortium MLCommons announced results from MLPerf Inference v2.1, which analyzes the performance of inference — the application of a trained machine learning model to new data. Inference allows for the intelligent enhancement of a vast array of applications and systems. Here are the results and […]

The post MLCommons: Latest MLPerf AI Benchmark Results Show Machine Learning Inference Advances appeared first on High-Performance Computing News Analysis | insideHPC.

]]>
90136
MLCommons Launches MLPerf Tiny AI Inference Benchmark https://insidehpc.com/2021/06/mlcommons-launches-mlperf-tiny-ai-inference-benchmark/ Wed, 16 Jun 2021 17:00:07 +0000 https://insidehpc.com/?p=87817

Today, open engineering consortium MLCommons released a new benchmark, MLPerf Tiny Inference to measure trained neural network AI inference performance for low-power devices in small form factors. MLPerf Tiny v0.5 is MLCommons’s first inference benchmark suite for embedded device machine learning, a growing field in which AI-driven sensor data analytics is performed in real-time, close […]

The post MLCommons Launches MLPerf Tiny AI Inference Benchmark appeared first on High-Performance Computing News Analysis | insideHPC.

]]>
87817
LeapMind Unveils Efficiera Ultra Low-Power AI Inference Accelerator IP https://insidehpc.com/2020/04/leapmind-unveils-efficiera-ultra-low-power-ai-inference-accelerator-ip/ Thu, 23 Apr 2020 07:13:30 +0000 https://insidehpc.com/?p=83803

Today LeapMind announced Efficiera, an ultra-low power AI inference accelerator IP for companies that design ASIC and FPGA circuits, and other related products. Efficiera will enable customers to develop cost-effective, low power edge devices and accelerate go-to-market of custom devices featuring AI capabilities. "This product enables the inclusion of deep learning capabilities in various edge devices that are technologically limited by power consumption and cost, such as consumer appliances (household electrical goods), industrial machinery (construction equipment), surveillance cameras, and broadcasting equipment as well as miniature machinery and robots with limited heat dissipation capabilities."

The post LeapMind Unveils Efficiera Ultra Low-Power AI Inference Accelerator IP appeared first on High-Performance Computing News Analysis | insideHPC.

]]>
83803
Gyrfalcon Acceleration Chips Speed SolidRun AI Inference Server https://insidehpc.com/2020/02/gyrfalcon-acceleration-chips-speed-solidrun-ai-inference-server/ Thu, 27 Feb 2020 08:02:10 +0000 https://insidehpc.com/?p=82916

Today SolidRun introduced a new Arm-based AI inference server optimized for the edge. Highly scalable and modular, the Janux GS31 supports today’s leading neural network frameworks and can be configured with up to 128 Gyrfalcon Lightspeeur SPR2803 AI acceleration chips for unrivaled inference performance for today’s most complex video AI models. “While GPU-based inference servers have seen significant traction for cloud-based applications, there is a growing need for edge-optimized solutions that offer powerful AI inference with less latency than cloud-based solutions. Working with Gyrfalcon and utilizing their industry-proven ASICs has allowed us to create a powerful, cost-effective solution for deploying AI at the Edge that offers seamless scalability.”

The post Gyrfalcon Acceleration Chips Speed SolidRun AI Inference Server appeared first on High-Performance Computing News Analysis | insideHPC.

]]>
82916
NVIDIA Tops MLPerf AI Inference Benchmarks https://insidehpc.com/2019/11/nvidia-tops-mlperf-ai-inference-benchmarks/ Thu, 07 Nov 2019 14:51:32 +0000 https://insidehpc.com/?p=80927

Today NVIDIA posted the fastest results on new benchmarks measuring the performance of AI inference workloads in data centers and at the edge — building on the company’s equally strong position in recent benchmarks measuring AI training. "NVIDIA topped all five benchmarks for both data center-focused scenarios (server and offline), with Turing GPUs providing the highest performance per processor among commercially available entries."

The post NVIDIA Tops MLPerf AI Inference Benchmarks appeared first on High-Performance Computing News Analysis | insideHPC.

]]>
80927
New MLPerf Benchmark Measures Machine Learning Inference Performance https://insidehpc.com/2019/06/new-mlperf-benchmark-measures-machine-learning-inference-performance/ Mon, 24 Jun 2019 15:30:16 +0000 https://insidehpc.com/?p=78537

Today a consortium involving over 40 leading companies and university researchers introduced MLPerf Inference v0.5, the first industry standard machine learning benchmark suite for measuring system performance and power efficiency. "Our goal is to create common and relevant metrics to assess new machine learning software frameworks, hardware accelerators, and cloud and edge computing platforms in real-life situations,” said David Kanter, co-chair of the MLPerf inference working group. “The inference benchmarks will establish a level playing field that even the smallest companies can use to compete.”

The post New MLPerf Benchmark Measures Machine Learning Inference Performance appeared first on High-Performance Computing News Analysis | insideHPC.

]]>
78537
Qualcomm to bring power-efficient AI Inference to the Cloud https://insidehpc.com/2019/04/qualcomm-to-bring-power-efficient-ai-inference-to-the-cloud/ Wed, 10 Apr 2019 15:26:30 +0000 https://insidehpc.com/?p=77253

Today Qualcomm announced that it is bringing the Company’s artificial intelligence expertise to the cloud with the Qualcomm Cloud AI 100. "Our all new Qualcomm Cloud AI 100 accelerator will significantly raise the bar for the AI inference processing relative to any combination of CPUs, GPUs, and/or FPGAs used in today’s data centers,” said Keith Kressin, senior vice president, product management, Qualcomm Technologies, Inc.

The post Qualcomm to bring power-efficient AI Inference to the Cloud appeared first on High-Performance Computing News Analysis | insideHPC.

]]>
77253