Cortical.io Semantic Folding Approach demonstrates 2,800x acceleration and 4,300x increase in energ

Substantial cost reduction for NLU implementations enables ubiquitous language intelligence
By: Cortical.io
 
SAN FRANCISCO - July 12, 2022 - PRLog -- Cortical.io announced its breakthrough prototype for classifying high volumes of unstructured text. Classifying documents or messages constitutes one of the most fundamental Natural Language Understanding (NLU) functions for business artificial intelligence (AI). The benchmark was carried out on two similar system setups using the same, off-the-shelve, dual AMD-Epyc server hardware. The "BERT" system, a transformer-based machine learning technique for natural language processing, was augmented by a NVidia GPU. The "Semantic Folding" approach utilized a cost comparable number of Xilinx Alveo FPGA accelerator cards.

The goal of the benchmark was to compare the throughput performance of the classification-inference engine of both systems. To measure performance, Cortical.io classified sixteen different sets of data including well-known data sets such as Enron (Kaminski, Farmer, and Lokay), DBPedia, IMDb, PubMed, Reuters (R8, R52), Ohsumed, Web of Science, BBC news text and others.

Staggering results were achieved by the simultaneous application of three distinct innovative steps:
  1. Improving the machine learning approach by applying Semantic Folding.
  2. Using tooling that enabled the concurrent implementation of software, hardware and networking aspects of the Semantic Folding approach.
  3. Using the parallelism of large gate arrays, practically implemented using FPGA technology in form of COTS datacenter hardware from Xilinx.

Benchmark results

BERT implemented in Python on an AMD Epyc Milan+NVIDA GPU

Performance 0.18 MB / Sec

Acceleration 1x

Power consumption 2,260 mwh / MB

Efficiency 1x

Semantic Folding implemented in Java on an AMD Epyc Milan

Performance 18.2 MB / Sec

Acceleration 100x

Power consumption 15 mwh / MB

Efficiency 150x

Semantic Folding implemented in binary on an AMD Epyc Milan+ 4 card Xilinx FPGA

Performance 528.30 MB / Sec

Acceleration 2856x

Power consumption 0.46 mwh / MB

Efficiency 4298x

Benchmark results show that with Semantic Folding, the operations costs can be reduced from several dollars per classifier to a fraction of a cent, making large-scale classification use cases for the first time commercially viable. Example real world workloads could be hate-speech detection for nearly three billion Facebook users or content filtering the Twitter firehose for hundreds of millions of users.

"Efficiency is the new precision in Artificial Intelligence," said Francisco Webber, CEO at Cortical.io. "While large industries are determined to use less energy, the AI and ML industry is headed in the opposite direction: growing its carbon footprint exponentially. The future of green computing hangs by the thread of high efficiency AI capabilities."

For more information, visit https://www.cortical.io.

Contact
Cortical.io
***@mrb-pr.com
End
Source:Cortical.io
Email:***@mrb-pr.com Email Verified
Tags:Natural Language
Industry:Technology
Location:San Francisco - California - United States
Subject:Reports
Account Email Address Verified     Account Phone Number Verified     Disclaimer     Report Abuse



Like PRLog?
9K2K1K
Click to Share