1. Home
  2. News
  3. Announcements
  4. Cortical.io Semantic Folding Approach demonstrates a 2,800x acceleration and 4,300x increase in energy efficiency over BERT

Cortical.io Semantic Folding Approach demonstrates a 2,800x acceleration and 4,300x increase in energy efficiency over BERT

July 12, 2022 – Cortical.io today announced its breakthrough prototype for classifying high volumes of unstructured text. Classifying documents or messages constitutes one of the most fundamental Natural Language Understanding (NLU) functions for business artificial intelligence (AI). The benchmark was carried out on two similar system setups using the same, off-the-shelf, dual AMD-Epyc server hardware. The “BERT” system, a transformer-based machine learning technique for natural language processing, was augmented by a NVidia GPU. The “Semantic Folding” approach utilized a cost comparable number of Xilinx Alveo FPGA accelerator cards.

The goal of the benchmark was to compare the throughput performance of the classification-inference engine of both systems. To measure performance, Cortical.io classified sixteen different sets of data including well-known data sets such as Enron (Kaminski, Farmer, and Lokay), DBPedia, IMDb, PubMed, Reuters (R8, R52), Ohsumed, Web of Science, BBC news text and others.

Staggering results were achieved by the simultaneous application of three distinct innovative steps:

  1. Improving the machine learning approach by applying Semantic Folding.
  2. Using tooling that enabled the concurrent implementation of software, hardware and networking aspects of the Semantic Folding approach.
  3. Using the parallelism of large gate arrays, practically implemented using FPGA technology in form of COTS datacenter hardware from Xilinx.

Benchmark results

BERTSemantic FoldingSemantic Folding
HardwareAMD Epyc Milan + NVidia GPU

(python)

AMD Epyc Milan (java)AMD Epyc Milan +

4 card Xilinx FPGA (binary)

MB / sec0.1818.42528.30
Acceleration1x100x2,865x
Energy / MB2260 mwh15 mwh0.46 mwh
Efficiency1x150x4,298x

 

Benchmark results show that with Semantic Folding, the operations costs can be reduced from several dollars per classifier to a fraction of a cent, making large-scale classification use cases for the first time commercially viable. Example real world workloads could be hate-speech detection for nearly three billion Facebook users or content filtering the Twitter firehose for hundreds of millions of users.

“Efficiency is the new precision in Artificial Intelligence,” said Francisco Webber, CEO at Cortical.io. “While large industries are determined to use less energy, the AI and ML industry is headed in the opposite direction: growing its carbon footprint exponentially. The future of green computing hangs by the thread of high efficiency AI capabilities.”

Read more about the differences between Semantic Folding and other approaches.

All trademarks are the property of their respective owners.

Stay informed!

Subscribe to our newsletter to keep track of what happens at Cortical.io.

You have Successfully Subscribed!