GigaMACS™ Patented Technology

Gigantor Technologies unveiled a novel computing architecture for sophisticated artificial intelligence models requiring real-time inference.

 

With nothing similar to GigaMACS™, the US Patent Office:

  • On Aug. 24, 2021, issued Patent no. 11,099,854 in only four months; 
  • On Feb. 22, 2022, issued Patent no. 11,256,981 in just seven months, and;
  • On April 01, 2022 approved Patent Application 17/570, 757 in 10 weeks.

 

 

 

 

     
GigaMACS™ accepts high-definition input pixels directly from the source & produces outputs with latency less than a millisecond.   GigaMACS™ is not dependent on RAM-to-RAM computations, so resource bottlenecks are eliminated.   Most cameras return 60 FPS; however, more powerful cameras can produce 1,000 FPS. GigaMACS™ will process input pixels as fast as the camera can deliver them.   Adding complexity to a large model is not a challenge. GigaMACS™ will run the inference without changing the model or slowing down as model size increases.

 

 

GigaMACS™ ResNet-34 Results

Gigantor Technologies tested a GigaMACS™ optimized model on ResNet-34. Ongoing tests are running the 34 nodes on circuits while accepting four (4) Inputs simultaneously.

 

 

We compared GigaMACS™ to common hardware on the market

Inference models running on standard market hardware struggle to process high-definition frames, so models typically use low resolution images for training and execution. When the same weights are applied across a high resolution frame, that hardware cannot keep up with the camera speed and consequently must drop frames. 

The test data describes the Frames per Second and the Latency for two image sizes: a lower resolution of 224x224x3 and a high-definition image of 1920x1080x3. The models used for testing are identical across all hardware and clearly demonstrates how much faster GigaMACS™ can process real-time video streams.   

However, please don’t take our word for it; look over the data and view the test results.

 

 

 

Understanding the Demo Videos

The videos show a simple MNIST inference model to identify numbers. The model is designed to identify moving objects just like a real-world scenario. Gigantor created this simple model for the purpose of demonstrating GigaMACS™ on Xilinx VU9P FPGA against other hardware on the market. But of course, GigaMACS performs at the same lightning speeds on ASICs or FPGAs from any vendor or source.

Unfortunately, nVidia’s hardware struggles to keep up. Even though the model is small, nVidia hardware drops frames. The GigaMACS™ demos show the inference smoothly processing 240 FPS almost instantly without losing any data. 

 

nVidia A100 vs. GigaMACS™

GigaMACS™ processes high-definition images at 240 FPS within a fraction of a millisecond compared to nVidia’s A100, which is processing 28 FPS with a 41-millisecond latency. 

The A100 is dropping 212 frames per second which means 88% of the data is lost. 

Watch nVidia A100 with HD Frames

 

 

 

Tesla V100 vs. GigaMACS™

nVidia’s Tesla V100, is processing HD images at 22 FPS with a 45-millisecond latency and LD micro-images at 44 FPS with 23.4-millisecond latency.

The V100 is dropping 218 FPS which means 90% of the data is lost. 

Watch nVidia Tesla V100 with HD Frames

 

 

 

 

Clock Speeds

 

GigaMACS™ hardware runs significantly slower and cooler while generating higher FPS with near-zero latency. All the nVidia hardware boards are running over a billion cycles to achieve fewer frames per second. GigaMACS™ runs at a sedate 125 MHz, yet still outperforms GPUs running over 1 GHz using 10 times the power and heat. 

 

 

 

 

GigaMACS™ Beats the Competition

Stop sacrificing model size, speed, and performance with standard hardware. Design larger and more complex models without losing data, lowering image resolution, or slowing down by using GigaMACS™.

GigaMACS™ accepts high-definition images as fast as the camera can deliver the frames. No changes are made to the model, so it operates identically to the original but much faster.

GigaMACS™ is the solution to artificial intelligence acceleration and superior model performance.