Not long ago, IBM Exploration extra a third enhancement to the mix: parallel tensors. The largest bottleneck in AI inferencing is memory. Managing a 70-billion parameter model needs no less than a hundred and fifty gigabytes of memory, practically 2 times just as much as a Nvidia A100 GPU retains. https://charlese789rle2.wikibriefing.com/user