This design enables organizations to train complex models like Llama-2 7B on standard off-the-shelf hardware while ensuring low-latency processing and high-endurance storage (100 DWPD) with SLC NAND.
. . .
According to the company, they have a technique similar to OS-level swapping, where less frequently used data is offloaded from RAM to storage to free up memory for active processes. Likewise, aiDAPTIV+ utilizes high-speed, high-capacity AI100E M.2 SSDs as aiDAPTIVCache, effectively extending GPU memory. Engineers specifically optimize these SSDs to handle the intensive demands of AI workloads, ensuring efficient data management and seamless model training.”- electronics-lab.com
Source: electronics-lab.com