Phison’s SSD strategy slashes AI training costs from $3 million to $100,000aiDAPTIV+ software shifts AI workloads from GPUs to SSDs efficientlySSDs could replace costly GPUs in massive AI model training
The development of AI models has become increasingly costly as their size and complexity grow, requiring massive computational resources with GPUs playing a central role in handling the workload.
Phison, a key player in portable SSDs, has unveiled a new solution that aims to drastically reduce the cost of training a 1 trillion parameter model by shifting some of the processing load from GPUs to SSDs, bringing the estimated $3 million operational expense down to just $100,000.
Phison’s strategy involves integrating its aiDAPTIV+ software with high-performance SSDs to handle some AI tool processing tasks traditionally managed by GPUs while also incorporating NVIDIA’s GH200 Superchip to enhance performance and keep costs manageable.
AI model growth and the trillion-parameter milestone
Phison expects the AI industry to reach the 1 trillion parameter milestone before 2026.
According to the company, model sizes have expanded rapidly, moving from 69 billion parameters in Llama 2 (2023) to 405 billion with Llama 3.1 (2024), followed by DeepSeek R3’s 671 billion parameters (2025).
If this pattern continues, a trillion-parameter model could be unveiled before the end of 2025, marking a significant leap in AI capabilities.
In addition, it believes that its solution can significantly reduce the number of GPUs needed to run large-scale AI models by shifting some of the processing tasks away from GPUs to the largest SSDs and this approach could bring down training costs to just 3% of current projections (97% savings), or less than 1/25 of the usual operating expenses.
Phison has already collaborated with Maingear to launch AI workstations powered by Intel Xeon W7-3455 CPUs, signaling its commitment to reshaping AI hardware.
As companies seek cost-effective ways to train massive AI models, innovations in SSD technology could play a crucial role in driving efficiency gains while external HDD options remain relevant for long-term data storage.
The push for cheaper AI training solutions gained momentum after DeepSeek made headlines earlier this year when its DeepSeek R1 model demonstrated that cutting-edge AI could be developed at a fraction of the usual cost, with 95% fewer chips and reportedly requiring only $6 million for training.
Via Tweaktown
You may also like
These are the best AI website builders aroundTake a look at the best firewalls on offer right nowSeagate teams with Nvidia to build an NVMe hard drive proof of concept, more than 3 years after its last effort