Discussion about this post

User's avatar
Utills' Thoughts and Ideas's avatar

I don’t buy the argument that AI requires access to all stages of data storage for inference. I can understand why this would be required for training but the point of inference is to act on the synthesised learning from all the tiers and then apply that to the current problem being solved.

The problem is more likely that while training you need fast access to the entire data collection across all tiers and there’s not enough slack to continually move data from cold and backup storage to the fast storage without adding a huge amount more of NAND.

Unless you’re saying that enterprises lack the ability to train (or fine tune) and therefore they are are largely running generic pre trained models without that learning having happened and therefore need access to all data while inferencing as the models they are reliant on have never encountered the data stuck in cold storage or backup.

That is likely the case but the right solution is to fine tune or train models first (and once) rather than constantly access expensive storage forever.

Kamal's avatar

Great article. What other constraints exist in the infra layer Main Street isn’t planning for?

1 more comment...

No posts

Ready for more?