low latency storage for AI inference