FAST '20 - Quiver An Informed Storage Cache for Deep Learning-FAST '20 - Quiver An Informed Storage Cache for Deep Learnin

AID:
CID:
视频图片:
作者头像:
弹幕地址:
视频描述:

热门回复:

  • 会喵喵的小汪:We introduce Quiver, an informed storage cache for deep learning training (DLT) jobs in a cluster of GPUs. Quiver employs domain-specific intelligence within the caching layer, to achieve much higher efficiency compared to a generic storage cache. First, Quiver uses a secure hash-based addressing to transparently reuse cached data across multiple jobs and even multiple users operating on the same dataset. Second, by co-designing with the deep learning framework (\eg, PyTorch), Quiver employs a technique of {\em substitutable cache hits} to get more value from the existing contents of the cache, thus avoiding cache thrashing when cache capacity is much smaller than the working set. Third, Quiver dynamically prioritizes cache allocation to jobs that benefit the most from the caching. With a prototype implementation in PyTorch, we show that Quiver can significantly improve throughput of deep learning workloads.
  • 会喵喵的小汪:《FAST '20 - Quiver 一种用于深度学习的信息存储缓存》 Abhishek Vijaya Kumar and Muthian Sivathanu, Microsoft Research India