A copy of this work was available on the public web and has been preserved in the Wayback Machine. The capture dates from 2020; you can also visit the original URL.
The file type is application/pdf
.
Soft-Label Dataset Distillation and Text Dataset Distillation
[article]
2020
arXiv
pre-print
Dataset distillation is a method for reducing dataset sizes by learning a small number of synthetic samples containing all the information of a large dataset. This has several benefits like speeding up model training, reducing energy consumption, and reducing required storage space. Currently, each synthetic sample is assigned a single 'hard' label, and also, dataset distillation can currently only be used with image data. We propose to simultaneously distill both images and their labels, thus
arXiv:1910.02551v3
fatcat:65ybpaulczbubepah7dchlh6vq