We consider distributed optimization over several devices, each sending incremental model updates to a central server. This setting is considered, for instance, in federated learning. Various schemes have been designed compress the order reduce overall communication cost. However, existing methods suffer from significant slowdown due additional variance ?>0 coming compression operator and as re...