Yes, I've read the paper and it's interesting, although pretty hairy. It's quite a bit more involved than just taking some samples in the time domain at random intervals. So again, I'm not sure we are all exactly thinking of this in the same manner nor for the same practical applications.
In all cases, and as some have mentioned, even though it may sound too trivial, if your time-domain samples happen to lie outside of some chunks of signal, you're just going to lose this information. Let's say we have a portion of signal that is zero everywhere except there is just a short pulse in between. Let's say all the random samples happen to lie only on zero. No way you're going to reconstruct the lost "pulse". Now if you can repeat random sampling on the same chunk of signal (window) enough times, you will eventually get enough samples. But for a *one-shot* record of limited density samples, you can't guarantee that. I think that is mostly what some of us meant. Now if you happen to know approximately where the information of interest will lie time-wise, that's a different story. Also, if you consider your signal in the frequency domain rather than in the time domain, it's also a bit different. But you'd have to get the spectrum first. Back to square one.
We are also confusing exact reconstruction of a bandwith-limited signal with lossy compression. Although both can lead to useful results, those are not exactly the same in the general case. It's sometimes more useful to know beforehand what you're going to lose spectral-wise than to know that what you're going to lose shouldn't matter much. Different use cases, although admittedly both can be close enough under specific assumptions. Of course I'm simplifying the concept but it's just to provoke some thoughts.
Also, in practice, getting randomly-distributed samples is not possible. You can just hope to approach randomness. And if you are in the digital domain, it will be pseudo-random anyway, given that you always deal with finite time resolution, even if you can get random generators from sophisticated analog stuff. Might be good enough, but it's not quite the theoretical randomness.
Anyway, looks like it has been discussed before in the following thread:
https://www.eevblog.com/forum/projects/shannon-and-nyquist-not-necessary/ For those interested, some papers:
https://statweb.stanford.edu/~donoho/Reports/2004/CompressedSensing091604.pdfhttps://www.researchgate.net/publication/268747805_Reconstruction_of_Sub-Nyquist_Random_Sampling_for_Sparse_and_Multi-Band_Signals