This is mostly a thought experiment, but it's also something I'd like to do real experiments with if I were to find a practical path forward.
Say you're at a festival or a conference, you and the rest of the crowd are all accessing the same scheduling info through the website, perhaps along with a chat/discussion app, all localized to the event, but every person is getting data coming back and forth from some data center in Utah and everybody's having a really shitty high-latency experience on the crappy network. Even though you're all accessing exactly the same data, you're all independently beaming across the country/world.
How can you overcome this problem, in a practical way? Meaning nobody has to really know anything about the underlying technology, they access the website as usual, but under the hood you treat them differently?
Things that come to mind are some potential peer-to-peer solutions, or perhaps a local server that's plenty powerful to handle all the traffic and keep a warm cache for all the traffic. I have a grip on the way this might be architected, but I'm really not sure where to begin on how this kind of thing might be implemented.
Again, this is more of a personal interest than anything, but if there were a practical implementation, I'd love to try something like this with dev.to at a conference or something. The site is already built to be serve a lot of widely shared caches served from the our CDN's edge-nodes.