Firefox is the only major browser not backed by a billionaire and our independence shapes everything we build. This independence allows us to prioritize bu
I’ve actually been thinking about this a lot. “Save Webpage” is useless nowadays because everything is loaded externally through scripts. What if it saved a timeline of requests and responses somehow and could play it back? This might require recording the entire JS state though… and so much more with browser APIs. Saving just the requests+responses as a cache would fail if the scripting was non-deterministic. Maybe it would make sense to literally save a “recording” of the HTML and CSS changes, playing back only the results of any network requests or JS?
This would be a whole new pipeline to make interactivity work. Emulating a server with cached responses would allow to reuse the JS part of websites and is easier to do. I have no doubt that some pages wouldn’t work and there would be a shitton of security considerations I can’t imagine.
I’ve actually been thinking about this a lot. “Save Webpage” is useless nowadays because everything is loaded externally through scripts. What if it saved a timeline of requests and responses somehow and could play it back? This might require recording the entire JS state though… and so much more with browser APIs. Saving just the requests+responses as a cache would fail if the scripting was non-deterministic. Maybe it would make sense to literally save a “recording” of the HTML and CSS changes, playing back only the results of any network requests or JS?
This would be a whole new pipeline to make interactivity work. Emulating a server with cached responses would allow to reuse the JS part of websites and is easier to do. I have no doubt that some pages wouldn’t work and there would be a shitton of security considerations I can’t imagine.