A theoretical question:
We all know about the pro's of minifying and combining javascript files in order to reduce HTTP requests to speed up a website. But when popular javascript libraries is used (jQuery for instance), it isn't too stupid to assume that these already have been downloaded to the clients computer from another page.
So what should be preferrered? How does the big guys in the industry handle it?
A) Combine and minify each script into a massive one and serve it from my own CDN.
B) Combine all "self-written" scripts into one file and utilize available CDN's of libraries where possible.
Thanks!
Omar al Zabir runs a site called pageflakes and has a slew of articles on how to improve performance. One of them in particular explains how to compress and combine all kinds of things at the server side before sending to the client.
You can also use css image sprites to reduce HTTP requests. They help a LOT for commonly used images.
*I certainly don't have a huge production site, but these things definitely helped save on bandwidth costs.
"Combine and minify each script into a massive one and serve it from my own CDN"
If you have a CDN, what are we talking about? :) You mean server, right?
"How does the big guys in the industry handle it?"
The big guys always use their own servers.
"...it isn't too stupid to assume that these already have been downloaded to the clients computer from another page."
Unfortunately it is. Facts:
I think it comes down to a couple of things:
There's also a difference between using popular Javascript packages, such as jQuery, and using a personal package, which only visitors who have visited your site will have.
The performance enhancement may occur from two places: 1) browser cache and 2) dns cache, which even if the file isn't stored locally, the dns server has a route that minimizes the request time, or even temporarily serves the file.
I would advise using a CDN and hosting the files locally. Depending on your resources (hardware/bandwidth), you might need to use a CDN anyhow. It'd be nice to use server side schedulers to check on the CDN status and reroute the path when applicable.
Also, take a reminder that some users choose to turn off their browser cache. So minifying your JS is always a plus. You should separate your JS into two files: 1) Needed on Load and 2) Not needed on load. Basically, get the necessary code out there first, to improve the perceived load time. Then load all the other extras (eg slideshows, color changers, etc).
One last point is to make use of the Expires headers, since none of this is important if you don't optimize that. That is what will really reduce the speed for returned visitors with cache enabled. YSlow is a nice Firefox addon that will help evaluate your load performance.
To answer your question: Reduce HTTP requests, but do your own evaluation on the file size of the JS.
(Being Extreme) You don't want one 10MB JS file, or your site will take too long to load. Nor do you want 1000 10KB files, because of the HTTP overhead. Again, use this to illustrate the point that you want a balance between size and number of files - and as I said earlier, package them into performance needed vs wanted.
I think the best approach is to use a minified 'application.js' file (containing all application specific javascript) and then use a service such as Google AJAX Libraries API (found here) to load jQuery, Prototype, etc.
Particular to your question about how the big guys of the industry handle client side scripts, you could always look and see. stackoverflow.com seems fine relying on Google's version of the jquery lib. Others most decidedly do not....
There's no particular reason that getting one script would be faster than getting it split. It happens because browser concurrent downloads are limited, but not by one.
I think the idea should be to handle synchronous UI scripts first, and then the "user activity response" scripts (like validation, etc).
All else given equal, option B looks as the best one.