On Sun, Nov 19, 2006 at 08:00:11PM -0800, keith smith wrote: > > About 18 years ago I was working on the file/server (not client/server) based desktop database that ran in DOS. > > The maker of the product suggested that we put no more than 300 files in each directory because DOS would slow and we would experience a degradation in performance. > > I was thinking of this and wondering if such a limit would exist in Linux in a web server environment. In other words if say I have 500 pages in one directory is that or some other quantity of pages going to cause the web server to slow? *nix has less of a problem with this than DOS did. For any filesystem and operating system there are parameters than can affect performance. You usually won't need to tweak these at all, even with a couple of thousand files. This falls under the Rules of Optimization (paraphrased from memory, sorry): 1. Don't do it. 2. Don't do it yet. 3. Profile. (optimize the right thing) 3. Measure before and after. (did you help or hurt?) If you want to see examples of a NEED to tweak, and how to do it, Google for what people have done for running NNTP servers (absolutely HUGE numbers of small files, which are added and removed continuously). *IF* you need to tweak, you must take into account your particular situation. -- Darrin Chandler | Phoenix BSD Users Group dwchandler@stilyagin.com | http://bsd.phoenix.az.us/ http://www.stilyagin.com/ | --------------------------------------------------- PLUG-discuss mailing list - PLUG-discuss@lists.plug.phoenix.az.us To subscribe, unsubscribe, or to change you mail settings: http://lists.PLUG.phoenix.az.us/mailman/listinfo/plug-discuss