too many sites, change server_names_hash_max_size?
maxbear at gmail.com
Fri Mar 6 11:51:33 MSK 2009
I also don't think it's related to nginx itself. I think it's the problem of
my wordpress supercache rewrite rule. It seems the rewrite rule is not
working correctly. I am working on that to see what's problem now. Thanks
On Fri, Mar 6, 2009 at 3:58 PM, Payam Chychi <pchychi at gmail.com> wrote:
> I dont think this is because of nginx, linux tends to utilize your RAM
> regardless of what your running (may or may not be performance
> I recently came a cross the same issue and after much reading it
> turned out to be the behavior or the OS and not the application
> On Fri, Mar 6, 2009 at 1:56 AM, Igor Sysoev <is at rambler-co.ru> wrote:
> > On Fri, Mar 06, 2009 at 01:20:33PM +0800, Max wrote:
> >> Hello,
> >> My nginx is hosting more than 200 sites. But when I reload it, it said I
> >> need to increase the following size:
> >> server_names_hash_max_size
> >> server_names_hash_bucket_size
> >> So I just increased it to:
> >> server_names_hash_max_size 1024;
> >> server_names_hash_bucket_size 128;
> >> After that, nginx can reload, but my server is always out of memory(4gb
> >> my server). So any body know if I change the above variables, will it
> >> the performance of the server?
> > No, the hash may take up to about 128 * 1024 bytes, i.e., just 128K.
> > What do you mean by "out of memory" ?
> > --
> > Igor Sysoev
> > http://sysoev.ru/en/
> Payam Tarverdyan Chychi
> Network Security Specialist / Network Engineer
-------------- next part --------------
An HTML attachment was scrubbed...
More information about the nginx