Apache optimization for images or static files - apache

I have several pages with about 200 images each. When I access them, Apache opens a lot of threads, using more than 1 Gb of RAM. I can see several "httpd" entries on "top" command, each using 0.6% of RAM.
All files are static, small JPG files. I'm using .htaccess for client side caching, but this is not enough since I have several new users each hour, non-cached ones.
My config:
KeepAlive On
MaxKeepAliveRequests 200
KeepAliveTimeout 30
StartServers 1
MinSpareServers 2
MaxSpareServers 4
ServerLimit 300
MaxClients 300
MaxRequestsPerChild 0
MaxRequestWorkers 300
What is the best way to serve lots of static files, per page, with low memory usage? It's a CentOS 7, Apache 2.4.6, almost in default config, except the directives above.
Thanks.

1GB Ram is quite nothing but Apache is depending also a lot about your CPU specs.
I am not sure how many CPU you have but I think the following setting need to be increased to
StartServers 5
MinSpareServers 5
MaxSpareServers 10
Also you can check the loading time using the developer tools in browser where you can check each image how much it takes to load.

Related

Apache2 config for Amazon EC2 c3.2xlarge instance

Currently, I'm using c3.2xlarge Amazon EC2 instance (8 vCPU and 15GB of RAM) with Apache2 installed on it. The server is under high load almost every time, so, sometimes it hangs up and CPU is overloaded for a couple minutes. Health Check reporting that instance is in unhealthy state. However, this is what I see running htop while server in a "working day" mode:
As you can see, the CPU is healthy, but only 2GB of RAM out of 15 is used.
This is mpm_prefork.conf file:
<IfModule mpm_prefork_module>
StartServers 20
MinSpareServers 20
MaxSpareServers 30
ServerLimit 360
MaxClients 360
MaxRequestWorkers 360
MaxConnectionsPerChild 1000
</IfModule>
And this is part of apache2.conf:
Timeout 200
KeepAlive On
MaxKeepAliveRequests 700
KeepAliveTimeout 5
I already changed ServerLimit and MaxClients from 150 to 360 and situation seems better (only 1GB of RAM was used before).
How can I optimize performance with prefork/apache config to archive more RAM usage and less CPU throttling?
Fond out the average amount of memory consumed by each Apache process Y (ps -ef | grep httpd). Subtract the amount of memory that other processes in your VM are consuming from 15GB (add some more buffer if you anticipate that the other processes may consume more memory in future). Let the difference be X. You can then set the MaxClients as approximately X/Y.

new server from Godaddy runny really slow

I purchased a new dedicated server from Godaddy yesterday. The website pages are loading really slow. I have 16 GB of Ram and i7 processor. I am trying to optimize my Apache server for high traffic 10K+ active users. Here is the old and new settings:
OLD:
<IfModule mpm_prefork_module>
StartServers 5
MinSpareServers 5
MaxSpareServers 10
MaxClients 150
MaxRequestsPerChild 0
</IfModule>
New:
<IfModule mpm_prefork_module>
StartServers 5
MinSpareServers 5
MaxSpareServers 10
MaxClients 2500
ServerLimit 2500
MaxRequestsPerChild 0
KeepAlive On
MaxKeepAliveRequests 100
KeepAliveTimeout 3
Timeout 30
</IfModule>
What are the best setting to solve my problem. Note the website is based php mysqli application. Also it has about 3 to 5 images on each page.
Have a look at this page - I use it all the time as a reference.
I also strongly recommend having a play with apache's benchmark tool once you've got it up and running. Keep an eye for a few days on the traffic intake and adjust settings accordingly. Every setup is different - the applications you run can use variable amounts of memory. Maybe the most commonly used page is a low resource page? Maybe people spend a lot of time on single pages? It's all conditional.
Good luck!

Configuring HAProxy+nginx+PHP-FPM to out perform Apache+mod_php

Edit
Running my OS on the VirtualBox was the issue. As soon as I made my OS the native on the disk, I was able to see the performance boost.
Original
I've read a lot of people recommend ditching Apache+mod_php for HAProxy+nginx+PHP-FPM. I'm trying to verify that it's a more efficient setup, but am not seeing the results people describe. Both siege and ab (Apache Benchmark) are showing that Apache at any number of concurrent connections is giving better responses per second, and can support more connections .
I'm running Ubuntu 11.04 server on VirtualBox. It has 10 gigs of space, and 1,344 megs of memory. I used apt-get for installing the programs mentioned above. Here are the related config files with just the important parts included.
haproxy.cfg
global
maxconn 4096
user haproxy
group haproxy
daemon
stats socket /var/run/haproxy.sock mode 0600 level admin
defaults
log global
mode http
option httplog
option dontlognull
retries 3
option redispatch
maxconn 2000
contimeout 5000
clitimeout 50000
srvtimeout 50000
listen tcpcluster *:80
mode tcp
option tcplog
balance roundrobin
server tcp01 192.168.1.199:8080 check
nginx.conf
worker_processes 2;
events {
worker_connections 768;
}
www.conf
pm = dynamic
pm.max_children = 10
pm.min_spare_servers = 2
pm.max_spare_servers = 4
pm.max_requests = 500
apache.conf
<IfModule mpm_prefork_module>
StartServers 5
MinSpareServers 5
MaxSpareServers 10
MaxClients 10
MaxRequestsPerChild 0
</IfModule>
<IfModule mpm_worker_module>
StartServers 2
MinSpareThreads 25
MaxSpareThreads 75
ThreadLimit 64
ThreadsPerChild 25
MaxClients 10
MaxRequestsPerChild 0
</IfModule>
<IfModule mpm_event_module>
StartServers 2
MinSpareThreads 25
MaxSpareThreads 75
ThreadLimit 64
ThreadsPerChild 25
MaxClients 10
MaxRequestsPerChild 0
</IfModule>
Given that PHP-FPM and Apache both have a maximum of 10 children, I would expect that any speed advantage would be visible. In every test I've run (always waiting until load is 0.01 before I run the test), Apache is always able to handle more request more efficiently.
Is there some other optimization that can be made so that the setup best suited to scale will outperform the setup that should not be more efficient?
Use haproxy as a connection concentrator : use "mode http" instead of "mode tcp", use "option http-server-close" and set a server maxconn value well below the worker connections value. You should cross a point where the lower concurrency brings much more performance with much lower RAM usage and better cache efficiency along the whole chain.
BTW, what are the numbers we're talking about ? Do they measure in hundreds or in thousands of requests per second ? Because clearly, the application server will make a real difference only in the higher loads. Obviously if the application runs very slowly, there is no reason to see a difference when replacing the server.

cannot launch more than 1000 apache processes though maxclients and serverlimits is set to 3000

we have a very strong server (32-cores cpu, 96GB ram) and have apache running in prefork mode. our apache2.conf file includes such settings :
<IfModule mpm_prefork_module>
StartServers 2
MinSpareServers 5
MaxSpareServers 20
ServerLimit 3000
MaxClients 3000
MaxRequestsPerChild 1000
</IfModule>
the problem is, when our website is under heavy load (when apache process count reaches 1000 to be precise) (or when setting StartServers beyond 1000), apache2 freezes and needs to be restarted. Yet there is still plenty of ram, cpu is underused and apache process count is far beyond maxclients.
My question is, what should i do to allow apache to reach the maxclients configured in the conf file ?
please consider we have already played with /etc/security/limits.conf to set max opened files and nprocs to 5000 (ulimit -a showed these values were well taken into account).
No errors are shown in /var/log/apache2/error.log
Your Apache server may have a compiled in hard limit. To change it you need to recompile your webserver. The default is 200000 which should be high enough - but packages from your linux distribution may differ.
I would rather recommend to get of static file serving from your webserver. Put an nginx or lighttp server in front of your apache. Let it serve static content (images, css, javascript, etc.) and forward dynamic request to your apache.

OSQA apache memory footprint

I have an OSQA (python / django q&a application) installation serving 8 different sites. The sites are all in development, receiving minimal traffic. The server is a virtual private server with 512 mb of ram.
Apache is only serving dynamic pages with mod_wsgi behind Nginx. I can't stop Apache consuming more and more memory with each request until the server chokes.
I experimented with the configuration parameters without much luck to minimize the memory footprint. With the following mpm_prefork parameters in apache2.conf:
StartServers 2
MinSpareServers 1
MaxSpareServers 4
MaxClients 4
MaxRequestsPerChild 100
2 apache processes start using 4 mb and after the first request there are 4 processes with each nearly 50 mb and with each new request those 4 processes climb steadily up to nearly 200 mb each.
I feel like there is something wrong going on. Any suggestions are greatly appreciated.
KeepAlive Off
MaxSpareThreads 3
MinSpareThreads 1
ServerLimit 3
SetEnvIf X-Forwarded-SSL on HTTPS=1
ThreadsPerChild 2
WSGIDaemonProcess osqaWSGI processes=2 python-path=/web/osqa_server:/web/osqa_server/lib/python2.6 threads=1 maximum-requests=550
WSGIProcessGroup osqaWSGI
Ran httperf against this with 10,000 concurrent hits and it was still standing.