Tuning orcharhino
Introduction to Performance Tuning
This document provides guidelines for tuning orcharhino for performance and scalability.
Performance Tuning Quickstart
You can tune your orcharhino Server based on expected managed host counts and hardware allocation using built in tuning profiles included in orcharhino that are available using the installation routine’s tuning flag. For more information, see Tuning orcharhino Server with Predefined Profiles in Installing orcharhino Server.
There are four sizes provided based on estimates of the number of managed hosts your orcharhino manages.
You can find the specific tuning settings for each profile in the configuration files contained in /usr/share/foreman-installer/config/foreman.hiera/tuning/sizes
.
Name | Number of managed hosts | Recommend RAM | Recommend Cores |
---|---|---|---|
default |
0 – 5000 |
20 GiB |
4 |
medium |
5000 – 10000 |
32 GiB |
8 |
large |
10000 – 20000 |
64 GiB |
16 |
extra-large |
20000 – 60000 |
128 GiB |
32 |
extra-extra-large |
60000+ |
256 GiB+ |
48+ |
-
Select an installation size:
default
,medium
,large
,extra-large
, orextra-extra-large
. The default value isdefault
. -
Run
orcharhino-installer
:# orcharhino-installer --tuning "My_Installation_Size"
-
Optional: Run a health check. For more information, see Applying configurations.
-
Optional: Tune the Ruby app server directly using the Puma Tuning section. For more information, see Puma Tunings.
Performance Tuning System Requirements
You can find the hardware and software requirements in Preparing your Environment for Installation in Installing orcharhino Server.
Configuring orcharhino for Performance
orcharhino comes with a number of components that communicate with each other. You can tune these components independently of each other to achieve the maximum possible performance for your scenario.
You will see no significant performance difference between orcharhino installed on Enterprise Linux 7 and Enterprise Linux 8.
Applying configurations
In following sections we suggest various tunables and how to apply them. Please always test changing these in non production environment first, with valid backup and with proper outage window as in most of the cases orcharhino restart is required.
It is also a good practice to setup a monitoring before applying any change as it will allow you to evaluate effect of the change. Our testing environment might be too far from what you will see although we are trying hard to mimic real world environment.
If you have changed some systemd service file, you need to notify systemd daemon to reload the configuration:
# systemctl daemon-reload
Restart orcharhino services:
# orcharhino-maintain service restart
If you have changed a configuration file such as /etc/foreman-installer/custom-hiera.yaml
, rerun installer to apply your changes:
# orcharhino-installer
If you need to rerun installer with some new options added:
# orcharhino-installer new options
Optional: After any change, run this quick orcharhino health-check:
# orcharhino-maintain health check
Benchmarking Disk Performance
We are working to update orcharhino-maintain
to only warn the users when its internal quick storage benchmark results in numbers below our recommended throughput.
Also working on an updated benchmark script you can run (which will likely be integrated into orcharhino-maintain
in the future) to get a more accurate real-world storage information.
|
This test does not use direct I/O and will utilize file caching as normal operations would.
You can find our first version of the script storage-benchmark. To execute it, just download the script to your orcharhino, make it executable, and run:
# ./storage-benchmark /var/lib/pulp
As noted in the README block in the script, generally you wish to see on average 100MB/sec or higher in the tests below:
-
Local SSD based storage should give values of 600MB/sec or higher.
-
Spinning disks should give values in the range of 100 – 200MB/sec or higher.
If you see values below this, please open a support ticket for assistance.
Enabling Tuned Profiles
Enterprise Linux 7 enables the tuned daemon by default during installation.
On bare-metal, ATIX AG recommends to run the throughput-performance
tuned profile on orcharhino Server and orcharhino Proxies.
On virtual machines, ATIX AG recommends to run the virtual-guest
profile.
-
Check if
tuned
is running:# systemctl status tuned
-
If
tuned
is not running, enable it:# systemctl enable --now tuned
-
Optional: View a list of available
tuned
profiles:# tuned-adm list
-
Enable a
tuned
profile depending on your scenario:# tuned-adm profile "My_Tuned_Profile"
Puma Tunings
Puma is a ruby application server which is used for serving the Foreman related requests to the clients. For any orcharhino configuration that is supposed to handle a large number of clients or frequent operations, it is important for the Puma to be tuned appropriately.
Puma Threads
Number of Puma threads (per Puma worker) is configured using two values: threads_min
and threads_max
.
Value of threads_min
determines how many threads each worker spawns at a worker start.
Then, as concurrent requests are coming and more threads is needed, worker will be spawning more and more workers up to threads_max
limit.
We recommend setting threads_min
to same value as threads_max
as having fewer Puma threads lead to higher memory usage on your orcharhino Server.
For example, we have compared these two setups using concurrent registrations test:
orcharhino VM with 8 CPUs, 40 GiB RAM | orcharhino VM with 8 CPUs, 40 GiB RAM |
---|---|
|
|
|
|
|
|
Setting the minimum Puma threads to 16
results in about 12% less memory usage as compared to threads_min=0
.
Puma Workers and Threads Auto-Tuning
If you do not provide any Puma workers and thread values with orcharhino-installer
or they are not present in your orcharhino configuration, the orcharhino-installer
configures a balanced number of workers.
It follows this formula:
min(CPU_COUNT * 1.5, RAM_IN_GB - 1.5)
This should be fine for most cases, but with some usage patterns tuning is needed to either limit the amount of resources dedicated to Puma (so other orcharhino components can use these) or for any other reason. Each Puma worker consumes around 1 GiB of RAM.
# cat /etc/systemd/system/foreman.service.d/installer.conf
# systemctl status foreman
Manually tuning Puma workers and threads count
If you decide not to rely on Puma Workers and Threads Auto Tuning, you can apply custom numbers for these tunables. In the example below we are using 2 workers, 5 and 5 threads:
# orcharhino-installer \ --foreman-foreman-service-puma-workers=2 \ --foreman-foreman-service-puma-threads-min=5 \ --foreman-foreman-service-puma-threads-max=5
Apply your changes to orcharhino Server. For more information, see Applying configurations.
Puma Workers and Threads Recommendations
In order to recommend thread and worker configurations for the different tuning profiles, we conducted Puma tuning testing on orcharhino with different tuning profiles. The main test used in this testing was concurrent registration with the following combinations along with different number of workers and threads. Our recommendation is based purely on concurrent registration performance, so it might not reflect your exact use-case. For example, if your setup is very content oriented with lots of publishes and promotes, you might want to limit resources consumed by Puma in favor of Pulp and PostgreSQL.
Name | Number of managed hosts | RAM | Cores | Recommended Puma Threads for both min & max | Recommended Puma Workers |
---|---|---|---|---|---|
default |
0 – 5000 |
20 GiB |
4 |
16 |
4 – 6 |
medium |
5000 – 10000 |
32 GiB |
8 |
16 |
8 – 12 |
large |
10000 – 20000 |
64 GiB |
16 |
16 |
12 – 18 |
extra-large |
20000 – 60000 |
128 GiB |
32 |
16 |
16 – 24 |
extra-extra-large |
60000+ |
256 GiB+ |
48+ |
16 |
20 – 26 |
Tuning number of workers is the more important aspect here and in some case we have seen up to 52% performance increase. Although installer uses 5 min/max threads by default, we recommend 16 threads with all the tuning profiles in the table above. That is because we have seen up to 23% performance increase with 16 threads (14% for 8 and 10% for 32) when compared to setup with 4 threads.
To figure out these recommendations we used concurrent registrations test case which is a very specific use-case. It can be different on your orcharhino which might have more balanced use-case (not only registrations). Keeping default 5 min/max threads is a good choice as well.
These are some of our measurements that lead us to these recommendations:
4 workers, 4 threads | 4 workers, 8 threads | 4 workers, 16 threads | 4 workers, 32 threads | |
---|---|---|---|---|
Improvement |
0% |
14% |
23% |
10% |
Use 4 – 6 workers on a default setup (4 CPUs) – we have seen about 25% higher performance with 5 workers when compared to 2 workers, but 8% lower performance with 8 workers when compared to 2 workers – see table below:
2 workers, 16 threads | 4 workers, 16 threads | 6 workers, 16 threads | 8 workers, 16 threads | |
---|---|---|---|---|
Improvement |
0% |
26% |
22% |
-8% |
Use 8 – 12 workers on a medium setup (8 CPUs) – see table below:
2 workers, 16 threads | 4 workers, 16 threads | 8 workers, 16 threads | 12 workers, 16 threads | 16 workers, 16 threads | |
---|---|---|---|---|---|
Improvement |
0% |
51% |
52% |
52% |
42% |
Use 16 – 24 workers on a 32 CPUs setup (this was tested on a 90 GiB RAM machine and memory turned out to be a factor here as system started swapping – proper extra-large should have 128 GiB), higher number of workers was problematic for higher registration concurrency levels we tested, so we cannot recommend it.
4 workers, 16 threads | 8 workers, 16 threads | 16 workers, 16 threads | 24 workers, 16 threads | 32 workers, 16 threads | 48 workers, 16 threads | |
---|---|---|---|---|---|---|
Improvement |
0% |
37% |
44% |
52% |
too many failures |
too many failures |
Configuring Puma Threads
More threads allow for lower time to register hosts in parallel. For example, we have compared these two setups:
orcharhino VM with 8 CPUs, 40 GiB RAM | orcharhino VM with 8 CPUs, 40 GiB RAM |
---|---|
|
|
|
|
|
|
Using more workers and the same total number of threads results in about 11% of speedup in highly concurrent registrations scenario. Moreover, adding more workers did not consume more CPU and RAM but gets more performance.
Configuring Puma DB Pool
The effective value of $db_pool
is automatically set to equal $foreman::foreman_service_puma_threads_max
.
It is the maximum of $foreman::db_pool
and $foreman::foreman_service_puma_threads_max
but both have default value 5, so any increase to the max threads above 5 automatically increases the database connection pool by the same amount.
If you encounter ActiveRecord::ConnectionTimeoutError: could not obtain a connection from the pool within 5.000 seconds (waited 5.006 seconds); all pooled connections were in use
error in /var/log/foreman/production.log
, you might want to increase this value.
# grep pool /etc/foreman/database.yml pool: 5
Manually tuning db_pool
If you decide not to rely on automatically configured value, you can apply custom number like this:
# orcharhino-installer --foreman-db-pool 10
Apply your changes to orcharhino Server. For more information, see Applying configurations.
Apache HTTPD Performance Tuning
Apache httpd forms a core part of the orcharhino and acts as a web server for handling the requests that are being made through the orcharhino management UI or exposed APIs. To increase the concurrency of the operations, httpd forms the first point where tuning can help to boost the performance of your orcharhino.
Configuring the Open Files Limit for Apache HTTPD
With the tuning in place, Apache httpd can easily open a lot of file descriptors on the server which may exceed the default limit of most of the Linux systems in place. To avoid any kind of issues that may arise as a result of exceeding max open files limit on the system, please create the following file and directory and set the contents of the file as specified in the below given example:
-
Set the maximum open files limit in
/etc/systemd/system/httpd.service.d/limits.conf
:[Service] LimitNOFILE=640000
-
Apply your changes to orcharhino Server. For more information, see Applying configurations.
Tuning Apache Httpd Child Processes
By default, httpd uses event request handling mechanism. When the number of requests to httpd exceeds the maximum number of child processes that can be launched to handle the incoming connections, httpd raises an HTTP 503 Service Unavailable error. Amidst httpd running out of processes to handle, the incoming connections can also result in multiple component failures on your orcharhino services side due to the dependency of some components on the availability of httpd processes.
You can adapt the configuration of httpd event to handle more concurrent requests based on your expected peak load.
Configuring these numbers in |
-
Modify the number of concurrent requests in '/etc/foreman-installer/custom-hiera.yaml' by changing or adding the following lines:
apache::mod::event::serverlimit: 64 apache::mod::event::maxrequestworkers: 1024 apache::mod::event::maxrequestsperchild: 4000
The example is identical to running
orcharhino-installer --tuning=medium
or higher on orcharhino Server. -
Apply your changes to orcharhino Server. For more information, see Applying configurations.
qdrouterd and qpid Tuning
Calculating the maximum open files limit for qdrouterd
In deployments using katello-agent
infrastructure with a large number of content hosts, it may be necessary to increase the maximum open files for qdrouterd.
Calculate the limit for open files in qdrouterd using this formula: (N x 3) + 100
, where N is the number of content hosts.
Each content host may consume up to three file descriptors in the router and 100 file descriptors are required to run the router itself.
The following settings permit orcharhino to scale up to 10.000 content hosts.
-
Set the maximum open files limit in
/etc/foreman-installer/custom-hiera.yaml
:qpid::router::open_file_limit: "My_Value"
The default value is
150100
. -
Apply your changes to orcharhino Server. For more information, see Applying configurations.
Calculating the maximum open files limit for qpidd
In deployments using katello-agent
infrastructure with a large number of content hosts, it may be necessary to increase the maximum open files for qpidd.
Calculate the limit for open files in qpidd using this formula: (N x 4) + 500
, where N is the number of content hosts.
A single content host can consume up to four file descriptors and 500 file descriptors are required for the operations of Broker (a component of qpidd).
-
Set the maximum open files limit in
/etc/foreman-installer/custom-hiera.yaml
:qpid::open_file_limit: "My_Value"
The default value is
65536
. -
Apply your changes to orcharhino Server. For more information, see Applying configurations.
Configuring the Maximum Asynchronous Input-Output Requests
In deployments using katello-agent
infrastructure with a large number of content hosts, it may be necessary to increase the maximum allowable concurrent AIO requests.
You can increase the maximum number of allowed concurrent AIO requests by increasing the kernel parameter fs.aio-max-nr
.
-
Set the value of
fs.aio-max-nr
to the desired maximum in a file in/etc/sysctl.d
:fs.aio-max-nr=My_Maximum_Concurrent_AIO_Requests
Ensure this number is bigger than 33 multiplied by the maximum number of content hosts you plan to register to orcharhino.
-
Apply the changes:
# sysctl -p
-
Optional: Reboot your orcharhino Server to ensure that this change is applied.
Storage Considerations
Ensure you provide enough storage space for /var/lib/qpidd
in advance when you are planning an installation that will use katello-agent
extensively.
On orcharhino Server, /var/lib/qpidd
requires 2MiB disk space per content host.
Configuring the QPID mgmt-pub-interval Parameter
You might see the following error in journal (use journalctl
command to access it) in Enterprise Linux 7:
orcharhino.example.com qpidd[92464]: [Broker] error Channel exception: not-attached: Channel 2 is not attached(/builddir/build/BUILD/qpid-cpp-0.30/src/qpid/amqp_0_10/SessionHandler.cpp: 39 orcharhino.example.com qpidd[92464]: [Protocol] error Connectionqpid.10.1.10.1:5671-10.1.10.1:53790 timed out: closing
This error message appears because qpid maintains management objects for queues, sessions, and connections and recycles them every ten seconds by default. The same object with the same ID is created, deleted, and created again. The old management object is not yet purged, which is why qpid throws this error.
-
Set the
mgmt-pub-interval
parameter in/etc/foreman-installer/custom-hiera.yaml
:qpid::mgmt_pub_interval: 5
-
Apply your changes to orcharhino Server. For more information, see Applying configurations.
Dynflow Tuning
Dynflow is the workflow management system and task orchestrator which is a orcharhino plugin and is used to execute the different tasks of orcharhino in an out-of-order execution manner. Under the conditions when there are a lot of clients checking in on orcharhino and running a number of tasks, Dynflow can take some help from an added tuning specifying how many executors can it launch.
For more information about the tunings involved related to Dynflow, see https://orcharhino.example.com/foreman_tasks/sidekiq
.
orcharhino contains a Dynflow service called dynflow-sidekiq
that performs tasks scheduled by Dynflow.
Sidekiq workers can be grouped into various queues to ensure lots of tasks of one type will not block execution of tasks of other type.
ATIX AG recommends to increase the number of sidekiq workers to scale the Foreman tasking system for bulk concurrent tasks, for example for multiple Content View publications and promotions, content synchronizations, and synchronizations to orcharhino Proxies. There are two options available:
-
You can increase the number of threads used by a worker (worker’s concurrency). This has limited impact for values larger than five due to Ruby implementation of the concurrency of threads.
-
You can increase the number of workers, which is recommended.
-
Increase the number of workers from one worker to three while remaining five threads/concurrency of each:
# orcharhino-installer --foreman-dynflow-worker-instances 3 # optionally, add --foreman-dynflow-worker-concurrency 5
-
Optional: Check if there are three worker services:
# systemctl -a | grep dynflow-sidekiq@worker-[0-9] dynflow-sidekiq@worker-1.service loaded active running Foreman jobs daemon - worker-1 on sidekiq dynflow-sidekiq@worker-2.service loaded active running Foreman jobs daemon - worker-2 on sidekiq dynflow-sidekiq@worker-3.service loaded active running Foreman jobs daemon - worker-3 on sidekiq
PostgreSQL Tuning
PostgreSQL is the primary SQL based database that is used by orcharhino for the storage of persistent context across a wide variety of tasks that orcharhino does. The database sees an extensive usage is usually working on to provide the orcharhino with the data which it needs for its smooth functioning. This makes PostgreSQL a heavily used process which if tuned can have a number of benefits on the overall operational response of orcharhino.
The PostgreSQL authors recommend disabling Transparent Hugepage on servers running PostgreSQL. For more information, see Disable Transparent Hugepage.
You can apply a set of tunings to PostgreSQL to improve its response times, which will modify the postgresql.conf
file.
-
Append
/etc/foreman-installer/custom-hiera.yaml
to tune PostgreSQL:postgresql::server::config_entries: max_connections: 1000 shared_buffers: 2GB work_mem: 8MB autovacuum_vacuum_cost_limit: 2000
You can use this to effectively tune down your orcharhino instance irrespective of a tuning profile.
-
Apply your changes to orcharhino Server. For more information, see Applying configurations.
In the above tuning configuration, there are a certain set of keys which we have altered:
-
max_connections
: The key defines the maximum number of connections that can be accepted by the PostgreSQL processes that are running. -
shared_buffers
: The shared buffers define the memory used by all the active connections inside PostgreSQL to store the data for the different database operations. An optimal value for this will vary between 2 GiB to a maximum of 25% of your total system memory depending upon the frequency of the operations being conducted on orcharhino. -
work_mem
: The work_mem is the memory that is allocated on per process basis for PostgreSQL and is used to store the intermediate results of the operations that are being performed by the process. Setting this value to 8 MB should be more than enough for most of the intensive operations on orcharhino. -
autovacuum_vacuum_cost_limit
: The key defines the cost limit value for the vacuuming operation inside the autovacuum process to clean up the dead tuples inside the database relations. The cost limit defines the number of tuples that can be processed in a single run by the process. ATIX AG recommends setting the value to2000
as it is for the medium, large, extra-large, and extra-extra-large profiles, based on the general load that orcharhino pushes on the PostgreSQL server process.
Benchmarking raw DB performance
PGbench utility (note you may need to resize PostgreSQL data directory /var/lib/pgsql
directory to 100 GiB or what does benchmark take to run) might be used to measure PostgreSQL performance on your system.
Use dnf install postgresql-contrib
to install it.
Choice of filesystem for PostgreSQL data directory might matter as well.
|
orcharhino Proxy Configuration Tuning
orcharhino Proxies are meant to offload part of orcharhino load and provide access to different networks related to distributing content to clients but they can also be used to execute remote execution jobs. What they cannot help with is anything which extensively uses orcharhino API as host registration or package profile update.
orcharhino Proxy Performance Tests
We have measured multiple test cases on multiple orcharhino Proxy configurations:
orcharhino Proxy HW configuration | CPUs | RAM |
---|---|---|
minimal |
4 |
12 GiB |
large |
8 |
24 GiB |
extra large |
16 |
46 GiB |
In a download test where we concurrently downloaded a 40MB repo of 2000 packages on 100, 200, .. 1000 hosts, we saw roughly 50% improvement in average download duration every time when we doubled orcharhino Proxy resources. For more precise numbers, see the table below.
Concurrent downloading hosts | Minimal (4 CPU and 12 GiB RAM) → Large (8 CPU and 24 GiB RAM) | Large (8 CPU and 24 GiB RAM) → Extra Large (16 CPU and 46 GiB RAM) | Minimal (4 CPU and 12 GiB RAM) → Extra Large (16 CPU and 46 GiB RAM) |
---|---|---|---|
Average Improvement |
~ 50% (e.g. for 700 concurrent downloads in average 9 seconds vs. 4.4 seconds per package) |
~ 40% (e.g. for 700 concurrent downloads in average 4.4 seconds vs. 2.5 seconds per package) |
~ 70% (e.g. for 700 concurrent downloads in average 9 seconds vs. 2.5 seconds per package) |
When we compared download performance from orcharhino Server vs. from orcharhino Proxy, we have seen only about 5% speedup, but that is expected as orcharhino Proxy’s main benefit is in getting content closer to geographically distributed clients (or clients in different networks) and in handling part of the load orcharhino Server would have to handle itself. In some smaller hardware configurations (8 CPUs and 24 GiB), orcharhino Server was not able to handle downloads from more than 500 concurrent clients, while a orcharhino Proxy with the same hardware configuration was able to service more than 1000 and possibly even more.
For concurrent registrations, a bottleneck is usually CPU speed, but all configs were able to handle even high concurrency without swapping.
Hardware resources used for orcharhino Proxy have only minimal impact on registration performance.
For example, orcharhino Proxy with 16 CPUs and 46 GiB RAM have at most a 9% registration speed improvement when compared to a orcharhino Proxy with 4 CPUs and 12 GiB RAM.
During periods of very high concurrency, you might experience timeouts in the orcharhino Proxy to orcharhino Server communication.
You can alleviate this by increasing the default timeout using the following tunable in /etc/foreman-installer/custom-hiera.yaml
:
+
apache::mod::proxy::proxy_timeout: 600
We have tested executing Remote Execution jobs via both SSH and Ansible backend on 500, 2000 and 4000 hosts. All configurations were able to handle all of the tests without errors, except for the smallest configuration (4 CPUs and 12 GiB memory) which failed to finish on all 4000 hosts.
In a sync test where we synced Enterprise Linux 6, 7, 8 BaseOS and 8 AppStream we have not seen significant differences among orcharhino Proxy configurations. This will be different for syncing a higher number of content views in parallel.
The text and illustrations on this page are licensed by ATIX AG under a Creative Commons Attribution–Share Alike 3.0 Unported ("CC-BY-SA") license. This page also contains text from the official Foreman documentation which uses the same license ("CC-BY-SA"). |