File bug-1066710_pacemaker-systemd-unit-files-enable-TasksMax=infinity.patch of Package pacemaker

commit 508ad52e7000a1af880353fee3d307b579b62cbb
Author: Ken Gaillot <kgaillot@redhat.com>
Date:   Wed Nov 15 17:06:19 2017 -0600

    Build: systemd unit files: enable TasksMax=infinity

diff --git a/lrmd/pacemaker_remote.service.in b/lrmd/pacemaker_remote.service.in
index d5717f6d0..833e94613 100644
--- a/lrmd/pacemaker_remote.service.in
+++ b/lrmd/pacemaker_remote.service.in
@@ -21,9 +21,12 @@ EnvironmentFile=-@CONFIGDIR@/sbd
 
 ExecStart=@sbindir@/pacemaker_remoted
 
-# Uncomment TasksMax if your systemd version supports it.
-# Only systemd v227 and above support this option.
-#TasksMax=infinity
+# Systemd v227 and above can limit the number of processes spawned by a
+# service. That is a bad idea for an HA cluster resource manager, so disable it
+# by default. The administrator can create a local override if they really want
+# a limit. If your systemd version does not support TasksMax, and you want to
+# get rid of the resulting log warnings, comment out this option.
+TasksMax=infinity
 
 # Pacemaker Remote can exit only after all managed services have shut down;
 # an HA database could conceivably take even longer than this 
diff --git a/mcp/pacemaker.service.in b/mcp/pacemaker.service.in
index 516de0f82..66182d135 100644
--- a/mcp/pacemaker.service.in
+++ b/mcp/pacemaker.service.in
@@ -41,9 +41,12 @@ SuccessExitStatus=100
 
 ExecStart=@sbindir@/pacemakerd -f
 
-# Uncomment TasksMax if your systemd version supports it.
-# Only systemd v227 and above support this option.
-#TasksMax=infinity
+# Systemd v227 and above can limit the number of processes spawned by a
+# service. That is a bad idea for an HA cluster resource manager, so disable it
+# by default. The administrator can create a local override if they really want
+# a limit. If your systemd version does not support TasksMax, and you want to
+# get rid of the resulting log warnings, comment out this option.
+TasksMax=infinity
 
 # If pacemakerd doesn't stop, it's probably waiting on a cluster
 # resource.  Sending -KILL will just get the node fenced