mirror of
https://github.com/samba-team/samba.git
synced 2024-12-27 03:21:53 +03:00
9a9b506d23
CTDB_SAMBA_SKIP_CONF_CHECK and CTDB_SAMBA_CHECK_PORTS. The first is used to tell ctdb to no longer monitoring if the smb.conf file is consistent or not. The second specifies which ports to check that smb is listening on instead of using testparm to figure this out. Since the net, testparm and smbstatus may block indefinitely in some configurations we must have a way to configure ctdb to NOT use any of these three commands in the scripts. These commands should thus never be used in scripts. Signed-off-by: Ronnie Sahlberg <ronniesahlberg@gmail.com> (This used to be ctdb commit 2fe52c7979ecd28250ec4ac195d3c3999916e573)
152 lines
5.7 KiB
Plaintext
152 lines
5.7 KiB
Plaintext
# Options to ctdbd. This is read by /etc/init.d/ctdb
|
|
|
|
# you must specify the location of a shared lock file across all the
|
|
# nodes. This must be on shared storage
|
|
# there is no default
|
|
# CTDB_RECOVERY_LOCK="/some/place/on/shared/storage"
|
|
|
|
# when doing IP takeover you also may specify what network interface
|
|
# to use by default for the public addresses. Otherwise you must
|
|
# specify an interface on each line of the public addresses file
|
|
# there is no default
|
|
# CTDB_PUBLIC_INTERFACE=eth0
|
|
|
|
# Should ctdb do IP takeover? If it should, then specify a file
|
|
# containing the list of public IP addresses that ctdb will manage
|
|
# Note that these IPs must be different from those in $NODES above
|
|
# there is no default.
|
|
# The syntax is one line per public address of the form :
|
|
# <ipaddress>/<netmask> <interface>
|
|
# Example: 10.1.1.1/24 eth0
|
|
#
|
|
# CTDB_PUBLIC_ADDRESSES=/etc/ctdb/public_addresses
|
|
|
|
# Should CTDB present the cluster using a single public ip address to clients
|
|
# and multiplex clients across all CONNECTED nodes ?
|
|
# This is based on LVS
|
|
# When this is enabled, the entire cluster will present one single ip address
|
|
# which clients will connect to.
|
|
# CTDB_LVS_PUBLIC_IP=10.1.1.1
|
|
|
|
|
|
# IPMUX : OBSOLETE use LVS instead
|
|
# Should ctdb implement a single public ip address across the entire cluster
|
|
# and multiplex incoming connections across the connected nodes
|
|
# When using a single public ip you must also specify the public interface!
|
|
# This makes all incoming traffic go through a single ctdb node which
|
|
# will then forward the packets out acros the other nodes. This will
|
|
# impact performance.
|
|
# CTDB_SINGLE_PUBLIC_IP=10.1.1.1
|
|
|
|
# should ctdb manage starting/stopping the Samba service for you?
|
|
# default is to not manage Samba
|
|
# CTDB_MANAGES_SAMBA=yes
|
|
|
|
# should we skip the check of whether the shares that samba is exporting
|
|
# is available or not. This may be useful if there are a huge number of shares
|
|
# since checking each one of them might take a long time.
|
|
# CTDB_SAMBA_SKIP_SHARE_CHECK="yes"
|
|
|
|
# should we skip checking smb.conf for consistency
|
|
# CTDB_SAMBA_SKIP_CONF_CHECK="yes"
|
|
|
|
# specify which ports we should check that there is a daemon listening to
|
|
# by default we use testparm and look in smb.conf to figure out.
|
|
# CTDB_SAMBA_CHECK_PORTS="445"
|
|
|
|
# should ctdb manage starting/stopping Winbind service?
|
|
# if left comented out then it will be autodetected based on smb.conf
|
|
# CTDB_MANAGES_WINBIND=yes
|
|
|
|
# should ctdb manage starting/stopping the VSFTPD service
|
|
# CTDB_MANAGES_VSFTPD=yes
|
|
|
|
# should ctdb manage starting/stopping the ISCSI service
|
|
# CTDB_MANAGES_ISCSI=yes
|
|
|
|
# should ctdb manage starting/stopping the NFS service
|
|
# CTDB_MANAGES_NFS=yes
|
|
|
|
|
|
# you may wish to raise the file descriptor limit for ctdb
|
|
# use a ulimit command here. ctdb needs one file descriptor per
|
|
# connected client (ie. one per connected client in Samba)
|
|
# ulimit -n 10000
|
|
|
|
# the NODES file must be specified or ctdb won't start
|
|
# it should contain a list of IPs that ctdb will use
|
|
# it must be exactly the same on all cluster nodes
|
|
# defaults to /etc/ctdb/nodes
|
|
# CTDB_NODES=/etc/ctdb/nodes
|
|
|
|
# the directory to put the local ctdb database files in
|
|
# defaults to /var/ctdb
|
|
# CTDB_DBDIR=/var/ctdb
|
|
|
|
# the directory to put the local persistent ctdb database files in
|
|
# defaults to /var/ctdb/persistent
|
|
# CTDB_DBDIR_PERSISTENT=/var/ctdb/persistent
|
|
|
|
# the directory where service specific event scripts are stored
|
|
# defaults to /etc/ctdb/events.d
|
|
# CTDB_EVENT_SCRIPT_DIR=/etc/ctdb/events.d
|
|
|
|
# the location of the local ctdb socket
|
|
# defaults to /tmp/ctdb.socket
|
|
# CTDB_SOCKET=/tmp/ctdb.socket
|
|
|
|
# what transport to use. Only tcp is currently supported
|
|
# defaults to tcp
|
|
# CTDB_TRANSPORT="tcp"
|
|
|
|
# When set, this variable makes ctdb monitor the amount of free memory
|
|
# in the system (the second number in the buffers/cache output from free -m).
|
|
# If the amount of free memory drops below this treshold the node will become
|
|
# unhealthy and ctdb and all managed services will be shutdown.
|
|
# Once this occurs, the administrator needs to find the reason for the OOM
|
|
# situation, rectify it and restart ctdb with "service ctdb start"
|
|
# The unit is MByte
|
|
# CTDB_MONITOR_FREE_MEMORY=100
|
|
|
|
# When set to yes, the CTDB node will start in DISABLED mode and not host
|
|
# any public ip addresses. The administrator needs to explicitely enable
|
|
# the node with "ctdb enable"
|
|
# CTDB_START_AS_DISABLED="yes"
|
|
|
|
# LMASTER and RECMASTER capabilities.
|
|
# By default all nodes are capable of both being LMASTER for records and
|
|
# also for taking the RECMASTER role and perform recovery.
|
|
# These parameters can be used to disable these two roles on a node.
|
|
# Note: If there are NO available nodes left in a cluster that can perform
|
|
# the RECMASTER role, the cluster will not be able to recover from a failure
|
|
# and will remain in RECOVERY mode until an RECMASTER capable node becomes
|
|
# available. Same for LMASTER.
|
|
# These parametersd are useful for scenarios where you have one "remote" node
|
|
# in a cluster and you do not want the remote node to be fully participating
|
|
# in the cluster and slow things down.
|
|
# For that case, set both roles to "no" for the remote node on the remote site
|
|
# but leave the roles default to "yes" on the primary nodes in the central
|
|
# datacentre.
|
|
# CTDB_CAPABILITY_RECMASTER=yes
|
|
# CTDB_CAPABILITY_LMASTER=yes
|
|
|
|
# where to log messages
|
|
# the default is /var/log/log.ctdb
|
|
# CTDB_LOGFILE=/var/log/log.ctdb
|
|
|
|
# what debug level to run at. Higher means more verbose
|
|
# the default is 0
|
|
# CTDB_DEBUGLEVEL=0
|
|
|
|
# set any default tuning options for ctdb
|
|
# use CTDB_SET_XXXX=value where XXXX is the name of the tuning
|
|
# variable
|
|
# for example
|
|
# CTDB_SET_TRAVERSETIMEOUT=60
|
|
# you can get a list of variables using "ctdb listvars"
|
|
|
|
|
|
# any other options you might want. Run ctdbd --help for a list
|
|
# CTDB_OPTIONS=
|
|
|