diff --git a/ctdb/doc/ctdb.7.xml b/ctdb/doc/ctdb.7.xml index c8b124d537b..db0a627984c 100644 --- a/ctdb/doc/ctdb.7.xml +++ b/ctdb/doc/ctdb.7.xml @@ -543,7 +543,7 @@ Node 3:/usr/local/etc/ctdb/public_addresses requests are going through the LVSMASTER bottleneck and the majority of the traffic volume (the data in the read replies) goes straight from the processing node back to the clients. For - read-intensive i/o patterns you can acheive very high throughput + read-intensive i/o patterns you can achieve very high throughput rates in this mode. diff --git a/python/samba/tests/kcc/__init__.py b/python/samba/tests/kcc/__init__.py index 5cf6a253b7b..2597388c167 100644 --- a/python/samba/tests/kcc/__init__.py +++ b/python/samba/tests/kcc/__init__.py @@ -69,7 +69,7 @@ class KCCTests(samba.tests.TestCase): def test_verify(self): """check that the KCC generates graphs that pass its own verify - option. This is not a spectacular acheivement when there are + option. This is not a spectacular achievement when there are only a couple of nodes to connect, but it shows something. """ my_kcc = kcc.KCC(unix_now, readonly=True, verify=True,