From 637c4304ca587e1a1b72700c3017abab24152752 Mon Sep 17 00:00:00 2001 From: Ben England Date: Tue, 19 Jul 2016 16:43:39 -0400 Subject: [PATCH] remove bad osd pool create cmd Mark's comments: Ran this through lots of cluster creations and performance tests last night. It appears to be working well in the tests that I have run. Closes #110 --- cluster/ceph.py | 4 ---- 1 file changed, 4 deletions(-) diff --git a/cluster/ceph.py b/cluster/ceph.py index 52cc2edc..01153731 100644 --- a/cluster/ceph.py +++ b/cluster/ceph.py @@ -482,10 +482,6 @@ def mkpool(self, name, profile_name, base_name=None): prefill_object_size = profile.get('prefill_object_size', 0) prefill_time = profile.get('prefill_time', 0) -# common.pdsh(settings.getnodes('head'), 'sudo ceph -c %s osd pool delete %s %s --yes-i-really-really-mean-it' % (self.tmp_conf, name, name)).communicate() - common.pdsh(settings.getnodes('head'), 'sudo %s -c %s osd pool create %s %d %d %s' % (self.ceph_cmd, self.tmp_conf, name, pg_size, pgp_size, erasure_profile), - continue_if_error=False).communicate() - if replication and replication == 'erasure': common.pdsh(settings.getnodes('head'), 'sudo %s -c %s osd pool create %s %d %d erasure %s' % (self.ceph_cmd, self.tmp_conf, name, pg_size, pgp_size, erasure_profile), continue_if_error=False).communicate()