[sheepdog] [PATCH 09/10] sheep/group: fix typo

Ruoyu liangry at ucweb.com
Tue Aug 5 09:30:30 CEST 2014


Signed-off-by: Ruoyu <liangry at ucweb.com>
---
 sheep/gateway.c  |  6 +++---
 sheep/group.c    | 10 +++++-----
 sheep/recovery.c | 14 +++++++-------
 3 files changed, 15 insertions(+), 15 deletions(-)

diff --git a/sheep/gateway.c b/sheep/gateway.c
index 8868bce..49ca12b 100644
--- a/sheep/gateway.c
+++ b/sheep/gateway.c
@@ -102,7 +102,7 @@ out:
 
 /*
  * We spread data strips of req along with its parity strips onto replica for
- * write opertaion. For read we only need to prepare data strip buffers.
+ * write operation. For read we only need to prepare data strip buffers.
  */
 static struct req_iter *prepare_erasure_requests(struct request *req, int *nr)
 {
@@ -392,7 +392,7 @@ again:
 		}
 
 		nr_sent = fi->nr_sent;
-		/* XXX Blinedly close all the connections */
+		/* XXX Blindly close all the connections */
 		for (i = 0; i < nr_sent; i++)
 			sockfd_cache_del(fi->ent[i].nid, fi->ent[i].sfd);
 
@@ -656,7 +656,7 @@ int gateway_write_obj(struct request *req)
 		goto out;
 
 	if (is_data_vid_update(hdr)) {
-		sd_debug("udpate reference counts, %" PRIx64, hdr->obj.oid);
+		sd_debug("update reference counts, %" PRIx64, hdr->obj.oid);
 		update_obj_refcnt(hdr, vids, new_vids, refs);
 	}
 out:
diff --git a/sheep/group.c b/sheep/group.c
index 39309eb..b4db954 100644
--- a/sheep/group.c
+++ b/sheep/group.c
@@ -434,8 +434,8 @@ static bool enough_nodes_gathered(struct cluster_info *cinfo,
 }
 
 /*
- * We have to use memcpy beause some cluster drivers like corosync can't support
- * to send the whole cluster_info structure.
+ * We have to use memcpy because some cluster drivers like corosync can't
+ * support to send the whole cluster_info structure.
  */
 static void cluster_info_copy(struct cluster_info *dst,
 			      const struct cluster_info *src)
@@ -1073,9 +1073,9 @@ static bool cluster_join_check(const struct cluster_info *cinfo)
 	/*
 	 * Sheepdog's recovery code assumes every node have the same epoch
 	 * history. But we don't check epoch history of joining node because:
-	 * 1. inconsist epoch history only happens in the network partition case
-	 *    for the corosync driver, but corosync driver will panic for such
-	 *    case to prevent epoch inconsistency.
+	 * 1. inconsistent epoch history only happens in the network partition
+	 *    case for the corosync driver, but corosync driver will panic for
+	 *    such case to prevent epoch inconsistency.
 	 * 2. checking epoch history with joining node is too expensive and is
 	 *    unneeded for zookeeper driver.
 	 *
diff --git a/sheep/recovery.c b/sheep/recovery.c
index 4648966..0a8f5eb 100644
--- a/sheep/recovery.c
+++ b/sheep/recovery.c
@@ -32,7 +32,7 @@ struct recovery_list_work {
 	uint64_t *oids;
 };
 
-/* for recoverying objects */
+/* for recovering objects */
 struct recovery_obj_work {
 	struct recovery_work base;
 
@@ -609,7 +609,7 @@ main_fn bool oid_in_recovery(uint64_t oid)
 
 	cur = rinfo->cur_vinfo;
 	if (sd_store->exist(oid, local_ec_index(cur, oid))) {
-		sd_debug("the object %" PRIx64 " is already recoverd", oid);
+		sd_debug("the object %" PRIx64 " is already recovered", oid);
 		return false;
 	}
 
@@ -659,7 +659,7 @@ main_fn bool oid_in_recovery(uint64_t oid)
 		sd_debug("%"PRIx64" is not in the recovery list", oid);
 		return false;
 	case RW_NOTIFY_COMPLETION:
-		sd_debug("the object %" PRIx64 " is already recoverd", oid);
+		sd_debug("the object %" PRIx64 " is already recovered", oid);
 		return false;
 	}
 
@@ -719,7 +719,7 @@ static inline bool run_next_rw(void)
 
 	nrinfo = uatomic_xchg_ptr(&next_rinfo, NULL);
 	/*
-	 * When md recovery supersed the reweight or node recovery, we need to
+	 * When md recovery supersedes the reweight or node recovery, we need to
 	 * notify completion.
 	 */
 	if (!nrinfo->notify_complete && cur->notify_complete)
@@ -936,7 +936,7 @@ static void finish_object_list(struct work *work)
 	 *    chances to be blocked for write and also improve reliability.
 	 * 3. For disk failure in node, this is similar to adding a node. All
 	 *    the data on the broken disk will be recovered on other disks in
-	 *    this node. Speedy recoery not only improve data reliability but
+	 *    this node. Speedy recovery not only improve data reliability but
 	 *    also cause less writing blocking on the lost data.
 	 *
 	 * We choose md_nr_disks() * 2 threads for recovery, no rationale.
@@ -1122,7 +1122,7 @@ int start_recovery(struct vnode_info *cur_vinfo, struct vnode_info *old_vinfo,
 		sd_debug("recovery skipped");
 
 		/*
-		 * This is necesary to invoke run_next_rw when
+		 * This is necessary to invoke run_next_rw when
 		 * recovery work is suspended.
 		 */
 		resume_suspended_recovery();
@@ -1163,7 +1163,7 @@ static void queue_recovery_work(struct recovery_info *rinfo)
 		rw->work.done = notify_recovery_completion_main;
 		break;
 	default:
-		panic("unknow recovery state %d", rinfo->state);
+		panic("unknown recovery state %d", rinfo->state);
 		break;
 	}
 
-- 
1.8.3.2





More information about the sheepdog mailing list