[sheepdog] [PATCH 07/10] sheep/oc: fix typo

Ruoyu liangry at ucweb.com
Tue Aug 5 09:30:28 CEST 2014


Signed-off-by: Ruoyu <liangry at ucweb.com>
---
 sheep/object_cache.c      | 16 ++++++++--------
 sheep/object_list_cache.c |  4 ++--
 2 files changed, 10 insertions(+), 10 deletions(-)

diff --git a/sheep/object_cache.c b/sheep/object_cache.c
index 728eacc..edfac6d 100644
--- a/sheep/object_cache.c
+++ b/sheep/object_cache.c
@@ -19,7 +19,7 @@
  *  0 - 31 (32 bits): data object space
  *  32 - 51 (20 bits): reserved
  *  52 - 59 (8 bits): object flag space
- *  60 - 63 (4 bits): object type indentifier space
+ *  60 - 63 (4 bits): object type identifier space
  */
 #define CACHE_VDI_SHIFT       63 /* if the entry is identified as VDI object */
 #define CACHE_CREATE_SHIFT    59 /* If the entry should be created at backend */
@@ -36,7 +36,7 @@
 
 struct global_cache {
 	uint32_t capacity; /* The real capacity of object cache of this node */
-	uatomic_bool in_reclaim; /* If the relcaimer is working */
+	uatomic_bool in_reclaim; /* If the reclaimer is working */
 };
 
 struct object_cache_entry {
@@ -887,7 +887,7 @@ static void push_object_done(struct work *work)
  *
  * 1. Don't grab cache lock tight so we can serve RW requests while pushing.
  *    It is okay for allow subsequent RW after FLUSH because we only need to
- *    garantee the dirty objects before FLUSH to be pushed.
+ *    grantee the dirty objects before FLUSH to be pushed.
  * 2. Use threaded AIO to boost push performance, such as fsync(2) from VM.
  */
 static int object_cache_push(struct object_cache *oc)
@@ -944,7 +944,7 @@ void object_cache_delete(uint32_t vid)
 	if (!cache)
 		return;
 
-	/* Firstly we free memeory */
+	/* Firstly we free memory */
 	sd_write_lock(&hashtable_lock[h]);
 	hlist_del(&cache->hash);
 	sd_rw_unlock(&hashtable_lock[h]);
@@ -1067,7 +1067,7 @@ bool bypass_object_cache(const struct request *req)
 			object_cache_flush_and_delete(cache);
 			return true;
 		} else  {
-			/* For read requet, we can read cache if any */
+			/* For read request, we can read cache if any */
 			uint64_t idx = object_cache_oid_to_idx(oid);
 
 			if (object_cache_lookup(cache, idx, false, false) == 0)
@@ -1245,8 +1245,8 @@ static int load_cache_object(struct object_cache *cache)
 
 		/*
 		 * We don't know VM's cache type after restarting, so we assume
-		 * that it is writeback and mark all the objects diry to avoid
-		 * false reclaim. Donot try to reclaim at loading phase becaue
+		 * that it is writeback and mark all the objects dirty to avoid
+		 * false reclaim. Don't try to reclaim at loading phase because
 		 * cluster isn't fully working.
 		 */
 		add_to_lru_cache(cache, idx, true);
@@ -1309,7 +1309,7 @@ int object_cache_remove(uint64_t oid)
 	/*
 	 * We assume no other thread will inc the refcount of this entry
 	 * before we call write_lock_cache(). object_cache_remove() is called
-	 * in the DISCARD context, which means nornamly no other read/write
+	 * in the DISCARD context, which means normally no other read/write
 	 * requests.
 	 */
 	assert(refcount_read(&entry->refcnt) == 1);
diff --git a/sheep/object_list_cache.c b/sheep/object_list_cache.c
index eefa40a..b9acaa0 100644
--- a/sheep/object_list_cache.c
+++ b/sheep/object_list_cache.c
@@ -105,7 +105,7 @@ int get_obj_list(const struct sd_req *hdr, struct sd_rsp *rsp, void *data)
 	if (obj_list_cache.tree_version == obj_list_cache.buf_version)
 		goto out;
 
-	/* if that fails grab a write lock for the usually nessecary update */
+	/* if that fails grab a write lock for the usually necessary update */
 	sd_rw_unlock(&obj_list_cache.lock);
 	sd_write_lock(&obj_list_cache.lock);
 	if (obj_list_cache.tree_version == obj_list_cache.buf_version)
@@ -143,7 +143,7 @@ static void objlist_deletion_work(struct work *work)
 	 * Before reclaiming the cache belonging to the VDI just deleted,
 	 * we should test whether the VDI is exist, because after some node
 	 * deleting it and before the notification is sent to all the node,
-	 * another node may issus a VDI creation event and reused the VDI id
+	 * another node may issue a VDI creation event and reused the VDI id
 	 * again, in which case we should not reclaim the cached entry.
 	 */
 	if (vdi_exist(vid)) {
-- 
1.8.3.2





More information about the sheepdog mailing list