Donate to e Foundation | Murena handsets with /e/OS | Own a part of Murena! Learn more

Commit 17891183 authored by Masanari Iida's avatar Masanari Iida Committed by Greg Kroah-Hartman
Browse files

staging: lustre: Fix typo in lustre/include part1



Fix typo in comments within lustre/include.

Signed-off-by: default avatarMasanari Iida <standby24x7@gmail.com>
Signed-off-by: default avatarGreg Kroah-Hartman <gregkh@linuxfoundation.org>
parent d0a0acc3
Loading
Loading
Loading
Loading
+1 −1
Original line number Diff line number Diff line
@@ -1185,7 +1185,7 @@ ldlm_handle2lock_long(const struct lustre_handle *h, __u64 flags)

/**
 * Update Lock Value Block Operations (LVBO) on a resource taking into account
 * data from reqest \a r
 * data from request \a r
 */
static inline int ldlm_res_lvbo_update(struct ldlm_resource *res,
				       struct ptlrpc_request *r, int increase)
+2 −2
Original line number Diff line number Diff line
@@ -147,7 +147,7 @@
#define ldlm_clear_test_lock(_l)        LDLM_CLEAR_FLAG((_l), 1ULL << 19)

/**
 * Immediatelly cancel such locks when they block some other locks. Send
 * Immediately cancel such locks when they block some other locks. Send
 * cancel notification to original lock holder, but expect no reply. This
 * is for clients (like liblustre) that cannot be expected to reliably
 * response to blocking AST. */
@@ -248,7 +248,7 @@

/**
 * A lock contributes to the known minimum size (KMS) calculation until it
 * has finished the part of its cancelation that performs write back on its
 * has finished the part of its cancellation that performs write back on its
 * dirty pages.  It can remain on the granted list during this whole time.
 * Threads racing to update the KMS after performing their writeback need
 * to know to exclude each other's locks from the calculation as they walk
+2 −2
Original line number Diff line number Diff line
@@ -192,9 +192,9 @@ struct obd_export {
	struct obd_import	*exp_imp_reverse;
	struct nid_stat	  *exp_nid_stats;
	struct lprocfs_stats     *exp_md_stats;
	/** Active connetion */
	/** Active connection */
	struct ptlrpc_connection *exp_connection;
	/** Connection count value from last succesful reconnect rpc */
	/** Connection count value from last successful reconnect rpc */
	__u32		     exp_conn_cnt;
	/** Hash list of all ldlm locks granted on this export */
	struct cfs_hash	       *exp_lock_hash;
+4 −4
Original line number Diff line number Diff line
@@ -339,7 +339,7 @@ struct lu_client_seq {
	struct mutex		lcs_mutex;

	/*
	 * Range of allowed for allocation sequeces. When using lu_client_seq on
	 * Range of allowed for allocation sequences. When using lu_client_seq on
	 * clients, this contains meta-sequence range. And for servers this
	 * contains super-sequence range.
	 */
@@ -398,7 +398,7 @@ struct lu_server_seq {
	/* LUSTRE_SEQ_SERVER or LUSTRE_SEQ_CONTROLLER */
	enum lu_mgr_type       lss_type;

	/* Client interafce to request controller */
	/* Client interface to request controller */
	struct lu_client_seq   *lss_cli;

	/* Mutex for protecting allocation */
@@ -568,14 +568,14 @@ fid_build_pdo_res_name(const struct lu_fid *fid, unsigned int hash,
 * finally, when we replace ost_id with FID in data stack.
 *
 * Currently, resid from the old client, whose res[0] = object_id,
 * res[1] = object_seq, is just oposite with Metatdata
 * res[1] = object_seq, is just opposite with Metatdata
 * resid, where, res[0] = fid->f_seq, res[1] = fid->f_oid.
 * To unifiy the resid identification, we will reverse the data
 * resid to keep it same with Metadata resid, i.e.
 *
 * For resid from the old client,
 *    res[0] = objid,  res[1] = 0, still keep the original order,
 *    for compatiblity.
 *    for compatibility.
 *
 * For new resid
 *    res will be built from normal FID directly, i.e. res[0] = f_seq,
+9 −9
Original line number Diff line number Diff line
@@ -445,7 +445,7 @@ struct ptlrpc_reply_state {
	lnet_handle_md_t       rs_md_h;
	atomic_t	   rs_refcount;

	/** Context for the sevice thread */
	/** Context for the service thread */
	struct ptlrpc_svc_ctx *rs_svc_ctx;
	/** Reply buffer (actually sent to the client), encoded if needed */
	struct lustre_msg     *rs_repbuf;       /* wrapper */
@@ -497,7 +497,7 @@ struct ptlrpc_request_pool {
	spinlock_t prp_lock;
	/** list of ptlrpc_request structs */
	struct list_head prp_req_list;
	/** Maximum message size that would fit into a rquest from this pool */
	/** Maximum message size that would fit into a request from this pool */
	int prp_rq_size;
	/** Function to allocate more requests for this pool */
	void (*prp_populate)(struct ptlrpc_request_pool *, int);
@@ -1351,7 +1351,7 @@ struct nrs_orr_data {
	 */
	enum nrs_orr_supp		od_supp;
	/**
	 * Round Robin quantum; the maxium number of RPCs that each request
	 * Round Robin quantum; the maximum number of RPCs that each request
	 * batch for each object or OST can have in a scheduling round.
	 */
	__u16				od_quantum;
@@ -1486,7 +1486,7 @@ struct ptlrpc_nrs_request {
		 */
		struct nrs_fifo_req	fifo;
		/**
		 * CRR-N request defintion
		 * CRR-N request definition
		 */
		struct nrs_crrn_req	crr;
		/** ORR and TRR share the same request definition */
@@ -1550,7 +1550,7 @@ struct ptlrpc_request {
	 * requests in time
	 */
	struct list_head rq_timed_list;
	/** server-side history, used for debuging purposes. */
	/** server-side history, used for debugging purposes. */
	struct list_head rq_history_list;
	/** server-side per-export list */
	struct list_head rq_exp_list;
@@ -1611,7 +1611,7 @@ struct ptlrpc_request {
	enum rq_phase rq_phase; /* one of RQ_PHASE_* */
	enum rq_phase rq_next_phase; /* one of RQ_PHASE_* to be used next */
	atomic_t rq_refcount;/* client-side refcount for SENT race,
				    server-side refcounf for multiple replies */
				    server-side refcount for multiple replies */

	/** Portal to which this request would be sent */
	short rq_request_portal;  /* XXX FIXME bug 249 */
@@ -1637,7 +1637,7 @@ struct ptlrpc_request {
	/** xid */
	__u64 rq_xid;
	/**
	 * List item to for replay list. Not yet commited requests get linked
	 * List item to for replay list. Not yet committed requests get linked
	 * there.
	 * Also see \a rq_replay comment above.
	 */
@@ -1952,7 +1952,7 @@ void _debug_req(struct ptlrpc_request *req,
	__attribute__ ((format (printf, 3, 4)));

/**
 * Helper that decides if we need to print request accordig to current debug
 * Helper that decides if we need to print request according to current debug
 * level settings
 */
#define debug_req(msgdata, mask, cdls, req, fmt, a...)			\
@@ -1966,7 +1966,7 @@ do { \
} while(0)

/**
 * This is the debug print function you need to use to print request sturucture
 * This is the debug print function you need to use to print request structure
 * content into lustre debug log.
 * for most callers (level is a constant) this is resolved at compile time */
#define DEBUG_REQ(level, req, fmt, args...)				   \
Loading