All of lore.kernel.org
 help / color / mirror / Atom feed
From: NeilBrown <neilb@suse.com>
To: lustre-devel@lists.lustre.org
Subject: [lustre-devel] [PATCH 4/4] lustre: clear up white space in osc header files.
Date: Mon, 10 Dec 2018 11:46:16 +1100	[thread overview]
Message-ID: <154440277682.29887.12055691371699460100.stgit@noble> (raw)
In-Reply-To: <154440246926.29887.1667505755325904791.stgit@noble>

Also change /** to /*.  In the kernel, /** should only be
used for kernel-doc comments.

Signed-off-by: NeilBrown <neilb@suse.com>
---
 .../staging/lustre/lustre/osc/osc_cl_internal.h    |  319 ++++++++++----------
 drivers/staging/lustre/lustre/osc/osc_internal.h   |   96 +++---
 2 files changed, 209 insertions(+), 206 deletions(-)

diff --git a/drivers/staging/lustre/lustre/osc/osc_cl_internal.h b/drivers/staging/lustre/lustre/osc/osc_cl_internal.h
index 4b0180946134..077a2b183634 100644
--- a/drivers/staging/lustre/lustre/osc/osc_cl_internal.h
+++ b/drivers/staging/lustre/lustre/osc/osc_cl_internal.h
@@ -48,62 +48,63 @@
 #include <cl_object.h>
 #include "osc_internal.h"
 
-/** \defgroup osc osc
+/* \defgroup osc osc
  *  @{
  */
 
 struct osc_extent;
 
-/**
+/*
  * State maintained by osc layer for each IO context.
  */
 struct osc_io {
-	/** super class */
-	struct cl_io_slice oi_cl;
-	/** true if this io is lockless. */
+	/* super class */
+	struct cl_io_slice	oi_cl;
+	/* true if this io is lockless. */
 	unsigned int		oi_lockless:1,
-	/** true if this io is counted as active IO */
+	/* true if this io is counted as active IO */
 				oi_is_active:1;
-	/** how many LRU pages are reserved for this IO */
+	/* how many LRU pages are reserved for this IO */
 	unsigned long		oi_lru_reserved;
 
-	/** active extents, we know how many bytes is going to be written,
-	 * so having an active extent will prevent it from being fragmented
+	/* active extents, we know how many bytes is going to be
+	 * written, so having an active extent will prevent it from
+	 * being fragmented
 	 */
-	struct osc_extent *oi_active;
-	/** partially truncated extent, we need to hold this extent to prevent
-	 * page writeback from happening.
+	struct osc_extent	*oi_active;
+	/* partially truncated extent, we need to hold this extent to
+	 * prevent page writeback from happening.
 	 */
-	struct osc_extent *oi_trunc;
+	struct osc_extent	*oi_trunc;
 
-	/** write osc_lock for this IO, used by osc_extent_find(). */
-	struct osc_lock   *oi_write_osclock;
-	struct obdo	oi_oa;
+	/* write osc_lock for this IO, used by osc_extent_find(). */
+	struct osc_lock		*oi_write_osclock;
+	struct obdo		oi_oa;
 	struct osc_async_cbargs {
-		bool		  opc_rpc_sent;
-		int	       opc_rc;
+		bool			opc_rpc_sent;
+		int			opc_rc;
 		struct completion	opc_sync;
-	} oi_cbarg;
+	}			oi_cbarg;
 };
 
-/**
+/*
  * State maintained by osc layer for the duration of a system call.
  */
 struct osc_session {
-	struct osc_io       os_io;
+	struct osc_io		os_io;
 };
 
 #define OTI_PVEC_SIZE 256
 struct osc_thread_info {
-	struct ldlm_res_id      oti_resname;
+	struct ldlm_res_id	oti_resname;
 	union ldlm_policy_data	oti_policy;
-	struct cl_lock_descr    oti_descr;
-	struct cl_attr	  oti_attr;
-	struct lustre_handle    oti_handle;
-	struct cl_page_list     oti_plist;
+	struct cl_lock_descr	oti_descr;
+	struct cl_attr		oti_attr;
+	struct lustre_handle	oti_handle;
+	struct cl_page_list	oti_plist;
 	struct cl_io		oti_io;
 	void			*oti_pvec[OTI_PVEC_SIZE];
-	/**
+	/*
 	 * Fields used by cl_lock_discard_pages().
 	 */
 	pgoff_t			oti_next_index;
@@ -114,44 +115,44 @@ struct osc_thread_info {
 };
 
 struct osc_object {
-	struct cl_object   oo_cl;
-	struct lov_oinfo  *oo_oinfo;
-	/**
+	struct cl_object	oo_cl;
+	struct lov_oinfo	*oo_oinfo;
+	/*
 	 * True if locking against this stripe got -EUSERS.
 	 */
-	int		oo_contended;
-	unsigned long	 oo_contention_time;
-	/**
+	int			oo_contended;
+	unsigned long		oo_contention_time;
+	/*
 	 * used by the osc to keep track of what objects to build into rpcs.
 	 * Protected by client_obd->cli_loi_list_lock.
 	 */
-	struct list_head	   oo_ready_item;
-	struct list_head	   oo_hp_ready_item;
-	struct list_head	   oo_write_item;
-	struct list_head	   oo_read_item;
+	struct list_head	oo_ready_item;
+	struct list_head	oo_hp_ready_item;
+	struct list_head	oo_write_item;
+	struct list_head	oo_read_item;
 
-	/**
+	/*
 	 * extent is a red black tree to manage (async) dirty pages.
 	 */
-	struct rb_root       oo_root;
-	/**
+	struct rb_root		oo_root;
+	/*
 	 * Manage write(dirty) extents.
 	 */
-	struct list_head	   oo_hp_exts; /* list of hp extents */
-	struct list_head	   oo_urgent_exts; /* list of writeback extents */
-	struct list_head	   oo_full_exts;
+	struct list_head	oo_hp_exts; /* list of hp extents */
+	struct list_head	oo_urgent_exts; /* list of writeback extents */
+	struct list_head	oo_full_exts;
 
-	struct list_head	   oo_reading_exts;
+	struct list_head	oo_reading_exts;
 
-	atomic_t	 oo_nr_reads;
-	atomic_t	 oo_nr_writes;
+	atomic_t		oo_nr_reads;
+	atomic_t		oo_nr_writes;
 
-	/** Protect extent tree. Will be used to protect
+	/* Protect extent tree. Will be used to protect
 	 * oo_{read|write}_pages soon.
 	 */
-	spinlock_t	    oo_lock;
+	spinlock_t		oo_lock;
 
-	/**
+	/*
 	 * Radix tree for caching pages
 	 */
 	struct radix_tree_root	oo_tree;
@@ -162,7 +163,7 @@ struct osc_object {
 	spinlock_t		oo_ol_spin;
 	struct list_head	oo_ol_list;
 
-	/** number of active IOs of this object */
+	/* number of active IOs of this object */
 	atomic_t		oo_nr_ios;
 	wait_queue_head_t	oo_io_waitq;
 };
@@ -207,7 +208,7 @@ enum osc_lock_state {
 	OLS_CANCELLED
 };
 
-/**
+/*
  * osc-private state of cl_lock.
  *
  * Interaction with DLM.
@@ -250,37 +251,37 @@ enum osc_lock_state {
  * future.
  */
 struct osc_lock {
-	struct cl_lock_slice     ols_cl;
-	/** Internal lock to protect states, etc. */
-	spinlock_t		ols_lock;
-	/** Owner sleeps on this channel for state change */
-	struct cl_sync_io	*ols_owner;
-	/** waiting list for this lock to be cancelled */
-	struct list_head	ols_waiting_list;
-	/** wait entry of ols_waiting_list */
-	struct list_head	ols_wait_entry;
-	/** list entry for osc_object::oo_ol_list */
-	struct list_head	ols_nextlock_oscobj;
-
-	/** underlying DLM lock */
-	struct ldlm_lock	*ols_dlmlock;
-	/** DLM flags with which osc_lock::ols_lock was enqueued */
-	__u64		    ols_flags;
-	/** osc_lock::ols_lock handle */
-	struct lustre_handle     ols_handle;
-	struct ldlm_enqueue_info ols_einfo;
-	enum osc_lock_state      ols_state;
-	/** lock value block */
-	struct ost_lvb		ols_lvb;
-
-	/**
+	struct cl_lock_slice		ols_cl;
+	/* Internal lock to protect states, etc. */
+	spinlock_t			ols_lock;
+	/* Owner sleeps on this channel for state change */
+	struct cl_sync_io		*ols_owner;
+	/* waiting list for this lock to be cancelled */
+	struct list_head		ols_waiting_list;
+	/* wait entry of ols_waiting_list */
+	struct list_head		ols_wait_entry;
+	/* list entry for osc_object::oo_ol_list */
+	struct list_head		ols_nextlock_oscobj;
+
+	/* underlying DLM lock */
+	struct ldlm_lock		*ols_dlmlock;
+	/* DLM flags with which osc_lock::ols_lock was enqueued */
+	__u64				ols_flags;
+	/* osc_lock::ols_lock handle */
+	struct lustre_handle		ols_handle;
+	struct ldlm_enqueue_info	ols_einfo;
+	enum osc_lock_state		ols_state;
+	/* lock value block */
+	struct ost_lvb			ols_lvb;
+
+	/*
 	 * true, if ldlm_lock_addref() was called against
 	 * osc_lock::ols_lock. This is used for sanity checking.
 	 *
 	 * \see osc_lock::ols_has_ref
 	 */
-	unsigned		  ols_hold :1,
-	/**
+	unsigned			ols_hold :1,
+	/*
 	 * this is much like osc_lock::ols_hold, except that this bit is
 	 * cleared _after_ reference in released in osc_lock_unuse(). This
 	 * fine distinction is needed because:
@@ -295,13 +296,13 @@ struct osc_lock {
 	 *       release lock reference (so that a flag is needed that is
 	 *       cleared before ldlm_lock_decref() is called).
 	 */
-				 ols_has_ref:1,
-	/**
+					ols_has_ref:1,
+	/*
 	 * inherit the lockless attribute from top level cl_io.
 	 * If true, osc_lock_enqueue is able to tolerate the -EUSERS error.
 	 */
-				 ols_locklessable:1,
-	/**
+					ols_locklessable:1,
+	/*
 	 * if set, the osc_lock is a glimpse lock. For glimpse locks, we treat
 	 * the EVAVAIL error as tolerable, this will make upper logic happy
 	 * to wait all glimpse locks to each OSTs to be completed.
@@ -309,53 +310,53 @@ struct osc_lock {
 	 * granted.
 	 * Glimpse lock should be destroyed immediately after use.
 	 */
-				 ols_glimpse:1,
-	/**
+					ols_glimpse:1,
+	/*
 	 * For async glimpse lock.
 	 */
-				 ols_agl:1;
+					ols_agl:1;
 };
 
-/**
+/*
  * Page state private for osc layer.
  */
 struct osc_page {
-	struct cl_page_slice  ops_cl;
-	/**
+	struct cl_page_slice	ops_cl;
+	/*
 	 * Page queues used by osc to detect when RPC can be formed.
 	 */
-	struct osc_async_page ops_oap;
-	/**
+	struct osc_async_page	ops_oap;
+	/*
 	 * An offset within page from which next transfer starts. This is used
 	 * by cl_page_clip() to submit partial page transfers.
 	 */
-	int		   ops_from;
-	/**
+	int			ops_from;
+	/*
 	 * An offset within page at which next transfer ends.
 	 *
 	 * \see osc_page::ops_from.
 	 */
-	int		   ops_to;
-	/**
+	int			ops_to;
+	/*
 	 * Boolean, true iff page is under transfer. Used for sanity checking.
 	 */
-	unsigned	      ops_transfer_pinned:1,
-	/**
+	unsigned		ops_transfer_pinned:1,
+	/*
 	 * in LRU?
 	 */
-			      ops_in_lru:1,
-	/**
+				ops_in_lru:1,
+	/*
 	 * Set if the page must be transferred with OBD_BRW_SRVLOCK.
 	 */
-			      ops_srvlock:1;
-	/**
+				ops_srvlock:1;
+	/*
 	 * lru page list. See osc_lru_{del|use}() in osc_page.c for usage.
 	 */
-	struct list_head	      ops_lru;
-	/**
+	struct list_head	ops_lru;
+	/*
 	 * Submit time - the time when the page is starting RPC. For debugging.
 	 */
-	unsigned long	    ops_submit_time;
+	unsigned long		ops_submit_time;
 };
 
 extern struct kmem_cache *osc_lock_kmem;
@@ -422,7 +423,7 @@ int  osc_object_is_contended(struct osc_object *obj);
 
 int  osc_lock_is_lockless(const struct osc_lock *olck);
 
-/*****************************************************************************
+/****************************************************************************
  *
  * Accessors.
  *
@@ -558,17 +559,17 @@ static inline int osc_io_srvlock(struct osc_io *oio)
 }
 
 enum osc_extent_state {
-	OES_INV       = 0, /** extent is just initialized or destroyed */
-	OES_ACTIVE    = 1, /** process is using this extent */
-	OES_CACHE     = 2, /** extent is ready for IO */
-	OES_LOCKING   = 3, /** locking page to prepare IO */
-	OES_LOCK_DONE = 4, /** locking finished, ready to send */
-	OES_RPC       = 5, /** in RPC */
-	OES_TRUNC     = 6, /** being truncated */
+	OES_INV       = 0, /* extent is just initialized or destroyed */
+	OES_ACTIVE    = 1, /* process is using this extent */
+	OES_CACHE     = 2, /* extent is ready for IO */
+	OES_LOCKING   = 3, /* locking page to prepare IO */
+	OES_LOCK_DONE = 4, /* locking finished, ready to send */
+	OES_RPC       = 5, /* in RPC */
+	OES_TRUNC     = 6, /* being truncated */
 	OES_STATE_MAX
 };
 
-/**
+/*
  * osc_extent data to manage dirty pages.
  * osc_extent has the following attributes:
  * 1. all pages in the same must be in one RPC in write back;
@@ -584,85 +585,85 @@ enum osc_extent_state {
  * page lock -> cl_loi_list_lock -> object lock(osc_object::oo_lock)
  */
 struct osc_extent {
-	/** red-black tree node */
-	struct rb_node     oe_node;
-	/** osc_object of this extent */
-	struct osc_object *oe_obj;
-	/** refcount, removed from red-black tree if reaches zero. */
-	atomic_t       oe_refc;
-	/** busy if non-zero */
-	atomic_t       oe_users;
-	/** link list of osc_object's oo_{hp|urgent|locking}_exts. */
-	struct list_head	 oe_link;
-	/** state of this extent */
+	/* red-black tree node */
+	struct rb_node		oe_node;
+	/* osc_object of this extent */
+	struct osc_object	*oe_obj;
+	/* refcount, removed from red-black tree if reaches zero. */
+	atomic_t		oe_refc;
+	/* busy if non-zero */
+	atomic_t		oe_users;
+	/* link list of osc_object's oo_{hp|urgent|locking}_exts. */
+	struct list_head	oe_link;
+	/* state of this extent */
 	enum osc_extent_state	oe_state;
-	/** flags for this extent. */
-	unsigned int       oe_intree:1,
-	/** 0 is write, 1 is read */
-			   oe_rw:1,
-	/** sync extent, queued by osc_queue_sync_pages() */
+	/* flags for this extent. */
+	unsigned int		oe_intree:1,
+	/* 0 is write, 1 is read */
+				oe_rw:1,
+	/* sync extent, queued by osc_queue_sync_pages() */
 				oe_sync:1,
-	/** set if this extent has partial, sync pages.
+	/* set if this extent has partial, sync pages.
 	 * Extents with partial page(s) can't merge with others in RPC
 	 */
 				oe_no_merge:1,
-			   oe_srvlock:1,
-			   oe_memalloc:1,
-	/** an ACTIVE extent is going to be truncated, so when this extent
+				oe_srvlock:1,
+				oe_memalloc:1,
+	/* an ACTIVE extent is going to be truncated, so when this extent
 	 * is released, it will turn into TRUNC state instead of CACHE.
 	 */
-			   oe_trunc_pending:1,
-	/** this extent should be written asap and someone may wait for the
+				oe_trunc_pending:1,
+	/* this extent should be written asap and someone may wait for the
 	 * write to finish. This bit is usually set along with urgent if
 	 * the extent was CACHE state.
 	 * fsync_wait extent can't be merged because new extent region may
 	 * exceed fsync range.
 	 */
-			   oe_fsync_wait:1,
-	/** covering lock is being canceled */
-			   oe_hp:1,
-	/** this extent should be written back asap. set if one of pages is
+				oe_fsync_wait:1,
+	/* covering lock is being canceled */
+				oe_hp:1,
+	/* this extent should be written back asap. set if one of pages is
 	 * called by page WB daemon, or sync write or reading requests.
 	 */
-			   oe_urgent:1;
-	/** how many grants allocated for this extent.
+				oe_urgent:1;
+	/* how many grants allocated for this extent.
 	 *  Grant allocated for this extent. There is no grant allocated
 	 *  for reading extents and sync write extents.
 	 */
-	unsigned int       oe_grants;
-	/** # of dirty pages in this extent */
-	unsigned int       oe_nr_pages;
-	/** list of pending oap pages. Pages in this list are NOT sorted. */
-	struct list_head	 oe_pages;
-	/** Since an extent has to be written out in atomic, this is used to
+	unsigned int		oe_grants;
+	/* # of dirty pages in this extent */
+	unsigned int		oe_nr_pages;
+	/* list of pending oap pages. Pages in this list are NOT sorted. */
+	struct list_head	oe_pages;
+	/* Since an extent has to be written out in atomic, this is used to
 	 * remember the next page need to be locked to write this extent out.
 	 * Not used right now.
 	 */
-	struct osc_page   *oe_next_page;
-	/** start and end index of this extent, include start and end
+	struct osc_page		*oe_next_page;
+	/* start and end index of this extent, include start and end
 	 * themselves. Page offset here is the page index of osc_pages.
 	 * oe_start is used as keyword for red-black tree.
 	 */
-	pgoff_t	    oe_start;
-	pgoff_t	    oe_end;
-	/** maximum ending index of this extent, this is limited by
+	pgoff_t			oe_start;
+	pgoff_t			oe_end;
+	/* maximum ending index of this extent, this is limited by
 	 * max_pages_per_rpc, lock extent and chunk size.
 	 */
-	pgoff_t	    oe_max_end;
-	/** waitqueue - for those who want to be notified if this extent's
+	pgoff_t			oe_max_end;
+	/* waitqueue - for those who want to be notified if this extent's
 	 * state has changed.
 	 */
 	wait_queue_head_t	oe_waitq;
-	/** lock covering this extent */
+	/* lock covering this extent */
 	struct ldlm_lock	*oe_dlmlock;
-	/** terminator of this extent. Must be true if this extent is in IO. */
+	/* terminator of this extent. Must be true if this extent is in IO. */
 	struct task_struct	*oe_owner;
-	/** return value of writeback. If somebody is waiting for this extent,
+	/* return value of writeback. If somebody is waiting for this extent,
 	 * this value can be known by outside world.
 	 */
-	int		oe_rc;
-	/** max pages per rpc when this extent was created */
-	unsigned int       oe_mppr;
+	int			oe_rc;
+	/* max pages per rpc when this extent was created */
+	unsigned int		oe_mppr;
 };
 
 int osc_extent_finish(const struct lu_env *env, struct osc_extent *ext,
@@ -677,6 +678,6 @@ typedef int (*osc_page_gang_cbt)(const struct lu_env *, struct cl_io *,
 int osc_page_gang_lookup(const struct lu_env *env, struct cl_io *io,
 			 struct osc_object *osc, pgoff_t start, pgoff_t end,
 			 osc_page_gang_cbt cb, void *cbdata);
-/** @} osc */
+/* @} osc */
 
 #endif /* OSC_CL_INTERNAL_H */
diff --git a/drivers/staging/lustre/lustre/osc/osc_internal.h b/drivers/staging/lustre/lustre/osc/osc_internal.h
index 55054d705a01..0de8a3ee826d 100644
--- a/drivers/staging/lustre/lustre/osc/osc_internal.h
+++ b/drivers/staging/lustre/lustre/osc/osc_internal.h
@@ -43,54 +43,56 @@ extern struct ptlrpc_request_pool *osc_rq_pool;
 struct lu_env;
 
 enum async_flags {
-	ASYNC_READY = 0x1, /* ap_make_ready will not be called before this
-			    * page is added to an rpc
-			    */
-	ASYNC_URGENT = 0x2, /* page must be put into an RPC before return */
-	ASYNC_COUNT_STABLE = 0x4, /* ap_refresh_count will not be called
-				   * to give the caller a chance to update
-				   * or cancel the size of the io
-				   */
+	ASYNC_READY		= 0x1, /* ap_make_ready will not be
+					* called before this page is
+					* added to an rpc
+					*/
+	ASYNC_URGENT		= 0x2, /* page must be put into an RPC
+					* before return */
+	ASYNC_COUNT_STABLE	= 0x4, /* ap_refresh_count will not be
+					* called to give the caller a
+					* chance to update or cancel
+					* the size of the io
+					*/
 	ASYNC_HP = 0x10,
 };
 
 struct osc_async_page {
-	int		     oap_magic;
-	unsigned short	  oap_cmd;
-	unsigned short	  oap_interrupted:1;
+	int				oap_magic;
+	unsigned short			oap_cmd;
+	unsigned short			oap_interrupted:1;
 
-	struct list_head	      oap_pending_item;
-	struct list_head	      oap_rpc_item;
+	struct list_head		oap_pending_item;
+	struct list_head		oap_rpc_item;
 
-	u64		 oap_obj_off;
-	unsigned int		oap_page_off;
-	enum async_flags	oap_async_flags;
+	u64				oap_obj_off;
+	unsigned int			oap_page_off;
+	enum async_flags		oap_async_flags;
 
-	struct brw_page	 oap_brw_page;
+	struct brw_page			oap_brw_page;
 
-	struct ptlrpc_request   *oap_request;
-	struct client_obd       *oap_cli;
-	struct osc_object       *oap_obj;
+	struct ptlrpc_request		*oap_request;
+	struct client_obd		*oap_cli;
+	struct osc_object		*oap_obj;
 
-	spinlock_t		 oap_lock;
+	spinlock_t			oap_lock;
 };
 
 #define oap_page	oap_brw_page.pg
-#define oap_count       oap_brw_page.count
-#define oap_brw_flags   oap_brw_page.flag
+#define oap_count	oap_brw_page.count
+#define oap_brw_flags	oap_brw_page.flag
 
 static inline struct osc_async_page *brw_page2oap(struct brw_page *pga)
 {
-	return (struct osc_async_page *)container_of(pga, struct osc_async_page,
-						     oap_brw_page);
+	return container_of(pga, struct osc_async_page, oap_brw_page);
 }
 
 struct osc_cache_waiter {
-	struct list_head	      ocw_entry;
-	wait_queue_head_t	     ocw_waitq;
-	struct osc_async_page  *ocw_oap;
-	int		     ocw_grant;
-	int		     ocw_rc;
+	struct list_head		ocw_entry;
+	wait_queue_head_t		ocw_waitq;
+	struct osc_async_page		*ocw_oap;
+	int				ocw_grant;
+	int				ocw_rc;
 };
 
 void osc_wake_cache_waiters(struct client_obd *cli);
@@ -166,19 +168,19 @@ static inline char *cli_name(struct client_obd *cli)
 }
 
 struct osc_device {
-	struct cl_device    od_cl;
-	struct obd_export  *od_exp;
+	struct cl_device	od_cl;
+	struct obd_export	*od_exp;
 
 	/* Write stats is actually protected by client_obd's lock. */
 	struct osc_stats {
-		u64	os_lockless_writes;	  /* by bytes */
-		u64	os_lockless_reads;	  /* by bytes */
-		u64	os_lockless_truncates;    /* by times */
-	} od_stats;
+		u64	os_lockless_writes;	/* by bytes */
+		u64	os_lockless_reads;	/* by bytes */
+		u64	os_lockless_truncates;	/* by times */
+	}			od_stats;
 
 	/* configuration item(s) */
-	int		 od_contention_time;
-	int		 od_lockless_truncate;
+	int			od_contention_time;
+	int			od_lockless_truncate;
 };
 
 static inline struct osc_device *obd2osc_dev(const struct obd_device *d)
@@ -190,10 +192,10 @@ extern struct lu_kmem_descr osc_caches[];
 
 extern struct kmem_cache *osc_quota_kmem;
 struct osc_quota_info {
-	/** linkage for quota hash table */
-	struct rhash_head oqi_hash;
-	u32		  oqi_id;
-	struct rcu_head	  rcu;
+	/* linkage for quota hash table */
+	struct rhash_head	oqi_hash;
+	u32			oqi_id;
+	struct rcu_head		rcu;
 };
 
 int osc_quota_setup(struct obd_device *obd);
@@ -207,16 +209,16 @@ void osc_inc_unstable_pages(struct ptlrpc_request *req);
 void osc_dec_unstable_pages(struct ptlrpc_request *req);
 bool osc_over_unstable_soft_limit(struct client_obd *cli);
 
-/**
+/*
  * Bit flags for osc_dlm_lock_at_pageoff().
  */
 enum osc_dap_flags {
-	/**
+	/*
 	 * Just check if the desired lock exists, it won't hold reference
 	 * count on lock.
 	 */
 	OSC_DAP_FL_TEST_LOCK	= BIT(0),
-	/**
+	/*
 	 * Return the lock even if it is being canceled.
 	 */
 	OSC_DAP_FL_CANCELING	= BIT(1),
@@ -228,9 +230,9 @@ struct ldlm_lock *osc_dlmlock_at_pgoff(const struct lu_env *env,
 
 int osc_object_invalidate(const struct lu_env *env, struct osc_object *osc);
 
-/** osc shrink list to link all osc client obd */
+/* osc shrink list to link all osc client obd */
 extern struct list_head osc_shrink_list;
-/** spin lock to protect osc_shrink_list */
+/* spin lock to protect osc_shrink_list */
 extern spinlock_t osc_shrink_lock;
 unsigned long osc_cache_shrink_count(struct shrinker *sk,
 				     struct shrink_control *sc);

  reply	other threads:[~2018-12-10  0:46 UTC|newest]

Thread overview: 13+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2018-12-10  0:46 [lustre-devel] [PATCH 0/4] some modest linux-lustre cleanups NeilBrown
2018-12-10  0:46 ` NeilBrown [this message]
2018-12-27  2:15   ` [lustre-devel] [PATCH 4/4] lustre: clear up white space in osc header files James Simmons
2018-12-10  0:46 ` [lustre-devel] [PATCH 3/4] lustre: use bit-locking in echo_client NeilBrown
2018-12-10  0:57   ` Patrick Farrell
2018-12-10  1:26     ` NeilBrown
2018-12-10  4:09       ` Patrick Farrell
2018-12-27  2:14   ` James Simmons
2018-12-10  0:46 ` [lustre-devel] [PATCH 1/4] lustre: lnet_startup_lndnet: avoid use-after-free NeilBrown
2018-12-27  2:13   ` James Simmons
2018-12-10  0:46 ` [lustre-devel] [PATCH 2/4] lustre: use GFP_NOFS when lli_och_mutex is held - again NeilBrown
2018-12-27  2:11   ` James Simmons
2018-12-27  2:13   ` James Simmons

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=154440277682.29887.12055691371699460100.stgit@noble \
    --to=neilb@suse.com \
    --cc=lustre-devel@lists.lustre.org \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.