[lustre-devel] [PATCH 4/4] lustre: clear up white space in osc header files.

NeilBrown neilb at suse.com
Sun Dec 9 16:46:16 PST 2018


Also change /** to /*.  In the kernel, /** should only be
used for kernel-doc comments.

Signed-off-by: NeilBrown <neilb at suse.com>
---
 .../staging/lustre/lustre/osc/osc_cl_internal.h    |  319 ++++++++++----------
 drivers/staging/lustre/lustre/osc/osc_internal.h   |   96 +++---
 2 files changed, 209 insertions(+), 206 deletions(-)

diff --git a/drivers/staging/lustre/lustre/osc/osc_cl_internal.h b/drivers/staging/lustre/lustre/osc/osc_cl_internal.h
index 4b0180946134..077a2b183634 100644
--- a/drivers/staging/lustre/lustre/osc/osc_cl_internal.h
+++ b/drivers/staging/lustre/lustre/osc/osc_cl_internal.h
@@ -48,62 +48,63 @@
 #include <cl_object.h>
 #include "osc_internal.h"
 
-/** \defgroup osc osc
+/* \defgroup osc osc
  *  @{
  */
 
 struct osc_extent;
 
-/**
+/*
  * State maintained by osc layer for each IO context.
  */
 struct osc_io {
-	/** super class */
-	struct cl_io_slice oi_cl;
-	/** true if this io is lockless. */
+	/* super class */
+	struct cl_io_slice	oi_cl;
+	/* true if this io is lockless. */
 	unsigned int		oi_lockless:1,
-	/** true if this io is counted as active IO */
+	/* true if this io is counted as active IO */
 				oi_is_active:1;
-	/** how many LRU pages are reserved for this IO */
+	/* how many LRU pages are reserved for this IO */
 	unsigned long		oi_lru_reserved;
 
-	/** active extents, we know how many bytes is going to be written,
-	 * so having an active extent will prevent it from being fragmented
+	/* active extents, we know how many bytes is going to be
+	 * written, so having an active extent will prevent it from
+	 * being fragmented
 	 */
-	struct osc_extent *oi_active;
-	/** partially truncated extent, we need to hold this extent to prevent
-	 * page writeback from happening.
+	struct osc_extent	*oi_active;
+	/* partially truncated extent, we need to hold this extent to
+	 * prevent page writeback from happening.
 	 */
-	struct osc_extent *oi_trunc;
+	struct osc_extent	*oi_trunc;
 
-	/** write osc_lock for this IO, used by osc_extent_find(). */
-	struct osc_lock   *oi_write_osclock;
-	struct obdo	oi_oa;
+	/* write osc_lock for this IO, used by osc_extent_find(). */
+	struct osc_lock		*oi_write_osclock;
+	struct obdo		oi_oa;
 	struct osc_async_cbargs {
-		bool		  opc_rpc_sent;
-		int	       opc_rc;
+		bool			opc_rpc_sent;
+		int			opc_rc;
 		struct completion	opc_sync;
-	} oi_cbarg;
+	}			oi_cbarg;
 };
 
-/**
+/*
  * State maintained by osc layer for the duration of a system call.
  */
 struct osc_session {
-	struct osc_io       os_io;
+	struct osc_io		os_io;
 };
 
 #define OTI_PVEC_SIZE 256
 struct osc_thread_info {
-	struct ldlm_res_id      oti_resname;
+	struct ldlm_res_id	oti_resname;
 	union ldlm_policy_data	oti_policy;
-	struct cl_lock_descr    oti_descr;
-	struct cl_attr	  oti_attr;
-	struct lustre_handle    oti_handle;
-	struct cl_page_list     oti_plist;
+	struct cl_lock_descr	oti_descr;
+	struct cl_attr		oti_attr;
+	struct lustre_handle	oti_handle;
+	struct cl_page_list	oti_plist;
 	struct cl_io		oti_io;
 	void			*oti_pvec[OTI_PVEC_SIZE];
-	/**
+	/*
 	 * Fields used by cl_lock_discard_pages().
 	 */
 	pgoff_t			oti_next_index;
@@ -114,44 +115,44 @@ struct osc_thread_info {
 };
 
 struct osc_object {
-	struct cl_object   oo_cl;
-	struct lov_oinfo  *oo_oinfo;
-	/**
+	struct cl_object	oo_cl;
+	struct lov_oinfo	*oo_oinfo;
+	/*
 	 * True if locking against this stripe got -EUSERS.
 	 */
-	int		oo_contended;
-	unsigned long	 oo_contention_time;
-	/**
+	int			oo_contended;
+	unsigned long		oo_contention_time;
+	/*
 	 * used by the osc to keep track of what objects to build into rpcs.
 	 * Protected by client_obd->cli_loi_list_lock.
 	 */
-	struct list_head	   oo_ready_item;
-	struct list_head	   oo_hp_ready_item;
-	struct list_head	   oo_write_item;
-	struct list_head	   oo_read_item;
+	struct list_head	oo_ready_item;
+	struct list_head	oo_hp_ready_item;
+	struct list_head	oo_write_item;
+	struct list_head	oo_read_item;
 
-	/**
+	/*
 	 * extent is a red black tree to manage (async) dirty pages.
 	 */
-	struct rb_root       oo_root;
-	/**
+	struct rb_root		oo_root;
+	/*
 	 * Manage write(dirty) extents.
 	 */
-	struct list_head	   oo_hp_exts; /* list of hp extents */
-	struct list_head	   oo_urgent_exts; /* list of writeback extents */
-	struct list_head	   oo_full_exts;
+	struct list_head	oo_hp_exts; /* list of hp extents */
+	struct list_head	oo_urgent_exts; /* list of writeback extents */
+	struct list_head	oo_full_exts;
 
-	struct list_head	   oo_reading_exts;
+	struct list_head	oo_reading_exts;
 
-	atomic_t	 oo_nr_reads;
-	atomic_t	 oo_nr_writes;
+	atomic_t		oo_nr_reads;
+	atomic_t		oo_nr_writes;
 
-	/** Protect extent tree. Will be used to protect
+	/* Protect extent tree. Will be used to protect
 	 * oo_{read|write}_pages soon.
 	 */
-	spinlock_t	    oo_lock;
+	spinlock_t		oo_lock;
 
-	/**
+	/*
 	 * Radix tree for caching pages
 	 */
 	struct radix_tree_root	oo_tree;
@@ -162,7 +163,7 @@ struct osc_object {
 	spinlock_t		oo_ol_spin;
 	struct list_head	oo_ol_list;
 
-	/** number of active IOs of this object */
+	/* number of active IOs of this object */
 	atomic_t		oo_nr_ios;
 	wait_queue_head_t	oo_io_waitq;
 };
@@ -207,7 +208,7 @@ enum osc_lock_state {
 	OLS_CANCELLED
 };
 
-/**
+/*
  * osc-private state of cl_lock.
  *
  * Interaction with DLM.
@@ -250,37 +251,37 @@ enum osc_lock_state {
  * future.
  */
 struct osc_lock {
-	struct cl_lock_slice     ols_cl;
-	/** Internal lock to protect states, etc. */
-	spinlock_t		ols_lock;
-	/** Owner sleeps on this channel for state change */
-	struct cl_sync_io	*ols_owner;
-	/** waiting list for this lock to be cancelled */
-	struct list_head	ols_waiting_list;
-	/** wait entry of ols_waiting_list */
-	struct list_head	ols_wait_entry;
-	/** list entry for osc_object::oo_ol_list */
-	struct list_head	ols_nextlock_oscobj;
-
-	/** underlying DLM lock */
-	struct ldlm_lock	*ols_dlmlock;
-	/** DLM flags with which osc_lock::ols_lock was enqueued */
-	__u64		    ols_flags;
-	/** osc_lock::ols_lock handle */
-	struct lustre_handle     ols_handle;
-	struct ldlm_enqueue_info ols_einfo;
-	enum osc_lock_state      ols_state;
-	/** lock value block */
-	struct ost_lvb		ols_lvb;
-
-	/**
+	struct cl_lock_slice		ols_cl;
+	/* Internal lock to protect states, etc. */
+	spinlock_t			ols_lock;
+	/* Owner sleeps on this channel for state change */
+	struct cl_sync_io		*ols_owner;
+	/* waiting list for this lock to be cancelled */
+	struct list_head		ols_waiting_list;
+	/* wait entry of ols_waiting_list */
+	struct list_head		ols_wait_entry;
+	/* list entry for osc_object::oo_ol_list */
+	struct list_head		ols_nextlock_oscobj;
+
+	/* underlying DLM lock */
+	struct ldlm_lock		*ols_dlmlock;
+	/* DLM flags with which osc_lock::ols_lock was enqueued */
+	__u64				ols_flags;
+	/* osc_lock::ols_lock handle */
+	struct lustre_handle		ols_handle;
+	struct ldlm_enqueue_info	ols_einfo;
+	enum osc_lock_state		ols_state;
+	/* lock value block */
+	struct ost_lvb			ols_lvb;
+
+	/*
 	 * true, if ldlm_lock_addref() was called against
 	 * osc_lock::ols_lock. This is used for sanity checking.
 	 *
 	 * \see osc_lock::ols_has_ref
 	 */
-	unsigned		  ols_hold :1,
-	/**
+	unsigned			ols_hold :1,
+	/*
 	 * this is much like osc_lock::ols_hold, except that this bit is
 	 * cleared _after_ reference in released in osc_lock_unuse(). This
 	 * fine distinction is needed because:
@@ -295,13 +296,13 @@ struct osc_lock {
 	 *       release lock reference (so that a flag is needed that is
 	 *       cleared before ldlm_lock_decref() is called).
 	 */
-				 ols_has_ref:1,
-	/**
+					ols_has_ref:1,
+	/*
 	 * inherit the lockless attribute from top level cl_io.
 	 * If true, osc_lock_enqueue is able to tolerate the -EUSERS error.
 	 */
-				 ols_locklessable:1,
-	/**
+					ols_locklessable:1,
+	/*
 	 * if set, the osc_lock is a glimpse lock. For glimpse locks, we treat
 	 * the EVAVAIL error as tolerable, this will make upper logic happy
 	 * to wait all glimpse locks to each OSTs to be completed.
@@ -309,53 +310,53 @@ struct osc_lock {
 	 * granted.
 	 * Glimpse lock should be destroyed immediately after use.
 	 */
-				 ols_glimpse:1,
-	/**
+					ols_glimpse:1,
+	/*
 	 * For async glimpse lock.
 	 */
-				 ols_agl:1;
+					ols_agl:1;
 };
 
-/**
+/*
  * Page state private for osc layer.
  */
 struct osc_page {
-	struct cl_page_slice  ops_cl;
-	/**
+	struct cl_page_slice	ops_cl;
+	/*
 	 * Page queues used by osc to detect when RPC can be formed.
 	 */
-	struct osc_async_page ops_oap;
-	/**
+	struct osc_async_page	ops_oap;
+	/*
 	 * An offset within page from which next transfer starts. This is used
 	 * by cl_page_clip() to submit partial page transfers.
 	 */
-	int		   ops_from;
-	/**
+	int			ops_from;
+	/*
 	 * An offset within page at which next transfer ends.
 	 *
 	 * \see osc_page::ops_from.
 	 */
-	int		   ops_to;
-	/**
+	int			ops_to;
+	/*
 	 * Boolean, true iff page is under transfer. Used for sanity checking.
 	 */
-	unsigned	      ops_transfer_pinned:1,
-	/**
+	unsigned		ops_transfer_pinned:1,
+	/*
 	 * in LRU?
 	 */
-			      ops_in_lru:1,
-	/**
+				ops_in_lru:1,
+	/*
 	 * Set if the page must be transferred with OBD_BRW_SRVLOCK.
 	 */
-			      ops_srvlock:1;
-	/**
+				ops_srvlock:1;
+	/*
 	 * lru page list. See osc_lru_{del|use}() in osc_page.c for usage.
 	 */
-	struct list_head	      ops_lru;
-	/**
+	struct list_head	ops_lru;
+	/*
 	 * Submit time - the time when the page is starting RPC. For debugging.
 	 */
-	unsigned long	    ops_submit_time;
+	unsigned long		ops_submit_time;
 };
 
 extern struct kmem_cache *osc_lock_kmem;
@@ -422,7 +423,7 @@ int  osc_object_is_contended(struct osc_object *obj);
 
 int  osc_lock_is_lockless(const struct osc_lock *olck);
 
-/*****************************************************************************
+/****************************************************************************
  *
  * Accessors.
  *
@@ -558,17 +559,17 @@ static inline int osc_io_srvlock(struct osc_io *oio)
 }
 
 enum osc_extent_state {
-	OES_INV       = 0, /** extent is just initialized or destroyed */
-	OES_ACTIVE    = 1, /** process is using this extent */
-	OES_CACHE     = 2, /** extent is ready for IO */
-	OES_LOCKING   = 3, /** locking page to prepare IO */
-	OES_LOCK_DONE = 4, /** locking finished, ready to send */
-	OES_RPC       = 5, /** in RPC */
-	OES_TRUNC     = 6, /** being truncated */
+	OES_INV       = 0, /* extent is just initialized or destroyed */
+	OES_ACTIVE    = 1, /* process is using this extent */
+	OES_CACHE     = 2, /* extent is ready for IO */
+	OES_LOCKING   = 3, /* locking page to prepare IO */
+	OES_LOCK_DONE = 4, /* locking finished, ready to send */
+	OES_RPC       = 5, /* in RPC */
+	OES_TRUNC     = 6, /* being truncated */
 	OES_STATE_MAX
 };
 
-/**
+/*
  * osc_extent data to manage dirty pages.
  * osc_extent has the following attributes:
  * 1. all pages in the same must be in one RPC in write back;
@@ -584,85 +585,85 @@ enum osc_extent_state {
  * page lock -> cl_loi_list_lock -> object lock(osc_object::oo_lock)
  */
 struct osc_extent {
-	/** red-black tree node */
-	struct rb_node     oe_node;
-	/** osc_object of this extent */
-	struct osc_object *oe_obj;
-	/** refcount, removed from red-black tree if reaches zero. */
-	atomic_t       oe_refc;
-	/** busy if non-zero */
-	atomic_t       oe_users;
-	/** link list of osc_object's oo_{hp|urgent|locking}_exts. */
-	struct list_head	 oe_link;
-	/** state of this extent */
+	/* red-black tree node */
+	struct rb_node		oe_node;
+	/* osc_object of this extent */
+	struct osc_object	*oe_obj;
+	/* refcount, removed from red-black tree if reaches zero. */
+	atomic_t		oe_refc;
+	/* busy if non-zero */
+	atomic_t		oe_users;
+	/* link list of osc_object's oo_{hp|urgent|locking}_exts. */
+	struct list_head	oe_link;
+	/* state of this extent */
 	enum osc_extent_state	oe_state;
-	/** flags for this extent. */
-	unsigned int       oe_intree:1,
-	/** 0 is write, 1 is read */
-			   oe_rw:1,
-	/** sync extent, queued by osc_queue_sync_pages() */
+	/* flags for this extent. */
+	unsigned int		oe_intree:1,
+	/* 0 is write, 1 is read */
+				oe_rw:1,
+	/* sync extent, queued by osc_queue_sync_pages() */
 				oe_sync:1,
-	/** set if this extent has partial, sync pages.
+	/* set if this extent has partial, sync pages.
 	 * Extents with partial page(s) can't merge with others in RPC
 	 */
 				oe_no_merge:1,
-			   oe_srvlock:1,
-			   oe_memalloc:1,
-	/** an ACTIVE extent is going to be truncated, so when this extent
+				oe_srvlock:1,
+				oe_memalloc:1,
+	/* an ACTIVE extent is going to be truncated, so when this extent
 	 * is released, it will turn into TRUNC state instead of CACHE.
 	 */
-			   oe_trunc_pending:1,
-	/** this extent should be written asap and someone may wait for the
+				oe_trunc_pending:1,
+	/* this extent should be written asap and someone may wait for the
 	 * write to finish. This bit is usually set along with urgent if
 	 * the extent was CACHE state.
 	 * fsync_wait extent can't be merged because new extent region may
 	 * exceed fsync range.
 	 */
-			   oe_fsync_wait:1,
-	/** covering lock is being canceled */
-			   oe_hp:1,
-	/** this extent should be written back asap. set if one of pages is
+				oe_fsync_wait:1,
+	/* covering lock is being canceled */
+				oe_hp:1,
+	/* this extent should be written back asap. set if one of pages is
 	 * called by page WB daemon, or sync write or reading requests.
 	 */
-			   oe_urgent:1;
-	/** how many grants allocated for this extent.
+				oe_urgent:1;
+	/* how many grants allocated for this extent.
 	 *  Grant allocated for this extent. There is no grant allocated
 	 *  for reading extents and sync write extents.
 	 */
-	unsigned int       oe_grants;
-	/** # of dirty pages in this extent */
-	unsigned int       oe_nr_pages;
-	/** list of pending oap pages. Pages in this list are NOT sorted. */
-	struct list_head	 oe_pages;
-	/** Since an extent has to be written out in atomic, this is used to
+	unsigned int		oe_grants;
+	/* # of dirty pages in this extent */
+	unsigned int		oe_nr_pages;
+	/* list of pending oap pages. Pages in this list are NOT sorted. */
+	struct list_head	oe_pages;
+	/* Since an extent has to be written out in atomic, this is used to
 	 * remember the next page need to be locked to write this extent out.
 	 * Not used right now.
 	 */
-	struct osc_page   *oe_next_page;
-	/** start and end index of this extent, include start and end
+	struct osc_page		*oe_next_page;
+	/* start and end index of this extent, include start and end
 	 * themselves. Page offset here is the page index of osc_pages.
 	 * oe_start is used as keyword for red-black tree.
 	 */
-	pgoff_t	    oe_start;
-	pgoff_t	    oe_end;
-	/** maximum ending index of this extent, this is limited by
+	pgoff_t			oe_start;
+	pgoff_t			oe_end;
+	/* maximum ending index of this extent, this is limited by
 	 * max_pages_per_rpc, lock extent and chunk size.
 	 */
-	pgoff_t	    oe_max_end;
-	/** waitqueue - for those who want to be notified if this extent's
+	pgoff_t			oe_max_end;
+	/* waitqueue - for those who want to be notified if this extent's
 	 * state has changed.
 	 */
 	wait_queue_head_t	oe_waitq;
-	/** lock covering this extent */
+	/* lock covering this extent */
 	struct ldlm_lock	*oe_dlmlock;
-	/** terminator of this extent. Must be true if this extent is in IO. */
+	/* terminator of this extent. Must be true if this extent is in IO. */
 	struct task_struct	*oe_owner;
-	/** return value of writeback. If somebody is waiting for this extent,
+	/* return value of writeback. If somebody is waiting for this extent,
 	 * this value can be known by outside world.
 	 */
-	int		oe_rc;
-	/** max pages per rpc when this extent was created */
-	unsigned int       oe_mppr;
+	int			oe_rc;
+	/* max pages per rpc when this extent was created */
+	unsigned int		oe_mppr;
 };
 
 int osc_extent_finish(const struct lu_env *env, struct osc_extent *ext,
@@ -677,6 +678,6 @@ typedef int (*osc_page_gang_cbt)(const struct lu_env *, struct cl_io *,
 int osc_page_gang_lookup(const struct lu_env *env, struct cl_io *io,
 			 struct osc_object *osc, pgoff_t start, pgoff_t end,
 			 osc_page_gang_cbt cb, void *cbdata);
-/** @} osc */
+/* @} osc */
 
 #endif /* OSC_CL_INTERNAL_H */
diff --git a/drivers/staging/lustre/lustre/osc/osc_internal.h b/drivers/staging/lustre/lustre/osc/osc_internal.h
index 55054d705a01..0de8a3ee826d 100644
--- a/drivers/staging/lustre/lustre/osc/osc_internal.h
+++ b/drivers/staging/lustre/lustre/osc/osc_internal.h
@@ -43,54 +43,56 @@ extern struct ptlrpc_request_pool *osc_rq_pool;
 struct lu_env;
 
 enum async_flags {
-	ASYNC_READY = 0x1, /* ap_make_ready will not be called before this
-			    * page is added to an rpc
-			    */
-	ASYNC_URGENT = 0x2, /* page must be put into an RPC before return */
-	ASYNC_COUNT_STABLE = 0x4, /* ap_refresh_count will not be called
-				   * to give the caller a chance to update
-				   * or cancel the size of the io
-				   */
+	ASYNC_READY		= 0x1, /* ap_make_ready will not be
+					* called before this page is
+					* added to an rpc
+					*/
+	ASYNC_URGENT		= 0x2, /* page must be put into an RPC
+					* before return */
+	ASYNC_COUNT_STABLE	= 0x4, /* ap_refresh_count will not be
+					* called to give the caller a
+					* chance to update or cancel
+					* the size of the io
+					*/
 	ASYNC_HP = 0x10,
 };
 
 struct osc_async_page {
-	int		     oap_magic;
-	unsigned short	  oap_cmd;
-	unsigned short	  oap_interrupted:1;
+	int				oap_magic;
+	unsigned short			oap_cmd;
+	unsigned short			oap_interrupted:1;
 
-	struct list_head	      oap_pending_item;
-	struct list_head	      oap_rpc_item;
+	struct list_head		oap_pending_item;
+	struct list_head		oap_rpc_item;
 
-	u64		 oap_obj_off;
-	unsigned int		oap_page_off;
-	enum async_flags	oap_async_flags;
+	u64				oap_obj_off;
+	unsigned int			oap_page_off;
+	enum async_flags		oap_async_flags;
 
-	struct brw_page	 oap_brw_page;
+	struct brw_page			oap_brw_page;
 
-	struct ptlrpc_request   *oap_request;
-	struct client_obd       *oap_cli;
-	struct osc_object       *oap_obj;
+	struct ptlrpc_request		*oap_request;
+	struct client_obd		*oap_cli;
+	struct osc_object		*oap_obj;
 
-	spinlock_t		 oap_lock;
+	spinlock_t			oap_lock;
 };
 
 #define oap_page	oap_brw_page.pg
-#define oap_count       oap_brw_page.count
-#define oap_brw_flags   oap_brw_page.flag
+#define oap_count	oap_brw_page.count
+#define oap_brw_flags	oap_brw_page.flag
 
 static inline struct osc_async_page *brw_page2oap(struct brw_page *pga)
 {
-	return (struct osc_async_page *)container_of(pga, struct osc_async_page,
-						     oap_brw_page);
+	return container_of(pga, struct osc_async_page, oap_brw_page);
 }
 
 struct osc_cache_waiter {
-	struct list_head	      ocw_entry;
-	wait_queue_head_t	     ocw_waitq;
-	struct osc_async_page  *ocw_oap;
-	int		     ocw_grant;
-	int		     ocw_rc;
+	struct list_head		ocw_entry;
+	wait_queue_head_t		ocw_waitq;
+	struct osc_async_page		*ocw_oap;
+	int				ocw_grant;
+	int				ocw_rc;
 };
 
 void osc_wake_cache_waiters(struct client_obd *cli);
@@ -166,19 +168,19 @@ static inline char *cli_name(struct client_obd *cli)
 }
 
 struct osc_device {
-	struct cl_device    od_cl;
-	struct obd_export  *od_exp;
+	struct cl_device	od_cl;
+	struct obd_export	*od_exp;
 
 	/* Write stats is actually protected by client_obd's lock. */
 	struct osc_stats {
-		u64	os_lockless_writes;	  /* by bytes */
-		u64	os_lockless_reads;	  /* by bytes */
-		u64	os_lockless_truncates;    /* by times */
-	} od_stats;
+		u64	os_lockless_writes;	/* by bytes */
+		u64	os_lockless_reads;	/* by bytes */
+		u64	os_lockless_truncates;	/* by times */
+	}			od_stats;
 
 	/* configuration item(s) */
-	int		 od_contention_time;
-	int		 od_lockless_truncate;
+	int			od_contention_time;
+	int			od_lockless_truncate;
 };
 
 static inline struct osc_device *obd2osc_dev(const struct obd_device *d)
@@ -190,10 +192,10 @@ extern struct lu_kmem_descr osc_caches[];
 
 extern struct kmem_cache *osc_quota_kmem;
 struct osc_quota_info {
-	/** linkage for quota hash table */
-	struct rhash_head oqi_hash;
-	u32		  oqi_id;
-	struct rcu_head	  rcu;
+	/* linkage for quota hash table */
+	struct rhash_head	oqi_hash;
+	u32			oqi_id;
+	struct rcu_head		rcu;
 };
 
 int osc_quota_setup(struct obd_device *obd);
@@ -207,16 +209,16 @@ void osc_inc_unstable_pages(struct ptlrpc_request *req);
 void osc_dec_unstable_pages(struct ptlrpc_request *req);
 bool osc_over_unstable_soft_limit(struct client_obd *cli);
 
-/**
+/*
  * Bit flags for osc_dlm_lock_at_pageoff().
  */
 enum osc_dap_flags {
-	/**
+	/*
 	 * Just check if the desired lock exists, it won't hold reference
 	 * count on lock.
 	 */
 	OSC_DAP_FL_TEST_LOCK	= BIT(0),
-	/**
+	/*
 	 * Return the lock even if it is being canceled.
 	 */
 	OSC_DAP_FL_CANCELING	= BIT(1),
@@ -228,9 +230,9 @@ struct ldlm_lock *osc_dlmlock_at_pgoff(const struct lu_env *env,
 
 int osc_object_invalidate(const struct lu_env *env, struct osc_object *osc);
 
-/** osc shrink list to link all osc client obd */
+/* osc shrink list to link all osc client obd */
 extern struct list_head osc_shrink_list;
-/** spin lock to protect osc_shrink_list */
+/* spin lock to protect osc_shrink_list */
 extern spinlock_t osc_shrink_lock;
 unsigned long osc_cache_shrink_count(struct shrinker *sk,
 				     struct shrink_control *sc);




More information about the lustre-devel mailing list