@@ -50,13 +50,13 @@ extern int rcutorture_runnable; /* for sysctl */
50
50
#endif /* #ifdef CONFIG_RCU_TORTURE_TEST */
51
51
52
52
#if defined(CONFIG_TREE_RCU ) || defined(CONFIG_TREE_PREEMPT_RCU )
53
- extern void rcutorture_record_test_transition (void );
54
- extern void rcutorture_record_progress (unsigned long vernum );
55
- extern void do_trace_rcu_torture_read (const char * rcutorturename ,
56
- struct rcu_head * rhp ,
57
- unsigned long secs ,
58
- unsigned long c_old ,
59
- unsigned long c );
53
+ void rcutorture_record_test_transition (void );
54
+ void rcutorture_record_progress (unsigned long vernum );
55
+ void do_trace_rcu_torture_read (const char * rcutorturename ,
56
+ struct rcu_head * rhp ,
57
+ unsigned long secs ,
58
+ unsigned long c_old ,
59
+ unsigned long c );
60
60
#else
61
61
static inline void rcutorture_record_test_transition (void )
62
62
{
@@ -65,11 +65,11 @@ static inline void rcutorture_record_progress(unsigned long vernum)
65
65
{
66
66
}
67
67
#ifdef CONFIG_RCU_TRACE
68
- extern void do_trace_rcu_torture_read (const char * rcutorturename ,
69
- struct rcu_head * rhp ,
70
- unsigned long secs ,
71
- unsigned long c_old ,
72
- unsigned long c );
68
+ void do_trace_rcu_torture_read (const char * rcutorturename ,
69
+ struct rcu_head * rhp ,
70
+ unsigned long secs ,
71
+ unsigned long c_old ,
72
+ unsigned long c );
73
73
#else
74
74
#define do_trace_rcu_torture_read (rcutorturename , rhp , secs , c_old , c ) \
75
75
do { } while (0)
@@ -118,8 +118,8 @@ extern void do_trace_rcu_torture_read(const char *rcutorturename,
118
118
* if CPU A and CPU B are the same CPU (but again only if the system has
119
119
* more than one CPU).
120
120
*/
121
- extern void call_rcu (struct rcu_head * head ,
122
- void (* func )(struct rcu_head * head ));
121
+ void call_rcu (struct rcu_head * head ,
122
+ void (* func )(struct rcu_head * head ));
123
123
124
124
#else /* #ifdef CONFIG_PREEMPT_RCU */
125
125
@@ -149,8 +149,8 @@ extern void call_rcu(struct rcu_head *head,
149
149
* See the description of call_rcu() for more detailed information on
150
150
* memory ordering guarantees.
151
151
*/
152
- extern void call_rcu_bh (struct rcu_head * head ,
153
- void (* func )(struct rcu_head * head ));
152
+ void call_rcu_bh (struct rcu_head * head ,
153
+ void (* func )(struct rcu_head * head ));
154
154
155
155
/**
156
156
* call_rcu_sched() - Queue an RCU for invocation after sched grace period.
@@ -171,16 +171,16 @@ extern void call_rcu_bh(struct rcu_head *head,
171
171
* See the description of call_rcu() for more detailed information on
172
172
* memory ordering guarantees.
173
173
*/
174
- extern void call_rcu_sched (struct rcu_head * head ,
175
- void (* func )(struct rcu_head * rcu ));
174
+ void call_rcu_sched (struct rcu_head * head ,
175
+ void (* func )(struct rcu_head * rcu ));
176
176
177
- extern void synchronize_sched (void );
177
+ void synchronize_sched (void );
178
178
179
179
#ifdef CONFIG_PREEMPT_RCU
180
180
181
- extern void __rcu_read_lock (void );
182
- extern void __rcu_read_unlock (void );
183
- extern void rcu_read_unlock_special (struct task_struct * t );
181
+ void __rcu_read_lock (void );
182
+ void __rcu_read_unlock (void );
183
+ void rcu_read_unlock_special (struct task_struct * t );
184
184
void synchronize_rcu (void );
185
185
186
186
/*
@@ -216,19 +216,19 @@ static inline int rcu_preempt_depth(void)
216
216
#endif /* #else #ifdef CONFIG_PREEMPT_RCU */
217
217
218
218
/* Internal to kernel */
219
- extern void rcu_init (void );
220
- extern void rcu_sched_qs (int cpu );
221
- extern void rcu_bh_qs (int cpu );
222
- extern void rcu_check_callbacks (int cpu , int user );
219
+ void rcu_init (void );
220
+ void rcu_sched_qs (int cpu );
221
+ void rcu_bh_qs (int cpu );
222
+ void rcu_check_callbacks (int cpu , int user );
223
223
struct notifier_block ;
224
- extern void rcu_idle_enter (void );
225
- extern void rcu_idle_exit (void );
226
- extern void rcu_irq_enter (void );
227
- extern void rcu_irq_exit (void );
224
+ void rcu_idle_enter (void );
225
+ void rcu_idle_exit (void );
226
+ void rcu_irq_enter (void );
227
+ void rcu_irq_exit (void );
228
228
229
229
#ifdef CONFIG_RCU_USER_QS
230
- extern void rcu_user_enter (void );
231
- extern void rcu_user_exit (void );
230
+ void rcu_user_enter (void );
231
+ void rcu_user_exit (void );
232
232
#else
233
233
static inline void rcu_user_enter (void ) { }
234
234
static inline void rcu_user_exit (void ) { }
@@ -262,7 +262,7 @@ static inline void rcu_user_hooks_switch(struct task_struct *prev,
262
262
} while (0)
263
263
264
264
#if defined(CONFIG_DEBUG_LOCK_ALLOC ) || defined(CONFIG_RCU_TRACE ) || defined(CONFIG_SMP )
265
- extern bool __rcu_is_watching (void );
265
+ bool __rcu_is_watching (void );
266
266
#endif /* #if defined(CONFIG_DEBUG_LOCK_ALLOC) || defined(CONFIG_RCU_TRACE) || defined(CONFIG_SMP) */
267
267
268
268
/*
@@ -289,8 +289,8 @@ void wait_rcu_gp(call_rcu_func_t crf);
289
289
* initialization.
290
290
*/
291
291
#ifdef CONFIG_DEBUG_OBJECTS_RCU_HEAD
292
- extern void init_rcu_head_on_stack (struct rcu_head * head );
293
- extern void destroy_rcu_head_on_stack (struct rcu_head * head );
292
+ void init_rcu_head_on_stack (struct rcu_head * head );
293
+ void destroy_rcu_head_on_stack (struct rcu_head * head );
294
294
#else /* !CONFIG_DEBUG_OBJECTS_RCU_HEAD */
295
295
static inline void init_rcu_head_on_stack (struct rcu_head * head )
296
296
{
@@ -363,7 +363,7 @@ static inline int rcu_read_lock_held(void)
363
363
* rcu_read_lock_bh_held() is defined out of line to avoid #include-file
364
364
* hell.
365
365
*/
366
- extern int rcu_read_lock_bh_held (void );
366
+ int rcu_read_lock_bh_held (void );
367
367
368
368
/**
369
369
* rcu_read_lock_sched_held() - might we be in RCU-sched read-side critical section?
@@ -449,7 +449,7 @@ static inline int rcu_read_lock_sched_held(void)
449
449
450
450
#ifdef CONFIG_PROVE_RCU
451
451
452
- extern int rcu_my_thread_group_empty (void );
452
+ int rcu_my_thread_group_empty (void );
453
453
454
454
/**
455
455
* rcu_lockdep_assert - emit lockdep splat if specified condition not met
@@ -1006,16 +1006,16 @@ static inline notrace void rcu_read_unlock_sched_notrace(void)
1006
1006
__kfree_rcu(&((ptr)->rcu_head), offsetof(typeof(*(ptr)), rcu_head))
1007
1007
1008
1008
#ifdef CONFIG_RCU_NOCB_CPU
1009
- extern bool rcu_is_nocb_cpu (int cpu );
1009
+ bool rcu_is_nocb_cpu (int cpu );
1010
1010
#else
1011
1011
static inline bool rcu_is_nocb_cpu (int cpu ) { return false; }
1012
1012
#endif /* #else #ifdef CONFIG_RCU_NOCB_CPU */
1013
1013
1014
1014
1015
1015
/* Only for use by adaptive-ticks code. */
1016
1016
#ifdef CONFIG_NO_HZ_FULL_SYSIDLE
1017
- extern bool rcu_sys_is_idle (void );
1018
- extern void rcu_sysidle_force_exit (void );
1017
+ bool rcu_sys_is_idle (void );
1018
+ void rcu_sysidle_force_exit (void );
1019
1019
#else /* #ifdef CONFIG_NO_HZ_FULL_SYSIDLE */
1020
1020
1021
1021
static inline bool rcu_sys_is_idle (void )
0 commit comments