Subversion Repositories shark

Rev

Go to most recent revision | Details | Last modification | View Log | RSS feed

Rev Author Line No. Line
422 giacomo 1
#ifndef _LINUX_SCHED_H
2
#define _LINUX_SCHED_H
3
 
4
#include <asm/param.h>  /* for HZ */
5
 
6
#include <linux/config.h>
7
#include <linux/capability.h>
8
#include <linux/threads.h>
9
#include <linux/kernel.h>
10
#include <linux/types.h>
11
#include <linux/timex.h>
12
#include <linux/jiffies.h>
13
#include <linux/rbtree.h>
14
#include <linux/thread_info.h>
15
#include <linux/cpumask.h>
16
 
17
#include <asm/system.h>
18
#include <asm/semaphore.h>
19
#include <asm/page.h>
20
#include <asm/ptrace.h>
21
#include <asm/mmu.h>
22
 
23
#include <linux/smp.h>
24
#include <linux/sem.h>
25
#include <linux/signal.h>
26
#include <linux/securebits.h>
27
#include <linux/fs_struct.h>
28
#include <linux/compiler.h>
29
#include <linux/completion.h>
30
#include <linux/pid.h>
31
#include <linux/percpu.h>
32
 
33
struct exec_domain;
34
 
35
/*
36
 * cloning flags:
37
 */
38
#define CSIGNAL         0x000000ff      /* signal mask to be sent at exit */
39
#define CLONE_VM        0x00000100      /* set if VM shared between processes */
40
#define CLONE_FS        0x00000200      /* set if fs info shared between processes */
41
#define CLONE_FILES     0x00000400      /* set if open files shared between processes */
42
#define CLONE_SIGHAND   0x00000800      /* set if signal handlers and blocked signals shared */
43
#define CLONE_IDLETASK  0x00001000      /* set if new pid should be 0 (kernel only)*/
44
#define CLONE_PTRACE    0x00002000      /* set if we want to let tracing continue on the child too */
45
#define CLONE_VFORK     0x00004000      /* set if the parent wants the child to wake it up on mm_release */
46
#define CLONE_PARENT    0x00008000      /* set if we want to have the same parent as the cloner */
47
#define CLONE_THREAD    0x00010000      /* Same thread group? */
48
#define CLONE_NEWNS     0x00020000      /* New namespace group? */
49
#define CLONE_SYSVSEM   0x00040000      /* share system V SEM_UNDO semantics */
50
#define CLONE_SETTLS    0x00080000      /* create a new TLS for the child */
51
#define CLONE_PARENT_SETTID     0x00100000      /* set the TID in the parent */
52
#define CLONE_CHILD_CLEARTID    0x00200000      /* clear the TID in the child */
53
#define CLONE_DETACHED          0x00400000      /* Not used - CLONE_THREAD implies detached uniquely */
54
#define CLONE_UNTRACED          0x00800000      /* set if the tracing process can't force CLONE_PTRACE on this clone */
55
#define CLONE_CHILD_SETTID      0x01000000      /* set the TID in the child */
56
#define CLONE_STOPPED           0x02000000      /* Start in stopped state */
57
 
58
/*
59
 * List of flags we want to share for kernel threads,
60
 * if only because they are not used by them anyway.
61
 */
62
#define CLONE_KERNEL    (CLONE_FS | CLONE_FILES | CLONE_SIGHAND)
63
 
64
/*
65
 * These are the constant used to fake the fixed-point load-average
66
 * counting. Some notes:
67
 *  - 11 bit fractions expand to 22 bits by the multiplies: this gives
68
 *    a load-average precision of 10 bits integer + 11 bits fractional
69
 *  - if you want to count load-averages more often, you need more
70
 *    precision, or rounding will get you. With 2-second counting freq,
71
 *    the EXP_n values would be 1981, 2034 and 2043 if still using only
72
 *    11 bit fractions.
73
 */
74
extern unsigned long avenrun[];         /* Load averages */
75
 
76
#define FSHIFT          11              /* nr of bits of precision */
77
#define FIXED_1         (1<<FSHIFT)     /* 1.0 as fixed-point */
78
#define LOAD_FREQ       (5*HZ)          /* 5 sec intervals */
79
#define EXP_1           1884            /* 1/exp(5sec/1min) as fixed-point */
80
#define EXP_5           2014            /* 1/exp(5sec/5min) */
81
#define EXP_15          2037            /* 1/exp(5sec/15min) */
82
 
83
#define CALC_LOAD(load,exp,n) \
84
        load *= exp; \
85
        load += n*(FIXED_1-exp); \
86
        load >>= FSHIFT;
87
 
88
#define CT_TO_SECS(x)   ((x) / HZ)
89
#define CT_TO_USECS(x)  (((x) % HZ) * 1000000/HZ)
90
 
91
extern int nr_threads;
92
extern int last_pid;
93
DECLARE_PER_CPU(unsigned long, process_counts);
94
extern int nr_processes(void);
95
extern unsigned long nr_running(void);
96
extern unsigned long nr_uninterruptible(void);
97
extern unsigned long nr_iowait(void);
98
 
99
#include <linux/time.h>
100
#include <linux/param.h>
101
#include <linux/resource.h>
102
#include <linux/timer.h>
103
 
104
#include <asm/processor.h>
105
 
106
#define TASK_RUNNING            0
107
#define TASK_INTERRUPTIBLE      1
108
#define TASK_UNINTERRUPTIBLE    2
109
#define TASK_STOPPED            4
110
#define TASK_ZOMBIE             8
111
#define TASK_DEAD               16
112
 
113
#define __set_task_state(tsk, state_value)              \
114
        do { (tsk)->state = (state_value); } while (0)
115
#define set_task_state(tsk, state_value)                \
116
        set_mb((tsk)->state, (state_value))
117
 
118
#define __set_current_state(state_value)                        \
119
        do { current->state = (state_value); } while (0)
120
#define set_current_state(state_value)          \
121
        set_mb(current->state, (state_value))
122
 
123
/*
124
 * Scheduling policies
125
 */
126
#define SCHED_NORMAL            0
127
#define SCHED_FIFO              1
128
#define SCHED_RR                2
129
 
130
struct sched_param {
131
        int sched_priority;
132
};
133
 
134
#ifdef __KERNEL__
135
 
136
#include <linux/spinlock.h>
137
 
138
/*
139
 * This serializes "schedule()" and also protects
140
 * the run-queue from deletions/modifications (but
141
 * _adding_ to the beginning of the run-queue has
142
 * a separate lock).
143
 */
144
extern rwlock_t tasklist_lock;
145
extern spinlock_t mmlist_lock;
146
 
147
typedef struct task_struct task_t;
148
 
149
extern void sched_init(void);
150
extern void init_idle(task_t *idle, int cpu);
151
 
152
extern void show_state(void);
153
extern void show_regs(struct pt_regs *);
154
 
155
/*
156
 * TASK is a pointer to the task whose backtrace we want to see (or NULL for current
157
 * task), SP is the stack pointer of the first frame that should be shown in the back
158
 * trace (or NULL if the entire call-chain of the task should be shown).
159
 */
160
extern void show_stack(struct task_struct *task, unsigned long *sp);
161
 
162
void io_schedule(void);
163
long io_schedule_timeout(long timeout);
164
 
165
extern void cpu_init (void);
166
extern void trap_init(void);
167
extern void update_process_times(int user);
168
extern void update_one_process(struct task_struct *p, unsigned long user,
169
                               unsigned long system, int cpu);
170
extern void scheduler_tick(int user_tick, int system);
171
extern unsigned long cache_decay_ticks;
172
 
173
 
174
#define MAX_SCHEDULE_TIMEOUT    LONG_MAX
175
extern signed long FASTCALL(schedule_timeout(signed long timeout));
176
asmlinkage void schedule(void);
177
 
178
struct namespace;
179
 
180
/* Maximum number of active map areas.. This is a random (large) number */
181
#define MAX_MAP_COUNT   (65536)
182
 
183
#include <linux/aio.h>
184
 
185
struct mm_struct {
186
        struct vm_area_struct * mmap;           /* list of VMAs */
187
        struct rb_root mm_rb;
188
        struct vm_area_struct * mmap_cache;     /* last find_vma result */
189
        unsigned long free_area_cache;          /* first hole */
190
        pgd_t * pgd;
191
        atomic_t mm_users;                      /* How many users with user space? */
192
        atomic_t mm_count;                      /* How many references to "struct mm_struct" (users count as 1) */
193
        int map_count;                          /* number of VMAs */
194
        struct rw_semaphore mmap_sem;
195
        spinlock_t page_table_lock;             /* Protects task page tables and mm->rss */
196
 
197
        struct list_head mmlist;                /* List of all active mm's.  These are globally strung
198
                                                 * together off init_mm.mmlist, and are protected
199
                                                 * by mmlist_lock
200
                                                 */
201
 
202
        unsigned long start_code, end_code, start_data, end_data;
203
        unsigned long start_brk, brk, start_stack;
204
        unsigned long arg_start, arg_end, env_start, env_end;
205
        unsigned long rss, total_vm, locked_vm;
206
        unsigned long def_flags;
207
        cpumask_t cpu_vm_mask;
208
        unsigned long swap_address;
209
 
210
        unsigned long saved_auxv[40]; /* for /proc/PID/auxv */
211
 
212
        unsigned dumpable:1;
213
#ifdef CONFIG_HUGETLB_PAGE
214
        int used_hugetlb;
215
#endif
216
        /* Architecture-specific MM context */
217
        mm_context_t context;
218
 
219
        /* coredumping support */
220
        int core_waiters;
221
        struct completion *core_startup_done, core_done;
222
 
223
        /* aio bits */
224
        rwlock_t                ioctx_list_lock;
225
        struct kioctx           *ioctx_list;
226
 
227
        struct kioctx           default_kioctx;
228
};
229
 
230
extern int mmlist_nr;
231
 
232
struct sighand_struct {
233
        atomic_t                count;
234
        struct k_sigaction      action[_NSIG];
235
        spinlock_t              siglock;
236
};
237
 
238
/*
239
 * NOTE! "signal_struct" does not have it's own
240
 * locking, because a shared signal_struct always
241
 * implies a shared sighand_struct, so locking
242
 * sighand_struct is always a proper superset of
243
 * the locking of signal_struct.
244
 */
245
struct signal_struct {
246
        atomic_t                count;
247
 
248
        /* current thread group signal load-balancing target: */
249
        task_t                  *curr_target;
250
 
251
        /* shared signal handling: */
252
        struct sigpending       shared_pending;
253
 
254
        /* thread group exit support */
255
        int                     group_exit;
256
        int                     group_exit_code;
257
        /* overloaded:
258
         * - notify group_exit_task when ->count is equal to notify_count
259
         * - everyone except group_exit_task is stopped during signal delivery
260
         *   of fatal signals, group_exit_task processes the signal.
261
         */
262
        struct task_struct      *group_exit_task;
263
        int                     notify_count;
264
 
265
        /* thread group stop support, overloads group_exit_code too */
266
        int                     group_stop_count;
267
};
268
 
269
/*
270
 * Priority of a process goes from 0..MAX_PRIO-1, valid RT
271
 * priority is 0..MAX_RT_PRIO-1, and SCHED_NORMAL tasks are
272
 * in the range MAX_RT_PRIO..MAX_PRIO-1. Priority values
273
 * are inverted: lower p->prio value means higher priority.
274
 *
275
 * The MAX_RT_USER_PRIO value allows the actual maximum
276
 * RT priority to be separate from the value exported to
277
 * user-space.  This allows kernel threads to set their
278
 * priority to a value higher than any user task. Note:
279
 * MAX_RT_PRIO must not be smaller than MAX_USER_RT_PRIO.
280
 */
281
 
282
#define MAX_USER_RT_PRIO        100
283
#define MAX_RT_PRIO             MAX_USER_RT_PRIO
284
 
285
#define MAX_PRIO                (MAX_RT_PRIO + 40)
286
 
287
#define rt_task(p)              ((p)->prio < MAX_RT_PRIO)
288
 
289
/*
290
 * Some day this will be a full-fledged user tracking system..
291
 */
292
struct user_struct {
293
        atomic_t __count;       /* reference count */
294
        atomic_t processes;     /* How many processes does this user have? */
295
        atomic_t files;         /* How many open files does this user have? */
296
 
297
        /* Hash table maintenance information */
298
        struct list_head uidhash_list;
299
        uid_t uid;
300
};
301
 
302
extern struct user_struct *find_user(uid_t);
303
 
304
extern struct user_struct root_user;
305
#define INIT_USER (&root_user)
306
 
307
typedef struct prio_array prio_array_t;
308
struct backing_dev_info;
309
struct reclaim_state;
310
 
311
/* POSIX.1b interval timer structure. */
312
struct k_itimer {
313
        struct list_head list;           /* free/ allocate list */
314
        spinlock_t it_lock;
315
        clockid_t it_clock;             /* which timer type */
316
        timer_t it_id;                  /* timer id */
317
        int it_overrun;                 /* overrun on pending signal  */
318
        int it_overrun_last;             /* overrun on last delivered signal */
319
        int it_requeue_pending;          /* waiting to requeue this timer */
320
        int it_sigev_notify;             /* notify word of sigevent struct */
321
        int it_sigev_signo;              /* signo word of sigevent struct */
322
        sigval_t it_sigev_value;         /* value word of sigevent struct */
323
        unsigned long it_incr;          /* interval specified in jiffies */
324
        struct task_struct *it_process; /* process to send signal to */
325
        struct timer_list it_timer;
326
        struct sigqueue *sigq;          /* signal queue entry. */
327
};
328
 
329
 
330
struct io_context;                      /* See blkdev.h */
331
void exit_io_context(void);
332
 
333
struct task_struct {
334
        volatile long state;    /* -1 unrunnable, 0 runnable, >0 stopped */
335
        struct thread_info *thread_info;
336
        atomic_t usage;
337
        unsigned long flags;    /* per process flags, defined below */
338
        unsigned long ptrace;
339
 
340
        int lock_depth;         /* Lock depth */
341
 
342
        int prio, static_prio;
343
        struct list_head run_list;
344
        prio_array_t *array;
345
 
346
        unsigned long sleep_avg;
347
        long interactive_credit;
348
        unsigned long long timestamp;
349
        int activated;
350
 
351
        unsigned long policy;
352
        cpumask_t cpus_allowed;
353
        unsigned int time_slice, first_time_slice;
354
 
355
        struct list_head tasks;
356
        struct list_head ptrace_children;
357
        struct list_head ptrace_list;
358
 
359
        struct mm_struct *mm, *active_mm;
360
 
361
/* task state */
362
        struct linux_binfmt *binfmt;
363
        int exit_code, exit_signal;
364
        int pdeath_signal;  /*  The signal sent when the parent dies  */
365
        /* ??? */
366
        unsigned long personality;
367
        int did_exec:1;
368
        pid_t pid;
369
        pid_t __pgrp;           /* Accessed via process_group() */
370
        pid_t tty_old_pgrp;
371
        pid_t session;
372
        pid_t tgid;
373
        /* boolean value for session group leader */
374
        int leader;
375
        /*
376
         * pointers to (original) parent process, youngest child, younger sibling,
377
         * older sibling, respectively.  (p->father can be replaced with
378
         * p->parent->pid)
379
         */
380
        struct task_struct *real_parent; /* real parent process (when being debugged) */
381
        struct task_struct *parent;     /* parent process */
382
        struct list_head children;      /* list of my children */
383
        struct list_head sibling;       /* linkage in my parent's children list */
384
        struct task_struct *group_leader;       /* threadgroup leader */
385
 
386
        /* PID/PID hash table linkage. */
387
        struct pid_link pids[PIDTYPE_MAX];
388
 
389
        wait_queue_head_t wait_chldexit;        /* for wait4() */
390
        struct completion *vfork_done;          /* for vfork() */
391
        int __user *set_child_tid;              /* CLONE_CHILD_SETTID */
392
        int __user *clear_child_tid;            /* CLONE_CHILD_CLEARTID */
393
 
394
        unsigned long rt_priority;
395
        unsigned long it_real_value, it_prof_value, it_virt_value;
396
        unsigned long it_real_incr, it_prof_incr, it_virt_incr;
397
        struct timer_list real_timer;
398
        struct list_head posix_timers; /* POSIX.1b Interval Timers */
399
        unsigned long utime, stime, cutime, cstime;
400
        unsigned long nvcsw, nivcsw, cnvcsw, cnivcsw; /* context switch counts */
401
        u64 start_time;
402
/* mm fault and swap info: this can arguably be seen as either mm-specific or thread-specific */
403
        unsigned long min_flt, maj_flt, nswap, cmin_flt, cmaj_flt, cnswap;
404
/* process credentials */
405
        uid_t uid,euid,suid,fsuid;
406
        gid_t gid,egid,sgid,fsgid;
407
        int ngroups;
408
        gid_t   groups[NGROUPS];
409
        kernel_cap_t   cap_effective, cap_inheritable, cap_permitted;
410
        int keep_capabilities:1;
411
        struct user_struct *user;
412
/* limits */
413
        struct rlimit rlim[RLIM_NLIMITS];
414
        unsigned short used_math;
415
        char comm[16];
416
/* file system info */
417
        int link_count, total_link_count;
418
        struct tty_struct *tty; /* NULL if no tty */
419
/* ipc stuff */
420
        struct sysv_sem sysvsem;
421
/* CPU-specific state of this task */
422
        struct thread_struct thread;
423
/* filesystem information */
424
        struct fs_struct *fs;
425
/* open file information */
426
        struct files_struct *files;
427
/* namespace */
428
        struct namespace *namespace;
429
/* signal handlers */
430
        struct signal_struct *signal;
431
        struct sighand_struct *sighand;
432
 
433
        sigset_t blocked, real_blocked;
434
        struct sigpending pending;
435
 
436
        unsigned long sas_ss_sp;
437
        size_t sas_ss_size;
438
        int (*notifier)(void *priv);
439
        void *notifier_data;
440
        sigset_t *notifier_mask;
441
 
442
        void *security;
443
 
444
/* Thread group tracking */
445
        u32 parent_exec_id;
446
        u32 self_exec_id;
447
/* Protection of (de-)allocation: mm, files, fs, tty */
448
        spinlock_t alloc_lock;
449
/* Protection of proc_dentry: nesting proc_lock, dcache_lock, write_lock_irq(&tasklist_lock); */
450
        spinlock_t proc_lock;
451
/* context-switch lock */
452
        spinlock_t switch_lock;
453
 
454
/* journalling filesystem info */
455
        void *journal_info;
456
 
457
/* VM state */
458
        struct reclaim_state *reclaim_state;
459
 
460
        struct dentry *proc_dentry;
461
        struct backing_dev_info *backing_dev_info;
462
 
463
        struct io_context *io_context;
464
 
465
        unsigned long ptrace_message;
466
        siginfo_t *last_siginfo; /* For ptrace use.  */
467
};
468
 
469
static inline pid_t process_group(struct task_struct *tsk)
470
{
471
        return tsk->group_leader->__pgrp;
472
}
473
 
474
extern void __put_task_struct(struct task_struct *tsk);
475
#define get_task_struct(tsk) do { atomic_inc(&(tsk)->usage); } while(0)
476
#define put_task_struct(tsk) \
477
do { if (atomic_dec_and_test(&(tsk)->usage)) __put_task_struct(tsk); } while(0)
478
 
479
/*
480
 * Per process flags
481
 */
482
#define PF_ALIGNWARN    0x00000001      /* Print alignment warning msgs */
483
                                        /* Not implemented yet, only for 486*/
484
#define PF_STARTING     0x00000002      /* being created */
485
#define PF_EXITING      0x00000004      /* getting shut down */
486
#define PF_DEAD         0x00000008      /* Dead */
487
#define PF_FORKNOEXEC   0x00000040      /* forked but didn't exec */
488
#define PF_SUPERPRIV    0x00000100      /* used super-user privileges */
489
#define PF_DUMPCORE     0x00000200      /* dumped core */
490
#define PF_SIGNALED     0x00000400      /* killed by a signal */
491
#define PF_MEMALLOC     0x00000800      /* Allocating memory */
492
#define PF_MEMDIE       0x00001000      /* Killed for out-of-memory */
493
#define PF_FLUSHER      0x00002000      /* responsible for disk writeback */
494
 
495
#define PF_FREEZE       0x00004000      /* this task should be frozen for suspend */
496
#define PF_IOTHREAD     0x00008000      /* this thread is needed for doing I/O to swap */
497
#define PF_FROZEN       0x00010000      /* frozen for system suspend */
498
#define PF_FSTRANS      0x00020000      /* inside a filesystem transaction */
499
#define PF_KSWAPD       0x00040000      /* I am kswapd */
500
#define PF_SWAPOFF      0x00080000      /* I am in swapoff */
501
#define PF_LESS_THROTTLE 0x00100000     /* Throttle me less: I clean memory */
502
#define PF_SYNCWRITE    0x00200000      /* I am doing a sync write */
503
 
504
#ifdef CONFIG_SMP
505
extern int set_cpus_allowed(task_t *p, cpumask_t new_mask);
506
#else
507
static inline int set_cpus_allowed(task_t *p, cpumask_t new_mask)
508
{
509
        return 0;
510
}
511
#endif
512
 
513
extern unsigned long long sched_clock(void);
514
 
515
#ifdef CONFIG_NUMA
516
extern void sched_balance_exec(void);
517
extern void node_nr_running_init(void);
518
#else
519
#define sched_balance_exec()   {}
520
#define node_nr_running_init() {}
521
#endif
522
 
523
extern void set_user_nice(task_t *p, long nice);
524
extern int task_prio(task_t *p);
525
extern int task_nice(task_t *p);
526
extern int task_curr(task_t *p);
527
extern int idle_cpu(int cpu);
528
 
529
void yield(void);
530
 
531
/*
532
 * The default (Linux) execution domain.
533
 */
534
extern struct exec_domain       default_exec_domain;
535
 
536
#ifndef INIT_THREAD_SIZE
537
# define INIT_THREAD_SIZE       2048*sizeof(long)
538
#endif
539
 
540
union thread_union {
541
        struct thread_info thread_info;
542
        unsigned long stack[INIT_THREAD_SIZE/sizeof(long)];
543
};
544
 
545
#ifndef __HAVE_ARCH_KSTACK_END
546
static inline int kstack_end(void *addr)
547
{
548
        /* Reliable end of stack detection:
549
         * Some APM bios versions misalign the stack
550
         */
551
        return !(((unsigned long)addr+sizeof(void*)-1) & (THREAD_SIZE-sizeof(void*)));
552
}
553
#endif
554
 
555
extern union thread_union init_thread_union;
556
extern struct task_struct init_task;
557
 
558
extern struct   mm_struct init_mm;
559
 
560
extern struct task_struct *find_task_by_pid(int pid);
561
extern void set_special_pids(pid_t session, pid_t pgrp);
562
extern void __set_special_pids(pid_t session, pid_t pgrp);
563
 
564
/* per-UID process charging. */
565
extern struct user_struct * alloc_uid(uid_t);
566
extern void free_uid(struct user_struct *);
567
extern void switch_uid(struct user_struct *);
568
 
569
#include <asm/current.h>
570
 
571
extern unsigned long itimer_ticks;
572
extern unsigned long itimer_next;
573
extern void do_timer(struct pt_regs *);
574
 
575
extern int FASTCALL(wake_up_state(struct task_struct * tsk, unsigned int state));
576
extern int FASTCALL(wake_up_process(struct task_struct * tsk));
577
#ifdef CONFIG_SMP
578
 extern void FASTCALL(kick_process(struct task_struct * tsk));
579
#else
580
 static inline void kick_process(struct task_struct *tsk) { }
581
#endif
582
extern void FASTCALL(wake_up_forked_process(struct task_struct * tsk));
583
extern void FASTCALL(sched_exit(task_t * p));
584
 
585
asmlinkage long sys_wait4(pid_t pid,unsigned int * stat_addr, int options, struct rusage * ru);
586
 
587
extern int in_group_p(gid_t);
588
extern int in_egroup_p(gid_t);
589
 
590
extern void proc_caches_init(void);
591
extern void flush_signals(struct task_struct *);
592
extern void flush_signal_handlers(struct task_struct *, int force_default);
593
extern int dequeue_signal(struct task_struct *tsk, sigset_t *mask, siginfo_t *info);
594
 
595
static inline int dequeue_signal_lock(struct task_struct *tsk, sigset_t *mask, siginfo_t *info)
596
{
597
        unsigned long flags;
598
        int ret;
599
 
600
        spin_lock_irqsave(&tsk->sighand->siglock, flags);
601
        ret = dequeue_signal(tsk, mask, info);
602
        spin_unlock_irqrestore(&tsk->sighand->siglock, flags);
603
 
604
        return ret;
605
}      
606
 
607
extern void block_all_signals(int (*notifier)(void *priv), void *priv,
608
                              sigset_t *mask);
609
extern void unblock_all_signals(void);
610
extern void release_task(struct task_struct * p);
611
extern int send_sig_info(int, struct siginfo *, struct task_struct *);
612
extern int send_group_sig_info(int, struct siginfo *, struct task_struct *);
613
extern int force_sig_info(int, struct siginfo *, struct task_struct *);
614
extern int __kill_pg_info(int sig, struct siginfo *info, pid_t pgrp);
615
extern int kill_pg_info(int, struct siginfo *, pid_t);
616
extern int kill_sl_info(int, struct siginfo *, pid_t);
617
extern int kill_proc_info(int, struct siginfo *, pid_t);
618
extern void notify_parent(struct task_struct *, int);
619
extern void do_notify_parent(struct task_struct *, int);
620
extern void force_sig(int, struct task_struct *);
621
extern void force_sig_specific(int, struct task_struct *);
622
extern int send_sig(int, struct task_struct *, int);
623
extern void zap_other_threads(struct task_struct *p);
624
extern int kill_pg(pid_t, int, int);
625
extern int kill_sl(pid_t, int, int);
626
extern int kill_proc(pid_t, int, int);
627
extern struct sigqueue *sigqueue_alloc(void);
628
extern void sigqueue_free(struct sigqueue *);
629
extern int send_sigqueue(int, struct sigqueue *,  struct task_struct *);
630
extern int send_group_sigqueue(int, struct sigqueue *,  struct task_struct *);
631
extern int do_sigaction(int, const struct k_sigaction *, struct k_sigaction *);
632
extern int do_sigaltstack(const stack_t __user *, stack_t __user *, unsigned long);
633
 
634
/* These can be the second arg to send_sig_info/send_group_sig_info.  */
635
#define SEND_SIG_NOINFO ((struct siginfo *) 0)
636
#define SEND_SIG_PRIV   ((struct siginfo *) 1)
637
#define SEND_SIG_FORCED ((struct siginfo *) 2)
638
 
639
/* True if we are on the alternate signal stack.  */
640
 
641
static inline int on_sig_stack(unsigned long sp)
642
{
643
        return (sp - current->sas_ss_sp < current->sas_ss_size);
644
}
645
 
646
static inline int sas_ss_flags(unsigned long sp)
647
{
648
        return (current->sas_ss_size == 0 ? SS_DISABLE
649
                : on_sig_stack(sp) ? SS_ONSTACK : 0);
650
}
651
 
652
 
653
#ifdef CONFIG_SECURITY
654
/* code is in security.c */
655
extern int capable(int cap);
656
#else
657
static inline int capable(int cap)
658
{
659
        if (cap_raised(current->cap_effective, cap)) {
660
                current->flags |= PF_SUPERPRIV;
661
                return 1;
662
        }
663
        return 0;
664
}
665
#endif
666
 
667
/*
668
 * Routines for handling mm_structs
669
 */
670
extern struct mm_struct * mm_alloc(void);
671
 
672
/* mmdrop drops the mm and the page tables */
673
extern inline void FASTCALL(__mmdrop(struct mm_struct *));
674
static inline void mmdrop(struct mm_struct * mm)
675
{
676
        if (atomic_dec_and_test(&mm->mm_count))
677
                __mmdrop(mm);
678
}
679
 
680
/* mmput gets rid of the mappings and all user-space */
681
extern void mmput(struct mm_struct *);
682
/* Grab a reference to the mm if its not already going away */
683
extern struct mm_struct *mmgrab(struct mm_struct *);
684
/* Remove the current tasks stale references to the old mm_struct */
685
extern void mm_release(struct task_struct *, struct mm_struct *);
686
 
687
extern int  copy_thread(int, unsigned long, unsigned long, unsigned long, struct task_struct *, struct pt_regs *);
688
extern void flush_thread(void);
689
extern void exit_thread(void);
690
 
691
extern void exit_mm(struct task_struct *);
692
extern void exit_files(struct task_struct *);
693
extern void exit_signal(struct task_struct *);
694
extern void __exit_signal(struct task_struct *);
695
extern void exit_sighand(struct task_struct *);
696
extern void __exit_sighand(struct task_struct *);
697
extern void exit_itimers(struct task_struct *);
698
 
699
extern NORET_TYPE void do_group_exit(int);
700
 
701
extern void reparent_to_init(void);
702
extern void daemonize(const char *, ...);
703
extern int allow_signal(int);
704
extern int disallow_signal(int);
705
extern task_t *child_reaper;
706
 
707
extern int do_execve(char *, char __user * __user *, char __user * __user *, struct pt_regs *);
708
extern long do_fork(unsigned long, unsigned long, struct pt_regs *, unsigned long, int __user *, int __user *);
709
extern struct task_struct * copy_process(unsigned long, unsigned long, struct pt_regs *, unsigned long, int __user *, int __user *);
710
 
711
#ifdef CONFIG_SMP
712
extern void wait_task_inactive(task_t * p);
713
#else
714
#define wait_task_inactive(p)   do { } while (0)
715
#endif
716
 
717
#define remove_parent(p)        list_del_init(&(p)->sibling)
718
#define add_parent(p, parent)   list_add_tail(&(p)->sibling,&(parent)->children)
719
 
720
#define REMOVE_LINKS(p) do {                                    \
721
        if (thread_group_leader(p))                             \
722
                list_del_init(&(p)->tasks);                     \
723
        remove_parent(p);                                       \
724
        } while (0)
725
 
726
#define SET_LINKS(p) do {                                       \
727
        if (thread_group_leader(p))                             \
728
                list_add_tail(&(p)->tasks,&init_task.tasks);    \
729
        add_parent(p, (p)->parent);                             \
730
        } while (0)
731
 
732
#define next_task(p)    list_entry((p)->tasks.next, struct task_struct, tasks)
733
#define prev_task(p)    list_entry((p)->tasks.prev, struct task_struct, tasks)
734
 
735
#define for_each_process(p) \
736
        for (p = &init_task ; (p = next_task(p)) != &init_task ; )
737
 
738
/*
739
 * Careful: do_each_thread/while_each_thread is a double loop so
740
 *          'break' will not work as expected - use goto instead.
741
 */
742
#define do_each_thread(g, t) \
743
        for (g = t = &init_task ; (g = t = next_task(g)) != &init_task ; ) do
744
 
745
#define while_each_thread(g, t) \
746
        while ((t = next_thread(t)) != g)
747
 
748
extern task_t * FASTCALL(next_thread(task_t *p));
749
 
750
#define thread_group_leader(p)  (p->pid == p->tgid)
751
 
752
static inline int thread_group_empty(task_t *p)
753
{
754
        struct pid *pid = p->pids[PIDTYPE_TGID].pidptr;
755
 
756
        return pid->task_list.next->next == &pid->task_list;
757
}
758
 
759
#define delay_group_leader(p) \
760
                (thread_group_leader(p) && !thread_group_empty(p))
761
 
762
extern void unhash_process(struct task_struct *p);
763
 
764
/* Protects ->fs, ->files, ->mm, and synchronises with wait4().
765
 * Nests both inside and outside of read_lock(&tasklist_lock).
766
 * It must not be nested with write_lock_irq(&tasklist_lock),
767
 * neither inside nor outside.
768
 */
769
static inline void task_lock(struct task_struct *p)
770
{
771
        spin_lock(&p->alloc_lock);
772
}
773
 
774
static inline void task_unlock(struct task_struct *p)
775
{
776
        spin_unlock(&p->alloc_lock);
777
}
778
 
779
/**
780
 * get_task_mm - acquire a reference to the task's mm
781
 *
782
 * Returns %NULL if the task has no mm. User must release
783
 * the mm via mmput() after use.
784
 */
785
static inline struct mm_struct * get_task_mm(struct task_struct * task)
786
{
787
        struct mm_struct * mm;
788
 
789
        task_lock(task);
790
        mm = task->mm;
791
        if (mm)
792
                mm = mmgrab(mm);
793
        task_unlock(task);
794
 
795
        return mm;
796
}
797
 
798
 
799
/* set thread flags in other task's structures
800
 * - see asm/thread_info.h for TIF_xxxx flags available
801
 */
802
static inline void set_tsk_thread_flag(struct task_struct *tsk, int flag)
803
{
804
        set_ti_thread_flag(tsk->thread_info,flag);
805
}
806
 
807
static inline void clear_tsk_thread_flag(struct task_struct *tsk, int flag)
808
{
809
        clear_ti_thread_flag(tsk->thread_info,flag);
810
}
811
 
812
static inline int test_and_set_tsk_thread_flag(struct task_struct *tsk, int flag)
813
{
814
        return test_and_set_ti_thread_flag(tsk->thread_info,flag);
815
}
816
 
817
static inline int test_and_clear_tsk_thread_flag(struct task_struct *tsk, int flag)
818
{
819
        return test_and_clear_ti_thread_flag(tsk->thread_info,flag);
820
}
821
 
822
static inline int test_tsk_thread_flag(struct task_struct *tsk, int flag)
823
{
824
        return test_ti_thread_flag(tsk->thread_info,flag);
825
}
826
 
827
static inline void set_tsk_need_resched(struct task_struct *tsk)
828
{
829
        set_tsk_thread_flag(tsk,TIF_NEED_RESCHED);
830
}
831
 
832
static inline void clear_tsk_need_resched(struct task_struct *tsk)
833
{
834
        clear_tsk_thread_flag(tsk,TIF_NEED_RESCHED);
835
}
836
 
837
static inline int signal_pending(struct task_struct *p)
838
{
839
        return unlikely(test_tsk_thread_flag(p,TIF_SIGPENDING));
840
}
841
 
842
static inline int need_resched(void)
843
{
844
        return unlikely(test_thread_flag(TIF_NEED_RESCHED));
845
}
846
 
847
extern void __cond_resched(void);
848
static inline void cond_resched(void)
849
{
850
        if (need_resched())
851
                __cond_resched();
852
}
853
 
854
/*
855
 * cond_resched_lock() - if a reschedule is pending, drop the given lock,
856
 * call schedule, and on return reacquire the lock.
857
 *
858
 * This works OK both with and without CONFIG_PREEMPT.  We do strange low-level
859
 * operations here to prevent schedule() from being called twice (once via
860
 * spin_unlock(), once by hand).
861
 */
862
static inline void cond_resched_lock(spinlock_t * lock)
863
{
864
        if (need_resched()) {
865
                _raw_spin_unlock(lock);
866
                preempt_enable_no_resched();
867
                __cond_resched();
868
                spin_lock(lock);
869
        }
870
}
871
 
872
/* Reevaluate whether the task has signals pending delivery.
873
   This is required every time the blocked sigset_t changes.
874
   callers must hold sighand->siglock.  */
875
 
876
extern FASTCALL(void recalc_sigpending_tsk(struct task_struct *t));
877
extern void recalc_sigpending(void);
878
 
879
extern void signal_wake_up(struct task_struct *t, int resume_stopped);
880
 
881
/*
882
 * Wrappers for p->thread_info->cpu access. No-op on UP.
883
 */
884
#ifdef CONFIG_SMP
885
 
886
static inline unsigned int task_cpu(struct task_struct *p)
887
{
888
        return p->thread_info->cpu;
889
}
890
 
891
static inline void set_task_cpu(struct task_struct *p, unsigned int cpu)
892
{
893
        p->thread_info->cpu = cpu;
894
}
895
 
896
#else
897
 
898
static inline unsigned int task_cpu(struct task_struct *p)
899
{
900
        return 0;
901
}
902
 
903
static inline void set_task_cpu(struct task_struct *p, unsigned int cpu)
904
{
905
}
906
 
907
#endif /* CONFIG_SMP */
908
 
909
#endif /* __KERNEL__ */
910
 
911
#endif