sem.c 55.9 KB
Newer Older
L
Linus Torvalds 已提交
1 2 3 4 5 6 7 8
/*
 * linux/ipc/sem.c
 * Copyright (C) 1992 Krishna Balasubramanian
 * Copyright (C) 1995 Eric Schenk, Bruno Haible
 *
 * /proc/sysvipc/sem support (c) 1999 Dragos Acostachioaie <dragos@iname.com>
 *
 * SMP-threaded, sysctl's added
9
 * (c) 1999 Manfred Spraul <manfred@colorfullife.com>
L
Linus Torvalds 已提交
10
 * Enforced range limit on SEM_UNDO
A
Alan Cox 已提交
11
 * (c) 2001 Red Hat Inc
L
Linus Torvalds 已提交
12 13
 * Lockless wakeup
 * (c) 2003 Manfred Spraul <manfred@colorfullife.com>
14 15
 * Further wakeup optimizations, documentation
 * (c) 2010 Manfred Spraul <manfred@colorfullife.com>
S
Steve Grubb 已提交
16 17 18
 *
 * support for audit of ipc object properties and permission changes
 * Dustin Kirkland <dustin.kirkland@us.ibm.com>
K
Kirill Korotaev 已提交
19 20 21 22
 *
 * namespaces support
 * OpenVZ, SWsoft Inc.
 * Pavel Emelianov <xemul@openvz.org>
23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49
 *
 * Implementation notes: (May 2010)
 * This file implements System V semaphores.
 *
 * User space visible behavior:
 * - FIFO ordering for semop() operations (just FIFO, not starvation
 *   protection)
 * - multiple semaphore operations that alter the same semaphore in
 *   one semop() are handled.
 * - sem_ctime (time of last semctl()) is updated in the IPC_SET, SETVAL and
 *   SETALL calls.
 * - two Linux specific semctl() commands: SEM_STAT, SEM_INFO.
 * - undo adjustments at process exit are limited to 0..SEMVMX.
 * - namespace are supported.
 * - SEMMSL, SEMMNS, SEMOPM and SEMMNI can be configured at runtine by writing
 *   to /proc/sys/kernel/sem.
 * - statistics about the usage are reported in /proc/sysvipc/sem.
 *
 * Internals:
 * - scalability:
 *   - all global variables are read-mostly.
 *   - semop() calls and semctl(RMID) are synchronized by RCU.
 *   - most operations do write operations (actually: spin_lock calls) to
 *     the per-semaphore array structure.
 *   Thus: Perfect SMP scaling between independent semaphore arrays.
 *         If multiple semaphores in one array are used, then cache line
 *         trashing on the semaphore array spinlock will limit the scaling.
50
 * - semncnt and semzcnt are calculated on demand in count_semcnt()
51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72
 * - the task that performs a successful semop() scans the list of all
 *   sleeping tasks and completes any pending operations that can be fulfilled.
 *   Semaphores are actively given to waiting tasks (necessary for FIFO).
 *   (see update_queue())
 * - To improve the scalability, the actual wake-up calls are performed after
 *   dropping all locks. (see wake_up_sem_queue_prepare(),
 *   wake_up_sem_queue_do())
 * - All work is done by the waker, the woken up task does not have to do
 *   anything - not even acquiring a lock or dropping a refcount.
 * - A woken up task may not even touch the semaphore array anymore, it may
 *   have been destroyed already by a semctl(RMID).
 * - The synchronizations between wake-ups due to a timeout/signal and a
 *   wake-up due to a completed semaphore operation is achieved by using an
 *   intermediate state (IN_WAKEUP).
 * - UNDO values are stored in an array (one per process and per
 *   semaphore array, lazily allocated). For backwards compatibility, multiple
 *   modes for the UNDO variables are supported (per process, per thread)
 *   (see copy_semundo, CLONE_SYSVSEM)
 * - There are two lists of the pending operations: a per-array list
 *   and per-semaphore list (stored in the array). This allows to achieve FIFO
 *   ordering without always scanning all pending operations.
 *   The worst-case behavior is nevertheless O(N^2) for N wakeups.
L
Linus Torvalds 已提交
73 74 75 76 77 78 79 80 81 82
 */

#include <linux/slab.h>
#include <linux/spinlock.h>
#include <linux/init.h>
#include <linux/proc_fs.h>
#include <linux/time.h>
#include <linux/security.h>
#include <linux/syscalls.h>
#include <linux/audit.h>
83
#include <linux/capability.h>
84
#include <linux/seq_file.h>
N
Nadia Derbey 已提交
85
#include <linux/rwsem.h>
K
Kirill Korotaev 已提交
86
#include <linux/nsproxy.h>
87
#include <linux/ipc_namespace.h>
I
Ingo Molnar 已提交
88

P
Paul McQuade 已提交
89
#include <linux/uaccess.h>
L
Linus Torvalds 已提交
90 91
#include "util.h"

92 93 94
/* One semaphore structure for each semaphore in the system. */
struct sem {
	int	semval;		/* current value */
95 96 97 98 99 100 101 102
	/*
	 * PID of the process that last modified the semaphore. For
	 * Linux, specifically these are:
	 *  - semop
	 *  - semctl, via SETVAL and SETALL.
	 *  - at task exit when performing undo adjustments (see exit_sem).
	 */
	int	sempid;
103
	spinlock_t	lock;	/* spinlock for fine-grained semtimedop */
104 105 106 107
	struct list_head pending_alter; /* pending single-sop operations */
					/* that alter the semaphore */
	struct list_head pending_const; /* pending single-sop operations */
					/* that do not alter the semaphore*/
108
	time_t	sem_otime;	/* candidate for sem_otime */
109
} ____cacheline_aligned_in_smp;
110 111 112 113 114 115 116 117 118

/* One queue for each sleeping process in the system. */
struct sem_queue {
	struct list_head	list;	 /* queue of pending operations */
	struct task_struct	*sleeper; /* this process */
	struct sem_undo		*undo;	 /* undo structure */
	int			pid;	 /* process id of requesting process */
	int			status;	 /* completion status of operation */
	struct sembuf		*sops;	 /* array of pending operations */
119
	struct sembuf		*blocking; /* the operation that blocked */
120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149
	int			nsops;	 /* number of operations */
	int			alter;	 /* does *sops alter the array? */
};

/* Each task has a list of undo requests. They are executed automatically
 * when the process exits.
 */
struct sem_undo {
	struct list_head	list_proc;	/* per-process list: *
						 * all undos from one process
						 * rcu protected */
	struct rcu_head		rcu;		/* rcu struct for sem_undo */
	struct sem_undo_list	*ulp;		/* back ptr to sem_undo_list */
	struct list_head	list_id;	/* per semaphore array list:
						 * all undos for one array */
	int			semid;		/* semaphore set identifier */
	short			*semadj;	/* array of adjustments */
						/* one per semaphore */
};

/* sem_undo_list controls shared access to the list of sem_undo structures
 * that may be shared among all a CLONE_SYSVSEM task group.
 */
struct sem_undo_list {
	atomic_t		refcnt;
	spinlock_t		lock;
	struct list_head	list_proc;
};


150
#define sem_ids(ns)	((ns)->ids[IPC_SEM_IDS])
K
Kirill Korotaev 已提交
151

N
Nadia Derbey 已提交
152
#define sem_checkid(sma, semid)	ipc_checkid(&sma->sem_perm, semid)
L
Linus Torvalds 已提交
153

N
Nadia Derbey 已提交
154
static int newary(struct ipc_namespace *, struct ipc_params *);
155
static void freeary(struct ipc_namespace *, struct kern_ipc_perm *);
L
Linus Torvalds 已提交
156
#ifdef CONFIG_PROC_FS
157
static int sysvipc_sem_proc_show(struct seq_file *s, void *it);
L
Linus Torvalds 已提交
158 159 160 161 162 163
#endif

#define SEMMSL_FAST	256 /* 512 bytes on stack */
#define SEMOPM_FAST	64  /* ~ 372 bytes on stack */

/*
164
 * Locking:
L
Linus Torvalds 已提交
165
 *	sem_undo.id_next,
166
 *	sem_array.complex_count,
167
 *	sem_array.pending{_alter,_cont},
168
 *	sem_array.sem_undo: global sem_lock() for read/write
L
Linus Torvalds 已提交
169
 *	sem_undo.proc_next: only "current" is allowed to read/write that field.
P
Paul McQuade 已提交
170
 *
171 172
 *	sem_array.sem_base[i].pending_{const,alter}:
 *		global or semaphore sem_lock() for read/write
L
Linus Torvalds 已提交
173 174
 */

K
Kirill Korotaev 已提交
175 176 177 178 179
#define sc_semmsl	sem_ctls[0]
#define sc_semmns	sem_ctls[1]
#define sc_semopm	sem_ctls[2]
#define sc_semmni	sem_ctls[3]

180
void sem_init_ns(struct ipc_namespace *ns)
K
Kirill Korotaev 已提交
181 182 183 184 185 186
{
	ns->sc_semmsl = SEMMSL;
	ns->sc_semmns = SEMMNS;
	ns->sc_semopm = SEMOPM;
	ns->sc_semmni = SEMMNI;
	ns->used_sems = 0;
187
	ipc_init_ids(&ns->ids[IPC_SEM_IDS]);
K
Kirill Korotaev 已提交
188 189
}

190
#ifdef CONFIG_IPC_NS
K
Kirill Korotaev 已提交
191 192
void sem_exit_ns(struct ipc_namespace *ns)
{
193
	free_ipcs(ns, &sem_ids(ns), freeary);
S
Serge E. Hallyn 已提交
194
	idr_destroy(&ns->ids[IPC_SEM_IDS].ipcs_idr);
K
Kirill Korotaev 已提交
195
}
196
#endif
L
Linus Torvalds 已提交
197

M
Manfred Spraul 已提交
198
void __init sem_init(void)
L
Linus Torvalds 已提交
199
{
200
	sem_init_ns(&init_ipc_ns);
201 202
	ipc_init_proc_interface("sysvipc/sem",
				"       key      semid perms      nsems   uid   gid  cuid  cgid      otime      ctime\n",
K
Kirill Korotaev 已提交
203
				IPC_SEM_IDS, sysvipc_sem_proc_show);
L
Linus Torvalds 已提交
204 205
}

206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234
/**
 * unmerge_queues - unmerge queues, if possible.
 * @sma: semaphore array
 *
 * The function unmerges the wait queues if complex_count is 0.
 * It must be called prior to dropping the global semaphore array lock.
 */
static void unmerge_queues(struct sem_array *sma)
{
	struct sem_queue *q, *tq;

	/* complex operations still around? */
	if (sma->complex_count)
		return;
	/*
	 * We will switch back to simple mode.
	 * Move all pending operation back into the per-semaphore
	 * queues.
	 */
	list_for_each_entry_safe(q, tq, &sma->pending_alter, list) {
		struct sem *curr;
		curr = &sma->sem_base[q->sops[0].sem_num];

		list_add_tail(&q->list, &curr->pending_alter);
	}
	INIT_LIST_HEAD(&sma->pending_alter);
}

/**
D
Davidlohr Bueso 已提交
235
 * merge_queues - merge single semop queues into global queue
236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252
 * @sma: semaphore array
 *
 * This function merges all per-semaphore queues into the global queue.
 * It is necessary to achieve FIFO ordering for the pending single-sop
 * operations when a multi-semop operation must sleep.
 * Only the alter operations must be moved, the const operations can stay.
 */
static void merge_queues(struct sem_array *sma)
{
	int i;
	for (i = 0; i < sma->sem_nsems; i++) {
		struct sem *sem = sma->sem_base + i;

		list_splice_init(&sem->pending_alter, &sma->pending_alter);
	}
}

D
Davidlohr Bueso 已提交
253 254 255 256 257 258 259 260 261
static void sem_rcu_free(struct rcu_head *head)
{
	struct ipc_rcu *p = container_of(head, struct ipc_rcu, rcu);
	struct sem_array *sma = ipc_rcu_to_struct(p);

	security_sem_free(sma);
	ipc_rcu_free(head);
}

262 263 264 265 266
/*
 * Wait until all currently ongoing simple ops have completed.
 * Caller must own sem_perm.lock.
 * New simple ops cannot start, because simple ops first check
 * that sem_perm.lock is free.
M
Manfred Spraul 已提交
267
 * that a) sem_perm.lock is free and b) complex_count is 0.
268 269 270 271 272 273
 */
static void sem_wait_array(struct sem_array *sma)
{
	int i;
	struct sem *sem;

M
Manfred Spraul 已提交
274 275 276 277 278 279 280
	if (sma->complex_count)  {
		/* The thread that increased sma->complex_count waited on
		 * all sem->lock locks. Thus we don't need to wait again.
		 */
		return;
	}

281 282 283 284 285 286
	for (i = 0; i < sma->sem_nsems; i++) {
		sem = sma->sem_base + i;
		spin_unlock_wait(&sem->lock);
	}
}

287 288 289 290 291 292 293 294 295 296
/*
 * If the request contains only one semaphore operation, and there are
 * no complex transactions pending, lock only the semaphore involved.
 * Otherwise, lock the entire semaphore array, since we either have
 * multiple semaphores in our own semops, or we need to look at
 * semaphores from other pending complex operations.
 */
static inline int sem_lock(struct sem_array *sma, struct sembuf *sops,
			      int nsops)
{
297
	struct sem *sem;
298

299 300 301
	if (nsops != 1) {
		/* Complex operation - acquire a full lock */
		ipc_lock_object(&sma->sem_perm);
302

303 304
		/* And wait until all simple ops that are processed
		 * right now have dropped their locks.
305
		 */
306 307 308 309 310 311 312 313 314 315 316 317 318 319 320 321 322 323 324 325
		sem_wait_array(sma);
		return -1;
	}

	/*
	 * Only one semaphore affected - try to optimize locking.
	 * The rules are:
	 * - optimized locking is possible if no complex operation
	 *   is either enqueued or processed right now.
	 * - The test for enqueued complex ops is simple:
	 *      sma->complex_count != 0
	 * - Testing for complex ops that are processed right now is
	 *   a bit more difficult. Complex ops acquire the full lock
	 *   and first wait that the running simple ops have completed.
	 *   (see above)
	 *   Thus: If we own a simple lock and the global lock is free
	 *	and complex_count is now 0, then it will stay 0 and
	 *	thus just locking sem->lock is sufficient.
	 */
	sem = sma->sem_base + sops->sem_num;
326

327
	if (sma->complex_count == 0) {
328
		/*
329 330
		 * It appears that no complex operation is around.
		 * Acquire the per-semaphore lock.
331
		 */
332 333 334 335
		spin_lock(&sem->lock);

		/* Then check that the global lock is free */
		if (!spin_is_locked(&sma->sem_perm.lock)) {
336
			/*
337 338
			 * We need a memory barrier with acquire semantics,
			 * otherwise we can race with another thread that does:
339 340 341
			 *	complex_count++;
			 *	spin_unlock(sem_perm.lock);
			 */
342
			smp_acquire__after_ctrl_dep();
343

344 345
			/*
			 * Now repeat the test of complex_count:
346 347 348 349 350 351 352
			 * It can't change anymore until we drop sem->lock.
			 * Thus: if is now 0, then it will stay 0.
			 */
			if (sma->complex_count == 0) {
				/* fast path successful! */
				return sops->sem_num;
			}
353
		}
354 355 356 357 358
		spin_unlock(&sem->lock);
	}

	/* slow path: acquire the full lock */
	ipc_lock_object(&sma->sem_perm);
359

360 361 362 363 364 365 366 367
	if (sma->complex_count == 0) {
		/* False alarm:
		 * There is no complex operation, thus we can switch
		 * back to the fast path.
		 */
		spin_lock(&sem->lock);
		ipc_unlock_object(&sma->sem_perm);
		return sops->sem_num;
368
	} else {
369 370
		/* Not a false alarm, thus complete the sequence for a
		 * full lock.
371
		 */
372 373
		sem_wait_array(sma);
		return -1;
374 375 376 377 378 379
	}
}

static inline void sem_unlock(struct sem_array *sma, int locknum)
{
	if (locknum == -1) {
380
		unmerge_queues(sma);
381
		ipc_unlock_object(&sma->sem_perm);
382 383 384 385 386 387
	} else {
		struct sem *sem = sma->sem_base + locknum;
		spin_unlock(&sem->lock);
	}
}

N
Nadia Derbey 已提交
388
/*
D
Davidlohr Bueso 已提交
389
 * sem_lock_(check_) routines are called in the paths where the rwsem
N
Nadia Derbey 已提交
390
 * is not held.
391 392
 *
 * The caller holds the RCU read lock.
N
Nadia Derbey 已提交
393
 */
394 395
static inline struct sem_array *sem_obtain_lock(struct ipc_namespace *ns,
			int id, struct sembuf *sops, int nsops, int *locknum)
396
{
397 398
	struct kern_ipc_perm *ipcp;
	struct sem_array *sma;
N
Nadia Derbey 已提交
399

D
Davidlohr Bueso 已提交
400
	ipcp = ipc_obtain_object_idr(&sem_ids(ns), id);
401 402
	if (IS_ERR(ipcp))
		return ERR_CAST(ipcp);
403

404 405
	sma = container_of(ipcp, struct sem_array, sem_perm);
	*locknum = sem_lock(sma, sops, nsops);
406 407 408 409

	/* ipc_rmid() may have already freed the ID while sem_lock
	 * was spinning: verify that the structure is still valid
	 */
410
	if (ipc_valid_object(ipcp))
411 412
		return container_of(ipcp, struct sem_array, sem_perm);

413
	sem_unlock(sma, *locknum);
414
	return ERR_PTR(-EINVAL);
415 416
}

417 418
static inline struct sem_array *sem_obtain_object(struct ipc_namespace *ns, int id)
{
D
Davidlohr Bueso 已提交
419
	struct kern_ipc_perm *ipcp = ipc_obtain_object_idr(&sem_ids(ns), id);
420 421 422 423 424 425 426 427 428 429 430 431 432 433

	if (IS_ERR(ipcp))
		return ERR_CAST(ipcp);

	return container_of(ipcp, struct sem_array, sem_perm);
}

static inline struct sem_array *sem_obtain_object_check(struct ipc_namespace *ns,
							int id)
{
	struct kern_ipc_perm *ipcp = ipc_obtain_object_check(&sem_ids(ns), id);

	if (IS_ERR(ipcp))
		return ERR_CAST(ipcp);
434

N
Nadia Derbey 已提交
435
	return container_of(ipcp, struct sem_array, sem_perm);
436 437
}

438 439
static inline void sem_lock_and_putref(struct sem_array *sma)
{
440
	sem_lock(sma, NULL, -1);
D
Davidlohr Bueso 已提交
441
	ipc_rcu_putref(sma, ipc_rcu_free);
442 443
}

N
Nadia Derbey 已提交
444 445 446 447 448
static inline void sem_rmid(struct ipc_namespace *ns, struct sem_array *s)
{
	ipc_rmid(&sem_ids(ns), &s->sem_perm);
}

L
Linus Torvalds 已提交
449 450 451 452 453
/*
 * Lockless wakeup algorithm:
 * Without the check/retry algorithm a lockless wakeup is possible:
 * - queue.status is initialized to -EINTR before blocking.
 * - wakeup is performed by
454
 *	* unlinking the queue entry from the pending list
L
Linus Torvalds 已提交
455 456 457 458 459 460
 *	* setting queue.status to IN_WAKEUP
 *	  This is the notification for the blocked thread that a
 *	  result value is imminent.
 *	* call wake_up_process
 *	* set queue.status to the final value.
 * - the previously blocked thread checks queue.status:
M
Manfred Spraul 已提交
461 462 463 464 465
 *	* if it's IN_WAKEUP, then it must wait until the value changes
 *	* if it's not -EINTR, then the operation was completed by
 *	  update_queue. semtimedop can return queue.status without
 *	  performing any operation on the sem array.
 *	* otherwise it must acquire the spinlock and check what's up.
L
Linus Torvalds 已提交
466 467 468 469 470 471 472 473 474 475 476 477 478 479 480 481 482
 *
 * The two-stage algorithm is necessary to protect against the following
 * races:
 * - if queue.status is set after wake_up_process, then the woken up idle
 *   thread could race forward and try (and fail) to acquire sma->lock
 *   before update_queue had a chance to set queue.status
 * - if queue.status is written before wake_up_process and if the
 *   blocked process is woken up by a signal between writing
 *   queue.status and the wake_up_process, then the woken up
 *   process could return from semtimedop and die by calling
 *   sys_exit before wake_up_process is called. Then wake_up_process
 *   will oops, because the task structure is already invalid.
 *   (yes, this happened on s390 with sysv msg).
 *
 */
#define IN_WAKEUP	1

N
Nadia Derbey 已提交
483 484 485 486 487
/**
 * newary - Create a new semaphore set
 * @ns: namespace
 * @params: ptr to the structure that contains key, semflg and nsems
 *
D
Davidlohr Bueso 已提交
488
 * Called with sem_ids.rwsem held (as a writer)
N
Nadia Derbey 已提交
489
 */
N
Nadia Derbey 已提交
490
static int newary(struct ipc_namespace *ns, struct ipc_params *params)
L
Linus Torvalds 已提交
491 492 493 494 495
{
	int id;
	int retval;
	struct sem_array *sma;
	int size;
N
Nadia Derbey 已提交
496 497 498
	key_t key = params->key;
	int nsems = params->u.nsems;
	int semflg = params->flg;
499
	int i;
L
Linus Torvalds 已提交
500 501 502

	if (!nsems)
		return -EINVAL;
K
Kirill Korotaev 已提交
503
	if (ns->used_sems + nsems > ns->sc_semmns)
L
Linus Torvalds 已提交
504 505
		return -ENOSPC;

M
Manfred Spraul 已提交
506
	size = sizeof(*sma) + nsems * sizeof(struct sem);
L
Linus Torvalds 已提交
507
	sma = ipc_rcu_alloc(size);
508
	if (!sma)
L
Linus Torvalds 已提交
509
		return -ENOMEM;
510

M
Manfred Spraul 已提交
511
	memset(sma, 0, size);
L
Linus Torvalds 已提交
512 513 514 515 516 517 518

	sma->sem_perm.mode = (semflg & S_IRWXUGO);
	sma->sem_perm.key = key;

	sma->sem_perm.security = NULL;
	retval = security_sem_alloc(sma);
	if (retval) {
D
Davidlohr Bueso 已提交
519
		ipc_rcu_putref(sma, ipc_rcu_free);
L
Linus Torvalds 已提交
520 521 522 523
		return retval;
	}

	sma->sem_base = (struct sem *) &sma[1];
524

525
	for (i = 0; i < nsems; i++) {
526 527
		INIT_LIST_HEAD(&sma->sem_base[i].pending_alter);
		INIT_LIST_HEAD(&sma->sem_base[i].pending_const);
528 529
		spin_lock_init(&sma->sem_base[i].lock);
	}
530 531

	sma->complex_count = 0;
532 533
	INIT_LIST_HEAD(&sma->pending_alter);
	INIT_LIST_HEAD(&sma->pending_const);
534
	INIT_LIST_HEAD(&sma->list_id);
L
Linus Torvalds 已提交
535 536
	sma->sem_nsems = nsems;
	sma->sem_ctime = get_seconds();
537 538 539 540 541 542 543 544

	id = ipc_addid(&sem_ids(ns), &sma->sem_perm, ns->sc_semmni);
	if (id < 0) {
		ipc_rcu_putref(sma, sem_rcu_free);
		return id;
	}
	ns->used_sems += nsems;

545
	sem_unlock(sma, -1);
546
	rcu_read_unlock();
L
Linus Torvalds 已提交
547

N
Nadia Derbey 已提交
548
	return sma->sem_perm.id;
L
Linus Torvalds 已提交
549 550
}

N
Nadia Derbey 已提交
551

N
Nadia Derbey 已提交
552
/*
D
Davidlohr Bueso 已提交
553
 * Called with sem_ids.rwsem and ipcp locked.
N
Nadia Derbey 已提交
554
 */
N
Nadia Derbey 已提交
555
static inline int sem_security(struct kern_ipc_perm *ipcp, int semflg)
N
Nadia Derbey 已提交
556
{
N
Nadia Derbey 已提交
557 558 559 560
	struct sem_array *sma;

	sma = container_of(ipcp, struct sem_array, sem_perm);
	return security_sem_associate(sma, semflg);
N
Nadia Derbey 已提交
561 562
}

N
Nadia Derbey 已提交
563
/*
D
Davidlohr Bueso 已提交
564
 * Called with sem_ids.rwsem and ipcp locked.
N
Nadia Derbey 已提交
565
 */
N
Nadia Derbey 已提交
566 567
static inline int sem_more_checks(struct kern_ipc_perm *ipcp,
				struct ipc_params *params)
N
Nadia Derbey 已提交
568
{
N
Nadia Derbey 已提交
569 570 571 572
	struct sem_array *sma;

	sma = container_of(ipcp, struct sem_array, sem_perm);
	if (params->u.nsems > sma->sem_nsems)
N
Nadia Derbey 已提交
573 574 575 576 577
		return -EINVAL;

	return 0;
}

578
SYSCALL_DEFINE3(semget, key_t, key, int, nsems, int, semflg)
L
Linus Torvalds 已提交
579
{
K
Kirill Korotaev 已提交
580
	struct ipc_namespace *ns;
M
Mathias Krause 已提交
581 582 583 584 585
	static const struct ipc_ops sem_ops = {
		.getnew = newary,
		.associate = sem_security,
		.more_checks = sem_more_checks,
	};
N
Nadia Derbey 已提交
586
	struct ipc_params sem_params;
K
Kirill Korotaev 已提交
587 588

	ns = current->nsproxy->ipc_ns;
L
Linus Torvalds 已提交
589

K
Kirill Korotaev 已提交
590
	if (nsems < 0 || nsems > ns->sc_semmsl)
L
Linus Torvalds 已提交
591
		return -EINVAL;
N
Nadia Derbey 已提交
592

N
Nadia Derbey 已提交
593 594 595
	sem_params.key = key;
	sem_params.flg = semflg;
	sem_params.u.nsems = nsems;
L
Linus Torvalds 已提交
596

N
Nadia Derbey 已提交
597
	return ipcget(ns, &sem_ids(ns), &sem_ops, &sem_params);
L
Linus Torvalds 已提交
598 599
}

600 601
/**
 * perform_atomic_semop - Perform (if possible) a semaphore operation
602
 * @sma: semaphore array
603
 * @q: struct sem_queue that describes the operation
604 605 606 607
 *
 * Returns 0 if the operation was possible.
 * Returns 1 if the operation is impossible, the caller must sleep.
 * Negative values are error codes.
L
Linus Torvalds 已提交
608
 */
609
static int perform_atomic_semop(struct sem_array *sma, struct sem_queue *q)
L
Linus Torvalds 已提交
610
{
611
	int result, sem_op, nsops, pid;
L
Linus Torvalds 已提交
612
	struct sembuf *sop;
M
Manfred Spraul 已提交
613
	struct sem *curr;
614 615 616 617 618 619
	struct sembuf *sops;
	struct sem_undo *un;

	sops = q->sops;
	nsops = q->nsops;
	un = q->undo;
L
Linus Torvalds 已提交
620 621 622 623 624

	for (sop = sops; sop < sops + nsops; sop++) {
		curr = sma->sem_base + sop->sem_num;
		sem_op = sop->sem_op;
		result = curr->semval;
625

L
Linus Torvalds 已提交
626 627 628 629 630 631 632 633
		if (!sem_op && result)
			goto would_block;

		result += sem_op;
		if (result < 0)
			goto would_block;
		if (result > SEMVMX)
			goto out_of_range;
634

L
Linus Torvalds 已提交
635 636
		if (sop->sem_flg & SEM_UNDO) {
			int undo = un->semadj[sop->sem_num] - sem_op;
637
			/* Exceeding the undo range is an error. */
L
Linus Torvalds 已提交
638 639
			if (undo < (-SEMAEM - 1) || undo > SEMAEM)
				goto out_of_range;
640
			un->semadj[sop->sem_num] = undo;
L
Linus Torvalds 已提交
641
		}
642

L
Linus Torvalds 已提交
643 644 645 646
		curr->semval = result;
	}

	sop--;
647
	pid = q->pid;
L
Linus Torvalds 已提交
648 649 650 651
	while (sop >= sops) {
		sma->sem_base[sop->sem_num].sempid = pid;
		sop--;
	}
652

L
Linus Torvalds 已提交
653 654 655 656 657 658 659
	return 0;

out_of_range:
	result = -ERANGE;
	goto undo;

would_block:
660 661
	q->blocking = sop;

L
Linus Torvalds 已提交
662 663 664 665 666 667 668 669
	if (sop->sem_flg & IPC_NOWAIT)
		result = -EAGAIN;
	else
		result = 1;

undo:
	sop--;
	while (sop >= sops) {
670 671 672 673
		sem_op = sop->sem_op;
		sma->sem_base[sop->sem_num].semval -= sem_op;
		if (sop->sem_flg & SEM_UNDO)
			un->semadj[sop->sem_num] += sem_op;
L
Linus Torvalds 已提交
674 675 676 677 678 679
		sop--;
	}

	return result;
}

680 681 682 683 684
/** wake_up_sem_queue_prepare(q, error): Prepare wake-up
 * @q: queue entry that must be signaled
 * @error: Error value for the signal
 *
 * Prepare the wake-up of the queue entry q.
N
Nick Piggin 已提交
685
 */
686 687
static void wake_up_sem_queue_prepare(struct list_head *pt,
				struct sem_queue *q, int error)
N
Nick Piggin 已提交
688
{
689 690 691 692 693 694 695
	if (list_empty(pt)) {
		/*
		 * Hold preempt off so that we don't get preempted and have the
		 * wakee busy-wait until we're scheduled back on.
		 */
		preempt_disable();
	}
N
Nick Piggin 已提交
696
	q->status = IN_WAKEUP;
697 698
	q->pid = error;

699
	list_add_tail(&q->list, pt);
700 701 702
}

/**
D
Davidlohr Bueso 已提交
703
 * wake_up_sem_queue_do - do the actual wake-up
704 705 706 707 708 709 710 711 712 713 714 715 716
 * @pt: list of tasks to be woken up
 *
 * Do the actual wake-up.
 * The function is called without any locks held, thus the semaphore array
 * could be destroyed already and the tasks can disappear as soon as the
 * status is set to the actual return code.
 */
static void wake_up_sem_queue_do(struct list_head *pt)
{
	struct sem_queue *q, *t;
	int did_something;

	did_something = !list_empty(pt);
717
	list_for_each_entry_safe(q, t, pt, list) {
718 719 720 721 722 723 724
		wake_up_process(q->sleeper);
		/* q can disappear immediately after writing q->status. */
		smp_wmb();
		q->status = q->pid;
	}
	if (did_something)
		preempt_enable();
N
Nick Piggin 已提交
725 726
}

727 728 729
static void unlink_queue(struct sem_array *sma, struct sem_queue *q)
{
	list_del(&q->list);
730
	if (q->nsops > 1)
731 732 733
		sma->complex_count--;
}

734 735 736 737 738 739 740
/** check_restart(sma, q)
 * @sma: semaphore array
 * @q: the operation that just completed
 *
 * update_queue is O(N^2) when it restarts scanning the whole queue of
 * waiting operations. Therefore this function checks if the restart is
 * really necessary. It is called after a previously waiting operation
741 742
 * modified the array.
 * Note that wait-for-zero operations are handled without restart.
743 744 745
 */
static int check_restart(struct sem_array *sma, struct sem_queue *q)
{
746 747
	/* pending complex alter operations are too difficult to analyse */
	if (!list_empty(&sma->pending_alter))
748 749 750 751 752 753
		return 1;

	/* we were a sleeping complex operation. Too difficult */
	if (q->nsops > 1)
		return 1;

754 755 756 757 758 759 760 761 762 763 764 765 766
	/* It is impossible that someone waits for the new value:
	 * - complex operations always restart.
	 * - wait-for-zero are handled seperately.
	 * - q is a previously sleeping simple operation that
	 *   altered the array. It must be a decrement, because
	 *   simple increments never sleep.
	 * - If there are older (higher priority) decrements
	 *   in the queue, then they have observed the original
	 *   semval value and couldn't proceed. The operation
	 *   decremented to value - thus they won't proceed either.
	 */
	return 0;
}
767

768
/**
D
Davidlohr Bueso 已提交
769
 * wake_const_ops - wake up non-alter tasks
770 771 772 773 774 775 776 777 778 779 780 781 782 783 784 785 786 787 788 789 790 791 792 793
 * @sma: semaphore array.
 * @semnum: semaphore that was modified.
 * @pt: list head for the tasks that must be woken up.
 *
 * wake_const_ops must be called after a semaphore in a semaphore array
 * was set to 0. If complex const operations are pending, wake_const_ops must
 * be called with semnum = -1, as well as with the number of each modified
 * semaphore.
 * The tasks that must be woken up are added to @pt. The return code
 * is stored in q->pid.
 * The function returns 1 if at least one operation was completed successfully.
 */
static int wake_const_ops(struct sem_array *sma, int semnum,
				struct list_head *pt)
{
	struct sem_queue *q;
	struct list_head *walk;
	struct list_head *pending_list;
	int semop_completed = 0;

	if (semnum == -1)
		pending_list = &sma->pending_const;
	else
		pending_list = &sma->sem_base[semnum].pending_const;
794

795 796 797 798 799 800 801
	walk = pending_list->next;
	while (walk != pending_list) {
		int error;

		q = container_of(walk, struct sem_queue, list);
		walk = walk->next;

802
		error = perform_atomic_semop(sma, q);
803 804 805 806 807 808 809 810 811 812 813 814 815 816 817

		if (error <= 0) {
			/* operation completed, remove from queue & wakeup */

			unlink_queue(sma, q);

			wake_up_sem_queue_prepare(pt, q, error);
			if (error == 0)
				semop_completed = 1;
		}
	}
	return semop_completed;
}

/**
D
Davidlohr Bueso 已提交
818
 * do_smart_wakeup_zero - wakeup all wait for zero tasks
819 820 821 822 823
 * @sma: semaphore array
 * @sops: operations that were performed
 * @nsops: number of operations
 * @pt: list head of the tasks that must be woken up.
 *
D
Davidlohr Bueso 已提交
824 825
 * Checks all required queue for wait-for-zero operations, based
 * on the actual changes that were performed on the semaphore array.
826 827 828 829 830 831 832 833 834 835 836 837 838 839 840 841 842 843 844 845 846 847 848
 * The function returns 1 if at least one operation was completed successfully.
 */
static int do_smart_wakeup_zero(struct sem_array *sma, struct sembuf *sops,
					int nsops, struct list_head *pt)
{
	int i;
	int semop_completed = 0;
	int got_zero = 0;

	/* first: the per-semaphore queues, if known */
	if (sops) {
		for (i = 0; i < nsops; i++) {
			int num = sops[i].sem_num;

			if (sma->sem_base[num].semval == 0) {
				got_zero = 1;
				semop_completed |= wake_const_ops(sma, num, pt);
			}
		}
	} else {
		/*
		 * No sops means modified semaphores not known.
		 * Assume all were changed.
849
		 */
850 851 852 853 854 855
		for (i = 0; i < sma->sem_nsems; i++) {
			if (sma->sem_base[i].semval == 0) {
				got_zero = 1;
				semop_completed |= wake_const_ops(sma, i, pt);
			}
		}
856 857
	}
	/*
858 859
	 * If one of the modified semaphores got 0,
	 * then check the global queue, too.
860
	 */
861 862
	if (got_zero)
		semop_completed |= wake_const_ops(sma, -1, pt);
863

864
	return semop_completed;
865 866
}

867 868

/**
D
Davidlohr Bueso 已提交
869
 * update_queue - look for tasks that can be completed.
870 871
 * @sma: semaphore array.
 * @semnum: semaphore that was modified.
872
 * @pt: list head for the tasks that must be woken up.
873 874
 *
 * update_queue must be called after a semaphore in a semaphore array
875 876 877
 * was modified. If multiple semaphores were modified, update_queue must
 * be called with semnum = -1, as well as with the number of each modified
 * semaphore.
878 879
 * The tasks that must be woken up are added to @pt. The return code
 * is stored in q->pid.
880 881
 * The function internally checks if const operations can now succeed.
 *
882
 * The function return 1 if at least one semop was completed successfully.
L
Linus Torvalds 已提交
883
 */
884
static int update_queue(struct sem_array *sma, int semnum, struct list_head *pt)
L
Linus Torvalds 已提交
885
{
886 887 888
	struct sem_queue *q;
	struct list_head *walk;
	struct list_head *pending_list;
889
	int semop_completed = 0;
890

891
	if (semnum == -1)
892
		pending_list = &sma->pending_alter;
893
	else
894
		pending_list = &sma->sem_base[semnum].pending_alter;
N
Nick Piggin 已提交
895 896

again:
897 898
	walk = pending_list->next;
	while (walk != pending_list) {
899
		int error, restart;
900

901
		q = container_of(walk, struct sem_queue, list);
902
		walk = walk->next;
L
Linus Torvalds 已提交
903

904 905
		/* If we are scanning the single sop, per-semaphore list of
		 * one semaphore and that semaphore is 0, then it is not
906
		 * necessary to scan further: simple increments
907 908 909 910
		 * that affect only one entry succeed immediately and cannot
		 * be in the  per semaphore pending queue, and decrements
		 * cannot be successful if the value is already 0.
		 */
911
		if (semnum != -1 && sma->sem_base[semnum].semval == 0)
912 913
			break;

914
		error = perform_atomic_semop(sma, q);
L
Linus Torvalds 已提交
915 916

		/* Does q->sleeper still need to sleep? */
N
Nick Piggin 已提交
917 918 919
		if (error > 0)
			continue;

920
		unlink_queue(sma, q);
N
Nick Piggin 已提交
921

922
		if (error) {
923
			restart = 0;
924 925
		} else {
			semop_completed = 1;
926
			do_smart_wakeup_zero(sma, q->sops, q->nsops, pt);
927
			restart = check_restart(sma, q);
928
		}
929

930
		wake_up_sem_queue_prepare(pt, q, error);
931
		if (restart)
N
Nick Piggin 已提交
932
			goto again;
L
Linus Torvalds 已提交
933
	}
934
	return semop_completed;
L
Linus Torvalds 已提交
935 936
}

937
/**
D
Davidlohr Bueso 已提交
938
 * set_semotime - set sem_otime
939 940 941 942 943 944 945 946 947 948 949 950 951 952 953 954
 * @sma: semaphore array
 * @sops: operations that modified the array, may be NULL
 *
 * sem_otime is replicated to avoid cache line trashing.
 * This function sets one instance to the current time.
 */
static void set_semotime(struct sem_array *sma, struct sembuf *sops)
{
	if (sops == NULL) {
		sma->sem_base[0].sem_otime = get_seconds();
	} else {
		sma->sem_base[sops[0].sem_num].sem_otime =
							get_seconds();
	}
}

955
/**
D
Davidlohr Bueso 已提交
956
 * do_smart_update - optimized update_queue
957 958 959
 * @sma: semaphore array
 * @sops: operations that were performed
 * @nsops: number of operations
960 961
 * @otime: force setting otime
 * @pt: list head of the tasks that must be woken up.
962
 *
963 964
 * do_smart_update() does the required calls to update_queue and wakeup_zero,
 * based on the actual changes that were performed on the semaphore array.
965 966 967
 * Note that the function does not do the actual wake-up: the caller is
 * responsible for calling wake_up_sem_queue_do(@pt).
 * It is safe to perform this call after dropping all locks.
968
 */
969 970
static void do_smart_update(struct sem_array *sma, struct sembuf *sops, int nsops,
			int otime, struct list_head *pt)
971 972 973
{
	int i;

974 975
	otime |= do_smart_wakeup_zero(sma, sops, nsops, pt);

976 977 978 979 980 981 982 983 984 985 986 987 988 989 990 991 992 993 994 995 996 997 998 999 1000 1001
	if (!list_empty(&sma->pending_alter)) {
		/* semaphore array uses the global queue - just process it. */
		otime |= update_queue(sma, -1, pt);
	} else {
		if (!sops) {
			/*
			 * No sops, thus the modified semaphores are not
			 * known. Check all.
			 */
			for (i = 0; i < sma->sem_nsems; i++)
				otime |= update_queue(sma, i, pt);
		} else {
			/*
			 * Check the semaphores that were increased:
			 * - No complex ops, thus all sleeping ops are
			 *   decrease.
			 * - if we decreased the value, then any sleeping
			 *   semaphore ops wont be able to run: If the
			 *   previous value was too small, then the new
			 *   value will be too small, too.
			 */
			for (i = 0; i < nsops; i++) {
				if (sops[i].sem_op > 0) {
					otime |= update_queue(sma,
							sops[i].sem_num, pt);
				}
1002
			}
1003
		}
1004
	}
1005 1006
	if (otime)
		set_semotime(sma, sops);
1007 1008
}

1009
/*
1010
 * check_qop: Test if a queued operation sleeps on the semaphore semnum
1011 1012 1013 1014
 */
static int check_qop(struct sem_array *sma, int semnum, struct sem_queue *q,
			bool count_zero)
{
1015
	struct sembuf *sop = q->blocking;
1016

1017 1018 1019 1020 1021 1022 1023 1024 1025 1026 1027
	/*
	 * Linux always (since 0.99.10) reported a task as sleeping on all
	 * semaphores. This violates SUS, therefore it was changed to the
	 * standard compliant behavior.
	 * Give the administrators a chance to notice that an application
	 * might misbehave because it relies on the Linux behavior.
	 */
	pr_info_once("semctl(GETNCNT/GETZCNT) is since 3.16 Single Unix Specification compliant.\n"
			"The task %s (%d) triggered the difference, watch for misbehavior.\n",
			current->comm, task_pid_nr(current));

1028 1029
	if (sop->sem_num != semnum)
		return 0;
1030

1031 1032 1033 1034 1035 1036
	if (count_zero && sop->sem_op == 0)
		return 1;
	if (!count_zero && sop->sem_op < 0)
		return 1;

	return 0;
1037 1038
}

L
Linus Torvalds 已提交
1039 1040 1041
/* The following counts are associated to each semaphore:
 *   semncnt        number of tasks waiting on semval being nonzero
 *   semzcnt        number of tasks waiting on semval being zero
1042 1043 1044
 *
 * Per definition, a task waits only on the semaphore of the first semop
 * that cannot proceed, even if additional operation would block, too.
L
Linus Torvalds 已提交
1045
 */
1046 1047
static int count_semcnt(struct sem_array *sma, ushort semnum,
			bool count_zero)
L
Linus Torvalds 已提交
1048
{
1049
	struct list_head *l;
M
Manfred Spraul 已提交
1050
	struct sem_queue *q;
1051
	int semcnt;
L
Linus Torvalds 已提交
1052

1053 1054 1055 1056 1057 1058
	semcnt = 0;
	/* First: check the simple operations. They are easy to evaluate */
	if (count_zero)
		l = &sma->sem_base[semnum].pending_const;
	else
		l = &sma->sem_base[semnum].pending_alter;
L
Linus Torvalds 已提交
1059

1060 1061 1062 1063 1064
	list_for_each_entry(q, l, list) {
		/* all task on a per-semaphore list sleep on exactly
		 * that semaphore
		 */
		semcnt++;
R
Rik van Riel 已提交
1065 1066
	}

1067
	/* Then: check the complex operations. */
1068
	list_for_each_entry(q, &sma->pending_alter, list) {
1069 1070 1071 1072 1073 1074
		semcnt += check_qop(sma, semnum, q, count_zero);
	}
	if (count_zero) {
		list_for_each_entry(q, &sma->pending_const, list) {
			semcnt += check_qop(sma, semnum, q, count_zero);
		}
1075
	}
1076
	return semcnt;
L
Linus Torvalds 已提交
1077 1078
}

D
Davidlohr Bueso 已提交
1079 1080
/* Free a semaphore set. freeary() is called with sem_ids.rwsem locked
 * as a writer and the spinlock for this semaphore set hold. sem_ids.rwsem
N
Nadia Derbey 已提交
1081
 * remains locked on exit.
L
Linus Torvalds 已提交
1082
 */
1083
static void freeary(struct ipc_namespace *ns, struct kern_ipc_perm *ipcp)
L
Linus Torvalds 已提交
1084
{
1085 1086
	struct sem_undo *un, *tu;
	struct sem_queue *q, *tq;
1087
	struct sem_array *sma = container_of(ipcp, struct sem_array, sem_perm);
1088
	struct list_head tasks;
1089
	int i;
L
Linus Torvalds 已提交
1090

1091
	/* Free the existing undo structures for this semaphore set.  */
1092
	ipc_assert_locked_object(&sma->sem_perm);
1093 1094 1095
	list_for_each_entry_safe(un, tu, &sma->list_id, list_id) {
		list_del(&un->list_id);
		spin_lock(&un->ulp->lock);
L
Linus Torvalds 已提交
1096
		un->semid = -1;
1097 1098
		list_del_rcu(&un->list_proc);
		spin_unlock(&un->ulp->lock);
1099
		kfree_rcu(un, rcu);
1100
	}
L
Linus Torvalds 已提交
1101 1102

	/* Wake up all pending processes and let them fail with EIDRM. */
1103
	INIT_LIST_HEAD(&tasks);
1104 1105 1106 1107 1108 1109
	list_for_each_entry_safe(q, tq, &sma->pending_const, list) {
		unlink_queue(sma, q);
		wake_up_sem_queue_prepare(&tasks, q, -EIDRM);
	}

	list_for_each_entry_safe(q, tq, &sma->pending_alter, list) {
1110
		unlink_queue(sma, q);
1111
		wake_up_sem_queue_prepare(&tasks, q, -EIDRM);
L
Linus Torvalds 已提交
1112
	}
1113 1114
	for (i = 0; i < sma->sem_nsems; i++) {
		struct sem *sem = sma->sem_base + i;
1115 1116 1117 1118 1119
		list_for_each_entry_safe(q, tq, &sem->pending_const, list) {
			unlink_queue(sma, q);
			wake_up_sem_queue_prepare(&tasks, q, -EIDRM);
		}
		list_for_each_entry_safe(q, tq, &sem->pending_alter, list) {
1120 1121 1122 1123
			unlink_queue(sma, q);
			wake_up_sem_queue_prepare(&tasks, q, -EIDRM);
		}
	}
L
Linus Torvalds 已提交
1124

N
Nadia Derbey 已提交
1125 1126
	/* Remove the semaphore set from the IDR */
	sem_rmid(ns, sma);
1127
	sem_unlock(sma, -1);
1128
	rcu_read_unlock();
L
Linus Torvalds 已提交
1129

1130
	wake_up_sem_queue_do(&tasks);
K
Kirill Korotaev 已提交
1131
	ns->used_sems -= sma->sem_nsems;
D
Davidlohr Bueso 已提交
1132
	ipc_rcu_putref(sma, sem_rcu_free);
L
Linus Torvalds 已提交
1133 1134 1135 1136
}

static unsigned long copy_semid_to_user(void __user *buf, struct semid64_ds *in, int version)
{
M
Manfred Spraul 已提交
1137
	switch (version) {
L
Linus Torvalds 已提交
1138 1139 1140 1141 1142 1143
	case IPC_64:
		return copy_to_user(buf, in, sizeof(*in));
	case IPC_OLD:
	    {
		struct semid_ds out;

1144 1145
		memset(&out, 0, sizeof(out));

L
Linus Torvalds 已提交
1146 1147 1148 1149 1150 1151 1152 1153 1154 1155 1156 1157 1158
		ipc64_perm_to_ipc_perm(&in->sem_perm, &out.sem_perm);

		out.sem_otime	= in->sem_otime;
		out.sem_ctime	= in->sem_ctime;
		out.sem_nsems	= in->sem_nsems;

		return copy_to_user(buf, &out, sizeof(out));
	    }
	default:
		return -EINVAL;
	}
}

1159 1160 1161 1162 1163 1164 1165 1166 1167 1168 1169 1170 1171 1172 1173
static time_t get_semotime(struct sem_array *sma)
{
	int i;
	time_t res;

	res = sma->sem_base[0].sem_otime;
	for (i = 1; i < sma->sem_nsems; i++) {
		time_t to = sma->sem_base[i].sem_otime;

		if (to > res)
			res = to;
	}
	return res;
}

1174
static int semctl_nolock(struct ipc_namespace *ns, int semid,
1175
			 int cmd, int version, void __user *p)
L
Linus Torvalds 已提交
1176
{
1177
	int err;
L
Linus Torvalds 已提交
1178 1179
	struct sem_array *sma;

M
Manfred Spraul 已提交
1180
	switch (cmd) {
L
Linus Torvalds 已提交
1181 1182 1183 1184 1185 1186 1187 1188 1189
	case IPC_INFO:
	case SEM_INFO:
	{
		struct seminfo seminfo;
		int max_id;

		err = security_sem_semctl(NULL, cmd);
		if (err)
			return err;
P
Paul McQuade 已提交
1190

M
Manfred Spraul 已提交
1191
		memset(&seminfo, 0, sizeof(seminfo));
K
Kirill Korotaev 已提交
1192 1193 1194 1195
		seminfo.semmni = ns->sc_semmni;
		seminfo.semmns = ns->sc_semmns;
		seminfo.semmsl = ns->sc_semmsl;
		seminfo.semopm = ns->sc_semopm;
L
Linus Torvalds 已提交
1196 1197 1198 1199
		seminfo.semvmx = SEMVMX;
		seminfo.semmnu = SEMMNU;
		seminfo.semmap = SEMMAP;
		seminfo.semume = SEMUME;
D
Davidlohr Bueso 已提交
1200
		down_read(&sem_ids(ns).rwsem);
L
Linus Torvalds 已提交
1201
		if (cmd == SEM_INFO) {
K
Kirill Korotaev 已提交
1202 1203
			seminfo.semusz = sem_ids(ns).in_use;
			seminfo.semaem = ns->used_sems;
L
Linus Torvalds 已提交
1204 1205 1206 1207
		} else {
			seminfo.semusz = SEMUSZ;
			seminfo.semaem = SEMAEM;
		}
N
Nadia Derbey 已提交
1208
		max_id = ipc_get_maxid(&sem_ids(ns));
D
Davidlohr Bueso 已提交
1209
		up_read(&sem_ids(ns).rwsem);
P
Paul McQuade 已提交
1210
		if (copy_to_user(p, &seminfo, sizeof(struct seminfo)))
L
Linus Torvalds 已提交
1211
			return -EFAULT;
M
Manfred Spraul 已提交
1212
		return (max_id < 0) ? 0 : max_id;
L
Linus Torvalds 已提交
1213
	}
1214
	case IPC_STAT:
L
Linus Torvalds 已提交
1215 1216 1217
	case SEM_STAT:
	{
		struct semid64_ds tbuf;
1218 1219 1220
		int id = 0;

		memset(&tbuf, 0, sizeof(tbuf));
L
Linus Torvalds 已提交
1221

1222
		rcu_read_lock();
1223
		if (cmd == SEM_STAT) {
1224 1225 1226 1227 1228
			sma = sem_obtain_object(ns, semid);
			if (IS_ERR(sma)) {
				err = PTR_ERR(sma);
				goto out_unlock;
			}
1229 1230
			id = sma->sem_perm.id;
		} else {
1231 1232 1233 1234 1235
			sma = sem_obtain_object_check(ns, semid);
			if (IS_ERR(sma)) {
				err = PTR_ERR(sma);
				goto out_unlock;
			}
1236
		}
L
Linus Torvalds 已提交
1237 1238

		err = -EACCES;
1239
		if (ipcperms(ns, &sma->sem_perm, S_IRUGO))
L
Linus Torvalds 已提交
1240 1241 1242 1243 1244 1245 1246
			goto out_unlock;

		err = security_sem_semctl(sma, cmd);
		if (err)
			goto out_unlock;

		kernel_to_ipc64_perm(&sma->sem_perm, &tbuf.sem_perm);
1247 1248 1249
		tbuf.sem_otime = get_semotime(sma);
		tbuf.sem_ctime = sma->sem_ctime;
		tbuf.sem_nsems = sma->sem_nsems;
1250
		rcu_read_unlock();
1251
		if (copy_semid_to_user(p, &tbuf, version))
L
Linus Torvalds 已提交
1252 1253 1254 1255 1256 1257 1258
			return -EFAULT;
		return id;
	}
	default:
		return -EINVAL;
	}
out_unlock:
1259
	rcu_read_unlock();
L
Linus Torvalds 已提交
1260 1261 1262
	return err;
}

1263 1264 1265 1266 1267
static int semctl_setval(struct ipc_namespace *ns, int semid, int semnum,
		unsigned long arg)
{
	struct sem_undo *un;
	struct sem_array *sma;
M
Manfred Spraul 已提交
1268
	struct sem *curr;
1269 1270 1271 1272 1273 1274 1275 1276 1277 1278 1279
	int err;
	struct list_head tasks;
	int val;
#if defined(CONFIG_64BIT) && defined(__BIG_ENDIAN)
	/* big-endian 64bit */
	val = arg >> 32;
#else
	/* 32bit or little-endian 64bit */
	val = arg;
#endif

1280 1281
	if (val > SEMVMX || val < 0)
		return -ERANGE;
1282 1283 1284

	INIT_LIST_HEAD(&tasks);

1285 1286 1287 1288 1289 1290 1291 1292 1293 1294 1295 1296 1297 1298 1299 1300 1301
	rcu_read_lock();
	sma = sem_obtain_object_check(ns, semid);
	if (IS_ERR(sma)) {
		rcu_read_unlock();
		return PTR_ERR(sma);
	}

	if (semnum < 0 || semnum >= sma->sem_nsems) {
		rcu_read_unlock();
		return -EINVAL;
	}


	if (ipcperms(ns, &sma->sem_perm, S_IWUGO)) {
		rcu_read_unlock();
		return -EACCES;
	}
1302 1303

	err = security_sem_semctl(sma, SETVAL);
1304 1305 1306 1307
	if (err) {
		rcu_read_unlock();
		return -EACCES;
	}
1308

1309
	sem_lock(sma, NULL, -1);
1310

1311
	if (!ipc_valid_object(&sma->sem_perm)) {
1312 1313 1314 1315 1316
		sem_unlock(sma, -1);
		rcu_read_unlock();
		return -EIDRM;
	}

1317 1318
	curr = &sma->sem_base[semnum];

1319
	ipc_assert_locked_object(&sma->sem_perm);
1320 1321 1322 1323 1324 1325 1326 1327
	list_for_each_entry(un, &sma->list_id, list_id)
		un->semadj[semnum] = 0;

	curr->semval = val;
	curr->sempid = task_tgid_vnr(current);
	sma->sem_ctime = get_seconds();
	/* maybe some queued-up processes were waiting for this */
	do_smart_update(sma, NULL, 0, 0, &tasks);
1328
	sem_unlock(sma, -1);
1329
	rcu_read_unlock();
1330
	wake_up_sem_queue_do(&tasks);
1331
	return 0;
1332 1333
}

K
Kirill Korotaev 已提交
1334
static int semctl_main(struct ipc_namespace *ns, int semid, int semnum,
1335
		int cmd, void __user *p)
L
Linus Torvalds 已提交
1336 1337
{
	struct sem_array *sma;
M
Manfred Spraul 已提交
1338
	struct sem *curr;
1339
	int err, nsems;
L
Linus Torvalds 已提交
1340
	ushort fast_sem_io[SEMMSL_FAST];
M
Manfred Spraul 已提交
1341
	ushort *sem_io = fast_sem_io;
1342
	struct list_head tasks;
L
Linus Torvalds 已提交
1343

1344 1345 1346 1347 1348 1349
	INIT_LIST_HEAD(&tasks);

	rcu_read_lock();
	sma = sem_obtain_object_check(ns, semid);
	if (IS_ERR(sma)) {
		rcu_read_unlock();
1350
		return PTR_ERR(sma);
1351
	}
L
Linus Torvalds 已提交
1352 1353 1354 1355

	nsems = sma->sem_nsems;

	err = -EACCES;
1356 1357
	if (ipcperms(ns, &sma->sem_perm, cmd == SETALL ? S_IWUGO : S_IRUGO))
		goto out_rcu_wakeup;
L
Linus Torvalds 已提交
1358 1359

	err = security_sem_semctl(sma, cmd);
1360 1361
	if (err)
		goto out_rcu_wakeup;
L
Linus Torvalds 已提交
1362 1363 1364 1365 1366

	err = -EACCES;
	switch (cmd) {
	case GETALL:
	{
1367
		ushort __user *array = p;
L
Linus Torvalds 已提交
1368 1369
		int i;

1370
		sem_lock(sma, NULL, -1);
1371
		if (!ipc_valid_object(&sma->sem_perm)) {
1372 1373 1374
			err = -EIDRM;
			goto out_unlock;
		}
M
Manfred Spraul 已提交
1375
		if (nsems > SEMMSL_FAST) {
1376 1377
			if (!ipc_rcu_getref(sma)) {
				err = -EIDRM;
1378
				goto out_unlock;
1379 1380
			}
			sem_unlock(sma, -1);
1381
			rcu_read_unlock();
L
Linus Torvalds 已提交
1382
			sem_io = ipc_alloc(sizeof(ushort)*nsems);
M
Manfred Spraul 已提交
1383
			if (sem_io == NULL) {
D
Davidlohr Bueso 已提交
1384
				ipc_rcu_putref(sma, ipc_rcu_free);
L
Linus Torvalds 已提交
1385 1386 1387
				return -ENOMEM;
			}

1388
			rcu_read_lock();
1389
			sem_lock_and_putref(sma);
1390
			if (!ipc_valid_object(&sma->sem_perm)) {
L
Linus Torvalds 已提交
1391
				err = -EIDRM;
1392
				goto out_unlock;
L
Linus Torvalds 已提交
1393
			}
1394
		}
L
Linus Torvalds 已提交
1395 1396
		for (i = 0; i < sma->sem_nsems; i++)
			sem_io[i] = sma->sem_base[i].semval;
1397
		sem_unlock(sma, -1);
1398
		rcu_read_unlock();
L
Linus Torvalds 已提交
1399
		err = 0;
M
Manfred Spraul 已提交
1400
		if (copy_to_user(array, sem_io, nsems*sizeof(ushort)))
L
Linus Torvalds 已提交
1401 1402 1403 1404 1405 1406 1407 1408
			err = -EFAULT;
		goto out_free;
	}
	case SETALL:
	{
		int i;
		struct sem_undo *un;

1409
		if (!ipc_rcu_getref(sma)) {
1410 1411
			err = -EIDRM;
			goto out_rcu_wakeup;
1412
		}
1413
		rcu_read_unlock();
L
Linus Torvalds 已提交
1414

M
Manfred Spraul 已提交
1415
		if (nsems > SEMMSL_FAST) {
L
Linus Torvalds 已提交
1416
			sem_io = ipc_alloc(sizeof(ushort)*nsems);
M
Manfred Spraul 已提交
1417
			if (sem_io == NULL) {
D
Davidlohr Bueso 已提交
1418
				ipc_rcu_putref(sma, ipc_rcu_free);
L
Linus Torvalds 已提交
1419 1420 1421 1422
				return -ENOMEM;
			}
		}

M
Manfred Spraul 已提交
1423
		if (copy_from_user(sem_io, p, nsems*sizeof(ushort))) {
D
Davidlohr Bueso 已提交
1424
			ipc_rcu_putref(sma, ipc_rcu_free);
L
Linus Torvalds 已提交
1425 1426 1427 1428 1429 1430
			err = -EFAULT;
			goto out_free;
		}

		for (i = 0; i < nsems; i++) {
			if (sem_io[i] > SEMVMX) {
D
Davidlohr Bueso 已提交
1431
				ipc_rcu_putref(sma, ipc_rcu_free);
L
Linus Torvalds 已提交
1432 1433 1434 1435
				err = -ERANGE;
				goto out_free;
			}
		}
1436
		rcu_read_lock();
1437
		sem_lock_and_putref(sma);
1438
		if (!ipc_valid_object(&sma->sem_perm)) {
L
Linus Torvalds 已提交
1439
			err = -EIDRM;
1440
			goto out_unlock;
L
Linus Torvalds 已提交
1441 1442
		}

1443
		for (i = 0; i < nsems; i++) {
L
Linus Torvalds 已提交
1444
			sma->sem_base[i].semval = sem_io[i];
1445 1446
			sma->sem_base[i].sempid = task_tgid_vnr(current);
		}
1447

1448
		ipc_assert_locked_object(&sma->sem_perm);
1449
		list_for_each_entry(un, &sma->list_id, list_id) {
L
Linus Torvalds 已提交
1450 1451
			for (i = 0; i < nsems; i++)
				un->semadj[i] = 0;
1452
		}
L
Linus Torvalds 已提交
1453 1454
		sma->sem_ctime = get_seconds();
		/* maybe some queued-up processes were waiting for this */
1455
		do_smart_update(sma, NULL, 0, 0, &tasks);
L
Linus Torvalds 已提交
1456 1457 1458
		err = 0;
		goto out_unlock;
	}
1459
	/* GETVAL, GETPID, GETNCTN, GETZCNT: fall-through */
L
Linus Torvalds 已提交
1460 1461
	}
	err = -EINVAL;
1462 1463
	if (semnum < 0 || semnum >= nsems)
		goto out_rcu_wakeup;
L
Linus Torvalds 已提交
1464

1465
	sem_lock(sma, NULL, -1);
1466
	if (!ipc_valid_object(&sma->sem_perm)) {
1467 1468 1469
		err = -EIDRM;
		goto out_unlock;
	}
L
Linus Torvalds 已提交
1470 1471 1472 1473 1474 1475 1476 1477 1478 1479
	curr = &sma->sem_base[semnum];

	switch (cmd) {
	case GETVAL:
		err = curr->semval;
		goto out_unlock;
	case GETPID:
		err = curr->sempid;
		goto out_unlock;
	case GETNCNT:
1480
		err = count_semcnt(sma, semnum, 0);
L
Linus Torvalds 已提交
1481 1482
		goto out_unlock;
	case GETZCNT:
1483
		err = count_semcnt(sma, semnum, 1);
L
Linus Torvalds 已提交
1484 1485
		goto out_unlock;
	}
1486

L
Linus Torvalds 已提交
1487
out_unlock:
1488
	sem_unlock(sma, -1);
1489
out_rcu_wakeup:
1490
	rcu_read_unlock();
1491
	wake_up_sem_queue_do(&tasks);
L
Linus Torvalds 已提交
1492
out_free:
M
Manfred Spraul 已提交
1493
	if (sem_io != fast_sem_io)
1494
		ipc_free(sem_io);
L
Linus Torvalds 已提交
1495 1496 1497
	return err;
}

1498 1499
static inline unsigned long
copy_semid_from_user(struct semid64_ds *out, void __user *buf, int version)
L
Linus Torvalds 已提交
1500
{
M
Manfred Spraul 已提交
1501
	switch (version) {
L
Linus Torvalds 已提交
1502
	case IPC_64:
1503
		if (copy_from_user(out, buf, sizeof(*out)))
L
Linus Torvalds 已提交
1504 1505 1506 1507 1508 1509
			return -EFAULT;
		return 0;
	case IPC_OLD:
	    {
		struct semid_ds tbuf_old;

M
Manfred Spraul 已提交
1510
		if (copy_from_user(&tbuf_old, buf, sizeof(tbuf_old)))
L
Linus Torvalds 已提交
1511 1512
			return -EFAULT;

1513 1514 1515
		out->sem_perm.uid	= tbuf_old.sem_perm.uid;
		out->sem_perm.gid	= tbuf_old.sem_perm.gid;
		out->sem_perm.mode	= tbuf_old.sem_perm.mode;
L
Linus Torvalds 已提交
1516 1517 1518 1519 1520 1521 1522 1523

		return 0;
	    }
	default:
		return -EINVAL;
	}
}

1524
/*
D
Davidlohr Bueso 已提交
1525
 * This function handles some semctl commands which require the rwsem
1526
 * to be held in write mode.
D
Davidlohr Bueso 已提交
1527
 * NOTE: no locks must be held, the rwsem is taken inside this function.
1528
 */
1529
static int semctl_down(struct ipc_namespace *ns, int semid,
1530
		       int cmd, int version, void __user *p)
L
Linus Torvalds 已提交
1531 1532 1533
{
	struct sem_array *sma;
	int err;
1534
	struct semid64_ds semid64;
L
Linus Torvalds 已提交
1535 1536
	struct kern_ipc_perm *ipcp;

M
Manfred Spraul 已提交
1537
	if (cmd == IPC_SET) {
1538
		if (copy_semid_from_user(&semid64, p, version))
L
Linus Torvalds 已提交
1539 1540
			return -EFAULT;
	}
S
Steve Grubb 已提交
1541

D
Davidlohr Bueso 已提交
1542
	down_write(&sem_ids(ns).rwsem);
1543 1544
	rcu_read_lock();

1545 1546
	ipcp = ipcctl_pre_down_nolock(ns, &sem_ids(ns), semid, cmd,
				      &semid64.sem_perm, 0);
1547 1548 1549 1550
	if (IS_ERR(ipcp)) {
		err = PTR_ERR(ipcp);
		goto out_unlock1;
	}
S
Steve Grubb 已提交
1551

1552
	sma = container_of(ipcp, struct sem_array, sem_perm);
L
Linus Torvalds 已提交
1553 1554

	err = security_sem_semctl(sma, cmd);
1555 1556
	if (err)
		goto out_unlock1;
L
Linus Torvalds 已提交
1557

1558
	switch (cmd) {
L
Linus Torvalds 已提交
1559
	case IPC_RMID:
1560
		sem_lock(sma, NULL, -1);
1561
		/* freeary unlocks the ipc object and rcu */
1562
		freeary(ns, ipcp);
1563
		goto out_up;
L
Linus Torvalds 已提交
1564
	case IPC_SET:
1565
		sem_lock(sma, NULL, -1);
1566 1567
		err = ipc_update_perm(&semid64.sem_perm, ipcp);
		if (err)
1568
			goto out_unlock0;
L
Linus Torvalds 已提交
1569 1570 1571 1572
		sma->sem_ctime = get_seconds();
		break;
	default:
		err = -EINVAL;
1573
		goto out_unlock1;
L
Linus Torvalds 已提交
1574 1575
	}

1576
out_unlock0:
1577
	sem_unlock(sma, -1);
1578
out_unlock1:
1579
	rcu_read_unlock();
1580
out_up:
D
Davidlohr Bueso 已提交
1581
	up_write(&sem_ids(ns).rwsem);
L
Linus Torvalds 已提交
1582 1583 1584
	return err;
}

1585
SYSCALL_DEFINE4(semctl, int, semid, int, semnum, int, cmd, unsigned long, arg)
L
Linus Torvalds 已提交
1586 1587
{
	int version;
K
Kirill Korotaev 已提交
1588
	struct ipc_namespace *ns;
1589
	void __user *p = (void __user *)arg;
L
Linus Torvalds 已提交
1590 1591 1592 1593 1594

	if (semid < 0)
		return -EINVAL;

	version = ipc_parse_version(&cmd);
K
Kirill Korotaev 已提交
1595
	ns = current->nsproxy->ipc_ns;
L
Linus Torvalds 已提交
1596

M
Manfred Spraul 已提交
1597
	switch (cmd) {
L
Linus Torvalds 已提交
1598 1599
	case IPC_INFO:
	case SEM_INFO:
1600
	case IPC_STAT:
L
Linus Torvalds 已提交
1601
	case SEM_STAT:
1602
		return semctl_nolock(ns, semid, cmd, version, p);
L
Linus Torvalds 已提交
1603 1604 1605 1606 1607 1608
	case GETALL:
	case GETVAL:
	case GETPID:
	case GETNCNT:
	case GETZCNT:
	case SETALL:
1609 1610 1611
		return semctl_main(ns, semid, semnum, cmd, p);
	case SETVAL:
		return semctl_setval(ns, semid, semnum, arg);
L
Linus Torvalds 已提交
1612 1613
	case IPC_RMID:
	case IPC_SET:
1614
		return semctl_down(ns, semid, cmd, version, p);
L
Linus Torvalds 已提交
1615 1616 1617 1618 1619 1620 1621 1622 1623 1624 1625 1626 1627 1628 1629 1630 1631 1632 1633 1634 1635 1636
	default:
		return -EINVAL;
	}
}

/* If the task doesn't already have a undo_list, then allocate one
 * here.  We guarantee there is only one thread using this undo list,
 * and current is THE ONE
 *
 * If this allocation and assignment succeeds, but later
 * portions of this code fail, there is no need to free the sem_undo_list.
 * Just let it stay associated with the task, and it'll be freed later
 * at exit time.
 *
 * This can block, so callers must hold no locks.
 */
static inline int get_undo_list(struct sem_undo_list **undo_listp)
{
	struct sem_undo_list *undo_list;

	undo_list = current->sysvsem.undo_list;
	if (!undo_list) {
1637
		undo_list = kzalloc(sizeof(*undo_list), GFP_KERNEL);
L
Linus Torvalds 已提交
1638 1639
		if (undo_list == NULL)
			return -ENOMEM;
I
Ingo Molnar 已提交
1640
		spin_lock_init(&undo_list->lock);
L
Linus Torvalds 已提交
1641
		atomic_set(&undo_list->refcnt, 1);
1642 1643
		INIT_LIST_HEAD(&undo_list->list_proc);

L
Linus Torvalds 已提交
1644 1645 1646 1647 1648 1649
		current->sysvsem.undo_list = undo_list;
	}
	*undo_listp = undo_list;
	return 0;
}

1650
static struct sem_undo *__lookup_undo(struct sem_undo_list *ulp, int semid)
L
Linus Torvalds 已提交
1651
{
1652
	struct sem_undo *un;
1653

1654 1655 1656
	list_for_each_entry_rcu(un, &ulp->list_proc, list_proc) {
		if (un->semid == semid)
			return un;
L
Linus Torvalds 已提交
1657
	}
1658
	return NULL;
L
Linus Torvalds 已提交
1659 1660
}

1661 1662 1663 1664
static struct sem_undo *lookup_undo(struct sem_undo_list *ulp, int semid)
{
	struct sem_undo *un;

M
Manfred Spraul 已提交
1665
	assert_spin_locked(&ulp->lock);
1666 1667 1668 1669 1670 1671 1672 1673 1674

	un = __lookup_undo(ulp, semid);
	if (un) {
		list_del_rcu(&un->list_proc);
		list_add_rcu(&un->list_proc, &ulp->list_proc);
	}
	return un;
}

1675
/**
D
Davidlohr Bueso 已提交
1676
 * find_alloc_undo - lookup (and if not present create) undo array
1677 1678 1679 1680 1681 1682
 * @ns: namespace
 * @semid: semaphore array id
 *
 * The function looks up (and if not present creates) the undo structure.
 * The size of the undo structure depends on the size of the semaphore
 * array, thus the alloc path is not that straightforward.
1683 1684
 * Lifetime-rules: sem_undo is rcu-protected, on success, the function
 * performs a rcu_read_lock().
1685 1686
 */
static struct sem_undo *find_alloc_undo(struct ipc_namespace *ns, int semid)
L
Linus Torvalds 已提交
1687 1688 1689 1690
{
	struct sem_array *sma;
	struct sem_undo_list *ulp;
	struct sem_undo *un, *new;
1691
	int nsems, error;
L
Linus Torvalds 已提交
1692 1693 1694 1695 1696

	error = get_undo_list(&ulp);
	if (error)
		return ERR_PTR(error);

1697
	rcu_read_lock();
1698
	spin_lock(&ulp->lock);
L
Linus Torvalds 已提交
1699
	un = lookup_undo(ulp, semid);
1700
	spin_unlock(&ulp->lock);
M
Manfred Spraul 已提交
1701
	if (likely(un != NULL))
L
Linus Torvalds 已提交
1702 1703 1704
		goto out;

	/* no undo structure around - allocate one. */
1705
	/* step 1: figure out the size of the semaphore array */
1706 1707 1708
	sma = sem_obtain_object_check(ns, semid);
	if (IS_ERR(sma)) {
		rcu_read_unlock();
J
Julia Lawall 已提交
1709
		return ERR_CAST(sma);
1710
	}
1711

L
Linus Torvalds 已提交
1712
	nsems = sma->sem_nsems;
1713 1714 1715 1716 1717
	if (!ipc_rcu_getref(sma)) {
		rcu_read_unlock();
		un = ERR_PTR(-EIDRM);
		goto out;
	}
1718
	rcu_read_unlock();
L
Linus Torvalds 已提交
1719

1720
	/* step 2: allocate new undo structure */
1721
	new = kzalloc(sizeof(struct sem_undo) + sizeof(short)*nsems, GFP_KERNEL);
L
Linus Torvalds 已提交
1722
	if (!new) {
D
Davidlohr Bueso 已提交
1723
		ipc_rcu_putref(sma, ipc_rcu_free);
L
Linus Torvalds 已提交
1724 1725 1726
		return ERR_PTR(-ENOMEM);
	}

1727
	/* step 3: Acquire the lock on semaphore array */
1728
	rcu_read_lock();
1729
	sem_lock_and_putref(sma);
1730
	if (!ipc_valid_object(&sma->sem_perm)) {
1731
		sem_unlock(sma, -1);
1732
		rcu_read_unlock();
L
Linus Torvalds 已提交
1733 1734 1735 1736
		kfree(new);
		un = ERR_PTR(-EIDRM);
		goto out;
	}
1737 1738 1739 1740 1741 1742 1743 1744 1745 1746
	spin_lock(&ulp->lock);

	/*
	 * step 4: check for races: did someone else allocate the undo struct?
	 */
	un = lookup_undo(ulp, semid);
	if (un) {
		kfree(new);
		goto success;
	}
1747 1748
	/* step 5: initialize & link new undo structure */
	new->semadj = (short *) &new[1];
1749
	new->ulp = ulp;
1750 1751
	new->semid = semid;
	assert_spin_locked(&ulp->lock);
1752
	list_add_rcu(&new->list_proc, &ulp->list_proc);
1753
	ipc_assert_locked_object(&sma->sem_perm);
1754
	list_add(&new->list_id, &sma->list_id);
1755
	un = new;
1756

1757
success:
1758
	spin_unlock(&ulp->lock);
1759
	sem_unlock(sma, -1);
L
Linus Torvalds 已提交
1760 1761 1762 1763
out:
	return un;
}

1764 1765

/**
D
Davidlohr Bueso 已提交
1766
 * get_queue_result - retrieve the result code from sem_queue
1767 1768 1769 1770 1771 1772 1773 1774 1775 1776 1777 1778 1779 1780 1781 1782 1783 1784 1785 1786 1787 1788 1789
 * @q: Pointer to queue structure
 *
 * Retrieve the return code from the pending queue. If IN_WAKEUP is found in
 * q->status, then we must loop until the value is replaced with the final
 * value: This may happen if a task is woken up by an unrelated event (e.g.
 * signal) and in parallel the task is woken up by another task because it got
 * the requested semaphores.
 *
 * The function can be called with or without holding the semaphore spinlock.
 */
static int get_queue_result(struct sem_queue *q)
{
	int error;

	error = q->status;
	while (unlikely(error == IN_WAKEUP)) {
		cpu_relax();
		error = q->status;
	}

	return error;
}

1790 1791
SYSCALL_DEFINE4(semtimedop, int, semid, struct sembuf __user *, tsops,
		unsigned, nsops, const struct timespec __user *, timeout)
L
Linus Torvalds 已提交
1792 1793 1794 1795
{
	int error = -EINVAL;
	struct sem_array *sma;
	struct sembuf fast_sops[SEMOPM_FAST];
M
Manfred Spraul 已提交
1796
	struct sembuf *sops = fast_sops, *sop;
L
Linus Torvalds 已提交
1797
	struct sem_undo *un;
1798
	int undos = 0, alter = 0, max, locknum;
L
Linus Torvalds 已提交
1799 1800
	struct sem_queue queue;
	unsigned long jiffies_left = 0;
K
Kirill Korotaev 已提交
1801
	struct ipc_namespace *ns;
1802
	struct list_head tasks;
K
Kirill Korotaev 已提交
1803 1804

	ns = current->nsproxy->ipc_ns;
L
Linus Torvalds 已提交
1805 1806 1807

	if (nsops < 1 || semid < 0)
		return -EINVAL;
K
Kirill Korotaev 已提交
1808
	if (nsops > ns->sc_semopm)
L
Linus Torvalds 已提交
1809
		return -E2BIG;
M
Manfred Spraul 已提交
1810 1811 1812
	if (nsops > SEMOPM_FAST) {
		sops = kmalloc(sizeof(*sops)*nsops, GFP_KERNEL);
		if (sops == NULL)
L
Linus Torvalds 已提交
1813 1814
			return -ENOMEM;
	}
M
Manfred Spraul 已提交
1815 1816
	if (copy_from_user(sops, tsops, nsops * sizeof(*tsops))) {
		error =  -EFAULT;
L
Linus Torvalds 已提交
1817 1818 1819 1820 1821 1822 1823 1824 1825 1826 1827 1828 1829 1830 1831 1832 1833 1834 1835 1836
		goto out_free;
	}
	if (timeout) {
		struct timespec _timeout;
		if (copy_from_user(&_timeout, timeout, sizeof(*timeout))) {
			error = -EFAULT;
			goto out_free;
		}
		if (_timeout.tv_sec < 0 || _timeout.tv_nsec < 0 ||
			_timeout.tv_nsec >= 1000000000L) {
			error = -EINVAL;
			goto out_free;
		}
		jiffies_left = timespec_to_jiffies(&_timeout);
	}
	max = 0;
	for (sop = sops; sop < sops + nsops; sop++) {
		if (sop->sem_num >= max)
			max = sop->sem_num;
		if (sop->sem_flg & SEM_UNDO)
1837 1838
			undos = 1;
		if (sop->sem_op != 0)
L
Linus Torvalds 已提交
1839 1840 1841
			alter = 1;
	}

1842 1843
	INIT_LIST_HEAD(&tasks);

L
Linus Torvalds 已提交
1844
	if (undos) {
1845
		/* On success, find_alloc_undo takes the rcu_read_lock */
1846
		un = find_alloc_undo(ns, semid);
L
Linus Torvalds 已提交
1847 1848 1849 1850
		if (IS_ERR(un)) {
			error = PTR_ERR(un);
			goto out_free;
		}
1851
	} else {
L
Linus Torvalds 已提交
1852
		un = NULL;
1853 1854
		rcu_read_lock();
	}
L
Linus Torvalds 已提交
1855

1856
	sma = sem_obtain_object_check(ns, semid);
1857
	if (IS_ERR(sma)) {
1858
		rcu_read_unlock();
1859
		error = PTR_ERR(sma);
L
Linus Torvalds 已提交
1860
		goto out_free;
1861 1862
	}

1863
	error = -EFBIG;
1864 1865
	if (max >= sma->sem_nsems)
		goto out_rcu_wakeup;
1866 1867

	error = -EACCES;
1868 1869
	if (ipcperms(ns, &sma->sem_perm, alter ? S_IWUGO : S_IRUGO))
		goto out_rcu_wakeup;
1870 1871

	error = security_sem_semop(sma, sops, nsops, alter);
1872 1873
	if (error)
		goto out_rcu_wakeup;
1874

1875 1876
	error = -EIDRM;
	locknum = sem_lock(sma, sops, nsops);
1877 1878 1879 1880 1881 1882 1883 1884 1885
	/*
	 * We eventually might perform the following check in a lockless
	 * fashion, considering ipc_valid_object() locking constraints.
	 * If nsops == 1 and there is no contention for sem_perm.lock, then
	 * only a per-semaphore lock is held and it's OK to proceed with the
	 * check below. More details on the fine grained locking scheme
	 * entangled here and why it's RMID race safe on comments at sem_lock()
	 */
	if (!ipc_valid_object(&sma->sem_perm))
1886
		goto out_unlock_free;
L
Linus Torvalds 已提交
1887
	/*
1888
	 * semid identifiers are not unique - find_alloc_undo may have
L
Linus Torvalds 已提交
1889
	 * allocated an undo structure, it was invalidated by an RMID
1890
	 * and now a new array with received the same id. Check and fail.
L
Lucas De Marchi 已提交
1891
	 * This case can be detected checking un->semid. The existence of
1892
	 * "un" itself is guaranteed by rcu.
L
Linus Torvalds 已提交
1893
	 */
1894 1895
	if (un && un->semid == -1)
		goto out_unlock_free;
1896

1897 1898 1899 1900 1901 1902 1903
	queue.sops = sops;
	queue.nsops = nsops;
	queue.undo = un;
	queue.pid = task_tgid_vnr(current);
	queue.alter = alter;

	error = perform_atomic_semop(sma, &queue);
1904 1905 1906 1907 1908
	if (error == 0) {
		/* If the operation was successful, then do
		 * the required updates.
		 */
		if (alter)
1909
			do_smart_update(sma, sops, nsops, 1, &tasks);
1910 1911
		else
			set_semotime(sma, sops);
L
Linus Torvalds 已提交
1912
	}
1913 1914
	if (error <= 0)
		goto out_unlock_free;
L
Linus Torvalds 已提交
1915 1916 1917 1918

	/* We need to sleep on this operation, so we put the current
	 * task into the pending queue and go to sleep.
	 */
P
Paul McQuade 已提交
1919

1920 1921 1922 1923
	if (nsops == 1) {
		struct sem *curr;
		curr = &sma->sem_base[sops->sem_num];

1924 1925 1926 1927 1928 1929 1930 1931 1932 1933
		if (alter) {
			if (sma->complex_count) {
				list_add_tail(&queue.list,
						&sma->pending_alter);
			} else {

				list_add_tail(&queue.list,
						&curr->pending_alter);
			}
		} else {
1934
			list_add_tail(&queue.list, &curr->pending_const);
1935
		}
1936
	} else {
1937 1938 1939
		if (!sma->complex_count)
			merge_queues(sma);

1940
		if (alter)
1941
			list_add_tail(&queue.list, &sma->pending_alter);
1942
		else
1943 1944
			list_add_tail(&queue.list, &sma->pending_const);

1945 1946 1947
		sma->complex_count++;
	}

L
Linus Torvalds 已提交
1948 1949
	queue.status = -EINTR;
	queue.sleeper = current;
1950 1951

sleep_again:
1952
	__set_current_state(TASK_INTERRUPTIBLE);
1953
	sem_unlock(sma, locknum);
1954
	rcu_read_unlock();
L
Linus Torvalds 已提交
1955 1956 1957 1958 1959 1960

	if (timeout)
		jiffies_left = schedule_timeout(jiffies_left);
	else
		schedule();

1961
	error = get_queue_result(&queue);
L
Linus Torvalds 已提交
1962 1963 1964

	if (error != -EINTR) {
		/* fast path: update_queue already obtained all requested
1965 1966 1967 1968 1969 1970 1971 1972
		 * resources.
		 * Perform a smp_mb(): User space could assume that semop()
		 * is a memory barrier: Without the mb(), the cpu could
		 * speculatively read in user space stale data that was
		 * overwritten by the previous owner of the semaphore.
		 */
		smp_mb();

L
Linus Torvalds 已提交
1973 1974 1975
		goto out_free;
	}

1976
	rcu_read_lock();
1977
	sma = sem_obtain_lock(ns, semid, sops, nsops, &locknum);
1978 1979 1980 1981 1982 1983 1984 1985 1986

	/*
	 * Wait until it's guaranteed that no wakeup_sem_queue_do() is ongoing.
	 */
	error = get_queue_result(&queue);

	/*
	 * Array removed? If yes, leave without sem_unlock().
	 */
1987
	if (IS_ERR(sma)) {
1988
		rcu_read_unlock();
L
Linus Torvalds 已提交
1989 1990 1991
		goto out_free;
	}

1992

L
Linus Torvalds 已提交
1993
	/*
1994 1995
	 * If queue.status != -EINTR we are woken up by another process.
	 * Leave without unlink_queue(), but with sem_unlock().
L
Linus Torvalds 已提交
1996
	 */
1997
	if (error != -EINTR)
L
Linus Torvalds 已提交
1998 1999 2000 2001 2002 2003 2004
		goto out_unlock_free;

	/*
	 * If an interrupt occurred we have to clean up the queue
	 */
	if (timeout && jiffies_left == 0)
		error = -EAGAIN;
2005 2006 2007 2008 2009 2010 2011

	/*
	 * If the wakeup was spurious, just retry
	 */
	if (error == -EINTR && !signal_pending(current))
		goto sleep_again;

2012
	unlink_queue(sma, &queue);
L
Linus Torvalds 已提交
2013 2014

out_unlock_free:
2015
	sem_unlock(sma, locknum);
2016
out_rcu_wakeup:
2017
	rcu_read_unlock();
2018
	wake_up_sem_queue_do(&tasks);
L
Linus Torvalds 已提交
2019
out_free:
M
Manfred Spraul 已提交
2020
	if (sops != fast_sops)
L
Linus Torvalds 已提交
2021 2022 2023 2024
		kfree(sops);
	return error;
}

2025 2026
SYSCALL_DEFINE3(semop, int, semid, struct sembuf __user *, tsops,
		unsigned, nsops)
L
Linus Torvalds 已提交
2027 2028 2029 2030 2031 2032 2033 2034 2035 2036 2037 2038 2039 2040 2041 2042 2043 2044 2045
{
	return sys_semtimedop(semid, tsops, nsops, NULL);
}

/* If CLONE_SYSVSEM is set, establish sharing of SEM_UNDO state between
 * parent and child tasks.
 */

int copy_semundo(unsigned long clone_flags, struct task_struct *tsk)
{
	struct sem_undo_list *undo_list;
	int error;

	if (clone_flags & CLONE_SYSVSEM) {
		error = get_undo_list(&undo_list);
		if (error)
			return error;
		atomic_inc(&undo_list->refcnt);
		tsk->sysvsem.undo_list = undo_list;
P
Paul McQuade 已提交
2046
	} else
L
Linus Torvalds 已提交
2047 2048 2049 2050 2051 2052 2053 2054 2055 2056 2057 2058 2059 2060 2061 2062 2063 2064 2065
		tsk->sysvsem.undo_list = NULL;

	return 0;
}

/*
 * add semadj values to semaphores, free undo structures.
 * undo structures are not freed when semaphore arrays are destroyed
 * so some of them may be out of date.
 * IMPLEMENTATION NOTE: There is some confusion over whether the
 * set of adjustments that needs to be done should be done in an atomic
 * manner or not. That is, if we are attempting to decrement the semval
 * should we queue up and wait until we can do so legally?
 * The original implementation attempted to do this (queue and wait).
 * The current implementation does not do so. The POSIX standard
 * and SVID should be consulted to determine what behavior is mandated.
 */
void exit_sem(struct task_struct *tsk)
{
2066
	struct sem_undo_list *ulp;
L
Linus Torvalds 已提交
2067

2068 2069
	ulp = tsk->sysvsem.undo_list;
	if (!ulp)
L
Linus Torvalds 已提交
2070
		return;
2071
	tsk->sysvsem.undo_list = NULL;
L
Linus Torvalds 已提交
2072

2073
	if (!atomic_dec_and_test(&ulp->refcnt))
L
Linus Torvalds 已提交
2074 2075
		return;

2076
	for (;;) {
L
Linus Torvalds 已提交
2077
		struct sem_array *sma;
2078
		struct sem_undo *un;
2079
		struct list_head tasks;
2080
		int semid, i;
2081

2082
		rcu_read_lock();
2083 2084
		un = list_entry_rcu(ulp->list_proc.next,
				    struct sem_undo, list_proc);
2085 2086 2087 2088 2089 2090 2091 2092 2093 2094 2095 2096 2097 2098
		if (&un->list_proc == &ulp->list_proc) {
			/*
			 * We must wait for freeary() before freeing this ulp,
			 * in case we raced with last sem_undo. There is a small
			 * possibility where we exit while freeary() didn't
			 * finish unlocking sem_undo_list.
			 */
			spin_unlock_wait(&ulp->lock);
			rcu_read_unlock();
			break;
		}
		spin_lock(&ulp->lock);
		semid = un->semid;
		spin_unlock(&ulp->lock);
2099

2100
		/* exit_sem raced with IPC_RMID, nothing to do */
2101 2102
		if (semid == -1) {
			rcu_read_unlock();
2103
			continue;
2104
		}
L
Linus Torvalds 已提交
2105

2106
		sma = sem_obtain_object_check(tsk->nsproxy->ipc_ns, semid);
2107
		/* exit_sem raced with IPC_RMID, nothing to do */
2108 2109
		if (IS_ERR(sma)) {
			rcu_read_unlock();
2110
			continue;
2111
		}
L
Linus Torvalds 已提交
2112

2113
		sem_lock(sma, NULL, -1);
2114
		/* exit_sem raced with IPC_RMID, nothing to do */
2115
		if (!ipc_valid_object(&sma->sem_perm)) {
2116 2117 2118 2119
			sem_unlock(sma, -1);
			rcu_read_unlock();
			continue;
		}
2120
		un = __lookup_undo(ulp, semid);
2121 2122 2123 2124
		if (un == NULL) {
			/* exit_sem raced with IPC_RMID+semget() that created
			 * exactly the same semid. Nothing to do.
			 */
2125
			sem_unlock(sma, -1);
2126
			rcu_read_unlock();
2127 2128 2129 2130
			continue;
		}

		/* remove un from the linked lists */
2131
		ipc_assert_locked_object(&sma->sem_perm);
2132 2133
		list_del(&un->list_id);

2134 2135 2136 2137
		/* we are the last process using this ulp, acquiring ulp->lock
		 * isn't required. Besides that, we are also protected against
		 * IPC_RMID as we hold sma->sem_perm lock now
		 */
2138 2139
		list_del_rcu(&un->list_proc);

2140 2141
		/* perform adjustments registered in un */
		for (i = 0; i < sma->sem_nsems; i++) {
M
Manfred Spraul 已提交
2142
			struct sem *semaphore = &sma->sem_base[i];
2143 2144
			if (un->semadj[i]) {
				semaphore->semval += un->semadj[i];
L
Linus Torvalds 已提交
2145 2146 2147 2148 2149 2150 2151 2152 2153 2154 2155
				/*
				 * Range checks of the new semaphore value,
				 * not defined by sus:
				 * - Some unices ignore the undo entirely
				 *   (e.g. HP UX 11i 11.22, Tru64 V5.1)
				 * - some cap the value (e.g. FreeBSD caps
				 *   at 0, but doesn't enforce SEMVMX)
				 *
				 * Linux caps the semaphore value, both at 0
				 * and at SEMVMX.
				 *
M
Manfred Spraul 已提交
2156
				 *	Manfred <manfred@colorfullife.com>
L
Linus Torvalds 已提交
2157
				 */
I
Ingo Molnar 已提交
2158 2159 2160 2161
				if (semaphore->semval < 0)
					semaphore->semval = 0;
				if (semaphore->semval > SEMVMX)
					semaphore->semval = SEMVMX;
2162
				semaphore->sempid = task_tgid_vnr(current);
L
Linus Torvalds 已提交
2163 2164 2165
			}
		}
		/* maybe some queued-up processes were waiting for this */
2166 2167
		INIT_LIST_HEAD(&tasks);
		do_smart_update(sma, NULL, 0, 1, &tasks);
2168
		sem_unlock(sma, -1);
2169
		rcu_read_unlock();
2170
		wake_up_sem_queue_do(&tasks);
2171

2172
		kfree_rcu(un, rcu);
L
Linus Torvalds 已提交
2173
	}
2174
	kfree(ulp);
L
Linus Torvalds 已提交
2175 2176 2177
}

#ifdef CONFIG_PROC_FS
2178
static int sysvipc_sem_proc_show(struct seq_file *s, void *it)
L
Linus Torvalds 已提交
2179
{
2180
	struct user_namespace *user_ns = seq_user_ns(s);
2181
	struct sem_array *sma = it;
2182 2183
	time_t sem_otime;

2184 2185 2186 2187 2188 2189 2190 2191
	/*
	 * The proc interface isn't aware of sem_lock(), it calls
	 * ipc_lock_object() directly (in sysvipc_find_ipc).
	 * In order to stay compatible with sem_lock(), we must wait until
	 * all simple semop() calls have left their critical regions.
	 */
	sem_wait_array(sma);

2192
	sem_otime = get_semotime(sma);
2193

2194 2195 2196 2197 2198 2199 2200 2201 2202 2203 2204 2205 2206 2207
	seq_printf(s,
		   "%10d %10d  %4o %10u %5u %5u %5u %5u %10lu %10lu\n",
		   sma->sem_perm.key,
		   sma->sem_perm.id,
		   sma->sem_perm.mode,
		   sma->sem_nsems,
		   from_kuid_munged(user_ns, sma->sem_perm.uid),
		   from_kgid_munged(user_ns, sma->sem_perm.gid),
		   from_kuid_munged(user_ns, sma->sem_perm.cuid),
		   from_kgid_munged(user_ns, sma->sem_perm.cgid),
		   sem_otime,
		   sma->sem_ctime);

	return 0;
L
Linus Torvalds 已提交
2208 2209
}
#endif