proc.c 25.7 KB
Newer Older
1 2
/*-------------------------------------------------------------------------
 *
3
 * proc.c
4
 *	  routines to manage per-process shared memory data structure
5
 *
B
Add:  
Bruce Momjian 已提交
6 7
 * Portions Copyright (c) 1996-2000, PostgreSQL, Inc
 * Portions Copyright (c) 1994, Regents of the University of California
8 9 10
 *
 *
 * IDENTIFICATION
11
 *	  $Header: /cvsroot/pgsql/src/backend/storage/lmgr/proc.c,v 1.78 2000/08/27 19:00:28 petere Exp $
12 13 14 15
 *
 *-------------------------------------------------------------------------
 */
/*
16 17
 *	Each postgres backend gets one of these.  We'll use it to
 *	clean up after the process should the process suddenly die.
18 19 20
 *
 *
 * Interface (a):
21 22 23
 *		ProcSleep(), ProcWakeup(), ProcWakeupNext(),
 *		ProcQueueAlloc() -- create a shm queue for sleeping processes
 *		ProcQueueInit() -- create a queue without allocing memory
24 25 26 27 28 29 30 31 32 33
 *
 * Locking and waiting for buffers can cause the backend to be
 * put to sleep.  Whoever releases the lock, etc. wakes the
 * process up again (and gives it an error code so it knows
 * whether it was awoken on an error condition).
 *
 * Interface (b):
 *
 * ProcReleaseLocks -- frees the locks associated with this process,
 * ProcKill -- destroys the shared memory state (and locks)
34
 *		associated with the process.
35 36
 *
 * 5/15/91 -- removed the buffer pool based lock chain in favor
37 38 39 40 41 42
 *		of a shared memory lock chain.	The write-protection is
 *		more expensive if the lock chain is in the buffer pool.
 *		The only reason I kept the lock chain in the buffer pool
 *		in the first place was to allow the lock table to grow larger
 *		than available shared memory and that isn't going to work
 *		without a lot of unimplemented support anyway.
43 44
 *
 * 4/7/95 -- instead of allocating a set of 1 semaphore per process, we
45 46 47 48
 *		allocate a semaphore from a set of PROC_NSEMS_PER_SET semaphores
 *		shared among backends (we keep a few sets of semaphores around).
 *		This is so that we can support more backends. (system-wide semaphore
 *		sets run out pretty fast.)				  -ay 4/95
49
 *
50
 * $Header: /cvsroot/pgsql/src/backend/storage/lmgr/proc.c,v 1.78 2000/08/27 19:00:28 petere Exp $
51 52 53
 */
#include <sys/time.h>
#include <unistd.h>
54
#include <signal.h>
55
#include <sys/types.h>
M
Marc G. Fournier 已提交
56

57
#if defined(solaris_sparc) || defined(__CYGWIN__)
58 59 60 61
#include <sys/ipc.h>
#include <sys/sem.h>
#endif

M
Marc G. Fournier 已提交
62
#include "postgres.h"
63
#include "miscadmin.h"
64 65


66
/* In Ultrix and QNX, sem.h must be included after ipc.h */
67
#include <sys/sem.h>
B
Bruce Momjian 已提交
68

69 70
#include "storage/proc.h"

71
void		HandleDeadLock(int signum);
72
static void ProcFreeAllSemaphores(void);
73
static bool GetOffWaitqueue(PROC *);
74

75
int DeadlockTimeout = 1000;
M
 
Marc G. Fournier 已提交
76

77 78 79 80 81 82 83
/* --------------------
 * Spin lock for manipulating the shared process data structure:
 * ProcGlobal.... Adding an extra spin lock seemed like the smallest
 * hack to get around reading and updating this structure in shared
 * memory. -mer 17 July 1991
 * --------------------
 */
84
SPINLOCK	ProcStructLock;
85 86 87

static PROC_HDR *ProcGlobal = NULL;

88
PROC	   *MyProc = NULL;
89

90
static void ProcKill(int exitStatus, int pid);
91
static void ProcGetNewSemKeyAndNum(IPCKey *key, int *semNum);
92
static void ProcFreeSem(IpcSemaphoreKey semKey, int semNum);
93

V
Vadim B. Mikheev 已提交
94 95
static char *DeadLockMessage = "Deadlock detected -- See the lock(l) manual page for a possible cause.";

96 97
/*
 * InitProcGlobal -
98
 *	  initializes the global process table. We put it here so that
99
 *	  the postmaster can do this initialization. (ProcFreeAllSemaphores needs
100 101 102
 *	  to read this table on exiting the postmaster. If we have the first
 *	  backend do this, starting up and killing the postmaster without
 *	  starting any backends will be a problem.)
103 104 105 106 107 108 109 110 111 112 113
 *
 *	  We also allocate all the per-process semaphores we will need to support
 *	  the requested number of backends.  We used to allocate semaphores
 *	  only when backends were actually started up, but that is bad because
 *	  it lets Postgres fail under load --- a lot of Unix systems are
 *	  (mis)configured with small limits on the number of semaphores, and
 *	  running out when trying to start another backend is a common failure.
 *	  So, now we grab enough semaphores to support the desired max number
 *	  of backends immediately at initialization --- if the sysadmin has set
 *	  MaxBackends higher than his kernel will support, he'll find out sooner
 *	  rather than later.
114 115
 */
void
116
InitProcGlobal(IPCKey key, int maxBackends)
117
{
118
	bool		found = false;
119

120 121
	/* attach to the free list */
	ProcGlobal = (PROC_HDR *)
122
		ShmemInitStruct("Proc Header", sizeof(PROC_HDR), &found);
123

124 125
	/* --------------------
	 * We're the first - initialize.
126 127
	 * XXX if found should ever be true, it is a sign of impending doom ...
	 * ought to complain if so?
128 129 130
	 * --------------------
	 */
	if (!found)
131
	{
132
		int			i;
133

134 135 136 137
		ProcGlobal->freeProcs = INVALID_OFFSET;
		ProcGlobal->currKey = IPCGetProcessSemaphoreInitKey(key);
		for (i = 0; i < MAX_PROC_SEMS / PROC_NSEMS_PER_SET; i++)
			ProcGlobal->freeSemMap[i] = 0;
138

B
Bruce Momjian 已提交
139 140 141
		/*
		 * Arrange to delete semas on exit --- set this up now so that we
		 * will clean up if pre-allocation fails...
142 143 144
		 */
		on_shmem_exit(ProcFreeAllSemaphores, NULL);

B
Bruce Momjian 已提交
145 146
		/*
		 * Pre-create the semaphores for the first maxBackends processes,
147 148 149
		 * unless we are running as a standalone backend.
		 */
		if (key != PrivateIPCKey)
150
		{
151
			for (i = 0;
B
Bruce Momjian 已提交
152
				 i < (maxBackends + PROC_NSEMS_PER_SET - 1) / PROC_NSEMS_PER_SET;
153 154 155 156 157 158 159 160 161
				 i++)
			{
				IPCKey		semKey = ProcGlobal->currKey + i;
				int			semId;

				semId = IpcSemaphoreCreate(semKey,
										   PROC_NSEMS_PER_SET,
										   IPCProtection,
										   IpcSemaphoreDefaultStartValue,
162 163 164
										   0);
				if (semId < 0)
					elog(FATAL, "InitProcGlobal: IpcSemaphoreCreate failed");
165 166 167
				/* mark this sema set allocated */
				ProcGlobal->freeSemMap[i] = (1 << PROC_NSEMS_PER_SET);
			}
168
		}
169 170 171 172 173 174 175 176 177 178 179
	}
}

/* ------------------------
 * InitProc -- create a per-process data structure for this process
 * used by the lock manager on semaphore queues.
 * ------------------------
 */
void
InitProcess(IPCKey key)
{
180 181 182
	bool		found = false;
	unsigned long location,
				myOffset;
183 184 185 186 187

	SpinAcquire(ProcStructLock);

	/* attach to the free list */
	ProcGlobal = (PROC_HDR *)
188
		ShmemInitStruct("Proc Header", sizeof(PROC_HDR), &found);
189
	if (!found)
190
	{
191
		/* this should not happen. InitProcGlobal() is called before this. */
192
		elog(STOP, "InitProcess: Proc Header uninitialized");
193
	}
194 195

	if (MyProc != NULL)
196
	{
197
		SpinRelease(ProcStructLock);
198
		elog(ERROR, "ProcInit: you already exist");
199
		return;
200
	}
201 202 203 204 205 206

	/* try to get a proc from the free list first */

	myOffset = ProcGlobal->freeProcs;

	if (myOffset != INVALID_OFFSET)
207
	{
208 209 210 211 212 213 214
		MyProc = (PROC *) MAKE_PTR(myOffset);
		ProcGlobal->freeProcs = MyProc->links.next;
	}
	else
	{

		/*
215 216 217 218
		 * have to allocate one.  We can't use the normal shmem index
		 * table mechanism because the proc structure is stored by PID
		 * instead of by a global name (need to look it up by PID when we
		 * cleanup dead processes).
219 220
		 */

221
		MyProc = (PROC *) ShmemAlloc(sizeof(PROC));
222
		if (!MyProc)
223
		{
224 225
			SpinRelease(ProcStructLock);
			elog(FATAL, "cannot create new proc: out of memory");
226
		}
227 228 229

		/* this cannot be initialized until after the buffer pool */
		SHMQueueInit(&(MyProc->lockQueue));
230
	}
231

232
	/*
233 234 235
	 * zero out the spin lock counts and set the sLocks field for
	 * ProcStructLock to 1 as we have acquired this spinlock above but
	 * didn't record it since we didn't have MyProc until now.
236
	 */
B
Bruce Momjian 已提交
237
	MemSet(MyProc->sLocks, 0, sizeof(MyProc->sLocks));
238 239 240 241 242
	MyProc->sLocks[ProcStructLock] = 1;


	if (IsUnderPostmaster)
	{
243 244 245 246
		IPCKey		semKey;
		int			semNum;
		int			semId;
		union semun semun;
247 248 249

		ProcGetNewSemKeyAndNum(&semKey, &semNum);

B
Bruce Momjian 已提交
250 251 252 253 254
		/*
		 * Note: because of the pre-allocation done in InitProcGlobal,
		 * this call should always attach to an existing semaphore. It
		 * will (try to) create a new group of semaphores only if the
		 * postmaster tries to start more backends than it said it would.
255
		 */
256 257 258 259
		semId = IpcSemaphoreCreate(semKey,
								   PROC_NSEMS_PER_SET,
								   IPCProtection,
								   IpcSemaphoreDefaultStartValue,
260
								   0);
261 262 263 264 265 266 267 268 269 270 271 272 273 274 275 276 277 278 279 280 281 282

		/*
		 * we might be reusing a semaphore that belongs to a dead backend.
		 * So be careful and reinitialize its value here.
		 */
		semun.val = IpcSemaphoreDefaultStartValue;
		semctl(semId, semNum, SETVAL, semun);

		IpcSemaphoreLock(semId, semNum, IpcExclusiveLock);
		MyProc->sem.semId = semId;
		MyProc->sem.semNum = semNum;
		MyProc->sem.semKey = semKey;
	}
	else
		MyProc->sem.semId = -1;

	/* ----------------------
	 * Release the lock.
	 * ----------------------
	 */
	SpinRelease(ProcStructLock);

B
Bruce Momjian 已提交
283
	MyProc->pid = MyProcPid;
284
	MyProc->databaseId = MyDatabaseId;
285
	MyProc->xid = InvalidTransactionId;
286
	MyProc->xmin = InvalidTransactionId;
287 288 289 290 291 292

	/* ----------------
	 * Start keeping spin lock stats from here on.	Any botch before
	 * this initialization is forever botched
	 * ----------------
	 */
B
Bruce Momjian 已提交
293
	MemSet(MyProc->sLocks, 0, MAX_SPINS * sizeof(*MyProc->sLocks));
294 295

	/* -------------------------
296
	 * Install ourselves in the shmem index table.	The name to
297 298 299 300 301 302
	 * use is determined by the OS-assigned process id.  That
	 * allows the cleanup process to find us after any untimely
	 * exit.
	 * -------------------------
	 */
	location = MAKE_OFFSET(MyProc);
B
Bruce Momjian 已提交
303
	if ((!ShmemPIDLookup(MyProcPid, &location)) || (location != MAKE_OFFSET(MyProc)))
304
		elog(STOP, "InitProc: ShmemPID table broken");
305 306 307 308

	MyProc->errType = NO_ERROR;
	SHMQueueElemInit(&(MyProc->links));

309
	on_shmem_exit(ProcKill, (caddr_t) MyProcPid);
310 311
}

H
Hiroshi Inoue 已提交
312 313 314 315
/* -----------------------
 * get off the wait queue
 * -----------------------
 */
316
static bool
H
Hiroshi Inoue 已提交
317 318
GetOffWaitqueue(PROC *proc)
{
319 320
	bool		getoffed = false;

H
Hiroshi Inoue 已提交
321 322 323
	LockLockTable();
	if (proc->links.next != INVALID_OFFSET)
	{
324
		int			lockmode = proc->token;
325
		LOCK	*waitLock = proc->waitLock;
326

327 328
		Assert(waitLock);
		Assert(waitLock->waitProcs.size > 0);
H
Hiroshi Inoue 已提交
329
		SHMQueueDelete(&(proc->links));
330 331 332 333 334 335 336 337 338
		--waitLock->waitProcs.size;
		Assert(waitLock->nHolding > 0);
		Assert(waitLock->nHolding > proc->waitLock->nActive);
		--waitLock->nHolding;
		Assert(waitLock->holders[lockmode] > 0);
		--waitLock->holders[lockmode];
		if (waitLock->activeHolders[lockmode] == waitLock->holders[lockmode])
			waitLock->waitMask &= ~(1 << lockmode);
		ProcLockWakeup(&(waitLock->waitProcs), LOCK_LOCKMETHOD(*waitLock), waitLock);
339
		getoffed = true;
H
Hiroshi Inoue 已提交
340 341 342 343
	}
	SHMQueueElemInit(&(proc->links));
	UnlockLockTable();

344
	return getoffed;
H
Hiroshi Inoue 已提交
345
}
346

347 348 349 350 351 352 353
/*
 * ProcReleaseLocks() -- release all locks associated with this process
 *
 */
void
ProcReleaseLocks()
{
354 355 356
	if (!MyProc)
		return;
	LockReleaseAll(1, &MyProc->lockQueue);
H
Hiroshi Inoue 已提交
357
	GetOffWaitqueue(MyProc);
358 359 360 361
}

/*
 * ProcRemove -
362 363 364 365 366
 *	  used by the postmaster to clean up the global tables. This also frees
 *	  up the semaphore used for the lmgr of the process. (We have to do
 *	  this is the postmaster instead of doing a IpcSemaphoreKill on exiting
 *	  the process because the semaphore set is shared among backends and
 *	  we don't want to remove other's semaphores on exit.)
367 368 369 370
 */
bool
ProcRemove(int pid)
{
371 372
	SHMEM_OFFSET location;
	PROC	   *proc;
373 374 375 376 377

	location = INVALID_OFFSET;

	location = ShmemPIDDestroy(pid);
	if (location == INVALID_OFFSET)
378
		return FALSE;
379 380 381 382 383 384 385 386 387 388 389
	proc = (PROC *) MAKE_PTR(location);

	SpinAcquire(ProcStructLock);

	ProcFreeSem(proc->sem.semKey, proc->sem.semNum);

	proc->links.next = ProcGlobal->freeProcs;
	ProcGlobal->freeProcs = MAKE_OFFSET(proc);

	SpinRelease(ProcStructLock);

390
	return TRUE;
391 392 393 394
}

/*
 * ProcKill() -- Destroy the per-proc data structure for
395
 *		this process. Release any of its held spin locks.
396 397 398 399
 */
static void
ProcKill(int exitStatus, int pid)
{
400 401
	PROC	   *proc;
	SHMEM_OFFSET location;
402 403 404 405 406 407 408 409 410 411

	/* --------------------
	 * If this is a FATAL exit the postmaster will have to kill all the
	 * existing backends and reinitialize shared memory.  So all we don't
	 * need to do anything here.
	 * --------------------
	 */
	if (exitStatus != 0)
		return;

B
Bruce Momjian 已提交
412
	ShmemPIDLookup(MyProcPid, &location);
413 414 415 416 417
	if (location == INVALID_OFFSET)
		return;

	proc = (PROC *) MAKE_PTR(location);

418 419 420
	Assert(proc == MyProc || pid != MyProcPid);

	MyProc = NULL;
421 422 423 424 425 426

	/* ---------------
	 * Assume one lock table.
	 * ---------------
	 */
	ProcReleaseSpins(proc);
M
 
Marc G. Fournier 已提交
427
	LockReleaseAll(DEFAULT_LOCKMETHOD, &proc->lockQueue);
428

429
#ifdef USER_LOCKS
430

M
 
Marc G. Fournier 已提交
431 432 433 434
	/*
	 * Assume we have a second lock table.
	 */
	LockReleaseAll(USER_LOCKMETHOD, &proc->lockQueue);
435 436
#endif

437 438 439 440
	/* ----------------
	 * get off the wait queue
	 * ----------------
	 */
H
Hiroshi Inoue 已提交
441
	GetOffWaitqueue(proc);
442 443

	return;
444 445 446 447
}

/*
 * ProcQueue package: routines for putting processes to sleep
448
 *		and  waking them up
449 450 451 452 453 454 455 456
 */

/*
 * ProcQueueAlloc -- alloc/attach to a shared memory process queue
 *
 * Returns: a pointer to the queue or NULL
 * Side Effects: Initializes the queue if we allocated one
 */
457
#ifdef NOT_USED
458
PROC_QUEUE *
459 460
ProcQueueAlloc(char *name)
{
461 462
	bool		found;
	PROC_QUEUE *queue = (PROC_QUEUE *)
463
		ShmemInitStruct(name, sizeof(PROC_QUEUE), &found);
464 465

	if (!queue)
466
		return NULL;
467 468
	if (!found)
		ProcQueueInit(queue);
469
	return queue;
470
}
471

472
#endif
473 474 475 476 477

/*
 * ProcQueueInit -- initialize a shared memory process queue
 */
void
478
ProcQueueInit(PROC_QUEUE *queue)
479
{
480 481
	SHMQueueInit(&(queue->links));
	queue->size = 0;
482 483 484
}


485 486 487 488
/*
 *	Handling cancel request while waiting for lock
 *
 */
489 490 491
static bool lockWaiting = false;
void
SetWaitingForLock(bool waiting)
492
{
493 494
	if (waiting == lockWaiting)
		return;
495
	lockWaiting = waiting;
496 497
	if (lockWaiting)
	{
498 499 500 501 502 503
		/* The lock was already released ? */
		if (MyProc->links.next == INVALID_OFFSET)
		{
			lockWaiting = false;
			return;
		}
504
		if (QueryCancel)		/* cancel request pending */
505 506 507 508 509 510 511 512
		{
			if (GetOffWaitqueue(MyProc))
			{
				lockWaiting = false;
				elog(ERROR, "Query cancel requested while waiting lock");
			}
		}
	}
513
}
514 515
void
LockWaitCancel(void)
516
{
517 518
	struct itimerval timeval,
				dummy;
519

520 521
	if (!lockWaiting)
		return;
522 523 524 525 526 527 528
	lockWaiting = false;
	/* Deadlock timer off */
	MemSet(&timeval, 0, sizeof(struct itimerval));
	setitimer(ITIMER_REAL, &timeval, &dummy);
	if (GetOffWaitqueue(MyProc))
		elog(ERROR, "Query cancel requested while waiting lock");
}
529 530 531 532 533 534 535 536 537

/*
 * ProcSleep -- put a process to sleep
 *
 * P() on the semaphore should put us to sleep.  The process
 * semaphore is cleared by default, so the first time we try
 * to acquire it, we sleep.
 *
 * ASSUME: that no one will fiddle with the queue until after
538
 *		we release the spin lock.
539 540 541 542
 *
 * NOTES: The process queue is now a priority queue for locking.
 */
int
543
ProcSleep(PROC_QUEUE *waitQueue,/* lock->waitProcs */
544
		  LOCKMETHODCTL *lockctl,
545
		  int token,			/* lockmode */
V
Vadim B. Mikheev 已提交
546
		  LOCK *lock)
547
{
548
	int			i;
V
Vadim B. Mikheev 已提交
549
	SPINLOCK	spinlock = lockctl->masterLock;
550
	PROC	   *proc;
V
Vadim B. Mikheev 已提交
551 552 553 554 555
	int			myMask = (1 << token);
	int			waitMask = lock->waitMask;
	int			aheadHolders[MAX_LOCKMODES];
	bool		selfConflict = (lockctl->conflictTab[token] & myMask),
				prevSame = false;
B
Bruce Momjian 已提交
556 557 558
	bool		deadlock_checked = false;
	struct itimerval timeval,
				dummy;
559

V
Vadim B. Mikheev 已提交
560 561 562
	MyProc->token = token;
	MyProc->waitLock = lock;

B
Bruce Momjian 已提交
563
	proc = (PROC *) MAKE_PTR(waitQueue->links.prev);
564

V
Vadim B. Mikheev 已提交
565 566 567
	/* if we don't conflict with any waiter - be first in queue */
	if (!(lockctl->conflictTab[token] & waitMask))
		goto ins;
568

V
Vadim B. Mikheev 已提交
569 570 571
	for (i = 1; i < MAX_LOCKMODES; i++)
		aheadHolders[i] = lock->activeHolders[i];
	(aheadHolders[token])++;
572

V
Vadim B. Mikheev 已提交
573 574 575 576 577 578 579 580 581 582 583 584 585 586 587 588 589 590 591 592
	for (i = 0; i < waitQueue->size; i++)
	{
		/* am I waiting for him ? */
		if (lockctl->conflictTab[token] & proc->holdLock)
		{
			/* is he waiting for me ? */
			if (lockctl->conflictTab[proc->token] & MyProc->holdLock)
			{
				MyProc->errType = STATUS_ERROR;
				elog(NOTICE, DeadLockMessage);
				goto rt;
			}
			/* being waiting for him - go past */
		}
		/* if he waits for me */
		else if (lockctl->conflictTab[proc->token] & MyProc->holdLock)
			break;
		/* if conflicting locks requested */
		else if (lockctl->conflictTab[proc->token] & myMask)
		{
B
Bruce Momjian 已提交
593

V
Vadim B. Mikheev 已提交
594
			/*
B
Bruce Momjian 已提交
595 596
			 * If I request non self-conflicting lock and there are others
			 * requesting the same lock just before me - stay here.
V
Vadim B. Mikheev 已提交
597 598 599 600
			 */
			if (!selfConflict && prevSame)
				break;
		}
B
Bruce Momjian 已提交
601

V
Vadim B. Mikheev 已提交
602
		/*
B
Bruce Momjian 已提交
603 604
		 * Last attempt to don't move any more: if we don't conflict with
		 * rest waiters in queue.
V
Vadim B. Mikheev 已提交
605 606 607
		 */
		else if (!(lockctl->conflictTab[token] & waitMask))
			break;
608

V
Vadim B. Mikheev 已提交
609 610 611
		prevSame = (proc->token == token);
		(aheadHolders[proc->token])++;
		if (aheadHolders[proc->token] == lock->holders[proc->token])
B
Bruce Momjian 已提交
612
			waitMask &= ~(1 << proc->token);
V
Vadim B. Mikheev 已提交
613 614
		proc = (PROC *) MAKE_PTR(proc->links.prev);
	}
615

V
Vadim B. Mikheev 已提交
616
ins:;
617 618 619 620 621 622
	/* -------------------
	 * assume that these two operations are atomic (because
	 * of the spinlock).
	 * -------------------
	 */
	SHMQueueInsertTL(&(proc->links), &(MyProc->links));
B
Bruce Momjian 已提交
623
	waitQueue->size++;
624

V
Vadim B. Mikheev 已提交
625
	lock->waitMask |= myMask;
626 627 628
	SpinRelease(spinlock);

	/* --------------
B
Bruce Momjian 已提交
629
	 * We set this so we can wake up periodically and check for a deadlock.
B
Bruce Momjian 已提交
630 631
	 * If a deadlock is detected, the handler releases the processes
	 * semaphore and aborts the current transaction.
B
Bruce Momjian 已提交
632 633 634
	 *
	 * Need to zero out struct to set the interval and the micro seconds fields
	 * to 0.
635 636
	 * --------------
	 */
B
Bruce Momjian 已提交
637
	MemSet(&timeval, 0, sizeof(struct itimerval));
638 639
	timeval.it_value.tv_sec = DeadlockTimeout / 1000;
	timeval.it_value.tv_usec = (DeadlockTimeout % 1000) * 1000;
640

641
	SetWaitingForLock(true);
B
Bruce Momjian 已提交
642 643
	do
	{
644
		MyProc->errType = NO_ERROR;		/* reset flag after deadlock check */
645

B
Bruce Momjian 已提交
646 647 648 649 650
		if (!deadlock_checked)
			if (setitimer(ITIMER_REAL, &timeval, &dummy))
				elog(FATAL, "ProcSleep: Unable to set timer for process wakeup");
		deadlock_checked = true;

B
Bruce Momjian 已提交
651 652 653 654 655 656
		/* --------------
		 * if someone wakes us between SpinRelease and IpcSemaphoreLock,
		 * IpcSemaphoreLock will not block.  The wakeup is "saved" by
		 * the semaphore implementation.
		 * --------------
		 */
M
 
Marc G. Fournier 已提交
657 658
		IpcSemaphoreLock(MyProc->sem.semId, MyProc->sem.semNum,
						 IpcExclusiveLock);
659 660
	} while (MyProc->errType == STATUS_NOT_FOUND);		/* sleep after deadlock
														 * check */
661
	lockWaiting = false;
662

B
Bruce Momjian 已提交
663 664 665 666 667
	/* ---------------
	 * We were awoken before a timeout - now disable the timer
	 * ---------------
	 */
	timeval.it_value.tv_sec = 0;
668
	timeval.it_value.tv_usec = 0;
B
Bruce Momjian 已提交
669 670 671
	if (setitimer(ITIMER_REAL, &timeval, &dummy))
		elog(FATAL, "ProcSleep: Unable to diable timer for process wakeup");

672 673 674 675 676 677 678
	/* ----------------
	 * We were assumed to be in a critical section when we went
	 * to sleep.
	 * ----------------
	 */
	SpinAcquire(spinlock);

V
Vadim B. Mikheev 已提交
679 680
rt:;

681
#ifdef LOCK_DEBUG
M
 
Marc G. Fournier 已提交
682
	/* Just to get meaningful debug messages from DumpLocks() */
683
	MyProc->waitLock = (LOCK *) NULL;
M
 
Marc G. Fournier 已提交
684 685
#endif

686
	return MyProc->errType;
687 688 689 690 691 692
}


/*
 * ProcWakeup -- wake up a process by releasing its private semaphore.
 *
693 694
 *	 remove the process from the wait queue and set its links invalid.
 *	 RETURN: the next process in the wait queue.
695
 */
B
Bruce Momjian 已提交
696
PROC *
697
ProcWakeup(PROC *proc, int errType)
698
{
699
	PROC	   *retProc;
700 701 702 703 704

	/* assume that spinlock has been acquired */

	if (proc->links.prev == INVALID_OFFSET ||
		proc->links.next == INVALID_OFFSET)
705
		return (PROC *) NULL;
706 707 708 709 710 711 712 713 714 715 716 717

	retProc = (PROC *) MAKE_PTR(proc->links.prev);

	/* you have to update waitLock->waitProcs.size yourself */
	SHMQueueDelete(&(proc->links));
	SHMQueueElemInit(&(proc->links));

	proc->errType = errType;

	IpcSemaphoreUnlock(proc->sem.semId, proc->sem.semNum, IpcExclusiveLock);

	return retProc;
718 719 720 721
}

/*
 * ProcLockWakeup -- routine for waking up processes when a lock is
722
 *		released.
723 724
 */
int
725
ProcLockWakeup(PROC_QUEUE *queue, LOCKMETHOD lockmethod, LOCK *lock)
726
{
727
	PROC	   *proc;
V
Vadim B. Mikheev 已提交
728 729
	int			count = 0;
	int			last_locktype = 0;
M
 
Marc G. Fournier 已提交
730 731 732
	int			queue_size = queue->size;

	Assert(queue->size >= 0);
733 734

	if (!queue->size)
735
		return STATUS_NOT_FOUND;
736 737

	proc = (PROC *) MAKE_PTR(queue->links.prev);
M
 
Marc G. Fournier 已提交
738 739
	while ((queue_size--) && (proc))
	{
740

M
 
Marc G. Fournier 已提交
741
		/*
742 743
		 * This proc will conflict as the previous one did, don't even
		 * try.
M
 
Marc G. Fournier 已提交
744 745 746 747 748
		 */
		if (proc->token == last_locktype)
			continue;

		/*
V
Vadim B. Mikheev 已提交
749
		 * Does this proc conflict with locks held by others ?
M
 
Marc G. Fournier 已提交
750 751
		 */
		if (LockResolveConflicts(lockmethod,
752
								 lock,
753
								 proc->token,
M
 
Marc G. Fournier 已提交
754 755 756
								 proc->xid,
								 (XIDLookupEnt *) NULL) != STATUS_OK)
		{
V
Vadim B. Mikheev 已提交
757 758
			if (count != 0)
				break;
M
 
Marc G. Fournier 已提交
759 760 761
			last_locktype = proc->token;
			continue;
		}
762 763 764 765 766 767 768

		/*
		 * there was a waiting process, grant it the lock before waking it
		 * up.	This will prevent another process from seizing the lock
		 * between the time we release the lock master (spinlock) and the
		 * time that the awoken process begins executing again.
		 */
769
		GrantLock(lock, proc->token);
770 771 772

		/*
		 * ProcWakeup removes proc from the lock waiting process queue and
773
		 * returns the next proc in chain.
774 775 776
		 */

		count++;
M
 
Marc G. Fournier 已提交
777 778
		queue->size--;
		proc = ProcWakeup(proc, NO_ERROR);
779
	}
780

M
 
Marc G. Fournier 已提交
781 782
	Assert(queue->size >= 0);

783
	if (count)
784
		return STATUS_OK;
785 786
	else
	{
787
		/* Something is still blocking us.	May have deadlocked. */
788 789 790 791 792
#ifdef LOCK_DEBUG
		if (lock->tag.lockmethod == USER_LOCKMETHOD ? Trace_userlocks : Trace_locks)
		{
			elog(DEBUG, "ProcLockWakeup: lock(%lx) can't wake up any process", MAKE_OFFSET(lock));
			if (Debug_deadlocks)
M
 
Marc G. Fournier 已提交
793
			DumpAllLocks();
794
		}
M
 
Marc G. Fournier 已提交
795
#endif
796
		return STATUS_NOT_FOUND;
M
 
Marc G. Fournier 已提交
797
	}
798 799 800
}

void
801
ProcAddLock(SHM_QUEUE *elem)
802
{
803
	SHMQueueInsertTL(&MyProc->lockQueue, elem);
804 805 806
}

/* --------------------
807
 * We only get to this routine if we got SIGALRM after DeadlockTimeout
B
Bruce Momjian 已提交
808 809
 * while waiting for a lock to be released by some other process.  If we have
 * a real deadlock, we must also indicate that I'm no longer waiting
810
 * on a lock so that other processes don't try to wake me up and screw
811 812 813
 * up my semaphore.
 * --------------------
 */
814
void
815
HandleDeadLock(int signum)
816
{
B
Bruce Momjian 已提交
817
	LOCK	   *mywaitlock;
818 819 820 821 822 823 824 825 826 827 828 829 830 831 832 833 834 835 836 837 838 839 840 841 842 843 844 845 846 847 848 849 850 851 852 853 854 855

	LockLockTable();

	/* ---------------------
	 * Check to see if we've been awoken by anyone in the interim.
	 *
	 * If we have we can return and resume our transaction -- happy day.
	 * Before we are awoken the process releasing the lock grants it to
	 * us so we know that we don't have to wait anymore.
	 *
	 * Damn these names are LONG! -mer
	 * ---------------------
	 */
	if (IpcSemaphoreGetCount(MyProc->sem.semId, MyProc->sem.semNum) ==
		IpcSemaphoreDefaultStartValue)
	{
		UnlockLockTable();
		return;
	}

	/*
	 * you would think this would be unnecessary, but...
	 *
	 * this also means we've been removed already.  in some ports (e.g.,
	 * sparc and aix) the semop(2) implementation is such that we can
	 * actually end up in this handler after someone has removed us from
	 * the queue and bopped the semaphore *but the test above fails to
	 * detect the semaphore update* (presumably something weird having to
	 * do with the order in which the semaphore wakeup signal and SIGALRM
	 * get handled).
	 */
	if (MyProc->links.prev == INVALID_OFFSET ||
		MyProc->links.next == INVALID_OFFSET)
	{
		UnlockLockTable();
		return;
	}

856 857 858
#ifdef LOCK_DEBUG
    if (Debug_deadlocks)
        DumpAllLocks();
859 860
#endif

B
Bruce Momjian 已提交
861 862
	MyProc->errType = STATUS_NOT_FOUND;
	if (!DeadLockCheck(MyProc, MyProc->waitLock))
B
Bruce Momjian 已提交
863 864 865 866 867 868 869
	{
		UnlockLockTable();
		return;
	}

	mywaitlock = MyProc->waitLock;

870 871 872 873
	/* ------------------------
	 * Get this process off the lock's wait queue
	 * ------------------------
	 */
B
Bruce Momjian 已提交
874
	Assert(mywaitlock->waitProcs.size > 0);
875
	lockWaiting = false;
B
Bruce Momjian 已提交
876
	--mywaitlock->waitProcs.size;
877 878 879 880 881 882 883 884
	SHMQueueDelete(&(MyProc->links));
	SHMQueueElemInit(&(MyProc->links));

	/* ------------------
	 * Unlock my semaphore so that the count is right for next time.
	 * I was awoken by a signal, not by someone unlocking my semaphore.
	 * ------------------
	 */
M
 
Marc G. Fournier 已提交
885 886
	IpcSemaphoreUnlock(MyProc->sem.semId, MyProc->sem.semNum,
					   IpcExclusiveLock);
887 888 889 890 891 892 893 894 895 896 897 898 899 900 901

	/* -------------
	 * Set MyProc->errType to STATUS_ERROR so that we abort after
	 * returning from this handler.
	 * -------------
	 */
	MyProc->errType = STATUS_ERROR;

	/*
	 * if this doesn't follow the IpcSemaphoreUnlock then we get lock
	 * table corruption ("LockReplace: xid table corrupted") due to race
	 * conditions.	i don't claim to understand this...
	 */
	UnlockLockTable();

V
Vadim B. Mikheev 已提交
902
	elog(NOTICE, DeadLockMessage);
903
	return;
904 905 906
}

void
907
ProcReleaseSpins(PROC *proc)
908
{
909
	int			i;
910 911 912 913 914 915 916

	if (!proc)
		proc = MyProc;

	if (!proc)
		return;
	for (i = 0; i < (int) MAX_SPINS; i++)
917
	{
918
		if (proc->sLocks[i])
919
		{
920 921
			Assert(proc->sLocks[i] == 1);
			SpinRelease(i);
922 923
		}
	}
H
 
Hiroshi Inoue 已提交
924
	AbortBufferIO();
925 926 927
}

/*****************************************************************************
928
 *
929 930 931 932
 *****************************************************************************/

/*
 * ProcGetNewSemKeyAndNum -
933 934 935 936
 *	  scan the free semaphore bitmap and allocate a single semaphore from
 *	  a semaphore set. (If the semaphore set doesn't exist yet,
 *	  IpcSemaphoreCreate will create it. Otherwise, we use the existing
 *	  semaphore set.)
937 938
 */
static void
939
ProcGetNewSemKeyAndNum(IPCKey *key, int *semNum)
940
{
941 942
	int			i;
	int32	   *freeSemMap = ProcGlobal->freeSemMap;
B
Bruce Momjian 已提交
943
	int32		fullmask = (1 << (PROC_NSEMS_PER_SET + 1)) - 1;
944

945 946 947 948
	/*
	 * we hold ProcStructLock when entering this routine. We scan through
	 * the bitmap to look for a free semaphore.
	 */
949

950 951
	for (i = 0; i < MAX_PROC_SEMS / PROC_NSEMS_PER_SET; i++)
	{
952 953
		int			mask = 1;
		int			j;
954 955

		if (freeSemMap[i] == fullmask)
956
			continue;			/* this set is fully allocated */
957 958 959 960 961 962 963

		for (j = 0; j < PROC_NSEMS_PER_SET; j++)
		{
			if ((freeSemMap[i] & mask) == 0)
			{

				/*
B
Bruce Momjian 已提交
964 965
				 * a free semaphore found. Mark it as allocated. Also set
				 * the bit indicating whole set is allocated.
966
				 */
967
				freeSemMap[i] |= mask + (1 << PROC_NSEMS_PER_SET);
968 969 970 971 972 973 974

				*key = ProcGlobal->currKey + i;
				*semNum = j;
				return;
			}
			mask <<= 1;
		}
975 976
	}

977
	/* if we reach here, all the semaphores are in use. */
978
	elog(ERROR, "InitProc: cannot allocate a free semaphore");
979 980 981 982
}

/*
 * ProcFreeSem -
983
 *	  free up our semaphore in the semaphore set.
984 985 986 987
 */
static void
ProcFreeSem(IpcSemaphoreKey semKey, int semNum)
{
988 989 990
	int			mask;
	int			i;
	int32	   *freeSemMap = ProcGlobal->freeSemMap;
991

992 993 994
	i = semKey - ProcGlobal->currKey;
	mask = ~(1 << semNum);
	freeSemMap[i] &= mask;
995

B
Bruce Momjian 已提交
996 997 998 999
	/*
	 * Formerly we'd release a semaphore set if it was now completely
	 * unused, but now we keep the semaphores to ensure we won't run out
	 * when starting new backends --- cf. InitProcGlobal.  Note that the
1000 1001 1002
	 * PROC_NSEMS_PER_SET+1'st bit of the freeSemMap entry remains set to
	 * indicate it is still allocated; ProcFreeAllSemaphores() needs that.
	 */
1003 1004 1005 1006
}

/*
 * ProcFreeAllSemaphores -
1007 1008 1009
 *	  called at shmem_exit time, ie when exiting the postmaster or
 *	  destroying shared state for a failed set of backends.
 *	  Free up all the semaphores allocated to the lmgrs of the backends.
1010
 */
1011
static void
1012 1013
ProcFreeAllSemaphores()
{
1014 1015
	int			i;
	int32	   *freeSemMap = ProcGlobal->freeSemMap;
1016

1017 1018 1019 1020 1021
	for (i = 0; i < MAX_PROC_SEMS / PROC_NSEMS_PER_SET; i++)
	{
		if (freeSemMap[i] != 0)
			IpcSemaphoreKill(ProcGlobal->currKey + i);
	}
1022
}