proc.c 25.7 KB
Newer Older
1 2
/*-------------------------------------------------------------------------
 *
3
 * proc.c
4
 *	  routines to manage per-process shared memory data structure
5
 *
B
Add:  
Bruce Momjian 已提交
6 7
 * Portions Copyright (c) 1996-2000, PostgreSQL, Inc
 * Portions Copyright (c) 1994, Regents of the University of California
8 9 10
 *
 *
 * IDENTIFICATION
11
 *	  $Header: /cvsroot/pgsql/src/backend/storage/lmgr/proc.c,v 1.70 2000/02/24 04:36:01 inoue Exp $
12 13 14 15
 *
 *-------------------------------------------------------------------------
 */
/*
16 17
 *	Each postgres backend gets one of these.  We'll use it to
 *	clean up after the process should the process suddenly die.
18 19 20
 *
 *
 * Interface (a):
21 22 23
 *		ProcSleep(), ProcWakeup(), ProcWakeupNext(),
 *		ProcQueueAlloc() -- create a shm queue for sleeping processes
 *		ProcQueueInit() -- create a queue without allocing memory
24 25 26 27 28 29 30 31 32 33
 *
 * Locking and waiting for buffers can cause the backend to be
 * put to sleep.  Whoever releases the lock, etc. wakes the
 * process up again (and gives it an error code so it knows
 * whether it was awoken on an error condition).
 *
 * Interface (b):
 *
 * ProcReleaseLocks -- frees the locks associated with this process,
 * ProcKill -- destroys the shared memory state (and locks)
34
 *		associated with the process.
35 36
 *
 * 5/15/91 -- removed the buffer pool based lock chain in favor
37 38 39 40 41 42
 *		of a shared memory lock chain.	The write-protection is
 *		more expensive if the lock chain is in the buffer pool.
 *		The only reason I kept the lock chain in the buffer pool
 *		in the first place was to allow the lock table to grow larger
 *		than available shared memory and that isn't going to work
 *		without a lot of unimplemented support anyway.
43 44
 *
 * 4/7/95 -- instead of allocating a set of 1 semaphore per process, we
45 46 47 48
 *		allocate a semaphore from a set of PROC_NSEMS_PER_SET semaphores
 *		shared among backends (we keep a few sets of semaphores around).
 *		This is so that we can support more backends. (system-wide semaphore
 *		sets run out pretty fast.)				  -ay 4/95
49
 *
50
 * $Header: /cvsroot/pgsql/src/backend/storage/lmgr/proc.c,v 1.70 2000/02/24 04:36:01 inoue Exp $
51 52 53
 */
#include <sys/time.h>
#include <unistd.h>
54
#include <signal.h>
55
#include <sys/types.h>
M
Marc G. Fournier 已提交
56

57
#if defined(solaris_sparc) || defined(__CYGWIN__)
58 59 60 61
#include <sys/ipc.h>
#include <sys/sem.h>
#endif

M
Marc G. Fournier 已提交
62
#include "postgres.h"
63
#include "miscadmin.h"
64
#include "libpq/pqsignal.h"
65 66


67 68
#include "storage/ipc.h"
/* In Ultrix and QNX, sem.h must be included after ipc.h */
69
#include <sys/sem.h>
B
Bruce Momjian 已提交
70

B
Bruce Momjian 已提交
71
#include "storage/lmgr.h"
72
#include "storage/proc.h"
M
 
Marc G. Fournier 已提交
73
#include "utils/trace.h"
74

75
void HandleDeadLock(SIGNAL_ARGS);
76
static void ProcFreeAllSemaphores(void);
77
static bool GetOffWaitqueue(PROC *);
78

M
 
Marc G. Fournier 已提交
79 80
#define DeadlockCheckTimer pg_options[OPT_DEADLOCKTIMEOUT]

81 82 83 84 85 86 87
/* --------------------
 * Spin lock for manipulating the shared process data structure:
 * ProcGlobal.... Adding an extra spin lock seemed like the smallest
 * hack to get around reading and updating this structure in shared
 * memory. -mer 17 July 1991
 * --------------------
 */
88
SPINLOCK	ProcStructLock;
89 90 91

static PROC_HDR *ProcGlobal = NULL;

92
PROC	   *MyProc = NULL;
93

94
static void ProcKill(int exitStatus, int pid);
95
static void ProcGetNewSemKeyAndNum(IPCKey *key, int *semNum);
96
static void ProcFreeSem(IpcSemaphoreKey semKey, int semNum);
97

V
Vadim B. Mikheev 已提交
98 99
static char *DeadLockMessage = "Deadlock detected -- See the lock(l) manual page for a possible cause.";

100 101
/*
 * InitProcGlobal -
102
 *	  initializes the global process table. We put it here so that
103
 *	  the postmaster can do this initialization. (ProcFreeAllSemaphores needs
104 105 106
 *	  to read this table on exiting the postmaster. If we have the first
 *	  backend do this, starting up and killing the postmaster without
 *	  starting any backends will be a problem.)
107 108 109 110 111 112 113 114 115 116 117
 *
 *	  We also allocate all the per-process semaphores we will need to support
 *	  the requested number of backends.  We used to allocate semaphores
 *	  only when backends were actually started up, but that is bad because
 *	  it lets Postgres fail under load --- a lot of Unix systems are
 *	  (mis)configured with small limits on the number of semaphores, and
 *	  running out when trying to start another backend is a common failure.
 *	  So, now we grab enough semaphores to support the desired max number
 *	  of backends immediately at initialization --- if the sysadmin has set
 *	  MaxBackends higher than his kernel will support, he'll find out sooner
 *	  rather than later.
118 119
 */
void
120
InitProcGlobal(IPCKey key, int maxBackends)
121
{
122
	bool		found = false;
123

124 125 126
	/* attach to the free list */
	ProcGlobal = (PROC_HDR *)
		ShmemInitStruct("Proc Header", (unsigned) sizeof(PROC_HDR), &found);
127

128 129
	/* --------------------
	 * We're the first - initialize.
130 131
	 * XXX if found should ever be true, it is a sign of impending doom ...
	 * ought to complain if so?
132 133 134
	 * --------------------
	 */
	if (!found)
135
	{
136
		int			i;
137

138 139 140 141
		ProcGlobal->freeProcs = INVALID_OFFSET;
		ProcGlobal->currKey = IPCGetProcessSemaphoreInitKey(key);
		for (i = 0; i < MAX_PROC_SEMS / PROC_NSEMS_PER_SET; i++)
			ProcGlobal->freeSemMap[i] = 0;
142

B
Bruce Momjian 已提交
143 144 145
		/*
		 * Arrange to delete semas on exit --- set this up now so that we
		 * will clean up if pre-allocation fails...
146 147 148
		 */
		on_shmem_exit(ProcFreeAllSemaphores, NULL);

B
Bruce Momjian 已提交
149 150
		/*
		 * Pre-create the semaphores for the first maxBackends processes,
151 152 153
		 * unless we are running as a standalone backend.
		 */
		if (key != PrivateIPCKey)
154
		{
155
			for (i = 0;
B
Bruce Momjian 已提交
156
				 i < (maxBackends + PROC_NSEMS_PER_SET - 1) / PROC_NSEMS_PER_SET;
157 158 159 160 161 162 163 164 165
				 i++)
			{
				IPCKey		semKey = ProcGlobal->currKey + i;
				int			semId;

				semId = IpcSemaphoreCreate(semKey,
										   PROC_NSEMS_PER_SET,
										   IPCProtection,
										   IpcSemaphoreDefaultStartValue,
166 167 168
										   0);
				if (semId < 0)
					elog(FATAL, "InitProcGlobal: IpcSemaphoreCreate failed");
169 170 171
				/* mark this sema set allocated */
				ProcGlobal->freeSemMap[i] = (1 << PROC_NSEMS_PER_SET);
			}
172
		}
173 174 175 176 177 178 179 180 181 182 183
	}
}

/* ------------------------
 * InitProc -- create a per-process data structure for this process
 * used by the lock manager on semaphore queues.
 * ------------------------
 */
void
InitProcess(IPCKey key)
{
184 185 186
	bool		found = false;
	unsigned long location,
				myOffset;
187 188 189 190 191 192 193

	SpinAcquire(ProcStructLock);

	/* attach to the free list */
	ProcGlobal = (PROC_HDR *)
		ShmemInitStruct("Proc Header", (unsigned) sizeof(PROC_HDR), &found);
	if (!found)
194
	{
195
		/* this should not happen. InitProcGlobal() is called before this. */
196
		elog(STOP, "InitProcess: Proc Header uninitialized");
197
	}
198 199

	if (MyProc != NULL)
200
	{
201
		SpinRelease(ProcStructLock);
202
		elog(ERROR, "ProcInit: you already exist");
203
		return;
204
	}
205 206 207 208 209 210

	/* try to get a proc from the free list first */

	myOffset = ProcGlobal->freeProcs;

	if (myOffset != INVALID_OFFSET)
211
	{
212 213 214 215 216 217 218
		MyProc = (PROC *) MAKE_PTR(myOffset);
		ProcGlobal->freeProcs = MyProc->links.next;
	}
	else
	{

		/*
219 220 221 222
		 * have to allocate one.  We can't use the normal shmem index
		 * table mechanism because the proc structure is stored by PID
		 * instead of by a global name (need to look it up by PID when we
		 * cleanup dead processes).
223 224 225 226
		 */

		MyProc = (PROC *) ShmemAlloc((unsigned) sizeof(PROC));
		if (!MyProc)
227
		{
228 229
			SpinRelease(ProcStructLock);
			elog(FATAL, "cannot create new proc: out of memory");
230
		}
231 232 233

		/* this cannot be initialized until after the buffer pool */
		SHMQueueInit(&(MyProc->lockQueue));
234
	}
235

236
	/*
237 238 239
	 * zero out the spin lock counts and set the sLocks field for
	 * ProcStructLock to 1 as we have acquired this spinlock above but
	 * didn't record it since we didn't have MyProc until now.
240
	 */
B
Bruce Momjian 已提交
241
	MemSet(MyProc->sLocks, 0, sizeof(MyProc->sLocks));
242 243 244 245 246
	MyProc->sLocks[ProcStructLock] = 1;


	if (IsUnderPostmaster)
	{
247 248 249 250
		IPCKey		semKey;
		int			semNum;
		int			semId;
		union semun semun;
251 252 253

		ProcGetNewSemKeyAndNum(&semKey, &semNum);

B
Bruce Momjian 已提交
254 255 256 257 258
		/*
		 * Note: because of the pre-allocation done in InitProcGlobal,
		 * this call should always attach to an existing semaphore. It
		 * will (try to) create a new group of semaphores only if the
		 * postmaster tries to start more backends than it said it would.
259
		 */
260 261 262 263
		semId = IpcSemaphoreCreate(semKey,
								   PROC_NSEMS_PER_SET,
								   IPCProtection,
								   IpcSemaphoreDefaultStartValue,
264
								   0);
265 266 267 268 269 270 271 272 273 274 275 276 277 278 279 280 281 282 283 284 285 286

		/*
		 * we might be reusing a semaphore that belongs to a dead backend.
		 * So be careful and reinitialize its value here.
		 */
		semun.val = IpcSemaphoreDefaultStartValue;
		semctl(semId, semNum, SETVAL, semun);

		IpcSemaphoreLock(semId, semNum, IpcExclusiveLock);
		MyProc->sem.semId = semId;
		MyProc->sem.semNum = semNum;
		MyProc->sem.semKey = semKey;
	}
	else
		MyProc->sem.semId = -1;

	/* ----------------------
	 * Release the lock.
	 * ----------------------
	 */
	SpinRelease(ProcStructLock);

B
Bruce Momjian 已提交
287
	MyProc->pid = MyProcPid;
288
	MyProc->databaseId = MyDatabaseId;
289
	MyProc->xid = InvalidTransactionId;
290
	MyProc->xmin = InvalidTransactionId;
291 292 293 294 295 296

	/* ----------------
	 * Start keeping spin lock stats from here on.	Any botch before
	 * this initialization is forever botched
	 * ----------------
	 */
B
Bruce Momjian 已提交
297
	MemSet(MyProc->sLocks, 0, MAX_SPINS * sizeof(*MyProc->sLocks));
298 299

	/* -------------------------
300
	 * Install ourselves in the shmem index table.	The name to
301 302 303 304 305 306
	 * use is determined by the OS-assigned process id.  That
	 * allows the cleanup process to find us after any untimely
	 * exit.
	 * -------------------------
	 */
	location = MAKE_OFFSET(MyProc);
B
Bruce Momjian 已提交
307
	if ((!ShmemPIDLookup(MyProcPid, &location)) || (location != MAKE_OFFSET(MyProc)))
308
		elog(STOP, "InitProc: ShmemPID table broken");
309 310 311 312

	MyProc->errType = NO_ERROR;
	SHMQueueElemInit(&(MyProc->links));

313
	on_shmem_exit(ProcKill, (caddr_t) MyProcPid);
314 315
}

H
Hiroshi Inoue 已提交
316 317 318 319
/* -----------------------
 * get off the wait queue
 * -----------------------
 */
320
static bool
H
Hiroshi Inoue 已提交
321 322
GetOffWaitqueue(PROC *proc)
{
323
	bool	getoffed = false;
H
Hiroshi Inoue 已提交
324 325 326 327 328 329 330 331 332 333 334 335 336 337 338
	LockLockTable();
	if (proc->links.next != INVALID_OFFSET)
	{
		int	lockmode = proc->token;
		Assert(proc->waitLock->waitProcs.size > 0);
		SHMQueueDelete(&(proc->links));
		--proc->waitLock->waitProcs.size;
		Assert(proc->waitLock->nHolding > 0);
		Assert(proc->waitLock->nHolding > proc->waitLock->nActive);
		--proc->waitLock->nHolding;
		Assert(proc->waitLock->holders[lockmode] > 0);
		--proc->waitLock->holders[lockmode];
		if (proc->waitLock->activeHolders[lockmode] ==
			proc->waitLock->holders[lockmode])
			proc->waitLock->waitMask &= ~(1 << lockmode);
339
		getoffed = true;
H
Hiroshi Inoue 已提交
340 341 342 343
	}
	SHMQueueElemInit(&(proc->links));
	UnlockLockTable();

344
	return getoffed;
H
Hiroshi Inoue 已提交
345
}
346 347 348 349 350 351 352
/*
 * ProcReleaseLocks() -- release all locks associated with this process
 *
 */
void
ProcReleaseLocks()
{
353 354 355
	if (!MyProc)
		return;
	LockReleaseAll(1, &MyProc->lockQueue);
H
Hiroshi Inoue 已提交
356
	GetOffWaitqueue(MyProc);
357 358 359 360
}

/*
 * ProcRemove -
361 362 363 364 365
 *	  used by the postmaster to clean up the global tables. This also frees
 *	  up the semaphore used for the lmgr of the process. (We have to do
 *	  this is the postmaster instead of doing a IpcSemaphoreKill on exiting
 *	  the process because the semaphore set is shared among backends and
 *	  we don't want to remove other's semaphores on exit.)
366 367 368 369
 */
bool
ProcRemove(int pid)
{
370 371
	SHMEM_OFFSET location;
	PROC	   *proc;
372 373 374 375 376

	location = INVALID_OFFSET;

	location = ShmemPIDDestroy(pid);
	if (location == INVALID_OFFSET)
377
		return FALSE;
378 379 380 381 382 383 384 385 386 387 388
	proc = (PROC *) MAKE_PTR(location);

	SpinAcquire(ProcStructLock);

	ProcFreeSem(proc->sem.semKey, proc->sem.semNum);

	proc->links.next = ProcGlobal->freeProcs;
	ProcGlobal->freeProcs = MAKE_OFFSET(proc);

	SpinRelease(ProcStructLock);

389
	return TRUE;
390 391 392 393
}

/*
 * ProcKill() -- Destroy the per-proc data structure for
394
 *		this process. Release any of its held spin locks.
395 396 397 398
 */
static void
ProcKill(int exitStatus, int pid)
{
399 400
	PROC	   *proc;
	SHMEM_OFFSET location;
401 402 403 404 405 406 407 408 409 410

	/* --------------------
	 * If this is a FATAL exit the postmaster will have to kill all the
	 * existing backends and reinitialize shared memory.  So all we don't
	 * need to do anything here.
	 * --------------------
	 */
	if (exitStatus != 0)
		return;

B
Bruce Momjian 已提交
411
	ShmemPIDLookup(MyProcPid, &location);
412 413 414 415 416
	if (location == INVALID_OFFSET)
		return;

	proc = (PROC *) MAKE_PTR(location);

417 418 419
	Assert(proc == MyProc || pid != MyProcPid);

	MyProc = NULL;
420 421 422 423 424 425

	/* ---------------
	 * Assume one lock table.
	 * ---------------
	 */
	ProcReleaseSpins(proc);
M
 
Marc G. Fournier 已提交
426
	LockReleaseAll(DEFAULT_LOCKMETHOD, &proc->lockQueue);
427

428
#ifdef USER_LOCKS
429

M
 
Marc G. Fournier 已提交
430 431 432 433
	/*
	 * Assume we have a second lock table.
	 */
	LockReleaseAll(USER_LOCKMETHOD, &proc->lockQueue);
434 435
#endif

436 437 438 439
	/* ----------------
	 * get off the wait queue
	 * ----------------
	 */
H
Hiroshi Inoue 已提交
440
	GetOffWaitqueue(proc);
441 442

	return;
443 444 445 446
}

/*
 * ProcQueue package: routines for putting processes to sleep
447
 *		and  waking them up
448 449 450 451 452 453 454 455
 */

/*
 * ProcQueueAlloc -- alloc/attach to a shared memory process queue
 *
 * Returns: a pointer to the queue or NULL
 * Side Effects: Initializes the queue if we allocated one
 */
456
#ifdef NOT_USED
457
PROC_QUEUE *
458 459
ProcQueueAlloc(char *name)
{
460 461
	bool		found;
	PROC_QUEUE *queue = (PROC_QUEUE *)
462 463 464
	ShmemInitStruct(name, (unsigned) sizeof(PROC_QUEUE), &found);

	if (!queue)
465
		return NULL;
466 467
	if (!found)
		ProcQueueInit(queue);
468
	return queue;
469
}
470

471
#endif
472 473 474 475 476

/*
 * ProcQueueInit -- initialize a shared memory process queue
 */
void
477
ProcQueueInit(PROC_QUEUE *queue)
478
{
479 480
	SHMQueueInit(&(queue->links));
	queue->size = 0;
481 482 483
}


484 485 486 487
/*
 *	Handling cancel request while waiting for lock
 *
 */
488 489 490
static bool	lockWaiting = false;
void	SetWaitingForLock(bool waiting)
{
491 492
	if (waiting == lockWaiting)
		return;
493
	lockWaiting = waiting;
494 495 496 497 498 499 500 501 502 503 504 505
	if (lockWaiting)
	{
		Assert(MyProc->links.next != INVALID_OFFSET);
		if (QueryCancel) /* cancel request pending */
		{
			if (GetOffWaitqueue(MyProc))
			{
				lockWaiting = false;
				elog(ERROR, "Query cancel requested while waiting lock");
			}
		}
	}
506 507 508 509 510 511 512 513 514 515 516 517 518
}
void	LockWaitCancel(void)
{
	struct itimerval timeval, dummy;

	if (!lockWaiting)	return;
	lockWaiting = false;
	/* Deadlock timer off */
	MemSet(&timeval, 0, sizeof(struct itimerval));
	setitimer(ITIMER_REAL, &timeval, &dummy);
	if (GetOffWaitqueue(MyProc))
		elog(ERROR, "Query cancel requested while waiting lock");
}
519 520 521 522 523 524 525 526 527

/*
 * ProcSleep -- put a process to sleep
 *
 * P() on the semaphore should put us to sleep.  The process
 * semaphore is cleared by default, so the first time we try
 * to acquire it, we sleep.
 *
 * ASSUME: that no one will fiddle with the queue until after
528
 *		we release the spin lock.
529 530 531 532
 *
 * NOTES: The process queue is now a priority queue for locking.
 */
int
533
ProcSleep(PROC_QUEUE *waitQueue,/* lock->waitProcs */
534
		  LOCKMETHODCTL *lockctl,
535
		  int token,			/* lockmode */
V
Vadim B. Mikheev 已提交
536
		  LOCK *lock)
537
{
538
	int			i;
V
Vadim B. Mikheev 已提交
539
	SPINLOCK	spinlock = lockctl->masterLock;
540
	PROC	   *proc;
V
Vadim B. Mikheev 已提交
541 542 543 544 545
	int			myMask = (1 << token);
	int			waitMask = lock->waitMask;
	int			aheadHolders[MAX_LOCKMODES];
	bool		selfConflict = (lockctl->conflictTab[token] & myMask),
				prevSame = false;
B
Bruce Momjian 已提交
546 547 548
	bool		deadlock_checked = false;
	struct itimerval timeval,
				dummy;
549

V
Vadim B. Mikheev 已提交
550 551 552
	MyProc->token = token;
	MyProc->waitLock = lock;

B
Bruce Momjian 已提交
553
	proc = (PROC *) MAKE_PTR(waitQueue->links.prev);
554

V
Vadim B. Mikheev 已提交
555 556 557
	/* if we don't conflict with any waiter - be first in queue */
	if (!(lockctl->conflictTab[token] & waitMask))
		goto ins;
558

V
Vadim B. Mikheev 已提交
559 560 561
	for (i = 1; i < MAX_LOCKMODES; i++)
		aheadHolders[i] = lock->activeHolders[i];
	(aheadHolders[token])++;
562

V
Vadim B. Mikheev 已提交
563 564 565 566 567 568 569 570 571 572 573 574 575 576 577 578 579 580 581 582
	for (i = 0; i < waitQueue->size; i++)
	{
		/* am I waiting for him ? */
		if (lockctl->conflictTab[token] & proc->holdLock)
		{
			/* is he waiting for me ? */
			if (lockctl->conflictTab[proc->token] & MyProc->holdLock)
			{
				MyProc->errType = STATUS_ERROR;
				elog(NOTICE, DeadLockMessage);
				goto rt;
			}
			/* being waiting for him - go past */
		}
		/* if he waits for me */
		else if (lockctl->conflictTab[proc->token] & MyProc->holdLock)
			break;
		/* if conflicting locks requested */
		else if (lockctl->conflictTab[proc->token] & myMask)
		{
B
Bruce Momjian 已提交
583

V
Vadim B. Mikheev 已提交
584
			/*
B
Bruce Momjian 已提交
585 586
			 * If I request non self-conflicting lock and there are others
			 * requesting the same lock just before me - stay here.
V
Vadim B. Mikheev 已提交
587 588 589 590
			 */
			if (!selfConflict && prevSame)
				break;
		}
B
Bruce Momjian 已提交
591

V
Vadim B. Mikheev 已提交
592
		/*
B
Bruce Momjian 已提交
593 594
		 * Last attempt to don't move any more: if we don't conflict with
		 * rest waiters in queue.
V
Vadim B. Mikheev 已提交
595 596 597
		 */
		else if (!(lockctl->conflictTab[token] & waitMask))
			break;
598

V
Vadim B. Mikheev 已提交
599 600 601
		prevSame = (proc->token == token);
		(aheadHolders[proc->token])++;
		if (aheadHolders[proc->token] == lock->holders[proc->token])
B
Bruce Momjian 已提交
602
			waitMask &= ~(1 << proc->token);
V
Vadim B. Mikheev 已提交
603 604
		proc = (PROC *) MAKE_PTR(proc->links.prev);
	}
605

V
Vadim B. Mikheev 已提交
606
ins:;
607 608 609 610 611 612
	/* -------------------
	 * assume that these two operations are atomic (because
	 * of the spinlock).
	 * -------------------
	 */
	SHMQueueInsertTL(&(proc->links), &(MyProc->links));
B
Bruce Momjian 已提交
613
	waitQueue->size++;
614

V
Vadim B. Mikheev 已提交
615
	lock->waitMask |= myMask;
616 617 618
	SpinRelease(spinlock);

	/* --------------
B
Bruce Momjian 已提交
619
	 * We set this so we can wake up periodically and check for a deadlock.
B
Bruce Momjian 已提交
620 621
	 * If a deadlock is detected, the handler releases the processes
	 * semaphore and aborts the current transaction.
B
Bruce Momjian 已提交
622 623 624
	 *
	 * Need to zero out struct to set the interval and the micro seconds fields
	 * to 0.
625 626
	 * --------------
	 */
B
Bruce Momjian 已提交
627 628 629
	MemSet(&timeval, 0, sizeof(struct itimerval));
	timeval.it_value.tv_sec = \
		(DeadlockCheckTimer ? DeadlockCheckTimer : DEADLOCK_CHECK_TIMER);
630

631
	SetWaitingForLock(true);
B
Bruce Momjian 已提交
632 633
	do
	{
634
		MyProc->errType = NO_ERROR;		/* reset flag after deadlock check */
635

B
Bruce Momjian 已提交
636 637 638 639 640
		if (!deadlock_checked)
			if (setitimer(ITIMER_REAL, &timeval, &dummy))
				elog(FATAL, "ProcSleep: Unable to set timer for process wakeup");
		deadlock_checked = true;

B
Bruce Momjian 已提交
641 642 643 644 645 646
		/* --------------
		 * if someone wakes us between SpinRelease and IpcSemaphoreLock,
		 * IpcSemaphoreLock will not block.  The wakeup is "saved" by
		 * the semaphore implementation.
		 * --------------
		 */
M
 
Marc G. Fournier 已提交
647 648
		IpcSemaphoreLock(MyProc->sem.semId, MyProc->sem.semNum,
						 IpcExclusiveLock);
649 650
	} while (MyProc->errType == STATUS_NOT_FOUND);		/* sleep after deadlock
														 * check */
651
	lockWaiting = false;
652

B
Bruce Momjian 已提交
653 654 655 656 657 658 659 660
	/* ---------------
	 * We were awoken before a timeout - now disable the timer
	 * ---------------
	 */
	timeval.it_value.tv_sec = 0;
	if (setitimer(ITIMER_REAL, &timeval, &dummy))
		elog(FATAL, "ProcSleep: Unable to diable timer for process wakeup");

661 662 663 664 665 666 667
	/* ----------------
	 * We were assumed to be in a critical section when we went
	 * to sleep.
	 * ----------------
	 */
	SpinAcquire(spinlock);

V
Vadim B. Mikheev 已提交
668 669
rt:;

M
 
Marc G. Fournier 已提交
670 671
#ifdef LOCK_MGR_DEBUG
	/* Just to get meaningful debug messages from DumpLocks() */
672
	MyProc->waitLock = (LOCK *) NULL;
M
 
Marc G. Fournier 已提交
673 674
#endif

675
	return MyProc->errType;
676 677 678 679 680 681
}


/*
 * ProcWakeup -- wake up a process by releasing its private semaphore.
 *
682 683
 *	 remove the process from the wait queue and set its links invalid.
 *	 RETURN: the next process in the wait queue.
684
 */
B
Bruce Momjian 已提交
685
PROC *
686
ProcWakeup(PROC *proc, int errType)
687
{
688
	PROC	   *retProc;
689 690 691 692 693

	/* assume that spinlock has been acquired */

	if (proc->links.prev == INVALID_OFFSET ||
		proc->links.next == INVALID_OFFSET)
694
		return (PROC *) NULL;
695 696 697 698 699 700 701 702 703 704 705 706

	retProc = (PROC *) MAKE_PTR(proc->links.prev);

	/* you have to update waitLock->waitProcs.size yourself */
	SHMQueueDelete(&(proc->links));
	SHMQueueElemInit(&(proc->links));

	proc->errType = errType;

	IpcSemaphoreUnlock(proc->sem.semId, proc->sem.semNum, IpcExclusiveLock);

	return retProc;
707 708 709 710
}

/*
 * ProcLockWakeup -- routine for waking up processes when a lock is
711
 *		released.
712 713
 */
int
714
ProcLockWakeup(PROC_QUEUE *queue, LOCKMETHOD lockmethod, LOCK *lock)
715
{
716
	PROC	   *proc;
V
Vadim B. Mikheev 已提交
717
	int			count = 0;
M
 
Marc G. Fournier 已提交
718
	int			trace_flag;
V
Vadim B. Mikheev 已提交
719
	int			last_locktype = 0;
M
 
Marc G. Fournier 已提交
720 721 722
	int			queue_size = queue->size;

	Assert(queue->size >= 0);
723 724

	if (!queue->size)
725
		return STATUS_NOT_FOUND;
726 727

	proc = (PROC *) MAKE_PTR(queue->links.prev);
M
 
Marc G. Fournier 已提交
728 729
	while ((queue_size--) && (proc))
	{
730

M
 
Marc G. Fournier 已提交
731
		/*
732 733
		 * This proc will conflict as the previous one did, don't even
		 * try.
M
 
Marc G. Fournier 已提交
734 735 736 737 738
		 */
		if (proc->token == last_locktype)
			continue;

		/*
V
Vadim B. Mikheev 已提交
739
		 * Does this proc conflict with locks held by others ?
M
 
Marc G. Fournier 已提交
740 741
		 */
		if (LockResolveConflicts(lockmethod,
742
								 lock,
743
								 proc->token,
M
 
Marc G. Fournier 已提交
744 745 746
								 proc->xid,
								 (XIDLookupEnt *) NULL) != STATUS_OK)
		{
V
Vadim B. Mikheev 已提交
747 748
			if (count != 0)
				break;
M
 
Marc G. Fournier 已提交
749 750 751
			last_locktype = proc->token;
			continue;
		}
752 753 754 755 756 757 758

		/*
		 * there was a waiting process, grant it the lock before waking it
		 * up.	This will prevent another process from seizing the lock
		 * between the time we release the lock master (spinlock) and the
		 * time that the awoken process begins executing again.
		 */
759
		GrantLock(lock, proc->token);
760 761 762

		/*
		 * ProcWakeup removes proc from the lock waiting process queue and
763
		 * returns the next proc in chain.
764 765 766
		 */

		count++;
M
 
Marc G. Fournier 已提交
767 768
		queue->size--;
		proc = ProcWakeup(proc, NO_ERROR);
769
	}
770

M
 
Marc G. Fournier 已提交
771 772
	Assert(queue->size >= 0);

773
	if (count)
774
		return STATUS_OK;
775 776
	else
	{
777
		/* Something is still blocking us.	May have deadlocked. */
M
 
Marc G. Fournier 已提交
778 779 780 781 782 783 784 785 786
		trace_flag = (lock->tag.lockmethod == USER_LOCKMETHOD) ? \
			TRACE_USERLOCKS : TRACE_LOCKS;
		TPRINTF(trace_flag,
				"ProcLockWakeup: lock(%x) can't wake up any process",
				MAKE_OFFSET(lock));
#ifdef DEADLOCK_DEBUG
		if (pg_options[trace_flag] >= 2)
			DumpAllLocks();
#endif
787
		return STATUS_NOT_FOUND;
M
 
Marc G. Fournier 已提交
788
	}
789 790 791
}

void
792
ProcAddLock(SHM_QUEUE *elem)
793
{
794
	SHMQueueInsertTL(&MyProc->lockQueue, elem);
795 796 797
}

/* --------------------
B
Bruce Momjian 已提交
798 799 800
 * We only get to this routine if we got SIGALRM after DEADLOCK_CHECK_TIMER
 * while waiting for a lock to be released by some other process.  If we have
 * a real deadlock, we must also indicate that I'm no longer waiting
801
 * on a lock so that other processes don't try to wake me up and screw
802 803 804
 * up my semaphore.
 * --------------------
 */
805 806
void
HandleDeadLock(SIGNAL_ARGS)
807
{
B
Bruce Momjian 已提交
808
	LOCK	   *mywaitlock;
809 810 811 812 813 814 815 816 817 818 819 820 821 822 823 824 825 826 827 828 829 830 831 832 833 834 835 836 837 838 839 840 841 842 843 844 845 846

	LockLockTable();

	/* ---------------------
	 * Check to see if we've been awoken by anyone in the interim.
	 *
	 * If we have we can return and resume our transaction -- happy day.
	 * Before we are awoken the process releasing the lock grants it to
	 * us so we know that we don't have to wait anymore.
	 *
	 * Damn these names are LONG! -mer
	 * ---------------------
	 */
	if (IpcSemaphoreGetCount(MyProc->sem.semId, MyProc->sem.semNum) ==
		IpcSemaphoreDefaultStartValue)
	{
		UnlockLockTable();
		return;
	}

	/*
	 * you would think this would be unnecessary, but...
	 *
	 * this also means we've been removed already.  in some ports (e.g.,
	 * sparc and aix) the semop(2) implementation is such that we can
	 * actually end up in this handler after someone has removed us from
	 * the queue and bopped the semaphore *but the test above fails to
	 * detect the semaphore update* (presumably something weird having to
	 * do with the order in which the semaphore wakeup signal and SIGALRM
	 * get handled).
	 */
	if (MyProc->links.prev == INVALID_OFFSET ||
		MyProc->links.next == INVALID_OFFSET)
	{
		UnlockLockTable();
		return;
	}

847
#ifdef DEADLOCK_DEBUG
M
 
Marc G. Fournier 已提交
848
	DumpAllLocks();
849 850
#endif

B
Bruce Momjian 已提交
851 852
	MyProc->errType = STATUS_NOT_FOUND;
	if (!DeadLockCheck(MyProc, MyProc->waitLock))
B
Bruce Momjian 已提交
853 854 855 856 857 858 859
	{
		UnlockLockTable();
		return;
	}

	mywaitlock = MyProc->waitLock;

860 861 862 863
	/* ------------------------
	 * Get this process off the lock's wait queue
	 * ------------------------
	 */
B
Bruce Momjian 已提交
864
	Assert(mywaitlock->waitProcs.size > 0);
865
	lockWaiting = false;
B
Bruce Momjian 已提交
866
	--mywaitlock->waitProcs.size;
867 868 869 870 871 872 873 874
	SHMQueueDelete(&(MyProc->links));
	SHMQueueElemInit(&(MyProc->links));

	/* ------------------
	 * Unlock my semaphore so that the count is right for next time.
	 * I was awoken by a signal, not by someone unlocking my semaphore.
	 * ------------------
	 */
M
 
Marc G. Fournier 已提交
875 876
	IpcSemaphoreUnlock(MyProc->sem.semId, MyProc->sem.semNum,
					   IpcExclusiveLock);
877 878 879 880 881 882 883 884 885 886 887 888 889 890 891

	/* -------------
	 * Set MyProc->errType to STATUS_ERROR so that we abort after
	 * returning from this handler.
	 * -------------
	 */
	MyProc->errType = STATUS_ERROR;

	/*
	 * if this doesn't follow the IpcSemaphoreUnlock then we get lock
	 * table corruption ("LockReplace: xid table corrupted") due to race
	 * conditions.	i don't claim to understand this...
	 */
	UnlockLockTable();

V
Vadim B. Mikheev 已提交
892
	elog(NOTICE, DeadLockMessage);
893
	return;
894 895 896
}

void
897
ProcReleaseSpins(PROC *proc)
898
{
899
	int			i;
900 901 902 903 904 905 906

	if (!proc)
		proc = MyProc;

	if (!proc)
		return;
	for (i = 0; i < (int) MAX_SPINS; i++)
907
	{
908
		if (proc->sLocks[i])
909
		{
910 911
			Assert(proc->sLocks[i] == 1);
			SpinRelease(i);
912 913
		}
	}
H
 
Hiroshi Inoue 已提交
914
	AbortBufferIO();
915 916 917
}

/*****************************************************************************
918
 *
919 920 921 922
 *****************************************************************************/

/*
 * ProcGetNewSemKeyAndNum -
923 924 925 926
 *	  scan the free semaphore bitmap and allocate a single semaphore from
 *	  a semaphore set. (If the semaphore set doesn't exist yet,
 *	  IpcSemaphoreCreate will create it. Otherwise, we use the existing
 *	  semaphore set.)
927 928
 */
static void
929
ProcGetNewSemKeyAndNum(IPCKey *key, int *semNum)
930
{
931 932
	int			i;
	int32	   *freeSemMap = ProcGlobal->freeSemMap;
B
Bruce Momjian 已提交
933
	int32		fullmask = (1 << (PROC_NSEMS_PER_SET + 1)) - 1;
934

935 936 937 938
	/*
	 * we hold ProcStructLock when entering this routine. We scan through
	 * the bitmap to look for a free semaphore.
	 */
939

940 941
	for (i = 0; i < MAX_PROC_SEMS / PROC_NSEMS_PER_SET; i++)
	{
942 943
		int			mask = 1;
		int			j;
944 945

		if (freeSemMap[i] == fullmask)
946
			continue;			/* this set is fully allocated */
947 948 949 950 951 952 953

		for (j = 0; j < PROC_NSEMS_PER_SET; j++)
		{
			if ((freeSemMap[i] & mask) == 0)
			{

				/*
B
Bruce Momjian 已提交
954 955
				 * a free semaphore found. Mark it as allocated. Also set
				 * the bit indicating whole set is allocated.
956
				 */
957
				freeSemMap[i] |= mask + (1 << PROC_NSEMS_PER_SET);
958 959 960 961 962 963 964

				*key = ProcGlobal->currKey + i;
				*semNum = j;
				return;
			}
			mask <<= 1;
		}
965 966
	}

967
	/* if we reach here, all the semaphores are in use. */
968
	elog(ERROR, "InitProc: cannot allocate a free semaphore");
969 970 971 972
}

/*
 * ProcFreeSem -
973
 *	  free up our semaphore in the semaphore set.
974 975 976 977
 */
static void
ProcFreeSem(IpcSemaphoreKey semKey, int semNum)
{
978 979 980
	int			mask;
	int			i;
	int32	   *freeSemMap = ProcGlobal->freeSemMap;
981

982 983 984
	i = semKey - ProcGlobal->currKey;
	mask = ~(1 << semNum);
	freeSemMap[i] &= mask;
985

B
Bruce Momjian 已提交
986 987 988 989
	/*
	 * Formerly we'd release a semaphore set if it was now completely
	 * unused, but now we keep the semaphores to ensure we won't run out
	 * when starting new backends --- cf. InitProcGlobal.  Note that the
990 991 992
	 * PROC_NSEMS_PER_SET+1'st bit of the freeSemMap entry remains set to
	 * indicate it is still allocated; ProcFreeAllSemaphores() needs that.
	 */
993 994 995 996
}

/*
 * ProcFreeAllSemaphores -
997 998 999
 *	  called at shmem_exit time, ie when exiting the postmaster or
 *	  destroying shared state for a failed set of backends.
 *	  Free up all the semaphores allocated to the lmgrs of the backends.
1000
 */
1001
static void
1002 1003
ProcFreeAllSemaphores()
{
1004 1005
	int			i;
	int32	   *freeSemMap = ProcGlobal->freeSemMap;
1006

1007 1008 1009 1010 1011
	for (i = 0; i < MAX_PROC_SEMS / PROC_NSEMS_PER_SET; i++)
	{
		if (freeSemMap[i] != 0)
			IpcSemaphoreKill(ProcGlobal->currKey + i);
	}
1012
}