spinlock.h 7.0 KB
Newer Older
1
/* SPDX-License-Identifier: GPL-2.0 */
L
Linus Torvalds 已提交
2 3 4 5 6 7 8 9 10 11 12 13 14
#ifndef _ASM_IA64_SPINLOCK_H
#define _ASM_IA64_SPINLOCK_H

/*
 * Copyright (C) 1998-2003 Hewlett-Packard Co
 *	David Mosberger-Tang <davidm@hpl.hp.com>
 * Copyright (C) 1999 Walt Drummond <drummond@valinux.com>
 *
 * This file is used for SMP configurations only.
 */

#include <linux/compiler.h>
#include <linux/kernel.h>
J
Jiri Slaby 已提交
15
#include <linux/bitops.h>
L
Linus Torvalds 已提交
16

A
Arun Sharma 已提交
17
#include <linux/atomic.h>
L
Linus Torvalds 已提交
18
#include <asm/intrinsics.h>
19 20
#include <asm/barrier.h>
#include <asm/processor.h>
L
Linus Torvalds 已提交
21

22
#define arch_spin_lock_init(x)			((x)->lock = 0)
L
Linus Torvalds 已提交
23 24

/*
25 26 27 28
 * Ticket locks are conceptually two parts, one indicating the current head of
 * the queue, and the other indicating the current tail. The lock is acquired
 * by atomically noting the tail and incrementing it by one (thus adding
 * ourself to the queue and noting our position), then waiting until the head
29
 * becomes equal to the initial value of the tail.
30 31
 * The pad bits in the middle are used to prevent the next_ticket number
 * overflowing into the now_serving number.
32
 *
33
 *   31             17  16    15  14                    0
34
 *  +----------------------------------------------------+
35
 *  |  now_serving     | padding |   next_ticket         |
36
 *  +----------------------------------------------------+
L
Linus Torvalds 已提交
37 38
 */

39 40 41
#define TICKET_SHIFT	17
#define TICKET_BITS	15
#define	TICKET_MASK	((1 << TICKET_BITS) - 1)
42

43
static __always_inline void __ticket_spin_lock(arch_spinlock_t *lock)
44
{
45
	int	*p = (int *)&lock->lock, ticket, serve;
46

47
	ticket = ia64_fetchadd(1, p, acq);
48

49
	if (!(((ticket >> TICKET_SHIFT) ^ ticket) & TICKET_MASK))
50 51
		return;

52 53 54 55 56 57 58
	ia64_invala();

	for (;;) {
		asm volatile ("ld4.c.nc %0=[%1]" : "=r"(serve) : "r"(p) : "memory");

		if (!(((serve >> TICKET_SHIFT) ^ ticket) & TICKET_MASK))
			return;
59
		cpu_relax();
60
	}
61 62
}

63
static __always_inline int __ticket_spin_trylock(arch_spinlock_t *lock)
64
{
65
	int tmp = READ_ONCE(lock->lock);
66

67 68
	if (!(((tmp >> TICKET_SHIFT) ^ tmp) & TICKET_MASK))
		return ia64_cmpxchg(acq, &lock->lock, tmp, tmp + 1, sizeof (tmp)) == tmp;
69 70 71
	return 0;
}

72
static __always_inline void __ticket_spin_unlock(arch_spinlock_t *lock)
73
{
74
	unsigned short	*p = (unsigned short *)&lock->lock + 1, tmp;
75

76 77
	/* This could be optimised with ARCH_HAS_MMIOWB */
	mmiowb();
78
	asm volatile ("ld2.bias %0=[%1]" : "=r"(tmp) : "r"(p));
79
	WRITE_ONCE(*p, (tmp + 2) & ~1);
80
}
L
Linus Torvalds 已提交
81

82
static inline int __ticket_spin_is_locked(arch_spinlock_t *lock)
L
Linus Torvalds 已提交
83
{
84
	long tmp = READ_ONCE(lock->lock);
85

86
	return !!(((tmp >> TICKET_SHIFT) ^ tmp) & TICKET_MASK);
L
Linus Torvalds 已提交
87
}
88

89
static inline int __ticket_spin_is_contended(arch_spinlock_t *lock)
90
{
91
	long tmp = READ_ONCE(lock->lock);
92

93
	return ((tmp - (tmp >> TICKET_SHIFT)) & TICKET_MASK) > 1;
94 95
}

96 97 98 99 100
static __always_inline int arch_spin_value_unlocked(arch_spinlock_t lock)
{
	return !(((lock.lock >> TICKET_SHIFT) ^ lock.lock) & TICKET_MASK);
}

101
static inline int arch_spin_is_locked(arch_spinlock_t *lock)
102 103 104 105
{
	return __ticket_spin_is_locked(lock);
}

106
static inline int arch_spin_is_contended(arch_spinlock_t *lock)
107 108 109
{
	return __ticket_spin_is_contended(lock);
}
110
#define arch_spin_is_contended	arch_spin_is_contended
111

112
static __always_inline void arch_spin_lock(arch_spinlock_t *lock)
113 114 115 116
{
	__ticket_spin_lock(lock);
}

117
static __always_inline int arch_spin_trylock(arch_spinlock_t *lock)
118 119 120 121
{
	return __ticket_spin_trylock(lock);
}

122
static __always_inline void arch_spin_unlock(arch_spinlock_t *lock)
123 124 125 126
{
	__ticket_spin_unlock(lock);
}

127
static __always_inline void arch_spin_lock_flags(arch_spinlock_t *lock,
128 129
						  unsigned long flags)
{
130
	arch_spin_lock(lock);
131
}
132
#define arch_spin_lock_flags	arch_spin_lock_flags
133

134 135 136
#ifdef ASM_SUPPORTED

static __always_inline void
137
arch_read_lock_flags(arch_rwlock_t *lock, unsigned long flags)
138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159
{
	__asm__ __volatile__ (
		"tbit.nz p6, p0 = %1,%2\n"
		"br.few 3f\n"
		"1:\n"
		"fetchadd4.rel r2 = [%0], -1;;\n"
		"(p6) ssm psr.i\n"
		"2:\n"
		"hint @pause\n"
		"ld4 r2 = [%0];;\n"
		"cmp4.lt p7,p0 = r2, r0\n"
		"(p7) br.cond.spnt.few 2b\n"
		"(p6) rsm psr.i\n"
		";;\n"
		"3:\n"
		"fetchadd4.acq r2 = [%0], 1;;\n"
		"cmp4.lt p7,p0 = r2, r0\n"
		"(p7) br.cond.spnt.few 1b\n"
		: : "r"(lock), "r"(flags), "i"(IA64_PSR_I_BIT)
		: "p6", "p7", "r2", "memory");
}

160
#define arch_read_lock_flags arch_read_lock_flags
161
#define arch_read_lock(lock) arch_read_lock_flags(lock, 0)
162 163 164

#else /* !ASM_SUPPORTED */

165
#define arch_read_lock_flags(rw, flags) arch_read_lock(rw)
166

167
#define arch_read_lock(rw)								\
L
Linus Torvalds 已提交
168
do {											\
169
	arch_rwlock_t *__read_lock_ptr = (rw);						\
L
Linus Torvalds 已提交
170 171 172 173 174 175 176 177
											\
	while (unlikely(ia64_fetchadd(1, (int *) __read_lock_ptr, acq) < 0)) {		\
		ia64_fetchadd(-1, (int *) __read_lock_ptr, rel);			\
		while (*(volatile int *)__read_lock_ptr < 0)				\
			cpu_relax();							\
	}										\
} while (0)

178 179
#endif /* !ASM_SUPPORTED */

180
#define arch_read_unlock(rw)					\
L
Linus Torvalds 已提交
181
do {								\
182
	arch_rwlock_t *__read_lock_ptr = (rw);			\
L
Linus Torvalds 已提交
183 184 185 186
	ia64_fetchadd(-1, (int *) __read_lock_ptr, rel);	\
} while (0)

#ifdef ASM_SUPPORTED
187 188

static __always_inline void
189
arch_write_lock_flags(arch_rwlock_t *lock, unsigned long flags)
190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212
{
	__asm__ __volatile__ (
		"tbit.nz p6, p0 = %1, %2\n"
		"mov ar.ccv = r0\n"
		"dep r29 = -1, r0, 31, 1\n"
		"br.few 3f;;\n"
		"1:\n"
		"(p6) ssm psr.i\n"
		"2:\n"
		"hint @pause\n"
		"ld4 r2 = [%0];;\n"
		"cmp4.eq p0,p7 = r0, r2\n"
		"(p7) br.cond.spnt.few 2b\n"
		"(p6) rsm psr.i\n"
		";;\n"
		"3:\n"
		"cmpxchg4.acq r2 = [%0], r29, ar.ccv;;\n"
		"cmp4.eq p0,p7 = r0, r2\n"
		"(p7) br.cond.spnt.few 1b;;\n"
		: : "r"(lock), "r"(flags), "i"(IA64_PSR_I_BIT)
		: "ar.ccv", "p6", "p7", "r2", "r29", "memory");
}

213
#define arch_write_lock_flags arch_write_lock_flags
214
#define arch_write_lock(rw) arch_write_lock_flags(rw, 0)
L
Linus Torvalds 已提交
215

216
#define arch_write_trylock(rw)							\
L
Linus Torvalds 已提交
217 218 219 220 221 222 223 224 225 226 227
({										\
	register long result;							\
										\
	__asm__ __volatile__ (							\
		"mov ar.ccv = r0\n"						\
		"dep r29 = -1, r0, 31, 1;;\n"					\
		"cmpxchg4.acq %0 = [%1], r29, ar.ccv\n"				\
		: "=r"(result) : "r"(rw) : "ar.ccv", "r29", "memory");		\
	(result == 0);								\
})

228
static inline void arch_write_unlock(arch_rwlock_t *x)
229 230 231 232 233 234
{
	u8 *y = (u8 *)x;
	barrier();
	asm volatile ("st1.rel.nta [%0] = r0\n\t" :: "r"(y+3) : "memory" );
}

L
Linus Torvalds 已提交
235 236
#else /* !ASM_SUPPORTED */

237
#define arch_write_lock(l)								\
L
Linus Torvalds 已提交
238 239 240 241 242 243 244 245 246 247
({											\
	__u64 ia64_val, ia64_set_val = ia64_dep_mi(-1, 0, 31, 1);			\
	__u32 *ia64_write_lock_ptr = (__u32 *) (l);					\
	do {										\
		while (*ia64_write_lock_ptr)						\
			ia64_barrier();							\
		ia64_val = ia64_cmpxchg4_acq(ia64_write_lock_ptr, ia64_set_val, 0);	\
	} while (ia64_val);								\
})

248
#define arch_write_trylock(rw)						\
L
Linus Torvalds 已提交
249 250 251 252 253 254 255
({									\
	__u64 ia64_val;							\
	__u64 ia64_set_val = ia64_dep_mi(-1, 0, 31,1);			\
	ia64_val = ia64_cmpxchg4_acq((__u32 *)(rw), ia64_set_val, 0);	\
	(ia64_val == 0);						\
})

256
static inline void arch_write_unlock(arch_rwlock_t *x)
257 258 259 260 261
{
	barrier();
	x->write_lock = 0;
}

L
Linus Torvalds 已提交
262 263
#endif /* !ASM_SUPPORTED */

264
static inline int arch_read_trylock(arch_rwlock_t *x)
265 266
{
	union {
267
		arch_rwlock_t lock;
268 269 270 271 272 273 274
		__u32 word;
	} old, new;
	old.lock = new.lock = *x;
	old.lock.write_lock = new.lock.write_lock = 0;
	++new.lock.read_counter;
	return (u32)ia64_cmpxchg4_acq((__u32 *)(x), new.word, old.word) == old.word;
}
L
Linus Torvalds 已提交
275 276

#endif /*  _ASM_IA64_SPINLOCK_H */