1/*	$NetBSD: psl.h,v 1.62.4.2 2023/09/09 15:01:24 martin Exp $ */
  2
  3/*
  4 * Copyright (c) 1992, 1993
  5 *	The Regents of the University of California.  All rights reserved.
  6 *
  7 * This software was developed by the Computer Systems Engineering group
  8 * at Lawrence Berkeley Laboratory under DARPA contract BG 91-66 and
  9 * contributed to Berkeley.
 10 *
 11 * All advertising materials mentioning features or use of this software
 12 * must display the following acknowledgement:
 13 *	This product includes software developed by the University of
 14 *	California, Lawrence Berkeley Laboratory.
 15 *
 16 * Redistribution and use in source and binary forms, with or without
 17 * modification, are permitted provided that the following conditions
 18 * are met:
 19 * 1. Redistributions of source code must retain the above copyright
 20 *    notice, this list of conditions and the following disclaimer.
 21 * 2. Redistributions in binary form must reproduce the above copyright
 22 *    notice, this list of conditions and the following disclaimer in the
 23 *    documentation and/or other materials provided with the distribution.
 24 * 3. Neither the name of the University nor the names of its contributors
 25 *    may be used to endorse or promote products derived from this software
 26 *    without specific prior written permission.
 27 *
 28 * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND
 29 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
 30 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
 31 * ARE DISCLAIMED.  IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
 32 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
 33 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
 34 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
 35 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
 36 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
 37 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
 38 * SUCH DAMAGE.
 39 *
 40 *	@(#)psl.h	8.1 (Berkeley) 6/11/93
 41 */
 42
 43#ifndef PSR_IMPL
 44
 45/*
 46 * SPARC Process Status Register (in psl.h for hysterical raisins).  This
 47 * doesn't exist on the V9.
 48 *
 49 * The picture in the Sun manuals looks like this:
 50 *	                                     1 1
 51 *	 31   28 27   24 23   20 19       14 3 2 11    8 7 6 5 4       0
 52 *	+-------+-------+-------+-----------+-+-+-------+-+-+-+---------+
 53 *	|  impl |  ver  |  icc  |  reserved |E|E|  pil  |S|P|E|   CWP   |
 54 *	|       |       |n z v c|           |C|F|       | |S|T|         |
 55 *	+-------+-------+-------+-----------+-+-+-------+-+-+-+---------+
 56 */
 57
 58#define PSR_IMPL	0xf0000000	/* implementation */
 59#define PSR_VER		0x0f000000	/* version */
 60#define PSR_ICC		0x00f00000	/* integer condition codes */
 61#define PSR_N		0x00800000	/* negative */
 62#define PSR_Z		0x00400000	/* zero */
 63#define PSR_O		0x00200000	/* overflow */
 64#define PSR_C		0x00100000	/* carry */
 65#define PSR_EC		0x00002000	/* coprocessor enable */
 66#define PSR_EF		0x00001000	/* FP enable */
 67#define PSR_PIL		0x00000f00	/* interrupt level */
 68#define PSR_S		0x00000080	/* supervisor (kernel) mode */
 69#define PSR_PS		0x00000040	/* previous supervisor mode (traps) */
 70#define PSR_ET		0x00000020	/* trap enable */
 71#define PSR_CWP		0x0000001f	/* current window pointer */
 72
 73#define PSR_BITS "\20\16EC\15EF\10S\7PS\6ET"
 74
 75/* Interesting spl()s */
 76#define PIL_BIO		5
 77#define PIL_VIDEO	5
 78#define PIL_TTY		6
 79#define PIL_LPT		6
 80#define PIL_NET		6
 81#define PIL_VM		7
 82#define	PIL_AUD		8
 83#define PIL_CLOCK	10
 84#define PIL_FD		11
 85#define PIL_SER		12
 86#define	PIL_STATCLOCK	14
 87#define PIL_HIGH	15
 88#define PIL_SCHED	PIL_CLOCK
 89#define PIL_LOCK	PIL_HIGH
 90
 91/* 
 92 * SPARC V9 CCR register
 93 */
 94
 95#define ICC_C	0x01L
 96#define ICC_V	0x02L
 97#define ICC_Z	0x04L
 98#define ICC_N	0x08L
 99#define XCC_SHIFT	4
100#define XCC_C	(ICC_C<<XCC_SHIFT)
101#define XCC_V	(ICC_V<<XCC_SHIFT)
102#define XCC_Z	(ICC_Z<<XCC_SHIFT)
103#define XCC_N	(ICC_N<<XCC_SHIFT)
104
105
106/*
107 * SPARC V9 PSTATE register (what replaces the PSR in V9)
108 *
109 * Here's the layout:
110 *
111 *    11   10    9     8   7  6   5     4     3     2     1   0
112 *  +------------------------------------------------------------+
113 *  | IG | MG | CLE | TLE | MM | RED | PEF | AM | PRIV | IE | AG |
114 *  +------------------------------------------------------------+
115 */
116
117#define PSTATE_IG	0x800	/* enable spitfire interrupt globals */
118#define PSTATE_MG	0x400	/* enable spitfire MMU globals */
119#define PSTATE_CLE	0x200	/* current little endian */
120#define PSTATE_TLE	0x100	/* traps little endian */
121#define PSTATE_MM	0x0c0	/* memory model */
122#define PSTATE_MM_TSO	0x000	/* total store order */
123#define PSTATE_MM_PSO	0x040	/* partial store order */
124#define PSTATE_MM_RMO	0x080	/* Relaxed memory order */
125#define PSTATE_RED	0x020	/* RED state */
126#define PSTATE_PEF	0x010	/* enable floating point */
127#define PSTATE_AM	0x008	/* 32-bit address masking */
128#define PSTATE_PRIV	0x004	/* privileged mode */
129#define PSTATE_IE	0x002	/* interrupt enable */
130#define PSTATE_AG	0x001	/* enable alternate globals */
131
132#define PSTATE_BITS "\20\14IG\13MG\12CLE\11TLE\10\7MM\6RED\5PEF\4AM\3PRIV\2IE\1AG"
133
134
135/*
136 * 32-bit code requires TSO or at best PSO since that's what's supported on
137 * SPARC V8 and earlier machines.
138 *
139 * 64-bit code sets the memory model in the ELF header.
140 *
141 * We're running kernel code in TSO for the moment so we don't need to worry
142 * about possible memory barrier bugs.
143 */
144
145#ifdef __arch64__
146#define PSTATE_PROM	(PSTATE_MM_TSO|PSTATE_PRIV)
147#define PSTATE_NUCLEUS	(PSTATE_MM_TSO|PSTATE_PRIV|PSTATE_AG)
148#define PSTATE_KERN	(PSTATE_MM_TSO|PSTATE_PRIV)
149#define PSTATE_INTR	(PSTATE_KERN|PSTATE_IE)
150#define PSTATE_USER32	(PSTATE_MM_TSO|PSTATE_AM|PSTATE_IE)
151#define PSTATE_USER	(PSTATE_MM_RMO|PSTATE_IE)
152#else
153#define PSTATE_PROM	(PSTATE_MM_TSO|PSTATE_PRIV)
154#define PSTATE_NUCLEUS	(PSTATE_MM_TSO|PSTATE_AM|PSTATE_PRIV|PSTATE_AG)
155#define PSTATE_KERN	(PSTATE_MM_TSO|PSTATE_AM|PSTATE_PRIV)
156#define PSTATE_INTR	(PSTATE_KERN|PSTATE_IE)
157#define PSTATE_USER32	(PSTATE_MM_TSO|PSTATE_AM|PSTATE_IE)
158#define PSTATE_USER	(PSTATE_MM_TSO|PSTATE_AM|PSTATE_IE)
159#endif
160
161
162/*
163 * SPARC V9 TSTATE register
164 *
165 *   39 32 31 24 23 20  19   8	7 5 4   0
166 *  +-----+-----+-----+--------+---+-----+
167 *  | CCR | ASI |  -  | PSTATE | - | CWP |
168 *  +-----+-----+-----+--------+---+-----+
169 */
170
171#define TSTATE_CWP		0x01f
172#define TSTATE_PSTATE		0xfff00
173#define TSTATE_PSTATE_SHIFT	8
174#define TSTATE_ASI		0xff000000LL
175#define TSTATE_ASI_SHIFT	24
176#define TSTATE_CCR		0xff00000000LL
177#define TSTATE_CCR_SHIFT	32
178
179#define PSRCC_TO_TSTATE(x)	(((int64_t)(x)&PSR_ICC)<<(TSTATE_CCR_SHIFT-20))
180#define TSTATECCR_TO_PSR(x)	(((x)&TSTATE_CCR)>>(TSTATE_CCR_SHIFT-20))
181
182/*
183 * These are here to simplify life.
184 */
185#define TSTATE_IG	(PSTATE_IG<<TSTATE_PSTATE_SHIFT)
186#define TSTATE_MG	(PSTATE_MG<<TSTATE_PSTATE_SHIFT)
187#define TSTATE_CLE	(PSTATE_CLE<<TSTATE_PSTATE_SHIFT)
188#define TSTATE_TLE	(PSTATE_TLE<<TSTATE_PSTATE_SHIFT)
189#define TSTATE_MM	(PSTATE_MM<<TSTATE_PSTATE_SHIFT)
190#define TSTATE_MM_TSO	(PSTATE_MM_TSO<<TSTATE_PSTATE_SHIFT)
191#define TSTATE_MM_PSO	(PSTATE_MM_PSO<<TSTATE_PSTATE_SHIFT)
192#define TSTATE_MM_RMO	(PSTATE_MM_RMO<<TSTATE_PSTATE_SHIFT)
193#define TSTATE_RED	(PSTATE_RED<<TSTATE_PSTATE_SHIFT)
194#define TSTATE_PEF	(PSTATE_PEF<<TSTATE_PSTATE_SHIFT)
195#define TSTATE_AM	(PSTATE_AM<<TSTATE_PSTATE_SHIFT)
196#define TSTATE_PRIV	(PSTATE_PRIV<<TSTATE_PSTATE_SHIFT)
197#define TSTATE_IE	(PSTATE_IE<<TSTATE_PSTATE_SHIFT)
198#define TSTATE_AG	(PSTATE_AG<<TSTATE_PSTATE_SHIFT)
199
200#define TSTATE_BITS "\20\14IG\13MG\12CLE\11TLE\10\7MM\6RED\5PEF\4AM\3PRIV\2IE\1AG"
201
202#define TSTATE_KERN	((PSTATE_KERN)<<TSTATE_PSTATE_SHIFT)
203#define TSTATE_USER	((PSTATE_USER)<<TSTATE_PSTATE_SHIFT)
204/*
205 * SPARC V9 VER version register.
206 *
207 *  63   48 47  32 31  24 23 16 15    8 7 5 4      0
208 * +-------+------+------+-----+-------+---+--------+
209 * | manuf | impl | mask |  -  | maxtl | - | maxwin |
210 * +-------+------+------+-----+-------+---+--------+
211 *
212 */
213
214#define VER_MANUF	0xffff000000000000LL
215#define VER_MANUF_SHIFT	48
216#define VER_IMPL	0x0000ffff00000000LL
217#define VER_IMPL_SHIFT	32
218#define VER_MASK	0x00000000ff000000LL
219#define VER_MASK_SHIFT	24
220#define VER_MAXTL	0x000000000000ff00LL
221#define VER_MAXTL_SHIFT	8
222#define VER_MAXWIN	0x000000000000001fLL
223
224#define MANUF_FUJITSU		0x04 /* Fujitsu SPARC64 */
225#define MANUF_SUN		0x17 /* Sun UltraSPARC */
226
227#define IMPL_SPARC64		0x01 /* SPARC64 */
228#define IMPL_SPARC64_II		0x02 /* SPARC64-II */
229#define IMPL_SPARC64_III	0x03 /* SPARC64-III */
230#define IMPL_SPARC64_IV		0x04 /* SPARC64-IV */
231#define IMPL_ZEUS		0x05 /* SPARC64-V */
232#define IMPL_OLYMPUS_C		0x06 /* SPARC64-VI */
233#define IMPL_JUPITER		0x07 /* SPARC64-VII */
234
235#define IMPL_SPITFIRE		0x10 /* UltraSPARC-I */
236#define IMPL_BLACKBIRD		0x11 /* UltraSPARC-II */
237#define IMPL_SABRE		0x12 /* UltraSPARC-IIi */
238#define IMPL_HUMMINGBIRD	0x13 /* UltraSPARC-IIe */
239#define IMPL_CHEETAH		0x14 /* UltraSPARC-III */
240#define IMPL_CHEETAH_PLUS	0x15 /* UltraSPARC-III+ */
241#define IMPL_JALAPENO		0x16 /* UltraSPARC-IIIi */
242#define IMPL_JAGUAR		0x18 /* UltraSPARC-IV */
243#define IMPL_PANTHER		0x19 /* UltraSPARC-IV+ */
244#define IMPL_SERRANO		0x22 /* UltraSPARC-IIIi+ */
245
246/*
247 * Here are a few things to help us transition between user and kernel mode:
248 */
249
250/* Memory models */
251#define KERN_MM		PSTATE_MM_TSO
252#define USER_MM		PSTATE_MM_RMO
253
254/* 
255 * Register window handlers.  These point to generic routines that check the
256 * stack pointer and then vector to the real handler.  We could optimize this
257 * if we could guarantee only 32-bit or 64-bit stacks.
258 */
259#define WSTATE_KERN	026
260#define WSTATE_USER	022
261
262#define CWP		0x01f
263
264/*
265 * UltraSPARC Ancillary State Registers
266 */
267#define SET_SOFTINT	%asr20	/* Set Software Interrupt register bits */
268#define CLEAR_SOFTINT	%asr21	/* Clear Software Interrupt register bits */
269#define SOFTINT		%asr22	/* Software Interrupt register */
270#define TICK_CMPR	%asr23	/* TICK Compare register */
271#define STICK		%asr24	/* STICK register */
272#define STICK_CMPR	%asr25	/* STICK Compare register */
273
274/* SOFTINT bit descriptions */
275#define TICK_INT	0x01		/* CPU clock timer interrupt */
276#define STICK_INT	(0x1<<16)	/* system clock timer interrupt */
277
278/* 64-byte alignment -- this seems the best place to put this. */
279#define SPARC64_BLOCK_SIZE	64
280#define SPARC64_BLOCK_ALIGN	0x3f
281
282
283#if (defined(_KERNEL) || defined(_KMEMUSER)) && !defined(_LOCORE)
284typedef uint8_t ipl_t;
285typedef struct {
286	ipl_t _ipl;
287} ipl_cookie_t;
288#endif /* _KERNEL|_KMEMUSER&!_LOCORE */
289
290#if defined(_KERNEL) && !defined(_LOCORE)
291
292#if defined(_KERNEL_OPT)
293#include "opt_sparc_arch.h"
294#endif
295
296/*
297 * Put "memory" to asm inline on sun4v to avoid issuing rdpr %ver
298 * before checking cputyp as a result of code moving by compiler
299 * optimization.
300 */
301#ifdef SUN4V
302#define constasm_clobbers "memory"
303#else
304#define constasm_clobbers
305#endif
306
307/*
308 * Inlines for manipulating privileged and ancillary state registers
309 */
310#define SPARC64_RDCONST_DEF(rd, name, reg, type)			\
311static __inline __constfunc type get##name(void)			\
312{									\
313	type _val;							\
314	__asm(#rd " %" #reg ",%0" : "=r" (_val) : : constasm_clobbers);	\
315	return _val;							\
316}
317#define SPARC64_RD_DEF(rd, name, reg, type)				\
318static __inline type get##name(void)					\
319{									\
320	type _val;							\
321	__asm volatile(#rd " %" #reg ",%0" : "=r" (_val));		\
322	return _val;							\
323}
324#define SPARC64_WR_DEF(wr, name, reg, type)				\
325static __inline void set##name(type _val)				\
326{									\
327	__asm volatile(#wr " %0,0,%" #reg : : "r" (_val) : "memory");	\
328}
329
330#ifdef __arch64__
331#define SPARC64_RDCONST64_DEF(rd, name, reg) \
332	SPARC64_RDCONST_DEF(rd, name, reg, uint64_t)
333#define SPARC64_RD64_DEF(rd, name, reg) SPARC64_RD_DEF(rd, name, reg, uint64_t)
334#define SPARC64_WR64_DEF(wr, name, reg) SPARC64_WR_DEF(wr, name, reg, uint64_t)
335#else
336#define SPARC64_RDCONST64_DEF(rd, name, reg)				\
337static __inline __constfunc uint64_t get##name(void)			\
338{									\
339	uint32_t _hi, _lo;						\
340	__asm(#rd " %" #reg ",%0; srl %0,0,%1; srlx %0,32,%0"		\
341		: "=r" (_hi), "=r" (_lo) : : constasm_clobbers);	\
342	return ((uint64_t)_hi << 32) | _lo;				\
343}
344#define SPARC64_RD64_DEF(rd, name, reg)					\
345static __inline uint64_t get##name(void)				\
346{									\
347	uint32_t _hi, _lo;						\
348	__asm volatile(#rd " %" #reg ",%0; srl %0,0,%1; srlx %0,32,%0"	\
349		: "=r" (_hi), "=r" (_lo));				\
350	return ((uint64_t)_hi << 32) | _lo;				\
351}
352#define SPARC64_WR64_DEF(wr, name, reg)					\
353static __inline void set##name(uint64_t _val)				\
354{									\
355	uint32_t _hi = _val >> 32, _lo = _val;				\
356	__asm volatile("sllx %1,32,%0; or %0,%2,%0; " #wr " %0,0,%" #reg\
357		       : "=&r" (_hi) /* scratch register */		\
358		       : "r" (_hi), "r" (_lo) : "memory");		\
359}
360#endif
361
362#define SPARC64_RDPR_DEF(name, reg, type) SPARC64_RD_DEF(rdpr, name, reg, type)
363#define SPARC64_WRPR_DEF(name, reg, type) SPARC64_WR_DEF(wrpr, name, reg, type)
364#define SPARC64_RDPR64_DEF(name, reg)	SPARC64_RD64_DEF(rdpr, name, reg)
365#define SPARC64_WRPR64_DEF(name, reg)	SPARC64_WR64_DEF(wrpr, name, reg)
366#define SPARC64_RDASR64_DEF(name, reg)	SPARC64_RD64_DEF(rd, name, reg)
367#define SPARC64_WRASR64_DEF(name, reg)	SPARC64_WR64_DEF(wr, name, reg)
368
369/* Tick Register (PR 4) */
370SPARC64_RDPR64_DEF(tick, %tick)			/* gettick() */
371SPARC64_WRPR64_DEF(tick, %tick)			/* settick() */
372
373/* Processor State Register (PR 6) */
374SPARC64_RDPR_DEF(pstate, %pstate, int)		/* getpstate() */
375SPARC64_WRPR_DEF(pstate, %pstate, int)		/* setpstate() */
376
377/* Trap Level Register (PR 7) */
378SPARC64_RDPR_DEF(tl, %tl, int)			/* gettl() */
379
380/* Current Window Pointer Register (PR 9) */
381SPARC64_RDPR_DEF(cwp, %cwp, int)		/* getcwp() */
382SPARC64_WRPR_DEF(cwp, %cwp, int)		/* setcwp() */
383
384/* Version Register (PR 31) */
385SPARC64_RDCONST64_DEF(rdpr, ver, %ver)		/* getver() */
386
387/* System Tick Register (ASR 24) */
388SPARC64_RDASR64_DEF(stick, STICK)		/* getstick() */
389SPARC64_WRASR64_DEF(stick, STICK)		/* setstick() */
390
391/* System Tick Compare Register (ASR 25) */
392SPARC64_RDASR64_DEF(stickcmpr, STICK_CMPR)	/* getstickcmpr() */
393
394/* Some simple macros to check the cpu type. */
395#define GETVER_CPU_MASK()	((getver() & VER_MASK) >> VER_MASK_SHIFT)
396#define GETVER_CPU_IMPL()	((getver() & VER_IMPL) >> VER_IMPL_SHIFT)
397#define GETVER_CPU_MANUF()	((getver() & VER_MANUF) >> VER_MANUF_SHIFT)
398#define CPU_IS_SPITFIRE()	(GETVER_CPU_IMPL() == IMPL_SPITFIRE)
399#define CPU_IS_HUMMINGBIRD()	(GETVER_CPU_IMPL() == IMPL_HUMMINGBIRD)
400#define CPU_IS_USIIIi()		((GETVER_CPU_IMPL() == IMPL_JALAPENO) || \
401				 (GETVER_CPU_IMPL() == IMPL_SERRANO))
402#define CPU_IS_USIII_UP()	(GETVER_CPU_IMPL() >= IMPL_CHEETAH)
403#define CPU_IS_SPARC64_V_UP()	(GETVER_CPU_MANUF() == MANUF_FUJITSU && \
404				 GETVER_CPU_IMPL() >= IMPL_ZEUS)
405
406static __inline int
407intr_disable(void)
408{
409	int pstate = getpstate();
410
411	setpstate(pstate & ~PSTATE_IE);
412	return pstate;
413}
414
415static __inline void
416intr_restore(int pstate)
417{
418	setpstate(pstate);
419}
420
421/*
422 * GCC pseudo-functions for manipulating PIL
423 */
424
425#ifdef SPLDEBUG
426void prom_printf(const char *fmt, ...);
427extern int printspl;
428#define SPLPRINT(x) \
429{ \
430	if (printspl) { \
431		int i = 10000000; \
432		prom_printf x ; \
433		while (i--) \
434			; \
435	} \
436}
437#define	SPL(name, newpil) \
438static __inline int name##X(const char* file, int line) \
439{ \
440	int oldpil; \
441	__asm volatile("rdpr %%pil,%0" : "=r" (oldpil)); \
442	SPLPRINT(("{%s:%d %d=>%d}", file, line, oldpil, newpil)); \
443	__asm volatile("wrpr %%g0,%0,%%pil" : : "n" (newpil) : "memory"); \
444	return (oldpil); \
445}
446/* A non-priority-decreasing version of SPL */
447#define	SPLHOLD(name, newpil) \
448static __inline int name##X(const char* file, int line) \
449{ \
450	int oldpil; \
451	__asm volatile("rdpr %%pil,%0" : "=r" (oldpil)); \
452	if (newpil <= oldpil) \
453		return oldpil; \
454	SPLPRINT(("{%s:%d %d->!d}", file, line, oldpil, newpil)); \
455	__asm volatile("wrpr %%g0,%0,%%pil" : : "n" (newpil) : "memory"); \
456	return (oldpil); \
457}
458
459#else
460#define SPLPRINT(x)	
461#define	SPL(name, newpil) \
462static __inline __always_inline int name(void) \
463{ \
464	int oldpil; \
465	__asm volatile("rdpr %%pil,%0" : "=r" (oldpil)); \
466	__asm volatile("wrpr %%g0,%0,%%pil" : : "n" (newpil) : "memory"); \
467	return (oldpil); \
468}
469/* A non-priority-decreasing version of SPL */
470#define	SPLHOLD(name, newpil) \
471static __inline __always_inline int name(void) \
472{ \
473	int oldpil; \
474	__asm volatile("rdpr %%pil,%0" : "=r" (oldpil)); \
475	if (newpil <= oldpil) \
476		return oldpil; \
477	__asm volatile("wrpr %%g0,%0,%%pil" : : "n" (newpil) : "memory"); \
478	return (oldpil); \
479}
480#endif
481
482static __inline ipl_cookie_t
483makeiplcookie(ipl_t ipl)
484{
485
486	return (ipl_cookie_t){._ipl = ipl};
487}
488
489static __inline int __attribute__((__unused__))
490splraiseipl(ipl_cookie_t icookie)
491{
492	int newpil = icookie._ipl;
493	int oldpil;
494
495	/*
496	 * NetBSD/sparc64's IPL_* constants equate directly to the
497	 * corresponding PIL_* names; no need to map them here.
498	 */
499	__asm volatile("rdpr %%pil,%0" : "=r" (oldpil));
500	if (newpil <= oldpil)
501		return (oldpil);
502	__asm volatile("wrpr %0,0,%%pil" : : "r" (newpil) : "memory");
503	return (oldpil);
504}
505
506SPL(spl0, 0)
507
508SPLHOLD(splsoftint, 1)
509#define	splsoftclock	splsoftint
510#define	splsoftnet	splsoftint
511
512SPLHOLD(splsoftserial, 4)
513
514/*
515 * Memory allocation (must be as high as highest network, tty, or disk device)
516 */
517SPLHOLD(splvm, PIL_VM)
518
519SPLHOLD(splsched, PIL_SCHED)
520
521SPLHOLD(splhigh, PIL_HIGH)
522
523/* splx does not have a return value */
524#ifdef SPLDEBUG
525#define	spl0()	spl0X(__FILE__, __LINE__)
526#define	splsoftint()	splsoftintX(__FILE__, __LINE__)
527#define	splsoftserial()	splsoftserialX(__FILE__, __LINE__)
528#define	splausoft()	splausoftX(__FILE__, __LINE__)
529#define	splfdsoft()	splfdsoftX(__FILE__, __LINE__)
530#define	splvm()		splvmX(__FILE__, __LINE__)
531#define	splclock()	splclockX(__FILE__, __LINE__)
532#define	splfd()		splfdX(__FILE__, __LINE__)
533#define	splzs()		splzsX(__FILE__, __LINE__)
534#define	splserial()	splzerialX(__FILE__, __LINE__)
535#define	splaudio()	splaudioX(__FILE__, __LINE__)
536#define	splstatclock()	splstatclockX(__FILE__, __LINE__)
537#define	splsched()	splschedX(__FILE__, __LINE__)
538#define	spllock()	spllockX(__FILE__, __LINE__)
539#define	splhigh()	splhighX(__FILE__, __LINE__)
540#define splx(x)		splxX((x),__FILE__, __LINE__)
541
542static __inline void splxX(int newpil, const char *file, int line)
543#else
544static __inline __always_inline void splx(int newpil)
545#endif
546{
547#ifdef SPLDEBUG
548	int pil;
549
550	__asm volatile("rdpr %%pil,%0" : "=r" (pil));
551	SPLPRINT(("{%d->%d}", pil, newpil));
552#endif
553	__asm volatile("wrpr %%g0,%0,%%pil" : : "rn" (newpil) : "memory");
554}
555#endif /* KERNEL && !_LOCORE */
556
557#endif /* PSR_IMPL */