1/*-
  2 * SPDX-License-Identifier: BSD-3-Clause AND BSD-4-Clause
  3 *
  4 * Copyright (C) 2006 Semihalf, Marian Balakowicz <m8@semihalf.com>
  5 * All rights reserved.
  6 *
  7 * Adapted for Freescale's e500 core CPUs.
  8 *
  9 * Redistribution and use in source and binary forms, with or without
 10 * modification, are permitted provided that the following conditions
 11 * are met:
 12 * 1. Redistributions of source code must retain the above copyright
 13 *    notice, this list of conditions and the following disclaimer.
 14 * 2. Redistributions in binary form must reproduce the above copyright
 15 *    notice, this list of conditions and the following disclaimer in the
 16 *    documentation and/or other materials provided with the distribution.
 17 * 3. The name of the author may not be used to endorse or promote products
 18 *    derived from this software without specific prior written permission.
 19 *
 20 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR
 21 * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
 22 * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.  IN
 23 * NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
 24 * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED
 25 * TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR
 26 * PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF
 27 * LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING
 28 * NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
 29 * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
 30 */
 31/*-
 32 * Copyright (C) 1995, 1996 Wolfgang Solfrank.
 33 * Copyright (C) 1995, 1996 TooLs GmbH.
 34 * All rights reserved.
 35 *
 36 * Redistribution and use in source and binary forms, with or without
 37 * modification, are permitted provided that the following conditions
 38 * are met:
 39 * 1. Redistributions of source code must retain the above copyright
 40 *    notice, this list of conditions and the following disclaimer.
 41 * 2. Redistributions in binary form must reproduce the above copyright
 42 *    notice, this list of conditions and the following disclaimer in the
 43 *    documentation and/or other materials provided with the distribution.
 44 * 3. All advertising materials mentioning features or use of this software
 45 *    must display the following acknowledgement:
 46 *	This product includes software developed by TooLs GmbH.
 47 * 4. The name of TooLs GmbH may not be used to endorse or promote products
 48 *    derived from this software without specific prior written permission.
 49 *
 50 * THIS SOFTWARE IS PROVIDED BY TOOLS GMBH ``AS IS'' AND ANY EXPRESS OR
 51 * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
 52 * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
 53 * IN NO EVENT SHALL TOOLS GMBH BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
 54 * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO,
 55 * PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS;
 56 * OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY,
 57 * WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR
 58 * OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF
 59 * ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
 60 *
 61 *	from: $NetBSD: pmap.h,v 1.17 2000/03/30 16:18:24 jdolecek Exp $
 62 */
 63
 64#ifndef	_MACHINE_PMAP_H_
 65#define	_MACHINE_PMAP_H_
 66
 67#include <sys/queue.h>
 68#include <sys/tree.h>
 69#include <sys/_cpuset.h>
 70#include <sys/_lock.h>
 71#include <sys/_mutex.h>
 72#include <machine/sr.h>
 73#include <machine/pte.h>
 74#include <machine/slb.h>
 75#include <machine/tlb.h>
 76#include <machine/vmparam.h>
 77#ifdef __powerpc64__
 78#include <vm/vm_radix.h>
 79#endif
 80
 81/*
 82 * The radix page table structure is described by levels 1-4.
 83 * See Fig 33. on p. 1002 of Power ISA v3.0B
 84 *
 85 * Page directories and tables must be size aligned.
 86 */
 87
 88/* Root page directory - 64k   -- each entry covers 512GB */
 89typedef uint64_t pml1_entry_t;
 90/* l2 page directory - 4k      -- each entry covers 1GB */
 91typedef uint64_t pml2_entry_t;
 92/* l3 page directory - 4k      -- each entry covers 2MB */
 93typedef uint64_t pml3_entry_t;
 94/* l4 page directory - 256B/4k -- each entry covers 64k/4k */
 95typedef uint64_t pml4_entry_t;
 96
 97typedef uint64_t pt_entry_t;
 98
 99struct pmap;
100typedef struct pmap *pmap_t;
101
102#define	PMAP_ENTER_QUICK_LOCKED	0x10000000
103
104#if !defined(NPMAPS)
105#define	NPMAPS		32768
106#endif /* !defined(NPMAPS) */
107
108struct	slbtnode;
109
110struct pvo_entry {
111	LIST_ENTRY(pvo_entry) pvo_vlink;	/* Link to common virt page */
112#ifndef __powerpc64__
113	LIST_ENTRY(pvo_entry) pvo_olink;	/* Link to overflow entry */
114#endif
115	union {
116		RB_ENTRY(pvo_entry) pvo_plink;	/* Link to pmap entries */
117		SLIST_ENTRY(pvo_entry) pvo_dlink; /* Link to delete enty */
118	};
119	struct {
120#ifndef __powerpc64__
121		/* 32-bit fields */
122		pte_t	    pte;
123#endif
124		/* 64-bit fields */
125		uintptr_t   slot;
126		vm_paddr_t  pa;
127		vm_prot_t   prot;
128	} pvo_pte;
129	pmap_t		pvo_pmap;		/* Owning pmap */
130	vm_offset_t	pvo_vaddr;		/* VA of entry */
131	uint64_t	pvo_vpn;		/* Virtual page number */
132};
133LIST_HEAD(pvo_head, pvo_entry);
134SLIST_HEAD(pvo_dlist, pvo_entry);
135RB_HEAD(pvo_tree, pvo_entry);
136int pvo_vaddr_compare(struct pvo_entry *, struct pvo_entry *);
137RB_PROTOTYPE(pvo_tree, pvo_entry, pvo_plink, pvo_vaddr_compare);
138
139/* Used by 32-bit PMAP */
140#define	PVO_PTEGIDX_MASK	0x007UL		/* which PTEG slot */
141#define	PVO_PTEGIDX_VALID	0x008UL		/* slot is valid */
142/* Used by 64-bit PMAP */
143#define	PVO_HID			0x008UL		/* PVO entry in alternate hash*/
144/* Used by both */
145#define	PVO_WIRED		0x010UL		/* PVO entry is wired */
146#define	PVO_MANAGED		0x020UL		/* PVO entry is managed */
147#define	PVO_BOOTSTRAP		0x080UL		/* PVO entry allocated during
148						   bootstrap */
149#define	PVO_DEAD		0x100UL		/* waiting to be deleted */
150#define	PVO_LARGE		0x200UL		/* large page */
151#define	PVO_VADDR(pvo)		((pvo)->pvo_vaddr & ~ADDR_POFF)
152#define	PVO_PTEGIDX_GET(pvo)	((pvo)->pvo_vaddr & PVO_PTEGIDX_MASK)
153#define	PVO_PTEGIDX_ISSET(pvo)	((pvo)->pvo_vaddr & PVO_PTEGIDX_VALID)
154#define	PVO_PTEGIDX_CLR(pvo)	\
155	((void)((pvo)->pvo_vaddr &= ~(PVO_PTEGIDX_VALID|PVO_PTEGIDX_MASK)))
156#define	PVO_PTEGIDX_SET(pvo, i)	\
157	((void)((pvo)->pvo_vaddr |= (i)|PVO_PTEGIDX_VALID))
158#define	PVO_VSID(pvo)		((pvo)->pvo_vpn >> 16)
159
160struct	pmap {
161	struct		pmap_statistics	pm_stats;
162	struct	mtx	pm_mtx;
163	cpuset_t	pm_active;
164	union {
165		struct {
166		    #ifdef __powerpc64__
167			struct slbtnode	*pm_slb_tree_root;
168			struct slb	**pm_slb;
169			int		pm_slb_len;
170		    #else
171			register_t	pm_sr[16];
172		    #endif
173
174			struct pmap	*pmap_phys;
175			struct pvo_tree pmap_pvo;
176		};
177#ifdef __powerpc64__
178		/* Radix support */
179		struct {
180			pml1_entry_t	*pm_pml1;	/* KVA of root page directory */
181			struct vm_radix	 pm_radix;	/* spare page table pages */
182			TAILQ_HEAD(,pv_chunk)	pm_pvchunk;	/* list of mappings in pmap */
183			uint64_t	pm_pid; /* PIDR value */
184			int pm_flags;
185		};
186#endif
187		struct {
188			/* TID to identify this pmap entries in TLB */
189			tlbtid_t	pm_tid[MAXCPU];
190
191#ifdef __powerpc64__
192			/*
193			 * Page table directory,
194			 * array of pointers to page directories.
195			 */
196			pte_t ****pm_root;
197#else
198			/*
199			 * Page table directory,
200			 * array of pointers to page tables.
201			 */
202			pte_t		**pm_pdir;
203
204			/* List of allocated ptbl bufs (ptbl kva regions). */
205			TAILQ_HEAD(, ptbl_buf)	pm_ptbl_list;
206#endif
207		};
208	} __aligned(CACHE_LINE_SIZE);
209};
210
211/*
212 * pv_entries are allocated in chunks per-process.  This avoids the
213 * need to track per-pmap assignments.
214 */
215#define	_NPCPV	126
216#define	_NPCM	howmany(_NPCPV, 64)
217
218#define	PV_CHUNK_HEADER							\
219	pmap_t			pc_pmap;				\
220	TAILQ_ENTRY(pv_chunk)	pc_list;				\
221	uint64_t		pc_map[_NPCM];	/* bitmap; 1 = free */	\
222	TAILQ_ENTRY(pv_chunk)	pc_lru;
223
224struct pv_entry {
225	pmap_t pv_pmap;
226	vm_offset_t pv_va;
227	TAILQ_ENTRY(pv_entry) pv_link;
228};
229typedef struct pv_entry *pv_entry_t;
230
231struct pv_chunk_header {
232	PV_CHUNK_HEADER
233};
234struct pv_chunk {
235	PV_CHUNK_HEADER
236	uint64_t	reserved;
237	struct pv_entry		pc_pventry[_NPCPV];
238};
239
240struct	md_page {
241	union {
242		struct {
243			volatile int32_t mdpg_attrs;
244			vm_memattr_t	 mdpg_cache_attrs;
245			struct	pvo_head mdpg_pvoh;
246			int		pv_gen;   /* (p) */
247		};
248		struct {
249			int			pv_tracked;
250		};
251	};
252	TAILQ_HEAD(, pv_entry)	pv_list;  /* (p) */
253};
254
255#ifdef AIM
256#define	pmap_page_get_memattr(m)	((m)->md.mdpg_cache_attrs)
257#else
258#define	pmap_page_get_memattr(m)	VM_MEMATTR_DEFAULT
259#endif /* AIM */
260
261/*
262 * Return the VSID corresponding to a given virtual address.
263 * If no VSID is currently defined, it will allocate one, and add
264 * it to a free slot if available.
265 *
266 * NB: The PMAP MUST be locked already.
267 */
268uint64_t va_to_vsid(pmap_t pm, vm_offset_t va);
269
270/* Lock-free, non-allocating lookup routines */
271uint64_t kernel_va_to_slbv(vm_offset_t va);
272struct slb *user_va_to_slb_entry(pmap_t pm, vm_offset_t va);
273
274uint64_t allocate_user_vsid(pmap_t pm, uint64_t esid, int large);
275void	free_vsid(pmap_t pm, uint64_t esid, int large);
276void	slb_insert_user(pmap_t pm, struct slb *slb);
277void	slb_insert_kernel(uint64_t slbe, uint64_t slbv);
278
279struct slbtnode *slb_alloc_tree(void);
280void     slb_free_tree(pmap_t pm);
281struct slb **slb_alloc_user_cache(void);
282void	slb_free_user_cache(struct slb **);
283
284extern	struct pmap kernel_pmap_store;
285#define	kernel_pmap	(&kernel_pmap_store)
286
287#ifdef _KERNEL
288
289#define	PMAP_LOCK(pmap)		mtx_lock(&(pmap)->pm_mtx)
290#define	PMAP_LOCK_ASSERT(pmap, type) \
291				mtx_assert(&(pmap)->pm_mtx, (type))
292#define	PMAP_LOCK_DESTROY(pmap)	mtx_destroy(&(pmap)->pm_mtx)
293#define	PMAP_LOCK_INIT(pmap)	mtx_init(&(pmap)->pm_mtx, \
294				    (pmap == kernel_pmap) ? "kernelpmap" : \
295				    "pmap", NULL, MTX_DEF | MTX_DUPOK)
296#define	PMAP_LOCKED(pmap)	mtx_owned(&(pmap)->pm_mtx)
297#define	PMAP_MTX(pmap)		(&(pmap)->pm_mtx)
298#define	PMAP_TRYLOCK(pmap)	mtx_trylock(&(pmap)->pm_mtx)
299#define	PMAP_UNLOCK(pmap)	mtx_unlock(&(pmap)->pm_mtx)
300
301#define	pmap_page_is_write_mapped(m)	(((m)->a.flags & PGA_WRITEABLE) != 0)
302
303#define	pmap_vm_page_alloc_check(m)
304
305void		pmap_bootstrap(vm_offset_t, vm_offset_t);
306void		pmap_kenter(vm_offset_t va, vm_paddr_t pa);
307void		pmap_kenter_attr(vm_offset_t va, vm_paddr_t pa, vm_memattr_t);
308void		pmap_kremove(vm_offset_t);
309void		*pmap_mapdev(vm_paddr_t, vm_size_t);
310void		*pmap_mapdev_attr(vm_paddr_t, vm_size_t, vm_memattr_t);
311void		pmap_unmapdev(void *, vm_size_t);
312void		pmap_page_set_memattr(vm_page_t, vm_memattr_t);
313int		pmap_change_attr(vm_offset_t, vm_size_t, vm_memattr_t);
314int		pmap_map_user_ptr(pmap_t pm, volatile const void *uaddr,
315		    void **kaddr, size_t ulen, size_t *klen);
316int		pmap_decode_kernel_ptr(vm_offset_t addr, int *is_user,
317		    vm_offset_t *decoded_addr);
318void		pmap_deactivate(struct thread *);
319vm_paddr_t	pmap_kextract(vm_offset_t);
320int		pmap_dev_direct_mapped(vm_paddr_t, vm_size_t);
321boolean_t	pmap_mmu_install(char *name, int prio);
322void		pmap_mmu_init(void);
323const char	*pmap_mmu_name(void);
324bool		pmap_ps_enabled(pmap_t pmap);
325int		pmap_nofault(pmap_t pmap, vm_offset_t va, vm_prot_t flags);
326boolean_t	pmap_page_is_mapped(vm_page_t m);
327#define	pmap_map_delete(pmap, sva, eva)	pmap_remove(pmap, sva, eva)
328
329void		pmap_page_array_startup(long count);
330
331#define	vtophys(va)	pmap_kextract((vm_offset_t)(va))
332
333extern	vm_offset_t virtual_avail;
334extern	vm_offset_t virtual_end;
335extern	caddr_t crashdumpmap;
336
337extern	vm_offset_t msgbuf_phys;
338
339extern	int pmap_bootstrapped;
340extern	int radix_mmu;
341extern	int superpages_enabled;
342
343#ifdef AIM
344void pmap_early_io_map_init(void);
345#endif
346vm_offset_t pmap_early_io_map(vm_paddr_t pa, vm_size_t size);
347void pmap_early_io_unmap(vm_offset_t va, vm_size_t size);
348void pmap_track_page(pmap_t pmap, vm_offset_t va);
349void pmap_page_print_mappings(vm_page_t m);
350void pmap_tlbie_all(void);
351
352static inline int
353pmap_vmspace_copy(pmap_t dst_pmap __unused, pmap_t src_pmap __unused)
354{
355
356	return (0);
357}
358
359#endif
360
361#endif /* !_MACHINE_PMAP_H_ */