Skip site navigation (1)Skip section navigation (2)
Date:      Sat, 2 Nov 2019 21:15:56 +0000 (UTC)
From:      Brandon Bergren <bdragon@FreeBSD.org>
To:        src-committers@freebsd.org, svn-src-all@freebsd.org, svn-src-head@freebsd.org
Subject:   svn commit: r354266 - in head/sys: conf powerpc/booke powerpc/conf/dpaa powerpc/include powerpc/mpc85xx powerpc/powerpc
Message-ID:  <201911022115.xA2LFuWi004298@repo.freebsd.org>

next in thread | raw e-mail | index | archive | help
Author: bdragon
Date: Sat Nov  2 21:15:56 2019
New Revision: 354266
URL: https://svnweb.freebsd.org/changeset/base/354266

Log:
  Add support for building Book-E kernels with clang/lld.
  
  This involved several changes:
  
  * Since lld does not like text relocations, replace SMP boot page text relocs
  in booke/locore.S with position-independent math, and track the virtual base
  in the SMP boot page header.
  
  * As some SPRs are interpreted differently on clang due to the way it handles
  platform-specific SPRs, switch m*dear and m*esr mnemonics out for regular
  m*spr. Add both forms of SPR_DEAR to spr.h so the correct encoding is selected.
  
  * Change some hardcoded 32 bit things in the boot page to be pointer-sized, and
  fix alignment.
  
  * Fix 64-bit build of booke/pmap.c when enabling pmap debugging.
  
  Additionally, I took the opportunity to document how the SMP boot page works.
  
  Approved by: jhibbits (mentor)
  Differential Revision: https://reviews.freebsd.org/D21999

Modified:
  head/sys/conf/Makefile.powerpc
  head/sys/powerpc/booke/locore.S
  head/sys/powerpc/booke/pmap.c
  head/sys/powerpc/booke/trap_subr.S
  head/sys/powerpc/conf/dpaa/config.dpaa
  head/sys/powerpc/include/spr.h
  head/sys/powerpc/mpc85xx/platform_mpc85xx.c
  head/sys/powerpc/powerpc/pmap_dispatch.c

Modified: head/sys/conf/Makefile.powerpc
==============================================================================
--- head/sys/conf/Makefile.powerpc	Sat Nov  2 21:02:54 2019	(r354265)
+++ head/sys/conf/Makefile.powerpc	Sat Nov  2 21:15:56 2019	(r354266)
@@ -42,6 +42,20 @@ CFLAGS+= -mabi=spe -D__SPE__
 CFLAGS+= -msoft-float
 CFLAGS.gcc+= -Wa,-many
 
+# Apply compiler-specific DPAA exceptions.
+.if "${COMPILER_TYPE}" == "clang"
+DPAAWARNFLAGS += \
+        -Wno-error=parentheses-equality \
+        -Wno-error=self-assign \
+        -Wno-error=incompatible-pointer-types-discards-qualifiers \
+        -Wno-error=non-literal-null-conversion \
+        -Wno-error=enum-conversion
+.elif "${COMPILER_TYPE}" == "gcc" && ${COMPILER_VERSION} >= 50200
+DPAAWARNFLAGS += \
+	-Wno-error=redundant-decls \
+	-Wno-error=int-in-bool-context
+.endif
+
 # Build position-independent kernel
 CFLAGS+= -fPIC
 LDFLAGS+= -pie

Modified: head/sys/powerpc/booke/locore.S
==============================================================================
--- head/sys/powerpc/booke/locore.S	Sat Nov  2 21:02:54 2019	(r354265)
+++ head/sys/powerpc/booke/locore.S	Sat Nov  2 21:15:56 2019	(r354266)
@@ -39,6 +39,10 @@
 #include <machine/vmparam.h>
 #include <machine/tlb.h>
 
+#ifdef _CALL_ELF
+.abiversion _CALL_ELF
+#endif
+
 #define TMPSTACKSZ	16384
 
 #ifdef __powerpc64__
@@ -76,6 +80,12 @@
 #define	WORD_SIZE	4
 #endif
 
+#ifdef __powerpc64__
+	/* Placate lld by creating a kboot stub. */
+        .section ".text.kboot", "x", @progbits
+        b __start
+#endif
+
 	.text
 	.globl	btext
 btext:
@@ -309,6 +319,9 @@ done_mapping:
 1:	mflr	%r3
 	ld	%r1,0(%r3)
 	add	%r1,%r1,%r3
+/*
+ * Relocate kernel
+ */
 	bl	1f
 	.llong _DYNAMIC-.
 1:	mflr	%r3
@@ -379,16 +392,63 @@ done_mapping:
 	.globl	__boot_page
 	.align	12
 __boot_page:
-	bl	1f
+	/*
+	 * The boot page is a special page of memory used during AP bringup.
+	 * Before the AP comes out of reset, the physical 4K page holding this
+	 * code is arranged to be mapped at 0xfffff000 by use of
+	 * platform-dependent registers.
+	 *
+	 * Alternatively, this page may be executed using an ePAPR-standardized
+	 * method -- writing to the address specified in "cpu-release-addr".
+	 *
+	 * In either case, execution begins at the last instruction of the
+	 * page, which is a branch back to the start of the page.
+	 *
+	 * The code in the page must do initial MMU setup and normalize the
+	 * TLBs for regular operation in the correct address space before
+	 * reading outside the page.
+	 *
+	 * This implementation accomplishes this by:
+	 * 1) Wiping TLB0 and all TLB1 entries but the one currently in use.
+	 * 2) Establishing a temporary 4K TLB1 mapping in AS=1, and switching
+	 *    to it with rfi. This entry must NOT be in TLB1 slot 0.
+	 *    (This is needed to give the code freedom to clean up AS=0.)
+	 * 3) Removing the initial TLB1 entry, leaving us with a single valid
+	 *    TLB1 entry, NOT in slot 0.
+	 * 4) Installing an AS0 entry in TLB1 slot 0 mapping the 64MB kernel
+	 *    segment at its final virtual address. A second rfi is done to
+	 *    switch to the final address space. At this point we can finally
+	 *    access the rest of the kernel segment safely.
+	 * 5) The temporary TLB1 AS=1 entry is removed, finally leaving us in
+	 *    a consistent (but minimal) state.
+	 * 6) Set up TOC, stack, and pcpu registers.
+	 * 7) Now that we can finally call C code, call pmap_boostrap_ap(),
+	 *    which finishes copying in the shared TLB1 entries.
+	 *
+	 * At this point, the MMU is fully set up, and we can proceed with
+	 * running the actual AP bootstrap code.
+	 *
+	 * Pieces of this code are also used for UP kernel, but in this case
+	 * the sections specific to boot page functionality are dropped by
+	 * the preprocessor.
+	 */
+#ifdef __powerpc64__
+	nop			/* PPC64 alignment word. 64-bit target. */
+#endif
+	bl	1f		/* 32-bit target. */
 
 	.globl	bp_trace
 bp_trace:
-	.long	0
+	ADDR(0)			/* Trace pointer (%r31). */
 
 	.globl	bp_kernload
 bp_kernload:
-	.long	0
+	ADDR(0)			/* Kern phys. load address. */
 
+	.globl	bp_virtaddr
+bp_virtaddr:
+	ADDR(0)			/* Virt. address of __boot_page. */
+
 /*
  * Initial configuration
  */
@@ -444,7 +504,7 @@ bp_kernload:
 	mfmsr	%r3
 	ori	%r3, %r3, (PSL_IS | PSL_DS)
 #ifdef __powerpc64__
-	oris	%r3, %r3, PSL_CM@h
+	oris	%r3, %r3, PSL_CM@h	/* Ensure we're in 64-bit after RFI */
 #endif
 	bl	3f
 3:	mflr	%r4
@@ -461,7 +521,7 @@ bp_kernload:
 	bl	tlb1_inval_entry
 
 /*
- * Setup final mapping in TLB1[1] and switch to it
+ * Setup final mapping in TLB1[0] and switch to it
  */
 	/* Final kernel mapping, map in 64 MB of RAM */
 	lis	%r3, MAS0_TLBSEL1@h	/* Select TLB1 */
@@ -481,31 +541,19 @@ bp_kernload:
 	isync
 
 	/* Retrieve kernel load [physical] address from bp_kernload */
+5:
+	mflr	%r3
 #ifdef __powerpc64__
-	b	0f
-	.align	3
-0:
-	nop
-#endif
-	bl 5f
-	ADDR(bp_kernload)
-	ADDR(__boot_page)
-5:	mflr	%r3
-#ifdef __powerpc64__
-	ld	%r4, 0(%r3)
-	ld	%r5, 8(%r3)
-	clrrdi	%r3, %r3, 12
+	clrrdi	%r3, %r3, PAGE_SHIFT	/* trunc_page(%r3) */
 #else
-	lwz	%r4, 0(%r3)
-	lwz	%r5, 4(%r3)
-	rlwinm	%r3, %r3, 0, 0, 19
+	clrrwi	%r3, %r3, PAGE_SHIFT	/* trunc_page(%r3) */
 #endif
-	sub	%r4, %r4, %r5	/* offset of bp_kernload within __boot_page */
-	lwzx	%r3, %r4, %r3
+	LOAD	%r4, (bp_kernload - __boot_page)(%r3)
+	LOAD	%r5, (bp_virtaddr - __boot_page)(%r3)
 
 	/* Set RPN and protection */
-	ori	%r3, %r3, (MAS3_SX | MAS3_SW | MAS3_SR)@l
-	mtspr	SPR_MAS3, %r3
+	ori	%r4, %r4, (MAS3_SX | MAS3_SW | MAS3_SR)@l
+	mtspr	SPR_MAS3, %r4
 	isync
 	li	%r4, 0
 	mtspr	SPR_MAS7, %r4
@@ -518,8 +566,8 @@ bp_kernload:
 	bl	6f
 6:	mflr	%r3
 	rlwinm	%r3, %r3, 0, 0xfff	/* Offset from boot page start */
-	add	%r3, %r3, %r5		/* Make this virtual address */
-	addi	%r3, %r3, (7f - 6b)
+	add	%r3, %r3, %r5		/* Make this a virtual address */
+	addi	%r3, %r3, (7f - 6b)	/* And figure out return address. */
 #ifdef __powerpc64__
 	lis	%r4, PSL_CM@h		/* Note AS=0 */
 #else
@@ -758,6 +806,11 @@ __boot_page_padding:
 	 */
 	.space	4092 - (__boot_page_padding - __boot_page)
 	b	__boot_page
+	/*
+	 * This is the end of the boot page.
+	 * During AP startup, the previous instruction is at 0xfffffffc
+	 * virtual (i.e. the reset vector.)
+	 */
 #endif /* SMP */
 
 /************************************************************************/
@@ -881,6 +934,7 @@ ENTRY(bpred_enable)
  * created.
  */
 ENTRY(get_spr)
+	/* Note: The spr number is patched at runtime */
 	mfspr	%r3, 0
 	blr
 
@@ -900,7 +954,9 @@ tmpstackbound:
 	.space 10240	/* XXX: this really should not be necessary */
 #ifdef __powerpc64__
 TOC_ENTRY(tmpstack)
+#ifdef SMP
 TOC_ENTRY(bp_kernload)
+#endif
 #endif
 
 /*

Modified: head/sys/powerpc/booke/pmap.c
==============================================================================
--- head/sys/powerpc/booke/pmap.c	Sat Nov  2 21:02:54 2019	(r354265)
+++ head/sys/powerpc/booke/pmap.c	Sat Nov  2 21:15:56 2019	(r354266)
@@ -1674,7 +1674,7 @@ mmu_booke_bootstrap(mmu_t mmu, vm_offset_t start, vm_o
 	/* Calculate corresponding physical addresses for the kernel region. */
 	phys_kernelend = kernload + kernsize;
 	debugf("kernel image and allocated data:\n");
-	debugf(" kernload    = 0x%09llx\n", (uint64_t)kernload);
+	debugf(" kernload    = 0x%09jx\n", (uintmax_t)kernload);
 	debugf(" kernstart   = 0x%"PRI0ptrX"\n", kernstart);
 	debugf(" kernsize    = 0x%"PRI0ptrX"\n", kernsize);
 
@@ -1859,9 +1859,9 @@ mmu_booke_bootstrap(mmu_t mmu, vm_offset_t start, vm_o
 	thread0.td_kstack = kstack0;
 	thread0.td_kstack_pages = kstack_pages;
 
-	debugf("kstack_sz = 0x%08x\n", kstack0_sz);
-	debugf("kstack0_phys at 0x%09llx - 0x%09llx\n",
-	    kstack0_phys, kstack0_phys + kstack0_sz);
+	debugf("kstack_sz = 0x%08jx\n", (uintmax_t)kstack0_sz);
+	debugf("kstack0_phys at 0x%09jx - 0x%09jx\n",
+	    (uintmax_t)kstack0_phys, (uintmax_t)kstack0_phys + kstack0_sz);
 	debugf("kstack0 at 0x%"PRI0ptrX" - 0x%"PRI0ptrX"\n",
 	    kstack0, kstack0 + kstack0_sz);
 	
@@ -4003,7 +4003,7 @@ tlb1_mapin_region(vm_offset_t va, vm_paddr_t pa, vm_si
 
 	for (idx = 0; idx < nents; idx++) {
 		pgsz = pgs[idx];
-		debugf("%u: %llx -> %jx, size=%jx\n", idx, pa,
+		debugf("%u: %jx -> %jx, size=%jx\n", idx, (uintmax_t)pa,
 		    (uintmax_t)va, (uintmax_t)pgsz);
 		tlb1_set_entry(va, pa, pgsz,
 		    _TLB_ENTRY_SHARED | _TLB_ENTRY_MEM);

Modified: head/sys/powerpc/booke/trap_subr.S
==============================================================================
--- head/sys/powerpc/booke/trap_subr.S	Sat Nov  2 21:02:54 2019	(r354265)
+++ head/sys/powerpc/booke/trap_subr.S	Sat Nov  2 21:15:56 2019	(r354266)
@@ -120,8 +120,8 @@
 	GET_CPUINFO(%r1);		/* Per-cpu structure */		\
 	STORE	%r30, (savearea+CPUSAVE_R30)(%r1);			\
 	STORE	%r31, (savearea+CPUSAVE_R31)(%r1); 			\
-	mfdear	%r30;		 					\
-	mfesr	%r31;							\
+	mfspr	%r30, SPR_DEAR;						\
+	mfspr	%r31, SPR_ESR;						\
 	STORE	%r30, (savearea+CPUSAVE_BOOKE_DEAR)(%r1); 		\
 	STORE	%r31, (savearea+CPUSAVE_BOOKE_ESR)(%r1); 		\
 	mfspr	%r30, isrr0;						\
@@ -143,8 +143,8 @@
 	GET_CPUINFO(%r1);		/* Per-cpu structure */		\
 	STORE	%r30, (savearea+CPUSAVE_R30)(%r1);			\
 	STORE	%r31, (savearea+CPUSAVE_R31)(%r1);			\
-	mfdear	%r30;							\
-	mfesr	%r31;							\
+	mfspr	%r30, SPR_DEAR;						\
+	mfspr	%r31, SPR_ESR;						\
 	STORE	%r30, (savearea+CPUSAVE_BOOKE_DEAR)(%r1);		\
 	STORE	%r31, (savearea+CPUSAVE_BOOKE_ESR)(%r1);		\
 	mfspr	%r30, isrr0;						\
@@ -375,9 +375,9 @@
 	rlwinm	outr, inr, 6, 23, 25;	/* 4 x TLBSAVE_LEN */
 #endif
 #define TLB_PROLOG							\
-	mtsprg4	%r1;			/* Save SP */			\
-	mtsprg5 %r28;							\
-	mtsprg6 %r29;							\
+	mtspr	SPR_SPRG4, %r1;			/* Save SP */		\
+	mtspr	SPR_SPRG5, %r28;					\
+	mtspr	SPR_SPRG6, %r29;					\
 	/* calculate TLB nesting level and TLBSAVE instance address */	\
 	GET_CPUINFO(%r1);	 	/* Per-cpu structure */		\
 	LOAD	%r28, PC_BOOKE_TLB_LEVEL(%r1);				\
@@ -388,8 +388,8 @@
 	add	%r1, %r1, %r29;		/* current TLBSAVE ptr */	\
 									\
 	/* save R20-31 */						\
-	mfsprg5 %r28;		 					\
-	mfsprg6 %r29;							\
+	mfspr	%r28, SPR_SPRG5;		 			\
+	mfspr	%r29, SPR_SPRG6;					\
 	TLB_SAVE_REGS(%r1);			\
 	/* save LR, CR */						\
 	mflr	%r30;		 					\
@@ -402,7 +402,7 @@
 	STORE	%r30, (TLBSAVE_BOOKE_SRR0)(%r1);	/* save SRR0 */	\
 	STORE	%r31, (TLBSAVE_BOOKE_SRR1)(%r1);	/* save SRR1 */	\
 	isync;								\
-	mfsprg4	%r1
+	mfspr	%r1, SPR_SPRG4
 
 /*
  * restores LR, CR, SRR0-1, R20-31 from the TLBSAVE area
@@ -410,7 +410,7 @@
  * same notes as for the TLB_PROLOG
  */
 #define TLB_RESTORE							\
-	mtsprg4	%r1;			/* Save SP */			\
+	mtspr	SPR_SPRG4, %r1;			/* Save SP */		\
 	GET_CPUINFO(%r1);	 	/* Per-cpu structure */		\
 	/* calculate TLB nesting level and TLBSAVE instance addr */	\
 	LOAD	%r28, PC_BOOKE_TLB_LEVEL(%r1);				\
@@ -432,7 +432,7 @@
 	mtsrr1	%r31;							\
 	/* restore R20-31 */						\
 	TLB_RESTORE_REGS(%r1);						\
-	mfsprg4	%r1
+	mfspr	%r1, SPR_SPRG4
 
 #ifdef SMP
 #define TLB_LOCK							\
@@ -693,7 +693,7 @@ INTERRUPT(int_data_tlb_error)
 	TLB_PROLOG
 	TLB_LOCK
 
-	mfdear	%r31
+	mfspr	%r31, SPR_DEAR
 
 	/*
 	 * Save MAS0-MAS2 registers. There might be another tlb miss during
@@ -1052,8 +1052,8 @@ ASENTRY_NOPROF(breakpoint)
 	mflr	%r31
 	mtsrr0	%r31
 
-	mfdear	%r30
-	mfesr	%r31
+	mfspr	%r30, SPR_DEAR
+	mfspr	%r31, SPR_ESR
 	STORE	%r30, (PC_DBSAVE+CPUSAVE_BOOKE_DEAR)(%r3)
 	STORE	%r31, (PC_DBSAVE+CPUSAVE_BOOKE_ESR)(%r3)
 

Modified: head/sys/powerpc/conf/dpaa/config.dpaa
==============================================================================
--- head/sys/powerpc/conf/dpaa/config.dpaa	Sat Nov  2 21:02:54 2019	(r354265)
+++ head/sys/powerpc/conf/dpaa/config.dpaa	Sat Nov  2 21:15:56 2019	(r354266)
@@ -2,7 +2,7 @@
 
 files		"dpaa/files.dpaa"
 
-makeoptions DPAA_COMPILE_CMD="${LINUXKPI_C} \
+makeoptions DPAA_COMPILE_CMD="${LINUXKPI_C} ${DPAAWARNFLAGS} \
 	-Wno-cast-qual -Wno-unused-function -Wno-init-self -fms-extensions \
 	-include $S/contrib/ncsw/build/dflags.h \
 	-Wno-error=missing-prototypes \

Modified: head/sys/powerpc/include/spr.h
==============================================================================
--- head/sys/powerpc/include/spr.h	Sat Nov  2 21:02:54 2019	(r354265)
+++ head/sys/powerpc/include/spr.h	Sat Nov  2 21:15:56 2019	(r354266)
@@ -504,7 +504,11 @@
 #define	SPR_HASH2		0x3d3	/* .68 Secondary Hash Address Register */
 #define	SPR_IMISS		0x3d4	/* .68 Instruction TLB Miss Address Register */
 #define	SPR_TLBMISS		0x3d4	/* .6. TLB Miss Address Register */
+#if defined(BOOKE_PPC4XX)
 #define	SPR_DEAR		0x3d5	/* 4.. Data Error Address Register */
+#else
+#define	SPR_DEAR		0x03d	/* ..8 Data Exception Address Register */
+#endif
 #define	SPR_ICMP		0x3d5	/* .68 Instruction TLB Compare Register */
 #define	SPR_PTEHI		0x3d5	/* .6. Instruction TLB Compare Register */
 #define	SPR_EVPR		0x3d6	/* 4.. Exception Vector Prefix Register */

Modified: head/sys/powerpc/mpc85xx/platform_mpc85xx.c
==============================================================================
--- head/sys/powerpc/mpc85xx/platform_mpc85xx.c	Sat Nov  2 21:02:54 2019	(r354265)
+++ head/sys/powerpc/mpc85xx/platform_mpc85xx.c	Sat Nov  2 21:15:56 2019	(r354266)
@@ -68,7 +68,8 @@ __FBSDID("$FreeBSD$");
 extern void *ap_pcpu;
 extern vm_paddr_t kernload;		/* Kernel physical load address */
 extern uint8_t __boot_page[];		/* Boot page body */
-extern uint32_t bp_kernload;
+extern vm_paddr_t bp_kernload;		/* Boot page copy of kernload */
+extern vm_offset_t bp_virtaddr;		/* Virtual address of boot page */
 extern vm_offset_t __startkernel;
 
 struct cpu_release {
@@ -354,11 +355,13 @@ mpc85xx_smp_start_cpu_epapr(platform_t plat, struct pc
 	pmap_kenter(rel_page, rel_pa & ~PAGE_MASK);
 	rel = (struct cpu_release *)rel_va;
 	bptr = pmap_kextract((uintptr_t)__boot_page);
+
 	cpu_flush_dcache(__DEVOLATILE(struct cpu_release *,rel), sizeof(*rel));
-	rel->pir = pc->pc_cpuid; __asm __volatile("sync");
-	rel->entry_h = (bptr >> 32);
-	rel->entry_l = bptr; __asm __volatile("sync");
+	rel->pir = pc->pc_cpuid; __asm __volatile("sync" ::: "memory");
+	rel->entry_h = (bptr >> 32); __asm __volatile("sync" ::: "memory");
 	cpu_flush_dcache(__DEVOLATILE(struct cpu_release *,rel), sizeof(*rel));
+	rel->entry_l = bptr & 0xffffffff; __asm __volatile("sync" ::: "memory");
+	cpu_flush_dcache(__DEVOLATILE(struct cpu_release *,rel), sizeof(*rel));
 	if (bootverbose)
 		printf("Waking up CPU %d via CPU release page %p\n",
 		    pc->pc_cpuid, rel);
@@ -397,11 +400,13 @@ mpc85xx_smp_start_cpu(platform_t plat, struct pcpu *pc
 		cpuid = pc->pc_cpuid + 24;
 	}
 	bp_kernload = kernload;
+	bp_virtaddr = (vm_offset_t)&__boot_page;
 	/*
-	 * bp_kernload is in the boot page.  Sync the cache because ePAPR
-	 * booting has the other core(s) already running.
+	 * bp_kernload and bp_virtaddr are in the boot page.  Sync the cache
+	 * because ePAPR booting has the other core(s) already running.
 	 */
 	cpu_flush_dcache(&bp_kernload, sizeof(bp_kernload));
+	cpu_flush_dcache(&bp_virtaddr, sizeof(bp_virtaddr));
 
 	ap_pcpu = pc;
 	__asm __volatile("msync; isync");

Modified: head/sys/powerpc/powerpc/pmap_dispatch.c
==============================================================================
--- head/sys/powerpc/powerpc/pmap_dispatch.c	Sat Nov  2 21:02:54 2019	(r354265)
+++ head/sys/powerpc/powerpc/pmap_dispatch.c	Sat Nov  2 21:15:56 2019	(r354266)
@@ -145,7 +145,7 @@ pmap_enter(pmap_t pmap, vm_offset_t va, vm_page_t p, v
     u_int flags, int8_t psind)
 {
 
-	CTR6(KTR_PMAP, "pmap_enter(%p, %#x, %p, %#x, %x, %d)", pmap, va,
+	CTR6(KTR_PMAP, "pmap_enter(%p, %#x, %p, %#x, %#x, %d)", pmap, va,
 	    p, prot, flags, psind);
 	return (MMU_ENTER(mmu_obj, pmap, va, p, prot, flags, psind));
 }



Want to link to this message? Use this URL: <https://mail-archive.FreeBSD.org/cgi/mid.cgi?201911022115.xA2LFuWi004298>