powerpc: Fix usage of register macros getting ready for %r0 change
authorMichael Neuling <mikey@neuling.org>
Mon, 25 Jun 2012 13:33:10 +0000 (13:33 +0000)
committerBenjamin Herrenschmidt <benh@kernel.crashing.org>
Tue, 10 Jul 2012 09:17:55 +0000 (19:17 +1000)
Anything that uses a constructed instruction (ie. from ppc-opcode.h),
need to use the new R0 macro, as %r0 is not going to work.

Also convert usages of macros where we are just determining an offset
(usually for a load/store), like:
std r14,STK_REG(r14)(r1)
Can't use STK_REG(r14) as %r14 doesn't work in the STK_REG macro since
it's just calculating an offset.

Signed-off-by: Michael Neuling <mikey@neuling.org>
Signed-off-by: Benjamin Herrenschmidt <benh@kernel.crashing.org>
25 files changed:
arch/powerpc/kernel/cpu_setup_a2.S
arch/powerpc/kernel/fpu.S
arch/powerpc/kernel/kvm.c
arch/powerpc/kernel/misc_64.S
arch/powerpc/kvm/book3s_hv_rmhandlers.S
arch/powerpc/kvm/book3s_interrupts.S
arch/powerpc/kvm/booke_interrupts.S
arch/powerpc/kvm/bookehv_interrupts.S
arch/powerpc/lib/checksum_64.S
arch/powerpc/lib/copypage_power7.S
arch/powerpc/lib/copyuser_64.S
arch/powerpc/lib/copyuser_power7.S
arch/powerpc/lib/hweight_64.S
arch/powerpc/lib/ldstfp.S
arch/powerpc/lib/mem_64.S
arch/powerpc/lib/memcpy_64.S
arch/powerpc/lib/memcpy_power7.S
arch/powerpc/mm/hash_low_64.S
arch/powerpc/mm/tlb_low_64e.S
arch/powerpc/mm/tlb_nohash_low.S
arch/powerpc/net/bpf_jit_comp.c
arch/powerpc/platforms/cell/beat_hvCall.S
arch/powerpc/platforms/powernv/opal-takeover.S
arch/powerpc/platforms/powernv/opal-wrappers.S
arch/powerpc/platforms/pseries/hvCall.S

index ebc62f4..95675a7 100644 (file)
@@ -100,19 +100,19 @@ _icswx_skip_guest:
        lis     r4,(MMUCR0_TLBSEL_I|MMUCR0_ECL)@h
        mtspr   SPRN_MMUCR0, r4
        li      r4,A2_IERAT_SIZE-1
-       PPC_ERATWE(r4,r4,3)
+       PPC_ERATWE(R4,R4,3)
 
        /* Now set the D-ERAT watermark to 31 */
        lis     r4,(MMUCR0_TLBSEL_D|MMUCR0_ECL)@h
        mtspr   SPRN_MMUCR0, r4
        li      r4,A2_DERAT_SIZE-1
-       PPC_ERATWE(r4,r4,3)
+       PPC_ERATWE(R4,R4,3)
 
        /* And invalidate the beast just in case. That won't get rid of
         * a bolted entry though it will be in LRU and so will go away eventually
         * but let's not bother for now
         */
-       PPC_ERATILX(0,0,0)
+       PPC_ERATILX(0,R0,R0)
 1:
        blr
 
index de36955..71c1c73 100644 (file)
@@ -106,7 +106,7 @@ END_FTR_SECTION_IFSET(CPU_FTR_VSX)
 #endif
        lfd     fr0,THREAD_FPSCR(r5)
        MTFSF_L(fr0)
-       REST_32FPVSRS(0, r4, r5)
+       REST_32FPVSRS(0, R4, R5)
 #ifndef CONFIG_SMP
        subi    r4,r5,THREAD
        fromreal(r4)
@@ -140,7 +140,7 @@ END_FTR_SECTION_IFSET(CPU_FTR_VSX)
        addi    r3,r3,THREAD            /* want THREAD of task */
        PPC_LL  r5,PT_REGS(r3)
        PPC_LCMPI       0,r5,0
-       SAVE_32FPVSRS(0, r4 ,r3)
+       SAVE_32FPVSRS(0, R4 ,R3)
        mffs    fr0
        stfd    fr0,THREAD_FPSCR(r3)
        beq     1f
index 62bdf23..02c167d 100644 (file)
@@ -302,7 +302,7 @@ static void kvm_patch_ins_wrtee(u32 *inst, u32 rt, int imm_one)
 
        if (imm_one) {
                p[kvm_emulate_wrtee_reg_offs] =
-                       KVM_INST_LI | __PPC_RT(30) | MSR_EE;
+                       KVM_INST_LI | __PPC_RT(R30) | MSR_EE;
        } else {
                /* Make clobbered registers work too */
                switch (get_rt(rt)) {
index 616921e..6ba08bc 100644 (file)
@@ -314,7 +314,7 @@ _GLOBAL(real_205_readb)
        mtmsrd  r0
        sync
        isync
-       LBZCIX(r3,0,r3)
+       LBZCIX(R3,0,R3)
        isync
        mtmsrd  r7
        sync
@@ -329,7 +329,7 @@ _GLOBAL(real_205_writeb)
        mtmsrd  r0
        sync
        isync
-       STBCIX(r3,0,r4)
+       STBCIX(R3,0,R4)
        isync
        mtmsrd  r7
        sync
index a1044f4..bc99015 100644 (file)
@@ -206,24 +206,24 @@ END_FTR_SECTION_IFSET(CPU_FTR_ARCH_201)
        /* Load up FP, VMX and VSX registers */
        bl      kvmppc_load_fp
 
-       ld      r14, VCPU_GPR(r14)(r4)
-       ld      r15, VCPU_GPR(r15)(r4)
-       ld      r16, VCPU_GPR(r16)(r4)
-       ld      r17, VCPU_GPR(r17)(r4)
-       ld      r18, VCPU_GPR(r18)(r4)
-       ld      r19, VCPU_GPR(r19)(r4)
-       ld      r20, VCPU_GPR(r20)(r4)
-       ld      r21, VCPU_GPR(r21)(r4)
-       ld      r22, VCPU_GPR(r22)(r4)
-       ld      r23, VCPU_GPR(r23)(r4)
-       ld      r24, VCPU_GPR(r24)(r4)
-       ld      r25, VCPU_GPR(r25)(r4)
-       ld      r26, VCPU_GPR(r26)(r4)
-       ld      r27, VCPU_GPR(r27)(r4)
-       ld      r28, VCPU_GPR(r28)(r4)
-       ld      r29, VCPU_GPR(r29)(r4)
-       ld      r30, VCPU_GPR(r30)(r4)
-       ld      r31, VCPU_GPR(r31)(r4)
+       ld      r14, VCPU_GPR(R14)(r4)
+       ld      r15, VCPU_GPR(R15)(r4)
+       ld      r16, VCPU_GPR(R16)(r4)
+       ld      r17, VCPU_GPR(R17)(r4)
+       ld      r18, VCPU_GPR(R18)(r4)
+       ld      r19, VCPU_GPR(R19)(r4)
+       ld      r20, VCPU_GPR(R20)(r4)
+       ld      r21, VCPU_GPR(R21)(r4)
+       ld      r22, VCPU_GPR(R22)(r4)
+       ld      r23, VCPU_GPR(R23)(r4)
+       ld      r24, VCPU_GPR(R24)(r4)
+       ld      r25, VCPU_GPR(R25)(r4)
+       ld      r26, VCPU_GPR(R26)(r4)
+       ld      r27, VCPU_GPR(R27)(r4)
+       ld      r28, VCPU_GPR(R28)(r4)
+       ld      r29, VCPU_GPR(R29)(r4)
+       ld      r30, VCPU_GPR(R30)(r4)
+       ld      r31, VCPU_GPR(R31)(r4)
 
 BEGIN_FTR_SECTION
        /* Switch DSCR to guest value */
@@ -547,21 +547,21 @@ fast_guest_return:
        mtlr    r5
        mtcr    r6
 
-       ld      r0, VCPU_GPR(r0)(r4)
-       ld      r1, VCPU_GPR(r1)(r4)
-       ld      r2, VCPU_GPR(r2)(r4)
-       ld      r3, VCPU_GPR(r3)(r4)
-       ld      r5, VCPU_GPR(r5)(r4)
-       ld      r6, VCPU_GPR(r6)(r4)
-       ld      r7, VCPU_GPR(r7)(r4)
-       ld      r8, VCPU_GPR(r8)(r4)
-       ld      r9, VCPU_GPR(r9)(r4)
-       ld      r10, VCPU_GPR(r10)(r4)
-       ld      r11, VCPU_GPR(r11)(r4)
-       ld      r12, VCPU_GPR(r12)(r4)
-       ld      r13, VCPU_GPR(r13)(r4)
-
-       ld      r4, VCPU_GPR(r4)(r4)
+       ld      r0, VCPU_GPR(R0)(r4)
+       ld      r1, VCPU_GPR(R1)(r4)
+       ld      r2, VCPU_GPR(R2)(r4)
+       ld      r3, VCPU_GPR(R3)(r4)
+       ld      r5, VCPU_GPR(R5)(r4)
+       ld      r6, VCPU_GPR(R6)(r4)
+       ld      r7, VCPU_GPR(R7)(r4)
+       ld      r8, VCPU_GPR(R8)(r4)
+       ld      r9, VCPU_GPR(R9)(r4)
+       ld      r10, VCPU_GPR(R10)(r4)
+       ld      r11, VCPU_GPR(R11)(r4)
+       ld      r12, VCPU_GPR(R12)(r4)
+       ld      r13, VCPU_GPR(R13)(r4)
+
+       ld      r4, VCPU_GPR(R4)(r4)
 
        hrfid
        b       .
@@ -590,22 +590,22 @@ kvmppc_interrupt:
 
        /* Save registers */
 
-       std     r0, VCPU_GPR(r0)(r9)
-       std     r1, VCPU_GPR(r1)(r9)
-       std     r2, VCPU_GPR(r2)(r9)
-       std     r3, VCPU_GPR(r3)(r9)
-       std     r4, VCPU_GPR(r4)(r9)
-       std     r5, VCPU_GPR(r5)(r9)
-       std     r6, VCPU_GPR(r6)(r9)
-       std     r7, VCPU_GPR(r7)(r9)
-       std     r8, VCPU_GPR(r8)(r9)
+       std     r0, VCPU_GPR(R0)(r9)
+       std     r1, VCPU_GPR(R1)(r9)
+       std     r2, VCPU_GPR(R2)(r9)
+       std     r3, VCPU_GPR(R3)(r9)
+       std     r4, VCPU_GPR(R4)(r9)
+       std     r5, VCPU_GPR(R5)(r9)
+       std     r6, VCPU_GPR(R6)(r9)
+       std     r7, VCPU_GPR(R7)(r9)
+       std     r8, VCPU_GPR(R8)(r9)
        ld      r0, HSTATE_HOST_R2(r13)
-       std     r0, VCPU_GPR(r9)(r9)
-       std     r10, VCPU_GPR(r10)(r9)
-       std     r11, VCPU_GPR(r11)(r9)
+       std     r0, VCPU_GPR(R9)(r9)
+       std     r10, VCPU_GPR(R10)(r9)
+       std     r11, VCPU_GPR(R11)(r9)
        ld      r3, HSTATE_SCRATCH0(r13)
        lwz     r4, HSTATE_SCRATCH1(r13)
-       std     r3, VCPU_GPR(r12)(r9)
+       std     r3, VCPU_GPR(R12)(r9)
        stw     r4, VCPU_CR(r9)
 
        /* Restore R1/R2 so we can handle faults */
@@ -626,7 +626,7 @@ kvmppc_interrupt:
 
        GET_SCRATCH0(r3)
        mflr    r4
-       std     r3, VCPU_GPR(r13)(r9)
+       std     r3, VCPU_GPR(R13)(r9)
        std     r4, VCPU_LR(r9)
 
        /* Unset guest mode */
@@ -968,24 +968,24 @@ BEGIN_FTR_SECTION
 END_FTR_SECTION_IFSET(CPU_FTR_ARCH_206)
 
        /* Save non-volatile GPRs */
-       std     r14, VCPU_GPR(r14)(r9)
-       std     r15, VCPU_GPR(r15)(r9)
-       std     r16, VCPU_GPR(r16)(r9)
-       std     r17, VCPU_GPR(r17)(r9)
-       std     r18, VCPU_GPR(r18)(r9)
-       std     r19, VCPU_GPR(r19)(r9)
-       std     r20, VCPU_GPR(r20)(r9)
-       std     r21, VCPU_GPR(r21)(r9)
-       std     r22, VCPU_GPR(r22)(r9)
-       std     r23, VCPU_GPR(r23)(r9)
-       std     r24, VCPU_GPR(r24)(r9)
-       std     r25, VCPU_GPR(r25)(r9)
-       std     r26, VCPU_GPR(r26)(r9)
-       std     r27, VCPU_GPR(r27)(r9)
-       std     r28, VCPU_GPR(r28)(r9)
-       std     r29, VCPU_GPR(r29)(r9)
-       std     r30, VCPU_GPR(r30)(r9)
-       std     r31, VCPU_GPR(r31)(r9)
+       std     r14, VCPU_GPR(R14)(r9)
+       std     r15, VCPU_GPR(R15)(r9)
+       std     r16, VCPU_GPR(R16)(r9)
+       std     r17, VCPU_GPR(R17)(r9)
+       std     r18, VCPU_GPR(R18)(r9)
+       std     r19, VCPU_GPR(R19)(r9)
+       std     r20, VCPU_GPR(R20)(r9)
+       std     r21, VCPU_GPR(R21)(r9)
+       std     r22, VCPU_GPR(R22)(r9)
+       std     r23, VCPU_GPR(R23)(r9)
+       std     r24, VCPU_GPR(R24)(r9)
+       std     r25, VCPU_GPR(R25)(r9)
+       std     r26, VCPU_GPR(R26)(r9)
+       std     r27, VCPU_GPR(R27)(r9)
+       std     r28, VCPU_GPR(R28)(r9)
+       std     r29, VCPU_GPR(R29)(r9)
+       std     r30, VCPU_GPR(R30)(r9)
+       std     r31, VCPU_GPR(R31)(r9)
 
        /* Save SPRGs */
        mfspr   r3, SPRN_SPRG0
@@ -1160,7 +1160,7 @@ kvmppc_hdsi:
        andi.   r0, r11, MSR_DR         /* data relocation enabled? */
        beq     3f
        clrrdi  r0, r4, 28
-       PPC_SLBFEE_DOT(r5, r0)          /* if so, look up SLB */
+       PPC_SLBFEE_DOT(R5, R0)          /* if so, look up SLB */
        bne     1f                      /* if no SLB entry found */
 4:     std     r4, VCPU_FAULT_DAR(r9)
        stw     r6, VCPU_FAULT_DSISR(r9)
@@ -1234,7 +1234,7 @@ kvmppc_hisi:
        andi.   r0, r11, MSR_IR         /* instruction relocation enabled? */
        beq     3f
        clrrdi  r0, r10, 28
-       PPC_SLBFEE_DOT(r5, r0)          /* if so, look up SLB */
+       PPC_SLBFEE_DOT(R5, R0)          /* if so, look up SLB */
        bne     1f                      /* if no SLB entry found */
 4:
        /* Search the hash table. */
@@ -1278,7 +1278,7 @@ kvmppc_hisi:
  */
        .globl  hcall_try_real_mode
 hcall_try_real_mode:
-       ld      r3,VCPU_GPR(r3)(r9)
+       ld      r3,VCPU_GPR(R3)(r9)
        andi.   r0,r11,MSR_PR
        bne     hcall_real_cont
        clrrdi  r3,r3,2
@@ -1291,12 +1291,12 @@ hcall_try_real_mode:
        add     r3,r3,r4
        mtctr   r3
        mr      r3,r9           /* get vcpu pointer */
-       ld      r4,VCPU_GPR(r4)(r9)
+       ld      r4,VCPU_GPR(R4)(r9)
        bctrl
        cmpdi   r3,H_TOO_HARD
        beq     hcall_real_fallback
        ld      r4,HSTATE_KVM_VCPU(r13)
-       std     r3,VCPU_GPR(r3)(r4)
+       std     r3,VCPU_GPR(R3)(r4)
        ld      r10,VCPU_PC(r4)
        ld      r11,VCPU_MSR(r4)
        b       fast_guest_return
@@ -1424,7 +1424,7 @@ _GLOBAL(kvmppc_h_cede)
        li      r0,0            /* set trap to 0 to say hcall is handled */
        stw     r0,VCPU_TRAP(r3)
        li      r0,H_SUCCESS
-       std     r0,VCPU_GPR(r3)(r3)
+       std     r0,VCPU_GPR(R3)(r3)
 BEGIN_FTR_SECTION
        b       2f              /* just send it up to host on 970 */
 END_FTR_SECTION_IFCLR(CPU_FTR_ARCH_206)
@@ -1443,7 +1443,7 @@ END_FTR_SECTION_IFCLR(CPU_FTR_ARCH_206)
        addi    r6,r5,VCORE_NAPPING_THREADS
 31:    lwarx   r4,0,r6
        or      r4,r4,r0
-       PPC_POPCNTW(r7,r4)
+       PPC_POPCNTW(R7,R4)
        cmpw    r7,r8
        bge     2f
        stwcx.  r4,0,r6
@@ -1464,24 +1464,24 @@ END_FTR_SECTION_IFCLR(CPU_FTR_ARCH_206)
  * DAR, DSISR, DABR, DABRX, DSCR, PMCx, MMCRx, SIAR, SDAR.
  */
        /* Save non-volatile GPRs */
-       std     r14, VCPU_GPR(r14)(r3)
-       std     r15, VCPU_GPR(r15)(r3)
-       std     r16, VCPU_GPR(r16)(r3)
-       std     r17, VCPU_GPR(r17)(r3)
-       std     r18, VCPU_GPR(r18)(r3)
-       std     r19, VCPU_GPR(r19)(r3)
-       std     r20, VCPU_GPR(r20)(r3)
-       std     r21, VCPU_GPR(r21)(r3)
-       std     r22, VCPU_GPR(r22)(r3)
-       std     r23, VCPU_GPR(r23)(r3)
-       std     r24, VCPU_GPR(r24)(r3)
-       std     r25, VCPU_GPR(r25)(r3)
-       std     r26, VCPU_GPR(r26)(r3)
-       std     r27, VCPU_GPR(r27)(r3)
-       std     r28, VCPU_GPR(r28)(r3)
-       std     r29, VCPU_GPR(r29)(r3)
-       std     r30, VCPU_GPR(r30)(r3)
-       std     r31, VCPU_GPR(r31)(r3)
+       std     r14, VCPU_GPR(R14)(r3)
+       std     r15, VCPU_GPR(R15)(r3)
+       std     r16, VCPU_GPR(R16)(r3)
+       std     r17, VCPU_GPR(R17)(r3)
+       std     r18, VCPU_GPR(R18)(r3)
+       std     r19, VCPU_GPR(R19)(r3)
+       std     r20, VCPU_GPR(R20)(r3)
+       std     r21, VCPU_GPR(R21)(r3)
+       std     r22, VCPU_GPR(R22)(r3)
+       std     r23, VCPU_GPR(R23)(r3)
+       std     r24, VCPU_GPR(R24)(r3)
+       std     r25, VCPU_GPR(R25)(r3)
+       std     r26, VCPU_GPR(R26)(r3)
+       std     r27, VCPU_GPR(R27)(r3)
+       std     r28, VCPU_GPR(R28)(r3)
+       std     r29, VCPU_GPR(R29)(r3)
+       std     r30, VCPU_GPR(R30)(r3)
+       std     r31, VCPU_GPR(R31)(r3)
 
        /* save FP state */
        bl      .kvmppc_save_fp
@@ -1513,24 +1513,24 @@ kvm_end_cede:
        bl      kvmppc_load_fp
 
        /* Load NV GPRS */
-       ld      r14, VCPU_GPR(r14)(r4)
-       ld      r15, VCPU_GPR(r15)(r4)
-       ld      r16, VCPU_GPR(r16)(r4)
-       ld      r17, VCPU_GPR(r17)(r4)
-       ld      r18, VCPU_GPR(r18)(r4)
-       ld      r19, VCPU_GPR(r19)(r4)
-       ld      r20, VCPU_GPR(r20)(r4)
-       ld      r21, VCPU_GPR(r21)(r4)
-       ld      r22, VCPU_GPR(r22)(r4)
-       ld      r23, VCPU_GPR(r23)(r4)
-       ld      r24, VCPU_GPR(r24)(r4)
-       ld      r25, VCPU_GPR(r25)(r4)
-       ld      r26, VCPU_GPR(r26)(r4)
-       ld      r27, VCPU_GPR(r27)(r4)
-       ld      r28, VCPU_GPR(r28)(r4)
-       ld      r29, VCPU_GPR(r29)(r4)
-       ld      r30, VCPU_GPR(r30)(r4)
-       ld      r31, VCPU_GPR(r31)(r4)
+       ld      r14, VCPU_GPR(R14)(r4)
+       ld      r15, VCPU_GPR(R15)(r4)
+       ld      r16, VCPU_GPR(R16)(r4)
+       ld      r17, VCPU_GPR(R17)(r4)
+       ld      r18, VCPU_GPR(R18)(r4)
+       ld      r19, VCPU_GPR(R19)(r4)
+       ld      r20, VCPU_GPR(R20)(r4)
+       ld      r21, VCPU_GPR(R21)(r4)
+       ld      r22, VCPU_GPR(R22)(r4)
+       ld      r23, VCPU_GPR(R23)(r4)
+       ld      r24, VCPU_GPR(R24)(r4)
+       ld      r25, VCPU_GPR(R25)(r4)
+       ld      r26, VCPU_GPR(R26)(r4)
+       ld      r27, VCPU_GPR(R27)(r4)
+       ld      r28, VCPU_GPR(R28)(r4)
+       ld      r29, VCPU_GPR(R29)(r4)
+       ld      r30, VCPU_GPR(R30)(r4)
+       ld      r31, VCPU_GPR(R31)(r4)
 
        /* clear our bit in vcore->napping_threads */
 33:    ld      r5,HSTATE_KVM_VCORE(r13)
@@ -1649,7 +1649,7 @@ BEGIN_FTR_SECTION
        reg = 0
        .rept   32
        li      r6,reg*16+VCPU_VSRS
-       STXVD2X(reg,r6,r3)
+       STXVD2X(reg,R6,R3)
        reg = reg + 1
        .endr
 FTR_SECTION_ELSE
@@ -1711,7 +1711,7 @@ BEGIN_FTR_SECTION
        reg = 0
        .rept   32
        li      r7,reg*16+VCPU_VSRS
-       LXVD2X(reg,r7,r4)
+       LXVD2X(reg,R7,R4)
        reg = reg + 1
        .endr
 FTR_SECTION_ELSE
index 3e35383..2ddab0f 100644 (file)
 
 #define VCPU_GPR(n)            (VCPU_GPRS + (n * ULONG_SIZE))
 #define VCPU_LOAD_NVGPRS(vcpu) \
-       PPC_LL  r14, VCPU_GPR(r14)(vcpu); \
-       PPC_LL  r15, VCPU_GPR(r15)(vcpu); \
-       PPC_LL  r16, VCPU_GPR(r16)(vcpu); \
-       PPC_LL  r17, VCPU_GPR(r17)(vcpu); \
-       PPC_LL  r18, VCPU_GPR(r18)(vcpu); \
-       PPC_LL  r19, VCPU_GPR(r19)(vcpu); \
-       PPC_LL  r20, VCPU_GPR(r20)(vcpu); \
-       PPC_LL  r21, VCPU_GPR(r21)(vcpu); \
-       PPC_LL  r22, VCPU_GPR(r22)(vcpu); \
-       PPC_LL  r23, VCPU_GPR(r23)(vcpu); \
-       PPC_LL  r24, VCPU_GPR(r24)(vcpu); \
-       PPC_LL  r25, VCPU_GPR(r25)(vcpu); \
-       PPC_LL  r26, VCPU_GPR(r26)(vcpu); \
-       PPC_LL  r27, VCPU_GPR(r27)(vcpu); \
-       PPC_LL  r28, VCPU_GPR(r28)(vcpu); \
-       PPC_LL  r29, VCPU_GPR(r29)(vcpu); \
-       PPC_LL  r30, VCPU_GPR(r30)(vcpu); \
-       PPC_LL  r31, VCPU_GPR(r31)(vcpu); \
+       PPC_LL  r14, VCPU_GPR(R14)(vcpu); \
+       PPC_LL  r15, VCPU_GPR(R15)(vcpu); \
+       PPC_LL  r16, VCPU_GPR(R16)(vcpu); \
+       PPC_LL  r17, VCPU_GPR(R17)(vcpu); \
+       PPC_LL  r18, VCPU_GPR(R18)(vcpu); \
+       PPC_LL  r19, VCPU_GPR(R19)(vcpu); \
+       PPC_LL  r20, VCPU_GPR(R20)(vcpu); \
+       PPC_LL  r21, VCPU_GPR(R21)(vcpu); \
+       PPC_LL  r22, VCPU_GPR(R22)(vcpu); \
+       PPC_LL  r23, VCPU_GPR(R23)(vcpu); \
+       PPC_LL  r24, VCPU_GPR(R24)(vcpu); \
+       PPC_LL  r25, VCPU_GPR(R25)(vcpu); \
+       PPC_LL  r26, VCPU_GPR(R26)(vcpu); \
+       PPC_LL  r27, VCPU_GPR(R27)(vcpu); \
+       PPC_LL  r28, VCPU_GPR(R28)(vcpu); \
+       PPC_LL  r29, VCPU_GPR(R29)(vcpu); \
+       PPC_LL  r30, VCPU_GPR(R30)(vcpu); \
+       PPC_LL  r31, VCPU_GPR(R31)(vcpu); \
 
 /*****************************************************************************
  *                                                                           *
@@ -131,24 +131,24 @@ kvmppc_handler_highmem:
        /* R7 = vcpu */
        PPC_LL  r7, GPR4(r1)
 
-       PPC_STL r14, VCPU_GPR(r14)(r7)
-       PPC_STL r15, VCPU_GPR(r15)(r7)
-       PPC_STL r16, VCPU_GPR(r16)(r7)
-       PPC_STL r17, VCPU_GPR(r17)(r7)
-       PPC_STL r18, VCPU_GPR(r18)(r7)
-       PPC_STL r19, VCPU_GPR(r19)(r7)
-       PPC_STL r20, VCPU_GPR(r20)(r7)
-       PPC_STL r21, VCPU_GPR(r21)(r7)
-       PPC_STL r22, VCPU_GPR(r22)(r7)
-       PPC_STL r23, VCPU_GPR(r23)(r7)
-       PPC_STL r24, VCPU_GPR(r24)(r7)
-       PPC_STL r25, VCPU_GPR(r25)(r7)
-       PPC_STL r26, VCPU_GPR(r26)(r7)
-       PPC_STL r27, VCPU_GPR(r27)(r7)
-       PPC_STL r28, VCPU_GPR(r28)(r7)
-       PPC_STL r29, VCPU_GPR(r29)(r7)
-       PPC_STL r30, VCPU_GPR(r30)(r7)
-       PPC_STL r31, VCPU_GPR(r31)(r7)
+       PPC_STL r14, VCPU_GPR(R14)(r7)
+       PPC_STL r15, VCPU_GPR(R15)(r7)
+       PPC_STL r16, VCPU_GPR(R16)(r7)
+       PPC_STL r17, VCPU_GPR(R17)(r7)
+       PPC_STL r18, VCPU_GPR(R18)(r7)
+       PPC_STL r19, VCPU_GPR(R19)(r7)
+       PPC_STL r20, VCPU_GPR(R20)(r7)
+       PPC_STL r21, VCPU_GPR(R21)(r7)
+       PPC_STL r22, VCPU_GPR(R22)(r7)
+       PPC_STL r23, VCPU_GPR(R23)(r7)
+       PPC_STL r24, VCPU_GPR(R24)(r7)
+       PPC_STL r25, VCPU_GPR(R25)(r7)
+       PPC_STL r26, VCPU_GPR(R26)(r7)
+       PPC_STL r27, VCPU_GPR(R27)(r7)
+       PPC_STL r28, VCPU_GPR(R28)(r7)
+       PPC_STL r29, VCPU_GPR(R29)(r7)
+       PPC_STL r30, VCPU_GPR(R30)(r7)
+       PPC_STL r31, VCPU_GPR(R31)(r7)
 
        /* Pass the exit number as 3rd argument to kvmppc_handle_exit */
        mr      r5, r12
index 8feec2f..e598a5a 100644 (file)
@@ -37,7 +37,7 @@
 #define HOST_CR         16
 #define HOST_NV_GPRS    20
 #define HOST_NV_GPR(n)  (HOST_NV_GPRS + ((n - 14) * 4))
-#define HOST_MIN_STACK_SIZE (HOST_NV_GPR(31) + 4)
+#define HOST_MIN_STACK_SIZE (HOST_NV_GPR(R31) + 4)
 #define HOST_STACK_SIZE (((HOST_MIN_STACK_SIZE + 15) / 16) * 16) /* Align. */
 #define HOST_STACK_LR   (HOST_STACK_SIZE + 4) /* In caller stack frame. */
 
@@ -58,8 +58,8 @@ _GLOBAL(kvmppc_handler_\ivor_nr)
        /* Get pointer to vcpu and record exit number. */
        mtspr   SPRN_SPRG_WSCRATCH0, r4
        mfspr   r4, SPRN_SPRG_RVCPU
-       stw     r5, VCPU_GPR(r5)(r4)
-       stw     r6, VCPU_GPR(r6)(r4)
+       stw     r5, VCPU_GPR(R5)(r4)
+       stw     r6, VCPU_GPR(R6)(r4)
        mfctr   r5
        lis     r6, kvmppc_resume_host@h
        stw     r5, VCPU_CTR(r4)
@@ -100,12 +100,12 @@ _GLOBAL(kvmppc_handler_len)
  *  r5: KVM exit number
  */
 _GLOBAL(kvmppc_resume_host)
-       stw     r3, VCPU_GPR(r3)(r4)
+       stw     r3, VCPU_GPR(R3)(r4)
        mfcr    r3
        stw     r3, VCPU_CR(r4)
-       stw     r7, VCPU_GPR(r7)(r4)
-       stw     r8, VCPU_GPR(r8)(r4)
-       stw     r9, VCPU_GPR(r9)(r4)
+       stw     r7, VCPU_GPR(R7)(r4)
+       stw     r8, VCPU_GPR(R8)(r4)
+       stw     r9, VCPU_GPR(R9)(r4)
 
        li      r6, 1
        slw     r6, r6, r5
@@ -135,23 +135,23 @@ _GLOBAL(kvmppc_resume_host)
        isync
        stw     r9, VCPU_LAST_INST(r4)
 
-       stw     r15, VCPU_GPR(r15)(r4)
-       stw     r16, VCPU_GPR(r16)(r4)
-       stw     r17, VCPU_GPR(r17)(r4)
-       stw     r18, VCPU_GPR(r18)(r4)
-       stw     r19, VCPU_GPR(r19)(r4)
-       stw     r20, VCPU_GPR(r20)(r4)
-       stw     r21, VCPU_GPR(r21)(r4)
-       stw     r22, VCPU_GPR(r22)(r4)
-       stw     r23, VCPU_GPR(r23)(r4)
-       stw     r24, VCPU_GPR(r24)(r4)
-       stw     r25, VCPU_GPR(r25)(r4)
-       stw     r26, VCPU_GPR(r26)(r4)
-       stw     r27, VCPU_GPR(r27)(r4)
-       stw     r28, VCPU_GPR(r28)(r4)
-       stw     r29, VCPU_GPR(r29)(r4)
-       stw     r30, VCPU_GPR(r30)(r4)
-       stw     r31, VCPU_GPR(r31)(r4)
+       stw     r15, VCPU_GPR(R15)(r4)
+       stw     r16, VCPU_GPR(R16)(r4)
+       stw     r17, VCPU_GPR(R17)(r4)
+       stw     r18, VCPU_GPR(R18)(r4)
+       stw     r19, VCPU_GPR(R19)(r4)
+       stw     r20, VCPU_GPR(R20)(r4)
+       stw     r21, VCPU_GPR(R21)(r4)
+       stw     r22, VCPU_GPR(R22)(r4)
+       stw     r23, VCPU_GPR(R23)(r4)
+       stw     r24, VCPU_GPR(R24)(r4)
+       stw     r25, VCPU_GPR(R25)(r4)
+       stw     r26, VCPU_GPR(R26)(r4)
+       stw     r27, VCPU_GPR(R27)(r4)
+       stw     r28, VCPU_GPR(R28)(r4)
+       stw     r29, VCPU_GPR(R29)(r4)
+       stw     r30, VCPU_GPR(R30)(r4)
+       stw     r31, VCPU_GPR(R31)(r4)
 ..skip_inst_copy:
 
        /* Also grab DEAR and ESR before the host can clobber them. */
@@ -169,20 +169,20 @@ _GLOBAL(kvmppc_resume_host)
 ..skip_esr:
 
        /* Save remaining volatile guest register state to vcpu. */
-       stw     r0, VCPU_GPR(r0)(r4)
-       stw     r1, VCPU_GPR(r1)(r4)
-       stw     r2, VCPU_GPR(r2)(r4)
-       stw     r10, VCPU_GPR(r10)(r4)
-       stw     r11, VCPU_GPR(r11)(r4)
-       stw     r12, VCPU_GPR(r12)(r4)
-       stw     r13, VCPU_GPR(r13)(r4)
-       stw     r14, VCPU_GPR(r14)(r4) /* We need a NV GPR below. */
+       stw     r0, VCPU_GPR(R0)(r4)
+       stw     r1, VCPU_GPR(R1)(r4)
+       stw     r2, VCPU_GPR(R2)(r4)
+       stw     r10, VCPU_GPR(R10)(r4)
+       stw     r11, VCPU_GPR(R11)(r4)
+       stw     r12, VCPU_GPR(R12)(r4)
+       stw     r13, VCPU_GPR(R13)(r4)
+       stw     r14, VCPU_GPR(R14)(r4) /* We need a NV GPR below. */
        mflr    r3
        stw     r3, VCPU_LR(r4)
        mfxer   r3
        stw     r3, VCPU_XER(r4)
        mfspr   r3, SPRN_SPRG_RSCRATCH0
-       stw     r3, VCPU_GPR(r4)(r4)
+       stw     r3, VCPU_GPR(R4)(r4)
        mfspr   r3, SPRN_SRR0
        stw     r3, VCPU_PC(r4)
 
@@ -214,28 +214,28 @@ _GLOBAL(kvmppc_resume_host)
 
        /* Restore vcpu pointer and the nonvolatiles we used. */
        mr      r4, r14
-       lwz     r14, VCPU_GPR(r14)(r4)
+       lwz     r14, VCPU_GPR(R14)(r4)
 
        /* Sometimes instruction emulation must restore complete GPR state. */
        andi.   r5, r3, RESUME_FLAG_NV
        beq     ..skip_nv_load
-       lwz     r15, VCPU_GPR(r15)(r4)
-       lwz     r16, VCPU_GPR(r16)(r4)
-       lwz     r17, VCPU_GPR(r17)(r4)
-       lwz     r18, VCPU_GPR(r18)(r4)
-       lwz     r19, VCPU_GPR(r19)(r4)
-       lwz     r20, VCPU_GPR(r20)(r4)
-       lwz     r21, VCPU_GPR(r21)(r4)
-       lwz     r22, VCPU_GPR(r22)(r4)
-       lwz     r23, VCPU_GPR(r23)(r4)
-       lwz     r24, VCPU_GPR(r24)(r4)
-       lwz     r25, VCPU_GPR(r25)(r4)
-       lwz     r26, VCPU_GPR(r26)(r4)
-       lwz     r27, VCPU_GPR(r27)(r4)
-       lwz     r28, VCPU_GPR(r28)(r4)
-       lwz     r29, VCPU_GPR(r29)(r4)
-       lwz     r30, VCPU_GPR(r30)(r4)
-       lwz     r31, VCPU_GPR(r31)(r4)
+       lwz     r15, VCPU_GPR(R15)(r4)
+       lwz     r16, VCPU_GPR(R16)(r4)
+       lwz     r17, VCPU_GPR(R17)(r4)
+       lwz     r18, VCPU_GPR(R18)(r4)
+       lwz     r19, VCPU_GPR(R19)(r4)
+       lwz     r20, VCPU_GPR(R20)(r4)
+       lwz     r21, VCPU_GPR(R21)(r4)
+       lwz     r22, VCPU_GPR(R22)(r4)
+       lwz     r23, VCPU_GPR(R23)(r4)
+       lwz     r24, VCPU_GPR(R24)(r4)
+       lwz     r25, VCPU_GPR(R25)(r4)
+       lwz     r26, VCPU_GPR(R26)(r4)
+       lwz     r27, VCPU_GPR(R27)(r4)
+       lwz     r28, VCPU_GPR(R28)(r4)
+       lwz     r29, VCPU_GPR(R29)(r4)
+       lwz     r30, VCPU_GPR(R30)(r4)
+       lwz     r31, VCPU_GPR(R31)(r4)
 ..skip_nv_load:
 
        /* Should we return to the guest? */
@@ -257,43 +257,43 @@ heavyweight_exit:
 
        /* We already saved guest volatile register state; now save the
         * non-volatiles. */
-       stw     r15, VCPU_GPR(r15)(r4)
-       stw     r16, VCPU_GPR(r16)(r4)
-       stw     r17, VCPU_GPR(r17)(r4)
-       stw     r18, VCPU_GPR(r18)(r4)
-       stw     r19, VCPU_GPR(r19)(r4)
-       stw     r20, VCPU_GPR(r20)(r4)
-       stw     r21, VCPU_GPR(r21)(r4)
-       stw     r22, VCPU_GPR(r22)(r4)
-       stw     r23, VCPU_GPR(r23)(r4)
-       stw     r24, VCPU_GPR(r24)(r4)
-       stw     r25, VCPU_GPR(r25)(r4)
-       stw     r26, VCPU_GPR(r26)(r4)
-       stw     r27, VCPU_GPR(r27)(r4)
-       stw     r28, VCPU_GPR(r28)(r4)
-       stw     r29, VCPU_GPR(r29)(r4)
-       stw     r30, VCPU_GPR(r30)(r4)
-       stw     r31, VCPU_GPR(r31)(r4)
+       stw     r15, VCPU_GPR(R15)(r4)
+       stw     r16, VCPU_GPR(R16)(r4)
+       stw     r17, VCPU_GPR(R17)(r4)
+       stw     r18, VCPU_GPR(R18)(r4)
+       stw     r19, VCPU_GPR(R19)(r4)
+       stw     r20, VCPU_GPR(R20)(r4)
+       stw     r21, VCPU_GPR(R21)(r4)
+       stw     r22, VCPU_GPR(R22)(r4)
+       stw     r23, VCPU_GPR(R23)(r4)
+       stw     r24, VCPU_GPR(R24)(r4)
+       stw     r25, VCPU_GPR(R25)(r4)
+       stw     r26, VCPU_GPR(R26)(r4)
+       stw     r27, VCPU_GPR(R27)(r4)
+       stw     r28, VCPU_GPR(R28)(r4)
+       stw     r29, VCPU_GPR(R29)(r4)
+       stw     r30, VCPU_GPR(R30)(r4)
+       stw     r31, VCPU_GPR(R31)(r4)
 
        /* Load host non-volatile register state from host stack. */
-       lwz     r14, HOST_NV_GPR(r14)(r1)
-       lwz     r15, HOST_NV_GPR(r15)(r1)
-       lwz     r16, HOST_NV_GPR(r16)(r1)
-       lwz     r17, HOST_NV_GPR(r17)(r1)
-       lwz     r18, HOST_NV_GPR(r18)(r1)
-       lwz     r19, HOST_NV_GPR(r19)(r1)
-       lwz     r20, HOST_NV_GPR(r20)(r1)
-       lwz     r21, HOST_NV_GPR(r21)(r1)
-       lwz     r22, HOST_NV_GPR(r22)(r1)
-       lwz     r23, HOST_NV_GPR(r23)(r1)
-       lwz     r24, HOST_NV_GPR(r24)(r1)
-       lwz     r25, HOST_NV_GPR(r25)(r1)
-       lwz     r26, HOST_NV_GPR(r26)(r1)
-       lwz     r27, HOST_NV_GPR(r27)(r1)
-       lwz     r28, HOST_NV_GPR(r28)(r1)
-       lwz     r29, HOST_NV_GPR(r29)(r1)
-       lwz     r30, HOST_NV_GPR(r30)(r1)
-       lwz     r31, HOST_NV_GPR(r31)(r1)
+       lwz     r14, HOST_NV_GPR(R14)(r1)
+       lwz     r15, HOST_NV_GPR(R15)(r1)
+       lwz     r16, HOST_NV_GPR(R16)(r1)
+       lwz     r17, HOST_NV_GPR(R17)(r1)
+       lwz     r18, HOST_NV_GPR(R18)(r1)
+       lwz     r19, HOST_NV_GPR(R19)(r1)
+       lwz     r20, HOST_NV_GPR(R20)(r1)
+       lwz     r21, HOST_NV_GPR(R21)(r1)
+       lwz     r22, HOST_NV_GPR(R22)(r1)
+       lwz     r23, HOST_NV_GPR(R23)(r1)
+       lwz     r24, HOST_NV_GPR(R24)(r1)
+       lwz     r25, HOST_NV_GPR(R25)(r1)
+       lwz     r26, HOST_NV_GPR(R26)(r1)
+       lwz     r27, HOST_NV_GPR(R27)(r1)
+       lwz     r28, HOST_NV_GPR(R28)(r1)
+       lwz     r29, HOST_NV_GPR(R29)(r1)
+       lwz     r30, HOST_NV_GPR(R30)(r1)
+       lwz     r31, HOST_NV_GPR(R31)(r1)
 
        /* Return to kvm_vcpu_run(). */
        lwz     r4, HOST_STACK_LR(r1)
@@ -321,44 +321,44 @@ _GLOBAL(__kvmppc_vcpu_run)
        stw     r5, HOST_CR(r1)
 
        /* Save host non-volatile register state to stack. */
-       stw     r14, HOST_NV_GPR(r14)(r1)
-       stw     r15, HOST_NV_GPR(r15)(r1)
-       stw     r16, HOST_NV_GPR(r16)(r1)
-       stw     r17, HOST_NV_GPR(r17)(r1)
-       stw     r18, HOST_NV_GPR(r18)(r1)
-       stw     r19, HOST_NV_GPR(r19)(r1)
-       stw     r20, HOST_NV_GPR(r20)(r1)
-       stw     r21, HOST_NV_GPR(r21)(r1)
-       stw     r22, HOST_NV_GPR(r22)(r1)
-       stw     r23, HOST_NV_GPR(r23)(r1)
-       stw     r24, HOST_NV_GPR(r24)(r1)
-       stw     r25, HOST_NV_GPR(r25)(r1)
-       stw     r26, HOST_NV_GPR(r26)(r1)
-       stw     r27, HOST_NV_GPR(r27)(r1)
-       stw     r28, HOST_NV_GPR(r28)(r1)
-       stw     r29, HOST_NV_GPR(r29)(r1)
-       stw     r30, HOST_NV_GPR(r30)(r1)
-       stw     r31, HOST_NV_GPR(r31)(r1)
+       stw     r14, HOST_NV_GPR(R14)(r1)
+       stw     r15, HOST_NV_GPR(R15)(r1)
+       stw     r16, HOST_NV_GPR(R16)(r1)
+       stw     r17, HOST_NV_GPR(R17)(r1)
+       stw     r18, HOST_NV_GPR(R18)(r1)
+       stw     r19, HOST_NV_GPR(R19)(r1)
+       stw     r20, HOST_NV_GPR(R20)(r1)
+       stw     r21, HOST_NV_GPR(R21)(r1)
+       stw     r22, HOST_NV_GPR(R22)(r1)
+       stw     r23, HOST_NV_GPR(R23)(r1)
+       stw     r24, HOST_NV_GPR(R24)(r1)
+       stw     r25, HOST_NV_GPR(R25)(r1)
+       stw     r26, HOST_NV_GPR(R26)(r1)
+       stw     r27, HOST_NV_GPR(R27)(r1)
+       stw     r28, HOST_NV_GPR(R28)(r1)
+       stw     r29, HOST_NV_GPR(R29)(r1)
+       stw     r30, HOST_NV_GPR(R30)(r1)
+       stw     r31, HOST_NV_GPR(R31)(r1)
 
        /* Load guest non-volatiles. */
-       lwz     r14, VCPU_GPR(r14)(r4)
-       lwz     r15, VCPU_GPR(r15)(r4)
-       lwz     r16, VCPU_GPR(r16)(r4)
-       lwz     r17, VCPU_GPR(r17)(r4)
-       lwz     r18, VCPU_GPR(r18)(r4)
-       lwz     r19, VCPU_GPR(r19)(r4)
-       lwz     r20, VCPU_GPR(r20)(r4)
-       lwz     r21, VCPU_GPR(r21)(r4)
-       lwz     r22, VCPU_GPR(r22)(r4)
-       lwz     r23, VCPU_GPR(r23)(r4)
-       lwz     r24, VCPU_GPR(r24)(r4)
-       lwz     r25, VCPU_GPR(r25)(r4)
-       lwz     r26, VCPU_GPR(r26)(r4)
-       lwz     r27, VCPU_GPR(r27)(r4)
-       lwz     r28, VCPU_GPR(r28)(r4)
-       lwz     r29, VCPU_GPR(r29)(r4)
-       lwz     r30, VCPU_GPR(r30)(r4)
-       lwz     r31, VCPU_GPR(r31)(r4)
+       lwz     r14, VCPU_GPR(R14)(r4)
+       lwz     r15, VCPU_GPR(R15)(r4)
+       lwz     r16, VCPU_GPR(R16)(r4)
+       lwz     r17, VCPU_GPR(R17)(r4)
+       lwz     r18, VCPU_GPR(R18)(r4)
+       lwz     r19, VCPU_GPR(R19)(r4)
+       lwz     r20, VCPU_GPR(R20)(r4)
+       lwz     r21, VCPU_GPR(R21)(r4)
+       lwz     r22, VCPU_GPR(R22)(r4)
+       lwz     r23, VCPU_GPR(R23)(r4)
+       lwz     r24, VCPU_GPR(R24)(r4)
+       lwz     r25, VCPU_GPR(R25)(r4)
+       lwz     r26, VCPU_GPR(R26)(r4)
+       lwz     r27, VCPU_GPR(R27)(r4)
+       lwz     r28, VCPU_GPR(R28)(r4)
+       lwz     r29, VCPU_GPR(R29)(r4)
+       lwz     r30, VCPU_GPR(R30)(r4)
+       lwz     r31, VCPU_GPR(R31)(r4)
 
 #ifdef CONFIG_SPE
        /* save host SPEFSCR and load guest SPEFSCR */
@@ -386,13 +386,13 @@ lightweight_exit:
 #endif
 
        /* Load some guest volatiles. */
-       lwz     r0, VCPU_GPR(r0)(r4)
-       lwz     r2, VCPU_GPR(r2)(r4)
-       lwz     r9, VCPU_GPR(r9)(r4)
-       lwz     r10, VCPU_GPR(r10)(r4)
-       lwz     r11, VCPU_GPR(r11)(r4)
-       lwz     r12, VCPU_GPR(r12)(r4)
-       lwz     r13, VCPU_GPR(r13)(r4)
+       lwz     r0, VCPU_GPR(R0)(r4)
+       lwz     r2, VCPU_GPR(R2)(r4)
+       lwz     r9, VCPU_GPR(R9)(r4)
+       lwz     r10, VCPU_GPR(R10)(r4)
+       lwz     r11, VCPU_GPR(R11)(r4)
+       lwz     r12, VCPU_GPR(R12)(r4)
+       lwz     r13, VCPU_GPR(R13)(r4)
        lwz     r3, VCPU_LR(r4)
        mtlr    r3
        lwz     r3, VCPU_XER(r4)
@@ -411,7 +411,7 @@ lightweight_exit:
 
        /* Can't switch the stack pointer until after IVPR is switched,
         * because host interrupt handlers would get confused. */
-       lwz     r1, VCPU_GPR(r1)(r4)
+       lwz     r1, VCPU_GPR(R1)(r4)
 
        /*
         * Host interrupt handlers may have clobbered these
@@ -449,10 +449,10 @@ lightweight_exit:
        mtcr    r5
        mtsrr0  r6
        mtsrr1  r7
-       lwz     r5, VCPU_GPR(r5)(r4)
-       lwz     r6, VCPU_GPR(r6)(r4)
-       lwz     r7, VCPU_GPR(r7)(r4)
-       lwz     r8, VCPU_GPR(r8)(r4)
+       lwz     r5, VCPU_GPR(R5)(r4)
+       lwz     r6, VCPU_GPR(R6)(r4)
+       lwz     r7, VCPU_GPR(R7)(r4)
+       lwz     r8, VCPU_GPR(R8)(r4)
 
        /* Clear any debug events which occurred since we disabled MSR[DE].
         * XXX This gives us a 3-instruction window in which a breakpoint
@@ -461,8 +461,8 @@ lightweight_exit:
        ori     r3, r3, 0xffff
        mtspr   SPRN_DBSR, r3
 
-       lwz     r3, VCPU_GPR(r3)(r4)
-       lwz     r4, VCPU_GPR(r4)(r4)
+       lwz     r3, VCPU_GPR(R3)(r4)
+       lwz     r4, VCPU_GPR(R4)(r4)
        rfi
 
 #ifdef CONFIG_SPE
index 6048a00..a623b1d 100644 (file)
  */
 .macro kvm_handler_common intno, srr0, flags
        /* Restore host stack pointer */
-       PPC_STL r1, VCPU_GPR(r1)(r4)
-       PPC_STL r2, VCPU_GPR(r2)(r4)
+       PPC_STL r1, VCPU_GPR(R1)(r4)
+       PPC_STL r2, VCPU_GPR(R2)(r4)
        PPC_LL  r1, VCPU_HOST_STACK(r4)
        PPC_LL  r2, HOST_R2(r1)
 
        mfspr   r10, SPRN_PID
        lwz     r8, VCPU_HOST_PID(r4)
        PPC_LL  r11, VCPU_SHARED(r4)
-       PPC_STL r14, VCPU_GPR(r14)(r4) /* We need a non-volatile GPR. */
+       PPC_STL r14, VCPU_GPR(R14)(r4) /* We need a non-volatile GPR. */
        li      r14, \intno
 
        stw     r10, VCPU_GUEST_PID(r4)
         */
 
        mfspr   r3, SPRN_EPLC   /* will already have correct ELPID and EGS */
-       PPC_STL r15, VCPU_GPR(r15)(r4)
-       PPC_STL r16, VCPU_GPR(r16)(r4)
-       PPC_STL r17, VCPU_GPR(r17)(r4)
-       PPC_STL r18, VCPU_GPR(r18)(r4)
-       PPC_STL r19, VCPU_GPR(r19)(r4)
+       PPC_STL r15, VCPU_GPR(R15)(r4)
+       PPC_STL r16, VCPU_GPR(R16)(r4)
+       PPC_STL r17, VCPU_GPR(R17)(r4)
+       PPC_STL r18, VCPU_GPR(R18)(r4)
+       PPC_STL r19, VCPU_GPR(R19)(r4)
        mr      r8, r3
-       PPC_STL r20, VCPU_GPR(r20)(r4)
+       PPC_STL r20, VCPU_GPR(R20)(r4)
        rlwimi  r8, r6, EPC_EAS_SHIFT - MSR_IR_LG, EPC_EAS
-       PPC_STL r21, VCPU_GPR(r21)(r4)
+       PPC_STL r21, VCPU_GPR(R21)(r4)
        rlwimi  r8, r6, EPC_EPR_SHIFT - MSR_PR_LG, EPC_EPR
-       PPC_STL r22, VCPU_GPR(r22)(r4)
+       PPC_STL r22, VCPU_GPR(R22)(r4)
        rlwimi  r8, r10, EPC_EPID_SHIFT, EPC_EPID
-       PPC_STL r23, VCPU_GPR(r23)(r4)
-       PPC_STL r24, VCPU_GPR(r24)(r4)
-       PPC_STL r25, VCPU_GPR(r25)(r4)
-       PPC_STL r26, VCPU_GPR(r26)(r4)
-       PPC_STL r27, VCPU_GPR(r27)(r4)
-       PPC_STL r28, VCPU_GPR(r28)(r4)
-       PPC_STL r29, VCPU_GPR(r29)(r4)
-       PPC_STL r30, VCPU_GPR(r30)(r4)
-       PPC_STL r31, VCPU_GPR(r31)(r4)
+       PPC_STL r23, VCPU_GPR(R23)(r4)
+       PPC_STL r24, VCPU_GPR(R24)(r4)
+       PPC_STL r25, VCPU_GPR(R25)(r4)
+       PPC_STL r26, VCPU_GPR(R26)(r4)
+       PPC_STL r27, VCPU_GPR(R27)(r4)
+       PPC_STL r28, VCPU_GPR(R28)(r4)
+       PPC_STL r29, VCPU_GPR(R29)(r4)
+       PPC_STL r30, VCPU_GPR(R30)(r4)
+       PPC_STL r31, VCPU_GPR(R31)(r4)
        mtspr   SPRN_EPLC, r8
 
        /* disable preemption, so we are sure we hit the fixup handler */
 .macro kvm_handler intno srr0, srr1, flags
 _GLOBAL(kvmppc_handler_\intno\()_\srr1)
        GET_VCPU(r11, r10)
-       PPC_STL r3, VCPU_GPR(r3)(r11)
+       PPC_STL r3, VCPU_GPR(R3)(r11)
        mfspr   r3, SPRN_SPRG_RSCRATCH0
-       PPC_STL r4, VCPU_GPR(r4)(r11)
+       PPC_STL r4, VCPU_GPR(R4)(r11)
        PPC_LL  r4, THREAD_NORMSAVE(0)(r10)
-       PPC_STL r5, VCPU_GPR(r5)(r11)
+       PPC_STL r5, VCPU_GPR(R5)(r11)
        stw     r13, VCPU_CR(r11)
        mfspr   r5, \srr0
-       PPC_STL r3, VCPU_GPR(r10)(r11)
+       PPC_STL r3, VCPU_GPR(R10)(r11)
        PPC_LL  r3, THREAD_NORMSAVE(2)(r10)
-       PPC_STL r6, VCPU_GPR(r6)(r11)
-       PPC_STL r4, VCPU_GPR(r11)(r11)
+       PPC_STL r6, VCPU_GPR(R6)(r11)
+       PPC_STL r4, VCPU_GPR(R11)(r11)
        mfspr   r6, \srr1
-       PPC_STL r7, VCPU_GPR(r7)(r11)
-       PPC_STL r8, VCPU_GPR(r8)(r11)
-       PPC_STL r9, VCPU_GPR(r9)(r11)
-       PPC_STL r3, VCPU_GPR(r13)(r11)
+       PPC_STL r7, VCPU_GPR(R7)(r11)
+       PPC_STL r8, VCPU_GPR(R8)(r11)
+       PPC_STL r9, VCPU_GPR(R9)(r11)
+       PPC_STL r3, VCPU_GPR(R13)(r11)
        mfctr   r7
-       PPC_STL r12, VCPU_GPR(r12)(r11)
+       PPC_STL r12, VCPU_GPR(R12)(r11)
        PPC_STL r7, VCPU_CTR(r11)
        mr      r4, r11
        kvm_handler_common \intno, \srr0, \flags
@@ -238,25 +238,25 @@ _GLOBAL(kvmppc_handler_\intno\()_\srr1)
 _GLOBAL(kvmppc_handler_\intno\()_\srr1)
        mfspr   r10, SPRN_SPRG_THREAD
        GET_VCPU(r11, r10)
-       PPC_STL r3, VCPU_GPR(r3)(r11)
+       PPC_STL r3, VCPU_GPR(R3)(r11)
        mfspr   r3, \scratch
-       PPC_STL r4, VCPU_GPR(r4)(r11)
+       PPC_STL r4, VCPU_GPR(R4)(r11)
        PPC_LL  r4, GPR9(r8)
-       PPC_STL r5, VCPU_GPR(r5)(r11)
+       PPC_STL r5, VCPU_GPR(R5)(r11)
        stw     r9, VCPU_CR(r11)
        mfspr   r5, \srr0
-       PPC_STL r3, VCPU_GPR(r8)(r11)
+       PPC_STL r3, VCPU_GPR(R8)(r11)
        PPC_LL  r3, GPR10(r8)
-       PPC_STL r6, VCPU_GPR(r6)(r11)
-       PPC_STL r4, VCPU_GPR(r9)(r11)
+       PPC_STL r6, VCPU_GPR(R6)(r11)
+       PPC_STL r4, VCPU_GPR(R9)(r11)
        mfspr   r6, \srr1
        PPC_LL  r4, GPR11(r8)
-       PPC_STL r7, VCPU_GPR(r7)(r11)
-       PPC_STL r3, VCPU_GPR(r10)(r11)
+       PPC_STL r7, VCPU_GPR(R7)(r11)
+       PPC_STL r3, VCPU_GPR(R10)(r11)
        mfctr   r7
-       PPC_STL r12, VCPU_GPR(r12)(r11)
-       PPC_STL r13, VCPU_GPR(r13)(r11)
-       PPC_STL r4, VCPU_GPR(r11)(r11)
+       PPC_STL r12, VCPU_GPR(R12)(r11)
+       PPC_STL r13, VCPU_GPR(R13)(r11)
+       PPC_STL r4, VCPU_GPR(R11)(r11)
        PPC_STL r7, VCPU_CTR(r11)
        mr      r4, r11
        kvm_handler_common \intno, \srr0, \flags
@@ -310,7 +310,7 @@ kvm_lvl_handler BOOKE_INTERRUPT_DEBUG, \
 _GLOBAL(kvmppc_resume_host)
        /* Save remaining volatile guest register state to vcpu. */
        mfspr   r3, SPRN_VRSAVE
-       PPC_STL r0, VCPU_GPR(r0)(r4)
+       PPC_STL r0, VCPU_GPR(R0)(r4)
        mflr    r5
        mfspr   r6, SPRN_SPRG4
        PPC_STL r5, VCPU_LR(r4)
@@ -358,27 +358,27 @@ _GLOBAL(kvmppc_resume_host)
 
        /* Restore vcpu pointer and the nonvolatiles we used. */
        mr      r4, r14
-       PPC_LL  r14, VCPU_GPR(r14)(r4)
+       PPC_LL  r14, VCPU_GPR(R14)(r4)
 
        andi.   r5, r3, RESUME_FLAG_NV
        beq     skip_nv_load
-       PPC_LL  r15, VCPU_GPR(r15)(r4)
-       PPC_LL  r16, VCPU_GPR(r16)(r4)
-       PPC_LL  r17, VCPU_GPR(r17)(r4)
-       PPC_LL  r18, VCPU_GPR(r18)(r4)
-       PPC_LL  r19, VCPU_GPR(r19)(r4)
-       PPC_LL  r20, VCPU_GPR(r20)(r4)
-       PPC_LL  r21, VCPU_GPR(r21)(r4)
-       PPC_LL  r22, VCPU_GPR(r22)(r4)
-       PPC_LL  r23, VCPU_GPR(r23)(r4)
-       PPC_LL  r24, VCPU_GPR(r24)(r4)
-       PPC_LL  r25, VCPU_GPR(r25)(r4)
-       PPC_LL  r26, VCPU_GPR(r26)(r4)
-       PPC_LL  r27, VCPU_GPR(r27)(r4)
-       PPC_LL  r28, VCPU_GPR(r28)(r4)
-       PPC_LL  r29, VCPU_GPR(r29)(r4)
-       PPC_LL  r30, VCPU_GPR(r30)(r4)
-       PPC_LL  r31, VCPU_GPR(r31)(r4)
+       PPC_LL  r15, VCPU_GPR(R15)(r4)
+       PPC_LL  r16, VCPU_GPR(R16)(r4)
+       PPC_LL  r17, VCPU_GPR(R17)(r4)
+       PPC_LL  r18, VCPU_GPR(R18)(r4)
+       PPC_LL  r19, VCPU_GPR(R19)(r4)
+       PPC_LL  r20, VCPU_GPR(R20)(r4)
+       PPC_LL  r21, VCPU_GPR(R21)(r4)
+       PPC_LL  r22, VCPU_GPR(R22)(r4)
+       PPC_LL  r23, VCPU_GPR(R23)(r4)
+       PPC_LL  r24, VCPU_GPR(R24)(r4)
+       PPC_LL  r25, VCPU_GPR(R25)(r4)
+       PPC_LL  r26, VCPU_GPR(R26)(r4)
+       PPC_LL  r27, VCPU_GPR(R27)(r4)
+       PPC_LL  r28, VCPU_GPR(R28)(r4)
+       PPC_LL  r29, VCPU_GPR(R29)(r4)
+       PPC_LL  r30, VCPU_GPR(R30)(r4)
+       PPC_LL  r31, VCPU_GPR(R31)(r4)
 skip_nv_load:
        /* Should we return to the guest? */
        andi.   r5, r3, RESUME_FLAG_HOST
@@ -396,23 +396,23 @@ heavyweight_exit:
         * non-volatiles.
         */
 
-       PPC_STL r15, VCPU_GPR(r15)(r4)
-       PPC_STL r16, VCPU_GPR(r16)(r4)
-       PPC_STL r17, VCPU_GPR(r17)(r4)
-       PPC_STL r18, VCPU_GPR(r18)(r4)
-       PPC_STL r19, VCPU_GPR(r19)(r4)
-       PPC_STL r20, VCPU_GPR(r20)(r4)
-       PPC_STL r21, VCPU_GPR(r21)(r4)
-       PPC_STL r22, VCPU_GPR(r22)(r4)
-       PPC_STL r23, VCPU_GPR(r23)(r4)
-       PPC_STL r24, VCPU_GPR(r24)(r4)
-       PPC_STL r25, VCPU_GPR(r25)(r4)
-       PPC_STL r26, VCPU_GPR(r26)(r4)
-       PPC_STL r27, VCPU_GPR(r27)(r4)
-       PPC_STL r28, VCPU_GPR(r28)(r4)
-       PPC_STL r29, VCPU_GPR(r29)(r4)
-       PPC_STL r30, VCPU_GPR(r30)(r4)
-       PPC_STL r31, VCPU_GPR(r31)(r4)
+       PPC_STL r15, VCPU_GPR(R15)(r4)
+       PPC_STL r16, VCPU_GPR(R16)(r4)
+       PPC_STL r17, VCPU_GPR(R17)(r4)
+       PPC_STL r18, VCPU_GPR(R18)(r4)
+       PPC_STL r19, VCPU_GPR(R19)(r4)
+       PPC_STL r20, VCPU_GPR(R20)(r4)
+       PPC_STL r21, VCPU_GPR(R21)(r4)
+       PPC_STL r22, VCPU_GPR(R22)(r4)
+       PPC_STL r23, VCPU_GPR(R23)(r4)
+       PPC_STL r24, VCPU_GPR(R24)(r4)
+       PPC_STL r25, VCPU_GPR(R25)(r4)
+       PPC_STL r26, VCPU_GPR(R26)(r4)
+       PPC_STL r27, VCPU_GPR(R27)(r4)
+       PPC_STL r28, VCPU_GPR(R28)(r4)
+       PPC_STL r29, VCPU_GPR(R29)(r4)
+       PPC_STL r30, VCPU_GPR(R30)(r4)
+       PPC_STL r31, VCPU_GPR(R31)(r4)
 
        /* Load host non-volatile register state from host stack. */
        PPC_LL  r14, HOST_NV_GPR(r14)(r1)
@@ -478,24 +478,24 @@ _GLOBAL(__kvmppc_vcpu_run)
        PPC_STL r31, HOST_NV_GPR(r31)(r1)
 
        /* Load guest non-volatiles. */
-       PPC_LL  r14, VCPU_GPR(r14)(r4)
-       PPC_LL  r15, VCPU_GPR(r15)(r4)
-       PPC_LL  r16, VCPU_GPR(r16)(r4)
-       PPC_LL  r17, VCPU_GPR(r17)(r4)
-       PPC_LL  r18, VCPU_GPR(r18)(r4)
-       PPC_LL  r19, VCPU_GPR(r19)(r4)
-       PPC_LL  r20, VCPU_GPR(r20)(r4)
-       PPC_LL  r21, VCPU_GPR(r21)(r4)
-       PPC_LL  r22, VCPU_GPR(r22)(r4)
-       PPC_LL  r23, VCPU_GPR(r23)(r4)
-       PPC_LL  r24, VCPU_GPR(r24)(r4)
-       PPC_LL  r25, VCPU_GPR(r25)(r4)
-       PPC_LL  r26, VCPU_GPR(r26)(r4)
-       PPC_LL  r27, VCPU_GPR(r27)(r4)
-       PPC_LL  r28, VCPU_GPR(r28)(r4)
-       PPC_LL  r29, VCPU_GPR(r29)(r4)
-       PPC_LL  r30, VCPU_GPR(r30)(r4)
-       PPC_LL  r31, VCPU_GPR(r31)(r4)
+       PPC_LL  r14, VCPU_GPR(R14)(r4)
+       PPC_LL  r15, VCPU_GPR(R15)(r4)
+       PPC_LL  r16, VCPU_GPR(R16)(r4)
+       PPC_LL  r17, VCPU_GPR(R17)(r4)
+       PPC_LL  r18, VCPU_GPR(R18)(r4)
+       PPC_LL  r19, VCPU_GPR(R19)(r4)
+       PPC_LL  r20, VCPU_GPR(R20)(r4)
+       PPC_LL  r21, VCPU_GPR(R21)(r4)
+       PPC_LL  r22, VCPU_GPR(R22)(r4)
+       PPC_LL  r23, VCPU_GPR(R23)(r4)
+       PPC_LL  r24, VCPU_GPR(R24)(r4)
+       PPC_LL  r25, VCPU_GPR(R25)(r4)
+       PPC_LL  r26, VCPU_GPR(R26)(r4)
+       PPC_LL  r27, VCPU_GPR(R27)(r4)
+       PPC_LL  r28, VCPU_GPR(R28)(r4)
+       PPC_LL  r29, VCPU_GPR(R29)(r4)
+       PPC_LL  r30, VCPU_GPR(R30)(r4)
+       PPC_LL  r31, VCPU_GPR(R31)(r4)
 
 
 lightweight_exit:
@@ -554,13 +554,13 @@ lightweight_exit:
        lwz     r7, VCPU_CR(r4)
        PPC_LL  r8, VCPU_PC(r4)
        PPC_LD(r9, VCPU_SHARED_MSR, r11)
-       PPC_LL  r0, VCPU_GPR(r0)(r4)
-       PPC_LL  r1, VCPU_GPR(r1)(r4)
-       PPC_LL  r2, VCPU_GPR(r2)(r4)
-       PPC_LL  r10, VCPU_GPR(r10)(r4)
-       PPC_LL  r11, VCPU_GPR(r11)(r4)
-       PPC_LL  r12, VCPU_GPR(r12)(r4)
-       PPC_LL  r13, VCPU_GPR(r13)(r4)
+       PPC_LL  r0, VCPU_GPR(R0)(r4)
+       PPC_LL  r1, VCPU_GPR(R1)(r4)
+       PPC_LL  r2, VCPU_GPR(R2)(r4)
+       PPC_LL  r10, VCPU_GPR(R10)(r4)
+       PPC_LL  r11, VCPU_GPR(R11)(r4)
+       PPC_LL  r12, VCPU_GPR(R12)(r4)
+       PPC_LL  r13, VCPU_GPR(R13)(r4)
        mtlr    r3
        mtxer   r5
        mtctr   r6
@@ -586,12 +586,12 @@ lightweight_exit:
        mtcr    r7
 
        /* Finish loading guest volatiles and jump to guest. */
-       PPC_LL  r5, VCPU_GPR(r5)(r4)
-       PPC_LL  r6, VCPU_GPR(r6)(r4)
-       PPC_LL  r7, VCPU_GPR(r7)(r4)
-       PPC_LL  r8, VCPU_GPR(r8)(r4)
-       PPC_LL  r9, VCPU_GPR(r9)(r4)
-
-       PPC_LL  r3, VCPU_GPR(r3)(r4)
-       PPC_LL  r4, VCPU_GPR(r4)(r4)
+       PPC_LL  r5, VCPU_GPR(R5)(r4)
+       PPC_LL  r6, VCPU_GPR(R6)(r4)
+       PPC_LL  r7, VCPU_GPR(R7)(r4)
+       PPC_LL  r8, VCPU_GPR(R8)(r4)
+       PPC_LL  r9, VCPU_GPR(R9)(r4)
+
+       PPC_LL  r3, VCPU_GPR(R3)(r4)
+       PPC_LL  r4, VCPU_GPR(R4)(r4)
        rfi
index 18245af..7ca6871 100644 (file)
@@ -114,9 +114,9 @@ _GLOBAL(csum_partial)
        mtctr   r6
 
        stdu    r1,-STACKFRAMESIZE(r1)
-       std     r14,STK_REG(r14)(r1)
-       std     r15,STK_REG(r15)(r1)
-       std     r16,STK_REG(r16)(r1)
+       std     r14,STK_REG(R14)(r1)
+       std     r15,STK_REG(R15)(r1)
+       std     r16,STK_REG(R16)(r1)
 
        ld      r6,0(r3)
        ld      r9,8(r3)
@@ -175,9 +175,9 @@ _GLOBAL(csum_partial)
        adde    r0,r0,r15
        adde    r0,r0,r16
 
-       ld      r14,STK_REG(r14)(r1)
-       ld      r15,STK_REG(r15)(r1)
-       ld      r16,STK_REG(r16)(r1)
+       ld      r14,STK_REG(R14)(r1)
+       ld      r15,STK_REG(R15)(r1)
+       ld      r16,STK_REG(R16)(r1)
        addi    r1,r1,STACKFRAMESIZE
 
        andi.   r4,r4,63
@@ -299,9 +299,9 @@ dest;       sth     r6,0(r4)
        mtctr   r6
 
        stdu    r1,-STACKFRAMESIZE(r1)
-       std     r14,STK_REG(r14)(r1)
-       std     r15,STK_REG(r15)(r1)
-       std     r16,STK_REG(r16)(r1)
+       std     r14,STK_REG(R14)(r1)
+       std     r15,STK_REG(R15)(r1)
+       std     r16,STK_REG(R16)(r1)
 
 source;        ld      r6,0(r3)
 source;        ld      r9,8(r3)
@@ -382,9 +382,9 @@ dest;       std     r16,56(r4)
        adde    r0,r0,r15
        adde    r0,r0,r16
 
-       ld      r14,STK_REG(r14)(r1)
-       ld      r15,STK_REG(r15)(r1)
-       ld      r16,STK_REG(r16)(r1)
+       ld      r14,STK_REG(R14)(r1)
+       ld      r15,STK_REG(R15)(r1)
+       ld      r16,STK_REG(R16)(r1)
        addi    r1,r1,STACKFRAMESIZE
 
        andi.   r5,r5,63
index 01e2b5d..a2126ce 100644 (file)
@@ -113,13 +113,13 @@ _GLOBAL(copypage_power7)
 #endif
 
 .Lnonvmx_copy:
-       std     r14,STK_REG(r14)(r1)
-       std     r15,STK_REG(r15)(r1)
-       std     r16,STK_REG(r16)(r1)
-       std     r17,STK_REG(r17)(r1)
-       std     r18,STK_REG(r18)(r1)
-       std     r19,STK_REG(r19)(r1)
-       std     r20,STK_REG(r20)(r1)
+       std     r14,STK_REG(R14)(r1)
+       std     r15,STK_REG(R15)(r1)
+       std     r16,STK_REG(R16)(r1)
+       std     r17,STK_REG(R17)(r1)
+       std     r18,STK_REG(R18)(r1)
+       std     r19,STK_REG(R19)(r1)
+       std     r20,STK_REG(R20)(r1)
 
 1:     ld      r0,0(r4)
        ld      r5,8(r4)
@@ -157,12 +157,12 @@ _GLOBAL(copypage_power7)
        addi    r3,r3,128
        bdnz    1b
 
-       ld      r14,STK_REG(r14)(r1)
-       ld      r15,STK_REG(r15)(r1)
-       ld      r16,STK_REG(r16)(r1)
-       ld      r17,STK_REG(r17)(r1)
-       ld      r18,STK_REG(r18)(r1)
-       ld      r19,STK_REG(r19)(r1)
-       ld      r20,STK_REG(r20)(r1)
+       ld      r14,STK_REG(R14)(r1)
+       ld      r15,STK_REG(R15)(r1)
+       ld      r16,STK_REG(R16)(r1)
+       ld      r17,STK_REG(R17)(r1)
+       ld      r18,STK_REG(R18)(r1)
+       ld      r19,STK_REG(R19)(r1)
+       ld      r20,STK_REG(R20)(r1)
        addi    r1,r1,STACKFRAMESIZE
        blr
index d73a590..f47d05a 100644 (file)
@@ -30,7 +30,7 @@ _GLOBAL(__copy_tofrom_user_base)
        dcbt    0,r4
        beq     .Lcopy_page_4K
        andi.   r6,r6,7
-       PPC_MTOCRF(0x01,r5)
+       PPC_MTOCRF(0x01,R5)
        blt     cr1,.Lshort_copy
 /* Below we want to nop out the bne if we're on a CPU that has the
  * CPU_FTR_UNALIGNED_LD_STD bit set and the CPU_FTR_CP_USE_DCBTZ bit
@@ -186,7 +186,7 @@ END_FTR_SECTION_IFCLR(CPU_FTR_UNALIGNED_LD_STD)
        blr
 
 .Ldst_unaligned:
-       PPC_MTOCRF(0x01,r6)             /* put #bytes to 8B bdry into cr7 */
+       PPC_MTOCRF(0x01,R6)             /* put #bytes to 8B bdry into cr7 */
        subf    r5,r6,r5
        li      r7,0
        cmpldi  cr1,r5,16
@@ -201,7 +201,7 @@ END_FTR_SECTION_IFCLR(CPU_FTR_UNALIGNED_LD_STD)
 2:     bf      cr7*4+1,3f
 37:    lwzx    r0,r7,r4
 83:    stwx    r0,r7,r3
-3:     PPC_MTOCRF(0x01,r5)
+3:     PPC_MTOCRF(0x01,R5)
        add     r4,r6,r4
        add     r3,r6,r3
        b       .Ldst_aligned
index 48e3f8c..c8680c0 100644 (file)
@@ -57,9 +57,9 @@
 
 
 .Ldo_err4:
-       ld      r16,STK_REG(r16)(r1)
-       ld      r15,STK_REG(r15)(r1)
-       ld      r14,STK_REG(r14)(r1)
+       ld      r16,STK_REG(R16)(r1)
+       ld      r15,STK_REG(R15)(r1)
+       ld      r14,STK_REG(R14)(r1)
 .Ldo_err3:
        bl      .exit_vmx_usercopy
        ld      r0,STACKFRAMESIZE+16(r1)
 #endif /* CONFIG_ALTIVEC */
 
 .Ldo_err2:
-       ld      r22,STK_REG(r22)(r1)
-       ld      r21,STK_REG(r21)(r1)
-       ld      r20,STK_REG(r20)(r1)
-       ld      r19,STK_REG(r19)(r1)
-       ld      r18,STK_REG(r18)(r1)
-       ld      r17,STK_REG(r17)(r1)
-       ld      r16,STK_REG(r16)(r1)
-       ld      r15,STK_REG(r15)(r1)
-       ld      r14,STK_REG(r14)(r1)
+       ld      r22,STK_REG(R22)(r1)
+       ld      r21,STK_REG(R21)(r1)
+       ld      r20,STK_REG(R20)(r1)
+       ld      r19,STK_REG(R19)(r1)
+       ld      r18,STK_REG(R18)(r1)
+       ld      r17,STK_REG(R17)(r1)
+       ld      r16,STK_REG(R16)(r1)
+       ld      r15,STK_REG(R15)(r1)
+       ld      r14,STK_REG(R14)(r1)
 .Lexit:
        addi    r1,r1,STACKFRAMESIZE
 .Ldo_err1:
@@ -137,15 +137,15 @@ err1;     stw     r0,0(r3)
 
        mflr    r0
        stdu    r1,-STACKFRAMESIZE(r1)
-       std     r14,STK_REG(r14)(r1)
-       std     r15,STK_REG(r15)(r1)
-       std     r16,STK_REG(r16)(r1)
-       std     r17,STK_REG(r17)(r1)
-       std     r18,STK_REG(r18)(r1)
-       std     r19,STK_REG(r19)(r1)
-       std     r20,STK_REG(r20)(r1)
-       std     r21,STK_REG(r21)(r1)
-       std     r22,STK_REG(r22)(r1)
+       std     r14,STK_REG(R14)(r1)
+       std     r15,STK_REG(R15)(r1)
+       std     r16,STK_REG(R16)(r1)
+       std     r17,STK_REG(R17)(r1)
+       std     r18,STK_REG(R18)(r1)
+       std     r19,STK_REG(R19)(r1)
+       std     r20,STK_REG(R20)(r1)
+       std     r21,STK_REG(R21)(r1)
+       std     r22,STK_REG(R22)(r1)
        std     r0,STACKFRAMESIZE+16(r1)
 
        srdi    r6,r5,7
@@ -192,15 +192,15 @@ err2;     std     r21,120(r3)
 
        clrldi  r5,r5,(64-7)
 
-       ld      r14,STK_REG(r14)(r1)
-       ld      r15,STK_REG(r15)(r1)
-       ld      r16,STK_REG(r16)(r1)
-       ld      r17,STK_REG(r17)(r1)
-       ld      r18,STK_REG(r18)(r1)
-       ld      r19,STK_REG(r19)(r1)
-       ld      r20,STK_REG(r20)(r1)
-       ld      r21,STK_REG(r21)(r1)
-       ld      r22,STK_REG(r22)(r1)
+       ld      r14,STK_REG(R14)(r1)
+       ld      r15,STK_REG(R15)(r1)
+       ld      r16,STK_REG(R16)(r1)
+       ld      r17,STK_REG(R17)(r1)
+       ld      r18,STK_REG(R18)(r1)
+       ld      r19,STK_REG(R19)(r1)
+       ld      r20,STK_REG(R20)(r1)
+       ld      r21,STK_REG(R21)(r1)
+       ld      r22,STK_REG(R22)(r1)
        addi    r1,r1,STACKFRAMESIZE
 
        /* Up to 127B to go */
@@ -440,9 +440,9 @@ err3;       stvx    vr0,r3,r11
 7:     sub     r5,r5,r6
        srdi    r6,r5,7
 
-       std     r14,STK_REG(r14)(r1)
-       std     r15,STK_REG(r15)(r1)
-       std     r16,STK_REG(r16)(r1)
+       std     r14,STK_REG(R14)(r1)
+       std     r15,STK_REG(R15)(r1)
+       std     r16,STK_REG(R16)(r1)
 
        li      r12,64
        li      r14,80
@@ -477,9 +477,9 @@ err4;       stvx    vr0,r3,r16
        addi    r3,r3,128
        bdnz    8b
 
-       ld      r14,STK_REG(r14)(r1)
-       ld      r15,STK_REG(r15)(r1)
-       ld      r16,STK_REG(r16)(r1)
+       ld      r14,STK_REG(R14)(r1)
+       ld      r15,STK_REG(R15)(r1)
+       ld      r16,STK_REG(R16)(r1)
 
        /* Up to 127B to go */
        clrldi  r5,r5,(64-7)
@@ -625,9 +625,9 @@ err3;       stvx    vr11,r3,r11
 7:     sub     r5,r5,r6
        srdi    r6,r5,7
 
-       std     r14,STK_REG(r14)(r1)
-       std     r15,STK_REG(r15)(r1)
-       std     r16,STK_REG(r16)(r1)
+       std     r14,STK_REG(R14)(r1)
+       std     r15,STK_REG(R15)(r1)
+       std     r16,STK_REG(R16)(r1)
 
        li      r12,64
        li      r14,80
@@ -670,9 +670,9 @@ err4;       stvx    vr15,r3,r16
        addi    r3,r3,128
        bdnz    8b
 
-       ld      r14,STK_REG(r14)(r1)
-       ld      r15,STK_REG(r15)(r1)
-       ld      r16,STK_REG(r16)(r1)
+       ld      r14,STK_REG(R14)(r1)
+       ld      r15,STK_REG(R15)(r1)
+       ld      r16,STK_REG(R16)(r1)
 
        /* Up to 127B to go */
        clrldi  r5,r5,(64-7)
index fda2786..9b96ff2 100644 (file)
@@ -28,7 +28,7 @@ BEGIN_FTR_SECTION
        nop
        nop
 FTR_SECTION_ELSE
-       PPC_POPCNTB(r3,r3)
+       PPC_POPCNTB(R3,R3)
        clrldi  r3,r3,64-8
        blr
 ALT_FTR_SECTION_END_IFCLR(CPU_FTR_POPCNTB)
@@ -42,14 +42,14 @@ BEGIN_FTR_SECTION
        nop
 FTR_SECTION_ELSE
   BEGIN_FTR_SECTION_NESTED(50)
-       PPC_POPCNTB(r3,r3)
+       PPC_POPCNTB(R3,R3)
        srdi    r4,r3,8
        add     r3,r4,r3
        clrldi  r3,r3,64-8
        blr
   FTR_SECTION_ELSE_NESTED(50)
        clrlwi  r3,r3,16
-       PPC_POPCNTW(r3,r3)
+       PPC_POPCNTW(R3,R3)
        clrldi  r3,r3,64-8
        blr
   ALT_FTR_SECTION_END_NESTED_IFCLR(CPU_FTR_POPCNTD, 50)
@@ -66,7 +66,7 @@ BEGIN_FTR_SECTION
        nop
 FTR_SECTION_ELSE
   BEGIN_FTR_SECTION_NESTED(51)
-       PPC_POPCNTB(r3,r3)
+       PPC_POPCNTB(R3,R3)
        srdi    r4,r3,16
        add     r3,r4,r3
        srdi    r4,r3,8
@@ -74,7 +74,7 @@ FTR_SECTION_ELSE
        clrldi  r3,r3,64-8
        blr
   FTR_SECTION_ELSE_NESTED(51)
-       PPC_POPCNTW(r3,r3)
+       PPC_POPCNTW(R3,R3)
        clrldi  r3,r3,64-8
        blr
   ALT_FTR_SECTION_END_NESTED_IFCLR(CPU_FTR_POPCNTD, 51)
@@ -93,7 +93,7 @@ BEGIN_FTR_SECTION
        nop
 FTR_SECTION_ELSE
   BEGIN_FTR_SECTION_NESTED(52)
-       PPC_POPCNTB(r3,r3)
+       PPC_POPCNTB(R3,R3)
        srdi    r4,r3,32
        add     r3,r4,r3
        srdi    r4,r3,16
@@ -103,7 +103,7 @@ FTR_SECTION_ELSE
        clrldi  r3,r3,64-8
        blr
   FTR_SECTION_ELSE_NESTED(52)
-       PPC_POPCNTD(r3,r3)
+       PPC_POPCNTD(R3,R3)
        clrldi  r3,r3,64-8
        blr
   ALT_FTR_SECTION_END_NESTED_IFCLR(CPU_FTR_POPCNTD, 52)
index 6a85380..3abae6b 100644 (file)
@@ -330,13 +330,13 @@ _GLOBAL(do_lxvd2x)
        MTMSRD(r7)
        isync
        beq     cr7,1f
-       STXVD2X(0,r1,r8)
+       STXVD2X(0,R1,R8)
 1:     li      r9,-EFAULT
-2:     LXVD2X(0,0,r4)
+2:     LXVD2X(0,0,R4)
        li      r9,0
 3:     beq     cr7,4f
        bl      put_vsr
-       LXVD2X(0,r1,r8)
+       LXVD2X(0,R1,R8)
 4:     PPC_LL  r0,STKFRM+PPC_LR_STKOFF(r1)
        mtlr    r0
        MTMSRD(r6)
@@ -358,13 +358,13 @@ _GLOBAL(do_stxvd2x)
        MTMSRD(r7)
        isync
        beq     cr7,1f
-       STXVD2X(0,r1,r8)
+       STXVD2X(0,R1,R8)
        bl      get_vsr
 1:     li      r9,-EFAULT
-2:     STXVD2X(0,0,r4)
+2:     STXVD2X(0,0,R4)
        li      r9,0
 3:     beq     cr7,4f
-       LXVD2X(0,r1,r8)
+       LXVD2X(0,R1,R8)
 4:     PPC_LL  r0,STKFRM+PPC_LR_STKOFF(r1)
        mtlr    r0
        MTMSRD(r6)
index f4fcb0b..886bfc7 100644 (file)
@@ -19,7 +19,7 @@ _GLOBAL(memset)
        rlwimi  r4,r4,16,0,15
        cmplw   cr1,r5,r0               /* do we get that far? */
        rldimi  r4,r4,32,0
-       PPC_MTOCRF(1,r0)
+       PPC_MTOCRF(1,R0)
        mr      r6,r3
        blt     cr1,8f
        beq+    3f                      /* if already 8-byte aligned */
@@ -49,7 +49,7 @@ _GLOBAL(memset)
        bdnz    4b
 5:     srwi.   r0,r5,3
        clrlwi  r5,r5,29
-       PPC_MTOCRF(1,r0)
+       PPC_MTOCRF(1,R0)
        beq     8f
        bf      29,6f
        std     r4,0(r6)
@@ -65,7 +65,7 @@ _GLOBAL(memset)
        std     r4,0(r6)
        addi    r6,r6,8
 8:     cmpwi   r5,0
-       PPC_MTOCRF(1,r5)
+       PPC_MTOCRF(1,R5)
        beqlr+
        bf      29,9f
        stw     r4,0(r6)
index d2bbbc8..0a87b37 100644 (file)
@@ -16,7 +16,7 @@ BEGIN_FTR_SECTION
 FTR_SECTION_ELSE
        b       memcpy_power7
 ALT_FTR_SECTION_END_IFCLR(CPU_FTR_VMX_COPY)
-       PPC_MTOCRF(0x01,r5)
+       PPC_MTOCRF(0x01,R5)
        cmpldi  cr1,r5,16
        neg     r6,r3           # LS 3 bits = # bytes to 8-byte dest bdry
        andi.   r6,r6,7
@@ -158,7 +158,7 @@ END_FTR_SECTION_IFCLR(CPU_FTR_UNALIGNED_LD_STD)
        blr
 
 .Ldst_unaligned:
-       PPC_MTOCRF(0x01,r6)             # put #bytes to 8B bdry into cr7
+       PPC_MTOCRF(0x01,R6)             # put #bytes to 8B bdry into cr7
        subf    r5,r6,r5
        li      r7,0
        cmpldi  cr1,r5,16
@@ -173,7 +173,7 @@ END_FTR_SECTION_IFCLR(CPU_FTR_UNALIGNED_LD_STD)
 2:     bf      cr7*4+1,3f
        lwzx    r0,r7,r4
        stwx    r0,r7,r3
-3:     PPC_MTOCRF(0x01,r5)
+3:     PPC_MTOCRF(0x01,R5)
        add     r4,r6,r4
        add     r3,r6,r3
        b       .Ldst_aligned
index 84674d8..04524a2 100644 (file)
@@ -69,15 +69,15 @@ _GLOBAL(memcpy_power7)
 
        mflr    r0
        stdu    r1,-STACKFRAMESIZE(r1)
-       std     r14,STK_REG(r14)(r1)
-       std     r15,STK_REG(r15)(r1)
-       std     r16,STK_REG(r16)(r1)
-       std     r17,STK_REG(r17)(r1)
-       std     r18,STK_REG(r18)(r1)
-       std     r19,STK_REG(r19)(r1)
-       std     r20,STK_REG(r20)(r1)
-       std     r21,STK_REG(r21)(r1)
-       std     r22,STK_REG(r22)(r1)
+       std     r14,STK_REG(R14)(r1)
+       std     r15,STK_REG(R15)(r1)
+       std     r16,STK_REG(R16)(r1)
+       std     r17,STK_REG(R17)(r1)
+       std     r18,STK_REG(R18)(r1)
+       std     r19,STK_REG(R19)(r1)
+       std     r20,STK_REG(R20)(r1)
+       std     r21,STK_REG(R21)(r1)
+       std     r22,STK_REG(R22)(r1)
        std     r0,STACKFRAMESIZE+16(r1)
 
        srdi    r6,r5,7
@@ -124,15 +124,15 @@ _GLOBAL(memcpy_power7)
 
        clrldi  r5,r5,(64-7)
 
-       ld      r14,STK_REG(r14)(r1)
-       ld      r15,STK_REG(r15)(r1)
-       ld      r16,STK_REG(r16)(r1)
-       ld      r17,STK_REG(r17)(r1)
-       ld      r18,STK_REG(r18)(r1)
-       ld      r19,STK_REG(r19)(r1)
-       ld      r20,STK_REG(r20)(r1)
-       ld      r21,STK_REG(r21)(r1)
-       ld      r22,STK_REG(r22)(r1)
+       ld      r14,STK_REG(R14)(r1)
+       ld      r15,STK_REG(R15)(r1)
+       ld      r16,STK_REG(R16)(r1)
+       ld      r17,STK_REG(R17)(r1)
+       ld      r18,STK_REG(R18)(r1)
+       ld      r19,STK_REG(R19)(r1)
+       ld      r20,STK_REG(R20)(r1)
+       ld      r21,STK_REG(R21)(r1)
+       ld      r22,STK_REG(R22)(r1)
        addi    r1,r1,STACKFRAMESIZE
 
        /* Up to 127B to go */
@@ -343,9 +343,9 @@ _GLOBAL(memcpy_power7)
 7:     sub     r5,r5,r6
        srdi    r6,r5,7
 
-       std     r14,STK_REG(r14)(r1)
-       std     r15,STK_REG(r15)(r1)
-       std     r16,STK_REG(r16)(r1)
+       std     r14,STK_REG(R14)(r1)
+       std     r15,STK_REG(R15)(r1)
+       std     r16,STK_REG(R16)(r1)
 
        li      r12,64
        li      r14,80
@@ -380,9 +380,9 @@ _GLOBAL(memcpy_power7)
        addi    r3,r3,128
        bdnz    8b
 
-       ld      r14,STK_REG(r14)(r1)
-       ld      r15,STK_REG(r15)(r1)
-       ld      r16,STK_REG(r16)(r1)
+       ld      r14,STK_REG(R14)(r1)
+       ld      r15,STK_REG(R15)(r1)
+       ld      r16,STK_REG(R16)(r1)
 
        /* Up to 127B to go */
        clrldi  r5,r5,(64-7)
@@ -529,9 +529,9 @@ _GLOBAL(memcpy_power7)
 7:     sub     r5,r5,r6
        srdi    r6,r5,7
 
-       std     r14,STK_REG(r14)(r1)
-       std     r15,STK_REG(r15)(r1)
-       std     r16,STK_REG(r16)(r1)
+       std     r14,STK_REG(R14)(r1)
+       std     r15,STK_REG(R15)(r1)
+       std     r16,STK_REG(R16)(r1)
 
        li      r12,64
        li      r14,80
@@ -574,9 +574,9 @@ _GLOBAL(memcpy_power7)
        addi    r3,r3,128
        bdnz    8b
 
-       ld      r14,STK_REG(r14)(r1)
-       ld      r15,STK_REG(r15)(r1)
-       ld      r16,STK_REG(r16)(r1)
+       ld      r14,STK_REG(R14)(r1)
+       ld      r15,STK_REG(R15)(r1)
+       ld      r16,STK_REG(R16)(r1)
 
        /* Up to 127B to go */
        clrldi  r5,r5,(64-7)
index a242b5d..113dcb0 100644 (file)
@@ -64,9 +64,9 @@ _GLOBAL(__hash_page_4K)
        std     r0,16(r1)
        stdu    r1,-STACKFRAMESIZE(r1)
        /* Save all params that we need after a function call */
-       std     r6,STK_PARM(r6)(r1)
-       std     r8,STK_PARM(r8)(r1)
-       std     r9,STK_PARM(r9)(r1)
+       std     r6,STK_PARM(R6)(r1)
+       std     r8,STK_PARM(R8)(r1)
+       std     r9,STK_PARM(R9)(r1)
        
        /* Save non-volatile registers.
         * r31 will hold "old PTE"
@@ -75,11 +75,11 @@ _GLOBAL(__hash_page_4K)
         * r28 is a hash value
         * r27 is hashtab mask (maybe dynamic patched instead ?)
         */
-       std     r27,STK_REG(r27)(r1)
-       std     r28,STK_REG(r28)(r1)
-       std     r29,STK_REG(r29)(r1)
-       std     r30,STK_REG(r30)(r1)
-       std     r31,STK_REG(r31)(r1)
+       std     r27,STK_REG(R27)(r1)
+       std     r28,STK_REG(R28)(r1)
+       std     r29,STK_REG(R29)(r1)
+       std     r30,STK_REG(R30)(r1)
+       std     r31,STK_REG(R31)(r1)
        
        /* Step 1:
         *
@@ -162,7 +162,7 @@ END_FTR_SECTION(CPU_FTR_NOEXECUTE|CPU_FTR_COHERENT_ICACHE, CPU_FTR_NOEXECUTE)
        /* At this point, r3 contains new PP bits, save them in
         * place of "access" in the param area (sic)
         */
-       std     r3,STK_PARM(r4)(r1)
+       std     r3,STK_PARM(R4)(r1)
 
        /* Get htab_hash_mask */
        ld      r4,htab_hash_mask@got(2)
@@ -192,11 +192,11 @@ htab_insert_pte:
        rldicr  r3,r0,3,63-3            /* r3 = (hash & mask) << 3 */
 
        /* Call ppc_md.hpte_insert */
-       ld      r6,STK_PARM(r4)(r1)     /* Retrieve new pp bits */
+       ld      r6,STK_PARM(R4)(r1)     /* Retrieve new pp bits */
        mr      r4,r29                  /* Retrieve va */
        li      r7,0                    /* !bolted, !secondary */
        li      r8,MMU_PAGE_4K          /* page size */
-       ld      r9,STK_PARM(r9)(r1)     /* segment size */
+       ld      r9,STK_PARM(R9)(r1)     /* segment size */
 _GLOBAL(htab_call_hpte_insert1)
        bl      .                       /* Patched by htab_finish_init() */
        cmpdi   0,r3,0
@@ -215,11 +215,11 @@ _GLOBAL(htab_call_hpte_insert1)
        rldicr  r3,r0,3,63-3    /* r0 = (~hash & mask) << 3 */
        
        /* Call ppc_md.hpte_insert */
-       ld      r6,STK_PARM(r4)(r1)     /* Retrieve new pp bits */
+       ld      r6,STK_PARM(R4)(r1)     /* Retrieve new pp bits */
        mr      r4,r29                  /* Retrieve va */
        li      r7,HPTE_V_SECONDARY     /* !bolted, secondary */
        li      r8,MMU_PAGE_4K          /* page size */
-       ld      r9,STK_PARM(r9)(r1)     /* segment size */
+       ld      r9,STK_PARM(R9)(r1)     /* segment size */
 _GLOBAL(htab_call_hpte_insert2)
        bl      .                       /* Patched by htab_finish_init() */
        cmpdi   0,r3,0
@@ -255,15 +255,15 @@ htab_pte_insert_ok:
         * (maybe add eieio may be good still ?)
         */
 htab_write_out_pte:
-       ld      r6,STK_PARM(r6)(r1)
+       ld      r6,STK_PARM(R6)(r1)
        std     r30,0(r6)
        li      r3, 0
 htab_bail:
-       ld      r27,STK_REG(r27)(r1)
-       ld      r28,STK_REG(r28)(r1)
-       ld      r29,STK_REG(r29)(r1)
-       ld      r30,STK_REG(r30)(r1)
-       ld      r31,STK_REG(r31)(r1)
+       ld      r27,STK_REG(R27)(r1)
+       ld      r28,STK_REG(R28)(r1)
+       ld      r29,STK_REG(R29)(r1)
+       ld      r30,STK_REG(R30)(r1)
+       ld      r31,STK_REG(R31)(r1)
        addi    r1,r1,STACKFRAMESIZE
        ld      r0,16(r1)
        mtlr    r0
@@ -288,8 +288,8 @@ htab_modify_pte:
        /* Call ppc_md.hpte_updatepp */
        mr      r5,r29                  /* va */
        li      r6,MMU_PAGE_4K          /* page size */
-       ld      r7,STK_PARM(r9)(r1)     /* segment size */
-       ld      r8,STK_PARM(r8)(r1)     /* get "local" param */
+       ld      r7,STK_PARM(R9)(r1)     /* segment size */
+       ld      r8,STK_PARM(R8)(r1)     /* get "local" param */
 _GLOBAL(htab_call_hpte_updatepp)
        bl      .                       /* Patched by htab_finish_init() */
 
@@ -312,7 +312,7 @@ htab_wrong_access:
 
 htab_pte_insert_failure:
        /* Bail out restoring old PTE */
-       ld      r6,STK_PARM(r6)(r1)
+       ld      r6,STK_PARM(R6)(r1)
        std     r31,0(r6)
        li      r3,-1
        b       htab_bail
@@ -340,9 +340,9 @@ _GLOBAL(__hash_page_4K)
        std     r0,16(r1)
        stdu    r1,-STACKFRAMESIZE(r1)
        /* Save all params that we need after a function call */
-       std     r6,STK_PARM(r6)(r1)
-       std     r8,STK_PARM(r8)(r1)
-       std     r9,STK_PARM(r9)(r1)
+       std     r6,STK_PARM(R6)(r1)
+       std     r8,STK_PARM(R8)(r1)
+       std     r9,STK_PARM(R9)(r1)
 
        /* Save non-volatile registers.
         * r31 will hold "old PTE"
@@ -353,13 +353,13 @@ _GLOBAL(__hash_page_4K)
         * r26 is the hidx mask
         * r25 is the index in combo page
         */
-       std     r25,STK_REG(r25)(r1)
-       std     r26,STK_REG(r26)(r1)
-       std     r27,STK_REG(r27)(r1)
-       std     r28,STK_REG(r28)(r1)
-       std     r29,STK_REG(r29)(r1)
-       std     r30,STK_REG(r30)(r1)
-       std     r31,STK_REG(r31)(r1)
+       std     r25,STK_REG(R25)(r1)
+       std     r26,STK_REG(R26)(r1)
+       std     r27,STK_REG(R27)(r1)
+       std     r28,STK_REG(R28)(r1)
+       std     r29,STK_REG(R29)(r1)
+       std     r30,STK_REG(R30)(r1)
+       std     r31,STK_REG(R31)(r1)
 
        /* Step 1:
         *
@@ -452,7 +452,7 @@ END_FTR_SECTION(CPU_FTR_NOEXECUTE|CPU_FTR_COHERENT_ICACHE, CPU_FTR_NOEXECUTE)
        /* At this point, r3 contains new PP bits, save them in
         * place of "access" in the param area (sic)
         */
-       std     r3,STK_PARM(r4)(r1)
+       std     r3,STK_PARM(R4)(r1)
 
        /* Get htab_hash_mask */
        ld      r4,htab_hash_mask@got(2)
@@ -473,7 +473,7 @@ END_FTR_SECTION(CPU_FTR_NOEXECUTE|CPU_FTR_COHERENT_ICACHE, CPU_FTR_NOEXECUTE)
        andis.  r0,r31,_PAGE_COMBO@h
        beq     htab_inval_old_hpte
 
-       ld      r6,STK_PARM(r6)(r1)
+       ld      r6,STK_PARM(R6)(r1)
        ori     r26,r6,0x8000           /* Load the hidx mask */
        ld      r26,0(r26)
        addi    r5,r25,36               /* Check actual HPTE_SUB bit, this */
@@ -495,11 +495,11 @@ htab_special_pfn:
        rldicr  r3,r0,3,63-3            /* r0 = (hash & mask) << 3 */
 
        /* Call ppc_md.hpte_insert */
-       ld      r6,STK_PARM(r4)(r1)     /* Retrieve new pp bits */
+       ld      r6,STK_PARM(R4)(r1)     /* Retrieve new pp bits */
        mr      r4,r29                  /* Retrieve va */
        li      r7,0                    /* !bolted, !secondary */
        li      r8,MMU_PAGE_4K          /* page size */
-       ld      r9,STK_PARM(r9)(r1)     /* segment size */
+       ld      r9,STK_PARM(R9)(r1)     /* segment size */
 _GLOBAL(htab_call_hpte_insert1)
        bl      .                       /* patched by htab_finish_init() */
        cmpdi   0,r3,0
@@ -522,11 +522,11 @@ _GLOBAL(htab_call_hpte_insert1)
        rldicr  r3,r0,3,63-3            /* r0 = (~hash & mask) << 3 */
 
        /* Call ppc_md.hpte_insert */
-       ld      r6,STK_PARM(r4)(r1)     /* Retrieve new pp bits */
+       ld      r6,STK_PARM(R4)(r1)     /* Retrieve new pp bits */
        mr      r4,r29                  /* Retrieve va */
        li      r7,HPTE_V_SECONDARY     /* !bolted, secondary */
        li      r8,MMU_PAGE_4K          /* page size */
-       ld      r9,STK_PARM(r9)(r1)     /* segment size */
+       ld      r9,STK_PARM(R9)(r1)     /* segment size */
 _GLOBAL(htab_call_hpte_insert2)
        bl      .                       /* patched by htab_finish_init() */
        cmpdi   0,r3,0
@@ -559,8 +559,8 @@ htab_inval_old_hpte:
        mr      r4,r31                  /* PTE.pte */
        li      r5,0                    /* PTE.hidx */
        li      r6,MMU_PAGE_64K         /* psize */
-       ld      r7,STK_PARM(r9)(r1)     /* ssize */
-       ld      r8,STK_PARM(r8)(r1)     /* local */
+       ld      r7,STK_PARM(R9)(r1)     /* ssize */
+       ld      r8,STK_PARM(R8)(r1)     /* local */
        bl      .flush_hash_page
        /* Clear out _PAGE_HPTE_SUB bits in the new linux PTE */
        lis     r0,_PAGE_HPTE_SUB@h
@@ -576,7 +576,7 @@ htab_pte_insert_ok:
        /* Insert slot number & secondary bit in PTE second half,
         * clear _PAGE_BUSY and set approriate HPTE slot bit
         */
-       ld      r6,STK_PARM(r6)(r1)
+       ld      r6,STK_PARM(R6)(r1)
        li      r0,_PAGE_BUSY
        andc    r30,r30,r0
        /* HPTE SUB bit */
@@ -597,13 +597,13 @@ htab_pte_insert_ok:
        std     r30,0(r6)
        li      r3, 0
 htab_bail:
-       ld      r25,STK_REG(r25)(r1)
-       ld      r26,STK_REG(r26)(r1)
-       ld      r27,STK_REG(r27)(r1)
-       ld      r28,STK_REG(r28)(r1)
-       ld      r29,STK_REG(r29)(r1)
-       ld      r30,STK_REG(r30)(r1)
-       ld      r31,STK_REG(r31)(r1)
+       ld      r25,STK_REG(R25)(r1)
+       ld      r26,STK_REG(R26)(r1)
+       ld      r27,STK_REG(R27)(r1)
+       ld      r28,STK_REG(R28)(r1)
+       ld      r29,STK_REG(R29)(r1)
+       ld      r30,STK_REG(R30)(r1)
+       ld      r31,STK_REG(R31)(r1)
        addi    r1,r1,STACKFRAMESIZE
        ld      r0,16(r1)
        mtlr    r0
@@ -630,8 +630,8 @@ htab_modify_pte:
        /* Call ppc_md.hpte_updatepp */
        mr      r5,r29                  /* va */
        li      r6,MMU_PAGE_4K          /* page size */
-       ld      r7,STK_PARM(r9)(r1)     /* segment size */
-       ld      r8,STK_PARM(r8)(r1)     /* get "local" param */
+       ld      r7,STK_PARM(R9)(r1)     /* segment size */
+       ld      r8,STK_PARM(R8)(r1)     /* get "local" param */
 _GLOBAL(htab_call_hpte_updatepp)
        bl      .                       /* patched by htab_finish_init() */
 
@@ -644,7 +644,7 @@ _GLOBAL(htab_call_hpte_updatepp)
        /* Clear the BUSY bit and Write out the PTE */
        li      r0,_PAGE_BUSY
        andc    r30,r30,r0
-       ld      r6,STK_PARM(r6)(r1)
+       ld      r6,STK_PARM(R6)(r1)
        std     r30,0(r6)
        li      r3,0
        b       htab_bail
@@ -657,7 +657,7 @@ htab_wrong_access:
 
 htab_pte_insert_failure:
        /* Bail out restoring old PTE */
-       ld      r6,STK_PARM(r6)(r1)
+       ld      r6,STK_PARM(R6)(r1)
        std     r31,0(r6)
        li      r3,-1
        b       htab_bail
@@ -677,9 +677,9 @@ _GLOBAL(__hash_page_64K)
        std     r0,16(r1)
        stdu    r1,-STACKFRAMESIZE(r1)
        /* Save all params that we need after a function call */
-       std     r6,STK_PARM(r6)(r1)
-       std     r8,STK_PARM(r8)(r1)
-       std     r9,STK_PARM(r9)(r1)
+       std     r6,STK_PARM(R6)(r1)
+       std     r8,STK_PARM(R8)(r1)
+       std     r9,STK_PARM(R9)(r1)
 
        /* Save non-volatile registers.
         * r31 will hold "old PTE"
@@ -688,11 +688,11 @@ _GLOBAL(__hash_page_64K)
         * r28 is a hash value
         * r27 is hashtab mask (maybe dynamic patched instead ?)
         */
-       std     r27,STK_REG(r27)(r1)
-       std     r28,STK_REG(r28)(r1)
-       std     r29,STK_REG(r29)(r1)
-       std     r30,STK_REG(r30)(r1)
-       std     r31,STK_REG(r31)(r1)
+       std     r27,STK_REG(R27)(r1)
+       std     r28,STK_REG(R28)(r1)
+       std     r29,STK_REG(R29)(r1)
+       std     r30,STK_REG(R30)(r1)
+       std     r31,STK_REG(R31)(r1)
 
        /* Step 1:
         *
@@ -780,7 +780,7 @@ END_FTR_SECTION(CPU_FTR_NOEXECUTE|CPU_FTR_COHERENT_ICACHE, CPU_FTR_NOEXECUTE)
        /* At this point, r3 contains new PP bits, save them in
         * place of "access" in the param area (sic)
         */
-       std     r3,STK_PARM(r4)(r1)
+       std     r3,STK_PARM(R4)(r1)
 
        /* Get htab_hash_mask */
        ld      r4,htab_hash_mask@got(2)
@@ -813,11 +813,11 @@ ht64_insert_pte:
        rldicr  r3,r0,3,63-3    /* r0 = (hash & mask) << 3 */
 
        /* Call ppc_md.hpte_insert */
-       ld      r6,STK_PARM(r4)(r1)     /* Retrieve new pp bits */
+       ld      r6,STK_PARM(R4)(r1)     /* Retrieve new pp bits */
        mr      r4,r29                  /* Retrieve va */
        li      r7,0                    /* !bolted, !secondary */
        li      r8,MMU_PAGE_64K
-       ld      r9,STK_PARM(r9)(r1)     /* segment size */
+       ld      r9,STK_PARM(R9)(r1)     /* segment size */
 _GLOBAL(ht64_call_hpte_insert1)
        bl      .                       /* patched by htab_finish_init() */
        cmpdi   0,r3,0
@@ -836,11 +836,11 @@ _GLOBAL(ht64_call_hpte_insert1)
        rldicr  r3,r0,3,63-3    /* r0 = (~hash & mask) << 3 */
 
        /* Call ppc_md.hpte_insert */
-       ld      r6,STK_PARM(r4)(r1)     /* Retrieve new pp bits */
+       ld      r6,STK_PARM(R4)(r1)     /* Retrieve new pp bits */
        mr      r4,r29                  /* Retrieve va */
        li      r7,HPTE_V_SECONDARY     /* !bolted, secondary */
        li      r8,MMU_PAGE_64K
-       ld      r9,STK_PARM(r9)(r1)     /* segment size */
+       ld      r9,STK_PARM(R9)(r1)     /* segment size */
 _GLOBAL(ht64_call_hpte_insert2)
        bl      .                       /* patched by htab_finish_init() */
        cmpdi   0,r3,0
@@ -876,15 +876,15 @@ ht64_pte_insert_ok:
         * (maybe add eieio may be good still ?)
         */
 ht64_write_out_pte:
-       ld      r6,STK_PARM(r6)(r1)
+       ld      r6,STK_PARM(R6)(r1)
        std     r30,0(r6)
        li      r3, 0
 ht64_bail:
-       ld      r27,STK_REG(r27)(r1)
-       ld      r28,STK_REG(r28)(r1)
-       ld      r29,STK_REG(r29)(r1)
-       ld      r30,STK_REG(r30)(r1)
-       ld      r31,STK_REG(r31)(r1)
+       ld      r27,STK_REG(R27)(r1)
+       ld      r28,STK_REG(R28)(r1)
+       ld      r29,STK_REG(R29)(r1)
+       ld      r30,STK_REG(R30)(r1)
+       ld      r31,STK_REG(R31)(r1)
        addi    r1,r1,STACKFRAMESIZE
        ld      r0,16(r1)
        mtlr    r0
@@ -909,8 +909,8 @@ ht64_modify_pte:
        /* Call ppc_md.hpte_updatepp */
        mr      r5,r29                  /* va */
        li      r6,MMU_PAGE_64K
-       ld      r7,STK_PARM(r9)(r1)     /* segment size */
-       ld      r8,STK_PARM(r8)(r1)     /* get "local" param */
+       ld      r7,STK_PARM(R9)(r1)     /* segment size */
+       ld      r8,STK_PARM(R8)(r1)     /* get "local" param */
 _GLOBAL(ht64_call_hpte_updatepp)
        bl      .                       /* patched by htab_finish_init() */
 
@@ -933,7 +933,7 @@ ht64_wrong_access:
 
 ht64_pte_insert_failure:
        /* Bail out restoring old PTE */
-       ld      r6,STK_PARM(r6)(r1)
+       ld      r6,STK_PARM(R6)(r1)
        std     r31,0(r6)
        li      r3,-1
        b       ht64_bail
index ff672bd..4b9e264 100644 (file)
@@ -126,7 +126,7 @@ BEGIN_MMU_FTR_SECTION
        /* Set the TLB reservation and search for existing entry. Then load
         * the entry.
         */
-       PPC_TLBSRX_DOT(0,r16)
+       PPC_TLBSRX_DOT(R0,R16)
        ldx     r14,r14,r15             /* grab pgd entry */
        beq     normal_tlb_miss_done    /* tlb exists already, bail */
 MMU_FTR_SECTION_ELSE
@@ -395,7 +395,7 @@ BEGIN_MMU_FTR_SECTION
        /* Set the TLB reservation and search for existing entry. Then load
         * the entry.
         */
-       PPC_TLBSRX_DOT(0,r16)
+       PPC_TLBSRX_DOT(R0,R16)
        ld      r14,0(r10)
        beq     normal_tlb_miss_done
 MMU_FTR_SECTION_ELSE
@@ -528,7 +528,7 @@ BEGIN_MMU_FTR_SECTION
        /* Search if we already have a TLB entry for that virtual address, and
         * if we do, bail out.
         */
-       PPC_TLBSRX_DOT(0,r16)
+       PPC_TLBSRX_DOT(R0,R16)
        beq     virt_page_table_tlb_miss_done
 END_MMU_FTR_SECTION_IFSET(MMU_FTR_USE_TLBRSRV)
 
@@ -779,7 +779,7 @@ htw_tlb_miss:
         *
         * MAS1:IND should be already set based on MAS4
         */
-       PPC_TLBSRX_DOT(0,r16)
+       PPC_TLBSRX_DOT(R0,R16)
        beq     htw_tlb_miss_done
 
        /* Now, we need to walk the page tables. First check if we are in
@@ -919,7 +919,7 @@ tlb_load_linear:
        mtspr   SPRN_MAS1,r15
 
        /* Already somebody there ? */
-       PPC_TLBSRX_DOT(0,r16)
+       PPC_TLBSRX_DOT(R0,R16)
        beq     tlb_load_linear_done
 
        /* Now we build the remaining MAS. MAS0 and 2 should be fine
index 7c63c0e..5a1285a 100644 (file)
@@ -313,7 +313,7 @@ BEGIN_MMU_FTR_SECTION
        mtspr   SPRN_MAS1,r4
        tlbwe
 MMU_FTR_SECTION_ELSE
-       PPC_TLBILX_VA(0,r3)
+       PPC_TLBILX_VA(R0,R3)
 ALT_MMU_FTR_SECTION_END_IFCLR(MMU_FTR_USE_TLBILX)
        msync
        isync
@@ -364,7 +364,7 @@ _GLOBAL(_tlbil_va)
        beq     1f
        rlwimi  r4,r6,MAS6_SIND_SHIFT,MAS6_SIND
 1:     mtspr   SPRN_MAS6,r4            /* assume AS=0 for now */
-       PPC_TLBILX_VA(0,r3)
+       PPC_TLBILX_VA(R0,R3)
        msync
        isync
        wrtee   r10
@@ -379,7 +379,7 @@ _GLOBAL(_tlbivax_bcast)
        beq     1f
        rlwimi  r4,r6,MAS6_SIND_SHIFT,MAS6_SIND
 1:     mtspr   SPRN_MAS6,r4            /* assume AS=0 for now */
-       PPC_TLBIVAX(0,r3)
+       PPC_TLBIVAX(R0,R3)
        eieio
        tlbsync
        sync
index 2dc8b14..dd11306 100644 (file)
@@ -39,7 +39,7 @@ static void bpf_jit_build_prologue(struct sk_filter *fp, u32 *image,
                /* Make stackframe */
                if (ctx->seen & SEEN_DATAREF) {
                        /* If we call any helpers (for loads), save LR */
-                       EMIT(PPC_INST_MFLR | __PPC_RT(0));
+                       EMIT(PPC_INST_MFLR | __PPC_RT(R0));
                        PPC_STD(0, 1, 16);
 
                        /* Back up non-volatile regs. */
@@ -56,7 +56,7 @@ static void bpf_jit_build_prologue(struct sk_filter *fp, u32 *image,
                                        PPC_STD(i, 1, -(8*(32-i)));
                        }
                }
-               EMIT(PPC_INST_STDU | __PPC_RS(1) | __PPC_RA(1) |
+               EMIT(PPC_INST_STDU | __PPC_RS(R1) | __PPC_RA(R1) |
                     (-BPF_PPC_STACKFRAME & 0xfffc));
        }
 
index 74c8174..079165e 100644 (file)
@@ -74,7 +74,7 @@ _GLOBAL(beat_hcall_norets8)
        mr      r6,r7
        mr      r7,r8
        mr      r8,r9
-       ld      r10,STK_PARM(r10)(r1)
+       ld      r10,STK_PARM(R10)(r1)
 
        HVSC                            /* invoke the hypervisor */
 
@@ -94,7 +94,7 @@ _GLOBAL(beat_hcall1)
 
        HCALL_INST_PRECALL
 
-       std     r4,STK_PARM(r4)(r1)     /* save ret buffer */
+       std     r4,STK_PARM(R4)(r1)     /* save ret buffer */
 
        mr      r11,r3
        mr      r3,r5
@@ -108,7 +108,7 @@ _GLOBAL(beat_hcall1)
 
        HCALL_INST_POSTCALL
 
-       ld      r12,STK_PARM(r4)(r1)
+       ld      r12,STK_PARM(R4)(r1)
        std     r4,  0(r12)
 
        lwz     r0,8(r1)
@@ -125,7 +125,7 @@ _GLOBAL(beat_hcall2)
 
        HCALL_INST_PRECALL
 
-       std     r4,STK_PARM(r4)(r1)     /* save ret buffer */
+       std     r4,STK_PARM(R4)(r1)     /* save ret buffer */
 
        mr      r11,r3
        mr      r3,r5
@@ -139,7 +139,7 @@ _GLOBAL(beat_hcall2)
 
        HCALL_INST_POSTCALL
 
-       ld      r12,STK_PARM(r4)(r1)
+       ld      r12,STK_PARM(R4)(r1)
        std     r4,  0(r12)
        std     r5,  8(r12)
 
@@ -157,7 +157,7 @@ _GLOBAL(beat_hcall3)
 
        HCALL_INST_PRECALL
 
-       std     r4,STK_PARM(r4)(r1)     /* save ret buffer */
+       std     r4,STK_PARM(R4)(r1)     /* save ret buffer */
 
        mr      r11,r3
        mr      r3,r5
@@ -171,7 +171,7 @@ _GLOBAL(beat_hcall3)
 
        HCALL_INST_POSTCALL
 
-       ld      r12,STK_PARM(r4)(r1)
+       ld      r12,STK_PARM(R4)(r1)
        std     r4,  0(r12)
        std     r5,  8(r12)
        std     r6, 16(r12)
@@ -190,7 +190,7 @@ _GLOBAL(beat_hcall4)
 
        HCALL_INST_PRECALL
 
-       std     r4,STK_PARM(r4)(r1)     /* save ret buffer */
+       std     r4,STK_PARM(R4)(r1)     /* save ret buffer */
 
        mr      r11,r3
        mr      r3,r5
@@ -204,7 +204,7 @@ _GLOBAL(beat_hcall4)
 
        HCALL_INST_POSTCALL
 
-       ld      r12,STK_PARM(r4)(r1)
+       ld      r12,STK_PARM(R4)(r1)
        std     r4,  0(r12)
        std     r5,  8(r12)
        std     r6, 16(r12)
@@ -224,7 +224,7 @@ _GLOBAL(beat_hcall5)
 
        HCALL_INST_PRECALL
 
-       std     r4,STK_PARM(r4)(r1)     /* save ret buffer */
+       std     r4,STK_PARM(R4)(r1)     /* save ret buffer */
 
        mr      r11,r3
        mr      r3,r5
@@ -238,7 +238,7 @@ _GLOBAL(beat_hcall5)
 
        HCALL_INST_POSTCALL
 
-       ld      r12,STK_PARM(r4)(r1)
+       ld      r12,STK_PARM(R4)(r1)
        std     r4,  0(r12)
        std     r5,  8(r12)
        std     r6, 16(r12)
@@ -259,7 +259,7 @@ _GLOBAL(beat_hcall6)
 
        HCALL_INST_PRECALL
 
-       std     r4,STK_PARM(r4)(r1)     /* save ret buffer */
+       std     r4,STK_PARM(R4)(r1)     /* save ret buffer */
 
        mr      r11,r3
        mr      r3,r5
@@ -273,7 +273,7 @@ _GLOBAL(beat_hcall6)
 
        HCALL_INST_POSTCALL
 
-       ld      r12,STK_PARM(r4)(r1)
+       ld      r12,STK_PARM(R4)(r1)
        std     r4,  0(r12)
        std     r5,  8(r12)
        std     r6, 16(r12)
index 77b48b2..1bb7768 100644 (file)
 _GLOBAL(opal_query_takeover)
        mfcr    r0
        stw     r0,8(r1)
-       std     r3,STK_PARAM(r3)(r1)
-       std     r4,STK_PARAM(r4)(r1)
+       std     r3,STK_PARAM(R3)(r1)
+       std     r4,STK_PARAM(R4)(r1)
        li      r3,H_HAL_TAKEOVER
        li      r4,H_HAL_TAKEOVER_QUERY_MAGIC
        HVSC
-       ld      r10,STK_PARAM(r3)(r1)
+       ld      r10,STK_PARAM(R3)(r1)
        std     r4,0(r10)
-       ld      r10,STK_PARAM(r4)(r1)
+       ld      r10,STK_PARAM(R4)(r1)
        std     r5,0(r10)
        lwz     r0,8(r1)
        mtcrf   0xff,r0
index 3bb07e5..afcddec 100644 (file)
@@ -32,7 +32,7 @@
        std     r12,PACASAVEDMSR(r13);  \
        andc    r12,r12,r0;             \
        mtmsrd  r12,1;                  \
-       LOAD_REG_ADDR(r0,.opal_return); \
+       LOAD_REG_ADDR(R0,.opal_return); \
        mtlr    r0;                     \
        li      r0,MSR_DR|MSR_IR;       \
        andc    r12,r12,r0;             \
index 3ce73d0..04b8efd 100644 (file)
@@ -40,28 +40,28 @@ END_FTR_SECTION(0, 1);                                              \
        cmpdi   r12,0;                                          \
        beq+    1f;                                             \
        mflr    r0;                                             \
-       std     r3,STK_PARM(r3)(r1);                            \
-       std     r4,STK_PARM(r4)(r1);                            \
-       std     r5,STK_PARM(r5)(r1);                            \
-       std     r6,STK_PARM(r6)(r1);                            \
-       std     r7,STK_PARM(r7)(r1);                            \
-       std     r8,STK_PARM(r8)(r1);                            \
-       std     r9,STK_PARM(r9)(r1);                            \
-       std     r10,STK_PARM(r10)(r1);                          \
+       std     r3,STK_PARM(R3)(r1);                            \
+       std     r4,STK_PARM(R4)(r1);                            \
+       std     r5,STK_PARM(R5)(r1);                            \
+       std     r6,STK_PARM(R6)(r1);                            \
+       std     r7,STK_PARM(R7)(r1);                            \
+       std     r8,STK_PARM(R8)(r1);                            \
+       std     r9,STK_PARM(R9)(r1);                            \
+       std     r10,STK_PARM(R10)(r1);                          \
        std     r0,16(r1);                                      \
        addi    r4,r1,STK_PARM(FIRST_REG);                      \
        stdu    r1,-STACK_FRAME_OVERHEAD(r1);                   \
        bl      .__trace_hcall_entry;                           \
        addi    r1,r1,STACK_FRAME_OVERHEAD;                     \
        ld      r0,16(r1);                                      \
-       ld      r3,STK_PARM(r3)(r1);                            \
-       ld      r4,STK_PARM(r4)(r1);                            \
-       ld      r5,STK_PARM(r5)(r1);                            \
-       ld      r6,STK_PARM(r6)(r1);                            \
-       ld      r7,STK_PARM(r7)(r1);                            \
-       ld      r8,STK_PARM(r8)(r1);                            \
-       ld      r9,STK_PARM(r9)(r1);                            \
-       ld      r10,STK_PARM(r10)(r1);                          \
+       ld      r3,STK_PARM(R3)(r1);                            \
+       ld      r4,STK_PARM(R4)(r1);                            \
+       ld      r5,STK_PARM(R5)(r1);                            \
+       ld      r6,STK_PARM(R6)(r1);                            \
+       ld      r7,STK_PARM(R7)(r1);                            \
+       ld      r8,STK_PARM(R8)(r1);                            \
+       ld      r9,STK_PARM(R9)(r1);                            \
+       ld      r10,STK_PARM(R10)(r1);                          \
        mtlr    r0;                                             \
 1:
 
@@ -79,8 +79,8 @@ END_FTR_SECTION(0, 1);                                                \
        cmpdi   r12,0;                                          \
        beq+    1f;                                             \
        mflr    r0;                                             \
-       ld      r6,STK_PARM(r3)(r1);                            \
-       std     r3,STK_PARM(r3)(r1);                            \
+       ld      r6,STK_PARM(R3)(r1);                            \
+       std     r3,STK_PARM(R3)(r1);                            \
        mr      r4,r3;                                          \
        mr      r3,r6;                                          \
        std     r0,16(r1);                                      \
@@ -88,7 +88,7 @@ END_FTR_SECTION(0, 1);                                                \
        bl      .__trace_hcall_exit;                            \
        addi    r1,r1,STACK_FRAME_OVERHEAD;                     \
        ld      r0,16(r1);                                      \
-       ld      r3,STK_PARM(r3)(r1);                            \
+       ld      r3,STK_PARM(R3)(r1);                            \
        mtlr    r0;                                             \
 1:
 
@@ -114,7 +114,7 @@ _GLOBAL(plpar_hcall_norets)
        mfcr    r0
        stw     r0,8(r1)
 
-       HCALL_INST_PRECALL(r4)
+       HCALL_INST_PRECALL(R4)
 
        HVSC                            /* invoke the hypervisor */
 
@@ -130,9 +130,9 @@ _GLOBAL(plpar_hcall)
        mfcr    r0
        stw     r0,8(r1)
 
-       HCALL_INST_PRECALL(r5)
+       HCALL_INST_PRECALL(R5)
 
-       std     r4,STK_PARM(r4)(r1)     /* Save ret buffer */
+       std     r4,STK_PARM(R4)(r1)     /* Save ret buffer */
 
        mr      r4,r5
        mr      r5,r6
@@ -143,7 +143,7 @@ _GLOBAL(plpar_hcall)
 
        HVSC                            /* invoke the hypervisor */
 
-       ld      r12,STK_PARM(r4)(r1)
+       ld      r12,STK_PARM(R4)(r1)
        std     r4,  0(r12)
        std     r5,  8(r12)
        std     r6, 16(r12)
@@ -168,7 +168,7 @@ _GLOBAL(plpar_hcall_raw)
        mfcr    r0
        stw     r0,8(r1)
 
-       std     r4,STK_PARM(r4)(r1)     /* Save ret buffer */
+       std     r4,STK_PARM(R4)(r1)     /* Save ret buffer */
 
        mr      r4,r5
        mr      r5,r6
@@ -179,7 +179,7 @@ _GLOBAL(plpar_hcall_raw)
 
        HVSC                            /* invoke the hypervisor */
 
-       ld      r12,STK_PARM(r4)(r1)
+       ld      r12,STK_PARM(R4)(r1)
        std     r4,  0(r12)
        std     r5,  8(r12)
        std     r6, 16(r12)
@@ -196,9 +196,9 @@ _GLOBAL(plpar_hcall9)
        mfcr    r0
        stw     r0,8(r1)
 
-       HCALL_INST_PRECALL(r5)
+       HCALL_INST_PRECALL(R5)
 
-       std     r4,STK_PARM(r4)(r1)     /* Save ret buffer */
+       std     r4,STK_PARM(R4)(r1)     /* Save ret buffer */
 
        mr      r4,r5
        mr      r5,r6
@@ -206,14 +206,14 @@ _GLOBAL(plpar_hcall9)
        mr      r7,r8
        mr      r8,r9
        mr      r9,r10
-       ld      r10,STK_PARM(r11)(r1)    /* put arg7 in R10 */
-       ld      r11,STK_PARM(r12)(r1)    /* put arg8 in R11 */
-       ld      r12,STK_PARM(r13)(r1)    /* put arg9 in R12 */
+       ld      r10,STK_PARM(R11)(r1)    /* put arg7 in R10 */
+       ld      r11,STK_PARM(R12)(r1)    /* put arg8 in R11 */
+       ld      r12,STK_PARM(R13)(r1)    /* put arg9 in R12 */
 
        HVSC                            /* invoke the hypervisor */
 
        mr      r0,r12
-       ld      r12,STK_PARM(r4)(r1)
+       ld      r12,STK_PARM(R4)(r1)
        std     r4,  0(r12)
        std     r5,  8(r12)
        std     r6, 16(r12)
@@ -238,7 +238,7 @@ _GLOBAL(plpar_hcall9_raw)
        mfcr    r0
        stw     r0,8(r1)
 
-       std     r4,STK_PARM(r4)(r1)     /* Save ret buffer */
+       std     r4,STK_PARM(R4)(r1)     /* Save ret buffer */
 
        mr      r4,r5
        mr      r5,r6
@@ -246,14 +246,14 @@ _GLOBAL(plpar_hcall9_raw)
        mr      r7,r8
        mr      r8,r9
        mr      r9,r10
-       ld      r10,STK_PARM(r11)(r1)    /* put arg7 in R10 */
-       ld      r11,STK_PARM(r12)(r1)    /* put arg8 in R11 */
-       ld      r12,STK_PARM(r13)(r1)    /* put arg9 in R12 */
+       ld      r10,STK_PARM(R11)(r1)    /* put arg7 in R10 */
+       ld      r11,STK_PARM(R12)(r1)    /* put arg8 in R11 */
+       ld      r12,STK_PARM(R13)(r1)    /* put arg9 in R12 */
 
        HVSC                            /* invoke the hypervisor */
 
        mr      r0,r12
-       ld      r12,STK_PARM(r4)(r1)
+       ld      r12,STK_PARM(R4)(r1)
        std     r4,  0(r12)
        std     r5,  8(r12)
        std     r6, 16(r12)