Donate to e Foundation | Murena handsets with /e/OS | Own a part of Murena! Learn more

Commit ffb03559 authored by Linus Torvalds's avatar Linus Torvalds
Browse files
* 'upstream' of git://ftp.linux-mips.org/pub/scm/upstream-linus:
  [MIPS] SMP: Fix use of cpumasks.
  [MIPS] Revert "[MIPS] tlbex.c: Cleanup __init usage."
  [MIPS] CFE: Add missing parenthesis.
parents bcd11eaa ece8a9e4
Loading
Loading
Loading
Loading
+3 −3
Original line number Original line Diff line number Diff line
@@ -379,7 +379,7 @@ void flush_tlb_mm(struct mm_struct *mm)
		unsigned int cpu;
		unsigned int cpu;


		cpu_clear(smp_processor_id(), mask);
		cpu_clear(smp_processor_id(), mask);
		for_each_online_cpu(cpu)
		for_each_cpu_mask(cpu, mask)
			if (cpu_context(cpu, mm))
			if (cpu_context(cpu, mm))
				cpu_context(cpu, mm) = 0;
				cpu_context(cpu, mm) = 0;
	}
	}
@@ -419,7 +419,7 @@ void flush_tlb_range(struct vm_area_struct *vma, unsigned long start, unsigned l
		unsigned int cpu;
		unsigned int cpu;


		cpu_clear(smp_processor_id(), mask);
		cpu_clear(smp_processor_id(), mask);
		for_each_online_cpu(cpu)
		for_each_cpu_mask(cpu, mask)
			if (cpu_context(cpu, mm))
			if (cpu_context(cpu, mm))
				cpu_context(cpu, mm) = 0;
				cpu_context(cpu, mm) = 0;
	}
	}
@@ -466,7 +466,7 @@ void flush_tlb_page(struct vm_area_struct *vma, unsigned long page)
		unsigned int cpu;
		unsigned int cpu;


		cpu_clear(smp_processor_id(), mask);
		cpu_clear(smp_processor_id(), mask);
		for_each_online_cpu(cpu)
		for_each_cpu_mask(cpu, mask)
			if (cpu_context(cpu, vma->vm_mm))
			if (cpu_context(cpu, vma->vm_mm))
				cpu_context(cpu, vma->vm_mm) = 0;
				cpu_context(cpu, vma->vm_mm) = 0;
	}
	}
+48 −48
Original line number Original line Diff line number Diff line
@@ -66,7 +66,7 @@ static inline int __maybe_unused r10000_llsc_war(void)
 * why; it's not an issue caused by the core RTL.
 * why; it's not an issue caused by the core RTL.
 *
 *
 */
 */
static int __init m4kc_tlbp_war(void)
static __init int __attribute__((unused)) m4kc_tlbp_war(void)
{
{
	return (current_cpu_data.processor_id & 0xffff00) ==
	return (current_cpu_data.processor_id & 0xffff00) ==
	       (PRID_COMP_MIPS | PRID_IMP_4KC);
	       (PRID_COMP_MIPS | PRID_IMP_4KC);
@@ -140,7 +140,7 @@ struct insn {
	 | (e) << RE_SH						\
	 | (e) << RE_SH						\
	 | (f) << FUNC_SH)
	 | (f) << FUNC_SH)


static struct insn insn_table[] __initdata = {
static __initdata struct insn insn_table[] = {
	{ insn_addiu, M(addiu_op, 0, 0, 0, 0, 0), RS | RT | SIMM },
	{ insn_addiu, M(addiu_op, 0, 0, 0, 0, 0), RS | RT | SIMM },
	{ insn_addu, M(spec_op, 0, 0, 0, 0, addu_op), RS | RT | RD },
	{ insn_addu, M(spec_op, 0, 0, 0, 0, addu_op), RS | RT | RD },
	{ insn_and, M(spec_op, 0, 0, 0, 0, and_op), RS | RT | RD },
	{ insn_and, M(spec_op, 0, 0, 0, 0, and_op), RS | RT | RD },
@@ -193,7 +193,7 @@ static struct insn insn_table[] __initdata = {


#undef M
#undef M


static u32 __init build_rs(u32 arg)
static __init u32 build_rs(u32 arg)
{
{
	if (arg & ~RS_MASK)
	if (arg & ~RS_MASK)
		printk(KERN_WARNING "TLB synthesizer field overflow\n");
		printk(KERN_WARNING "TLB synthesizer field overflow\n");
@@ -201,7 +201,7 @@ static u32 __init build_rs(u32 arg)
	return (arg & RS_MASK) << RS_SH;
	return (arg & RS_MASK) << RS_SH;
}
}


static u32 __init build_rt(u32 arg)
static __init u32 build_rt(u32 arg)
{
{
	if (arg & ~RT_MASK)
	if (arg & ~RT_MASK)
		printk(KERN_WARNING "TLB synthesizer field overflow\n");
		printk(KERN_WARNING "TLB synthesizer field overflow\n");
@@ -209,7 +209,7 @@ static u32 __init build_rt(u32 arg)
	return (arg & RT_MASK) << RT_SH;
	return (arg & RT_MASK) << RT_SH;
}
}


static u32 __init build_rd(u32 arg)
static __init u32 build_rd(u32 arg)
{
{
	if (arg & ~RD_MASK)
	if (arg & ~RD_MASK)
		printk(KERN_WARNING "TLB synthesizer field overflow\n");
		printk(KERN_WARNING "TLB synthesizer field overflow\n");
@@ -217,7 +217,7 @@ static u32 __init build_rd(u32 arg)
	return (arg & RD_MASK) << RD_SH;
	return (arg & RD_MASK) << RD_SH;
}
}


static u32 __init build_re(u32 arg)
static __init u32 build_re(u32 arg)
{
{
	if (arg & ~RE_MASK)
	if (arg & ~RE_MASK)
		printk(KERN_WARNING "TLB synthesizer field overflow\n");
		printk(KERN_WARNING "TLB synthesizer field overflow\n");
@@ -225,7 +225,7 @@ static u32 __init build_re(u32 arg)
	return (arg & RE_MASK) << RE_SH;
	return (arg & RE_MASK) << RE_SH;
}
}


static u32 __init build_simm(s32 arg)
static __init u32 build_simm(s32 arg)
{
{
	if (arg > 0x7fff || arg < -0x8000)
	if (arg > 0x7fff || arg < -0x8000)
		printk(KERN_WARNING "TLB synthesizer field overflow\n");
		printk(KERN_WARNING "TLB synthesizer field overflow\n");
@@ -233,7 +233,7 @@ static u32 __init build_simm(s32 arg)
	return arg & 0xffff;
	return arg & 0xffff;
}
}


static u32 __init build_uimm(u32 arg)
static __init u32 build_uimm(u32 arg)
{
{
	if (arg & ~IMM_MASK)
	if (arg & ~IMM_MASK)
		printk(KERN_WARNING "TLB synthesizer field overflow\n");
		printk(KERN_WARNING "TLB synthesizer field overflow\n");
@@ -241,7 +241,7 @@ static u32 __init build_uimm(u32 arg)
	return arg & IMM_MASK;
	return arg & IMM_MASK;
}
}


static u32 __init build_bimm(s32 arg)
static __init u32 build_bimm(s32 arg)
{
{
	if (arg > 0x1ffff || arg < -0x20000)
	if (arg > 0x1ffff || arg < -0x20000)
		printk(KERN_WARNING "TLB synthesizer field overflow\n");
		printk(KERN_WARNING "TLB synthesizer field overflow\n");
@@ -252,7 +252,7 @@ static u32 __init build_bimm(s32 arg)
	return ((arg < 0) ? (1 << 15) : 0) | ((arg >> 2) & 0x7fff);
	return ((arg < 0) ? (1 << 15) : 0) | ((arg >> 2) & 0x7fff);
}
}


static u32 __init build_jimm(u32 arg)
static __init u32 build_jimm(u32 arg)
{
{
	if (arg & ~((JIMM_MASK) << 2))
	if (arg & ~((JIMM_MASK) << 2))
		printk(KERN_WARNING "TLB synthesizer field overflow\n");
		printk(KERN_WARNING "TLB synthesizer field overflow\n");
@@ -260,7 +260,7 @@ static u32 __init build_jimm(u32 arg)
	return (arg >> 2) & JIMM_MASK;
	return (arg >> 2) & JIMM_MASK;
}
}


static u32 __init build_func(u32 arg)
static __init u32 build_func(u32 arg)
{
{
	if (arg & ~FUNC_MASK)
	if (arg & ~FUNC_MASK)
		printk(KERN_WARNING "TLB synthesizer field overflow\n");
		printk(KERN_WARNING "TLB synthesizer field overflow\n");
@@ -268,7 +268,7 @@ static u32 __init build_func(u32 arg)
	return arg & FUNC_MASK;
	return arg & FUNC_MASK;
}
}


static u32 __init build_set(u32 arg)
static __init u32 build_set(u32 arg)
{
{
	if (arg & ~SET_MASK)
	if (arg & ~SET_MASK)
		printk(KERN_WARNING "TLB synthesizer field overflow\n");
		printk(KERN_WARNING "TLB synthesizer field overflow\n");
@@ -315,69 +315,69 @@ static void __init build_insn(u32 **buf, enum opcode opc, ...)
}
}


#define I_u1u2u3(op)						\
#define I_u1u2u3(op)						\
	static inline void i##op(u32 **buf, unsigned int a,	\
	static inline void __init i##op(u32 **buf, unsigned int a,	\
	 	unsigned int b, unsigned int c)			\
	 	unsigned int b, unsigned int c)			\
	{							\
	{							\
		build_insn(buf, insn##op, a, b, c);		\
		build_insn(buf, insn##op, a, b, c);		\
	}
	}


#define I_u2u1u3(op)						\
#define I_u2u1u3(op)						\
	static inline void i##op(u32 **buf, unsigned int a,	\
	static inline void __init i##op(u32 **buf, unsigned int a,	\
	 	unsigned int b, unsigned int c)			\
	 	unsigned int b, unsigned int c)			\
	{							\
	{							\
		build_insn(buf, insn##op, b, a, c);		\
		build_insn(buf, insn##op, b, a, c);		\
	}
	}


#define I_u3u1u2(op)						\
#define I_u3u1u2(op)						\
	static inline void i##op(u32 **buf, unsigned int a,	\
	static inline void __init i##op(u32 **buf, unsigned int a,	\
	 	unsigned int b, unsigned int c)			\
	 	unsigned int b, unsigned int c)			\
	{							\
	{							\
		build_insn(buf, insn##op, b, c, a);		\
		build_insn(buf, insn##op, b, c, a);		\
	}
	}


#define I_u1u2s3(op)						\
#define I_u1u2s3(op)						\
	static inline void i##op(u32 **buf, unsigned int a,	\
	static inline void __init i##op(u32 **buf, unsigned int a,	\
	 	unsigned int b, signed int c)			\
	 	unsigned int b, signed int c)			\
	{							\
	{							\
		build_insn(buf, insn##op, a, b, c);		\
		build_insn(buf, insn##op, a, b, c);		\
	}
	}


#define I_u2s3u1(op)						\
#define I_u2s3u1(op)						\
	static inline void i##op(u32 **buf, unsigned int a,	\
	static inline void __init i##op(u32 **buf, unsigned int a,	\
	 	signed int b, unsigned int c)			\
	 	signed int b, unsigned int c)			\
	{							\
	{							\
		build_insn(buf, insn##op, c, a, b);		\
		build_insn(buf, insn##op, c, a, b);		\
	}
	}


#define I_u2u1s3(op)						\
#define I_u2u1s3(op)						\
	static inline void i##op(u32 **buf, unsigned int a,	\
	static inline void __init i##op(u32 **buf, unsigned int a,	\
	 	unsigned int b, signed int c)			\
	 	unsigned int b, signed int c)			\
	{							\
	{							\
		build_insn(buf, insn##op, b, a, c);		\
		build_insn(buf, insn##op, b, a, c);		\
	}
	}


#define I_u1u2(op)						\
#define I_u1u2(op)						\
	static inline void i##op(u32 **buf, unsigned int a,	\
	static inline void __init i##op(u32 **buf, unsigned int a,	\
	 	unsigned int b)					\
	 	unsigned int b)					\
	{							\
	{							\
		build_insn(buf, insn##op, a, b);		\
		build_insn(buf, insn##op, a, b);		\
	}
	}


#define I_u1s2(op)						\
#define I_u1s2(op)						\
	static inline void i##op(u32 **buf, unsigned int a,	\
	static inline void __init i##op(u32 **buf, unsigned int a,	\
	 	signed int b)					\
	 	signed int b)					\
	{							\
	{							\
		build_insn(buf, insn##op, a, b);		\
		build_insn(buf, insn##op, a, b);		\
	}
	}


#define I_u1(op)						\
#define I_u1(op)						\
	static inline void i##op(u32 **buf, unsigned int a)	\
	static inline void __init i##op(u32 **buf, unsigned int a)	\
	{							\
	{							\
		build_insn(buf, insn##op, a);			\
		build_insn(buf, insn##op, a);			\
	}
	}


#define I_0(op)							\
#define I_0(op)							\
	static inline void i##op(u32 **buf)		\
	static inline void __init i##op(u32 **buf)		\
	{							\
	{							\
		build_insn(buf, insn##op);			\
		build_insn(buf, insn##op);			\
	}
	}
@@ -457,7 +457,7 @@ struct label {
	enum label_id lab;
	enum label_id lab;
};
};


static void __init build_label(struct label **lab, u32 *addr,
static __init void build_label(struct label **lab, u32 *addr,
			       enum label_id l)
			       enum label_id l)
{
{
	(*lab)->addr = addr;
	(*lab)->addr = addr;
@@ -526,34 +526,34 @@ L_LA(_r3000_write_probe_fail)
#define i_ehb(buf) i_sll(buf, 0, 0, 3)
#define i_ehb(buf) i_sll(buf, 0, 0, 3)


#ifdef CONFIG_64BIT
#ifdef CONFIG_64BIT
static int __init __maybe_unused in_compat_space_p(long addr)
static __init int __maybe_unused in_compat_space_p(long addr)
{
{
	/* Is this address in 32bit compat space? */
	/* Is this address in 32bit compat space? */
	return (((addr) & 0xffffffff00000000L) == 0xffffffff00000000L);
	return (((addr) & 0xffffffff00000000L) == 0xffffffff00000000L);
}
}


static int __init __maybe_unused rel_highest(long val)
static __init int __maybe_unused rel_highest(long val)
{
{
	return ((((val + 0x800080008000L) >> 48) & 0xffff) ^ 0x8000) - 0x8000;
	return ((((val + 0x800080008000L) >> 48) & 0xffff) ^ 0x8000) - 0x8000;
}
}


static int __init __maybe_unused rel_higher(long val)
static __init int __maybe_unused rel_higher(long val)
{
{
	return ((((val + 0x80008000L) >> 32) & 0xffff) ^ 0x8000) - 0x8000;
	return ((((val + 0x80008000L) >> 32) & 0xffff) ^ 0x8000) - 0x8000;
}
}
#endif
#endif


static int __init rel_hi(long val)
static __init int rel_hi(long val)
{
{
	return ((((val + 0x8000L) >> 16) & 0xffff) ^ 0x8000) - 0x8000;
	return ((((val + 0x8000L) >> 16) & 0xffff) ^ 0x8000) - 0x8000;
}
}


static int __init rel_lo(long val)
static __init int rel_lo(long val)
{
{
	return ((val & 0xffff) ^ 0x8000) - 0x8000;
	return ((val & 0xffff) ^ 0x8000) - 0x8000;
}
}


static void __init i_LA_mostly(u32 **buf, unsigned int rs, long addr)
static __init void i_LA_mostly(u32 **buf, unsigned int rs, long addr)
{
{
#ifdef CONFIG_64BIT
#ifdef CONFIG_64BIT
	if (!in_compat_space_p(addr)) {
	if (!in_compat_space_p(addr)) {
@@ -571,7 +571,7 @@ static void __init i_LA_mostly(u32 **buf, unsigned int rs, long addr)
		i_lui(buf, rs, rel_hi(addr));
		i_lui(buf, rs, rel_hi(addr));
}
}


static void __init __maybe_unused i_LA(u32 **buf, unsigned int rs,
static __init void __maybe_unused i_LA(u32 **buf, unsigned int rs,
					     long addr)
					     long addr)
{
{
	i_LA_mostly(buf, rs, addr);
	i_LA_mostly(buf, rs, addr);
@@ -589,7 +589,7 @@ struct reloc {
	enum label_id lab;
	enum label_id lab;
};
};


static void __init r_mips_pc16(struct reloc **rel, u32 *addr,
static __init void r_mips_pc16(struct reloc **rel, u32 *addr,
			       enum label_id l)
			       enum label_id l)
{
{
	(*rel)->addr = addr;
	(*rel)->addr = addr;
@@ -614,7 +614,7 @@ static inline void __resolve_relocs(struct reloc *rel, struct label *lab)
	}
	}
}
}


static void __init resolve_relocs(struct reloc *rel, struct label *lab)
static __init void resolve_relocs(struct reloc *rel, struct label *lab)
{
{
	struct label *l;
	struct label *l;


@@ -624,7 +624,7 @@ static void __init resolve_relocs(struct reloc *rel, struct label *lab)
				__resolve_relocs(rel, l);
				__resolve_relocs(rel, l);
}
}


static void __init move_relocs(struct reloc *rel, u32 *first, u32 *end,
static __init void move_relocs(struct reloc *rel, u32 *first, u32 *end,
			       long off)
			       long off)
{
{
	for (; rel->lab != label_invalid; rel++)
	for (; rel->lab != label_invalid; rel++)
@@ -632,7 +632,7 @@ static void __init move_relocs(struct reloc *rel, u32 *first, u32 *end,
			rel->addr += off;
			rel->addr += off;
}
}


static void __init move_labels(struct label *lab, u32 *first, u32 *end,
static __init void move_labels(struct label *lab, u32 *first, u32 *end,
			       long off)
			       long off)
{
{
	for (; lab->lab != label_invalid; lab++)
	for (; lab->lab != label_invalid; lab++)
@@ -640,7 +640,7 @@ static void __init move_labels(struct label *lab, u32 *first, u32 *end,
			lab->addr += off;
			lab->addr += off;
}
}


static void __init copy_handler(struct reloc *rel, struct label *lab,
static __init void copy_handler(struct reloc *rel, struct label *lab,
				u32 *first, u32 *end, u32 *target)
				u32 *first, u32 *end, u32 *target)
{
{
	long off = (long)(target - first);
	long off = (long)(target - first);
@@ -651,7 +651,7 @@ static void __init copy_handler(struct reloc *rel, struct label *lab,
	move_labels(lab, first, end, off);
	move_labels(lab, first, end, off);
}
}


static int __init __maybe_unused insn_has_bdelay(struct reloc *rel,
static __init int __maybe_unused insn_has_bdelay(struct reloc *rel,
						       u32 *addr)
						       u32 *addr)
{
{
	for (; rel->lab != label_invalid; rel++) {
	for (; rel->lab != label_invalid; rel++) {
@@ -743,11 +743,11 @@ il_bgez(u32 **p, struct reloc **r, unsigned int reg, enum label_id l)
 * We deliberately chose a buffer size of 128, so we won't scribble
 * We deliberately chose a buffer size of 128, so we won't scribble
 * over anything important on overflow before we panic.
 * over anything important on overflow before we panic.
 */
 */
static u32 tlb_handler[128] __initdata;
static __initdata u32 tlb_handler[128];


/* simply assume worst case size for labels and relocs */
/* simply assume worst case size for labels and relocs */
static struct label labels[128] __initdata;
static __initdata struct label labels[128];
static struct reloc relocs[128] __initdata;
static __initdata struct reloc relocs[128];


/*
/*
 * The R3000 TLB handler is simple.
 * The R3000 TLB handler is simple.
@@ -801,7 +801,7 @@ static void __init build_r3000_tlb_refill_handler(void)
 * other one.To keep things simple, we first assume linear space,
 * other one.To keep things simple, we first assume linear space,
 * then we relocate it to the final handler layout as needed.
 * then we relocate it to the final handler layout as needed.
 */
 */
static u32 final_handler[64] __initdata;
static __initdata u32 final_handler[64];


/*
/*
 * Hazards
 * Hazards
@@ -825,7 +825,7 @@ static u32 final_handler[64] __initdata;
 *
 *
 * As if we MIPS hackers wouldn't know how to nop pipelines happy ...
 * As if we MIPS hackers wouldn't know how to nop pipelines happy ...
 */
 */
static void __init __maybe_unused build_tlb_probe_entry(u32 **p)
static __init void __maybe_unused build_tlb_probe_entry(u32 **p)
{
{
	switch (current_cpu_type()) {
	switch (current_cpu_type()) {
	/* Found by experiment: R4600 v2.0 needs this, too.  */
	/* Found by experiment: R4600 v2.0 needs this, too.  */
@@ -849,7 +849,7 @@ static void __init __maybe_unused build_tlb_probe_entry(u32 **p)
 */
 */
enum tlb_write_entry { tlb_random, tlb_indexed };
enum tlb_write_entry { tlb_random, tlb_indexed };


static void __init build_tlb_write_entry(u32 **p, struct label **l,
static __init void build_tlb_write_entry(u32 **p, struct label **l,
					 struct reloc **r,
					 struct reloc **r,
					 enum tlb_write_entry wmode)
					 enum tlb_write_entry wmode)
{
{
@@ -993,7 +993,7 @@ static void __init build_tlb_write_entry(u32 **p, struct label **l,
 * TMP and PTR are scratch.
 * TMP and PTR are scratch.
 * TMP will be clobbered, PTR will hold the pmd entry.
 * TMP will be clobbered, PTR will hold the pmd entry.
 */
 */
static void __init
static __init void
build_get_pmde64(u32 **p, struct label **l, struct reloc **r,
build_get_pmde64(u32 **p, struct label **l, struct reloc **r,
		 unsigned int tmp, unsigned int ptr)
		 unsigned int tmp, unsigned int ptr)
{
{
@@ -1054,7 +1054,7 @@ build_get_pmde64(u32 **p, struct label **l, struct reloc **r,
 * BVADDR is the faulting address, PTR is scratch.
 * BVADDR is the faulting address, PTR is scratch.
 * PTR will hold the pgd for vmalloc.
 * PTR will hold the pgd for vmalloc.
 */
 */
static void __init
static __init void
build_get_pgd_vmalloc64(u32 **p, struct label **l, struct reloc **r,
build_get_pgd_vmalloc64(u32 **p, struct label **l, struct reloc **r,
			unsigned int bvaddr, unsigned int ptr)
			unsigned int bvaddr, unsigned int ptr)
{
{
@@ -1118,7 +1118,7 @@ build_get_pgd_vmalloc64(u32 **p, struct label **l, struct reloc **r,
 * TMP and PTR are scratch.
 * TMP and PTR are scratch.
 * TMP will be clobbered, PTR will hold the pgd entry.
 * TMP will be clobbered, PTR will hold the pgd entry.
 */
 */
static void __init __maybe_unused
static __init void __maybe_unused
build_get_pgde32(u32 **p, unsigned int tmp, unsigned int ptr)
build_get_pgde32(u32 **p, unsigned int tmp, unsigned int ptr)
{
{
	long pgdc = (long)pgd_current;
	long pgdc = (long)pgd_current;
@@ -1153,7 +1153,7 @@ build_get_pgde32(u32 **p, unsigned int tmp, unsigned int ptr)


#endif /* !CONFIG_64BIT */
#endif /* !CONFIG_64BIT */


static void __init build_adjust_context(u32 **p, unsigned int ctx)
static __init void build_adjust_context(u32 **p, unsigned int ctx)
{
{
	unsigned int shift = 4 - (PTE_T_LOG2 + 1) + PAGE_SHIFT - 12;
	unsigned int shift = 4 - (PTE_T_LOG2 + 1) + PAGE_SHIFT - 12;
	unsigned int mask = (PTRS_PER_PTE / 2 - 1) << (PTE_T_LOG2 + 1);
	unsigned int mask = (PTRS_PER_PTE / 2 - 1) << (PTE_T_LOG2 + 1);
@@ -1179,7 +1179,7 @@ static void __init build_adjust_context(u32 **p, unsigned int ctx)
	i_andi(p, ctx, ctx, mask);
	i_andi(p, ctx, ctx, mask);
}
}


static void __init build_get_ptep(u32 **p, unsigned int tmp, unsigned int ptr)
static __init void build_get_ptep(u32 **p, unsigned int tmp, unsigned int ptr)
{
{
	/*
	/*
	 * Bug workaround for the Nevada. It seems as if under certain
	 * Bug workaround for the Nevada. It seems as if under certain
@@ -1204,7 +1204,7 @@ static void __init build_get_ptep(u32 **p, unsigned int tmp, unsigned int ptr)
	i_ADDU(p, ptr, ptr, tmp); /* add in offset */
	i_ADDU(p, ptr, ptr, tmp); /* add in offset */
}
}


static void __init build_update_entries(u32 **p, unsigned int tmp,
static __init void build_update_entries(u32 **p, unsigned int tmp,
					unsigned int ptep)
					unsigned int ptep)
{
{
	/*
	/*
+1 −1
Original line number Original line Diff line number Diff line
@@ -154,7 +154,7 @@ int64_t cfe_getticks(void);
#define cfe_readblk(a, b, c, d)		__cfe_readblk(a, b, c, d)
#define cfe_readblk(a, b, c, d)		__cfe_readblk(a, b, c, d)
#define cfe_setenv(a, b)		__cfe_setenv(a, b)
#define cfe_setenv(a, b)		__cfe_setenv(a, b)
#define cfe_write(a, b, c)		__cfe_write(a, b, c)
#define cfe_write(a, b, c)		__cfe_write(a, b, c)
#define cfe_writeblk(a, b, c, d		__cfe_writeblk(a, b, c, d)
#define cfe_writeblk(a, b, c, d)	__cfe_writeblk(a, b, c, d)
#endif				/* CFE_API_IMPL_NAMESPACE */
#endif				/* CFE_API_IMPL_NAMESPACE */


int cfe_close(int handle);
int cfe_close(int handle);