Donate to e Foundation | Murena handsets with /e/OS | Own a part of Murena! Learn more

Commit 144cf864 authored by Cong Wang's avatar Cong Wang
Browse files

frv: remove the second parameter of kmap_atomic_primary()



All callers of kmap_atomic_primary() use __KM_CACHE, so it can be
removed safely, and __kmap_atomic_primary() only check if 'type' if
__KM_CACHE or not, so 'type' can be changed to a boolean as well.

Ditto for kunmap_atomic_primary()/__kunmap_atomic_primary().

Acked-by: default avatarGeert Uytterhoeven <geert@linux-m68k.org>
Signed-off-by: default avatarCong Wang <amwang@redhat.com>
parent 906adea1
Loading
Loading
Loading
Loading
+9 −25
Original line number Diff line number Diff line
@@ -76,15 +76,16 @@ extern struct page *kmap_atomic_to_page(void *ptr);

#ifndef __ASSEMBLY__

#define __kmap_atomic_primary(type, paddr, ampr)						\
#define __kmap_atomic_primary(cached, paddr, ampr)						\
({												\
	unsigned long damlr, dampr;								\
												\
	dampr = paddr | xAMPRx_L | xAMPRx_M | xAMPRx_S | xAMPRx_SS_16Kb | xAMPRx_V;		\
												\
	if (type != __KM_CACHE)									\
	if (!cached)										\
		asm volatile("movgs %0,dampr"#ampr :: "r"(dampr) : "memory");			\
	else											\
		/* cache flush page attachment point */						\
		asm volatile("movgs %0,iampr"#ampr"\n"						\
			     "movgs %0,dampr"#ampr"\n"						\
			     :: "r"(dampr) : "memory"						\
@@ -112,29 +113,20 @@ extern struct page *kmap_atomic_to_page(void *ptr);
	(void *) damlr;										  \
})

static inline void *kmap_atomic_primary(struct page *page, enum km_type type)
static inline void *kmap_atomic_primary(struct page *page)
{
	unsigned long paddr;

	pagefault_disable();
	paddr = page_to_phys(page);

	switch (type) {
        case 0:		return __kmap_atomic_primary(0, paddr, 2);
        case 1:		return __kmap_atomic_primary(1, paddr, 3);
        case 2:		return __kmap_atomic_primary(2, paddr, 4);
        case 3:		return __kmap_atomic_primary(3, paddr, 5);

	default:
		BUG();
		return NULL;
	}
        return __kmap_atomic_primary(1, paddr, 2);
}

#define __kunmap_atomic_primary(type, ampr)				\
#define __kunmap_atomic_primary(cached, ampr)				\
do {									\
	asm volatile("movgs gr0,dampr"#ampr"\n" ::: "memory");		\
	if (type == __KM_CACHE)						\
	if (cached)							\
		asm volatile("movgs gr0,iampr"#ampr"\n" ::: "memory");	\
} while(0)

@@ -143,17 +135,9 @@ do { \
	asm volatile("tlbpr %0,gr0,#4,#1" : : "r"(vaddr) : "memory");	\
} while(0)

static inline void kunmap_atomic_primary(void *kvaddr, enum km_type type)
static inline void kunmap_atomic_primary(void *kvaddr)
{
	switch (type) {
        case 0:		__kunmap_atomic_primary(0, 2);	break;
        case 1:		__kunmap_atomic_primary(1, 3);	break;
        case 2:		__kunmap_atomic_primary(2, 4);	break;
        case 3:		__kunmap_atomic_primary(3, 5);	break;

	default:
		BUG();
	}
        __kunmap_atomic_primary(1, 2);
	pagefault_enable();
}

+2 −2
Original line number Diff line number Diff line
@@ -62,14 +62,14 @@ int dma_map_sg(struct device *dev, struct scatterlist *sg, int nents,
	dampr2 = __get_DAMPR(2);

	for (i = 0; i < nents; i++) {
		vaddr = kmap_atomic_primary(sg_page(&sg[i]), __KM_CACHE);
		vaddr = kmap_atomic_primary(sg_page(&sg[i]));

		frv_dcache_writeback((unsigned long) vaddr,
				     (unsigned long) vaddr + PAGE_SIZE);

	}

	kunmap_atomic_primary(vaddr, __KM_CACHE);
	kunmap_atomic_primary(vaddr);
	if (dampr2) {
		__set_DAMPR(2, dampr2);
		__set_IAMPR(2, dampr2);
+4 −4
Original line number Diff line number Diff line
@@ -26,11 +26,11 @@ void flush_dcache_page(struct page *page)

	dampr2 = __get_DAMPR(2);

	vaddr = kmap_atomic_primary(page, __KM_CACHE);
	vaddr = kmap_atomic_primary(page);

	frv_dcache_writeback((unsigned long) vaddr, (unsigned long) vaddr + PAGE_SIZE);

	kunmap_atomic_primary(vaddr, __KM_CACHE);
	kunmap_atomic_primary(vaddr);

	if (dampr2) {
		__set_DAMPR(2, dampr2);
@@ -54,12 +54,12 @@ void flush_icache_user_range(struct vm_area_struct *vma, struct page *page,

	dampr2 = __get_DAMPR(2);

	vaddr = kmap_atomic_primary(page, __KM_CACHE);
	vaddr = kmap_atomic_primary(page);

	start = (start & ~PAGE_MASK) | (unsigned long) vaddr;
	frv_cache_wback_inv(start, start + len);

	kunmap_atomic_primary(vaddr, __KM_CACHE);
	kunmap_atomic_primary(vaddr);

	if (dampr2) {
		__set_DAMPR(2, dampr2);
+10 −10
Original line number Diff line number Diff line
@@ -50,11 +50,11 @@ void *kmap_atomic(struct page *page)
	/*
	 * The first 4 primary maps are reserved for architecture code
	 */
	case 0:		return __kmap_atomic_primary(4, paddr, 6);
	case 1:		return __kmap_atomic_primary(5, paddr, 7);
	case 2:		return __kmap_atomic_primary(6, paddr, 8);
	case 3:		return __kmap_atomic_primary(7, paddr, 9);
	case 4:		return __kmap_atomic_primary(8, paddr, 10);
	case 0:		return __kmap_atomic_primary(0, paddr, 6);
	case 1:		return __kmap_atomic_primary(0, paddr, 7);
	case 2:		return __kmap_atomic_primary(0, paddr, 8);
	case 3:		return __kmap_atomic_primary(0, paddr, 9);
	case 4:		return __kmap_atomic_primary(0, paddr, 10);

	case 5 ... 5 + NR_TLB_LINES - 1:
		return __kmap_atomic_secondary(type - 5, paddr);
@@ -70,11 +70,11 @@ void __kunmap_atomic(void *kvaddr)
{
	int type = kmap_atomic_idx();
	switch (type) {
	case 0:		__kunmap_atomic_primary(4, 6);	break;
	case 1:		__kunmap_atomic_primary(5, 7);	break;
	case 2:		__kunmap_atomic_primary(6, 8);	break;
	case 3:		__kunmap_atomic_primary(7, 9);	break;
	case 4:		__kunmap_atomic_primary(8, 10);	break;
	case 0:		__kunmap_atomic_primary(0, 6);	break;
	case 1:		__kunmap_atomic_primary(0, 7);	break;
	case 2:		__kunmap_atomic_primary(0, 8);	break;
	case 3:		__kunmap_atomic_primary(0, 9);	break;
	case 4:		__kunmap_atomic_primary(0, 10);	break;

	case 5 ... 5 + NR_TLB_LINES - 1:
		__kunmap_atomic_secondary(type - 5, kvaddr);