Donate to e Foundation | Murena handsets with /e/OS | Own a part of Murena! Learn more

Commit 267b50fe authored by Linus Torvalds's avatar Linus Torvalds
Browse files
Pull s390 fixes from Martin Schwidefsky:
 "Bug fixes for 3.6-rc7, including some important patches for large page
  related memory management issues."

* 'for-linus' of git://git.kernel.org/pub/scm/linux/kernel/git/s390/linux:
  s390/dasd: fix read unit address configuration loop
  s390/dasd: fix pathgroup race
  s390/mm: fix user access page-table walk code
  s390/hwcaps: do not report high gprs for 31 bit kernel
  s390/cio: invalidate cdev pointer before deregistration
  s390/cio: fix IO subchannel event race
  s390/dasd: move wake_up call
  s390/hugetlb: use direct TLB flushing for hugetlbfs pages
  s390/mm: fix deadlock in unmap_hugepage_range()
parents 8ca7de91 03429f34
Loading
Loading
Loading
Loading
+10 −14
Original line number Original line Diff line number Diff line
@@ -66,16 +66,6 @@ static inline pte_t huge_ptep_get(pte_t *ptep)
	return pte;
	return pte;
}
}


static inline pte_t huge_ptep_get_and_clear(struct mm_struct *mm,
					    unsigned long addr, pte_t *ptep)
{
	pte_t pte = huge_ptep_get(ptep);

	mm->context.flush_mm = 1;
	pmd_clear((pmd_t *) ptep);
	return pte;
}

static inline void __pmd_csp(pmd_t *pmdp)
static inline void __pmd_csp(pmd_t *pmdp)
{
{
	register unsigned long reg2 asm("2") = pmd_val(*pmdp);
	register unsigned long reg2 asm("2") = pmd_val(*pmdp);
@@ -117,6 +107,15 @@ static inline void huge_ptep_invalidate(struct mm_struct *mm,
		__pmd_csp(pmdp);
		__pmd_csp(pmdp);
}
}


static inline pte_t huge_ptep_get_and_clear(struct mm_struct *mm,
					    unsigned long addr, pte_t *ptep)
{
	pte_t pte = huge_ptep_get(ptep);

	huge_ptep_invalidate(mm, addr, ptep);
	return pte;
}

#define huge_ptep_set_access_flags(__vma, __addr, __ptep, __entry, __dirty) \
#define huge_ptep_set_access_flags(__vma, __addr, __ptep, __entry, __dirty) \
({									    \
({									    \
	int __changed = !pte_same(huge_ptep_get(__ptep), __entry);	    \
	int __changed = !pte_same(huge_ptep_get(__ptep), __entry);	    \
@@ -131,9 +130,6 @@ static inline void huge_ptep_invalidate(struct mm_struct *mm,
({									\
({									\
	pte_t __pte = huge_ptep_get(__ptep);				\
	pte_t __pte = huge_ptep_get(__ptep);				\
	if (pte_write(__pte)) {						\
	if (pte_write(__pte)) {						\
		(__mm)->context.flush_mm = 1;				\
		if (atomic_read(&(__mm)->context.attach_count) > 1 ||	\
		    (__mm) != current->active_mm)			\
		huge_ptep_invalidate(__mm, __addr, __ptep);		\
		huge_ptep_invalidate(__mm, __addr, __ptep);		\
		set_huge_pte_at(__mm, __addr, __ptep,			\
		set_huge_pte_at(__mm, __addr, __ptep,			\
				huge_pte_wrprotect(__pte));		\
				huge_pte_wrprotect(__pte));		\
+0 −2
Original line number Original line Diff line number Diff line
@@ -90,12 +90,10 @@ static inline void __tlb_flush_mm(struct mm_struct * mm)


static inline void __tlb_flush_mm_cond(struct mm_struct * mm)
static inline void __tlb_flush_mm_cond(struct mm_struct * mm)
{
{
	spin_lock(&mm->page_table_lock);
	if (mm->context.flush_mm) {
	if (mm->context.flush_mm) {
		__tlb_flush_mm(mm);
		__tlb_flush_mm(mm);
		mm->context.flush_mm = 0;
		mm->context.flush_mm = 0;
	}
	}
	spin_unlock(&mm->page_table_lock);
}
}


/*
/*
+2 −0
Original line number Original line Diff line number Diff line
@@ -974,11 +974,13 @@ static void __init setup_hwcaps(void)
	if (MACHINE_HAS_HPAGE)
	if (MACHINE_HAS_HPAGE)
		elf_hwcap |= HWCAP_S390_HPAGE;
		elf_hwcap |= HWCAP_S390_HPAGE;


#if defined(CONFIG_64BIT)
	/*
	/*
	 * 64-bit register support for 31-bit processes
	 * 64-bit register support for 31-bit processes
	 * HWCAP_S390_HIGH_GPRS is bit 9.
	 * HWCAP_S390_HIGH_GPRS is bit 9.
	 */
	 */
	elf_hwcap |= HWCAP_S390_HIGH_GPRS;
	elf_hwcap |= HWCAP_S390_HIGH_GPRS;
#endif


	get_cpu_id(&cpu_id);
	get_cpu_id(&cpu_id);
	switch (cpu_id.machine) {
	switch (cpu_id.machine) {
+63 −79
Original line number Original line Diff line number Diff line
@@ -2,69 +2,82 @@
 *  User access functions based on page table walks for enhanced
 *  User access functions based on page table walks for enhanced
 *  system layout without hardware support.
 *  system layout without hardware support.
 *
 *
 *    Copyright IBM Corp. 2006
 *    Copyright IBM Corp. 2006, 2012
 *    Author(s): Gerald Schaefer (gerald.schaefer@de.ibm.com)
 *    Author(s): Gerald Schaefer (gerald.schaefer@de.ibm.com)
 */
 */


#include <linux/errno.h>
#include <linux/errno.h>
#include <linux/hardirq.h>
#include <linux/hardirq.h>
#include <linux/mm.h>
#include <linux/mm.h>
#include <linux/hugetlb.h>
#include <asm/uaccess.h>
#include <asm/uaccess.h>
#include <asm/futex.h>
#include <asm/futex.h>
#include "uaccess.h"
#include "uaccess.h"


static inline pte_t *follow_table(struct mm_struct *mm, unsigned long addr)

/*
 * Returns kernel address for user virtual address. If the returned address is
 * >= -4095 (IS_ERR_VALUE(x) returns true), a fault has occured and the address
 * contains the (negative) exception code.
 */
static __always_inline unsigned long follow_table(struct mm_struct *mm,
						  unsigned long addr, int write)
{
{
	pgd_t *pgd;
	pgd_t *pgd;
	pud_t *pud;
	pud_t *pud;
	pmd_t *pmd;
	pmd_t *pmd;
	pte_t *ptep;


	pgd = pgd_offset(mm, addr);
	pgd = pgd_offset(mm, addr);
	if (pgd_none(*pgd) || unlikely(pgd_bad(*pgd)))
	if (pgd_none(*pgd) || unlikely(pgd_bad(*pgd)))
		return (pte_t *) 0x3a;
		return -0x3aUL;


	pud = pud_offset(pgd, addr);
	pud = pud_offset(pgd, addr);
	if (pud_none(*pud) || unlikely(pud_bad(*pud)))
	if (pud_none(*pud) || unlikely(pud_bad(*pud)))
		return (pte_t *) 0x3b;
		return -0x3bUL;


	pmd = pmd_offset(pud, addr);
	pmd = pmd_offset(pud, addr);
	if (pmd_none(*pmd) || unlikely(pmd_bad(*pmd)))
	if (pmd_none(*pmd))
		return (pte_t *) 0x10;
		return -0x10UL;
	if (pmd_huge(*pmd)) {
		if (write && (pmd_val(*pmd) & _SEGMENT_ENTRY_RO))
			return -0x04UL;
		return (pmd_val(*pmd) & HPAGE_MASK) + (addr & ~HPAGE_MASK);
	}
	if (unlikely(pmd_bad(*pmd)))
		return -0x10UL;

	ptep = pte_offset_map(pmd, addr);
	if (!pte_present(*ptep))
		return -0x11UL;
	if (write && !pte_write(*ptep))
		return -0x04UL;


	return pte_offset_map(pmd, addr);
	return (pte_val(*ptep) & PAGE_MASK) + (addr & ~PAGE_MASK);
}
}


static __always_inline size_t __user_copy_pt(unsigned long uaddr, void *kptr,
static __always_inline size_t __user_copy_pt(unsigned long uaddr, void *kptr,
					     size_t n, int write_user)
					     size_t n, int write_user)
{
{
	struct mm_struct *mm = current->mm;
	struct mm_struct *mm = current->mm;
	unsigned long offset, pfn, done, size;
	unsigned long offset, done, size, kaddr;
	pte_t *pte;
	void *from, *to;
	void *from, *to;


	done = 0;
	done = 0;
retry:
retry:
	spin_lock(&mm->page_table_lock);
	spin_lock(&mm->page_table_lock);
	do {
	do {
		pte = follow_table(mm, uaddr);
		kaddr = follow_table(mm, uaddr, write_user);
		if ((unsigned long) pte < 0x1000)
		if (IS_ERR_VALUE(kaddr))
			goto fault;
			goto fault;
		if (!pte_present(*pte)) {
			pte = (pte_t *) 0x11;
			goto fault;
		} else if (write_user && !pte_write(*pte)) {
			pte = (pte_t *) 0x04;
			goto fault;
		}


		pfn = pte_pfn(*pte);
		offset = uaddr & ~PAGE_MASK;
		offset = uaddr & (PAGE_SIZE - 1);
		size = min(n - done, PAGE_SIZE - offset);
		size = min(n - done, PAGE_SIZE - offset);
		if (write_user) {
		if (write_user) {
			to = (void *)((pfn << PAGE_SHIFT) + offset);
			to = (void *) kaddr;
			from = kptr + done;
			from = kptr + done;
		} else {
		} else {
			from = (void *)((pfn << PAGE_SHIFT) + offset);
			from = (void *) kaddr;
			to = kptr + done;
			to = kptr + done;
		}
		}
		memcpy(to, from, size);
		memcpy(to, from, size);
@@ -75,7 +88,7 @@ retry:
	return n - done;
	return n - done;
fault:
fault:
	spin_unlock(&mm->page_table_lock);
	spin_unlock(&mm->page_table_lock);
	if (__handle_fault(uaddr, (unsigned long) pte, write_user))
	if (__handle_fault(uaddr, -kaddr, write_user))
		return n - done;
		return n - done;
	goto retry;
	goto retry;
}
}
@@ -84,27 +97,22 @@ fault:
 * Do DAT for user address by page table walk, return kernel address.
 * Do DAT for user address by page table walk, return kernel address.
 * This function needs to be called with current->mm->page_table_lock held.
 * This function needs to be called with current->mm->page_table_lock held.
 */
 */
static __always_inline unsigned long __dat_user_addr(unsigned long uaddr)
static __always_inline unsigned long __dat_user_addr(unsigned long uaddr,
						     int write)
{
{
	struct mm_struct *mm = current->mm;
	struct mm_struct *mm = current->mm;
	unsigned long pfn;
	unsigned long kaddr;
	pte_t *pte;
	int rc;
	int rc;


retry:
retry:
	pte = follow_table(mm, uaddr);
	kaddr = follow_table(mm, uaddr, write);
	if ((unsigned long) pte < 0x1000)
	if (IS_ERR_VALUE(kaddr))
		goto fault;
	if (!pte_present(*pte)) {
		pte = (pte_t *) 0x11;
		goto fault;
		goto fault;
	}


	pfn = pte_pfn(*pte);
	return kaddr;
	return (pfn << PAGE_SHIFT) + (uaddr & (PAGE_SIZE - 1));
fault:
fault:
	spin_unlock(&mm->page_table_lock);
	spin_unlock(&mm->page_table_lock);
	rc = __handle_fault(uaddr, (unsigned long) pte, 0);
	rc = __handle_fault(uaddr, -kaddr, write);
	spin_lock(&mm->page_table_lock);
	spin_lock(&mm->page_table_lock);
	if (!rc)
	if (!rc)
		goto retry;
		goto retry;
@@ -159,11 +167,9 @@ static size_t clear_user_pt(size_t n, void __user *to)


static size_t strnlen_user_pt(size_t count, const char __user *src)
static size_t strnlen_user_pt(size_t count, const char __user *src)
{
{
	char *addr;
	unsigned long uaddr = (unsigned long) src;
	unsigned long uaddr = (unsigned long) src;
	struct mm_struct *mm = current->mm;
	struct mm_struct *mm = current->mm;
	unsigned long offset, pfn, done, len;
	unsigned long offset, done, len, kaddr;
	pte_t *pte;
	size_t len_str;
	size_t len_str;


	if (segment_eq(get_fs(), KERNEL_DS))
	if (segment_eq(get_fs(), KERNEL_DS))
@@ -172,19 +178,13 @@ static size_t strnlen_user_pt(size_t count, const char __user *src)
retry:
retry:
	spin_lock(&mm->page_table_lock);
	spin_lock(&mm->page_table_lock);
	do {
	do {
		pte = follow_table(mm, uaddr);
		kaddr = follow_table(mm, uaddr, 0);
		if ((unsigned long) pte < 0x1000)
		if (IS_ERR_VALUE(kaddr))
			goto fault;
		if (!pte_present(*pte)) {
			pte = (pte_t *) 0x11;
			goto fault;
			goto fault;
		}


		pfn = pte_pfn(*pte);
		offset = uaddr & ~PAGE_MASK;
		offset = uaddr & (PAGE_SIZE-1);
		addr = (char *)(pfn << PAGE_SHIFT) + offset;
		len = min(count - done, PAGE_SIZE - offset);
		len = min(count - done, PAGE_SIZE - offset);
		len_str = strnlen(addr, len);
		len_str = strnlen((char *) kaddr, len);
		done += len_str;
		done += len_str;
		uaddr += len_str;
		uaddr += len_str;
	} while ((len_str == len) && (done < count));
	} while ((len_str == len) && (done < count));
@@ -192,7 +192,7 @@ retry:
	return done + 1;
	return done + 1;
fault:
fault:
	spin_unlock(&mm->page_table_lock);
	spin_unlock(&mm->page_table_lock);
	if (__handle_fault(uaddr, (unsigned long) pte, 0))
	if (__handle_fault(uaddr, -kaddr, 0))
		return 0;
		return 0;
	goto retry;
	goto retry;
}
}
@@ -225,11 +225,10 @@ static size_t copy_in_user_pt(size_t n, void __user *to,
			      const void __user *from)
			      const void __user *from)
{
{
	struct mm_struct *mm = current->mm;
	struct mm_struct *mm = current->mm;
	unsigned long offset_from, offset_to, offset_max, pfn_from, pfn_to,
	unsigned long offset_max, uaddr, done, size, error_code;
		      uaddr, done, size, error_code;
	unsigned long uaddr_from = (unsigned long) from;
	unsigned long uaddr_from = (unsigned long) from;
	unsigned long uaddr_to = (unsigned long) to;
	unsigned long uaddr_to = (unsigned long) to;
	pte_t *pte_from, *pte_to;
	unsigned long kaddr_to, kaddr_from;
	int write_user;
	int write_user;


	if (segment_eq(get_fs(), KERNEL_DS)) {
	if (segment_eq(get_fs(), KERNEL_DS)) {
@@ -242,38 +241,23 @@ retry:
	do {
	do {
		write_user = 0;
		write_user = 0;
		uaddr = uaddr_from;
		uaddr = uaddr_from;
		pte_from = follow_table(mm, uaddr_from);
		kaddr_from = follow_table(mm, uaddr_from, 0);
		error_code = (unsigned long) pte_from;
		error_code = kaddr_from;
		if (error_code < 0x1000)
		if (IS_ERR_VALUE(error_code))
			goto fault;
		if (!pte_present(*pte_from)) {
			error_code = 0x11;
			goto fault;
			goto fault;
		}


		write_user = 1;
		write_user = 1;
		uaddr = uaddr_to;
		uaddr = uaddr_to;
		pte_to = follow_table(mm, uaddr_to);
		kaddr_to = follow_table(mm, uaddr_to, 1);
		error_code = (unsigned long) pte_to;
		error_code = (unsigned long) kaddr_to;
		if (error_code < 0x1000)
		if (IS_ERR_VALUE(error_code))
			goto fault;
		if (!pte_present(*pte_to)) {
			error_code = 0x11;
			goto fault;
			goto fault;
		} else if (!pte_write(*pte_to)) {
			error_code = 0x04;
			goto fault;
		}


		pfn_from = pte_pfn(*pte_from);
		offset_max = max(uaddr_from & ~PAGE_MASK,
		pfn_to = pte_pfn(*pte_to);
				 uaddr_to & ~PAGE_MASK);
		offset_from = uaddr_from & (PAGE_SIZE-1);
		offset_to = uaddr_from & (PAGE_SIZE-1);
		offset_max = max(offset_from, offset_to);
		size = min(n - done, PAGE_SIZE - offset_max);
		size = min(n - done, PAGE_SIZE - offset_max);


		memcpy((void *)(pfn_to << PAGE_SHIFT) + offset_to,
		memcpy((void *) kaddr_to, (void *) kaddr_from, size);
		       (void *)(pfn_from << PAGE_SHIFT) + offset_from, size);
		done += size;
		done += size;
		uaddr_from += size;
		uaddr_from += size;
		uaddr_to += size;
		uaddr_to += size;
@@ -282,7 +266,7 @@ retry:
	return n - done;
	return n - done;
fault:
fault:
	spin_unlock(&mm->page_table_lock);
	spin_unlock(&mm->page_table_lock);
	if (__handle_fault(uaddr, error_code, write_user))
	if (__handle_fault(uaddr, -error_code, write_user))
		return n - done;
		return n - done;
	goto retry;
	goto retry;
}
}
@@ -341,7 +325,7 @@ int futex_atomic_op_pt(int op, u32 __user *uaddr, int oparg, int *old)
		return __futex_atomic_op_pt(op, uaddr, oparg, old);
		return __futex_atomic_op_pt(op, uaddr, oparg, old);
	spin_lock(&current->mm->page_table_lock);
	spin_lock(&current->mm->page_table_lock);
	uaddr = (u32 __force __user *)
	uaddr = (u32 __force __user *)
		__dat_user_addr((__force unsigned long) uaddr);
		__dat_user_addr((__force unsigned long) uaddr, 1);
	if (!uaddr) {
	if (!uaddr) {
		spin_unlock(&current->mm->page_table_lock);
		spin_unlock(&current->mm->page_table_lock);
		return -EFAULT;
		return -EFAULT;
@@ -378,7 +362,7 @@ int futex_atomic_cmpxchg_pt(u32 *uval, u32 __user *uaddr,
		return __futex_atomic_cmpxchg_pt(uval, uaddr, oldval, newval);
		return __futex_atomic_cmpxchg_pt(uval, uaddr, oldval, newval);
	spin_lock(&current->mm->page_table_lock);
	spin_lock(&current->mm->page_table_lock);
	uaddr = (u32 __force __user *)
	uaddr = (u32 __force __user *)
		__dat_user_addr((__force unsigned long) uaddr);
		__dat_user_addr((__force unsigned long) uaddr, 1);
	if (!uaddr) {
	if (!uaddr) {
		spin_unlock(&current->mm->page_table_lock);
		spin_unlock(&current->mm->page_table_lock);
		return -EFAULT;
		return -EFAULT;
+14 −3
Original line number Original line Diff line number Diff line
@@ -534,11 +534,11 @@ static void dasd_change_state(struct dasd_device *device)
	if (rc)
	if (rc)
		device->target = device->state;
		device->target = device->state;


	if (device->state == device->target)
		wake_up(&dasd_init_waitq);

	/* let user-space know that the device status changed */
	/* let user-space know that the device status changed */
	kobject_uevent(&device->cdev->dev.kobj, KOBJ_CHANGE);
	kobject_uevent(&device->cdev->dev.kobj, KOBJ_CHANGE);

	if (device->state == device->target)
		wake_up(&dasd_init_waitq);
}
}


/*
/*
@@ -2157,6 +2157,7 @@ static int _dasd_sleep_on(struct dasd_ccw_req *maincqr, int interruptible)
		    test_bit(DASD_CQR_FLAGS_FAILFAST, &cqr->flags) &&
		    test_bit(DASD_CQR_FLAGS_FAILFAST, &cqr->flags) &&
		    (!dasd_eer_enabled(device))) {
		    (!dasd_eer_enabled(device))) {
			cqr->status = DASD_CQR_FAILED;
			cqr->status = DASD_CQR_FAILED;
			cqr->intrc = -EAGAIN;
			continue;
			continue;
		}
		}
		/* Don't try to start requests if device is stopped */
		/* Don't try to start requests if device is stopped */
@@ -3270,6 +3271,16 @@ void dasd_generic_path_event(struct ccw_device *cdev, int *path_event)
			dasd_schedule_device_bh(device);
			dasd_schedule_device_bh(device);
		}
		}
		if (path_event[chp] & PE_PATHGROUP_ESTABLISHED) {
		if (path_event[chp] & PE_PATHGROUP_ESTABLISHED) {
			if (!(device->path_data.opm & eventlpm) &&
			    !(device->path_data.tbvpm & eventlpm)) {
				/*
				 * we can not establish a pathgroup on an
				 * unavailable path, so trigger a path
				 * verification first
				 */
				device->path_data.tbvpm |= eventlpm;
				dasd_schedule_device_bh(device);
			}
			DBF_DEV_EVENT(DBF_WARNING, device, "%s",
			DBF_DEV_EVENT(DBF_WARNING, device, "%s",
				      "Pathgroup re-established\n");
				      "Pathgroup re-established\n");
			if (device->discipline->kick_validate)
			if (device->discipline->kick_validate)
Loading