Donate to e Foundation | Murena handsets with /e/OS | Own a part of Murena! Learn more

Commit 0afacde3 authored by Arnd Bergmann's avatar Arnd Bergmann Committed by Paul Mackerras
Browse files

[POWERPC] spufs: allow isolated mode apps by starting the SPE loader



This patch adds general support for isolated mode SPE apps.

Isolated apps are started indirectly, by a dedicated loader "kernel".
This patch starts the loader when spe_create is invoked with the
ISOLATE flag. We do this at spe_create time to allow libspe to pass the
isolated app in before calling spe_run.

The loader is read from the device tree, at the location
"/spu-isolation/loader". If the loader is not present, an attempt to
start an isolated SPE binary will fail with -ENODEV.

Update: loader needs to be correctly aligned - copy to a kmalloced buf.
Update: remove workaround for systemsim/spurom 'L-bit' bug, which has
        been fixed.
Update: don't write to runcntl on spu_run_init: SPU is already running.
Update: do spu_setup_isolated earlier

Tested on systemsim.

Signed-off-by: default avatarJeremy Kerr <jk@ozlabs.org>
Signed-off-by: default avatarArnd Bergmann <arnd.bergmann@de.ibm.com>
Signed-off-by: default avatarPaul Mackerras <paulus@samba.org>
parent eb758ce5
Loading
Loading
Loading
Loading
+24 −11
Original line number Original line Diff line number Diff line
@@ -89,15 +89,10 @@ static int __spu_trap_data_seg(struct spu *spu, unsigned long ea)
		printk("%s: invalid access during switch!\n", __func__);
		printk("%s: invalid access during switch!\n", __func__);
		return 1;
		return 1;
	}
	}
	if (!mm || (REGION_ID(ea) != USER_REGION_ID)) {
		/* Future: support kernel segments so that drivers
		 * can use SPUs.
		 */
		pr_debug("invalid region access at %016lx\n", ea);
		return 1;
	}

	esid = (ea & ESID_MASK) | SLB_ESID_V;
	esid = (ea & ESID_MASK) | SLB_ESID_V;

	switch(REGION_ID(ea)) {
	case USER_REGION_ID:
#ifdef CONFIG_HUGETLB_PAGE
#ifdef CONFIG_HUGETLB_PAGE
		if (in_hugepage_area(mm->context, ea))
		if (in_hugepage_area(mm->context, ea))
			llp = mmu_psize_defs[mmu_huge_psize].sllp;
			llp = mmu_psize_defs[mmu_huge_psize].sllp;
@@ -106,6 +101,24 @@ static int __spu_trap_data_seg(struct spu *spu, unsigned long ea)
			llp = mmu_psize_defs[mmu_virtual_psize].sllp;
			llp = mmu_psize_defs[mmu_virtual_psize].sllp;
		vsid = (get_vsid(mm->context.id, ea) << SLB_VSID_SHIFT) |
		vsid = (get_vsid(mm->context.id, ea) << SLB_VSID_SHIFT) |
				SLB_VSID_USER | llp;
				SLB_VSID_USER | llp;
		break;
	case VMALLOC_REGION_ID:
		llp = mmu_psize_defs[mmu_virtual_psize].sllp;
		vsid = (get_kernel_vsid(ea) << SLB_VSID_SHIFT) |
			SLB_VSID_KERNEL | llp;
		break;
	case KERNEL_REGION_ID:
		llp = mmu_psize_defs[mmu_linear_psize].sllp;
		vsid = (get_kernel_vsid(ea) << SLB_VSID_SHIFT) |
			SLB_VSID_KERNEL | llp;
		break;
	default:
		/* Future: support kernel segments so that drivers
		 * can use SPUs.
		 */
		pr_debug("invalid region access at %016lx\n", ea);
		return 1;
	}


	out_be64(&priv2->slb_index_W, spu->slb_replace);
	out_be64(&priv2->slb_index_W, spu->slb_replace);
	out_be64(&priv2->slb_vsid_RW, vsid);
	out_be64(&priv2->slb_vsid_RW, vsid);
+117 −0
Original line number Original line Diff line number Diff line
@@ -33,6 +33,8 @@
#include <linux/slab.h>
#include <linux/slab.h>
#include <linux/parser.h>
#include <linux/parser.h>


#include <asm/prom.h>
#include <asm/spu_priv1.h>
#include <asm/io.h>
#include <asm/io.h>
#include <asm/semaphore.h>
#include <asm/semaphore.h>
#include <asm/spu.h>
#include <asm/spu.h>
@@ -41,6 +43,7 @@
#include "spufs.h"
#include "spufs.h"


static kmem_cache_t *spufs_inode_cache;
static kmem_cache_t *spufs_inode_cache;
static char *isolated_loader;


static struct inode *
static struct inode *
spufs_alloc_inode(struct super_block *sb)
spufs_alloc_inode(struct super_block *sb)
@@ -232,6 +235,89 @@ struct file_operations spufs_context_fops = {
	.fsync		= simple_sync_file,
	.fsync		= simple_sync_file,
};
};


static int spu_setup_isolated(struct spu_context *ctx)
{
	int ret;
	u64 __iomem *mfc_cntl;
	u64 sr1;
	u32 status;
	unsigned long timeout;
	const u32 status_loading = SPU_STATUS_RUNNING
		| SPU_STATUS_ISOLATED_STATE | SPU_STATUS_ISOLATED_LOAD_STATUS;

	if (!isolated_loader)
		return -ENODEV;

	if ((ret = spu_acquire_runnable(ctx)) != 0)
		return ret;

	mfc_cntl = &ctx->spu->priv2->mfc_control_RW;

	/* purge the MFC DMA queue to ensure no spurious accesses before we
	 * enter kernel mode */
	timeout = jiffies + HZ;
	out_be64(mfc_cntl, MFC_CNTL_PURGE_DMA_REQUEST);
	while ((in_be64(mfc_cntl) & MFC_CNTL_PURGE_DMA_STATUS_MASK)
			!= MFC_CNTL_PURGE_DMA_COMPLETE) {
		if (time_after(jiffies, timeout)) {
			printk(KERN_ERR "%s: timeout flushing MFC DMA queue\n",
					__FUNCTION__);
			ret = -EIO;
			goto out_unlock;
		}
		cond_resched();
	}

	/* put the SPE in kernel mode to allow access to the loader */
	sr1 = spu_mfc_sr1_get(ctx->spu);
	sr1 &= ~MFC_STATE1_PROBLEM_STATE_MASK;
	spu_mfc_sr1_set(ctx->spu, sr1);

	/* start the loader */
	ctx->ops->signal1_write(ctx, (unsigned long)isolated_loader >> 32);
	ctx->ops->signal2_write(ctx,
			(unsigned long)isolated_loader & 0xffffffff);

	ctx->ops->runcntl_write(ctx,
			SPU_RUNCNTL_RUNNABLE | SPU_RUNCNTL_ISOLATE);

	ret = 0;
	timeout = jiffies + HZ;
	while (((status = ctx->ops->status_read(ctx)) & status_loading) ==
				status_loading) {
		if (time_after(jiffies, timeout)) {
			printk(KERN_ERR "%s: timeout waiting for loader\n",
					__FUNCTION__);
			ret = -EIO;
			goto out_drop_priv;
		}
		cond_resched();
	}

	if (!(status & SPU_STATUS_RUNNING)) {
		/* If isolated LOAD has failed: run SPU, we will get a stop-and
		 * signal later. */
		pr_debug("%s: isolated LOAD failed\n", __FUNCTION__);
		ctx->ops->runcntl_write(ctx, SPU_RUNCNTL_RUNNABLE);
		ret = -EACCES;

	} else if (!(status & SPU_STATUS_ISOLATED_STATE)) {
		/* This isn't allowed by the CBEA, but check anyway */
		pr_debug("%s: SPU fell out of isolated mode?\n", __FUNCTION__);
		ctx->ops->runcntl_write(ctx, SPU_RUNCNTL_STOP);
		ret = -EINVAL;
	}

out_drop_priv:
	/* Finished accessing the loader. Drop kernel mode */
	sr1 |= MFC_STATE1_PROBLEM_STATE_MASK;
	spu_mfc_sr1_set(ctx->spu, sr1);

out_unlock:
	up_write(&ctx->state_sema);
	return ret;
}

static int
static int
spufs_mkdir(struct inode *dir, struct dentry *dentry, unsigned int flags,
spufs_mkdir(struct inode *dir, struct dentry *dentry, unsigned int flags,
		int mode)
		int mode)
@@ -255,6 +341,11 @@ spufs_mkdir(struct inode *dir, struct dentry *dentry, unsigned int flags,
		goto out_iput;
		goto out_iput;


	ctx->flags = flags;
	ctx->flags = flags;
	if (flags & SPU_CREATE_ISOLATE) {
		ret = spu_setup_isolated(ctx);
		if (ret)
			goto out_iput;
	}


	inode->i_op = &spufs_dir_inode_operations;
	inode->i_op = &spufs_dir_inode_operations;
	inode->i_fop = &simple_dir_operations;
	inode->i_fop = &simple_dir_operations;
@@ -555,6 +646,30 @@ spufs_parse_options(char *options, struct inode *root)
	return 1;
	return 1;
}
}


static void
spufs_init_isolated_loader(void)
{
	struct device_node *dn;
	const char *loader;
	int size;

	dn = of_find_node_by_path("/spu-isolation");
	if (!dn)
		return;

	loader = get_property(dn, "loader", &size);
	if (!loader)
		return;

	/* kmalloc should align on a 16 byte boundary..* */
	isolated_loader = kmalloc(size, GFP_KERNEL);
	if (!isolated_loader)
		return;

	memcpy(isolated_loader, loader, size);
	printk(KERN_INFO "spufs: SPU isolation mode enabled\n");
}

static int
static int
spufs_create_root(struct super_block *sb, void *data)
spufs_create_root(struct super_block *sb, void *data)
{
{
@@ -640,6 +755,8 @@ static int __init spufs_init(void)
	ret = register_spu_syscalls(&spufs_calls);
	ret = register_spu_syscalls(&spufs_calls);
	if (ret)
	if (ret)
		goto out_fs;
		goto out_fs;

	spufs_init_isolated_loader();
	return 0;
	return 0;
out_fs:
out_fs:
	unregister_filesystem(&spufs_type);
	unregister_filesystem(&spufs_type);
+7 −5
Original line number Original line Diff line number Diff line
#define DEBUG

#include <linux/wait.h>
#include <linux/wait.h>
#include <linux/ptrace.h>
#include <linux/ptrace.h>


@@ -56,12 +58,12 @@ static inline int spu_run_init(struct spu_context *ctx, u32 * npc)
	if ((ret = spu_acquire_runnable(ctx)) != 0)
	if ((ret = spu_acquire_runnable(ctx)) != 0)
		return ret;
		return ret;


	if (ctx->flags & SPU_CREATE_ISOLATE)
	/* if we're in isolated mode, we would have started the SPU
		runcntl |= SPU_RUNCNTL_ISOLATE;
	 * earlier, so don't do it again now. */
	else
	if (!(ctx->flags & SPU_CREATE_ISOLATE)) {
		ctx->ops->npc_write(ctx, *npc);
		ctx->ops->npc_write(ctx, *npc);

		ctx->ops->runcntl_write(ctx, runcntl);
		ctx->ops->runcntl_write(ctx, runcntl);
	}
	return 0;
	return 0;
}
}