|  | /* | 
|  | * This file is subject to the terms and conditions of the GNU General Public | 
|  | * License.  See the file "COPYING" in the main directory of this archive | 
|  | * for more details. | 
|  | * | 
|  | * Copyright (C) 2004, 2005 MIPS Technologies, Inc.  All rights reserved. | 
|  | * Copyright (C) 2013 Imagination Technologies Ltd. | 
|  | */ | 
|  | #include <linux/kernel.h> | 
|  | #include <linux/device.h> | 
|  | #include <linux/fs.h> | 
|  | #include <linux/slab.h> | 
|  | #include <linux/export.h> | 
|  |  | 
|  | #include <asm/mipsregs.h> | 
|  | #include <asm/mipsmtregs.h> | 
|  | #include <asm/mips_mt.h> | 
|  | #include <asm/vpe.h> | 
|  |  | 
|  | static int major; | 
|  |  | 
|  | /* The number of TCs and VPEs physically available on the core */ | 
|  | static int hw_tcs, hw_vpes; | 
|  |  | 
|  | /* We are prepared so configure and start the VPE... */ | 
|  | int vpe_run(struct vpe *v) | 
|  | { | 
|  | unsigned long flags, val, dmt_flag; | 
|  | struct vpe_notifications *notifier; | 
|  | unsigned int vpeflags; | 
|  | struct tc *t; | 
|  |  | 
|  | /* check we are the Master VPE */ | 
|  | local_irq_save(flags); | 
|  | val = read_c0_vpeconf0(); | 
|  | if (!(val & VPECONF0_MVP)) { | 
|  | pr_warn("VPE loader: only Master VPE's are able to config MT\n"); | 
|  | local_irq_restore(flags); | 
|  |  | 
|  | return -1; | 
|  | } | 
|  |  | 
|  | dmt_flag = dmt(); | 
|  | vpeflags = dvpe(); | 
|  |  | 
|  | if (list_empty(&v->tc)) { | 
|  | evpe(vpeflags); | 
|  | emt(dmt_flag); | 
|  | local_irq_restore(flags); | 
|  |  | 
|  | pr_warn("VPE loader: No TC's associated with VPE %d\n", | 
|  | v->minor); | 
|  |  | 
|  | return -ENOEXEC; | 
|  | } | 
|  |  | 
|  | t = list_first_entry(&v->tc, struct tc, tc); | 
|  |  | 
|  | /* Put MVPE's into 'configuration state' */ | 
|  | set_c0_mvpcontrol(MVPCONTROL_VPC); | 
|  |  | 
|  | settc(t->index); | 
|  |  | 
|  | /* should check it is halted, and not activated */ | 
|  | if ((read_tc_c0_tcstatus() & TCSTATUS_A) || | 
|  | !(read_tc_c0_tchalt() & TCHALT_H)) { | 
|  | evpe(vpeflags); | 
|  | emt(dmt_flag); | 
|  | local_irq_restore(flags); | 
|  |  | 
|  | pr_warn("VPE loader: TC %d is already active!\n", | 
|  | t->index); | 
|  |  | 
|  | return -ENOEXEC; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Write the address we want it to start running from in the TCPC | 
|  | * register. | 
|  | */ | 
|  | write_tc_c0_tcrestart((unsigned long)v->__start); | 
|  | write_tc_c0_tccontext((unsigned long)0); | 
|  |  | 
|  | /* | 
|  | * Mark the TC as activated, not interrupt exempt and not dynamically | 
|  | * allocatable | 
|  | */ | 
|  | val = read_tc_c0_tcstatus(); | 
|  | val = (val & ~(TCSTATUS_DA | TCSTATUS_IXMT)) | TCSTATUS_A; | 
|  | write_tc_c0_tcstatus(val); | 
|  |  | 
|  | write_tc_c0_tchalt(read_tc_c0_tchalt() & ~TCHALT_H); | 
|  |  | 
|  | /* | 
|  | * The sde-kit passes 'memsize' to __start in $a3, so set something | 
|  | * here...  Or set $a3 to zero and define DFLT_STACK_SIZE and | 
|  | * DFLT_HEAP_SIZE when you compile your program | 
|  | */ | 
|  | mttgpr(6, v->ntcs); | 
|  | mttgpr(7, physical_memsize); | 
|  |  | 
|  | /* set up VPE1 */ | 
|  | /* | 
|  | * bind the TC to VPE 1 as late as possible so we only have the final | 
|  | * VPE registers to set up, and so an EJTAG probe can trigger on it | 
|  | */ | 
|  | write_tc_c0_tcbind((read_tc_c0_tcbind() & ~TCBIND_CURVPE) | 1); | 
|  |  | 
|  | write_vpe_c0_vpeconf0(read_vpe_c0_vpeconf0() & ~(VPECONF0_VPA)); | 
|  |  | 
|  | back_to_back_c0_hazard(); | 
|  |  | 
|  | /* Set up the XTC bit in vpeconf0 to point at our tc */ | 
|  | write_vpe_c0_vpeconf0((read_vpe_c0_vpeconf0() & ~(VPECONF0_XTC)) | 
|  | | (t->index << VPECONF0_XTC_SHIFT)); | 
|  |  | 
|  | back_to_back_c0_hazard(); | 
|  |  | 
|  | /* enable this VPE */ | 
|  | write_vpe_c0_vpeconf0(read_vpe_c0_vpeconf0() | VPECONF0_VPA); | 
|  |  | 
|  | /* clear out any left overs from a previous program */ | 
|  | write_vpe_c0_status(0); | 
|  | write_vpe_c0_cause(0); | 
|  |  | 
|  | /* take system out of configuration state */ | 
|  | clear_c0_mvpcontrol(MVPCONTROL_VPC); | 
|  |  | 
|  | /* | 
|  | * SMVP kernels manage VPE enable independently, but uniprocessor | 
|  | * kernels need to turn it on, even if that wasn't the pre-dvpe() state. | 
|  | */ | 
|  | #ifdef CONFIG_SMP | 
|  | evpe(vpeflags); | 
|  | #else | 
|  | evpe(EVPE_ENABLE); | 
|  | #endif | 
|  | emt(dmt_flag); | 
|  | local_irq_restore(flags); | 
|  |  | 
|  | list_for_each_entry(notifier, &v->notify, list) | 
|  | notifier->start(VPE_MODULE_MINOR); | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | void cleanup_tc(struct tc *tc) | 
|  | { | 
|  | unsigned long flags; | 
|  | unsigned int mtflags, vpflags; | 
|  | int tmp; | 
|  |  | 
|  | local_irq_save(flags); | 
|  | mtflags = dmt(); | 
|  | vpflags = dvpe(); | 
|  | /* Put MVPE's into 'configuration state' */ | 
|  | set_c0_mvpcontrol(MVPCONTROL_VPC); | 
|  |  | 
|  | settc(tc->index); | 
|  | tmp = read_tc_c0_tcstatus(); | 
|  |  | 
|  | /* mark not allocated and not dynamically allocatable */ | 
|  | tmp &= ~(TCSTATUS_A | TCSTATUS_DA); | 
|  | tmp |= TCSTATUS_IXMT;	/* interrupt exempt */ | 
|  | write_tc_c0_tcstatus(tmp); | 
|  |  | 
|  | write_tc_c0_tchalt(TCHALT_H); | 
|  | mips_ihb(); | 
|  |  | 
|  | clear_c0_mvpcontrol(MVPCONTROL_VPC); | 
|  | evpe(vpflags); | 
|  | emt(mtflags); | 
|  | local_irq_restore(flags); | 
|  | } | 
|  |  | 
|  | /* module wrapper entry points */ | 
|  | /* give me a vpe */ | 
|  | void *vpe_alloc(void) | 
|  | { | 
|  | int i; | 
|  | struct vpe *v; | 
|  |  | 
|  | /* find a vpe */ | 
|  | for (i = 1; i < MAX_VPES; i++) { | 
|  | v = get_vpe(i); | 
|  | if (v != NULL) { | 
|  | v->state = VPE_STATE_INUSE; | 
|  | return v; | 
|  | } | 
|  | } | 
|  | return NULL; | 
|  | } | 
|  | EXPORT_SYMBOL(vpe_alloc); | 
|  |  | 
|  | /* start running from here */ | 
|  | int vpe_start(void *vpe, unsigned long start) | 
|  | { | 
|  | struct vpe *v = vpe; | 
|  |  | 
|  | v->__start = start; | 
|  | return vpe_run(v); | 
|  | } | 
|  | EXPORT_SYMBOL(vpe_start); | 
|  |  | 
|  | /* halt it for now */ | 
|  | int vpe_stop(void *vpe) | 
|  | { | 
|  | struct vpe *v = vpe; | 
|  | struct tc *t; | 
|  | unsigned int evpe_flags; | 
|  |  | 
|  | evpe_flags = dvpe(); | 
|  |  | 
|  | t = list_entry(v->tc.next, struct tc, tc); | 
|  | if (t != NULL) { | 
|  | settc(t->index); | 
|  | write_vpe_c0_vpeconf0(read_vpe_c0_vpeconf0() & ~VPECONF0_VPA); | 
|  | } | 
|  |  | 
|  | evpe(evpe_flags); | 
|  |  | 
|  | return 0; | 
|  | } | 
|  | EXPORT_SYMBOL(vpe_stop); | 
|  |  | 
|  | /* I've done with it thank you */ | 
|  | int vpe_free(void *vpe) | 
|  | { | 
|  | struct vpe *v = vpe; | 
|  | struct tc *t; | 
|  | unsigned int evpe_flags; | 
|  |  | 
|  | t = list_entry(v->tc.next, struct tc, tc); | 
|  | if (t == NULL) | 
|  | return -ENOEXEC; | 
|  |  | 
|  | evpe_flags = dvpe(); | 
|  |  | 
|  | /* Put MVPE's into 'configuration state' */ | 
|  | set_c0_mvpcontrol(MVPCONTROL_VPC); | 
|  |  | 
|  | settc(t->index); | 
|  | write_vpe_c0_vpeconf0(read_vpe_c0_vpeconf0() & ~VPECONF0_VPA); | 
|  |  | 
|  | /* halt the TC */ | 
|  | write_tc_c0_tchalt(TCHALT_H); | 
|  | mips_ihb(); | 
|  |  | 
|  | /* mark the TC unallocated */ | 
|  | write_tc_c0_tcstatus(read_tc_c0_tcstatus() & ~TCSTATUS_A); | 
|  |  | 
|  | v->state = VPE_STATE_UNUSED; | 
|  |  | 
|  | clear_c0_mvpcontrol(MVPCONTROL_VPC); | 
|  | evpe(evpe_flags); | 
|  |  | 
|  | return 0; | 
|  | } | 
|  | EXPORT_SYMBOL(vpe_free); | 
|  |  | 
|  | static ssize_t store_kill(struct device *dev, struct device_attribute *attr, | 
|  | const char *buf, size_t len) | 
|  | { | 
|  | struct vpe *vpe = get_vpe(aprp_cpu_index()); | 
|  | struct vpe_notifications *notifier; | 
|  |  | 
|  | list_for_each_entry(notifier, &vpe->notify, list) | 
|  | notifier->stop(aprp_cpu_index()); | 
|  |  | 
|  | release_progmem(vpe->load_addr); | 
|  | cleanup_tc(get_tc(aprp_cpu_index())); | 
|  | vpe_stop(vpe); | 
|  | vpe_free(vpe); | 
|  |  | 
|  | return len; | 
|  | } | 
|  | static DEVICE_ATTR(kill, S_IWUSR, NULL, store_kill); | 
|  |  | 
|  | static ssize_t ntcs_show(struct device *cd, struct device_attribute *attr, | 
|  | char *buf) | 
|  | { | 
|  | struct vpe *vpe = get_vpe(aprp_cpu_index()); | 
|  |  | 
|  | return sprintf(buf, "%d\n", vpe->ntcs); | 
|  | } | 
|  |  | 
|  | static ssize_t ntcs_store(struct device *dev, struct device_attribute *attr, | 
|  | const char *buf, size_t len) | 
|  | { | 
|  | struct vpe *vpe = get_vpe(aprp_cpu_index()); | 
|  | unsigned long new; | 
|  | int ret; | 
|  |  | 
|  | ret = kstrtoul(buf, 0, &new); | 
|  | if (ret < 0) | 
|  | return ret; | 
|  |  | 
|  | if (new == 0 || new > (hw_tcs - aprp_cpu_index())) | 
|  | return -EINVAL; | 
|  |  | 
|  | vpe->ntcs = new; | 
|  |  | 
|  | return len; | 
|  | } | 
|  | static DEVICE_ATTR_RW(ntcs); | 
|  |  | 
|  | static struct attribute *vpe_attrs[] = { | 
|  | &dev_attr_kill.attr, | 
|  | &dev_attr_ntcs.attr, | 
|  | NULL, | 
|  | }; | 
|  | ATTRIBUTE_GROUPS(vpe); | 
|  |  | 
|  | static void vpe_device_release(struct device *cd) | 
|  | { | 
|  | kfree(cd); | 
|  | } | 
|  |  | 
|  | static struct class vpe_class = { | 
|  | .name = "vpe", | 
|  | .owner = THIS_MODULE, | 
|  | .dev_release = vpe_device_release, | 
|  | .dev_groups = vpe_groups, | 
|  | }; | 
|  |  | 
|  | static struct device vpe_device; | 
|  |  | 
|  | int __init vpe_module_init(void) | 
|  | { | 
|  | unsigned int mtflags, vpflags; | 
|  | unsigned long flags, val; | 
|  | struct vpe *v = NULL; | 
|  | struct tc *t; | 
|  | int tc, err; | 
|  |  | 
|  | if (!cpu_has_mipsmt) { | 
|  | pr_warn("VPE loader: not a MIPS MT capable processor\n"); | 
|  | return -ENODEV; | 
|  | } | 
|  |  | 
|  | if (vpelimit == 0) { | 
|  | pr_warn("No VPEs reserved for AP/SP, not initialize VPE loader\n" | 
|  | "Pass maxvpes=<n> argument as kernel argument\n"); | 
|  |  | 
|  | return -ENODEV; | 
|  | } | 
|  |  | 
|  | if (aprp_cpu_index() == 0) { | 
|  | pr_warn("No TCs reserved for AP/SP, not initialize VPE loader\n" | 
|  | "Pass maxtcs=<n> argument as kernel argument\n"); | 
|  |  | 
|  | return -ENODEV; | 
|  | } | 
|  |  | 
|  | major = register_chrdev(0, VPE_MODULE_NAME, &vpe_fops); | 
|  | if (major < 0) { | 
|  | pr_warn("VPE loader: unable to register character device\n"); | 
|  | return major; | 
|  | } | 
|  |  | 
|  | err = class_register(&vpe_class); | 
|  | if (err) { | 
|  | pr_err("vpe_class registration failed\n"); | 
|  | goto out_chrdev; | 
|  | } | 
|  |  | 
|  | device_initialize(&vpe_device); | 
|  | vpe_device.class	= &vpe_class, | 
|  | vpe_device.parent	= NULL, | 
|  | dev_set_name(&vpe_device, "vpe1"); | 
|  | vpe_device.devt = MKDEV(major, VPE_MODULE_MINOR); | 
|  | err = device_add(&vpe_device); | 
|  | if (err) { | 
|  | pr_err("Adding vpe_device failed\n"); | 
|  | goto out_class; | 
|  | } | 
|  |  | 
|  | local_irq_save(flags); | 
|  | mtflags = dmt(); | 
|  | vpflags = dvpe(); | 
|  |  | 
|  | /* Put MVPE's into 'configuration state' */ | 
|  | set_c0_mvpcontrol(MVPCONTROL_VPC); | 
|  |  | 
|  | val = read_c0_mvpconf0(); | 
|  | hw_tcs = (val & MVPCONF0_PTC) + 1; | 
|  | hw_vpes = ((val & MVPCONF0_PVPE) >> MVPCONF0_PVPE_SHIFT) + 1; | 
|  |  | 
|  | for (tc = aprp_cpu_index(); tc < hw_tcs; tc++) { | 
|  | /* | 
|  | * Must re-enable multithreading temporarily or in case we | 
|  | * reschedule send IPIs or similar we might hang. | 
|  | */ | 
|  | clear_c0_mvpcontrol(MVPCONTROL_VPC); | 
|  | evpe(vpflags); | 
|  | emt(mtflags); | 
|  | local_irq_restore(flags); | 
|  | t = alloc_tc(tc); | 
|  | if (!t) { | 
|  | err = -ENOMEM; | 
|  | goto out_dev; | 
|  | } | 
|  |  | 
|  | local_irq_save(flags); | 
|  | mtflags = dmt(); | 
|  | vpflags = dvpe(); | 
|  | set_c0_mvpcontrol(MVPCONTROL_VPC); | 
|  |  | 
|  | /* VPE's */ | 
|  | if (tc < hw_tcs) { | 
|  | settc(tc); | 
|  |  | 
|  | v = alloc_vpe(tc); | 
|  | if (v == NULL) { | 
|  | pr_warn("VPE: unable to allocate VPE\n"); | 
|  | goto out_reenable; | 
|  | } | 
|  |  | 
|  | v->ntcs = hw_tcs - aprp_cpu_index(); | 
|  |  | 
|  | /* add the tc to the list of this vpe's tc's. */ | 
|  | list_add(&t->tc, &v->tc); | 
|  |  | 
|  | /* deactivate all but vpe0 */ | 
|  | if (tc >= aprp_cpu_index()) { | 
|  | unsigned long tmp = read_vpe_c0_vpeconf0(); | 
|  |  | 
|  | tmp &= ~VPECONF0_VPA; | 
|  |  | 
|  | /* master VPE */ | 
|  | tmp |= VPECONF0_MVP; | 
|  | write_vpe_c0_vpeconf0(tmp); | 
|  | } | 
|  |  | 
|  | /* disable multi-threading with TC's */ | 
|  | write_vpe_c0_vpecontrol(read_vpe_c0_vpecontrol() & | 
|  | ~VPECONTROL_TE); | 
|  |  | 
|  | if (tc >= vpelimit) { | 
|  | /* | 
|  | * Set config to be the same as vpe0, | 
|  | * particularly kseg0 coherency alg | 
|  | */ | 
|  | write_vpe_c0_config(read_c0_config()); | 
|  | } | 
|  | } | 
|  |  | 
|  | /* TC's */ | 
|  | t->pvpe = v;	/* set the parent vpe */ | 
|  |  | 
|  | if (tc >= aprp_cpu_index()) { | 
|  | unsigned long tmp; | 
|  |  | 
|  | settc(tc); | 
|  |  | 
|  | /* | 
|  | * A TC that is bound to any other VPE gets bound to | 
|  | * VPE0, ideally I'd like to make it homeless but it | 
|  | * doesn't appear to let me bind a TC to a non-existent | 
|  | * VPE. Which is perfectly reasonable. | 
|  | * | 
|  | * The (un)bound state is visible to an EJTAG probe so | 
|  | * may notify GDB... | 
|  | */ | 
|  | tmp = read_tc_c0_tcbind(); | 
|  | if (tmp & TCBIND_CURVPE) { | 
|  | /* tc is bound >vpe0 */ | 
|  | write_tc_c0_tcbind(tmp & ~TCBIND_CURVPE); | 
|  |  | 
|  | t->pvpe = get_vpe(0);	/* set the parent vpe */ | 
|  | } | 
|  |  | 
|  | /* halt the TC */ | 
|  | write_tc_c0_tchalt(TCHALT_H); | 
|  | mips_ihb(); | 
|  |  | 
|  | tmp = read_tc_c0_tcstatus(); | 
|  |  | 
|  | /* mark not activated and not dynamically allocatable */ | 
|  | tmp &= ~(TCSTATUS_A | TCSTATUS_DA); | 
|  | tmp |= TCSTATUS_IXMT;	/* interrupt exempt */ | 
|  | write_tc_c0_tcstatus(tmp); | 
|  | } | 
|  | } | 
|  |  | 
|  | out_reenable: | 
|  | /* release config state */ | 
|  | clear_c0_mvpcontrol(MVPCONTROL_VPC); | 
|  |  | 
|  | evpe(vpflags); | 
|  | emt(mtflags); | 
|  | local_irq_restore(flags); | 
|  |  | 
|  | return 0; | 
|  |  | 
|  | out_dev: | 
|  | device_del(&vpe_device); | 
|  |  | 
|  | out_class: | 
|  | class_unregister(&vpe_class); | 
|  |  | 
|  | out_chrdev: | 
|  | unregister_chrdev(major, VPE_MODULE_NAME); | 
|  |  | 
|  | return err; | 
|  | } | 
|  |  | 
|  | void __exit vpe_module_exit(void) | 
|  | { | 
|  | struct vpe *v, *n; | 
|  |  | 
|  | device_del(&vpe_device); | 
|  | class_unregister(&vpe_class); | 
|  | unregister_chrdev(major, VPE_MODULE_NAME); | 
|  |  | 
|  | /* No locking needed here */ | 
|  | list_for_each_entry_safe(v, n, &vpecontrol.vpe_list, list) { | 
|  | if (v->state != VPE_STATE_UNUSED) | 
|  | release_vpe(v); | 
|  | } | 
|  | } |