mirror of
git://projects.qi-hardware.com/openwrt-xburst.git
synced 2024-11-01 23:42:49 +02:00
4f72aeefd8
git-svn-id: svn://svn.openwrt.org/openwrt/trunk@31336 3c298f89-4303-0410-b956-a3cf2f4a3e73
1199 lines
32 KiB
Diff
1199 lines
32 KiB
Diff
From 948bb4dd94209332253d2b69c28f44e2bb11f518 Mon Sep 17 00:00:00 2001
|
|
From: John Crispin <blogic@openwrt.org>
|
|
Date: Thu, 29 Sep 2011 20:30:40 +0200
|
|
Subject: [PATCH 53/70] MIPS: lantiq: VPE extensions
|
|
|
|
---
|
|
arch/mips/Kconfig | 22 +++
|
|
arch/mips/include/asm/mipsmtregs.h | 54 +++++++
|
|
arch/mips/kernel/Makefile | 3 +-
|
|
arch/mips/kernel/mips-mt.c | 97 +++++++++++--
|
|
arch/mips/kernel/mtsched_proc.c | 279 ++++++++++++++++++++++++++++++++++++
|
|
arch/mips/kernel/perf_proc.c | 191 ++++++++++++++++++++++++
|
|
arch/mips/kernel/proc.c | 17 +++
|
|
arch/mips/kernel/smtc.c | 7 +
|
|
arch/mips/kernel/vpe.c | 250 ++++++++++++++++++++++++++++++++-
|
|
9 files changed, 905 insertions(+), 15 deletions(-)
|
|
create mode 100644 arch/mips/kernel/mtsched_proc.c
|
|
create mode 100644 arch/mips/kernel/perf_proc.c
|
|
|
|
--- a/arch/mips/Kconfig
|
|
+++ b/arch/mips/Kconfig
|
|
@@ -1947,6 +1947,28 @@ config MIPS_VPE_LOADER
|
|
Includes a loader for loading an elf relocatable object
|
|
onto another VPE and running it.
|
|
|
|
+config IFX_VPE_EXT
|
|
+ bool "IFX APRP Extensions"
|
|
+ depends on MIPS_VPE_LOADER
|
|
+ default y
|
|
+ help
|
|
+ IFX included extensions in APRP
|
|
+
|
|
+config PERFCTRS
|
|
+ bool "34K Performance counters"
|
|
+ depends on MIPS_MT && PROC_FS
|
|
+ default n
|
|
+ help
|
|
+ 34K Performance counter through /proc
|
|
+
|
|
+config MTSCHED
|
|
+ bool "Support mtsched priority configuration for TCs"
|
|
+ depends on MIPS_MT && PROC_FS
|
|
+ default y
|
|
+ help
|
|
+ Support for mtsched priority configuration for TCs through
|
|
+ /proc/mips/mtsched
|
|
+
|
|
config MIPS_MT_SMTC_IM_BACKSTOP
|
|
bool "Use per-TC register bits as backstop for inhibited IM bits"
|
|
depends on MIPS_MT_SMTC
|
|
--- a/arch/mips/include/asm/mipsmtregs.h
|
|
+++ b/arch/mips/include/asm/mipsmtregs.h
|
|
@@ -28,14 +28,34 @@
|
|
#define read_c0_vpeconf0() __read_32bit_c0_register($1, 2)
|
|
#define write_c0_vpeconf0(val) __write_32bit_c0_register($1, 2, val)
|
|
|
|
+#define read_c0_vpeconf1() __read_32bit_c0_register($1, 3)
|
|
+#define write_c0_vpeconf1(val) __write_32bit_c0_register($1, 3, val)
|
|
+
|
|
+#define read_c0_vpeschedule() __read_32bit_c0_register($1, 5)
|
|
+#define write_c0_vpeschedule(val) __write_32bit_c0_register($1, 5, val)
|
|
+
|
|
+#define read_c0_vpeschefback() __read_32bit_c0_register($1, 6)
|
|
+#define write_c0_vpeschefback(val) __write_32bit_c0_register($1, 6, val)
|
|
+
|
|
+#define read_c0_vpeopt() __read_32bit_c0_register($1, 7)
|
|
+#define write_c0_vpeopt(val) __write_32bit_c0_register($1, 7, val)
|
|
+
|
|
#define read_c0_tcstatus() __read_32bit_c0_register($2, 1)
|
|
#define write_c0_tcstatus(val) __write_32bit_c0_register($2, 1, val)
|
|
|
|
#define read_c0_tcbind() __read_32bit_c0_register($2, 2)
|
|
+#define write_c0_tcbind(val) __write_32bit_c0_register($2, 2, val)
|
|
|
|
#define read_c0_tccontext() __read_32bit_c0_register($2, 5)
|
|
#define write_c0_tccontext(val) __write_32bit_c0_register($2, 5, val)
|
|
|
|
+#define read_c0_tcschedule() __read_32bit_c0_register($2, 6)
|
|
+#define write_c0_tcschedule(val) __write_32bit_c0_register($2, 6, val)
|
|
+
|
|
+#define read_c0_tcschefback() __read_32bit_c0_register($2, 7)
|
|
+#define write_c0_tcschefback(val) __write_32bit_c0_register($2, 7, val)
|
|
+
|
|
+
|
|
#else /* Assembly */
|
|
/*
|
|
* Macros for use in assembly language code
|
|
@@ -74,6 +94,8 @@
|
|
#define MVPCONTROL_STLB_SHIFT 2
|
|
#define MVPCONTROL_STLB (_ULCAST_(1) << MVPCONTROL_STLB_SHIFT)
|
|
|
|
+#define MVPCONTROL_CPA_SHIFT 3
|
|
+#define MVPCONTROL_CPA (_ULCAST_(1) << MVPCONTROL_CPA_SHIFT)
|
|
|
|
/* MVPConf0 fields */
|
|
#define MVPCONF0_PTC_SHIFT 0
|
|
@@ -84,6 +106,8 @@
|
|
#define MVPCONF0_TCA ( _ULCAST_(1) << MVPCONF0_TCA_SHIFT)
|
|
#define MVPCONF0_PTLBE_SHIFT 16
|
|
#define MVPCONF0_PTLBE (_ULCAST_(0x3ff) << MVPCONF0_PTLBE_SHIFT)
|
|
+#define MVPCONF0_PCP_SHIFT 27
|
|
+#define MVPCONF0_PCP (_ULCAST_(1) << MVPCONF0_PCP_SHIFT)
|
|
#define MVPCONF0_TLBS_SHIFT 29
|
|
#define MVPCONF0_TLBS (_ULCAST_(1) << MVPCONF0_TLBS_SHIFT)
|
|
#define MVPCONF0_M_SHIFT 31
|
|
@@ -121,9 +145,25 @@
|
|
#define VPECONF0_VPA (_ULCAST_(1) << VPECONF0_VPA_SHIFT)
|
|
#define VPECONF0_MVP_SHIFT 1
|
|
#define VPECONF0_MVP (_ULCAST_(1) << VPECONF0_MVP_SHIFT)
|
|
+#define VPECONF0_ICS_SHIFT 16
|
|
+#define VPECONF0_ICS (_ULCAST_(1) << VPECONF0_ICS_SHIFT)
|
|
+#define VPECONF0_DCS_SHIFT 17
|
|
+#define VPECONF0_DCS (_ULCAST_(1) << VPECONF0_DCS_SHIFT)
|
|
#define VPECONF0_XTC_SHIFT 21
|
|
#define VPECONF0_XTC (_ULCAST_(0xff) << VPECONF0_XTC_SHIFT)
|
|
|
|
+/* VPEOpt fields */
|
|
+#define VPEOPT_DWX_SHIFT 0
|
|
+#define VPEOPT_IWX_SHIFT 8
|
|
+#define VPEOPT_IWX0 ( _ULCAST_(0x1) << VPEOPT_IWX_SHIFT)
|
|
+#define VPEOPT_IWX1 ( _ULCAST_(0x2) << VPEOPT_IWX_SHIFT)
|
|
+#define VPEOPT_IWX2 ( _ULCAST_(0x4) << VPEOPT_IWX_SHIFT)
|
|
+#define VPEOPT_IWX3 ( _ULCAST_(0x8) << VPEOPT_IWX_SHIFT)
|
|
+#define VPEOPT_DWX0 ( _ULCAST_(0x1) << VPEOPT_DWX_SHIFT)
|
|
+#define VPEOPT_DWX1 ( _ULCAST_(0x2) << VPEOPT_DWX_SHIFT)
|
|
+#define VPEOPT_DWX2 ( _ULCAST_(0x4) << VPEOPT_DWX_SHIFT)
|
|
+#define VPEOPT_DWX3 ( _ULCAST_(0x8) << VPEOPT_DWX_SHIFT)
|
|
+
|
|
/* TCStatus fields (per TC) */
|
|
#define TCSTATUS_TASID (_ULCAST_(0xff))
|
|
#define TCSTATUS_IXMT_SHIFT 10
|
|
@@ -350,6 +390,14 @@ do { \
|
|
#define write_vpe_c0_vpecontrol(val) mttc0(1, 1, val)
|
|
#define read_vpe_c0_vpeconf0() mftc0(1, 2)
|
|
#define write_vpe_c0_vpeconf0(val) mttc0(1, 2, val)
|
|
+#define read_vpe_c0_vpeschedule() mftc0(1, 5)
|
|
+#define write_vpe_c0_vpeschedule(val) mttc0(1, 5, val)
|
|
+#define read_vpe_c0_vpeschefback() mftc0(1, 6)
|
|
+#define write_vpe_c0_vpeschefback(val) mttc0(1, 6, val)
|
|
+#define read_vpe_c0_vpeopt() mftc0(1, 7)
|
|
+#define write_vpe_c0_vpeopt(val) mttc0(1, 7, val)
|
|
+#define read_vpe_c0_wired() mftc0(6, 0)
|
|
+#define write_vpe_c0_wired(val) mttc0(6, 0, val)
|
|
#define read_vpe_c0_count() mftc0(9, 0)
|
|
#define write_vpe_c0_count(val) mttc0(9, 0, val)
|
|
#define read_vpe_c0_status() mftc0(12, 0)
|
|
@@ -381,6 +429,12 @@ do { \
|
|
#define write_tc_c0_tchalt(val) mttc0(2, 4, val)
|
|
#define read_tc_c0_tccontext() mftc0(2, 5)
|
|
#define write_tc_c0_tccontext(val) mttc0(2, 5, val)
|
|
+#define read_tc_c0_tcschedule() mftc0(2, 6)
|
|
+#define write_tc_c0_tcschedule(val) mttc0(2, 6, val)
|
|
+#define read_tc_c0_tcschefback() mftc0(2, 7)
|
|
+#define write_tc_c0_tcschefback(val) mttc0(2, 7, val)
|
|
+#define read_tc_c0_entryhi() mftc0(10, 0)
|
|
+#define write_tc_c0_entryhi(val) mttc0(10, 0, val)
|
|
|
|
/* GPR */
|
|
#define read_tc_gpr_sp() mftgpr(29)
|
|
--- a/arch/mips/kernel/Makefile
|
|
+++ b/arch/mips/kernel/Makefile
|
|
@@ -90,7 +90,8 @@ obj-$(CONFIG_MIPS32_O32) += binfmt_elfo3
|
|
|
|
obj-$(CONFIG_KGDB) += kgdb.o
|
|
obj-$(CONFIG_PROC_FS) += proc.o
|
|
-
|
|
+obj-$(CONFIG_MTSCHED) += mtsched_proc.o
|
|
+obj-$(CONFIG_PERFCTRS) += perf_proc.o
|
|
obj-$(CONFIG_64BIT) += cpu-bugs64.o
|
|
|
|
obj-$(CONFIG_I8253) += i8253.o
|
|
--- a/arch/mips/kernel/mips-mt.c
|
|
+++ b/arch/mips/kernel/mips-mt.c
|
|
@@ -21,26 +21,96 @@
|
|
#include <asm/cacheflush.h>
|
|
|
|
int vpelimit;
|
|
-
|
|
static int __init maxvpes(char *str)
|
|
{
|
|
get_option(&str, &vpelimit);
|
|
-
|
|
return 1;
|
|
}
|
|
-
|
|
__setup("maxvpes=", maxvpes);
|
|
|
|
int tclimit;
|
|
-
|
|
static int __init maxtcs(char *str)
|
|
{
|
|
get_option(&str, &tclimit);
|
|
+ return 1;
|
|
+}
|
|
+__setup("maxtcs=", maxtcs);
|
|
|
|
+#ifdef CONFIG_IFX_VPE_EXT
|
|
+int stlb;
|
|
+static int __init istlbshared(char *str)
|
|
+{
|
|
+ get_option(&str, &stlb);
|
|
return 1;
|
|
}
|
|
+__setup("vpe_tlb_shared=", istlbshared);
|
|
|
|
-__setup("maxtcs=", maxtcs);
|
|
+int vpe0_wired;
|
|
+static int __init vpe0wired(char *str)
|
|
+{
|
|
+ get_option(&str, &vpe0_wired);
|
|
+ return 1;
|
|
+}
|
|
+__setup("vpe0_wired_tlb_entries=", vpe0wired);
|
|
+
|
|
+int vpe1_wired;
|
|
+static int __init vpe1wired(char *str)
|
|
+{
|
|
+ get_option(&str, &vpe1_wired);
|
|
+ return 1;
|
|
+}
|
|
+__setup("vpe1_wired_tlb_entries=", vpe1wired);
|
|
+
|
|
+#ifdef CONFIG_MIPS_MT_SMTC
|
|
+extern int nostlb;
|
|
+#endif
|
|
+void configure_tlb(void)
|
|
+{
|
|
+ int vpeflags, tcflags, tlbsiz;
|
|
+ unsigned int config1val;
|
|
+ vpeflags = dvpe();
|
|
+ tcflags = dmt();
|
|
+ write_c0_vpeconf0((read_c0_vpeconf0() | VPECONF0_MVP));
|
|
+ write_c0_mvpcontrol((read_c0_mvpcontrol() | MVPCONTROL_VPC));
|
|
+ mips_ihb();
|
|
+ //printk("stlb = %d, vpe0_wired = %d vpe1_wired=%d\n", stlb,vpe0_wired, vpe1_wired);
|
|
+ if (stlb) {
|
|
+ if (!(read_c0_mvpconf0() & MVPCONF0_TLBS)) {
|
|
+ emt(tcflags);
|
|
+ evpe(vpeflags);
|
|
+ return;
|
|
+ }
|
|
+
|
|
+ write_c0_mvpcontrol(read_c0_mvpcontrol() | MVPCONTROL_STLB);
|
|
+ write_c0_wired(vpe0_wired + vpe1_wired);
|
|
+ if (((read_vpe_c0_config() & MIPS_CONF_MT) >> 7) == 1) {
|
|
+ config1val = read_vpe_c0_config1();
|
|
+ tlbsiz = (((config1val >> 25) & 0x3f) + 1);
|
|
+ if (tlbsiz > 64)
|
|
+ tlbsiz = 64;
|
|
+ cpu_data[0].tlbsize = tlbsiz;
|
|
+ current_cpu_data.tlbsize = tlbsiz;
|
|
+ }
|
|
+
|
|
+ }
|
|
+ else {
|
|
+ write_c0_mvpcontrol(read_c0_mvpcontrol() & ~MVPCONTROL_STLB);
|
|
+ write_c0_wired(vpe0_wired);
|
|
+ }
|
|
+
|
|
+ ehb();
|
|
+ write_c0_mvpcontrol((read_c0_mvpcontrol() & ~MVPCONTROL_VPC));
|
|
+ ehb();
|
|
+ local_flush_tlb_all();
|
|
+
|
|
+ printk("Wired TLB entries for Linux read_c0_wired() = %d\n", read_c0_wired());
|
|
+#ifdef CONFIG_MIPS_MT_SMTC
|
|
+ nostlb = !stlb;
|
|
+#endif
|
|
+ emt(tcflags);
|
|
+ evpe(vpeflags);
|
|
+}
|
|
+#endif
|
|
|
|
/*
|
|
* Dump new MIPS MT state for the core. Does not leave TCs halted.
|
|
@@ -78,18 +148,18 @@ void mips_mt_regdump(unsigned long mvpct
|
|
if ((read_tc_c0_tcbind() & TCBIND_CURVPE) == i) {
|
|
printk(" VPE %d\n", i);
|
|
printk(" VPEControl : %08lx\n",
|
|
- read_vpe_c0_vpecontrol());
|
|
+ read_vpe_c0_vpecontrol());
|
|
printk(" VPEConf0 : %08lx\n",
|
|
- read_vpe_c0_vpeconf0());
|
|
+ read_vpe_c0_vpeconf0());
|
|
printk(" VPE%d.Status : %08lx\n",
|
|
- i, read_vpe_c0_status());
|
|
+ i, read_vpe_c0_status());
|
|
printk(" VPE%d.EPC : %08lx %pS\n",
|
|
- i, read_vpe_c0_epc(),
|
|
- (void *) read_vpe_c0_epc());
|
|
+ i, read_vpe_c0_epc(),
|
|
+ (void *) read_vpe_c0_epc());
|
|
printk(" VPE%d.Cause : %08lx\n",
|
|
- i, read_vpe_c0_cause());
|
|
+ i, read_vpe_c0_cause());
|
|
printk(" VPE%d.Config7 : %08lx\n",
|
|
- i, read_vpe_c0_config7());
|
|
+ i, read_vpe_c0_config7());
|
|
break; /* Next VPE */
|
|
}
|
|
}
|
|
@@ -287,6 +357,9 @@ void mips_mt_set_cpuoptions(void)
|
|
printk("Mapped %ld ITC cells starting at 0x%08x\n",
|
|
((itcblkgrn & 0x7fe00000) >> 20), itc_base);
|
|
}
|
|
+#ifdef CONFIG_IFX_VPE_EXT
|
|
+ configure_tlb();
|
|
+#endif
|
|
}
|
|
|
|
/*
|
|
--- /dev/null
|
|
+++ b/arch/mips/kernel/mtsched_proc.c
|
|
@@ -0,0 +1,279 @@
|
|
+/*
|
|
+ * /proc hooks for MIPS MT scheduling policy management for 34K cores
|
|
+ *
|
|
+ * This program is free software; you can distribute it and/or modify it
|
|
+ * under the terms of the GNU General Public License (Version 2) as
|
|
+ * published by the Free Software Foundation.
|
|
+ *
|
|
+ * This program is distributed in the hope it will be useful, but WITHOUT
|
|
+ * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
|
|
+ * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
|
|
+ * for more details.
|
|
+ *
|
|
+ * You should have received a copy of the GNU General Public License along
|
|
+ * with this program; if not, write to the Free Software Foundation, Inc.,
|
|
+ * 59 Temple Place - Suite 330, Boston MA 02111-1307, USA.
|
|
+ *
|
|
+ * Copyright (C) 2006 Mips Technologies, Inc
|
|
+ */
|
|
+
|
|
+#include <linux/kernel.h>
|
|
+
|
|
+#include <asm/cpu.h>
|
|
+#include <asm/processor.h>
|
|
+#include <asm/system.h>
|
|
+#include <asm/mipsregs.h>
|
|
+#include <asm/mipsmtregs.h>
|
|
+#include <asm/uaccess.h>
|
|
+#include <linux/proc_fs.h>
|
|
+
|
|
+static struct proc_dir_entry *mtsched_proc;
|
|
+
|
|
+#ifndef CONFIG_MIPS_MT_SMTC
|
|
+#define NTCS 2
|
|
+#else
|
|
+#define NTCS NR_CPUS
|
|
+#endif
|
|
+#define NVPES 2
|
|
+
|
|
+int lastvpe = 1;
|
|
+int lasttc = 8;
|
|
+
|
|
+static int proc_read_mtsched(char *page, char **start, off_t off,
|
|
+ int count, int *eof, void *data)
|
|
+{
|
|
+ int totalen = 0;
|
|
+ int len;
|
|
+
|
|
+ int i;
|
|
+ int vpe;
|
|
+ int mytc;
|
|
+ unsigned long flags;
|
|
+ unsigned int mtflags;
|
|
+ unsigned int haltstate;
|
|
+ unsigned int vpes_checked[NVPES];
|
|
+ unsigned int vpeschedule[NVPES];
|
|
+ unsigned int vpeschefback[NVPES];
|
|
+ unsigned int tcschedule[NTCS];
|
|
+ unsigned int tcschefback[NTCS];
|
|
+
|
|
+ /* Dump the state of the MIPS MT scheduling policy manager */
|
|
+ /* Inititalize control state */
|
|
+ for(i = 0; i < NVPES; i++) {
|
|
+ vpes_checked[i] = 0;
|
|
+ vpeschedule[i] = 0;
|
|
+ vpeschefback[i] = 0;
|
|
+ }
|
|
+ for(i = 0; i < NTCS; i++) {
|
|
+ tcschedule[i] = 0;
|
|
+ tcschefback[i] = 0;
|
|
+ }
|
|
+
|
|
+ /* Disable interrupts and multithreaded issue */
|
|
+ local_irq_save(flags);
|
|
+ mtflags = dvpe();
|
|
+
|
|
+ /* Then go through the TCs, halt 'em, and extract the values */
|
|
+ mytc = (read_c0_tcbind() & TCBIND_CURTC) >> TCBIND_CURTC_SHIFT;
|
|
+ for(i = 0; i < NTCS; i++) {
|
|
+ if(i == mytc) {
|
|
+ /* No need to halt ourselves! */
|
|
+ tcschedule[i] = read_c0_tcschedule();
|
|
+ tcschefback[i] = read_c0_tcschefback();
|
|
+ /* If VPE bound to TC hasn't been checked, do it */
|
|
+ vpe = read_c0_tcbind() & TCBIND_CURVPE;
|
|
+ if(!vpes_checked[vpe]) {
|
|
+ vpeschedule[vpe] = read_c0_vpeschedule();
|
|
+ vpeschefback[vpe] = read_c0_vpeschefback();
|
|
+ vpes_checked[vpe] = 1;
|
|
+ }
|
|
+ } else {
|
|
+ settc(i);
|
|
+ haltstate = read_tc_c0_tchalt();
|
|
+ write_tc_c0_tchalt(TCHALT_H);
|
|
+ mips_ihb();
|
|
+ tcschedule[i] = read_tc_c0_tcschedule();
|
|
+ tcschefback[i] = read_tc_c0_tcschefback();
|
|
+ /* If VPE bound to TC hasn't been checked, do it */
|
|
+ vpe = read_tc_c0_tcbind() & TCBIND_CURVPE;
|
|
+ if(!vpes_checked[vpe]) {
|
|
+ vpeschedule[vpe] = read_vpe_c0_vpeschedule();
|
|
+ vpeschefback[vpe] = read_vpe_c0_vpeschefback();
|
|
+ vpes_checked[vpe] = 1;
|
|
+ }
|
|
+ if(!haltstate) write_tc_c0_tchalt(0);
|
|
+ }
|
|
+ }
|
|
+ /* Re-enable MT and interrupts */
|
|
+ evpe(mtflags);
|
|
+ local_irq_restore(flags);
|
|
+
|
|
+ for(vpe=0; vpe < NVPES; vpe++) {
|
|
+ len = sprintf(page, "VPE[%d].VPEschedule = 0x%08x\n",
|
|
+ vpe, vpeschedule[vpe]);
|
|
+ totalen += len;
|
|
+ page += len;
|
|
+ len = sprintf(page, "VPE[%d].VPEschefback = 0x%08x\n",
|
|
+ vpe, vpeschefback[vpe]);
|
|
+ totalen += len;
|
|
+ page += len;
|
|
+ }
|
|
+ for(i=0; i < NTCS; i++) {
|
|
+ len = sprintf(page, "TC[%d].TCschedule = 0x%08x\n",
|
|
+ i, tcschedule[i]);
|
|
+ totalen += len;
|
|
+ page += len;
|
|
+ len = sprintf(page, "TC[%d].TCschefback = 0x%08x\n",
|
|
+ i, tcschefback[i]);
|
|
+ totalen += len;
|
|
+ page += len;
|
|
+ }
|
|
+ return totalen;
|
|
+}
|
|
+
|
|
+/*
|
|
+ * Write to perf counter registers based on text input
|
|
+ */
|
|
+
|
|
+#define TXTBUFSZ 100
|
|
+
|
|
+static int proc_write_mtsched(struct file *file, const char *buffer,
|
|
+ unsigned long count, void *data)
|
|
+{
|
|
+ int len = 0;
|
|
+ char mybuf[TXTBUFSZ];
|
|
+ /* At most, we will set up 9 TCs and 2 VPEs, 11 entries in all */
|
|
+ char entity[1]; //, entity1[1];
|
|
+ int number[1];
|
|
+ unsigned long value[1];
|
|
+ int nparsed = 0 , index = 0;
|
|
+ unsigned long flags;
|
|
+ unsigned int mtflags;
|
|
+ unsigned int haltstate;
|
|
+ unsigned int tcbindval;
|
|
+
|
|
+ if(count >= TXTBUFSZ) len = TXTBUFSZ-1;
|
|
+ else len = count;
|
|
+ memset(mybuf,0,TXTBUFSZ);
|
|
+ if(copy_from_user(mybuf, buffer, len)) return -EFAULT;
|
|
+
|
|
+ nparsed = sscanf(mybuf, "%c%d %lx",
|
|
+ &entity[0] ,&number[0], &value[0]);
|
|
+
|
|
+ /*
|
|
+ * Having acquired the inputs, which might have
|
|
+ * generated exceptions and preemptions,
|
|
+ * program the registers.
|
|
+ */
|
|
+ /* Disable interrupts and multithreaded issue */
|
|
+ local_irq_save(flags);
|
|
+ mtflags = dvpe();
|
|
+
|
|
+ if(entity[index] == 't' ) {
|
|
+ /* Set TCSchedule or TCScheFBack of specified TC */
|
|
+ if(number[index] > NTCS) goto skip;
|
|
+ /* If it's our own TC, do it direct */
|
|
+ if(number[index] ==
|
|
+ ((read_c0_tcbind() & TCBIND_CURTC)
|
|
+ >> TCBIND_CURTC_SHIFT)) {
|
|
+ if(entity[index] == 't')
|
|
+ write_c0_tcschedule(value[index]);
|
|
+ else
|
|
+ write_c0_tcschefback(value[index]);
|
|
+ } else {
|
|
+ /* Otherwise, we do it via MTTR */
|
|
+ settc(number[index]);
|
|
+ haltstate = read_tc_c0_tchalt();
|
|
+ write_tc_c0_tchalt(TCHALT_H);
|
|
+ mips_ihb();
|
|
+ if(entity[index] == 't')
|
|
+ write_tc_c0_tcschedule(value[index]);
|
|
+ else
|
|
+ write_tc_c0_tcschefback(value[index]);
|
|
+ mips_ihb();
|
|
+ if(!haltstate) write_tc_c0_tchalt(0);
|
|
+ }
|
|
+ } else if(entity[index] == 'v') {
|
|
+ /* Set VPESchedule of specified VPE */
|
|
+ if(number[index] > NVPES) goto skip;
|
|
+ tcbindval = read_c0_tcbind();
|
|
+ /* Are we doing this to our current VPE? */
|
|
+ if((tcbindval & TCBIND_CURVPE) == number[index]) {
|
|
+ /* Then life is simple */
|
|
+ write_c0_vpeschedule(value[index]);
|
|
+ } else {
|
|
+ /*
|
|
+ * Bind ourselves to the other VPE long enough
|
|
+ * to program the bind value.
|
|
+ */
|
|
+ write_c0_tcbind((tcbindval & ~TCBIND_CURVPE)
|
|
+ | number[index]);
|
|
+ mips_ihb();
|
|
+ write_c0_vpeschedule(value[index]);
|
|
+ mips_ihb();
|
|
+ /* Restore previous binding */
|
|
+ write_c0_tcbind(tcbindval);
|
|
+ mips_ihb();
|
|
+ }
|
|
+ }
|
|
+
|
|
+ else if(entity[index] == 'r') {
|
|
+ unsigned int vpes_checked[2], vpe ,i , mytc;
|
|
+ vpes_checked[0] = vpes_checked[1] = 0;
|
|
+
|
|
+ /* Then go through the TCs, halt 'em, and extract the values */
|
|
+ mytc = (read_c0_tcbind() & TCBIND_CURTC) >> TCBIND_CURTC_SHIFT;
|
|
+
|
|
+ for(i = 0; i < NTCS; i++) {
|
|
+ if(i == mytc) {
|
|
+ /* No need to halt ourselves! */
|
|
+ write_c0_vpeschefback(0);
|
|
+ write_c0_tcschefback(0);
|
|
+ } else {
|
|
+ settc(i);
|
|
+ haltstate = read_tc_c0_tchalt();
|
|
+ write_tc_c0_tchalt(TCHALT_H);
|
|
+ mips_ihb();
|
|
+ write_tc_c0_tcschefback(0);
|
|
+ /* If VPE bound to TC hasn't been checked, do it */
|
|
+ vpe = read_tc_c0_tcbind() & TCBIND_CURVPE;
|
|
+ if(!vpes_checked[vpe]) {
|
|
+ write_vpe_c0_vpeschefback(0);
|
|
+ vpes_checked[vpe] = 1;
|
|
+ }
|
|
+ if(!haltstate) write_tc_c0_tchalt(0);
|
|
+ }
|
|
+ }
|
|
+ }
|
|
+ else {
|
|
+ printk ("\n Usage : <t/v><0/1> <Hex Value>\n Example : t0 0x01\n");
|
|
+ }
|
|
+
|
|
+skip:
|
|
+ /* Re-enable MT and interrupts */
|
|
+ evpe(mtflags);
|
|
+ local_irq_restore(flags);
|
|
+ return (len);
|
|
+}
|
|
+
|
|
+static int __init init_mtsched_proc(void)
|
|
+{
|
|
+ extern struct proc_dir_entry *get_mips_proc_dir(void);
|
|
+ struct proc_dir_entry *mips_proc_dir;
|
|
+
|
|
+ if (!cpu_has_mipsmt) {
|
|
+ printk("mtsched: not a MIPS MT capable processor\n");
|
|
+ return -ENODEV;
|
|
+ }
|
|
+
|
|
+ mips_proc_dir = get_mips_proc_dir();
|
|
+
|
|
+ mtsched_proc = create_proc_entry("mtsched", 0644, mips_proc_dir);
|
|
+ mtsched_proc->read_proc = proc_read_mtsched;
|
|
+ mtsched_proc->write_proc = proc_write_mtsched;
|
|
+
|
|
+ return 0;
|
|
+}
|
|
+
|
|
+/* Automagically create the entry */
|
|
+module_init(init_mtsched_proc);
|
|
--- /dev/null
|
|
+++ b/arch/mips/kernel/perf_proc.c
|
|
@@ -0,0 +1,191 @@
|
|
+/*
|
|
+ * /proc hooks for CPU performance counter support for SMTC kernel
|
|
+ * (and ultimately others)
|
|
+ * Copyright (C) 2006 Mips Technologies, Inc
|
|
+ */
|
|
+
|
|
+#include <linux/kernel.h>
|
|
+
|
|
+#include <asm/cpu.h>
|
|
+#include <asm/processor.h>
|
|
+#include <asm/system.h>
|
|
+#include <asm/mipsregs.h>
|
|
+#include <asm/uaccess.h>
|
|
+#include <linux/proc_fs.h>
|
|
+
|
|
+/*
|
|
+ * /proc diagnostic and statistics hooks
|
|
+ */
|
|
+
|
|
+
|
|
+/* Internal software-extended event counters */
|
|
+
|
|
+static unsigned long long extencount[4] = {0,0,0,0};
|
|
+
|
|
+static struct proc_dir_entry *perf_proc;
|
|
+
|
|
+static int proc_read_perf(char *page, char **start, off_t off,
|
|
+ int count, int *eof, void *data)
|
|
+{
|
|
+ int totalen = 0;
|
|
+ int len;
|
|
+
|
|
+ len = sprintf(page, "PerfCnt[0].Ctl : 0x%08x\n", read_c0_perfctrl0());
|
|
+ totalen += len;
|
|
+ page += len;
|
|
+ len = sprintf(page, "PerfCnt[0].Cnt : %Lu\n",
|
|
+ extencount[0] + (unsigned long long)((unsigned)read_c0_perfcntr0()));
|
|
+ totalen += len;
|
|
+ page += len;
|
|
+ len = sprintf(page, "PerfCnt[1].Ctl : 0x%08x\n", read_c0_perfctrl1());
|
|
+ totalen += len;
|
|
+ page += len;
|
|
+ len = sprintf(page, "PerfCnt[1].Cnt : %Lu\n",
|
|
+ extencount[1] + (unsigned long long)((unsigned)read_c0_perfcntr1()));
|
|
+ totalen += len;
|
|
+ page += len;
|
|
+ len = sprintf(page, "PerfCnt[2].Ctl : 0x%08x\n", read_c0_perfctrl2());
|
|
+ totalen += len;
|
|
+ page += len;
|
|
+ len = sprintf(page, "PerfCnt[2].Cnt : %Lu\n",
|
|
+ extencount[2] + (unsigned long long)((unsigned)read_c0_perfcntr2()));
|
|
+ totalen += len;
|
|
+ page += len;
|
|
+ len = sprintf(page, "PerfCnt[3].Ctl : 0x%08x\n", read_c0_perfctrl3());
|
|
+ totalen += len;
|
|
+ page += len;
|
|
+ len = sprintf(page, "PerfCnt[3].Cnt : %Lu\n",
|
|
+ extencount[3] + (unsigned long long)((unsigned)read_c0_perfcntr3()));
|
|
+ totalen += len;
|
|
+ page += len;
|
|
+
|
|
+ return totalen;
|
|
+}
|
|
+
|
|
+/*
|
|
+ * Write to perf counter registers based on text input
|
|
+ */
|
|
+
|
|
+#define TXTBUFSZ 100
|
|
+
|
|
+static int proc_write_perf(struct file *file, const char *buffer,
|
|
+ unsigned long count, void *data)
|
|
+{
|
|
+ int len;
|
|
+ int nparsed;
|
|
+ int index;
|
|
+ char mybuf[TXTBUFSZ];
|
|
+
|
|
+ int which[4];
|
|
+ unsigned long control[4];
|
|
+ long long ctrdata[4];
|
|
+
|
|
+ if(count >= TXTBUFSZ) len = TXTBUFSZ-1;
|
|
+ else len = count;
|
|
+ memset(mybuf,0,TXTBUFSZ);
|
|
+ if(copy_from_user(mybuf, buffer, len)) return -EFAULT;
|
|
+
|
|
+ nparsed = sscanf(mybuf,
|
|
+ "%d %lx %Ld %d %lx %Ld %d %lx %Ld %d %lx %Ld",
|
|
+ &which[0], &control[0], &ctrdata[0],
|
|
+ &which[1], &control[1], &ctrdata[1],
|
|
+ &which[2], &control[2], &ctrdata[2],
|
|
+ &which[3], &control[3], &ctrdata[3]);
|
|
+
|
|
+ for(index = 0; nparsed >= 3; index++) {
|
|
+ switch (which[index]) {
|
|
+ case 0:
|
|
+ write_c0_perfctrl0(control[index]);
|
|
+ if(ctrdata[index] != -1) {
|
|
+ extencount[0] = (unsigned long long)ctrdata[index];
|
|
+ write_c0_perfcntr0((unsigned long)0);
|
|
+ }
|
|
+ break;
|
|
+ case 1:
|
|
+ write_c0_perfctrl1(control[index]);
|
|
+ if(ctrdata[index] != -1) {
|
|
+ extencount[1] = (unsigned long long)ctrdata[index];
|
|
+ write_c0_perfcntr1((unsigned long)0);
|
|
+ }
|
|
+ break;
|
|
+ case 2:
|
|
+ write_c0_perfctrl2(control[index]);
|
|
+ if(ctrdata[index] != -1) {
|
|
+ extencount[2] = (unsigned long long)ctrdata[index];
|
|
+ write_c0_perfcntr2((unsigned long)0);
|
|
+ }
|
|
+ break;
|
|
+ case 3:
|
|
+ write_c0_perfctrl3(control[index]);
|
|
+ if(ctrdata[index] != -1) {
|
|
+ extencount[3] = (unsigned long long)ctrdata[index];
|
|
+ write_c0_perfcntr3((unsigned long)0);
|
|
+ }
|
|
+ break;
|
|
+ }
|
|
+ nparsed -= 3;
|
|
+ }
|
|
+ return (len);
|
|
+}
|
|
+
|
|
+extern int (*perf_irq)(void);
|
|
+
|
|
+/*
|
|
+ * Invoked when timer interrupt vector picks up a perf counter overflow
|
|
+ */
|
|
+
|
|
+static int perf_proc_irq(void)
|
|
+{
|
|
+ unsigned long snapshot;
|
|
+
|
|
+ /*
|
|
+ * It would be nice to do this as a loop, but we don't have
|
|
+ * indirect access to CP0 registers.
|
|
+ */
|
|
+ snapshot = read_c0_perfcntr0();
|
|
+ if ((long)snapshot < 0) {
|
|
+ extencount[0] +=
|
|
+ (unsigned long long)((unsigned)read_c0_perfcntr0());
|
|
+ write_c0_perfcntr0(0);
|
|
+ }
|
|
+ snapshot = read_c0_perfcntr1();
|
|
+ if ((long)snapshot < 0) {
|
|
+ extencount[1] +=
|
|
+ (unsigned long long)((unsigned)read_c0_perfcntr1());
|
|
+ write_c0_perfcntr1(0);
|
|
+ }
|
|
+ snapshot = read_c0_perfcntr2();
|
|
+ if ((long)snapshot < 0) {
|
|
+ extencount[2] +=
|
|
+ (unsigned long long)((unsigned)read_c0_perfcntr2());
|
|
+ write_c0_perfcntr2(0);
|
|
+ }
|
|
+ snapshot = read_c0_perfcntr3();
|
|
+ if ((long)snapshot < 0) {
|
|
+ extencount[3] +=
|
|
+ (unsigned long long)((unsigned)read_c0_perfcntr3());
|
|
+ write_c0_perfcntr3(0);
|
|
+ }
|
|
+ return 0;
|
|
+}
|
|
+
|
|
+static int __init init_perf_proc(void)
|
|
+{
|
|
+ extern struct proc_dir_entry *get_mips_proc_dir(void);
|
|
+
|
|
+ struct proc_dir_entry *mips_proc_dir = get_mips_proc_dir();
|
|
+
|
|
+ write_c0_perfcntr0(0);
|
|
+ write_c0_perfcntr1(0);
|
|
+ write_c0_perfcntr2(0);
|
|
+ write_c0_perfcntr3(0);
|
|
+ perf_proc = create_proc_entry("perf", 0644, mips_proc_dir);
|
|
+ perf_proc->read_proc = proc_read_perf;
|
|
+ perf_proc->write_proc = proc_write_perf;
|
|
+ perf_irq = perf_proc_irq;
|
|
+
|
|
+ return 0;
|
|
+}
|
|
+
|
|
+/* Automagically create the entry */
|
|
+module_init(init_perf_proc);
|
|
--- a/arch/mips/kernel/proc.c
|
|
+++ b/arch/mips/kernel/proc.c
|
|
@@ -7,6 +7,7 @@
|
|
#include <linux/kernel.h>
|
|
#include <linux/sched.h>
|
|
#include <linux/seq_file.h>
|
|
+#include <linux/proc_fs.h>
|
|
#include <asm/bootinfo.h>
|
|
#include <asm/cpu.h>
|
|
#include <asm/cpu-features.h>
|
|
@@ -110,3 +111,19 @@ const struct seq_operations cpuinfo_op =
|
|
.stop = c_stop,
|
|
.show = show_cpuinfo,
|
|
};
|
|
+
|
|
+/*
|
|
+ * Support for MIPS/local /proc hooks in /proc/mips/
|
|
+ */
|
|
+
|
|
+static struct proc_dir_entry *mips_proc = NULL;
|
|
+
|
|
+struct proc_dir_entry *get_mips_proc_dir(void)
|
|
+{
|
|
+ /*
|
|
+ * This ought not to be preemptable.
|
|
+ */
|
|
+ if(mips_proc == NULL)
|
|
+ mips_proc = proc_mkdir("mips", NULL);
|
|
+ return(mips_proc);
|
|
+}
|
|
--- a/arch/mips/kernel/smtc.c
|
|
+++ b/arch/mips/kernel/smtc.c
|
|
@@ -1334,6 +1334,13 @@ void smtc_get_new_mmu_context(struct mm_
|
|
asid = asid_cache(cpu);
|
|
|
|
do {
|
|
+#ifdef CONFIG_IFX_VPE_EXT
|
|
+ /* If TLB is shared between AP and RP (AP is running SMTC),
|
|
+ leave out max ASID i.e., ASID_MASK for RP
|
|
+ */
|
|
+ if (!nostlb && ((asid & ASID_MASK) == (ASID_MASK - 1)))
|
|
+ asid++;
|
|
+#endif
|
|
if (!((asid += ASID_INC) & ASID_MASK) ) {
|
|
if (cpu_has_vtag_icache)
|
|
flush_icache_all();
|
|
--- a/arch/mips/kernel/vpe.c
|
|
+++ b/arch/mips/kernel/vpe.c
|
|
@@ -75,6 +75,58 @@ static struct kspd_notifications kspd_ev
|
|
static int kspd_events_reqd;
|
|
#endif
|
|
|
|
+#ifdef CONFIG_IFX_VPE_EXT
|
|
+static int is_sdepgm;
|
|
+extern int stlb;
|
|
+extern int vpe0_wired;
|
|
+extern int vpe1_wired;
|
|
+unsigned int vpe1_load_addr;
|
|
+
|
|
+static int __init load_address(char *str)
|
|
+{
|
|
+ get_option(&str, &vpe1_load_addr);
|
|
+ return 1;
|
|
+}
|
|
+__setup("vpe1_load_addr=", load_address);
|
|
+
|
|
+#include <asm/mipsmtregs.h>
|
|
+#define write_vpe_c0_wired(val) mttc0(6, 0, val)
|
|
+
|
|
+#ifndef COMMAND_LINE_SIZE
|
|
+# define COMMAND_LINE_SIZE 512
|
|
+#endif
|
|
+
|
|
+char command_line[COMMAND_LINE_SIZE * 2];
|
|
+
|
|
+static unsigned int vpe1_mem;
|
|
+static int __init vpe1mem(char *str)
|
|
+{
|
|
+ vpe1_mem = memparse(str, &str);
|
|
+ return 1;
|
|
+}
|
|
+__setup("vpe1_mem=", vpe1mem);
|
|
+
|
|
+uint32_t vpe1_wdog_ctr;
|
|
+static int __init wdog_ctr(char *str)
|
|
+{
|
|
+ get_option(&str, &vpe1_wdog_ctr);
|
|
+ return 1;
|
|
+}
|
|
+
|
|
+__setup("vpe1_wdog_ctr_addr=", wdog_ctr);
|
|
+EXPORT_SYMBOL(vpe1_wdog_ctr);
|
|
+
|
|
+uint32_t vpe1_wdog_timeout;
|
|
+static int __init wdog_timeout(char *str)
|
|
+{
|
|
+ get_option(&str, &vpe1_wdog_timeout);
|
|
+ return 1;
|
|
+}
|
|
+
|
|
+__setup("vpe1_wdog_timeout=", wdog_timeout);
|
|
+EXPORT_SYMBOL(vpe1_wdog_timeout);
|
|
+
|
|
+#endif
|
|
/* grab the likely amount of memory we will need. */
|
|
#ifdef CONFIG_MIPS_VPE_LOADER_TOM
|
|
#define P_SIZE (2 * 1024 * 1024)
|
|
@@ -267,6 +319,13 @@ static void *alloc_progmem(unsigned long
|
|
void *addr;
|
|
|
|
#ifdef CONFIG_MIPS_VPE_LOADER_TOM
|
|
+#ifdef CONFIG_IFX_VPE_EXT
|
|
+ if (vpe1_load_addr) {
|
|
+ memset((void *)vpe1_load_addr, 0, len);
|
|
+ return (void *)vpe1_load_addr;
|
|
+ }
|
|
+#endif
|
|
+
|
|
/*
|
|
* This means you must tell Linux to use less memory than you
|
|
* physically have, for example by passing a mem= boot argument.
|
|
@@ -745,6 +804,12 @@ static int vpe_run(struct vpe * v)
|
|
}
|
|
|
|
/* Write the address we want it to start running from in the TCPC register. */
|
|
+#if defined(CONFIG_IFX_VPE_EXT) && 0
|
|
+ if (stlb)
|
|
+ write_vpe_c0_wired(vpe0_wired + vpe1_wired);
|
|
+ else
|
|
+ write_vpe_c0_wired(vpe1_wired);
|
|
+#endif
|
|
write_tc_c0_tcrestart((unsigned long)v->__start);
|
|
write_tc_c0_tccontext((unsigned long)0);
|
|
|
|
@@ -758,6 +823,20 @@ static int vpe_run(struct vpe * v)
|
|
|
|
write_tc_c0_tchalt(read_tc_c0_tchalt() & ~TCHALT_H);
|
|
|
|
+#if defined(CONFIG_IFX_VPE_EXT) && 0
|
|
+ /*
|
|
+ * $a2 & $a3 are used to pass command line parameters to VPE1. $a2
|
|
+ * points to the start of the command line string and $a3 points to
|
|
+ * the end of the string. This convention is identical to the Linux
|
|
+ * kernel boot parameter passing mechanism. Please note that $a3 is
|
|
+ * used to pass physical memory size or 0 in SDE tool kit. So, if you
|
|
+ * are passing comand line parameters through $a2 & $a3 SDE programs
|
|
+ * don't work as desired.
|
|
+ */
|
|
+ mttgpr(6, command_line);
|
|
+ mttgpr(7, (command_line + strlen(command_line)));
|
|
+ if (is_sdepgm)
|
|
+#endif
|
|
/*
|
|
* The sde-kit passes 'memsize' to __start in $a3, so set something
|
|
* here... Or set $a3 to zero and define DFLT_STACK_SIZE and
|
|
@@ -832,6 +911,9 @@ static int find_vpe_symbols(struct vpe *
|
|
if ( (v->__start == 0) || (v->shared_ptr == NULL))
|
|
return -1;
|
|
|
|
+#ifdef CONFIG_IFX_VPE_EXT
|
|
+ is_sdepgm = 1;
|
|
+#endif
|
|
return 0;
|
|
}
|
|
|
|
@@ -993,6 +1075,15 @@ static int vpe_elfload(struct vpe * v)
|
|
(unsigned long)v->load_addr + v->len);
|
|
|
|
if ((find_vpe_symbols(v, sechdrs, symindex, strtab, &mod)) < 0) {
|
|
+#ifdef CONFIG_IFX_VPE_EXT
|
|
+ if (vpe1_load_addr) {
|
|
+ /* Conversion to KSEG1 is required ??? */
|
|
+ v->__start = KSEG1ADDR(vpe1_load_addr);
|
|
+ is_sdepgm = 0;
|
|
+ return 0;
|
|
+ }
|
|
+#endif
|
|
+
|
|
if (v->__start == 0) {
|
|
printk(KERN_WARNING "VPE loader: program does not contain "
|
|
"a __start symbol\n");
|
|
@@ -1063,6 +1154,9 @@ static int vpe_open(struct inode *inode,
|
|
struct vpe_notifications *not;
|
|
struct vpe *v;
|
|
int ret;
|
|
+#ifdef CONFIG_IFX_VPE_EXT
|
|
+ int progsize;
|
|
+#endif
|
|
|
|
if (minor != iminor(inode)) {
|
|
/* assume only 1 device at the moment. */
|
|
@@ -1088,7 +1182,12 @@ static int vpe_open(struct inode *inode,
|
|
release_progmem(v->load_addr);
|
|
cleanup_tc(get_tc(tclimit));
|
|
}
|
|
-
|
|
+#ifdef CONFIG_IFX_VPE_EXT
|
|
+ progsize = (vpe1_mem != 0) ? vpe1_mem : P_SIZE;
|
|
+ //printk("progsize = %x\n", progsize);
|
|
+ v->pbuffer = vmalloc(progsize);
|
|
+ v->plen = progsize;
|
|
+#else
|
|
/* this of-course trashes what was there before... */
|
|
v->pbuffer = vmalloc(P_SIZE);
|
|
if (!v->pbuffer) {
|
|
@@ -1096,11 +1195,14 @@ static int vpe_open(struct inode *inode,
|
|
return -ENOMEM;
|
|
}
|
|
v->plen = P_SIZE;
|
|
+#endif
|
|
v->load_addr = NULL;
|
|
v->len = 0;
|
|
|
|
+#if 0
|
|
v->uid = filp->f_cred->fsuid;
|
|
v->gid = filp->f_cred->fsgid;
|
|
+#endif
|
|
|
|
#ifdef CONFIG_MIPS_APSP_KSPD
|
|
/* get kspd to tell us when a syscall_exit happens */
|
|
@@ -1348,6 +1450,133 @@ static void kspd_sp_exit( int sp_id)
|
|
cleanup_tc(get_tc(sp_id));
|
|
}
|
|
#endif
|
|
+#ifdef CONFIG_IFX_VPE_EXT
|
|
+int32_t vpe1_sw_start(void* sw_start_addr, uint32_t tcmask, uint32_t flags)
|
|
+{
|
|
+ enum vpe_state state;
|
|
+ struct vpe *v = get_vpe(tclimit);
|
|
+ struct vpe_notifications *not;
|
|
+
|
|
+ if (tcmask || flags) {
|
|
+ printk(KERN_WARNING "Currently tcmask and flags should be 0.\
|
|
+ other values not supported\n");
|
|
+ return -1;
|
|
+ }
|
|
+
|
|
+ state = xchg(&v->state, VPE_STATE_INUSE);
|
|
+ if (state != VPE_STATE_UNUSED) {
|
|
+ vpe_stop(v);
|
|
+
|
|
+ list_for_each_entry(not, &v->notify, list) {
|
|
+ not->stop(tclimit);
|
|
+ }
|
|
+ }
|
|
+
|
|
+ v->__start = (unsigned long)sw_start_addr;
|
|
+ is_sdepgm = 0;
|
|
+
|
|
+ if (!vpe_run(v)) {
|
|
+ printk(KERN_DEBUG "VPE loader: VPE1 running successfully\n");
|
|
+ return 0;
|
|
+ }
|
|
+ return -1;
|
|
+}
|
|
+
|
|
+EXPORT_SYMBOL(vpe1_sw_start);
|
|
+
|
|
+int32_t vpe1_sw_stop(uint32_t flags)
|
|
+{
|
|
+ struct vpe *v = get_vpe(tclimit);
|
|
+
|
|
+ if (!vpe_free(v)) {
|
|
+ printk(KERN_DEBUG "RP Stopped\n");
|
|
+ return 0;
|
|
+ }
|
|
+ else
|
|
+ return -1;
|
|
+}
|
|
+
|
|
+EXPORT_SYMBOL(vpe1_sw_stop);
|
|
+
|
|
+uint32_t vpe1_get_load_addr (uint32_t flags)
|
|
+{
|
|
+ return vpe1_load_addr;
|
|
+}
|
|
+
|
|
+EXPORT_SYMBOL(vpe1_get_load_addr);
|
|
+
|
|
+uint32_t vpe1_get_max_mem (uint32_t flags)
|
|
+{
|
|
+ if (!vpe1_mem)
|
|
+ return P_SIZE;
|
|
+ else
|
|
+ return vpe1_mem;
|
|
+}
|
|
+
|
|
+EXPORT_SYMBOL(vpe1_get_max_mem);
|
|
+
|
|
+void* vpe1_get_cmdline_argument(void)
|
|
+{
|
|
+ return saved_command_line;
|
|
+}
|
|
+
|
|
+EXPORT_SYMBOL(vpe1_get_cmdline_argument);
|
|
+
|
|
+int32_t vpe1_set_boot_param(char *field, char *value, char flags)
|
|
+{
|
|
+ char *ptr, string[64];
|
|
+ int start_off, end_off;
|
|
+ if (!field)
|
|
+ return -1;
|
|
+ strcpy(string, field);
|
|
+ if (value) {
|
|
+ strcat(string, "=");
|
|
+ strcat(string, value);
|
|
+ strcat(command_line, " ");
|
|
+ strcat(command_line, string);
|
|
+ }
|
|
+ else {
|
|
+ ptr = strstr(command_line, string);
|
|
+ if (ptr) {
|
|
+ start_off = ptr - command_line;
|
|
+ ptr += strlen(string);
|
|
+ while ((*ptr != ' ') && (*ptr != '\0'))
|
|
+ ptr++;
|
|
+ end_off = ptr - command_line;
|
|
+ command_line[start_off] = '\0';
|
|
+ strcat (command_line, command_line+end_off);
|
|
+ }
|
|
+ }
|
|
+ return 0;
|
|
+}
|
|
+
|
|
+EXPORT_SYMBOL(vpe1_set_boot_param);
|
|
+
|
|
+int32_t vpe1_get_boot_param(char *field, char **value, char flags)
|
|
+{
|
|
+ char *ptr, string[64];
|
|
+ int i = 0;
|
|
+ if (!field)
|
|
+ return -1;
|
|
+ if ((ptr = strstr(command_line, field))) {
|
|
+ ptr += strlen(field) + 1; /* including = */
|
|
+ while ((*ptr != ' ') && (*ptr != '\0'))
|
|
+ string[i++] = *ptr++;
|
|
+ string[i] = '\0';
|
|
+ *value = kmalloc((strlen(string) + 1), GFP_KERNEL);
|
|
+ if (*value != NULL)
|
|
+ strcpy(*value, string);
|
|
+ }
|
|
+ else
|
|
+ *value = NULL;
|
|
+
|
|
+ return 0;
|
|
+}
|
|
+
|
|
+EXPORT_SYMBOL(vpe1_get_boot_param);
|
|
+
|
|
+extern void configure_tlb(void);
|
|
+#endif
|
|
|
|
static ssize_t store_kill(struct device *dev, struct device_attribute *attr,
|
|
const char *buf, size_t len)
|
|
@@ -1429,6 +1658,18 @@ static int __init vpe_module_init(void)
|
|
printk("VPE loader: not a MIPS MT capable processor\n");
|
|
return -ENODEV;
|
|
}
|
|
+#ifdef CONFIG_IFX_VPE_EXT
|
|
+#ifndef CONFIG_MIPS_MT_SMTC
|
|
+ configure_tlb();
|
|
+#endif
|
|
+#endif
|
|
+
|
|
+#ifndef CONFIG_MIPS_MT_SMTC
|
|
+ if (!vpelimit)
|
|
+ vpelimit = 1;
|
|
+ if (!tclimit)
|
|
+ tclimit = 1;
|
|
+#endif
|
|
|
|
if (vpelimit == 0) {
|
|
printk(KERN_WARNING "No VPEs reserved for AP/SP, not "
|
|
@@ -1473,10 +1714,12 @@ static int __init vpe_module_init(void)
|
|
mtflags = dmt();
|
|
vpflags = dvpe();
|
|
|
|
+ back_to_back_c0_hazard();
|
|
+
|
|
/* Put MVPE's into 'configuration state' */
|
|
set_c0_mvpcontrol(MVPCONTROL_VPC);
|
|
|
|
- /* dump_mtregs(); */
|
|
+ dump_mtregs();
|
|
|
|
val = read_c0_mvpconf0();
|
|
hw_tcs = (val & MVPCONF0_PTC) + 1;
|
|
@@ -1488,6 +1731,7 @@ static int __init vpe_module_init(void)
|
|
* reschedule send IPIs or similar we might hang.
|
|
*/
|
|
clear_c0_mvpcontrol(MVPCONTROL_VPC);
|
|
+ back_to_back_c0_hazard();
|
|
evpe(vpflags);
|
|
emt(mtflags);
|
|
local_irq_restore(flags);
|
|
@@ -1513,6 +1757,7 @@ static int __init vpe_module_init(void)
|
|
}
|
|
|
|
v->ntcs = hw_tcs - tclimit;
|
|
+ write_tc_c0_tcbind((read_tc_c0_tcbind() & ~TCBIND_CURVPE) | 1);
|
|
|
|
/* add the tc to the list of this vpe's tc's. */
|
|
list_add(&t->tc, &v->tc);
|
|
@@ -1581,6 +1826,7 @@ static int __init vpe_module_init(void)
|
|
out_reenable:
|
|
/* release config state */
|
|
clear_c0_mvpcontrol(MVPCONTROL_VPC);
|
|
+ back_to_back_c0_hazard();
|
|
|
|
evpe(vpflags);
|
|
emt(mtflags);
|