7112 lines
200 KiB
Diff
7112 lines
200 KiB
Diff
diff --git a/Documentation/arm64/booting.txt b/Documentation/arm64/booting.txt
|
|
index 85af34d..f3c05b5 100644
|
|
--- a/Documentation/arm64/booting.txt
|
|
+++ b/Documentation/arm64/booting.txt
|
|
@@ -168,6 +168,14 @@ Before jumping into the kernel, the following conditions must be met:
|
|
the kernel image will be entered must be initialised by software at a
|
|
higher exception level to prevent execution in an UNKNOWN state.
|
|
|
|
+ For systems with a GICv3 interrupt controller:
|
|
+ - If EL3 is present:
|
|
+ ICC_SRE_EL3.Enable (bit 3) must be initialiased to 0b1.
|
|
+ ICC_SRE_EL3.SRE (bit 0) must be initialised to 0b1.
|
|
+ - If the kernel is entered at EL1:
|
|
+ ICC.SRE_EL2.Enable (bit 3) must be initialised to 0b1
|
|
+ ICC_SRE_EL2.SRE (bit 0) must be initialised to 0b1.
|
|
+
|
|
The requirements described above for CPU mode, caches, MMUs, architected
|
|
timers, coherency and system registers apply to all CPUs. All CPUs must
|
|
enter the kernel in the same exception level.
|
|
diff --git a/Documentation/devicetree/bindings/arm/gic-v3.txt b/Documentation/devicetree/bindings/arm/gic-v3.txt
|
|
new file mode 100644
|
|
index 0000000..33cd05e
|
|
--- /dev/null
|
|
+++ b/Documentation/devicetree/bindings/arm/gic-v3.txt
|
|
@@ -0,0 +1,79 @@
|
|
+* ARM Generic Interrupt Controller, version 3
|
|
+
|
|
+AArch64 SMP cores are often associated with a GICv3, providing Private
|
|
+Peripheral Interrupts (PPI), Shared Peripheral Interrupts (SPI),
|
|
+Software Generated Interrupts (SGI), and Locality-specific Peripheral
|
|
+Interrupts (LPI).
|
|
+
|
|
+Main node required properties:
|
|
+
|
|
+- compatible : should at least contain "arm,gic-v3".
|
|
+- interrupt-controller : Identifies the node as an interrupt controller
|
|
+- #interrupt-cells : Specifies the number of cells needed to encode an
|
|
+ interrupt source. Must be a single cell with a value of at least 3.
|
|
+
|
|
+ The 1st cell is the interrupt type; 0 for SPI interrupts, 1 for PPI
|
|
+ interrupts. Other values are reserved for future use.
|
|
+
|
|
+ The 2nd cell contains the interrupt number for the interrupt type.
|
|
+ SPI interrupts are in the range [0-987]. PPI interrupts are in the
|
|
+ range [0-15].
|
|
+
|
|
+ The 3rd cell is the flags, encoded as follows:
|
|
+ bits[3:0] trigger type and level flags.
|
|
+ 1 = edge triggered
|
|
+ 4 = level triggered
|
|
+
|
|
+ Cells 4 and beyond are reserved for future use. When the 1st cell
|
|
+ has a value of 0 or 1, cells 4 and beyond act as padding, and may be
|
|
+ ignored. It is recommended that padding cells have a value of 0.
|
|
+
|
|
+- reg : Specifies base physical address(s) and size of the GIC
|
|
+ registers, in the following order:
|
|
+ - GIC Distributor interface (GICD)
|
|
+ - GIC Redistributors (GICR), one range per redistributor region
|
|
+ - GIC CPU interface (GICC)
|
|
+ - GIC Hypervisor interface (GICH)
|
|
+ - GIC Virtual CPU interface (GICV)
|
|
+
|
|
+ GICC, GICH and GICV are optional.
|
|
+
|
|
+- interrupts : Interrupt source of the VGIC maintenance interrupt.
|
|
+
|
|
+Optional
|
|
+
|
|
+- redistributor-stride : If using padding pages, specifies the stride
|
|
+ of consecutive redistributors. Must be a multiple of 64kB.
|
|
+
|
|
+- #redistributor-regions: The number of independent contiguous regions
|
|
+ occupied by the redistributors. Required if more than one such
|
|
+ region is present.
|
|
+
|
|
+Examples:
|
|
+
|
|
+ gic: interrupt-controller@2cf00000 {
|
|
+ compatible = "arm,gic-v3";
|
|
+ #interrupt-cells = <3>;
|
|
+ interrupt-controller;
|
|
+ reg = <0x0 0x2f000000 0 0x10000>, // GICD
|
|
+ <0x0 0x2f100000 0 0x200000>, // GICR
|
|
+ <0x0 0x2c000000 0 0x2000>, // GICC
|
|
+ <0x0 0x2c010000 0 0x2000>, // GICH
|
|
+ <0x0 0x2c020000 0 0x2000>; // GICV
|
|
+ interrupts = <1 9 4>;
|
|
+ };
|
|
+
|
|
+ gic: interrupt-controller@2c010000 {
|
|
+ compatible = "arm,gic-v3";
|
|
+ #interrupt-cells = <3>;
|
|
+ interrupt-controller;
|
|
+ redistributor-stride = <0x0 0x40000>; // 256kB stride
|
|
+ #redistributor-regions = <2>;
|
|
+ reg = <0x0 0x2c010000 0 0x10000>, // GICD
|
|
+ <0x0 0x2d000000 0 0x800000>, // GICR 1: CPUs 0-31
|
|
+ <0x0 0x2e000000 0 0x800000>; // GICR 2: CPUs 32-63
|
|
+ <0x0 0x2c040000 0 0x2000>, // GICC
|
|
+ <0x0 0x2c060000 0 0x2000>, // GICH
|
|
+ <0x0 0x2c080000 0 0x2000>; // GICV
|
|
+ interrupts = <1 9 4>;
|
|
+ };
|
|
diff --git a/Documentation/devicetree/bindings/net/apm-xgene-enet.txt b/Documentation/devicetree/bindings/net/apm-xgene-enet.txt
|
|
new file mode 100644
|
|
index 0000000..3e2a295
|
|
--- /dev/null
|
|
+++ b/Documentation/devicetree/bindings/net/apm-xgene-enet.txt
|
|
@@ -0,0 +1,72 @@
|
|
+APM X-Gene SoC Ethernet nodes
|
|
+
|
|
+Ethernet nodes are defined to describe on-chip ethernet interfaces in
|
|
+APM X-Gene SoC.
|
|
+
|
|
+Required properties:
|
|
+- compatible: Should be "apm,xgene-enet"
|
|
+- reg: Address and length of the register set for the device. It contains the
|
|
+ information of registers in the same order as described by reg-names
|
|
+- reg-names: Should contain the register set names
|
|
+ "enet_csr": Ethernet control and status register address space
|
|
+ "ring_csr": Descriptor ring control and status register address space
|
|
+ "ring_cmd": Descriptor ring command register address space
|
|
+- interrupts: Ethernet main interrupt
|
|
+- clocks: Reference to the clock entry.
|
|
+- local-mac-address: MAC address assigned to this device
|
|
+- phy-connection-type: Interface type between ethernet device and PHY device
|
|
+- phy-handle: Reference to a PHY node connected to this device
|
|
+
|
|
+- mdio: Device tree subnode with the following required
|
|
+ properties:
|
|
+
|
|
+ - compatible: Must be "apm,xgene-mdio".
|
|
+ - #address-cells: Must be <1>.
|
|
+ - #size-cells: Must be <0>.
|
|
+
|
|
+ For the phy on the mdio bus, there must be a node with the following
|
|
+ fields:
|
|
+
|
|
+ - compatible: PHY identifier. Please refer ./phy.txt for the format.
|
|
+ - reg: The ID number for the phy.
|
|
+
|
|
+Optional properties:
|
|
+- status : Should be "ok" or "disabled" for enabled/disabled.
|
|
+ Default is "ok".
|
|
+
|
|
+
|
|
+Example:
|
|
+ menetclk: menetclk {
|
|
+ compatible = "apm,xgene-device-clock";
|
|
+ clock-output-names = "menetclk";
|
|
+ status = "ok";
|
|
+ };
|
|
+
|
|
+ menet: ethernet@17020000 {
|
|
+ compatible = "apm,xgene-enet";
|
|
+ status = "disabled";
|
|
+ reg = <0x0 0x17020000 0x0 0xd100>,
|
|
+ <0x0 0X17030000 0x0 0X400>,
|
|
+ <0x0 0X10000000 0x0 0X200>;
|
|
+ reg-names = "enet_csr", "ring_csr", "ring_cmd";
|
|
+ interrupts = <0x0 0x3c 0x4>;
|
|
+ clocks = <&menetclk 0>;
|
|
+ local-mac-address = [00 01 73 00 00 01];
|
|
+ phy-connection-type = "rgmii";
|
|
+ phy-handle = <&menetphy>;
|
|
+ mdio {
|
|
+ compatible = "apm,xgene-mdio";
|
|
+ #address-cells = <1>;
|
|
+ #size-cells = <0>;
|
|
+ menetphy: menetphy@3 {
|
|
+ compatible = "ethernet-phy-id001c.c915";
|
|
+ reg = <0x3>;
|
|
+ };
|
|
+
|
|
+ };
|
|
+ };
|
|
+
|
|
+/* Board-specific peripheral configurations */
|
|
+&menet {
|
|
+ status = "ok";
|
|
+};
|
|
diff --git a/Documentation/devicetree/bindings/pci/xgene-pci.txt b/Documentation/devicetree/bindings/pci/xgene-pci.txt
|
|
new file mode 100644
|
|
index 0000000..e19fdb8
|
|
--- /dev/null
|
|
+++ b/Documentation/devicetree/bindings/pci/xgene-pci.txt
|
|
@@ -0,0 +1,52 @@
|
|
+* AppliedMicro X-Gene PCIe interface
|
|
+
|
|
+Required properties:
|
|
+- device_type: set to "pci"
|
|
+- compatible: should contain "apm,xgene-pcie" to identify the core.
|
|
+- reg: A list of physical base address and length for each set of controller
|
|
+ registers. Must contain an entry for each entry in the reg-names
|
|
+ property.
|
|
+- reg-names: Must include the following entries:
|
|
+ "csr": controller configuration registers.
|
|
+ "cfg": pcie configuration space registers.
|
|
+- #address-cells: set to <3>
|
|
+- #size-cells: set to <2>
|
|
+- ranges: ranges for the outbound memory, I/O regions.
|
|
+- dma-ranges: ranges for the inbound memory regions.
|
|
+- #interrupt-cells: set to <1>
|
|
+- interrupt-map-mask and interrupt-map: standard PCI properties
|
|
+ to define the mapping of the PCIe interface to interrupt
|
|
+ numbers.
|
|
+- clocks: from common clock binding: handle to pci clock.
|
|
+
|
|
+Optional properties:
|
|
+- status: Either "ok" or "disabled".
|
|
+
|
|
+Example:
|
|
+
|
|
+SoC specific DT Entry:
|
|
+ pcie0: pcie@1f2b0000 {
|
|
+ status = "disabled";
|
|
+ device_type = "pci";
|
|
+ compatible = "apm,xgene-storm-pcie", "apm,xgene-pcie";
|
|
+ #interrupt-cells = <1>;
|
|
+ #size-cells = <2>;
|
|
+ #address-cells = <3>;
|
|
+ reg = < 0x00 0x1f2b0000 0x0 0x00010000 /* Controller registers */
|
|
+ 0xe0 0xd0000000 0x0 0x00200000>; /* PCI config space */
|
|
+ reg-names = "csr", "cfg";
|
|
+ ranges = <0x01000000 0x00 0x00000000 0xe0 0x00000000 0x00 0x00010000 /* io */
|
|
+ 0x02000000 0x00 0x10000000 0xe0 0x10000000 0x00 0x80000000>; /* mem */
|
|
+ dma-ranges = <0x42000000 0x40 0x00000000 0x40 0x00000000 0x40 0x00000000>;
|
|
+ interrupt-map-mask = <0x0 0x0 0x0 0x7>;
|
|
+ interrupt-map = <0x0 0x0 0x0 0x1 &gic 0x0 0xc2 0x1
|
|
+ 0x0 0x0 0x0 0x2 &gic 0x0 0xc3 0x1
|
|
+ 0x0 0x0 0x0 0x3 &gic 0x0 0xc4 0x1
|
|
+ 0x0 0x0 0x0 0x4 &gic 0x0 0xc5 0x1>;
|
|
+ clocks = <&pcie0clk 0>;
|
|
+ };
|
|
+
|
|
+Board specific DT Entry:
|
|
+ &pcie0 {
|
|
+ status = "ok";
|
|
+ };
|
|
diff --git a/MAINTAINERS b/MAINTAINERS
|
|
index 0f066dd..5a10717 100644
|
|
--- a/MAINTAINERS
|
|
+++ b/MAINTAINERS
|
|
@@ -701,6 +701,14 @@ S: Maintained
|
|
F: drivers/net/appletalk/
|
|
F: net/appletalk/
|
|
|
|
+APPLIED MICRO (APM) X-GENE SOC ETHERNET DRIVER
|
|
+M: Iyappan Subramanian <isubramanian@apm.com>
|
|
+M: Keyur Chudgar <kchudgar@apm.com>
|
|
+M: Ravi Patel <rapatel@apm.com>
|
|
+S: Supported
|
|
+F: drivers/net/ethernet/apm/xgene/
|
|
+F: Documentation/devicetree/bindings/net/apm-xgene-enet.txt
|
|
+
|
|
APTINA CAMERA SENSOR PLL
|
|
M: Laurent Pinchart <Laurent.pinchart@ideasonboard.com>
|
|
L: linux-media@vger.kernel.org
|
|
@@ -6861,6 +6869,13 @@ S: Maintained
|
|
F: Documentation/devicetree/bindings/pci/host-generic-pci.txt
|
|
F: drivers/pci/host/pci-host-generic.c
|
|
|
|
+PCI DRIVER FOR APPLIEDMICRO XGENE
|
|
+M: Tanmay Inamdar <tinamdar@apm.com>
|
|
+L: linux-pci@vger.kernel.org
|
|
+L: linux-arm-kernel@lists.infradead.org
|
|
+S: Maintained
|
|
+F: drivers/pci/host/pci-xgene.c
|
|
+
|
|
PCMCIA SUBSYSTEM
|
|
P: Linux PCMCIA Team
|
|
L: linux-pcmcia@lists.infradead.org
|
|
diff --git a/arch/arm/include/asm/kvm_host.h b/arch/arm/include/asm/kvm_host.h
|
|
index 193ceaf..d6d5227 100644
|
|
--- a/arch/arm/include/asm/kvm_host.h
|
|
+++ b/arch/arm/include/asm/kvm_host.h
|
|
@@ -225,6 +225,11 @@ static inline int kvm_arch_dev_ioctl_check_extension(long ext)
|
|
return 0;
|
|
}
|
|
|
|
+static inline void vgic_arch_setup(const struct vgic_params *vgic)
|
|
+{
|
|
+ BUG_ON(vgic->type != VGIC_V2);
|
|
+}
|
|
+
|
|
int kvm_perf_init(void);
|
|
int kvm_perf_teardown(void);
|
|
|
|
diff --git a/arch/arm/kernel/asm-offsets.c b/arch/arm/kernel/asm-offsets.c
|
|
index 85598b5..713e807 100644
|
|
--- a/arch/arm/kernel/asm-offsets.c
|
|
+++ b/arch/arm/kernel/asm-offsets.c
|
|
@@ -182,13 +182,13 @@ int main(void)
|
|
DEFINE(VCPU_HYP_PC, offsetof(struct kvm_vcpu, arch.fault.hyp_pc));
|
|
#ifdef CONFIG_KVM_ARM_VGIC
|
|
DEFINE(VCPU_VGIC_CPU, offsetof(struct kvm_vcpu, arch.vgic_cpu));
|
|
- DEFINE(VGIC_CPU_HCR, offsetof(struct vgic_cpu, vgic_hcr));
|
|
- DEFINE(VGIC_CPU_VMCR, offsetof(struct vgic_cpu, vgic_vmcr));
|
|
- DEFINE(VGIC_CPU_MISR, offsetof(struct vgic_cpu, vgic_misr));
|
|
- DEFINE(VGIC_CPU_EISR, offsetof(struct vgic_cpu, vgic_eisr));
|
|
- DEFINE(VGIC_CPU_ELRSR, offsetof(struct vgic_cpu, vgic_elrsr));
|
|
- DEFINE(VGIC_CPU_APR, offsetof(struct vgic_cpu, vgic_apr));
|
|
- DEFINE(VGIC_CPU_LR, offsetof(struct vgic_cpu, vgic_lr));
|
|
+ DEFINE(VGIC_V2_CPU_HCR, offsetof(struct vgic_cpu, vgic_v2.vgic_hcr));
|
|
+ DEFINE(VGIC_V2_CPU_VMCR, offsetof(struct vgic_cpu, vgic_v2.vgic_vmcr));
|
|
+ DEFINE(VGIC_V2_CPU_MISR, offsetof(struct vgic_cpu, vgic_v2.vgic_misr));
|
|
+ DEFINE(VGIC_V2_CPU_EISR, offsetof(struct vgic_cpu, vgic_v2.vgic_eisr));
|
|
+ DEFINE(VGIC_V2_CPU_ELRSR, offsetof(struct vgic_cpu, vgic_v2.vgic_elrsr));
|
|
+ DEFINE(VGIC_V2_CPU_APR, offsetof(struct vgic_cpu, vgic_v2.vgic_apr));
|
|
+ DEFINE(VGIC_V2_CPU_LR, offsetof(struct vgic_cpu, vgic_v2.vgic_lr));
|
|
DEFINE(VGIC_CPU_NR_LR, offsetof(struct vgic_cpu, nr_lr));
|
|
#ifdef CONFIG_KVM_ARM_TIMER
|
|
DEFINE(VCPU_TIMER_CNTV_CTL, offsetof(struct kvm_vcpu, arch.timer_cpu.cntv_ctl));
|
|
diff --git a/arch/arm/kvm/Makefile b/arch/arm/kvm/Makefile
|
|
index 789bca9..f7057ed 100644
|
|
--- a/arch/arm/kvm/Makefile
|
|
+++ b/arch/arm/kvm/Makefile
|
|
@@ -21,4 +21,5 @@ obj-y += kvm-arm.o init.o interrupts.o
|
|
obj-y += arm.o handle_exit.o guest.o mmu.o emulate.o reset.o
|
|
obj-y += coproc.o coproc_a15.o coproc_a7.o mmio.o psci.o perf.o
|
|
obj-$(CONFIG_KVM_ARM_VGIC) += $(KVM)/arm/vgic.o
|
|
+obj-$(CONFIG_KVM_ARM_VGIC) += $(KVM)/arm/vgic-v2.o
|
|
obj-$(CONFIG_KVM_ARM_TIMER) += $(KVM)/arm/arch_timer.o
|
|
diff --git a/arch/arm/kvm/interrupts_head.S b/arch/arm/kvm/interrupts_head.S
|
|
index 76af9302..e4eaf30 100644
|
|
--- a/arch/arm/kvm/interrupts_head.S
|
|
+++ b/arch/arm/kvm/interrupts_head.S
|
|
@@ -421,14 +421,14 @@ vcpu .req r0 @ vcpu pointer always in r0
|
|
ldr r9, [r2, #GICH_ELRSR1]
|
|
ldr r10, [r2, #GICH_APR]
|
|
|
|
- str r3, [r11, #VGIC_CPU_HCR]
|
|
- str r4, [r11, #VGIC_CPU_VMCR]
|
|
- str r5, [r11, #VGIC_CPU_MISR]
|
|
- str r6, [r11, #VGIC_CPU_EISR]
|
|
- str r7, [r11, #(VGIC_CPU_EISR + 4)]
|
|
- str r8, [r11, #VGIC_CPU_ELRSR]
|
|
- str r9, [r11, #(VGIC_CPU_ELRSR + 4)]
|
|
- str r10, [r11, #VGIC_CPU_APR]
|
|
+ str r3, [r11, #VGIC_V2_CPU_HCR]
|
|
+ str r4, [r11, #VGIC_V2_CPU_VMCR]
|
|
+ str r5, [r11, #VGIC_V2_CPU_MISR]
|
|
+ str r6, [r11, #VGIC_V2_CPU_EISR]
|
|
+ str r7, [r11, #(VGIC_V2_CPU_EISR + 4)]
|
|
+ str r8, [r11, #VGIC_V2_CPU_ELRSR]
|
|
+ str r9, [r11, #(VGIC_V2_CPU_ELRSR + 4)]
|
|
+ str r10, [r11, #VGIC_V2_CPU_APR]
|
|
|
|
/* Clear GICH_HCR */
|
|
mov r5, #0
|
|
@@ -436,7 +436,7 @@ vcpu .req r0 @ vcpu pointer always in r0
|
|
|
|
/* Save list registers */
|
|
add r2, r2, #GICH_LR0
|
|
- add r3, r11, #VGIC_CPU_LR
|
|
+ add r3, r11, #VGIC_V2_CPU_LR
|
|
ldr r4, [r11, #VGIC_CPU_NR_LR]
|
|
1: ldr r6, [r2], #4
|
|
str r6, [r3], #4
|
|
@@ -463,9 +463,9 @@ vcpu .req r0 @ vcpu pointer always in r0
|
|
add r11, vcpu, #VCPU_VGIC_CPU
|
|
|
|
/* We only restore a minimal set of registers */
|
|
- ldr r3, [r11, #VGIC_CPU_HCR]
|
|
- ldr r4, [r11, #VGIC_CPU_VMCR]
|
|
- ldr r8, [r11, #VGIC_CPU_APR]
|
|
+ ldr r3, [r11, #VGIC_V2_CPU_HCR]
|
|
+ ldr r4, [r11, #VGIC_V2_CPU_VMCR]
|
|
+ ldr r8, [r11, #VGIC_V2_CPU_APR]
|
|
|
|
str r3, [r2, #GICH_HCR]
|
|
str r4, [r2, #GICH_VMCR]
|
|
@@ -473,7 +473,7 @@ vcpu .req r0 @ vcpu pointer always in r0
|
|
|
|
/* Restore list registers */
|
|
add r2, r2, #GICH_LR0
|
|
- add r3, r11, #VGIC_CPU_LR
|
|
+ add r3, r11, #VGIC_V2_CPU_LR
|
|
ldr r4, [r11, #VGIC_CPU_NR_LR]
|
|
1: ldr r6, [r3], #4
|
|
str r6, [r2], #4
|
|
diff --git a/arch/arm64/Kconfig b/arch/arm64/Kconfig
|
|
index b0f9c9d..be54ce2 100644
|
|
--- a/arch/arm64/Kconfig
|
|
+++ b/arch/arm64/Kconfig
|
|
@@ -80,7 +80,7 @@ config MMU
|
|
def_bool y
|
|
|
|
config NO_IOPORT_MAP
|
|
- def_bool y
|
|
+ def_bool y if !PCI
|
|
|
|
config STACKTRACE_SUPPORT
|
|
def_bool y
|
|
@@ -155,6 +155,23 @@ menu "Bus support"
|
|
config ARM_AMBA
|
|
bool
|
|
|
|
+config PCI
|
|
+ bool "PCI support"
|
|
+ help
|
|
+ This feature enables support for PCIe bus system. If you say Y
|
|
+ here, the kernel will include drivers and infrastructure code
|
|
+ to support PCIe bus devices.
|
|
+
|
|
+config PCI_DOMAINS
|
|
+ def_bool PCI
|
|
+
|
|
+config PCI_SYSCALL
|
|
+ def_bool PCI
|
|
+
|
|
+source "drivers/pci/Kconfig"
|
|
+source "drivers/pci/pcie/Kconfig"
|
|
+source "drivers/pci/hotplug/Kconfig"
|
|
+
|
|
endmenu
|
|
|
|
menu "Kernel Features"
|
|
diff --git a/arch/arm64/boot/dts/apm-mustang.dts b/arch/arm64/boot/dts/apm-mustang.dts
|
|
index 6541962..0cb67fc 100644
|
|
--- a/arch/arm64/boot/dts/apm-mustang.dts
|
|
+++ b/arch/arm64/boot/dts/apm-mustang.dts
|
|
@@ -28,3 +28,15 @@
|
|
&serial0 {
|
|
status = "ok";
|
|
};
|
|
+
|
|
+&menet {
|
|
+ status = "ok";
|
|
+};
|
|
+
|
|
+&pcie0clk {
|
|
+ status = "ok";
|
|
+};
|
|
+
|
|
+&pcie0 {
|
|
+ status = "ok";
|
|
+};
|
|
diff --git a/arch/arm64/boot/dts/apm-storm.dtsi b/arch/arm64/boot/dts/apm-storm.dtsi
|
|
index 40aa96c..fb2ee54 100644
|
|
--- a/arch/arm64/boot/dts/apm-storm.dtsi
|
|
+++ b/arch/arm64/boot/dts/apm-storm.dtsi
|
|
@@ -24,56 +24,56 @@
|
|
compatible = "apm,potenza", "arm,armv8";
|
|
reg = <0x0 0x000>;
|
|
enable-method = "spin-table";
|
|
- cpu-release-addr = <0x1 0x0000fff8>;
|
|
+ cpu-release-addr = <0x40 0x0000f000>;
|
|
};
|
|
cpu@001 {
|
|
device_type = "cpu";
|
|
compatible = "apm,potenza", "arm,armv8";
|
|
reg = <0x0 0x001>;
|
|
enable-method = "spin-table";
|
|
- cpu-release-addr = <0x1 0x0000fff8>;
|
|
+ cpu-release-addr = <0x40 0x0000f000>;
|
|
};
|
|
cpu@100 {
|
|
device_type = "cpu";
|
|
compatible = "apm,potenza", "arm,armv8";
|
|
reg = <0x0 0x100>;
|
|
enable-method = "spin-table";
|
|
- cpu-release-addr = <0x1 0x0000fff8>;
|
|
+ cpu-release-addr = <0x40 0x0000f000>;
|
|
};
|
|
cpu@101 {
|
|
device_type = "cpu";
|
|
compatible = "apm,potenza", "arm,armv8";
|
|
reg = <0x0 0x101>;
|
|
enable-method = "spin-table";
|
|
- cpu-release-addr = <0x1 0x0000fff8>;
|
|
+ cpu-release-addr = <0x40 0x0000f000>;
|
|
};
|
|
cpu@200 {
|
|
device_type = "cpu";
|
|
compatible = "apm,potenza", "arm,armv8";
|
|
reg = <0x0 0x200>;
|
|
enable-method = "spin-table";
|
|
- cpu-release-addr = <0x1 0x0000fff8>;
|
|
+ cpu-release-addr = <0x40 0x0000f000>;
|
|
};
|
|
cpu@201 {
|
|
device_type = "cpu";
|
|
compatible = "apm,potenza", "arm,armv8";
|
|
reg = <0x0 0x201>;
|
|
enable-method = "spin-table";
|
|
- cpu-release-addr = <0x1 0x0000fff8>;
|
|
+ cpu-release-addr = <0x40 0x0000f000>;
|
|
};
|
|
cpu@300 {
|
|
device_type = "cpu";
|
|
compatible = "apm,potenza", "arm,armv8";
|
|
reg = <0x0 0x300>;
|
|
enable-method = "spin-table";
|
|
- cpu-release-addr = <0x1 0x0000fff8>;
|
|
+ cpu-release-addr = <0x40 0x0000f000>;
|
|
};
|
|
cpu@301 {
|
|
device_type = "cpu";
|
|
compatible = "apm,potenza", "arm,armv8";
|
|
reg = <0x0 0x301>;
|
|
enable-method = "spin-table";
|
|
- cpu-release-addr = <0x1 0x0000fff8>;
|
|
+ cpu-release-addr = <0x40 0x0000f000>;
|
|
};
|
|
};
|
|
|
|
@@ -97,6 +97,11 @@
|
|
clock-frequency = <50000000>;
|
|
};
|
|
|
|
+ pmu {
|
|
+ compatible = "arm,armv8-pmuv3";
|
|
+ interrupts = <1 12 0xff04>;
|
|
+ };
|
|
+
|
|
soc {
|
|
compatible = "simple-bus";
|
|
#address-cells = <2>;
|
|
@@ -167,14 +172,13 @@
|
|
clock-output-names = "ethclk";
|
|
};
|
|
|
|
- eth8clk: eth8clk {
|
|
+ menetclk: menetclk {
|
|
compatible = "apm,xgene-device-clock";
|
|
#clock-cells = <1>;
|
|
clocks = <ðclk 0>;
|
|
- clock-names = "eth8clk";
|
|
reg = <0x0 0x1702C000 0x0 0x1000>;
|
|
reg-names = "csr-reg";
|
|
- clock-output-names = "eth8clk";
|
|
+ clock-output-names = "menetclk";
|
|
};
|
|
|
|
sataphy1clk: sataphy1clk@1f21c000 {
|
|
@@ -270,6 +274,161 @@
|
|
enable-mask = <0x2>;
|
|
clock-output-names = "rtcclk";
|
|
};
|
|
+
|
|
+ pcie0clk: pcie0clk@1f2bc000 {
|
|
+ status = "disabled";
|
|
+ compatible = "apm,xgene-device-clock";
|
|
+ #clock-cells = <1>;
|
|
+ clocks = <&socplldiv2 0>;
|
|
+ reg = <0x0 0x1f2bc000 0x0 0x1000>;
|
|
+ reg-names = "csr-reg";
|
|
+ clock-output-names = "pcie0clk";
|
|
+ };
|
|
+
|
|
+ pcie1clk: pcie1clk@1f2cc000 {
|
|
+ status = "disabled";
|
|
+ compatible = "apm,xgene-device-clock";
|
|
+ #clock-cells = <1>;
|
|
+ clocks = <&socplldiv2 0>;
|
|
+ reg = <0x0 0x1f2cc000 0x0 0x1000>;
|
|
+ reg-names = "csr-reg";
|
|
+ clock-output-names = "pcie1clk";
|
|
+ };
|
|
+
|
|
+ pcie2clk: pcie2clk@1f2dc000 {
|
|
+ status = "disabled";
|
|
+ compatible = "apm,xgene-device-clock";
|
|
+ #clock-cells = <1>;
|
|
+ clocks = <&socplldiv2 0>;
|
|
+ reg = <0x0 0x1f2dc000 0x0 0x1000>;
|
|
+ reg-names = "csr-reg";
|
|
+ clock-output-names = "pcie2clk";
|
|
+ };
|
|
+
|
|
+ pcie3clk: pcie3clk@1f50c000 {
|
|
+ status = "disabled";
|
|
+ compatible = "apm,xgene-device-clock";
|
|
+ #clock-cells = <1>;
|
|
+ clocks = <&socplldiv2 0>;
|
|
+ reg = <0x0 0x1f50c000 0x0 0x1000>;
|
|
+ reg-names = "csr-reg";
|
|
+ clock-output-names = "pcie3clk";
|
|
+ };
|
|
+
|
|
+ pcie4clk: pcie4clk@1f51c000 {
|
|
+ status = "disabled";
|
|
+ compatible = "apm,xgene-device-clock";
|
|
+ #clock-cells = <1>;
|
|
+ clocks = <&socplldiv2 0>;
|
|
+ reg = <0x0 0x1f51c000 0x0 0x1000>;
|
|
+ reg-names = "csr-reg";
|
|
+ clock-output-names = "pcie4clk";
|
|
+ };
|
|
+ };
|
|
+
|
|
+ pcie0: pcie@1f2b0000 {
|
|
+ status = "disabled";
|
|
+ device_type = "pci";
|
|
+ compatible = "apm,xgene-storm-pcie", "apm,xgene-pcie";
|
|
+ #interrupt-cells = <1>;
|
|
+ #size-cells = <2>;
|
|
+ #address-cells = <3>;
|
|
+ reg = < 0x00 0x1f2b0000 0x0 0x00010000 /* Controller registers */
|
|
+ 0xe0 0xd0000000 0x0 0x00200000>; /* PCI config space */
|
|
+ reg-names = "csr", "cfg";
|
|
+ ranges = <0x01000000 0x00 0x00000000 0xe0 0x00000000 0x00 0x00010000 /* io */
|
|
+ 0x02000000 0x00 0x10000000 0xe0 0x10000000 0x00 0x80000000>; /* mem */
|
|
+ dma-ranges = <0x42000000 0x40 0x00000000 0x40 0x00000000 0x40 0x00000000>;
|
|
+ interrupt-map-mask = <0x0 0x0 0x0 0x7>;
|
|
+ interrupt-map = <0x0 0x0 0x0 0x1 &gic 0x0 0xc2 0x1
|
|
+ 0x0 0x0 0x0 0x2 &gic 0x0 0xc3 0x1
|
|
+ 0x0 0x0 0x0 0x3 &gic 0x0 0xc4 0x1
|
|
+ 0x0 0x0 0x0 0x4 &gic 0x0 0xc5 0x1>;
|
|
+ clocks = <&pcie0clk 0>;
|
|
+ };
|
|
+
|
|
+ pcie1: pcie@1f2c0000 {
|
|
+ status = "disabled";
|
|
+ device_type = "pci";
|
|
+ compatible = "apm,xgene-storm-pcie", "apm,xgene-pcie";
|
|
+ #interrupt-cells = <1>;
|
|
+ #size-cells = <2>;
|
|
+ #address-cells = <3>;
|
|
+ reg = < 0x00 0x1f2c0000 0x0 0x00010000 /* Controller registers */
|
|
+ 0xd0 0xd0000000 0x0 0x00200000>; /* PCI config space */
|
|
+ reg-names = "csr", "cfg";
|
|
+ ranges = <0x01000000 0x0 0x00000000 0xd0 0x00000000 0x00 0x00010000 /* io */
|
|
+ 0x02000000 0x0 0x10000000 0xd0 0x10000000 0x00 0x80000000>; /* mem */
|
|
+ dma-ranges = <0x42000000 0x40 0x00000000 0x40 0x00000000 0x40 0x00000000>;
|
|
+ interrupt-map-mask = <0x0 0x0 0x0 0x7>;
|
|
+ interrupt-map = <0x0 0x0 0x0 0x1 &gic 0x0 0xc8 0x1
|
|
+ 0x0 0x0 0x0 0x2 &gic 0x0 0xc9 0x1
|
|
+ 0x0 0x0 0x0 0x3 &gic 0x0 0xca 0x1
|
|
+ 0x0 0x0 0x0 0x4 &gic 0x0 0xcb 0x1>;
|
|
+ clocks = <&pcie1clk 0>;
|
|
+ };
|
|
+
|
|
+ pcie2: pcie@1f2d0000 {
|
|
+ status = "disabled";
|
|
+ device_type = "pci";
|
|
+ compatible = "apm,xgene-storm-pcie", "apm,xgene-pcie";
|
|
+ #interrupt-cells = <1>;
|
|
+ #size-cells = <2>;
|
|
+ #address-cells = <3>;
|
|
+ reg = < 0x00 0x1f2d0000 0x0 0x00010000 /* Controller registers */
|
|
+ 0x90 0xd0000000 0x0 0x00200000>; /* PCI config space */
|
|
+ reg-names = "csr", "cfg";
|
|
+ ranges = <0x01000000 0x0 0x00000000 0x90 0x00000000 0x0 0x00010000 /* io */
|
|
+ 0x02000000 0x0 0x10000000 0x90 0x10000000 0x0 0x80000000>; /* mem */
|
|
+ dma-ranges = <0x42000000 0x40 0x00000000 0x40 0x00000000 0x40 0x00000000>;
|
|
+ interrupt-map-mask = <0x0 0x0 0x0 0x7>;
|
|
+ interrupt-map = <0x0 0x0 0x0 0x1 &gic 0x0 0xce 0x1
|
|
+ 0x0 0x0 0x0 0x2 &gic 0x0 0xcf 0x1
|
|
+ 0x0 0x0 0x0 0x3 &gic 0x0 0xd0 0x1
|
|
+ 0x0 0x0 0x0 0x4 &gic 0x0 0xd1 0x1>;
|
|
+ clocks = <&pcie2clk 0>;
|
|
+ };
|
|
+
|
|
+ pcie3: pcie@1f500000 {
|
|
+ status = "disabled";
|
|
+ device_type = "pci";
|
|
+ compatible = "apm,xgene-storm-pcie", "apm,xgene-pcie";
|
|
+ #interrupt-cells = <1>;
|
|
+ #size-cells = <2>;
|
|
+ #address-cells = <3>;
|
|
+ reg = < 0x00 0x1f500000 0x0 0x00010000 /* Controller registers */
|
|
+ 0xa0 0xd0000000 0x0 0x00200000>; /* PCI config space */
|
|
+ reg-names = "csr", "cfg";
|
|
+ ranges = <0x01000000 0x0 0x00000000 0xa0 0x00000000 0x0 0x00010000 /* io */
|
|
+ 0x02000000 0x0 0x10000000 0xa0 0x10000000 0x0 0x80000000>; /* mem */
|
|
+ dma-ranges = <0x42000000 0x40 0x00000000 0x40 0x00000000 0x40 0x00000000>;
|
|
+ interrupt-map-mask = <0x0 0x0 0x0 0x7>;
|
|
+ interrupt-map = <0x0 0x0 0x0 0x1 &gic 0x0 0xd4 0x1
|
|
+ 0x0 0x0 0x0 0x2 &gic 0x0 0xd5 0x1
|
|
+ 0x0 0x0 0x0 0x3 &gic 0x0 0xd6 0x1
|
|
+ 0x0 0x0 0x0 0x4 &gic 0x0 0xd7 0x1>;
|
|
+ clocks = <&pcie3clk 0>;
|
|
+ };
|
|
+
|
|
+ pcie4: pcie@1f510000 {
|
|
+ status = "disabled";
|
|
+ device_type = "pci";
|
|
+ compatible = "apm,xgene-storm-pcie", "apm,xgene-pcie";
|
|
+ #interrupt-cells = <1>;
|
|
+ #size-cells = <2>;
|
|
+ #address-cells = <3>;
|
|
+ reg = < 0x00 0x1f510000 0x0 0x00010000 /* Controller registers */
|
|
+ 0xc0 0xd0000000 0x0 0x00200000>; /* PCI config space */
|
|
+ reg-names = "csr", "cfg";
|
|
+ ranges = <0x01000000 0x0 0x00000000 0xc0 0x00000000 0x0 0x00010000 /* io */
|
|
+ 0x02000000 0x0 0x10000000 0xc0 0x10000000 0x0 0x80000000>; /* mem */
|
|
+ dma-ranges = <0x42000000 0x40 0x00000000 0x40 0x00000000 0x40 0x00000000>;
|
|
+ interrupt-map-mask = <0x0 0x0 0x0 0x7>;
|
|
+ interrupt-map = <0x0 0x0 0x0 0x1 &gic 0x0 0xda 0x1
|
|
+ 0x0 0x0 0x0 0x2 &gic 0x0 0xdb 0x1
|
|
+ 0x0 0x0 0x0 0x3 &gic 0x0 0xdc 0x1
|
|
+ 0x0 0x0 0x0 0x4 &gic 0x0 0xdd 0x1>;
|
|
+ clocks = <&pcie4clk 0>;
|
|
};
|
|
|
|
serial0: serial@1c020000 {
|
|
@@ -278,7 +437,7 @@
|
|
compatible = "ns16550a";
|
|
reg = <0 0x1c020000 0x0 0x1000>;
|
|
reg-shift = <2>;
|
|
- clock-frequency = <10000000>; /* Updated by bootloader */
|
|
+ clock-frequency = <50000000>; /* Updated by bootloader */
|
|
interrupt-parent = <&gic>;
|
|
interrupts = <0x0 0x4c 0x4>;
|
|
};
|
|
@@ -397,5 +556,30 @@
|
|
#clock-cells = <1>;
|
|
clocks = <&rtcclk 0>;
|
|
};
|
|
+
|
|
+ menet: ethernet@17020000 {
|
|
+ compatible = "apm,xgene-enet";
|
|
+ status = "disabled";
|
|
+ reg = <0x0 0x17020000 0x0 0xd100>,
|
|
+ <0x0 0X17030000 0x0 0X400>,
|
|
+ <0x0 0X10000000 0x0 0X200>;
|
|
+ reg-names = "enet_csr", "ring_csr", "ring_cmd";
|
|
+ interrupts = <0x0 0x3c 0x4>;
|
|
+ dma-coherent;
|
|
+ clocks = <&menetclk 0>;
|
|
+ local-mac-address = [00 00 00 00 00 00];
|
|
+ phy-connection-type = "rgmii";
|
|
+ phy-handle = <&menetphy>;
|
|
+ mdio {
|
|
+ compatible = "apm,xgene-mdio";
|
|
+ #address-cells = <1>;
|
|
+ #size-cells = <0>;
|
|
+ menetphy: menetphy@3 {
|
|
+ compatible = "ethernet-phy-id001c.c915";
|
|
+ reg = <0x3>;
|
|
+ };
|
|
+
|
|
+ };
|
|
+ };
|
|
};
|
|
};
|
|
diff --git a/arch/arm64/include/asm/Kbuild b/arch/arm64/include/asm/Kbuild
|
|
index 0b3fcf8..07cb417 100644
|
|
--- a/arch/arm64/include/asm/Kbuild
|
|
+++ b/arch/arm64/include/asm/Kbuild
|
|
@@ -29,6 +29,7 @@ generic-y += mman.h
|
|
generic-y += msgbuf.h
|
|
generic-y += mutex.h
|
|
generic-y += pci.h
|
|
+generic-y += pci-bridge.h
|
|
generic-y += poll.h
|
|
generic-y += preempt.h
|
|
generic-y += resource.h
|
|
diff --git a/arch/arm64/include/asm/elf.h b/arch/arm64/include/asm/elf.h
|
|
index 01d3aab..8186df6 100644
|
|
--- a/arch/arm64/include/asm/elf.h
|
|
+++ b/arch/arm64/include/asm/elf.h
|
|
@@ -114,7 +114,8 @@ typedef struct user_fpsimd_state elf_fpregset_t;
|
|
*/
|
|
#define elf_check_arch(x) ((x)->e_machine == EM_AARCH64)
|
|
|
|
-#define elf_read_implies_exec(ex,stk) (stk != EXSTACK_DISABLE_X)
|
|
+#define elf_read_implies_exec(ex,stk) (test_thread_flag(TIF_32BIT) \
|
|
+ ? (stk == EXSTACK_ENABLE_X) : 0)
|
|
|
|
#define CORE_DUMP_USE_REGSET
|
|
#define ELF_EXEC_PAGESIZE PAGE_SIZE
|
|
diff --git a/arch/arm64/include/asm/io.h b/arch/arm64/include/asm/io.h
|
|
index e0ecdcf..dc34039 100644
|
|
--- a/arch/arm64/include/asm/io.h
|
|
+++ b/arch/arm64/include/asm/io.h
|
|
@@ -121,7 +121,8 @@ static inline u64 __raw_readq(const volatile void __iomem *addr)
|
|
/*
|
|
* I/O port access primitives.
|
|
*/
|
|
-#define IO_SPACE_LIMIT 0xffff
|
|
+#define arch_has_dev_port() (1)
|
|
+#define IO_SPACE_LIMIT 0x1ffffff
|
|
#define PCI_IOBASE ((void __iomem *)(MODULES_VADDR - SZ_32M))
|
|
|
|
static inline u8 inb(unsigned long addr)
|
|
diff --git a/arch/arm64/include/asm/kvm_arm.h b/arch/arm64/include/asm/kvm_arm.h
|
|
index 3d69030..cc83520 100644
|
|
--- a/arch/arm64/include/asm/kvm_arm.h
|
|
+++ b/arch/arm64/include/asm/kvm_arm.h
|
|
@@ -76,9 +76,10 @@
|
|
*/
|
|
#define HCR_GUEST_FLAGS (HCR_TSC | HCR_TSW | HCR_TWE | HCR_TWI | HCR_VM | \
|
|
HCR_TVM | HCR_BSU_IS | HCR_FB | HCR_TAC | \
|
|
- HCR_AMO | HCR_IMO | HCR_FMO | \
|
|
- HCR_SWIO | HCR_TIDCP | HCR_RW)
|
|
+ HCR_AMO | HCR_SWIO | HCR_TIDCP | HCR_RW)
|
|
#define HCR_VIRT_EXCP_MASK (HCR_VA | HCR_VI | HCR_VF)
|
|
+#define HCR_INT_OVERRIDE (HCR_FMO | HCR_IMO)
|
|
+
|
|
|
|
/* Hyp System Control Register (SCTLR_EL2) bits */
|
|
#define SCTLR_EL2_EE (1 << 25)
|
|
diff --git a/arch/arm64/include/asm/kvm_asm.h b/arch/arm64/include/asm/kvm_asm.h
|
|
index 9fcd54b..a28c35b 100644
|
|
--- a/arch/arm64/include/asm/kvm_asm.h
|
|
+++ b/arch/arm64/include/asm/kvm_asm.h
|
|
@@ -18,6 +18,8 @@
|
|
#ifndef __ARM_KVM_ASM_H__
|
|
#define __ARM_KVM_ASM_H__
|
|
|
|
+#include <asm/virt.h>
|
|
+
|
|
/*
|
|
* 0 is reserved as an invalid value.
|
|
* Order *must* be kept in sync with the hyp switch code.
|
|
@@ -96,13 +98,21 @@ extern char __kvm_hyp_init_end[];
|
|
|
|
extern char __kvm_hyp_vector[];
|
|
|
|
-extern char __kvm_hyp_code_start[];
|
|
-extern char __kvm_hyp_code_end[];
|
|
+#define __kvm_hyp_code_start __hyp_text_start
|
|
+#define __kvm_hyp_code_end __hyp_text_end
|
|
|
|
extern void __kvm_flush_vm_context(void);
|
|
extern void __kvm_tlb_flush_vmid_ipa(struct kvm *kvm, phys_addr_t ipa);
|
|
|
|
extern int __kvm_vcpu_run(struct kvm_vcpu *vcpu);
|
|
+
|
|
+extern u64 __vgic_v3_get_ich_vtr_el2(void);
|
|
+
|
|
+extern char __save_vgic_v2_state[];
|
|
+extern char __restore_vgic_v2_state[];
|
|
+extern char __save_vgic_v3_state[];
|
|
+extern char __restore_vgic_v3_state[];
|
|
+
|
|
#endif
|
|
|
|
#endif /* __ARM_KVM_ASM_H__ */
|
|
diff --git a/arch/arm64/include/asm/kvm_host.h b/arch/arm64/include/asm/kvm_host.h
|
|
index 92242ce..4ae9213 100644
|
|
--- a/arch/arm64/include/asm/kvm_host.h
|
|
+++ b/arch/arm64/include/asm/kvm_host.h
|
|
@@ -200,4 +200,32 @@ static inline void __cpu_init_hyp_mode(phys_addr_t boot_pgd_ptr,
|
|
hyp_stack_ptr, vector_ptr);
|
|
}
|
|
|
|
+struct vgic_sr_vectors {
|
|
+ void *save_vgic;
|
|
+ void *restore_vgic;
|
|
+};
|
|
+
|
|
+static inline void vgic_arch_setup(const struct vgic_params *vgic)
|
|
+{
|
|
+ extern struct vgic_sr_vectors __vgic_sr_vectors;
|
|
+
|
|
+ switch(vgic->type)
|
|
+ {
|
|
+ case VGIC_V2:
|
|
+ __vgic_sr_vectors.save_vgic = __save_vgic_v2_state;
|
|
+ __vgic_sr_vectors.restore_vgic = __restore_vgic_v2_state;
|
|
+ break;
|
|
+
|
|
+#ifdef CONFIG_ARM_GIC_V3
|
|
+ case VGIC_V3:
|
|
+ __vgic_sr_vectors.save_vgic = __save_vgic_v3_state;
|
|
+ __vgic_sr_vectors.restore_vgic = __restore_vgic_v3_state;
|
|
+ break;
|
|
+#endif
|
|
+
|
|
+ default:
|
|
+ BUG();
|
|
+ }
|
|
+}
|
|
+
|
|
#endif /* __ARM64_KVM_HOST_H__ */
|
|
diff --git a/arch/arm64/include/asm/kvm_mmu.h b/arch/arm64/include/asm/kvm_mmu.h
|
|
index 7d29847..d7f77ff 100644
|
|
--- a/arch/arm64/include/asm/kvm_mmu.h
|
|
+++ b/arch/arm64/include/asm/kvm_mmu.h
|
|
@@ -122,8 +122,16 @@ static inline void kvm_set_s2pmd_writable(pmd_t *pmd)
|
|
}
|
|
|
|
#define kvm_pgd_addr_end(addr, end) pgd_addr_end(addr, end)
|
|
-#define kvm_pud_addr_end(addr, end) pud_addr_end(addr, end)
|
|
-#define kvm_pmd_addr_end(addr, end) pmd_addr_end(addr, end)
|
|
+
|
|
+#define kvm_pud_addr_end(addr, end) \
|
|
+({ unsigned long __boundary = ((addr) + PUD_SIZE) & PUD_MASK; \
|
|
+ (__boundary - 1 < (end) - 1)? __boundary: (end); \
|
|
+})
|
|
+
|
|
+#define kvm_pmd_addr_end(addr, end) \
|
|
+({ unsigned long __boundary = ((addr) + PMD_SIZE) & PMD_MASK; \
|
|
+ (__boundary - 1 < (end) - 1)? __boundary: (end); \
|
|
+})
|
|
|
|
struct kvm;
|
|
|
|
diff --git a/arch/arm64/include/asm/pci.h b/arch/arm64/include/asm/pci.h
|
|
new file mode 100644
|
|
index 0000000..3f7856e
|
|
--- /dev/null
|
|
+++ b/arch/arm64/include/asm/pci.h
|
|
@@ -0,0 +1,49 @@
|
|
+#ifndef __ASM_PCI_H
|
|
+#define __ASM_PCI_H
|
|
+#ifdef __KERNEL__
|
|
+
|
|
+#include <linux/types.h>
|
|
+#include <linux/slab.h>
|
|
+#include <linux/dma-mapping.h>
|
|
+
|
|
+#include <asm/io.h>
|
|
+#include <asm-generic/pci-bridge.h>
|
|
+#include <asm-generic/pci-dma-compat.h>
|
|
+
|
|
+#define PCIBIOS_MIN_IO 0x1000
|
|
+#define PCIBIOS_MIN_MEM 0
|
|
+
|
|
+struct pci_host_bridge *find_pci_host_bridge(struct pci_bus *bus);
|
|
+
|
|
+/*
|
|
+ * Set to 1 if the kernel should re-assign all PCI bus numbers
|
|
+ */
|
|
+#define pcibios_assign_all_busses() \
|
|
+ (pci_has_flag(PCI_REASSIGN_ALL_BUS))
|
|
+
|
|
+/*
|
|
+ * PCI address space differs from physical memory address space
|
|
+ */
|
|
+#define PCI_DMA_BUS_IS_PHYS (0)
|
|
+
|
|
+extern int isa_dma_bridge_buggy;
|
|
+
|
|
+#ifdef CONFIG_PCI
|
|
+static inline int pci_domain_nr(struct pci_bus *bus)
|
|
+{
|
|
+ struct pci_host_bridge *bridge = find_pci_host_bridge(bus);
|
|
+
|
|
+ if (bridge)
|
|
+ return bridge->domain_nr;
|
|
+
|
|
+ return 0;
|
|
+}
|
|
+
|
|
+static inline int pci_proc_domain(struct pci_bus *bus)
|
|
+{
|
|
+ return 1;
|
|
+}
|
|
+#endif /* CONFIG_PCI */
|
|
+
|
|
+#endif /* __KERNEL__ */
|
|
+#endif /* __ASM_PCI_H */
|
|
diff --git a/arch/arm64/include/asm/virt.h b/arch/arm64/include/asm/virt.h
|
|
index 215ad46..7a5df52 100644
|
|
--- a/arch/arm64/include/asm/virt.h
|
|
+++ b/arch/arm64/include/asm/virt.h
|
|
@@ -50,6 +50,10 @@ static inline bool is_hyp_mode_mismatched(void)
|
|
return __boot_cpu_mode[0] != __boot_cpu_mode[1];
|
|
}
|
|
|
|
+/* The section containing the hypervisor text */
|
|
+extern char __hyp_text_start[];
|
|
+extern char __hyp_text_end[];
|
|
+
|
|
#endif /* __ASSEMBLY__ */
|
|
|
|
#endif /* ! __ASM__VIRT_H */
|
|
diff --git a/arch/arm64/kernel/Makefile b/arch/arm64/kernel/Makefile
|
|
index df7ef87..1ed5a06 100644
|
|
--- a/arch/arm64/kernel/Makefile
|
|
+++ b/arch/arm64/kernel/Makefile
|
|
@@ -29,6 +29,7 @@ arm64-obj-$(CONFIG_ARM64_CPU_SUSPEND) += sleep.o suspend.o
|
|
arm64-obj-$(CONFIG_JUMP_LABEL) += jump_label.o
|
|
arm64-obj-$(CONFIG_KGDB) += kgdb.o
|
|
arm64-obj-$(CONFIG_EFI) += efi.o efi-stub.o efi-entry.o
|
|
+arm64-obj-$(CONFIG_PCI) += pci.o
|
|
|
|
obj-y += $(arm64-obj-y) vdso/
|
|
obj-m += $(arm64-obj-m)
|
|
diff --git a/arch/arm64/kernel/asm-offsets.c b/arch/arm64/kernel/asm-offsets.c
|
|
index 646f888..e74654c 100644
|
|
--- a/arch/arm64/kernel/asm-offsets.c
|
|
+++ b/arch/arm64/kernel/asm-offsets.c
|
|
@@ -129,13 +129,24 @@ int main(void)
|
|
DEFINE(KVM_TIMER_ENABLED, offsetof(struct kvm, arch.timer.enabled));
|
|
DEFINE(VCPU_KVM, offsetof(struct kvm_vcpu, kvm));
|
|
DEFINE(VCPU_VGIC_CPU, offsetof(struct kvm_vcpu, arch.vgic_cpu));
|
|
- DEFINE(VGIC_CPU_HCR, offsetof(struct vgic_cpu, vgic_hcr));
|
|
- DEFINE(VGIC_CPU_VMCR, offsetof(struct vgic_cpu, vgic_vmcr));
|
|
- DEFINE(VGIC_CPU_MISR, offsetof(struct vgic_cpu, vgic_misr));
|
|
- DEFINE(VGIC_CPU_EISR, offsetof(struct vgic_cpu, vgic_eisr));
|
|
- DEFINE(VGIC_CPU_ELRSR, offsetof(struct vgic_cpu, vgic_elrsr));
|
|
- DEFINE(VGIC_CPU_APR, offsetof(struct vgic_cpu, vgic_apr));
|
|
- DEFINE(VGIC_CPU_LR, offsetof(struct vgic_cpu, vgic_lr));
|
|
+ DEFINE(VGIC_SAVE_FN, offsetof(struct vgic_sr_vectors, save_vgic));
|
|
+ DEFINE(VGIC_RESTORE_FN, offsetof(struct vgic_sr_vectors, restore_vgic));
|
|
+ DEFINE(VGIC_SR_VECTOR_SZ, sizeof(struct vgic_sr_vectors));
|
|
+ DEFINE(VGIC_V2_CPU_HCR, offsetof(struct vgic_cpu, vgic_v2.vgic_hcr));
|
|
+ DEFINE(VGIC_V2_CPU_VMCR, offsetof(struct vgic_cpu, vgic_v2.vgic_vmcr));
|
|
+ DEFINE(VGIC_V2_CPU_MISR, offsetof(struct vgic_cpu, vgic_v2.vgic_misr));
|
|
+ DEFINE(VGIC_V2_CPU_EISR, offsetof(struct vgic_cpu, vgic_v2.vgic_eisr));
|
|
+ DEFINE(VGIC_V2_CPU_ELRSR, offsetof(struct vgic_cpu, vgic_v2.vgic_elrsr));
|
|
+ DEFINE(VGIC_V2_CPU_APR, offsetof(struct vgic_cpu, vgic_v2.vgic_apr));
|
|
+ DEFINE(VGIC_V2_CPU_LR, offsetof(struct vgic_cpu, vgic_v2.vgic_lr));
|
|
+ DEFINE(VGIC_V3_CPU_HCR, offsetof(struct vgic_cpu, vgic_v3.vgic_hcr));
|
|
+ DEFINE(VGIC_V3_CPU_VMCR, offsetof(struct vgic_cpu, vgic_v3.vgic_vmcr));
|
|
+ DEFINE(VGIC_V3_CPU_MISR, offsetof(struct vgic_cpu, vgic_v3.vgic_misr));
|
|
+ DEFINE(VGIC_V3_CPU_EISR, offsetof(struct vgic_cpu, vgic_v3.vgic_eisr));
|
|
+ DEFINE(VGIC_V3_CPU_ELRSR, offsetof(struct vgic_cpu, vgic_v3.vgic_elrsr));
|
|
+ DEFINE(VGIC_V3_CPU_AP0R, offsetof(struct vgic_cpu, vgic_v3.vgic_ap0r));
|
|
+ DEFINE(VGIC_V3_CPU_AP1R, offsetof(struct vgic_cpu, vgic_v3.vgic_ap1r));
|
|
+ DEFINE(VGIC_V3_CPU_LR, offsetof(struct vgic_cpu, vgic_v3.vgic_lr));
|
|
DEFINE(VGIC_CPU_NR_LR, offsetof(struct vgic_cpu, nr_lr));
|
|
DEFINE(KVM_VTTBR, offsetof(struct kvm, arch.vttbr));
|
|
DEFINE(KVM_VGIC_VCTRL, offsetof(struct kvm, arch.vgic.vctrl_base));
|
|
diff --git a/arch/arm64/kernel/efi-stub.c b/arch/arm64/kernel/efi-stub.c
|
|
index 1317fef..d27dd98 100644
|
|
--- a/arch/arm64/kernel/efi-stub.c
|
|
+++ b/arch/arm64/kernel/efi-stub.c
|
|
@@ -28,20 +28,16 @@ efi_status_t handle_kernel_image(efi_system_table_t *sys_table,
|
|
kernel_size = _edata - _text;
|
|
if (*image_addr != (dram_base + TEXT_OFFSET)) {
|
|
kernel_memsize = kernel_size + (_end - _edata);
|
|
- status = efi_relocate_kernel(sys_table, image_addr,
|
|
- kernel_size, kernel_memsize,
|
|
- dram_base + TEXT_OFFSET,
|
|
- PAGE_SIZE);
|
|
+ status = efi_low_alloc(sys_table, kernel_memsize + TEXT_OFFSET,
|
|
+ SZ_2M, reserve_addr);
|
|
if (status != EFI_SUCCESS) {
|
|
pr_efi_err(sys_table, "Failed to relocate kernel\n");
|
|
return status;
|
|
}
|
|
- if (*image_addr != (dram_base + TEXT_OFFSET)) {
|
|
- pr_efi_err(sys_table, "Failed to alloc kernel memory\n");
|
|
- efi_free(sys_table, kernel_memsize, *image_addr);
|
|
- return EFI_LOAD_ERROR;
|
|
- }
|
|
- *image_size = kernel_memsize;
|
|
+ memcpy((void *)*reserve_addr + TEXT_OFFSET, (void *)*image_addr,
|
|
+ kernel_size);
|
|
+ *image_addr = *reserve_addr + TEXT_OFFSET;
|
|
+ *reserve_size = kernel_memsize + TEXT_OFFSET;
|
|
}
|
|
|
|
|
|
diff --git a/arch/arm64/kernel/efi.c b/arch/arm64/kernel/efi.c
|
|
index e72f310..72ee260 100644
|
|
--- a/arch/arm64/kernel/efi.c
|
|
+++ b/arch/arm64/kernel/efi.c
|
|
@@ -475,3 +475,14 @@ err_unmap:
|
|
return -1;
|
|
}
|
|
early_initcall(arm64_enter_virtual_mode);
|
|
+
|
|
+/*
|
|
+ * If nothing else is handling pm_power_off, use EFI
|
|
+ *
|
|
+ * This is called from a late_initcall after other mechanisms
|
|
+ * have had a chance to register a handler.
|
|
+ */
|
|
+bool efi_poweroff_required(void)
|
|
+{
|
|
+ return pm_power_off == NULL;
|
|
+}
|
|
diff --git a/arch/arm64/kernel/head.S b/arch/arm64/kernel/head.S
|
|
index 144f105..b6ca95a 100644
|
|
--- a/arch/arm64/kernel/head.S
|
|
+++ b/arch/arm64/kernel/head.S
|
|
@@ -151,7 +151,7 @@ optional_header:
|
|
.short 0x20b // PE32+ format
|
|
.byte 0x02 // MajorLinkerVersion
|
|
.byte 0x14 // MinorLinkerVersion
|
|
- .long _edata - stext // SizeOfCode
|
|
+ .long _end - stext // SizeOfCode
|
|
.long 0 // SizeOfInitializedData
|
|
.long 0 // SizeOfUninitializedData
|
|
.long efi_stub_entry - efi_head // AddressOfEntryPoint
|
|
@@ -169,7 +169,7 @@ extra_header_fields:
|
|
.short 0 // MinorSubsystemVersion
|
|
.long 0 // Win32VersionValue
|
|
|
|
- .long _edata - efi_head // SizeOfImage
|
|
+ .long _end - efi_head // SizeOfImage
|
|
|
|
// Everything before the kernel image is considered part of the header
|
|
.long stext - efi_head // SizeOfHeaders
|
|
@@ -216,7 +216,7 @@ section_table:
|
|
.byte 0
|
|
.byte 0
|
|
.byte 0 // end of 0 padding of section name
|
|
- .long _edata - stext // VirtualSize
|
|
+ .long _end - stext // VirtualSize
|
|
.long stext - efi_head // VirtualAddress
|
|
.long _edata - stext // SizeOfRawData
|
|
.long stext - efi_head // PointerToRawData
|
|
diff --git a/arch/arm64/kernel/pci.c b/arch/arm64/kernel/pci.c
|
|
new file mode 100644
|
|
index 0000000..955d6d1
|
|
--- /dev/null
|
|
+++ b/arch/arm64/kernel/pci.c
|
|
@@ -0,0 +1,38 @@
|
|
+/*
|
|
+ * Code borrowed from powerpc/kernel/pci-common.c
|
|
+ *
|
|
+ * Copyright (C) 2003 Anton Blanchard <anton@au.ibm.com>, IBM
|
|
+ * Copyright (C) 2014 ARM Ltd.
|
|
+ *
|
|
+ * This program is free software; you can redistribute it and/or
|
|
+ * modify it under the terms of the GNU General Public License
|
|
+ * version 2 as published by the Free Software Foundation.
|
|
+ *
|
|
+ */
|
|
+
|
|
+#include <linux/init.h>
|
|
+#include <linux/io.h>
|
|
+#include <linux/kernel.h>
|
|
+#include <linux/mm.h>
|
|
+#include <linux/of_pci.h>
|
|
+#include <linux/of_platform.h>
|
|
+#include <linux/slab.h>
|
|
+
|
|
+#include <asm/pci-bridge.h>
|
|
+
|
|
+/*
|
|
+ * Called after each bus is probed, but before its children are examined
|
|
+ */
|
|
+void pcibios_fixup_bus(struct pci_bus *bus)
|
|
+{
|
|
+ /* nothing to do, expected to be removed in the future */
|
|
+}
|
|
+
|
|
+/*
|
|
+ * We don't have to worry about legacy ISA devices, so nothing to do here
|
|
+ */
|
|
+resource_size_t pcibios_align_resource(void *data, const struct resource *res,
|
|
+ resource_size_t size, resource_size_t align)
|
|
+{
|
|
+ return res->start;
|
|
+}
|
|
diff --git a/arch/arm64/kernel/process.c b/arch/arm64/kernel/process.c
|
|
index 1309d64..8ff5208 100644
|
|
--- a/arch/arm64/kernel/process.c
|
|
+++ b/arch/arm64/kernel/process.c
|
|
@@ -43,6 +43,7 @@
|
|
#include <linux/hw_breakpoint.h>
|
|
#include <linux/personality.h>
|
|
#include <linux/notifier.h>
|
|
+#include <linux/efi.h>
|
|
|
|
#include <asm/compat.h>
|
|
#include <asm/cacheflush.h>
|
|
@@ -182,6 +183,11 @@ void machine_restart(char *cmd)
|
|
arm_pm_restart(reboot_mode, cmd);
|
|
|
|
/*
|
|
+ * If all else fails, try EFI
|
|
+ */
|
|
+ efi_reboot(reboot_mode, cmd);
|
|
+
|
|
+ /*
|
|
* Whoops - the architecture was unable to reboot.
|
|
*/
|
|
printk("Reboot failed -- System halted\n");
|
|
diff --git a/arch/arm64/kernel/smp_spin_table.c b/arch/arm64/kernel/smp_spin_table.c
|
|
index 0347d38..70181c1 100644
|
|
--- a/arch/arm64/kernel/smp_spin_table.c
|
|
+++ b/arch/arm64/kernel/smp_spin_table.c
|
|
@@ -20,6 +20,7 @@
|
|
#include <linux/init.h>
|
|
#include <linux/of.h>
|
|
#include <linux/smp.h>
|
|
+#include <linux/types.h>
|
|
|
|
#include <asm/cacheflush.h>
|
|
#include <asm/cpu_ops.h>
|
|
@@ -65,12 +66,21 @@ static int smp_spin_table_cpu_init(struct device_node *dn, unsigned int cpu)
|
|
|
|
static int smp_spin_table_cpu_prepare(unsigned int cpu)
|
|
{
|
|
- void **release_addr;
|
|
+ __le64 __iomem *release_addr;
|
|
|
|
if (!cpu_release_addr[cpu])
|
|
return -ENODEV;
|
|
|
|
- release_addr = __va(cpu_release_addr[cpu]);
|
|
+ /*
|
|
+ * The cpu-release-addr may or may not be inside the linear mapping.
|
|
+ * As ioremap_cache will either give us a new mapping or reuse the
|
|
+ * existing linear mapping, we can use it to cover both cases. In
|
|
+ * either case the memory will be MT_NORMAL.
|
|
+ */
|
|
+ release_addr = ioremap_cache(cpu_release_addr[cpu],
|
|
+ sizeof(*release_addr));
|
|
+ if (!release_addr)
|
|
+ return -ENOMEM;
|
|
|
|
/*
|
|
* We write the release address as LE regardless of the native
|
|
@@ -79,15 +89,16 @@ static int smp_spin_table_cpu_prepare(unsigned int cpu)
|
|
* boot-loader's endianess before jumping. This is mandated by
|
|
* the boot protocol.
|
|
*/
|
|
- release_addr[0] = (void *) cpu_to_le64(__pa(secondary_holding_pen));
|
|
-
|
|
- __flush_dcache_area(release_addr, sizeof(release_addr[0]));
|
|
+ writeq_relaxed(__pa(secondary_holding_pen), release_addr);
|
|
+ __flush_dcache_area(release_addr, sizeof(*release_addr));
|
|
|
|
/*
|
|
* Send an event to wake up the secondary CPU.
|
|
*/
|
|
sev();
|
|
|
|
+ iounmap(release_addr);
|
|
+
|
|
return 0;
|
|
}
|
|
|
|
diff --git a/arch/arm64/kvm/Makefile b/arch/arm64/kvm/Makefile
|
|
index 72a9fd5..32a0961 100644
|
|
--- a/arch/arm64/kvm/Makefile
|
|
+++ b/arch/arm64/kvm/Makefile
|
|
@@ -20,4 +20,8 @@ kvm-$(CONFIG_KVM_ARM_HOST) += hyp.o hyp-init.o handle_exit.o
|
|
kvm-$(CONFIG_KVM_ARM_HOST) += guest.o reset.o sys_regs.o sys_regs_generic_v8.o
|
|
|
|
kvm-$(CONFIG_KVM_ARM_VGIC) += $(KVM)/arm/vgic.o
|
|
+kvm-$(CONFIG_KVM_ARM_VGIC) += $(KVM)/arm/vgic-v2.o
|
|
+kvm-$(CONFIG_KVM_ARM_VGIC) += vgic-v2-switch.o
|
|
+kvm-$(CONFIG_KVM_ARM_VGIC) += $(KVM)/arm/vgic-v3.o
|
|
+kvm-$(CONFIG_KVM_ARM_VGIC) += vgic-v3-switch.o
|
|
kvm-$(CONFIG_KVM_ARM_TIMER) += $(KVM)/arm/arch_timer.o
|
|
diff --git a/arch/arm64/kvm/hyp.S b/arch/arm64/kvm/hyp.S
|
|
index b0d1512..5945f3b 100644
|
|
--- a/arch/arm64/kvm/hyp.S
|
|
+++ b/arch/arm64/kvm/hyp.S
|
|
@@ -16,7 +16,6 @@
|
|
*/
|
|
|
|
#include <linux/linkage.h>
|
|
-#include <linux/irqchip/arm-gic.h>
|
|
|
|
#include <asm/assembler.h>
|
|
#include <asm/memory.h>
|
|
@@ -36,9 +35,6 @@
|
|
.pushsection .hyp.text, "ax"
|
|
.align PAGE_SHIFT
|
|
|
|
-__kvm_hyp_code_start:
|
|
- .globl __kvm_hyp_code_start
|
|
-
|
|
.macro save_common_regs
|
|
// x2: base address for cpu context
|
|
// x3: tmp register
|
|
@@ -339,11 +335,8 @@ __kvm_hyp_code_start:
|
|
.endm
|
|
|
|
.macro activate_traps
|
|
- ldr x2, [x0, #VCPU_IRQ_LINES]
|
|
- ldr x1, [x0, #VCPU_HCR_EL2]
|
|
- orr x2, x2, x1
|
|
- msr hcr_el2, x2
|
|
-
|
|
+ ldr x2, [x0, #VCPU_HCR_EL2]
|
|
+ msr hcr_el2, x2
|
|
ldr x2, =(CPTR_EL2_TTA)
|
|
msr cptr_el2, x2
|
|
|
|
@@ -379,100 +372,33 @@ __kvm_hyp_code_start:
|
|
.endm
|
|
|
|
/*
|
|
- * Save the VGIC CPU state into memory
|
|
- * x0: Register pointing to VCPU struct
|
|
- * Do not corrupt x1!!!
|
|
+ * Call into the vgic backend for state saving
|
|
*/
|
|
.macro save_vgic_state
|
|
- /* Get VGIC VCTRL base into x2 */
|
|
- ldr x2, [x0, #VCPU_KVM]
|
|
- kern_hyp_va x2
|
|
- ldr x2, [x2, #KVM_VGIC_VCTRL]
|
|
- kern_hyp_va x2
|
|
- cbz x2, 2f // disabled
|
|
-
|
|
- /* Compute the address of struct vgic_cpu */
|
|
- add x3, x0, #VCPU_VGIC_CPU
|
|
-
|
|
- /* Save all interesting registers */
|
|
- ldr w4, [x2, #GICH_HCR]
|
|
- ldr w5, [x2, #GICH_VMCR]
|
|
- ldr w6, [x2, #GICH_MISR]
|
|
- ldr w7, [x2, #GICH_EISR0]
|
|
- ldr w8, [x2, #GICH_EISR1]
|
|
- ldr w9, [x2, #GICH_ELRSR0]
|
|
- ldr w10, [x2, #GICH_ELRSR1]
|
|
- ldr w11, [x2, #GICH_APR]
|
|
-CPU_BE( rev w4, w4 )
|
|
-CPU_BE( rev w5, w5 )
|
|
-CPU_BE( rev w6, w6 )
|
|
-CPU_BE( rev w7, w7 )
|
|
-CPU_BE( rev w8, w8 )
|
|
-CPU_BE( rev w9, w9 )
|
|
-CPU_BE( rev w10, w10 )
|
|
-CPU_BE( rev w11, w11 )
|
|
-
|
|
- str w4, [x3, #VGIC_CPU_HCR]
|
|
- str w5, [x3, #VGIC_CPU_VMCR]
|
|
- str w6, [x3, #VGIC_CPU_MISR]
|
|
- str w7, [x3, #VGIC_CPU_EISR]
|
|
- str w8, [x3, #(VGIC_CPU_EISR + 4)]
|
|
- str w9, [x3, #VGIC_CPU_ELRSR]
|
|
- str w10, [x3, #(VGIC_CPU_ELRSR + 4)]
|
|
- str w11, [x3, #VGIC_CPU_APR]
|
|
-
|
|
- /* Clear GICH_HCR */
|
|
- str wzr, [x2, #GICH_HCR]
|
|
-
|
|
- /* Save list registers */
|
|
- add x2, x2, #GICH_LR0
|
|
- ldr w4, [x3, #VGIC_CPU_NR_LR]
|
|
- add x3, x3, #VGIC_CPU_LR
|
|
-1: ldr w5, [x2], #4
|
|
-CPU_BE( rev w5, w5 )
|
|
- str w5, [x3], #4
|
|
- sub w4, w4, #1
|
|
- cbnz w4, 1b
|
|
-2:
|
|
+ adr x24, __vgic_sr_vectors
|
|
+ ldr x24, [x24, VGIC_SAVE_FN]
|
|
+ kern_hyp_va x24
|
|
+ blr x24
|
|
+ mrs x24, hcr_el2
|
|
+ mov x25, #HCR_INT_OVERRIDE
|
|
+ neg x25, x25
|
|
+ and x24, x24, x25
|
|
+ msr hcr_el2, x24
|
|
.endm
|
|
|
|
/*
|
|
- * Restore the VGIC CPU state from memory
|
|
- * x0: Register pointing to VCPU struct
|
|
+ * Call into the vgic backend for state restoring
|
|
*/
|
|
.macro restore_vgic_state
|
|
- /* Get VGIC VCTRL base into x2 */
|
|
- ldr x2, [x0, #VCPU_KVM]
|
|
- kern_hyp_va x2
|
|
- ldr x2, [x2, #KVM_VGIC_VCTRL]
|
|
- kern_hyp_va x2
|
|
- cbz x2, 2f // disabled
|
|
-
|
|
- /* Compute the address of struct vgic_cpu */
|
|
- add x3, x0, #VCPU_VGIC_CPU
|
|
-
|
|
- /* We only restore a minimal set of registers */
|
|
- ldr w4, [x3, #VGIC_CPU_HCR]
|
|
- ldr w5, [x3, #VGIC_CPU_VMCR]
|
|
- ldr w6, [x3, #VGIC_CPU_APR]
|
|
-CPU_BE( rev w4, w4 )
|
|
-CPU_BE( rev w5, w5 )
|
|
-CPU_BE( rev w6, w6 )
|
|
-
|
|
- str w4, [x2, #GICH_HCR]
|
|
- str w5, [x2, #GICH_VMCR]
|
|
- str w6, [x2, #GICH_APR]
|
|
-
|
|
- /* Restore list registers */
|
|
- add x2, x2, #GICH_LR0
|
|
- ldr w4, [x3, #VGIC_CPU_NR_LR]
|
|
- add x3, x3, #VGIC_CPU_LR
|
|
-1: ldr w5, [x3], #4
|
|
-CPU_BE( rev w5, w5 )
|
|
- str w5, [x2], #4
|
|
- sub w4, w4, #1
|
|
- cbnz w4, 1b
|
|
-2:
|
|
+ mrs x24, hcr_el2
|
|
+ ldr x25, [x0, #VCPU_IRQ_LINES]
|
|
+ orr x24, x24, #HCR_INT_OVERRIDE
|
|
+ orr x24, x24, x25
|
|
+ msr hcr_el2, x24
|
|
+ adr x24, __vgic_sr_vectors
|
|
+ ldr x24, [x24, #VGIC_RESTORE_FN]
|
|
+ kern_hyp_va x24
|
|
+ blr x24
|
|
.endm
|
|
|
|
.macro save_timer_state
|
|
@@ -653,6 +579,12 @@ ENTRY(__kvm_flush_vm_context)
|
|
ret
|
|
ENDPROC(__kvm_flush_vm_context)
|
|
|
|
+ // struct vgic_sr_vectors __vgi_sr_vectors;
|
|
+ .align 3
|
|
+ENTRY(__vgic_sr_vectors)
|
|
+ .skip VGIC_SR_VECTOR_SZ
|
|
+ENDPROC(__vgic_sr_vectors)
|
|
+
|
|
__kvm_hyp_panic:
|
|
// Guess the context by looking at VTTBR:
|
|
// If zero, then we're already a host.
|
|
@@ -880,7 +812,4 @@ ENTRY(__kvm_hyp_vector)
|
|
ventry el1_error_invalid // Error 32-bit EL1
|
|
ENDPROC(__kvm_hyp_vector)
|
|
|
|
-__kvm_hyp_code_end:
|
|
- .globl __kvm_hyp_code_end
|
|
-
|
|
.popsection
|
|
diff --git a/arch/arm64/kvm/vgic-v2-switch.S b/arch/arm64/kvm/vgic-v2-switch.S
|
|
new file mode 100644
|
|
index 0000000..ae21177
|
|
--- /dev/null
|
|
+++ b/arch/arm64/kvm/vgic-v2-switch.S
|
|
@@ -0,0 +1,133 @@
|
|
+/*
|
|
+ * Copyright (C) 2012,2013 - ARM Ltd
|
|
+ * Author: Marc Zyngier <marc.zyngier@arm.com>
|
|
+ *
|
|
+ * This program is free software; you can redistribute it and/or modify
|
|
+ * it under the terms of the GNU General Public License version 2 as
|
|
+ * published by the Free Software Foundation.
|
|
+ *
|
|
+ * This program is distributed in the hope that it will be useful,
|
|
+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
+ * GNU General Public License for more details.
|
|
+ *
|
|
+ * You should have received a copy of the GNU General Public License
|
|
+ * along with this program. If not, see <http://www.gnu.org/licenses/>.
|
|
+ */
|
|
+
|
|
+#include <linux/linkage.h>
|
|
+#include <linux/irqchip/arm-gic.h>
|
|
+
|
|
+#include <asm/assembler.h>
|
|
+#include <asm/memory.h>
|
|
+#include <asm/asm-offsets.h>
|
|
+#include <asm/kvm.h>
|
|
+#include <asm/kvm_asm.h>
|
|
+#include <asm/kvm_arm.h>
|
|
+#include <asm/kvm_mmu.h>
|
|
+
|
|
+ .text
|
|
+ .pushsection .hyp.text, "ax"
|
|
+
|
|
+/*
|
|
+ * Save the VGIC CPU state into memory
|
|
+ * x0: Register pointing to VCPU struct
|
|
+ * Do not corrupt x1!!!
|
|
+ */
|
|
+ENTRY(__save_vgic_v2_state)
|
|
+__save_vgic_v2_state:
|
|
+ /* Get VGIC VCTRL base into x2 */
|
|
+ ldr x2, [x0, #VCPU_KVM]
|
|
+ kern_hyp_va x2
|
|
+ ldr x2, [x2, #KVM_VGIC_VCTRL]
|
|
+ kern_hyp_va x2
|
|
+ cbz x2, 2f // disabled
|
|
+
|
|
+ /* Compute the address of struct vgic_cpu */
|
|
+ add x3, x0, #VCPU_VGIC_CPU
|
|
+
|
|
+ /* Save all interesting registers */
|
|
+ ldr w4, [x2, #GICH_HCR]
|
|
+ ldr w5, [x2, #GICH_VMCR]
|
|
+ ldr w6, [x2, #GICH_MISR]
|
|
+ ldr w7, [x2, #GICH_EISR0]
|
|
+ ldr w8, [x2, #GICH_EISR1]
|
|
+ ldr w9, [x2, #GICH_ELRSR0]
|
|
+ ldr w10, [x2, #GICH_ELRSR1]
|
|
+ ldr w11, [x2, #GICH_APR]
|
|
+CPU_BE( rev w4, w4 )
|
|
+CPU_BE( rev w5, w5 )
|
|
+CPU_BE( rev w6, w6 )
|
|
+CPU_BE( rev w7, w7 )
|
|
+CPU_BE( rev w8, w8 )
|
|
+CPU_BE( rev w9, w9 )
|
|
+CPU_BE( rev w10, w10 )
|
|
+CPU_BE( rev w11, w11 )
|
|
+
|
|
+ str w4, [x3, #VGIC_V2_CPU_HCR]
|
|
+ str w5, [x3, #VGIC_V2_CPU_VMCR]
|
|
+ str w6, [x3, #VGIC_V2_CPU_MISR]
|
|
+ str w7, [x3, #VGIC_V2_CPU_EISR]
|
|
+ str w8, [x3, #(VGIC_V2_CPU_EISR + 4)]
|
|
+ str w9, [x3, #VGIC_V2_CPU_ELRSR]
|
|
+ str w10, [x3, #(VGIC_V2_CPU_ELRSR + 4)]
|
|
+ str w11, [x3, #VGIC_V2_CPU_APR]
|
|
+
|
|
+ /* Clear GICH_HCR */
|
|
+ str wzr, [x2, #GICH_HCR]
|
|
+
|
|
+ /* Save list registers */
|
|
+ add x2, x2, #GICH_LR0
|
|
+ ldr w4, [x3, #VGIC_CPU_NR_LR]
|
|
+ add x3, x3, #VGIC_V2_CPU_LR
|
|
+1: ldr w5, [x2], #4
|
|
+CPU_BE( rev w5, w5 )
|
|
+ str w5, [x3], #4
|
|
+ sub w4, w4, #1
|
|
+ cbnz w4, 1b
|
|
+2:
|
|
+ ret
|
|
+ENDPROC(__save_vgic_v2_state)
|
|
+
|
|
+/*
|
|
+ * Restore the VGIC CPU state from memory
|
|
+ * x0: Register pointing to VCPU struct
|
|
+ */
|
|
+ENTRY(__restore_vgic_v2_state)
|
|
+__restore_vgic_v2_state:
|
|
+ /* Get VGIC VCTRL base into x2 */
|
|
+ ldr x2, [x0, #VCPU_KVM]
|
|
+ kern_hyp_va x2
|
|
+ ldr x2, [x2, #KVM_VGIC_VCTRL]
|
|
+ kern_hyp_va x2
|
|
+ cbz x2, 2f // disabled
|
|
+
|
|
+ /* Compute the address of struct vgic_cpu */
|
|
+ add x3, x0, #VCPU_VGIC_CPU
|
|
+
|
|
+ /* We only restore a minimal set of registers */
|
|
+ ldr w4, [x3, #VGIC_V2_CPU_HCR]
|
|
+ ldr w5, [x3, #VGIC_V2_CPU_VMCR]
|
|
+ ldr w6, [x3, #VGIC_V2_CPU_APR]
|
|
+CPU_BE( rev w4, w4 )
|
|
+CPU_BE( rev w5, w5 )
|
|
+CPU_BE( rev w6, w6 )
|
|
+
|
|
+ str w4, [x2, #GICH_HCR]
|
|
+ str w5, [x2, #GICH_VMCR]
|
|
+ str w6, [x2, #GICH_APR]
|
|
+
|
|
+ /* Restore list registers */
|
|
+ add x2, x2, #GICH_LR0
|
|
+ ldr w4, [x3, #VGIC_CPU_NR_LR]
|
|
+ add x3, x3, #VGIC_V2_CPU_LR
|
|
+1: ldr w5, [x3], #4
|
|
+CPU_BE( rev w5, w5 )
|
|
+ str w5, [x2], #4
|
|
+ sub w4, w4, #1
|
|
+ cbnz w4, 1b
|
|
+2:
|
|
+ ret
|
|
+ENDPROC(__restore_vgic_v2_state)
|
|
+
|
|
+ .popsection
|
|
diff --git a/arch/arm64/kvm/vgic-v3-switch.S b/arch/arm64/kvm/vgic-v3-switch.S
|
|
new file mode 100644
|
|
index 0000000..d160469
|
|
--- /dev/null
|
|
+++ b/arch/arm64/kvm/vgic-v3-switch.S
|
|
@@ -0,0 +1,267 @@
|
|
+/*
|
|
+ * Copyright (C) 2012,2013 - ARM Ltd
|
|
+ * Author: Marc Zyngier <marc.zyngier@arm.com>
|
|
+ *
|
|
+ * This program is free software; you can redistribute it and/or modify
|
|
+ * it under the terms of the GNU General Public License version 2 as
|
|
+ * published by the Free Software Foundation.
|
|
+ *
|
|
+ * This program is distributed in the hope that it will be useful,
|
|
+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
+ * GNU General Public License for more details.
|
|
+ *
|
|
+ * You should have received a copy of the GNU General Public License
|
|
+ * along with this program. If not, see <http://www.gnu.org/licenses/>.
|
|
+ */
|
|
+
|
|
+#include <linux/linkage.h>
|
|
+#include <linux/irqchip/arm-gic-v3.h>
|
|
+
|
|
+#include <asm/assembler.h>
|
|
+#include <asm/memory.h>
|
|
+#include <asm/asm-offsets.h>
|
|
+#include <asm/kvm.h>
|
|
+#include <asm/kvm_asm.h>
|
|
+#include <asm/kvm_arm.h>
|
|
+
|
|
+ .text
|
|
+ .pushsection .hyp.text, "ax"
|
|
+
|
|
+/*
|
|
+ * We store LRs in reverse order to let the CPU deal with streaming
|
|
+ * access. Use this macro to make it look saner...
|
|
+ */
|
|
+#define LR_OFFSET(n) (VGIC_V3_CPU_LR + (15 - n) * 8)
|
|
+
|
|
+/*
|
|
+ * Save the VGIC CPU state into memory
|
|
+ * x0: Register pointing to VCPU struct
|
|
+ * Do not corrupt x1!!!
|
|
+ */
|
|
+.macro save_vgic_v3_state
|
|
+ // Compute the address of struct vgic_cpu
|
|
+ add x3, x0, #VCPU_VGIC_CPU
|
|
+
|
|
+ // Make sure stores to the GIC via the memory mapped interface
|
|
+ // are now visible to the system register interface
|
|
+ dsb st
|
|
+
|
|
+ // Save all interesting registers
|
|
+ mrs_s x4, ICH_HCR_EL2
|
|
+ mrs_s x5, ICH_VMCR_EL2
|
|
+ mrs_s x6, ICH_MISR_EL2
|
|
+ mrs_s x7, ICH_EISR_EL2
|
|
+ mrs_s x8, ICH_ELSR_EL2
|
|
+
|
|
+ str w4, [x3, #VGIC_V3_CPU_HCR]
|
|
+ str w5, [x3, #VGIC_V3_CPU_VMCR]
|
|
+ str w6, [x3, #VGIC_V3_CPU_MISR]
|
|
+ str w7, [x3, #VGIC_V3_CPU_EISR]
|
|
+ str w8, [x3, #VGIC_V3_CPU_ELRSR]
|
|
+
|
|
+ msr_s ICH_HCR_EL2, xzr
|
|
+
|
|
+ mrs_s x21, ICH_VTR_EL2
|
|
+ mvn w22, w21
|
|
+ ubfiz w23, w22, 2, 4 // w23 = (15 - ListRegs) * 4
|
|
+
|
|
+ adr x24, 1f
|
|
+ add x24, x24, x23
|
|
+ br x24
|
|
+
|
|
+1:
|
|
+ mrs_s x20, ICH_LR15_EL2
|
|
+ mrs_s x19, ICH_LR14_EL2
|
|
+ mrs_s x18, ICH_LR13_EL2
|
|
+ mrs_s x17, ICH_LR12_EL2
|
|
+ mrs_s x16, ICH_LR11_EL2
|
|
+ mrs_s x15, ICH_LR10_EL2
|
|
+ mrs_s x14, ICH_LR9_EL2
|
|
+ mrs_s x13, ICH_LR8_EL2
|
|
+ mrs_s x12, ICH_LR7_EL2
|
|
+ mrs_s x11, ICH_LR6_EL2
|
|
+ mrs_s x10, ICH_LR5_EL2
|
|
+ mrs_s x9, ICH_LR4_EL2
|
|
+ mrs_s x8, ICH_LR3_EL2
|
|
+ mrs_s x7, ICH_LR2_EL2
|
|
+ mrs_s x6, ICH_LR1_EL2
|
|
+ mrs_s x5, ICH_LR0_EL2
|
|
+
|
|
+ adr x24, 1f
|
|
+ add x24, x24, x23
|
|
+ br x24
|
|
+
|
|
+1:
|
|
+ str x20, [x3, #LR_OFFSET(15)]
|
|
+ str x19, [x3, #LR_OFFSET(14)]
|
|
+ str x18, [x3, #LR_OFFSET(13)]
|
|
+ str x17, [x3, #LR_OFFSET(12)]
|
|
+ str x16, [x3, #LR_OFFSET(11)]
|
|
+ str x15, [x3, #LR_OFFSET(10)]
|
|
+ str x14, [x3, #LR_OFFSET(9)]
|
|
+ str x13, [x3, #LR_OFFSET(8)]
|
|
+ str x12, [x3, #LR_OFFSET(7)]
|
|
+ str x11, [x3, #LR_OFFSET(6)]
|
|
+ str x10, [x3, #LR_OFFSET(5)]
|
|
+ str x9, [x3, #LR_OFFSET(4)]
|
|
+ str x8, [x3, #LR_OFFSET(3)]
|
|
+ str x7, [x3, #LR_OFFSET(2)]
|
|
+ str x6, [x3, #LR_OFFSET(1)]
|
|
+ str x5, [x3, #LR_OFFSET(0)]
|
|
+
|
|
+ tbnz w21, #29, 6f // 6 bits
|
|
+ tbz w21, #30, 5f // 5 bits
|
|
+ // 7 bits
|
|
+ mrs_s x20, ICH_AP0R3_EL2
|
|
+ str w20, [x3, #(VGIC_V3_CPU_AP0R + 3*4)]
|
|
+ mrs_s x19, ICH_AP0R2_EL2
|
|
+ str w19, [x3, #(VGIC_V3_CPU_AP0R + 2*4)]
|
|
+6: mrs_s x18, ICH_AP0R1_EL2
|
|
+ str w18, [x3, #(VGIC_V3_CPU_AP0R + 1*4)]
|
|
+5: mrs_s x17, ICH_AP0R0_EL2
|
|
+ str w17, [x3, #VGIC_V3_CPU_AP0R]
|
|
+
|
|
+ tbnz w21, #29, 6f // 6 bits
|
|
+ tbz w21, #30, 5f // 5 bits
|
|
+ // 7 bits
|
|
+ mrs_s x20, ICH_AP1R3_EL2
|
|
+ str w20, [x3, #(VGIC_V3_CPU_AP1R + 3*4)]
|
|
+ mrs_s x19, ICH_AP1R2_EL2
|
|
+ str w19, [x3, #(VGIC_V3_CPU_AP1R + 2*4)]
|
|
+6: mrs_s x18, ICH_AP1R1_EL2
|
|
+ str w18, [x3, #(VGIC_V3_CPU_AP1R + 1*4)]
|
|
+5: mrs_s x17, ICH_AP1R0_EL2
|
|
+ str w17, [x3, #VGIC_V3_CPU_AP1R]
|
|
+
|
|
+ // Restore SRE_EL1 access and re-enable SRE at EL1.
|
|
+ mrs_s x5, ICC_SRE_EL2
|
|
+ orr x5, x5, #ICC_SRE_EL2_ENABLE
|
|
+ msr_s ICC_SRE_EL2, x5
|
|
+ isb
|
|
+ mov x5, #1
|
|
+ msr_s ICC_SRE_EL1, x5
|
|
+.endm
|
|
+
|
|
+/*
|
|
+ * Restore the VGIC CPU state from memory
|
|
+ * x0: Register pointing to VCPU struct
|
|
+ */
|
|
+.macro restore_vgic_v3_state
|
|
+ // Disable SRE_EL1 access. Necessary, otherwise
|
|
+ // ICH_VMCR_EL2.VFIQEn becomes one, and FIQ happens...
|
|
+ msr_s ICC_SRE_EL1, xzr
|
|
+ isb
|
|
+
|
|
+ // Compute the address of struct vgic_cpu
|
|
+ add x3, x0, #VCPU_VGIC_CPU
|
|
+
|
|
+ // Restore all interesting registers
|
|
+ ldr w4, [x3, #VGIC_V3_CPU_HCR]
|
|
+ ldr w5, [x3, #VGIC_V3_CPU_VMCR]
|
|
+
|
|
+ msr_s ICH_HCR_EL2, x4
|
|
+ msr_s ICH_VMCR_EL2, x5
|
|
+
|
|
+ mrs_s x21, ICH_VTR_EL2
|
|
+
|
|
+ tbnz w21, #29, 6f // 6 bits
|
|
+ tbz w21, #30, 5f // 5 bits
|
|
+ // 7 bits
|
|
+ ldr w20, [x3, #(VGIC_V3_CPU_AP1R + 3*4)]
|
|
+ msr_s ICH_AP1R3_EL2, x20
|
|
+ ldr w19, [x3, #(VGIC_V3_CPU_AP1R + 2*4)]
|
|
+ msr_s ICH_AP1R2_EL2, x19
|
|
+6: ldr w18, [x3, #(VGIC_V3_CPU_AP1R + 1*4)]
|
|
+ msr_s ICH_AP1R1_EL2, x18
|
|
+5: ldr w17, [x3, #VGIC_V3_CPU_AP1R]
|
|
+ msr_s ICH_AP1R0_EL2, x17
|
|
+
|
|
+ tbnz w21, #29, 6f // 6 bits
|
|
+ tbz w21, #30, 5f // 5 bits
|
|
+ // 7 bits
|
|
+ ldr w20, [x3, #(VGIC_V3_CPU_AP0R + 3*4)]
|
|
+ msr_s ICH_AP0R3_EL2, x20
|
|
+ ldr w19, [x3, #(VGIC_V3_CPU_AP0R + 2*4)]
|
|
+ msr_s ICH_AP0R2_EL2, x19
|
|
+6: ldr w18, [x3, #(VGIC_V3_CPU_AP0R + 1*4)]
|
|
+ msr_s ICH_AP0R1_EL2, x18
|
|
+5: ldr w17, [x3, #VGIC_V3_CPU_AP0R]
|
|
+ msr_s ICH_AP0R0_EL2, x17
|
|
+
|
|
+ and w22, w21, #0xf
|
|
+ mvn w22, w21
|
|
+ ubfiz w23, w22, 2, 4 // w23 = (15 - ListRegs) * 4
|
|
+
|
|
+ adr x24, 1f
|
|
+ add x24, x24, x23
|
|
+ br x24
|
|
+
|
|
+1:
|
|
+ ldr x20, [x3, #LR_OFFSET(15)]
|
|
+ ldr x19, [x3, #LR_OFFSET(14)]
|
|
+ ldr x18, [x3, #LR_OFFSET(13)]
|
|
+ ldr x17, [x3, #LR_OFFSET(12)]
|
|
+ ldr x16, [x3, #LR_OFFSET(11)]
|
|
+ ldr x15, [x3, #LR_OFFSET(10)]
|
|
+ ldr x14, [x3, #LR_OFFSET(9)]
|
|
+ ldr x13, [x3, #LR_OFFSET(8)]
|
|
+ ldr x12, [x3, #LR_OFFSET(7)]
|
|
+ ldr x11, [x3, #LR_OFFSET(6)]
|
|
+ ldr x10, [x3, #LR_OFFSET(5)]
|
|
+ ldr x9, [x3, #LR_OFFSET(4)]
|
|
+ ldr x8, [x3, #LR_OFFSET(3)]
|
|
+ ldr x7, [x3, #LR_OFFSET(2)]
|
|
+ ldr x6, [x3, #LR_OFFSET(1)]
|
|
+ ldr x5, [x3, #LR_OFFSET(0)]
|
|
+
|
|
+ adr x24, 1f
|
|
+ add x24, x24, x23
|
|
+ br x24
|
|
+
|
|
+1:
|
|
+ msr_s ICH_LR15_EL2, x20
|
|
+ msr_s ICH_LR14_EL2, x19
|
|
+ msr_s ICH_LR13_EL2, x18
|
|
+ msr_s ICH_LR12_EL2, x17
|
|
+ msr_s ICH_LR11_EL2, x16
|
|
+ msr_s ICH_LR10_EL2, x15
|
|
+ msr_s ICH_LR9_EL2, x14
|
|
+ msr_s ICH_LR8_EL2, x13
|
|
+ msr_s ICH_LR7_EL2, x12
|
|
+ msr_s ICH_LR6_EL2, x11
|
|
+ msr_s ICH_LR5_EL2, x10
|
|
+ msr_s ICH_LR4_EL2, x9
|
|
+ msr_s ICH_LR3_EL2, x8
|
|
+ msr_s ICH_LR2_EL2, x7
|
|
+ msr_s ICH_LR1_EL2, x6
|
|
+ msr_s ICH_LR0_EL2, x5
|
|
+
|
|
+ // Ensure that the above will have reached the
|
|
+ // (re)distributors. This ensure the guest will read
|
|
+ // the correct values from the memory-mapped interface.
|
|
+ isb
|
|
+ dsb sy
|
|
+
|
|
+ // Prevent the guest from touching the GIC system registers
|
|
+ mrs_s x5, ICC_SRE_EL2
|
|
+ and x5, x5, #~ICC_SRE_EL2_ENABLE
|
|
+ msr_s ICC_SRE_EL2, x5
|
|
+.endm
|
|
+
|
|
+ENTRY(__save_vgic_v3_state)
|
|
+ save_vgic_v3_state
|
|
+ ret
|
|
+ENDPROC(__save_vgic_v3_state)
|
|
+
|
|
+ENTRY(__restore_vgic_v3_state)
|
|
+ restore_vgic_v3_state
|
|
+ ret
|
|
+ENDPROC(__restore_vgic_v3_state)
|
|
+
|
|
+ENTRY(__vgic_v3_get_ich_vtr_el2)
|
|
+ mrs_s x0, ICH_VTR_EL2
|
|
+ ret
|
|
+ENDPROC(__vgic_v3_get_ich_vtr_el2)
|
|
+
|
|
+ .popsection
|
|
diff --git a/arch/ia64/kernel/time.c b/arch/ia64/kernel/time.c
|
|
index 71c52bc..a149c67 100644
|
|
--- a/arch/ia64/kernel/time.c
|
|
+++ b/arch/ia64/kernel/time.c
|
|
@@ -384,21 +384,6 @@ static struct irqaction timer_irqaction = {
|
|
.name = "timer"
|
|
};
|
|
|
|
-static struct platform_device rtc_efi_dev = {
|
|
- .name = "rtc-efi",
|
|
- .id = -1,
|
|
-};
|
|
-
|
|
-static int __init rtc_init(void)
|
|
-{
|
|
- if (platform_device_register(&rtc_efi_dev) < 0)
|
|
- printk(KERN_ERR "unable to register rtc device...\n");
|
|
-
|
|
- /* not necessarily an error */
|
|
- return 0;
|
|
-}
|
|
-module_init(rtc_init);
|
|
-
|
|
void read_persistent_clock(struct timespec *ts)
|
|
{
|
|
efi_gettimeofday(ts);
|
|
diff --git a/drivers/ata/ahci_xgene.c b/drivers/ata/ahci_xgene.c
|
|
index bc28111..00e5a0c 100644
|
|
--- a/drivers/ata/ahci_xgene.c
|
|
+++ b/drivers/ata/ahci_xgene.c
|
|
@@ -134,7 +134,8 @@ static unsigned int xgene_ahci_qc_issue(struct ata_queued_cmd *qc)
|
|
struct xgene_ahci_context *ctx = hpriv->plat_data;
|
|
int rc = 0;
|
|
|
|
- if (unlikely(ctx->last_cmd[ap->port_no] == ATA_CMD_ID_ATA))
|
|
+ if (unlikely(ctx->last_cmd[ap->port_no] == ATA_CMD_ID_ATA ||
|
|
+ ctx->last_cmd[ap->port_no] == ATA_CMD_SMART))
|
|
xgene_ahci_restart_engine(ap);
|
|
|
|
rc = ahci_qc_issue(qc);
|
|
diff --git a/drivers/irqchip/irq-gic.c b/drivers/irqchip/irq-gic.c
|
|
index 9c1f883..1ddfdde 100644
|
|
--- a/drivers/irqchip/irq-gic.c
|
|
+++ b/drivers/irqchip/irq-gic.c
|
|
@@ -383,6 +383,7 @@ static void gic_cpu_init(struct gic_chip_data *gic)
|
|
void __iomem *dist_base = gic_data_dist_base(gic);
|
|
void __iomem *base = gic_data_cpu_base(gic);
|
|
unsigned int cpu_mask, cpu = smp_processor_id();
|
|
+ unsigned int ctrl_mask;
|
|
int i;
|
|
|
|
/*
|
|
@@ -403,13 +404,29 @@ static void gic_cpu_init(struct gic_chip_data *gic)
|
|
gic_cpu_config(dist_base, NULL);
|
|
|
|
writel_relaxed(0xf0, base + GIC_CPU_PRIMASK);
|
|
- writel_relaxed(1, base + GIC_CPU_CTRL);
|
|
+
|
|
+ ctrl_mask = readl(base + GIC_CPU_CTRL);
|
|
+
|
|
+ /* Mask out the gic v2 bypass bits */
|
|
+ ctrl_mask &= 0x1e0;
|
|
+
|
|
+ /* Enable group 0 */
|
|
+ ctrl_mask |= 0x1;
|
|
+ writel_relaxed(ctrl_mask, base + GIC_CPU_CTRL);
|
|
}
|
|
|
|
void gic_cpu_if_down(void)
|
|
{
|
|
+ unsigned int ctrl_mask;
|
|
void __iomem *cpu_base = gic_data_cpu_base(&gic_data[0]);
|
|
- writel_relaxed(0, cpu_base + GIC_CPU_CTRL);
|
|
+
|
|
+ ctrl_mask = readl(cpu_base + GIC_CPU_CTRL);
|
|
+ /*
|
|
+ * Disable grp enable bit, leave the bypass bits alone as changing
|
|
+ * them could leave the system unstable
|
|
+ */
|
|
+ ctrl_mask &= 0x1e0;
|
|
+ writel_relaxed(ctrl_mask, cpu_base + GIC_CPU_CTRL);
|
|
}
|
|
|
|
#ifdef CONFIG_CPU_PM
|
|
@@ -520,6 +537,7 @@ static void gic_cpu_restore(unsigned int gic_nr)
|
|
{
|
|
int i;
|
|
u32 *ptr;
|
|
+ unsigned int ctrl_mask;
|
|
void __iomem *dist_base;
|
|
void __iomem *cpu_base;
|
|
|
|
@@ -544,7 +562,15 @@ static void gic_cpu_restore(unsigned int gic_nr)
|
|
writel_relaxed(0xa0a0a0a0, dist_base + GIC_DIST_PRI + i * 4);
|
|
|
|
writel_relaxed(0xf0, cpu_base + GIC_CPU_PRIMASK);
|
|
- writel_relaxed(1, cpu_base + GIC_CPU_CTRL);
|
|
+
|
|
+ ctrl_mask = readl(cpu_base + GIC_CPU_CTRL);
|
|
+
|
|
+ /* Mask out the gic v2 bypass bits */
|
|
+ ctrl_mask &= 0x1e0;
|
|
+
|
|
+ /* Enable group 0 */
|
|
+ ctrl_mask |= 0x1;
|
|
+ writel_relaxed(ctrl_mask, cpu_base + GIC_CPU_CTRL);
|
|
}
|
|
|
|
static int gic_notifier(struct notifier_block *self, unsigned long cmd, void *v)
|
|
diff --git a/drivers/net/ethernet/Kconfig b/drivers/net/ethernet/Kconfig
|
|
index edb7186..dc7406c 100644
|
|
--- a/drivers/net/ethernet/Kconfig
|
|
+++ b/drivers/net/ethernet/Kconfig
|
|
@@ -24,6 +24,7 @@ source "drivers/net/ethernet/allwinner/Kconfig"
|
|
source "drivers/net/ethernet/alteon/Kconfig"
|
|
source "drivers/net/ethernet/altera/Kconfig"
|
|
source "drivers/net/ethernet/amd/Kconfig"
|
|
+source "drivers/net/ethernet/apm/Kconfig"
|
|
source "drivers/net/ethernet/apple/Kconfig"
|
|
source "drivers/net/ethernet/arc/Kconfig"
|
|
source "drivers/net/ethernet/atheros/Kconfig"
|
|
diff --git a/drivers/net/ethernet/Makefile b/drivers/net/ethernet/Makefile
|
|
index 58de333..224a018 100644
|
|
--- a/drivers/net/ethernet/Makefile
|
|
+++ b/drivers/net/ethernet/Makefile
|
|
@@ -10,6 +10,7 @@ obj-$(CONFIG_NET_VENDOR_ALLWINNER) += allwinner/
|
|
obj-$(CONFIG_NET_VENDOR_ALTEON) += alteon/
|
|
obj-$(CONFIG_ALTERA_TSE) += altera/
|
|
obj-$(CONFIG_NET_VENDOR_AMD) += amd/
|
|
+obj-$(CONFIG_NET_XGENE) += apm/
|
|
obj-$(CONFIG_NET_VENDOR_APPLE) += apple/
|
|
obj-$(CONFIG_NET_VENDOR_ARC) += arc/
|
|
obj-$(CONFIG_NET_VENDOR_ATHEROS) += atheros/
|
|
diff --git a/drivers/net/ethernet/apm/Kconfig b/drivers/net/ethernet/apm/Kconfig
|
|
new file mode 100644
|
|
index 0000000..ec63d70
|
|
--- /dev/null
|
|
+++ b/drivers/net/ethernet/apm/Kconfig
|
|
@@ -0,0 +1 @@
|
|
+source "drivers/net/ethernet/apm/xgene/Kconfig"
|
|
diff --git a/drivers/net/ethernet/apm/Makefile b/drivers/net/ethernet/apm/Makefile
|
|
new file mode 100644
|
|
index 0000000..65ce32a
|
|
--- /dev/null
|
|
+++ b/drivers/net/ethernet/apm/Makefile
|
|
@@ -0,0 +1,5 @@
|
|
+#
|
|
+# Makefile for APM X-GENE Ethernet driver.
|
|
+#
|
|
+
|
|
+obj-$(CONFIG_NET_XGENE) += xgene/
|
|
diff --git a/drivers/net/ethernet/apm/xgene/Kconfig b/drivers/net/ethernet/apm/xgene/Kconfig
|
|
new file mode 100644
|
|
index 0000000..616dff6
|
|
--- /dev/null
|
|
+++ b/drivers/net/ethernet/apm/xgene/Kconfig
|
|
@@ -0,0 +1,9 @@
|
|
+config NET_XGENE
|
|
+ tristate "APM X-Gene SoC Ethernet Driver"
|
|
+ select PHYLIB
|
|
+ help
|
|
+ This is the Ethernet driver for the on-chip ethernet interface on the
|
|
+ APM X-Gene SoC.
|
|
+
|
|
+ To compile this driver as a module, choose M here. This module will
|
|
+ be called xgene_enet.
|
|
diff --git a/drivers/net/ethernet/apm/xgene/Makefile b/drivers/net/ethernet/apm/xgene/Makefile
|
|
new file mode 100644
|
|
index 0000000..c643e8a
|
|
--- /dev/null
|
|
+++ b/drivers/net/ethernet/apm/xgene/Makefile
|
|
@@ -0,0 +1,6 @@
|
|
+#
|
|
+# Makefile for APM X-Gene Ethernet Driver.
|
|
+#
|
|
+
|
|
+xgene-enet-objs := xgene_enet_hw.o xgene_enet_main.o xgene_enet_ethtool.o
|
|
+obj-$(CONFIG_NET_XGENE) += xgene-enet.o
|
|
diff --git a/drivers/net/ethernet/apm/xgene/xgene_enet_ethtool.c b/drivers/net/ethernet/apm/xgene/xgene_enet_ethtool.c
|
|
new file mode 100644
|
|
index 0000000..63f2aa5
|
|
--- /dev/null
|
|
+++ b/drivers/net/ethernet/apm/xgene/xgene_enet_ethtool.c
|
|
@@ -0,0 +1,125 @@
|
|
+/* Applied Micro X-Gene SoC Ethernet Driver
|
|
+ *
|
|
+ * Copyright (c) 2014, Applied Micro Circuits Corporation
|
|
+ * Authors: Iyappan Subramanian <isubramanian@apm.com>
|
|
+ *
|
|
+ * This program is free software; you can redistribute it and/or modify it
|
|
+ * under the terms of the GNU General Public License as published by the
|
|
+ * Free Software Foundation; either version 2 of the License, or (at your
|
|
+ * option) any later version.
|
|
+ *
|
|
+ * This program is distributed in the hope that it will be useful,
|
|
+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
+ * GNU General Public License for more details.
|
|
+ *
|
|
+ * You should have received a copy of the GNU General Public License
|
|
+ * along with this program. If not, see <http://www.gnu.org/licenses/>.
|
|
+ */
|
|
+
|
|
+#include <linux/ethtool.h>
|
|
+#include "xgene_enet_main.h"
|
|
+
|
|
+struct xgene_gstrings_stats {
|
|
+ char name[ETH_GSTRING_LEN];
|
|
+ int offset;
|
|
+};
|
|
+
|
|
+#define XGENE_STAT(m) { #m, offsetof(struct xgene_enet_pdata, stats.m) }
|
|
+
|
|
+static const struct xgene_gstrings_stats gstrings_stats[] = {
|
|
+ XGENE_STAT(rx_packets),
|
|
+ XGENE_STAT(tx_packets),
|
|
+ XGENE_STAT(rx_bytes),
|
|
+ XGENE_STAT(tx_bytes),
|
|
+ XGENE_STAT(rx_errors),
|
|
+ XGENE_STAT(tx_errors),
|
|
+ XGENE_STAT(rx_length_errors),
|
|
+ XGENE_STAT(rx_crc_errors),
|
|
+ XGENE_STAT(rx_frame_errors),
|
|
+ XGENE_STAT(rx_fifo_errors)
|
|
+};
|
|
+
|
|
+#define XGENE_STATS_LEN ARRAY_SIZE(gstrings_stats)
|
|
+
|
|
+static void xgene_get_drvinfo(struct net_device *ndev,
|
|
+ struct ethtool_drvinfo *info)
|
|
+{
|
|
+ struct xgene_enet_pdata *pdata = netdev_priv(ndev);
|
|
+ struct platform_device *pdev = pdata->pdev;
|
|
+
|
|
+ strcpy(info->driver, "xgene_enet");
|
|
+ strcpy(info->version, XGENE_DRV_VERSION);
|
|
+ snprintf(info->fw_version, ETHTOOL_FWVERS_LEN, "N/A");
|
|
+ sprintf(info->bus_info, "%s", pdev->name);
|
|
+}
|
|
+
|
|
+static int xgene_get_settings(struct net_device *ndev, struct ethtool_cmd *cmd)
|
|
+{
|
|
+ struct xgene_enet_pdata *pdata = netdev_priv(ndev);
|
|
+ struct phy_device *phydev = pdata->phy_dev;
|
|
+
|
|
+ if (phydev == NULL)
|
|
+ return -ENODEV;
|
|
+
|
|
+ return phy_ethtool_gset(phydev, cmd);
|
|
+}
|
|
+
|
|
+static int xgene_set_settings(struct net_device *ndev, struct ethtool_cmd *cmd)
|
|
+{
|
|
+ struct xgene_enet_pdata *pdata = netdev_priv(ndev);
|
|
+ struct phy_device *phydev = pdata->phy_dev;
|
|
+
|
|
+ if (phydev == NULL)
|
|
+ return -ENODEV;
|
|
+
|
|
+ return phy_ethtool_sset(phydev, cmd);
|
|
+}
|
|
+
|
|
+static void xgene_get_strings(struct net_device *ndev, u32 stringset, u8 *data)
|
|
+{
|
|
+ int i;
|
|
+ u8 *p = data;
|
|
+
|
|
+ if (stringset != ETH_SS_STATS)
|
|
+ return;
|
|
+
|
|
+ for (i = 0; i < XGENE_STATS_LEN; i++) {
|
|
+ memcpy(p, gstrings_stats[i].name, ETH_GSTRING_LEN);
|
|
+ p += ETH_GSTRING_LEN;
|
|
+ }
|
|
+}
|
|
+
|
|
+static int xgene_get_sset_count(struct net_device *ndev, int sset)
|
|
+{
|
|
+ if (sset != ETH_SS_STATS)
|
|
+ return -EINVAL;
|
|
+
|
|
+ return XGENE_STATS_LEN;
|
|
+}
|
|
+
|
|
+static void xgene_get_ethtool_stats(struct net_device *ndev,
|
|
+ struct ethtool_stats *dummy,
|
|
+ u64 *data)
|
|
+{
|
|
+ void *pdata = netdev_priv(ndev);
|
|
+ int i;
|
|
+
|
|
+ for (i = 0; i < XGENE_STATS_LEN; i++)
|
|
+ *data++ = *(u64 *)(pdata + gstrings_stats[i].offset);
|
|
+}
|
|
+
|
|
+static const struct ethtool_ops xgene_ethtool_ops = {
|
|
+ .get_drvinfo = xgene_get_drvinfo,
|
|
+ .get_settings = xgene_get_settings,
|
|
+ .set_settings = xgene_set_settings,
|
|
+ .get_link = ethtool_op_get_link,
|
|
+ .get_strings = xgene_get_strings,
|
|
+ .get_sset_count = xgene_get_sset_count,
|
|
+ .get_ethtool_stats = xgene_get_ethtool_stats
|
|
+};
|
|
+
|
|
+void xgene_enet_set_ethtool_ops(struct net_device *ndev)
|
|
+{
|
|
+ ndev->ethtool_ops = &xgene_ethtool_ops;
|
|
+}
|
|
diff --git a/drivers/net/ethernet/apm/xgene/xgene_enet_hw.c b/drivers/net/ethernet/apm/xgene/xgene_enet_hw.c
|
|
new file mode 100644
|
|
index 0000000..e52af60
|
|
--- /dev/null
|
|
+++ b/drivers/net/ethernet/apm/xgene/xgene_enet_hw.c
|
|
@@ -0,0 +1,747 @@
|
|
+/* Applied Micro X-Gene SoC Ethernet Driver
|
|
+ *
|
|
+ * Copyright (c) 2014, Applied Micro Circuits Corporation
|
|
+ * Authors: Iyappan Subramanian <isubramanian@apm.com>
|
|
+ * Ravi Patel <rapatel@apm.com>
|
|
+ * Keyur Chudgar <kchudgar@apm.com>
|
|
+ *
|
|
+ * This program is free software; you can redistribute it and/or modify it
|
|
+ * under the terms of the GNU General Public License as published by the
|
|
+ * Free Software Foundation; either version 2 of the License, or (at your
|
|
+ * option) any later version.
|
|
+ *
|
|
+ * This program is distributed in the hope that it will be useful,
|
|
+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
+ * GNU General Public License for more details.
|
|
+ *
|
|
+ * You should have received a copy of the GNU General Public License
|
|
+ * along with this program. If not, see <http://www.gnu.org/licenses/>.
|
|
+ */
|
|
+
|
|
+#include "xgene_enet_main.h"
|
|
+#include "xgene_enet_hw.h"
|
|
+
|
|
+static void xgene_enet_ring_init(struct xgene_enet_desc_ring *ring)
|
|
+{
|
|
+ u32 *ring_cfg = ring->state;
|
|
+ u64 addr = ring->dma;
|
|
+ enum xgene_enet_ring_cfgsize cfgsize = ring->cfgsize;
|
|
+
|
|
+ ring_cfg[4] |= (1 << SELTHRSH_POS) &
|
|
+ CREATE_MASK(SELTHRSH_POS, SELTHRSH_LEN);
|
|
+ ring_cfg[3] |= ACCEPTLERR;
|
|
+ ring_cfg[2] |= QCOHERENT;
|
|
+
|
|
+ addr >>= 8;
|
|
+ ring_cfg[2] |= (addr << RINGADDRL_POS) &
|
|
+ CREATE_MASK_ULL(RINGADDRL_POS, RINGADDRL_LEN);
|
|
+ addr >>= RINGADDRL_LEN;
|
|
+ ring_cfg[3] |= addr & CREATE_MASK_ULL(RINGADDRH_POS, RINGADDRH_LEN);
|
|
+ ring_cfg[3] |= ((u32) cfgsize << RINGSIZE_POS) &
|
|
+ CREATE_MASK(RINGSIZE_POS, RINGSIZE_LEN);
|
|
+}
|
|
+
|
|
+static void xgene_enet_ring_set_type(struct xgene_enet_desc_ring *ring)
|
|
+{
|
|
+ u32 *ring_cfg = ring->state;
|
|
+ bool is_bufpool;
|
|
+ u32 val;
|
|
+
|
|
+ is_bufpool = xgene_enet_is_bufpool(ring->id);
|
|
+ val = (is_bufpool) ? RING_BUFPOOL : RING_REGULAR;
|
|
+ ring_cfg[4] |= (val << RINGTYPE_POS) &
|
|
+ CREATE_MASK(RINGTYPE_POS, RINGTYPE_LEN);
|
|
+
|
|
+ if (is_bufpool) {
|
|
+ ring_cfg[3] |= (BUFPOOL_MODE << RINGMODE_POS) &
|
|
+ CREATE_MASK(RINGMODE_POS, RINGMODE_LEN);
|
|
+ }
|
|
+}
|
|
+
|
|
+static void xgene_enet_ring_set_recombbuf(struct xgene_enet_desc_ring *ring)
|
|
+{
|
|
+ u32 *ring_cfg = ring->state;
|
|
+
|
|
+ ring_cfg[3] |= RECOMBBUF;
|
|
+ ring_cfg[3] |= (0xf << RECOMTIMEOUTL_POS) &
|
|
+ CREATE_MASK(RECOMTIMEOUTL_POS, RECOMTIMEOUTL_LEN);
|
|
+ ring_cfg[4] |= 0x7 & CREATE_MASK(RECOMTIMEOUTH_POS, RECOMTIMEOUTH_LEN);
|
|
+}
|
|
+
|
|
+static void xgene_enet_ring_wr32(struct xgene_enet_desc_ring *ring,
|
|
+ u32 offset, u32 data)
|
|
+{
|
|
+ struct xgene_enet_pdata *pdata = netdev_priv(ring->ndev);
|
|
+
|
|
+ iowrite32(data, pdata->ring_csr_addr + offset);
|
|
+}
|
|
+
|
|
+static void xgene_enet_ring_rd32(struct xgene_enet_desc_ring *ring,
|
|
+ u32 offset, u32 *data)
|
|
+{
|
|
+ struct xgene_enet_pdata *pdata = netdev_priv(ring->ndev);
|
|
+
|
|
+ *data = ioread32(pdata->ring_csr_addr + offset);
|
|
+}
|
|
+
|
|
+static void xgene_enet_write_ring_state(struct xgene_enet_desc_ring *ring)
|
|
+{
|
|
+ int i;
|
|
+
|
|
+ xgene_enet_ring_wr32(ring, CSR_RING_CONFIG, ring->num);
|
|
+ for (i = 0; i < NUM_RING_CONFIG; i++) {
|
|
+ xgene_enet_ring_wr32(ring, CSR_RING_WR_BASE + (i * 4),
|
|
+ ring->state[i]);
|
|
+ }
|
|
+}
|
|
+
|
|
+static void xgene_enet_clr_ring_state(struct xgene_enet_desc_ring *ring)
|
|
+{
|
|
+ memset(ring->state, 0, sizeof(u32) * NUM_RING_CONFIG);
|
|
+ xgene_enet_write_ring_state(ring);
|
|
+}
|
|
+
|
|
+static void xgene_enet_set_ring_state(struct xgene_enet_desc_ring *ring)
|
|
+{
|
|
+ xgene_enet_ring_set_type(ring);
|
|
+
|
|
+ if (xgene_enet_ring_owner(ring->id) == RING_OWNER_ETH0)
|
|
+ xgene_enet_ring_set_recombbuf(ring);
|
|
+
|
|
+ xgene_enet_ring_init(ring);
|
|
+ xgene_enet_write_ring_state(ring);
|
|
+}
|
|
+
|
|
+static void xgene_enet_set_ring_id(struct xgene_enet_desc_ring *ring)
|
|
+{
|
|
+ u32 ring_id_val, ring_id_buf;
|
|
+ bool is_bufpool;
|
|
+
|
|
+ is_bufpool = xgene_enet_is_bufpool(ring->id);
|
|
+
|
|
+ ring_id_val = ring->id & GENMASK(9, 0);
|
|
+ ring_id_val |= OVERWRITE;
|
|
+
|
|
+ ring_id_buf = (ring->num << 9) & GENMASK(18, 9);
|
|
+ ring_id_buf |= PREFETCH_BUF_EN;
|
|
+ if (is_bufpool)
|
|
+ ring_id_buf |= IS_BUFFER_POOL;
|
|
+
|
|
+ xgene_enet_ring_wr32(ring, CSR_RING_ID, ring_id_val);
|
|
+ xgene_enet_ring_wr32(ring, CSR_RING_ID_BUF, ring_id_buf);
|
|
+}
|
|
+
|
|
+static void xgene_enet_clr_desc_ring_id(struct xgene_enet_desc_ring *ring)
|
|
+{
|
|
+ u32 ring_id;
|
|
+
|
|
+ ring_id = ring->id | OVERWRITE;
|
|
+ xgene_enet_ring_wr32(ring, CSR_RING_ID, ring_id);
|
|
+ xgene_enet_ring_wr32(ring, CSR_RING_ID_BUF, 0);
|
|
+}
|
|
+
|
|
+struct xgene_enet_desc_ring *xgene_enet_setup_ring(
|
|
+ struct xgene_enet_desc_ring *ring)
|
|
+{
|
|
+ u32 size = ring->size;
|
|
+ u32 i, data;
|
|
+ u64 *desc;
|
|
+ bool is_bufpool;
|
|
+
|
|
+ xgene_enet_clr_ring_state(ring);
|
|
+ xgene_enet_set_ring_state(ring);
|
|
+ xgene_enet_set_ring_id(ring);
|
|
+
|
|
+ ring->slots = xgene_enet_get_numslots(ring->id, size);
|
|
+
|
|
+ is_bufpool = xgene_enet_is_bufpool(ring->id);
|
|
+ if (is_bufpool || xgene_enet_ring_owner(ring->id) != RING_OWNER_CPU)
|
|
+ return ring;
|
|
+
|
|
+ for (i = 0; i < ring->slots; i++) {
|
|
+ desc = (u64 *)&ring->raw_desc[i];
|
|
+ desc[EMPTY_SLOT_INDEX] = EMPTY_SLOT;
|
|
+ }
|
|
+
|
|
+ xgene_enet_ring_rd32(ring, CSR_RING_NE_INT_MODE, &data);
|
|
+ data |= BIT(31 - xgene_enet_ring_bufnum(ring->id));
|
|
+ xgene_enet_ring_wr32(ring, CSR_RING_NE_INT_MODE, data);
|
|
+
|
|
+ return ring;
|
|
+}
|
|
+
|
|
+void xgene_enet_clear_ring(struct xgene_enet_desc_ring *ring)
|
|
+{
|
|
+ u32 data;
|
|
+ bool is_bufpool;
|
|
+
|
|
+ is_bufpool = xgene_enet_is_bufpool(ring->id);
|
|
+ if (is_bufpool || xgene_enet_ring_owner(ring->id) != RING_OWNER_CPU)
|
|
+ goto out;
|
|
+
|
|
+ xgene_enet_ring_rd32(ring, CSR_RING_NE_INT_MODE, &data);
|
|
+ data &= ~BIT(31 - xgene_enet_ring_bufnum(ring->id));
|
|
+ xgene_enet_ring_wr32(ring, CSR_RING_NE_INT_MODE, data);
|
|
+
|
|
+out:
|
|
+ xgene_enet_clr_desc_ring_id(ring);
|
|
+ xgene_enet_clr_ring_state(ring);
|
|
+}
|
|
+
|
|
+void xgene_enet_parse_error(struct xgene_enet_desc_ring *ring,
|
|
+ struct xgene_enet_pdata *pdata,
|
|
+ enum xgene_enet_err_code status)
|
|
+{
|
|
+ struct rtnl_link_stats64 *stats = &pdata->stats;
|
|
+
|
|
+ switch (status) {
|
|
+ case INGRESS_CRC:
|
|
+ stats->rx_crc_errors++;
|
|
+ break;
|
|
+ case INGRESS_CHECKSUM:
|
|
+ case INGRESS_CHECKSUM_COMPUTE:
|
|
+ stats->rx_errors++;
|
|
+ break;
|
|
+ case INGRESS_TRUNC_FRAME:
|
|
+ stats->rx_frame_errors++;
|
|
+ break;
|
|
+ case INGRESS_PKT_LEN:
|
|
+ stats->rx_length_errors++;
|
|
+ break;
|
|
+ case INGRESS_PKT_UNDER:
|
|
+ stats->rx_frame_errors++;
|
|
+ break;
|
|
+ case INGRESS_FIFO_OVERRUN:
|
|
+ stats->rx_fifo_errors++;
|
|
+ break;
|
|
+ default:
|
|
+ break;
|
|
+ }
|
|
+}
|
|
+
|
|
+static void xgene_enet_wr_csr(struct xgene_enet_pdata *pdata,
|
|
+ u32 offset, u32 val)
|
|
+{
|
|
+ void __iomem *addr = pdata->eth_csr_addr + offset;
|
|
+
|
|
+ iowrite32(val, addr);
|
|
+}
|
|
+
|
|
+static void xgene_enet_wr_ring_if(struct xgene_enet_pdata *pdata,
|
|
+ u32 offset, u32 val)
|
|
+{
|
|
+ void __iomem *addr = pdata->eth_ring_if_addr + offset;
|
|
+
|
|
+ iowrite32(val, addr);
|
|
+}
|
|
+
|
|
+static void xgene_enet_wr_diag_csr(struct xgene_enet_pdata *pdata,
|
|
+ u32 offset, u32 val)
|
|
+{
|
|
+ void __iomem *addr = pdata->eth_diag_csr_addr + offset;
|
|
+
|
|
+ iowrite32(val, addr);
|
|
+}
|
|
+
|
|
+static void xgene_enet_wr_mcx_csr(struct xgene_enet_pdata *pdata,
|
|
+ u32 offset, u32 val)
|
|
+{
|
|
+ void __iomem *addr = pdata->mcx_mac_csr_addr + offset;
|
|
+
|
|
+ iowrite32(val, addr);
|
|
+}
|
|
+
|
|
+static bool xgene_enet_wr_indirect(void __iomem *addr, void __iomem *wr,
|
|
+ void __iomem *cmd, void __iomem *cmd_done,
|
|
+ u32 wr_addr, u32 wr_data)
|
|
+{
|
|
+ u32 done;
|
|
+ u8 wait = 10;
|
|
+
|
|
+ iowrite32(wr_addr, addr);
|
|
+ iowrite32(wr_data, wr);
|
|
+ iowrite32(XGENE_ENET_WR_CMD, cmd);
|
|
+
|
|
+ /* wait for write command to complete */
|
|
+ while (!(done = ioread32(cmd_done)) && wait--)
|
|
+ udelay(1);
|
|
+
|
|
+ if (!done)
|
|
+ return false;
|
|
+
|
|
+ iowrite32(0, cmd);
|
|
+
|
|
+ return true;
|
|
+}
|
|
+
|
|
+static void xgene_enet_wr_mcx_mac(struct xgene_enet_pdata *pdata,
|
|
+ u32 wr_addr, u32 wr_data)
|
|
+{
|
|
+ void __iomem *addr, *wr, *cmd, *cmd_done;
|
|
+ bool ret;
|
|
+
|
|
+ addr = pdata->mcx_mac_addr + MAC_ADDR_REG_OFFSET;
|
|
+ wr = pdata->mcx_mac_addr + MAC_WRITE_REG_OFFSET;
|
|
+ cmd = pdata->mcx_mac_addr + MAC_COMMAND_REG_OFFSET;
|
|
+ cmd_done = pdata->mcx_mac_addr + MAC_COMMAND_DONE_REG_OFFSET;
|
|
+
|
|
+ ret = xgene_enet_wr_indirect(addr, wr, cmd, cmd_done, wr_addr, wr_data);
|
|
+ if (!ret)
|
|
+ netdev_err(pdata->ndev, "MCX mac write failed, addr: %04x\n",
|
|
+ wr_addr);
|
|
+}
|
|
+
|
|
+static void xgene_enet_rd_csr(struct xgene_enet_pdata *pdata,
|
|
+ u32 offset, u32 *val)
|
|
+{
|
|
+ void __iomem *addr = pdata->eth_csr_addr + offset;
|
|
+
|
|
+ *val = ioread32(addr);
|
|
+}
|
|
+
|
|
+static void xgene_enet_rd_diag_csr(struct xgene_enet_pdata *pdata,
|
|
+ u32 offset, u32 *val)
|
|
+{
|
|
+ void __iomem *addr = pdata->eth_diag_csr_addr + offset;
|
|
+
|
|
+ *val = ioread32(addr);
|
|
+}
|
|
+
|
|
+static void xgene_enet_rd_mcx_csr(struct xgene_enet_pdata *pdata,
|
|
+ u32 offset, u32 *val)
|
|
+{
|
|
+ void __iomem *addr = pdata->mcx_mac_csr_addr + offset;
|
|
+
|
|
+ *val = ioread32(addr);
|
|
+}
|
|
+
|
|
+static bool xgene_enet_rd_indirect(void __iomem *addr, void __iomem *rd,
|
|
+ void __iomem *cmd, void __iomem *cmd_done,
|
|
+ u32 rd_addr, u32 *rd_data)
|
|
+{
|
|
+ u32 done;
|
|
+ u8 wait = 10;
|
|
+
|
|
+ iowrite32(rd_addr, addr);
|
|
+ iowrite32(XGENE_ENET_RD_CMD, cmd);
|
|
+
|
|
+ /* wait for read command to complete */
|
|
+ while (!(done = ioread32(cmd_done)) && wait--)
|
|
+ udelay(1);
|
|
+
|
|
+ if (!done)
|
|
+ return false;
|
|
+
|
|
+ *rd_data = ioread32(rd);
|
|
+ iowrite32(0, cmd);
|
|
+
|
|
+ return true;
|
|
+}
|
|
+
|
|
+static void xgene_enet_rd_mcx_mac(struct xgene_enet_pdata *pdata,
|
|
+ u32 rd_addr, u32 *rd_data)
|
|
+{
|
|
+ void __iomem *addr, *rd, *cmd, *cmd_done;
|
|
+ bool ret;
|
|
+
|
|
+ addr = pdata->mcx_mac_addr + MAC_ADDR_REG_OFFSET;
|
|
+ rd = pdata->mcx_mac_addr + MAC_READ_REG_OFFSET;
|
|
+ cmd = pdata->mcx_mac_addr + MAC_COMMAND_REG_OFFSET;
|
|
+ cmd_done = pdata->mcx_mac_addr + MAC_COMMAND_DONE_REG_OFFSET;
|
|
+
|
|
+ ret = xgene_enet_rd_indirect(addr, rd, cmd, cmd_done, rd_addr, rd_data);
|
|
+ if (!ret)
|
|
+ netdev_err(pdata->ndev, "MCX mac read failed, addr: %04x\n",
|
|
+ rd_addr);
|
|
+}
|
|
+
|
|
+static int xgene_mii_phy_write(struct xgene_enet_pdata *pdata, int phy_id,
|
|
+ u32 reg, u16 data)
|
|
+{
|
|
+ u32 addr = 0, wr_data = 0;
|
|
+ u32 done;
|
|
+ u8 wait = 10;
|
|
+
|
|
+ PHY_ADDR_SET(&addr, phy_id);
|
|
+ REG_ADDR_SET(&addr, reg);
|
|
+ xgene_enet_wr_mcx_mac(pdata, MII_MGMT_ADDRESS_ADDR, addr);
|
|
+
|
|
+ PHY_CONTROL_SET(&wr_data, data);
|
|
+ xgene_enet_wr_mcx_mac(pdata, MII_MGMT_CONTROL_ADDR, wr_data);
|
|
+ do {
|
|
+ usleep_range(5, 10);
|
|
+ xgene_enet_rd_mcx_mac(pdata, MII_MGMT_INDICATORS_ADDR, &done);
|
|
+ } while ((done & BUSY_MASK) && wait--);
|
|
+
|
|
+ if (done & BUSY_MASK) {
|
|
+ netdev_err(pdata->ndev, "MII_MGMT write failed\n");
|
|
+ return -1;
|
|
+ }
|
|
+
|
|
+ return 0;
|
|
+}
|
|
+
|
|
+static int xgene_mii_phy_read(struct xgene_enet_pdata *pdata,
|
|
+ u8 phy_id, u32 reg)
|
|
+{
|
|
+ u32 addr = 0;
|
|
+ u32 data, done;
|
|
+ u8 wait = 10;
|
|
+
|
|
+ PHY_ADDR_SET(&addr, phy_id);
|
|
+ REG_ADDR_SET(&addr, reg);
|
|
+ xgene_enet_wr_mcx_mac(pdata, MII_MGMT_ADDRESS_ADDR, addr);
|
|
+ xgene_enet_wr_mcx_mac(pdata, MII_MGMT_COMMAND_ADDR, READ_CYCLE_MASK);
|
|
+ do {
|
|
+ usleep_range(5, 10);
|
|
+ xgene_enet_rd_mcx_mac(pdata, MII_MGMT_INDICATORS_ADDR, &done);
|
|
+ } while ((done & BUSY_MASK) && wait--);
|
|
+
|
|
+ if (done & BUSY_MASK) {
|
|
+ netdev_err(pdata->ndev, "MII_MGMT read failed\n");
|
|
+ return -1;
|
|
+ }
|
|
+
|
|
+ xgene_enet_rd_mcx_mac(pdata, MII_MGMT_STATUS_ADDR, &data);
|
|
+ xgene_enet_wr_mcx_mac(pdata, MII_MGMT_COMMAND_ADDR, 0);
|
|
+
|
|
+ return data;
|
|
+}
|
|
+
|
|
+void xgene_gmac_set_mac_addr(struct xgene_enet_pdata *pdata)
|
|
+{
|
|
+ u32 addr0, addr1;
|
|
+ u8 *dev_addr = pdata->ndev->dev_addr;
|
|
+
|
|
+ addr0 = (dev_addr[3] << 24) | (dev_addr[2] << 16) |
|
|
+ (dev_addr[1] << 8) | dev_addr[0];
|
|
+ addr1 = (dev_addr[5] << 24) | (dev_addr[4] << 16);
|
|
+ addr1 |= pdata->phy_addr & 0xFFFF;
|
|
+
|
|
+ xgene_enet_wr_mcx_mac(pdata, STATION_ADDR0_ADDR, addr0);
|
|
+ xgene_enet_wr_mcx_mac(pdata, STATION_ADDR1_ADDR, addr1);
|
|
+}
|
|
+
|
|
+static int xgene_enet_ecc_init(struct xgene_enet_pdata *pdata)
|
|
+{
|
|
+ struct net_device *ndev = pdata->ndev;
|
|
+ u32 data;
|
|
+ u8 wait = 10;
|
|
+
|
|
+ xgene_enet_wr_diag_csr(pdata, ENET_CFG_MEM_RAM_SHUTDOWN_ADDR, 0x0);
|
|
+ do {
|
|
+ usleep_range(100, 110);
|
|
+ xgene_enet_rd_diag_csr(pdata, ENET_BLOCK_MEM_RDY_ADDR, &data);
|
|
+ } while ((data != 0xffffffff) && wait--);
|
|
+
|
|
+ if (data != 0xffffffff) {
|
|
+ netdev_err(ndev, "Failed to release memory from shutdown\n");
|
|
+ return -ENODEV;
|
|
+ }
|
|
+
|
|
+ return 0;
|
|
+}
|
|
+
|
|
+void xgene_gmac_reset(struct xgene_enet_pdata *pdata)
|
|
+{
|
|
+ xgene_enet_wr_mcx_mac(pdata, MAC_CONFIG_1_ADDR, SOFT_RESET1);
|
|
+ xgene_enet_wr_mcx_mac(pdata, MAC_CONFIG_1_ADDR, 0);
|
|
+}
|
|
+
|
|
+void xgene_gmac_init(struct xgene_enet_pdata *pdata, int speed)
|
|
+{
|
|
+ u32 value, mc2;
|
|
+ u32 intf_ctl, rgmii;
|
|
+ u32 icm0, icm2;
|
|
+
|
|
+ xgene_gmac_reset(pdata);
|
|
+
|
|
+ xgene_enet_rd_mcx_csr(pdata, ICM_CONFIG0_REG_0_ADDR, &icm0);
|
|
+ xgene_enet_rd_mcx_csr(pdata, ICM_CONFIG2_REG_0_ADDR, &icm2);
|
|
+ xgene_enet_rd_mcx_mac(pdata, MAC_CONFIG_2_ADDR, &mc2);
|
|
+ xgene_enet_rd_mcx_mac(pdata, INTERFACE_CONTROL_ADDR, &intf_ctl);
|
|
+ xgene_enet_rd_csr(pdata, RGMII_REG_0_ADDR, &rgmii);
|
|
+
|
|
+ switch (speed) {
|
|
+ case SPEED_10:
|
|
+ ENET_INTERFACE_MODE2_SET(&mc2, 1);
|
|
+ CFG_MACMODE_SET(&icm0, 0);
|
|
+ CFG_WAITASYNCRD_SET(&icm2, 500);
|
|
+ rgmii &= ~CFG_SPEED_1250;
|
|
+ break;
|
|
+ case SPEED_100:
|
|
+ ENET_INTERFACE_MODE2_SET(&mc2, 1);
|
|
+ intf_ctl |= ENET_LHD_MODE;
|
|
+ CFG_MACMODE_SET(&icm0, 1);
|
|
+ CFG_WAITASYNCRD_SET(&icm2, 80);
|
|
+ rgmii &= ~CFG_SPEED_1250;
|
|
+ break;
|
|
+ default:
|
|
+ ENET_INTERFACE_MODE2_SET(&mc2, 2);
|
|
+ intf_ctl |= ENET_GHD_MODE;
|
|
+ CFG_TXCLK_MUXSEL0_SET(&rgmii, 4);
|
|
+ xgene_enet_rd_csr(pdata, DEBUG_REG_ADDR, &value);
|
|
+ value |= CFG_BYPASS_UNISEC_TX | CFG_BYPASS_UNISEC_RX;
|
|
+ xgene_enet_wr_csr(pdata, DEBUG_REG_ADDR, value);
|
|
+ break;
|
|
+ }
|
|
+
|
|
+ mc2 |= FULL_DUPLEX2;
|
|
+ xgene_enet_wr_mcx_mac(pdata, MAC_CONFIG_2_ADDR, mc2);
|
|
+ xgene_enet_wr_mcx_mac(pdata, INTERFACE_CONTROL_ADDR, intf_ctl);
|
|
+
|
|
+ xgene_gmac_set_mac_addr(pdata);
|
|
+
|
|
+ /* Adjust MDC clock frequency */
|
|
+ xgene_enet_rd_mcx_mac(pdata, MII_MGMT_CONFIG_ADDR, &value);
|
|
+ MGMT_CLOCK_SEL_SET(&value, 7);
|
|
+ xgene_enet_wr_mcx_mac(pdata, MII_MGMT_CONFIG_ADDR, value);
|
|
+
|
|
+ /* Enable drop if bufpool not available */
|
|
+ xgene_enet_rd_csr(pdata, RSIF_CONFIG_REG_ADDR, &value);
|
|
+ value |= CFG_RSIF_FPBUFF_TIMEOUT_EN;
|
|
+ xgene_enet_wr_csr(pdata, RSIF_CONFIG_REG_ADDR, value);
|
|
+
|
|
+ /* Rtype should be copied from FP */
|
|
+ xgene_enet_wr_csr(pdata, RSIF_RAM_DBG_REG0_ADDR, 0);
|
|
+ xgene_enet_wr_csr(pdata, RGMII_REG_0_ADDR, rgmii);
|
|
+
|
|
+ /* Rx-Tx traffic resume */
|
|
+ xgene_enet_wr_csr(pdata, CFG_LINK_AGGR_RESUME_0_ADDR, TX_PORT0);
|
|
+
|
|
+ xgene_enet_wr_mcx_csr(pdata, ICM_CONFIG0_REG_0_ADDR, icm0);
|
|
+ xgene_enet_wr_mcx_csr(pdata, ICM_CONFIG2_REG_0_ADDR, icm2);
|
|
+
|
|
+ xgene_enet_rd_mcx_csr(pdata, RX_DV_GATE_REG_0_ADDR, &value);
|
|
+ value &= ~TX_DV_GATE_EN0;
|
|
+ value &= ~RX_DV_GATE_EN0;
|
|
+ value |= RESUME_RX0;
|
|
+ xgene_enet_wr_mcx_csr(pdata, RX_DV_GATE_REG_0_ADDR, value);
|
|
+
|
|
+ xgene_enet_wr_csr(pdata, CFG_BYPASS_ADDR, RESUME_TX);
|
|
+}
|
|
+
|
|
+static void xgene_enet_config_ring_if_assoc(struct xgene_enet_pdata *pdata)
|
|
+{
|
|
+ u32 val = 0xffffffff;
|
|
+
|
|
+ xgene_enet_wr_ring_if(pdata, ENET_CFGSSQMIWQASSOC_ADDR, val);
|
|
+ xgene_enet_wr_ring_if(pdata, ENET_CFGSSQMIFPQASSOC_ADDR, val);
|
|
+ xgene_enet_wr_ring_if(pdata, ENET_CFGSSQMIQMLITEWQASSOC_ADDR, val);
|
|
+ xgene_enet_wr_ring_if(pdata, ENET_CFGSSQMIQMLITEFPQASSOC_ADDR, val);
|
|
+}
|
|
+
|
|
+void xgene_enet_cle_bypass(struct xgene_enet_pdata *pdata,
|
|
+ u32 dst_ring_num, u16 bufpool_id)
|
|
+{
|
|
+ u32 cb;
|
|
+ u32 fpsel;
|
|
+
|
|
+ fpsel = xgene_enet_ring_bufnum(bufpool_id) - 0x20;
|
|
+
|
|
+ xgene_enet_rd_csr(pdata, CLE_BYPASS_REG0_0_ADDR, &cb);
|
|
+ cb |= CFG_CLE_BYPASS_EN0;
|
|
+ CFG_CLE_IP_PROTOCOL0_SET(&cb, 3);
|
|
+ xgene_enet_wr_csr(pdata, CLE_BYPASS_REG0_0_ADDR, cb);
|
|
+
|
|
+ xgene_enet_rd_csr(pdata, CLE_BYPASS_REG1_0_ADDR, &cb);
|
|
+ CFG_CLE_DSTQID0_SET(&cb, dst_ring_num);
|
|
+ CFG_CLE_FPSEL0_SET(&cb, fpsel);
|
|
+ xgene_enet_wr_csr(pdata, CLE_BYPASS_REG1_0_ADDR, cb);
|
|
+}
|
|
+
|
|
+void xgene_gmac_rx_enable(struct xgene_enet_pdata *pdata)
|
|
+{
|
|
+ u32 data;
|
|
+
|
|
+ xgene_enet_rd_mcx_mac(pdata, MAC_CONFIG_1_ADDR, &data);
|
|
+ xgene_enet_wr_mcx_mac(pdata, MAC_CONFIG_1_ADDR, data | RX_EN);
|
|
+}
|
|
+
|
|
+void xgene_gmac_tx_enable(struct xgene_enet_pdata *pdata)
|
|
+{
|
|
+ u32 data;
|
|
+
|
|
+ xgene_enet_rd_mcx_mac(pdata, MAC_CONFIG_1_ADDR, &data);
|
|
+ xgene_enet_wr_mcx_mac(pdata, MAC_CONFIG_1_ADDR, data | TX_EN);
|
|
+}
|
|
+
|
|
+void xgene_gmac_rx_disable(struct xgene_enet_pdata *pdata)
|
|
+{
|
|
+ u32 data;
|
|
+
|
|
+ xgene_enet_rd_mcx_mac(pdata, MAC_CONFIG_1_ADDR, &data);
|
|
+ xgene_enet_wr_mcx_mac(pdata, MAC_CONFIG_1_ADDR, data & ~RX_EN);
|
|
+}
|
|
+
|
|
+void xgene_gmac_tx_disable(struct xgene_enet_pdata *pdata)
|
|
+{
|
|
+ u32 data;
|
|
+
|
|
+ xgene_enet_rd_mcx_mac(pdata, MAC_CONFIG_1_ADDR, &data);
|
|
+ xgene_enet_wr_mcx_mac(pdata, MAC_CONFIG_1_ADDR, data & ~TX_EN);
|
|
+}
|
|
+
|
|
+void xgene_enet_reset(struct xgene_enet_pdata *pdata)
|
|
+{
|
|
+ u32 val;
|
|
+
|
|
+ clk_prepare_enable(pdata->clk);
|
|
+ clk_disable_unprepare(pdata->clk);
|
|
+ clk_prepare_enable(pdata->clk);
|
|
+ xgene_enet_ecc_init(pdata);
|
|
+ xgene_enet_config_ring_if_assoc(pdata);
|
|
+
|
|
+ /* Enable auto-incr for scanning */
|
|
+ xgene_enet_rd_mcx_mac(pdata, MII_MGMT_CONFIG_ADDR, &val);
|
|
+ val |= SCAN_AUTO_INCR;
|
|
+ MGMT_CLOCK_SEL_SET(&val, 1);
|
|
+ xgene_enet_wr_mcx_mac(pdata, MII_MGMT_CONFIG_ADDR, val);
|
|
+}
|
|
+
|
|
+void xgene_gport_shutdown(struct xgene_enet_pdata *pdata)
|
|
+{
|
|
+ clk_disable_unprepare(pdata->clk);
|
|
+}
|
|
+
|
|
+static int xgene_enet_mdio_read(struct mii_bus *bus, int mii_id, int regnum)
|
|
+{
|
|
+ struct xgene_enet_pdata *pdata = bus->priv;
|
|
+ u32 val;
|
|
+
|
|
+ val = xgene_mii_phy_read(pdata, mii_id, regnum);
|
|
+ netdev_dbg(pdata->ndev, "mdio_rd: bus=%d reg=%d val=%x\n",
|
|
+ mii_id, regnum, val);
|
|
+
|
|
+ return val;
|
|
+}
|
|
+
|
|
+static int xgene_enet_mdio_write(struct mii_bus *bus, int mii_id, int regnum,
|
|
+ u16 val)
|
|
+{
|
|
+ struct xgene_enet_pdata *pdata = bus->priv;
|
|
+ int ret;
|
|
+
|
|
+ netdev_dbg(pdata->ndev, "mdio_wr: bus=%d reg=%d val=%x\n",
|
|
+ mii_id, regnum, val);
|
|
+ ret = xgene_mii_phy_write(pdata, mii_id, regnum, val);
|
|
+
|
|
+ return ret;
|
|
+}
|
|
+
|
|
+static void xgene_enet_adjust_link(struct net_device *ndev)
|
|
+{
|
|
+ struct xgene_enet_pdata *pdata = netdev_priv(ndev);
|
|
+ struct phy_device *phydev = pdata->phy_dev;
|
|
+
|
|
+ if (phydev->link) {
|
|
+ if (pdata->phy_speed != phydev->speed) {
|
|
+ xgene_gmac_init(pdata, phydev->speed);
|
|
+ xgene_gmac_rx_enable(pdata);
|
|
+ xgene_gmac_tx_enable(pdata);
|
|
+ pdata->phy_speed = phydev->speed;
|
|
+ phy_print_status(phydev);
|
|
+ }
|
|
+ } else {
|
|
+ xgene_gmac_rx_disable(pdata);
|
|
+ xgene_gmac_tx_disable(pdata);
|
|
+ pdata->phy_speed = SPEED_UNKNOWN;
|
|
+ phy_print_status(phydev);
|
|
+ }
|
|
+}
|
|
+
|
|
+static int xgene_enet_phy_connect(struct net_device *ndev)
|
|
+{
|
|
+ struct xgene_enet_pdata *pdata = netdev_priv(ndev);
|
|
+ struct device_node *phy_np;
|
|
+ struct phy_device *phy_dev;
|
|
+ struct device *dev = &pdata->pdev->dev;
|
|
+
|
|
+ phy_np = of_parse_phandle(dev->of_node, "phy-handle", 0);
|
|
+ if (!phy_np) {
|
|
+ netdev_dbg(ndev, "No phy-handle found\n");
|
|
+ return -ENODEV;
|
|
+ }
|
|
+
|
|
+ phy_dev = of_phy_connect(ndev, phy_np, &xgene_enet_adjust_link,
|
|
+ 0, pdata->phy_mode);
|
|
+ if (!phy_dev) {
|
|
+ netdev_err(ndev, "Could not connect to PHY\n");
|
|
+ return -ENODEV;
|
|
+ }
|
|
+
|
|
+ pdata->phy_speed = SPEED_UNKNOWN;
|
|
+ phy_dev->supported &= ~SUPPORTED_10baseT_Half &
|
|
+ ~SUPPORTED_100baseT_Half &
|
|
+ ~SUPPORTED_1000baseT_Half;
|
|
+ phy_dev->advertising = phy_dev->supported;
|
|
+ pdata->phy_dev = phy_dev;
|
|
+
|
|
+ return 0;
|
|
+}
|
|
+
|
|
+int xgene_enet_mdio_config(struct xgene_enet_pdata *pdata)
|
|
+{
|
|
+ struct net_device *ndev = pdata->ndev;
|
|
+ struct device *dev = &pdata->pdev->dev;
|
|
+ struct device_node *child_np;
|
|
+ struct device_node *mdio_np = NULL;
|
|
+ struct mii_bus *mdio_bus;
|
|
+ int ret;
|
|
+
|
|
+ for_each_child_of_node(dev->of_node, child_np) {
|
|
+ if (of_device_is_compatible(child_np, "apm,xgene-mdio")) {
|
|
+ mdio_np = child_np;
|
|
+ break;
|
|
+ }
|
|
+ }
|
|
+
|
|
+ if (!mdio_np) {
|
|
+ netdev_dbg(ndev, "No mdio node in the dts\n");
|
|
+ return -1;
|
|
+ }
|
|
+
|
|
+ mdio_bus = mdiobus_alloc();
|
|
+ if (!mdio_bus)
|
|
+ return -ENOMEM;
|
|
+
|
|
+ mdio_bus->name = "APM X-Gene MDIO bus";
|
|
+ mdio_bus->read = xgene_enet_mdio_read;
|
|
+ mdio_bus->write = xgene_enet_mdio_write;
|
|
+ snprintf(mdio_bus->id, MII_BUS_ID_SIZE, "%s-%s", "xgene-mii",
|
|
+ ndev->name);
|
|
+
|
|
+ mdio_bus->priv = pdata;
|
|
+ mdio_bus->parent = &ndev->dev;
|
|
+
|
|
+ ret = of_mdiobus_register(mdio_bus, mdio_np);
|
|
+ if (ret) {
|
|
+ netdev_err(ndev, "Failed to register MDIO bus\n");
|
|
+ goto err;
|
|
+ }
|
|
+ pdata->mdio_bus = mdio_bus;
|
|
+
|
|
+ ret = xgene_enet_phy_connect(ndev);
|
|
+ if (ret)
|
|
+ goto err;
|
|
+
|
|
+ return ret;
|
|
+
|
|
+err:
|
|
+ mdiobus_free(mdio_bus);
|
|
+
|
|
+ return ret;
|
|
+}
|
|
+
|
|
+int xgene_enet_mdio_remove(struct xgene_enet_pdata *pdata)
|
|
+{
|
|
+ struct mii_bus *mdio_bus;
|
|
+
|
|
+ mdio_bus = pdata->mdio_bus;
|
|
+ mdiobus_unregister(mdio_bus);
|
|
+ mdiobus_free(mdio_bus);
|
|
+ pdata->mdio_bus = NULL;
|
|
+
|
|
+ return 0;
|
|
+}
|
|
diff --git a/drivers/net/ethernet/apm/xgene/xgene_enet_hw.h b/drivers/net/ethernet/apm/xgene/xgene_enet_hw.h
|
|
new file mode 100644
|
|
index 0000000..2041313
|
|
--- /dev/null
|
|
+++ b/drivers/net/ethernet/apm/xgene/xgene_enet_hw.h
|
|
@@ -0,0 +1,375 @@
|
|
+/* Applied Micro X-Gene SoC Ethernet Driver
|
|
+ *
|
|
+ * Copyright (c) 2014, Applied Micro Circuits Corporation
|
|
+ * Authors: Iyappan Subramanian <isubramanian@apm.com>
|
|
+ * Ravi Patel <rapatel@apm.com>
|
|
+ * Keyur Chudgar <kchudgar@apm.com>
|
|
+ *
|
|
+ * This program is free software; you can redistribute it and/or modify it
|
|
+ * under the terms of the GNU General Public License as published by the
|
|
+ * Free Software Foundation; either version 2 of the License, or (at your
|
|
+ * option) any later version.
|
|
+ *
|
|
+ * This program is distributed in the hope that it will be useful,
|
|
+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
+ * GNU General Public License for more details.
|
|
+ *
|
|
+ * You should have received a copy of the GNU General Public License
|
|
+ * along with this program. If not, see <http://www.gnu.org/licenses/>.
|
|
+ */
|
|
+
|
|
+#ifndef __XGENE_ENET_HW_H__
|
|
+#define __XGENE_ENET_HW_H__
|
|
+
|
|
+#include "xgene_enet_main.h"
|
|
+
|
|
+struct xgene_enet_pdata;
|
|
+struct xgene_enet_stats;
|
|
+
|
|
+/* clears and then set bits */
|
|
+static inline void xgene_set_bits(u32 *dst, u32 val, u32 start, u32 len)
|
|
+{
|
|
+ u32 end = start + len - 1;
|
|
+ u32 mask = GENMASK(end, start);
|
|
+
|
|
+ *dst &= ~mask;
|
|
+ *dst |= (val << start) & mask;
|
|
+}
|
|
+
|
|
+static inline u32 xgene_get_bits(u32 val, u32 start, u32 end)
|
|
+{
|
|
+ return (val & GENMASK(end, start)) >> start;
|
|
+}
|
|
+
|
|
+#define CSR_RING_ID 0x0008
|
|
+#define OVERWRITE BIT(31)
|
|
+#define IS_BUFFER_POOL BIT(20)
|
|
+#define PREFETCH_BUF_EN BIT(21)
|
|
+#define CSR_RING_ID_BUF 0x000c
|
|
+#define CSR_RING_NE_INT_MODE 0x017c
|
|
+#define CSR_RING_CONFIG 0x006c
|
|
+#define CSR_RING_WR_BASE 0x0070
|
|
+#define NUM_RING_CONFIG 5
|
|
+#define BUFPOOL_MODE 3
|
|
+#define RM3 3
|
|
+#define INC_DEC_CMD_ADDR 0x002c
|
|
+#define UDP_HDR_SIZE 2
|
|
+#define BUF_LEN_CODE_2K 0x5000
|
|
+
|
|
+#define CREATE_MASK(pos, len) GENMASK((pos)+(len)-1, (pos))
|
|
+#define CREATE_MASK_ULL(pos, len) GENMASK_ULL((pos)+(len)-1, (pos))
|
|
+
|
|
+/* Empty slot soft signature */
|
|
+#define EMPTY_SLOT_INDEX 1
|
|
+#define EMPTY_SLOT ~0ULL
|
|
+
|
|
+#define WORK_DESC_SIZE 32
|
|
+#define BUFPOOL_DESC_SIZE 16
|
|
+
|
|
+#define RING_OWNER_MASK GENMASK(9, 6)
|
|
+#define RING_BUFNUM_MASK GENMASK(5, 0)
|
|
+
|
|
+#define SELTHRSH_POS 3
|
|
+#define SELTHRSH_LEN 3
|
|
+#define RINGADDRL_POS 5
|
|
+#define RINGADDRL_LEN 27
|
|
+#define RINGADDRH_POS 0
|
|
+#define RINGADDRH_LEN 6
|
|
+#define RINGSIZE_POS 23
|
|
+#define RINGSIZE_LEN 3
|
|
+#define RINGTYPE_POS 19
|
|
+#define RINGTYPE_LEN 2
|
|
+#define RINGMODE_POS 20
|
|
+#define RINGMODE_LEN 3
|
|
+#define RECOMTIMEOUTL_POS 28
|
|
+#define RECOMTIMEOUTL_LEN 3
|
|
+#define RECOMTIMEOUTH_POS 0
|
|
+#define RECOMTIMEOUTH_LEN 2
|
|
+#define NUMMSGSINQ_POS 1
|
|
+#define NUMMSGSINQ_LEN 16
|
|
+#define ACCEPTLERR BIT(19)
|
|
+#define QCOHERENT BIT(4)
|
|
+#define RECOMBBUF BIT(27)
|
|
+
|
|
+#define BLOCK_ETH_CSR_OFFSET 0x2000
|
|
+#define BLOCK_ETH_RING_IF_OFFSET 0x9000
|
|
+#define BLOCK_ETH_CLKRST_CSR_OFFSET 0xC000
|
|
+#define BLOCK_ETH_DIAG_CSR_OFFSET 0xD000
|
|
+
|
|
+#define BLOCK_ETH_MAC_OFFSET 0x0000
|
|
+#define BLOCK_ETH_STATS_OFFSET 0x0014
|
|
+#define BLOCK_ETH_MAC_CSR_OFFSET 0x2800
|
|
+
|
|
+#define MAC_ADDR_REG_OFFSET 0x00
|
|
+#define MAC_COMMAND_REG_OFFSET 0x04
|
|
+#define MAC_WRITE_REG_OFFSET 0x08
|
|
+#define MAC_READ_REG_OFFSET 0x0c
|
|
+#define MAC_COMMAND_DONE_REG_OFFSET 0x10
|
|
+
|
|
+#define STAT_ADDR_REG_OFFSET 0x00
|
|
+#define STAT_COMMAND_REG_OFFSET 0x04
|
|
+#define STAT_WRITE_REG_OFFSET 0x08
|
|
+#define STAT_READ_REG_OFFSET 0x0c
|
|
+#define STAT_COMMAND_DONE_REG_OFFSET 0x10
|
|
+
|
|
+#define MII_MGMT_CONFIG_ADDR 0x20
|
|
+#define MII_MGMT_COMMAND_ADDR 0x24
|
|
+#define MII_MGMT_ADDRESS_ADDR 0x28
|
|
+#define MII_MGMT_CONTROL_ADDR 0x2c
|
|
+#define MII_MGMT_STATUS_ADDR 0x30
|
|
+#define MII_MGMT_INDICATORS_ADDR 0x34
|
|
+
|
|
+#define BUSY_MASK BIT(0)
|
|
+#define READ_CYCLE_MASK BIT(0)
|
|
+#define PHY_CONTROL_SET(dst, val) xgene_set_bits(dst, val, 0, 16)
|
|
+
|
|
+#define ENET_SPARE_CFG_REG_ADDR 0x0750
|
|
+#define RSIF_CONFIG_REG_ADDR 0x0010
|
|
+#define RSIF_RAM_DBG_REG0_ADDR 0x0048
|
|
+#define RGMII_REG_0_ADDR 0x07e0
|
|
+#define CFG_LINK_AGGR_RESUME_0_ADDR 0x07c8
|
|
+#define DEBUG_REG_ADDR 0x0700
|
|
+#define CFG_BYPASS_ADDR 0x0294
|
|
+#define CLE_BYPASS_REG0_0_ADDR 0x0490
|
|
+#define CLE_BYPASS_REG1_0_ADDR 0x0494
|
|
+#define CFG_RSIF_FPBUFF_TIMEOUT_EN BIT(31)
|
|
+#define RESUME_TX BIT(0)
|
|
+#define CFG_SPEED_1250 BIT(24)
|
|
+#define TX_PORT0 BIT(0)
|
|
+#define CFG_BYPASS_UNISEC_TX BIT(2)
|
|
+#define CFG_BYPASS_UNISEC_RX BIT(1)
|
|
+#define CFG_CLE_BYPASS_EN0 BIT(31)
|
|
+#define CFG_TXCLK_MUXSEL0_SET(dst, val) xgene_set_bits(dst, val, 29, 3)
|
|
+
|
|
+#define CFG_CLE_IP_PROTOCOL0_SET(dst, val) xgene_set_bits(dst, val, 16, 2)
|
|
+#define CFG_CLE_DSTQID0_SET(dst, val) xgene_set_bits(dst, val, 0, 12)
|
|
+#define CFG_CLE_FPSEL0_SET(dst, val) xgene_set_bits(dst, val, 16, 4)
|
|
+#define CFG_MACMODE_SET(dst, val) xgene_set_bits(dst, val, 18, 2)
|
|
+#define CFG_WAITASYNCRD_SET(dst, val) xgene_set_bits(dst, val, 0, 16)
|
|
+#define ICM_CONFIG0_REG_0_ADDR 0x0400
|
|
+#define ICM_CONFIG2_REG_0_ADDR 0x0410
|
|
+#define RX_DV_GATE_REG_0_ADDR 0x05fc
|
|
+#define TX_DV_GATE_EN0 BIT(2)
|
|
+#define RX_DV_GATE_EN0 BIT(1)
|
|
+#define RESUME_RX0 BIT(0)
|
|
+#define ENET_CFGSSQMIWQASSOC_ADDR 0xe0
|
|
+#define ENET_CFGSSQMIFPQASSOC_ADDR 0xdc
|
|
+#define ENET_CFGSSQMIQMLITEFPQASSOC_ADDR 0xf0
|
|
+#define ENET_CFGSSQMIQMLITEWQASSOC_ADDR 0xf4
|
|
+#define ENET_CFG_MEM_RAM_SHUTDOWN_ADDR 0x70
|
|
+#define ENET_BLOCK_MEM_RDY_ADDR 0x74
|
|
+#define MAC_CONFIG_1_ADDR 0x00
|
|
+#define MAC_CONFIG_2_ADDR 0x04
|
|
+#define MAX_FRAME_LEN_ADDR 0x10
|
|
+#define INTERFACE_CONTROL_ADDR 0x38
|
|
+#define STATION_ADDR0_ADDR 0x40
|
|
+#define STATION_ADDR1_ADDR 0x44
|
|
+#define PHY_ADDR_SET(dst, val) xgene_set_bits(dst, val, 8, 5)
|
|
+#define REG_ADDR_SET(dst, val) xgene_set_bits(dst, val, 0, 5)
|
|
+#define ENET_INTERFACE_MODE2_SET(dst, val) xgene_set_bits(dst, val, 8, 2)
|
|
+#define MGMT_CLOCK_SEL_SET(dst, val) xgene_set_bits(dst, val, 0, 3)
|
|
+#define SOFT_RESET1 BIT(31)
|
|
+#define TX_EN BIT(0)
|
|
+#define RX_EN BIT(2)
|
|
+#define ENET_LHD_MODE BIT(25)
|
|
+#define ENET_GHD_MODE BIT(26)
|
|
+#define FULL_DUPLEX2 BIT(0)
|
|
+#define SCAN_AUTO_INCR BIT(5)
|
|
+#define TBYT_ADDR 0x38
|
|
+#define TPKT_ADDR 0x39
|
|
+#define TDRP_ADDR 0x45
|
|
+#define TFCS_ADDR 0x47
|
|
+#define TUND_ADDR 0x4a
|
|
+
|
|
+#define TSO_IPPROTO_TCP 1
|
|
+#define FULL_DUPLEX 2
|
|
+
|
|
+#define USERINFO_POS 0
|
|
+#define USERINFO_LEN 32
|
|
+#define FPQNUM_POS 32
|
|
+#define FPQNUM_LEN 12
|
|
+#define LERR_POS 60
|
|
+#define LERR_LEN 3
|
|
+#define STASH_POS 52
|
|
+#define STASH_LEN 2
|
|
+#define BUFDATALEN_POS 48
|
|
+#define BUFDATALEN_LEN 12
|
|
+#define DATAADDR_POS 0
|
|
+#define DATAADDR_LEN 42
|
|
+#define COHERENT_POS 63
|
|
+#define HENQNUM_POS 48
|
|
+#define HENQNUM_LEN 12
|
|
+#define TYPESEL_POS 44
|
|
+#define TYPESEL_LEN 4
|
|
+#define ETHHDR_POS 12
|
|
+#define IC_POS 35 /* Insert CRC */
|
|
+#define TCPHDR_POS 0
|
|
+#define TCPHDR_LEN 6
|
|
+#define IPHDR_POS 6
|
|
+#define IPHDR_LEN 6
|
|
+#define EC_POS 22 /* Enable checksum */
|
|
+#define IS_POS 24 /* IP protocol select */
|
|
+
|
|
+#define DATAADDR_MASK CREATE_MASK_ULL(DATAADDR_POS, DATAADDR_LEN)
|
|
+#define BUFDATALEN_MASK CREATE_MASK_ULL(BUFDATALEN_POS, BUFDATALEN_LEN)
|
|
+#define USERINFO_MASK CREATE_MASK_ULL(USERINFO_POS, USERINFO_LEN)
|
|
+#define FPQNUM_MASK CREATE_MASK_ULL(FPQNUM_POS, FPQNUM_LEN)
|
|
+#define LERR_MASK CREATE_MASK_ULL(LERR_POS, LERR_LEN)
|
|
+#define STASHING_MASK CREATE_MASK_ULL(STASH_POS, STASH_LEN)
|
|
+#define COHERENT_MASK BIT_ULL(COHERENT_POS)
|
|
+#define HENQNUM_MASK CREATE_MASK_ULL(HENQNUM_POS, HENQNUM_LEN)
|
|
+#define TCPHDR_MASK CREATE_MASK(TCPHDR_POS, TCPHDR_LEN)
|
|
+#define IPHDR_MASK CREATE_MASK(IPHDR_POS, IPHDR_LEN)
|
|
+#define EC_MASK BIT(EC_POS)
|
|
+#define IS_MASK BIT(IS_POS)
|
|
+#define INSERT_CRC BIT_ULL(IC_POS)
|
|
+#define TYPE_ETH_WORK_MESSAGE BIT_ULL(44)
|
|
+
|
|
+struct xgene_enet_raw_desc {
|
|
+ u64 m0;
|
|
+ u64 m1;
|
|
+ u64 m2;
|
|
+ u64 m3;
|
|
+};
|
|
+
|
|
+struct xgene_enet_raw_desc16 {
|
|
+ u64 m0;
|
|
+ u64 m1;
|
|
+};
|
|
+
|
|
+static inline void xgene_enet_cpu_to_le64(void *desc_ptr, int count)
|
|
+{
|
|
+ u64 *desc = desc_ptr;
|
|
+ int i;
|
|
+
|
|
+ for (i = 0; i < count; i++)
|
|
+ desc[i] = cpu_to_le64(desc[i]);
|
|
+}
|
|
+
|
|
+static inline void xgene_enet_le64_to_cpu(void *desc_ptr, int count)
|
|
+{
|
|
+ u64 *desc = desc_ptr;
|
|
+ int i;
|
|
+
|
|
+ for (i = 0; i < count; i++)
|
|
+ desc[i] = le64_to_cpu(desc[i]);
|
|
+}
|
|
+
|
|
+static inline void xgene_enet_desc16_to_le64(void *desc_ptr)
|
|
+{
|
|
+ u64 *desc;
|
|
+
|
|
+ desc = desc_ptr;
|
|
+ desc[1] = cpu_to_le64(desc[1]);
|
|
+}
|
|
+
|
|
+static inline void xgene_enet_le64_to_desc16(void *desc_ptr)
|
|
+{
|
|
+ u64 *desc;
|
|
+
|
|
+ desc = desc_ptr;
|
|
+ desc[1] = le64_to_cpu(desc[1]);
|
|
+}
|
|
+
|
|
+enum xgene_enet_ring_cfgsize {
|
|
+ RING_CFGSIZE_512B,
|
|
+ RING_CFGSIZE_2KB,
|
|
+ RING_CFGSIZE_16KB,
|
|
+ RING_CFGSIZE_64KB,
|
|
+ RING_CFGSIZE_512KB,
|
|
+ RING_CFGSIZE_INVALID
|
|
+};
|
|
+
|
|
+enum xgene_enet_ring_type {
|
|
+ RING_DISABLED,
|
|
+ RING_REGULAR,
|
|
+ RING_BUFPOOL
|
|
+};
|
|
+
|
|
+enum xgene_ring_owner {
|
|
+ RING_OWNER_ETH0,
|
|
+ RING_OWNER_CPU = 15,
|
|
+ RING_OWNER_INVALID
|
|
+};
|
|
+
|
|
+enum xgene_enet_ring_bufnum {
|
|
+ RING_BUFNUM_REGULAR = 0x0,
|
|
+ RING_BUFNUM_BUFPOOL = 0x20,
|
|
+ RING_BUFNUM_INVALID
|
|
+};
|
|
+
|
|
+enum xgene_enet_cmd {
|
|
+ XGENE_ENET_WR_CMD = BIT(31),
|
|
+ XGENE_ENET_RD_CMD = BIT(30)
|
|
+};
|
|
+
|
|
+enum xgene_enet_err_code {
|
|
+ HBF_READ_DATA = 3,
|
|
+ HBF_LL_READ = 4,
|
|
+ BAD_WORK_MSG = 6,
|
|
+ BUFPOOL_TIMEOUT = 15,
|
|
+ INGRESS_CRC = 16,
|
|
+ INGRESS_CHECKSUM = 17,
|
|
+ INGRESS_TRUNC_FRAME = 18,
|
|
+ INGRESS_PKT_LEN = 19,
|
|
+ INGRESS_PKT_UNDER = 20,
|
|
+ INGRESS_FIFO_OVERRUN = 21,
|
|
+ INGRESS_CHECKSUM_COMPUTE = 26,
|
|
+ ERR_CODE_INVALID
|
|
+};
|
|
+
|
|
+static inline enum xgene_ring_owner xgene_enet_ring_owner(u16 id)
|
|
+{
|
|
+ return (id & RING_OWNER_MASK) >> 6;
|
|
+}
|
|
+
|
|
+static inline u8 xgene_enet_ring_bufnum(u16 id)
|
|
+{
|
|
+ return id & RING_BUFNUM_MASK;
|
|
+}
|
|
+
|
|
+static inline bool xgene_enet_is_bufpool(u16 id)
|
|
+{
|
|
+ return ((id & RING_BUFNUM_MASK) >= 0x20) ? true : false;
|
|
+}
|
|
+
|
|
+static inline u16 xgene_enet_get_numslots(u16 id, u32 size)
|
|
+{
|
|
+ bool is_bufpool = xgene_enet_is_bufpool(id);
|
|
+
|
|
+ return (is_bufpool) ? size / BUFPOOL_DESC_SIZE :
|
|
+ size / WORK_DESC_SIZE;
|
|
+}
|
|
+
|
|
+struct xgene_enet_desc_ring *xgene_enet_setup_ring(
|
|
+ struct xgene_enet_desc_ring *ring);
|
|
+void xgene_enet_clear_ring(struct xgene_enet_desc_ring *ring);
|
|
+
|
|
+void xgene_set_tx_desc(struct xgene_enet_desc_ring *ring,
|
|
+ struct xgene_enet_raw_desc *raw_desc);
|
|
+void xgene_get_desc(struct xgene_enet_desc_ring *ring,
|
|
+ struct xgene_enet_raw_desc *raw_desc);
|
|
+void xgene_get_bufpool_desc(struct xgene_enet_desc_ring *ring,
|
|
+ struct xgene_enet_raw_desc16 *raw_desc);
|
|
+void xgene_enet_parse_error(struct xgene_enet_desc_ring *ring,
|
|
+ struct xgene_enet_pdata *pdata,
|
|
+ enum xgene_enet_err_code status);
|
|
+
|
|
+void xgene_enet_reset(struct xgene_enet_pdata *priv);
|
|
+void xgene_gmac_reset(struct xgene_enet_pdata *priv);
|
|
+void xgene_gmac_init(struct xgene_enet_pdata *priv, int speed);
|
|
+void xgene_gmac_tx_enable(struct xgene_enet_pdata *priv);
|
|
+void xgene_gmac_rx_enable(struct xgene_enet_pdata *priv);
|
|
+void xgene_gmac_tx_disable(struct xgene_enet_pdata *priv);
|
|
+void xgene_gmac_rx_disable(struct xgene_enet_pdata *priv);
|
|
+void xgene_gmac_set_mac_addr(struct xgene_enet_pdata *pdata);
|
|
+void xgene_enet_cle_bypass(struct xgene_enet_pdata *pdata,
|
|
+ u32 dst_ring_num, u16 bufpool_id);
|
|
+void xgene_gport_shutdown(struct xgene_enet_pdata *priv);
|
|
+void xgene_gmac_get_tx_stats(struct xgene_enet_pdata *pdata);
|
|
+
|
|
+int xgene_enet_mdio_config(struct xgene_enet_pdata *pdata);
|
|
+int xgene_enet_mdio_remove(struct xgene_enet_pdata *pdata);
|
|
+
|
|
+#endif /* __XGENE_ENET_HW_H__ */
|
|
diff --git a/drivers/net/ethernet/apm/xgene/xgene_enet_main.c b/drivers/net/ethernet/apm/xgene/xgene_enet_main.c
|
|
new file mode 100644
|
|
index 0000000..756523a
|
|
--- /dev/null
|
|
+++ b/drivers/net/ethernet/apm/xgene/xgene_enet_main.c
|
|
@@ -0,0 +1,962 @@
|
|
+/* Applied Micro X-Gene SoC Ethernet Driver
|
|
+ *
|
|
+ * Copyright (c) 2014, Applied Micro Circuits Corporation
|
|
+ * Authors: Iyappan Subramanian <isubramanian@apm.com>
|
|
+ * Ravi Patel <rapatel@apm.com>
|
|
+ * Keyur Chudgar <kchudgar@apm.com>
|
|
+ *
|
|
+ * This program is free software; you can redistribute it and/or modify it
|
|
+ * under the terms of the GNU General Public License as published by the
|
|
+ * Free Software Foundation; either version 2 of the License, or (at your
|
|
+ * option) any later version.
|
|
+ *
|
|
+ * This program is distributed in the hope that it will be useful,
|
|
+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
+ * GNU General Public License for more details.
|
|
+ *
|
|
+ * You should have received a copy of the GNU General Public License
|
|
+ * along with this program. If not, see <http://www.gnu.org/licenses/>.
|
|
+ */
|
|
+
|
|
+#include "xgene_enet_main.h"
|
|
+#include "xgene_enet_hw.h"
|
|
+
|
|
+static void xgene_enet_init_bufpool(struct xgene_enet_desc_ring *buf_pool)
|
|
+{
|
|
+ struct xgene_enet_raw_desc16 *raw_desc;
|
|
+ int i;
|
|
+
|
|
+ for (i = 0; i < buf_pool->slots; i++) {
|
|
+ raw_desc = &buf_pool->raw_desc16[i];
|
|
+
|
|
+ /* Hardware expects descriptor in little endian format */
|
|
+ raw_desc->m0 = cpu_to_le64(i |
|
|
+ (((u64)buf_pool->dst_ring_num << FPQNUM_POS) &
|
|
+ FPQNUM_MASK) | STASHING_MASK);
|
|
+ }
|
|
+}
|
|
+
|
|
+static struct device *ndev_to_dev(struct net_device *ndev)
|
|
+{
|
|
+ return ndev->dev.parent;
|
|
+}
|
|
+
|
|
+static int xgene_enet_refill_bufpool(struct xgene_enet_desc_ring *buf_pool,
|
|
+ u32 nbuf)
|
|
+{
|
|
+ struct sk_buff *skb;
|
|
+ struct xgene_enet_raw_desc16 *raw_desc;
|
|
+ struct net_device *ndev;
|
|
+ struct device *dev;
|
|
+ dma_addr_t dma_addr;
|
|
+ u32 tail = buf_pool->tail;
|
|
+ u32 slots = buf_pool->slots - 1;
|
|
+ u16 bufdatalen, len;
|
|
+ int i;
|
|
+
|
|
+ ndev = buf_pool->ndev;
|
|
+ dev = ndev_to_dev(buf_pool->ndev);
|
|
+ bufdatalen = BUF_LEN_CODE_2K | (SKB_BUFFER_SIZE & GENMASK(11, 0));
|
|
+ len = XGENE_ENET_MAX_MTU;
|
|
+
|
|
+ for (i = 0; i < nbuf; i++) {
|
|
+ raw_desc = &buf_pool->raw_desc16[tail];
|
|
+
|
|
+ skb = netdev_alloc_skb_ip_align(ndev, len);
|
|
+ if (unlikely(!skb))
|
|
+ return -ENOMEM;
|
|
+ buf_pool->rx_skb[tail] = skb;
|
|
+
|
|
+ dma_addr = dma_map_single(dev, skb->data, len, DMA_FROM_DEVICE);
|
|
+ if (dma_mapping_error(dev, dma_addr)) {
|
|
+ netdev_err(ndev, "DMA mapping error\n");
|
|
+ dev_kfree_skb_any(skb);
|
|
+ return -EINVAL;
|
|
+ }
|
|
+
|
|
+ raw_desc->m1 = cpu_to_le64((dma_addr & DATAADDR_MASK) |
|
|
+ (((u64)bufdatalen << BUFDATALEN_POS) &
|
|
+ BUFDATALEN_MASK) | COHERENT_MASK);
|
|
+ tail = (tail + 1) & slots;
|
|
+ }
|
|
+
|
|
+ iowrite32(nbuf, buf_pool->cmd);
|
|
+ buf_pool->tail = tail;
|
|
+
|
|
+ return 0;
|
|
+}
|
|
+
|
|
+static u16 xgene_enet_dst_ring_num(struct xgene_enet_desc_ring *ring)
|
|
+{
|
|
+ struct xgene_enet_pdata *pdata = netdev_priv(ring->ndev);
|
|
+
|
|
+ return ((u16)pdata->rm << 10) | ring->num;
|
|
+}
|
|
+
|
|
+static u8 xgene_enet_hdr_len(const void *data)
|
|
+{
|
|
+ const struct ethhdr *eth = data;
|
|
+
|
|
+ return (eth->h_proto == htons(ETH_P_8021Q)) ? VLAN_ETH_HLEN : ETH_HLEN;
|
|
+}
|
|
+
|
|
+static u32 xgene_enet_ring_len(struct xgene_enet_desc_ring *ring)
|
|
+{
|
|
+ u32 *cmd_base = ring->cmd_base;
|
|
+ u32 ring_state, num_msgs;
|
|
+
|
|
+ ring_state = ioread32(&cmd_base[1]);
|
|
+ num_msgs = ring_state & CREATE_MASK(NUMMSGSINQ_POS, NUMMSGSINQ_LEN);
|
|
+
|
|
+ return num_msgs >> NUMMSGSINQ_POS;
|
|
+}
|
|
+
|
|
+static void xgene_enet_delete_bufpool(struct xgene_enet_desc_ring *buf_pool)
|
|
+{
|
|
+ struct xgene_enet_raw_desc16 *raw_desc;
|
|
+ u32 slots = buf_pool->slots - 1;
|
|
+ u32 tail = buf_pool->tail;
|
|
+ u32 userinfo;
|
|
+ int i, len;
|
|
+
|
|
+ len = xgene_enet_ring_len(buf_pool);
|
|
+ for (i = 0; i < len; i++) {
|
|
+ tail = (tail - 1) & slots;
|
|
+ raw_desc = &buf_pool->raw_desc16[tail];
|
|
+
|
|
+ /* Hardware stores descriptor in little endian format */
|
|
+ userinfo = le64_to_cpu(raw_desc->m0) & USERINFO_MASK;
|
|
+ dev_kfree_skb_any(buf_pool->rx_skb[userinfo]);
|
|
+ }
|
|
+
|
|
+ iowrite32(-len, buf_pool->cmd);
|
|
+ buf_pool->tail = tail;
|
|
+}
|
|
+
|
|
+static irqreturn_t xgene_enet_rx_irq(const int irq, void *data)
|
|
+{
|
|
+ struct xgene_enet_desc_ring *rx_ring = data;
|
|
+
|
|
+ if (napi_schedule_prep(&rx_ring->napi)) {
|
|
+ disable_irq_nosync(irq);
|
|
+ __napi_schedule(&rx_ring->napi);
|
|
+ }
|
|
+
|
|
+ return IRQ_HANDLED;
|
|
+}
|
|
+
|
|
+static int xgene_enet_tx_completion(struct xgene_enet_desc_ring *cp_ring,
|
|
+ struct xgene_enet_raw_desc *raw_desc)
|
|
+{
|
|
+ struct sk_buff *skb;
|
|
+ struct device *dev;
|
|
+ u16 skb_index;
|
|
+ u8 status;
|
|
+ int ret = 0;
|
|
+
|
|
+ skb_index = raw_desc->m0 & USERINFO_MASK;
|
|
+ skb = cp_ring->cp_skb[skb_index];
|
|
+
|
|
+ dev = ndev_to_dev(cp_ring->ndev);
|
|
+ dma_unmap_single(dev, raw_desc->m1 & DATAADDR_MASK,
|
|
+ (raw_desc->m1 & BUFDATALEN_MASK) >> BUFDATALEN_POS,
|
|
+ DMA_TO_DEVICE);
|
|
+
|
|
+ /* Checking for error */
|
|
+ status = (raw_desc->m0 & LERR_MASK) >> LERR_POS;
|
|
+ if (unlikely(status > 2)) {
|
|
+ xgene_enet_parse_error(cp_ring, netdev_priv(cp_ring->ndev),
|
|
+ status);
|
|
+ ret = -1;
|
|
+ }
|
|
+
|
|
+ if (likely(skb)) {
|
|
+ dev_kfree_skb_any(skb);
|
|
+ } else {
|
|
+ netdev_err(cp_ring->ndev, "completion skb is NULL\n");
|
|
+ ret = -1;
|
|
+ }
|
|
+
|
|
+ return ret;
|
|
+}
|
|
+
|
|
+static u64 xgene_enet_work_msg(struct sk_buff *skb)
|
|
+{
|
|
+ struct iphdr *iph;
|
|
+ u8 l3hlen, l4hlen = 0;
|
|
+ u8 csum_enable = 0;
|
|
+ u8 proto = 0;
|
|
+ u8 ethhdr;
|
|
+ u64 hopinfo;
|
|
+
|
|
+ if (unlikely(skb->protocol != htons(ETH_P_IP)) &&
|
|
+ unlikely(skb->protocol != htons(ETH_P_8021Q)))
|
|
+ goto out;
|
|
+
|
|
+ if (unlikely(!(skb->dev->features & NETIF_F_IP_CSUM)))
|
|
+ goto out;
|
|
+
|
|
+ iph = ip_hdr(skb);
|
|
+ if (unlikely(ip_is_fragment(iph)))
|
|
+ goto out;
|
|
+
|
|
+ if (likely(iph->protocol == IPPROTO_TCP)) {
|
|
+ l4hlen = tcp_hdrlen(skb) >> 2;
|
|
+ csum_enable = 1;
|
|
+ proto = TSO_IPPROTO_TCP;
|
|
+ } else if (iph->protocol == IPPROTO_UDP) {
|
|
+ l4hlen = UDP_HDR_SIZE;
|
|
+ csum_enable = 1;
|
|
+ }
|
|
+out:
|
|
+ l3hlen = ip_hdrlen(skb) >> 2;
|
|
+ ethhdr = xgene_enet_hdr_len(skb->data);
|
|
+ hopinfo = (l4hlen & TCPHDR_MASK) |
|
|
+ ((l3hlen << IPHDR_POS) & IPHDR_MASK) |
|
|
+ (ethhdr << ETHHDR_POS) |
|
|
+ (csum_enable << EC_POS) |
|
|
+ (proto << IS_POS) |
|
|
+ INSERT_CRC |
|
|
+ TYPE_ETH_WORK_MESSAGE;
|
|
+
|
|
+ return hopinfo;
|
|
+}
|
|
+
|
|
+static int xgene_enet_setup_tx_desc(struct xgene_enet_desc_ring *tx_ring,
|
|
+ struct sk_buff *skb)
|
|
+{
|
|
+ struct device *dev = ndev_to_dev(tx_ring->ndev);
|
|
+ struct xgene_enet_raw_desc *raw_desc;
|
|
+ dma_addr_t dma_addr;
|
|
+ u16 tail = tx_ring->tail;
|
|
+ u64 hopinfo;
|
|
+
|
|
+ raw_desc = &tx_ring->raw_desc[tail];
|
|
+ memset(raw_desc, 0, sizeof(struct xgene_enet_raw_desc));
|
|
+
|
|
+ dma_addr = dma_map_single(dev, skb->data, skb->len, DMA_TO_DEVICE);
|
|
+ if (dma_mapping_error(dev, dma_addr)) {
|
|
+ netdev_err(tx_ring->ndev, "DMA mapping error\n");
|
|
+ return -EINVAL;
|
|
+ }
|
|
+
|
|
+ /* Hardware expects descriptor in little endian format */
|
|
+ raw_desc->m0 = cpu_to_le64(tail);
|
|
+ raw_desc->m1 = cpu_to_le64((dma_addr & DATAADDR_MASK) |
|
|
+ (((u64)skb->len << BUFDATALEN_POS) & BUFDATALEN_MASK) |
|
|
+ COHERENT_MASK);
|
|
+ hopinfo = xgene_enet_work_msg(skb);
|
|
+ raw_desc->m3 = cpu_to_le64(
|
|
+ (((u64)tx_ring->dst_ring_num << HENQNUM_POS) &
|
|
+ HENQNUM_MASK) | hopinfo);
|
|
+ tx_ring->cp_ring->cp_skb[tail] = skb;
|
|
+
|
|
+ return 0;
|
|
+}
|
|
+
|
|
+static netdev_tx_t xgene_enet_start_xmit(struct sk_buff *skb,
|
|
+ struct net_device *ndev)
|
|
+{
|
|
+ struct xgene_enet_pdata *pdata = netdev_priv(ndev);
|
|
+ struct xgene_enet_desc_ring *tx_ring = pdata->tx_ring;
|
|
+ struct xgene_enet_desc_ring *cp_ring = tx_ring->cp_ring;
|
|
+ u32 tx_level, cq_level;
|
|
+
|
|
+ tx_level = xgene_enet_ring_len(tx_ring);
|
|
+ cq_level = xgene_enet_ring_len(cp_ring);
|
|
+ if (unlikely(tx_level > pdata->tx_qcnt_hi ||
|
|
+ cq_level > pdata->cp_qcnt_hi)) {
|
|
+ netif_stop_queue(ndev);
|
|
+ return NETDEV_TX_BUSY;
|
|
+ }
|
|
+
|
|
+ if (xgene_enet_setup_tx_desc(tx_ring, skb)) {
|
|
+ dev_kfree_skb_any(skb);
|
|
+ return NETDEV_TX_OK;
|
|
+ }
|
|
+
|
|
+ iowrite32(1, tx_ring->cmd);
|
|
+ skb_tx_timestamp(skb);
|
|
+ tx_ring->tail = (tx_ring->tail + 1) & (tx_ring->slots - 1);
|
|
+
|
|
+ pdata->stats.tx_packets++;
|
|
+ pdata->stats.tx_bytes += skb->len;
|
|
+
|
|
+ return NETDEV_TX_OK;
|
|
+}
|
|
+
|
|
+static void xgene_enet_skip_csum(struct sk_buff *skb)
|
|
+{
|
|
+ struct iphdr *iph = ip_hdr(skb);
|
|
+
|
|
+ if (!ip_is_fragment(iph) ||
|
|
+ (iph->protocol != IPPROTO_TCP && iph->protocol != IPPROTO_UDP)) {
|
|
+ skb->ip_summed = CHECKSUM_UNNECESSARY;
|
|
+ }
|
|
+}
|
|
+
|
|
+static int xgene_enet_rx_frame(struct xgene_enet_desc_ring *rx_ring,
|
|
+ struct xgene_enet_raw_desc *raw_desc)
|
|
+{
|
|
+ struct net_device *ndev;
|
|
+ struct xgene_enet_pdata *pdata;
|
|
+ struct device *dev;
|
|
+ struct xgene_enet_desc_ring *buf_pool;
|
|
+ u32 datalen, skb_index;
|
|
+ struct sk_buff *skb;
|
|
+ u8 status;
|
|
+ int ret = 0;
|
|
+
|
|
+ ndev = rx_ring->ndev;
|
|
+ pdata = netdev_priv(ndev);
|
|
+ dev = ndev_to_dev(rx_ring->ndev);
|
|
+ buf_pool = rx_ring->buf_pool;
|
|
+
|
|
+ dma_unmap_single(dev, raw_desc->m1 & DATAADDR_MASK, XGENE_ENET_MAX_MTU,
|
|
+ DMA_FROM_DEVICE);
|
|
+ skb_index = raw_desc->m0 & USERINFO_MASK;
|
|
+ skb = buf_pool->rx_skb[skb_index];
|
|
+
|
|
+ /* checking for error */
|
|
+ status = (raw_desc->m0 & LERR_MASK) >> LERR_POS;
|
|
+ if (unlikely(status > 2)) {
|
|
+ dev_kfree_skb_any(skb);
|
|
+ xgene_enet_parse_error(rx_ring, netdev_priv(rx_ring->ndev),
|
|
+ status);
|
|
+ pdata->stats.rx_dropped++;
|
|
+ ret = -1;
|
|
+ goto out;
|
|
+ }
|
|
+
|
|
+ /* strip off CRC as HW isn't doing this */
|
|
+ datalen = (raw_desc->m1 & BUFDATALEN_MASK) >> BUFDATALEN_POS;
|
|
+ datalen -= 4;
|
|
+ prefetch(skb->data - NET_IP_ALIGN);
|
|
+ skb_put(skb, datalen);
|
|
+
|
|
+ skb_checksum_none_assert(skb);
|
|
+ skb->protocol = eth_type_trans(skb, ndev);
|
|
+ if (likely((ndev->features & NETIF_F_IP_CSUM) &&
|
|
+ skb->protocol == htons(ETH_P_IP))) {
|
|
+ xgene_enet_skip_csum(skb);
|
|
+ }
|
|
+
|
|
+ pdata->stats.rx_packets++;
|
|
+ pdata->stats.rx_bytes += datalen;
|
|
+ napi_gro_receive(&rx_ring->napi, skb);
|
|
+out:
|
|
+ if (--rx_ring->nbufpool == 0) {
|
|
+ ret = xgene_enet_refill_bufpool(buf_pool, NUM_BUFPOOL);
|
|
+ rx_ring->nbufpool = NUM_BUFPOOL;
|
|
+ }
|
|
+
|
|
+ return ret;
|
|
+}
|
|
+
|
|
+static bool is_rx_desc(struct xgene_enet_raw_desc *raw_desc)
|
|
+{
|
|
+ /* Hardware stores descriptor in little endian format */
|
|
+ raw_desc->m0 = le64_to_cpu(raw_desc->m0);
|
|
+ raw_desc->m1 = le64_to_cpu(raw_desc->m1);
|
|
+ return ((raw_desc->m0 & FPQNUM_MASK) >> FPQNUM_POS) ? true : false;
|
|
+}
|
|
+
|
|
+static int xgene_enet_process_ring(struct xgene_enet_desc_ring *ring,
|
|
+ int budget)
|
|
+{
|
|
+ struct xgene_enet_pdata *pdata = netdev_priv(ring->ndev);
|
|
+ struct xgene_enet_raw_desc *raw_desc;
|
|
+ u16 head = ring->head;
|
|
+ u16 slots = ring->slots - 1;
|
|
+ int ret, count = 0;
|
|
+
|
|
+ do {
|
|
+ raw_desc = &ring->raw_desc[head];
|
|
+ if (unlikely(((u64 *)raw_desc)[EMPTY_SLOT_INDEX] == EMPTY_SLOT))
|
|
+ break;
|
|
+
|
|
+ if (is_rx_desc(raw_desc))
|
|
+ ret = xgene_enet_rx_frame(ring, raw_desc);
|
|
+ else
|
|
+ ret = xgene_enet_tx_completion(ring, raw_desc);
|
|
+ ((u64 *)raw_desc)[EMPTY_SLOT_INDEX] = EMPTY_SLOT;
|
|
+
|
|
+ head = (head + 1) & slots;
|
|
+ count++;
|
|
+
|
|
+ if (ret)
|
|
+ break;
|
|
+ } while (--budget);
|
|
+
|
|
+ if (likely(count)) {
|
|
+ iowrite32(-count, ring->cmd);
|
|
+ ring->head = head;
|
|
+
|
|
+ if (netif_queue_stopped(ring->ndev)) {
|
|
+ if (xgene_enet_ring_len(ring) < pdata->cp_qcnt_low)
|
|
+ netif_wake_queue(ring->ndev);
|
|
+ }
|
|
+ }
|
|
+
|
|
+ return budget;
|
|
+}
|
|
+
|
|
+static int xgene_enet_napi(struct napi_struct *napi, const int budget)
|
|
+{
|
|
+ struct xgene_enet_desc_ring *ring;
|
|
+ int processed;
|
|
+
|
|
+ ring = container_of(napi, struct xgene_enet_desc_ring, napi);
|
|
+ processed = xgene_enet_process_ring(ring, budget);
|
|
+
|
|
+ if (processed != budget) {
|
|
+ napi_complete(napi);
|
|
+ enable_irq(ring->irq);
|
|
+ }
|
|
+
|
|
+ return processed;
|
|
+}
|
|
+
|
|
+static void xgene_enet_timeout(struct net_device *ndev)
|
|
+{
|
|
+ struct xgene_enet_pdata *pdata = netdev_priv(ndev);
|
|
+
|
|
+ xgene_gmac_reset(pdata);
|
|
+}
|
|
+
|
|
+static int xgene_enet_register_irq(struct net_device *ndev)
|
|
+{
|
|
+ struct xgene_enet_pdata *pdata = netdev_priv(ndev);
|
|
+ struct device *dev = ndev_to_dev(ndev);
|
|
+ int ret;
|
|
+
|
|
+ ret = devm_request_irq(dev, pdata->rx_ring->irq, xgene_enet_rx_irq,
|
|
+ IRQF_SHARED, ndev->name, pdata->rx_ring);
|
|
+ if (ret) {
|
|
+ netdev_err(ndev, "rx%d interrupt request failed\n",
|
|
+ pdata->rx_ring->irq);
|
|
+ }
|
|
+
|
|
+ return ret;
|
|
+}
|
|
+
|
|
+static void xgene_enet_free_irq(struct net_device *ndev)
|
|
+{
|
|
+ struct xgene_enet_pdata *pdata;
|
|
+ struct device *dev;
|
|
+
|
|
+ pdata = netdev_priv(ndev);
|
|
+ dev = ndev_to_dev(ndev);
|
|
+ devm_free_irq(dev, pdata->rx_ring->irq, pdata->rx_ring);
|
|
+}
|
|
+
|
|
+static int xgene_enet_open(struct net_device *ndev)
|
|
+{
|
|
+ struct xgene_enet_pdata *pdata = netdev_priv(ndev);
|
|
+ int ret;
|
|
+
|
|
+ xgene_gmac_tx_enable(pdata);
|
|
+ xgene_gmac_rx_enable(pdata);
|
|
+
|
|
+ ret = xgene_enet_register_irq(ndev);
|
|
+ if (ret)
|
|
+ return ret;
|
|
+ napi_enable(&pdata->rx_ring->napi);
|
|
+
|
|
+ if (pdata->phy_dev)
|
|
+ phy_start(pdata->phy_dev);
|
|
+
|
|
+ netif_start_queue(ndev);
|
|
+
|
|
+ return ret;
|
|
+}
|
|
+
|
|
+static int xgene_enet_close(struct net_device *ndev)
|
|
+{
|
|
+ struct xgene_enet_pdata *pdata = netdev_priv(ndev);
|
|
+
|
|
+ netif_stop_queue(ndev);
|
|
+
|
|
+ if (pdata->phy_dev)
|
|
+ phy_stop(pdata->phy_dev);
|
|
+
|
|
+ napi_disable(&pdata->rx_ring->napi);
|
|
+ xgene_enet_free_irq(ndev);
|
|
+ xgene_enet_process_ring(pdata->rx_ring, -1);
|
|
+
|
|
+ xgene_gmac_tx_disable(pdata);
|
|
+ xgene_gmac_rx_disable(pdata);
|
|
+
|
|
+ return 0;
|
|
+}
|
|
+
|
|
+static void xgene_enet_delete_ring(struct xgene_enet_desc_ring *ring)
|
|
+{
|
|
+ struct xgene_enet_pdata *pdata;
|
|
+ struct device *dev;
|
|
+
|
|
+ pdata = netdev_priv(ring->ndev);
|
|
+ dev = ndev_to_dev(ring->ndev);
|
|
+
|
|
+ xgene_enet_clear_ring(ring);
|
|
+ dma_free_coherent(dev, ring->size, ring->desc_addr, ring->dma);
|
|
+}
|
|
+
|
|
+static void xgene_enet_delete_desc_rings(struct xgene_enet_pdata *pdata)
|
|
+{
|
|
+ struct xgene_enet_desc_ring *buf_pool;
|
|
+
|
|
+ if (pdata->tx_ring) {
|
|
+ xgene_enet_delete_ring(pdata->tx_ring);
|
|
+ pdata->tx_ring = NULL;
|
|
+ }
|
|
+
|
|
+ if (pdata->rx_ring) {
|
|
+ buf_pool = pdata->rx_ring->buf_pool;
|
|
+ xgene_enet_delete_bufpool(buf_pool);
|
|
+ xgene_enet_delete_ring(buf_pool);
|
|
+ xgene_enet_delete_ring(pdata->rx_ring);
|
|
+ pdata->rx_ring = NULL;
|
|
+ }
|
|
+}
|
|
+
|
|
+static int xgene_enet_get_ring_size(struct device *dev,
|
|
+ enum xgene_enet_ring_cfgsize cfgsize)
|
|
+{
|
|
+ int size = -EINVAL;
|
|
+
|
|
+ switch (cfgsize) {
|
|
+ case RING_CFGSIZE_512B:
|
|
+ size = 0x200;
|
|
+ break;
|
|
+ case RING_CFGSIZE_2KB:
|
|
+ size = 0x800;
|
|
+ break;
|
|
+ case RING_CFGSIZE_16KB:
|
|
+ size = 0x4000;
|
|
+ break;
|
|
+ case RING_CFGSIZE_64KB:
|
|
+ size = 0x10000;
|
|
+ break;
|
|
+ case RING_CFGSIZE_512KB:
|
|
+ size = 0x80000;
|
|
+ break;
|
|
+ default:
|
|
+ dev_err(dev, "Unsupported cfg ring size %d\n", cfgsize);
|
|
+ break;
|
|
+ }
|
|
+
|
|
+ return size;
|
|
+}
|
|
+
|
|
+static void xgene_enet_free_desc_ring(struct xgene_enet_desc_ring *ring)
|
|
+{
|
|
+ struct device *dev;
|
|
+
|
|
+ if (!ring)
|
|
+ return;
|
|
+
|
|
+ dev = ndev_to_dev(ring->ndev);
|
|
+
|
|
+ if (ring->desc_addr) {
|
|
+ xgene_enet_clear_ring(ring);
|
|
+ dma_free_coherent(dev, ring->size, ring->desc_addr, ring->dma);
|
|
+ }
|
|
+ devm_kfree(dev, ring);
|
|
+}
|
|
+
|
|
+static void xgene_enet_free_desc_rings(struct xgene_enet_pdata *pdata)
|
|
+{
|
|
+ struct device *dev = &pdata->pdev->dev;
|
|
+ struct xgene_enet_desc_ring *ring;
|
|
+
|
|
+ ring = pdata->tx_ring;
|
|
+ if (ring && ring->cp_ring && ring->cp_ring->cp_skb)
|
|
+ devm_kfree(dev, ring->cp_ring->cp_skb);
|
|
+ xgene_enet_free_desc_ring(ring);
|
|
+
|
|
+ ring = pdata->rx_ring;
|
|
+ if (ring && ring->buf_pool && ring->buf_pool->rx_skb)
|
|
+ devm_kfree(dev, ring->buf_pool->rx_skb);
|
|
+ xgene_enet_free_desc_ring(ring->buf_pool);
|
|
+ xgene_enet_free_desc_ring(ring);
|
|
+}
|
|
+
|
|
+static struct xgene_enet_desc_ring *xgene_enet_create_desc_ring(
|
|
+ struct net_device *ndev, u32 ring_num,
|
|
+ enum xgene_enet_ring_cfgsize cfgsize, u32 ring_id)
|
|
+{
|
|
+ struct xgene_enet_desc_ring *ring;
|
|
+ struct xgene_enet_pdata *pdata = netdev_priv(ndev);
|
|
+ struct device *dev = ndev_to_dev(ndev);
|
|
+ u32 size;
|
|
+
|
|
+ ring = devm_kzalloc(dev, sizeof(struct xgene_enet_desc_ring),
|
|
+ GFP_KERNEL);
|
|
+ if (!ring)
|
|
+ return NULL;
|
|
+
|
|
+ ring->ndev = ndev;
|
|
+ ring->num = ring_num;
|
|
+ ring->cfgsize = cfgsize;
|
|
+ ring->id = ring_id;
|
|
+
|
|
+ size = xgene_enet_get_ring_size(dev, cfgsize);
|
|
+ ring->desc_addr = dma_zalloc_coherent(dev, size, &ring->dma,
|
|
+ GFP_KERNEL);
|
|
+ if (!ring->desc_addr) {
|
|
+ devm_kfree(dev, ring);
|
|
+ return NULL;
|
|
+ }
|
|
+ ring->size = size;
|
|
+
|
|
+ ring->cmd_base = pdata->ring_cmd_addr + (ring->num << 6);
|
|
+ ring->cmd = ring->cmd_base + INC_DEC_CMD_ADDR;
|
|
+ pdata->rm = RM3;
|
|
+ ring = xgene_enet_setup_ring(ring);
|
|
+ netdev_dbg(ndev, "ring info: num=%d size=%d id=%d slots=%d\n",
|
|
+ ring->num, ring->size, ring->id, ring->slots);
|
|
+
|
|
+ return ring;
|
|
+}
|
|
+
|
|
+static u16 xgene_enet_get_ring_id(enum xgene_ring_owner owner, u8 bufnum)
|
|
+{
|
|
+ return (owner << 6) | (bufnum & GENMASK(5, 0));
|
|
+}
|
|
+
|
|
+static int xgene_enet_create_desc_rings(struct net_device *ndev)
|
|
+{
|
|
+ struct xgene_enet_pdata *pdata = netdev_priv(ndev);
|
|
+ struct device *dev = ndev_to_dev(ndev);
|
|
+ struct xgene_enet_desc_ring *rx_ring, *tx_ring, *cp_ring;
|
|
+ struct xgene_enet_desc_ring *buf_pool = NULL;
|
|
+ u8 cpu_bufnum = 0, eth_bufnum = 0;
|
|
+ u8 bp_bufnum = 0x20;
|
|
+ u16 ring_id, ring_num = 0;
|
|
+ int ret;
|
|
+
|
|
+ /* allocate rx descriptor ring */
|
|
+ ring_id = xgene_enet_get_ring_id(RING_OWNER_CPU, cpu_bufnum++);
|
|
+ rx_ring = xgene_enet_create_desc_ring(ndev, ring_num++,
|
|
+ RING_CFGSIZE_16KB, ring_id);
|
|
+ if (!rx_ring) {
|
|
+ ret = -ENOMEM;
|
|
+ goto err;
|
|
+ }
|
|
+
|
|
+ /* allocate buffer pool for receiving packets */
|
|
+ ring_id = xgene_enet_get_ring_id(RING_OWNER_ETH0, bp_bufnum++);
|
|
+ buf_pool = xgene_enet_create_desc_ring(ndev, ring_num++,
|
|
+ RING_CFGSIZE_2KB, ring_id);
|
|
+ if (!buf_pool) {
|
|
+ ret = -ENOMEM;
|
|
+ goto err;
|
|
+ }
|
|
+
|
|
+ rx_ring->nbufpool = NUM_BUFPOOL;
|
|
+ rx_ring->buf_pool = buf_pool;
|
|
+ rx_ring->irq = pdata->rx_irq;
|
|
+ buf_pool->rx_skb = devm_kcalloc(dev, buf_pool->slots,
|
|
+ sizeof(struct sk_buff *), GFP_KERNEL);
|
|
+ if (!buf_pool->rx_skb) {
|
|
+ ret = -ENOMEM;
|
|
+ goto err;
|
|
+ }
|
|
+
|
|
+ buf_pool->dst_ring_num = xgene_enet_dst_ring_num(buf_pool);
|
|
+ rx_ring->buf_pool = buf_pool;
|
|
+ pdata->rx_ring = rx_ring;
|
|
+
|
|
+ /* allocate tx descriptor ring */
|
|
+ ring_id = xgene_enet_get_ring_id(RING_OWNER_ETH0, eth_bufnum++);
|
|
+ tx_ring = xgene_enet_create_desc_ring(ndev, ring_num++,
|
|
+ RING_CFGSIZE_16KB, ring_id);
|
|
+ if (!tx_ring) {
|
|
+ ret = -ENOMEM;
|
|
+ goto err;
|
|
+ }
|
|
+ pdata->tx_ring = tx_ring;
|
|
+
|
|
+ cp_ring = pdata->rx_ring;
|
|
+ cp_ring->cp_skb = devm_kcalloc(dev, tx_ring->slots,
|
|
+ sizeof(struct sk_buff *), GFP_KERNEL);
|
|
+ if (!cp_ring->cp_skb) {
|
|
+ ret = -ENOMEM;
|
|
+ goto err;
|
|
+ }
|
|
+ pdata->tx_ring->cp_ring = cp_ring;
|
|
+ pdata->tx_ring->dst_ring_num = xgene_enet_dst_ring_num(cp_ring);
|
|
+
|
|
+ pdata->tx_qcnt_hi = pdata->tx_ring->slots / 2;
|
|
+ pdata->cp_qcnt_hi = pdata->rx_ring->slots / 2;
|
|
+ pdata->cp_qcnt_low = pdata->cp_qcnt_hi / 2;
|
|
+
|
|
+ return 0;
|
|
+
|
|
+err:
|
|
+ xgene_enet_free_desc_rings(pdata);
|
|
+ return ret;
|
|
+}
|
|
+
|
|
+static struct rtnl_link_stats64 *xgene_enet_get_stats64(
|
|
+ struct net_device *ndev,
|
|
+ struct rtnl_link_stats64 *storage)
|
|
+{
|
|
+ struct xgene_enet_pdata *pdata = netdev_priv(ndev);
|
|
+ struct rtnl_link_stats64 *stats = &pdata->stats;
|
|
+
|
|
+ spin_lock(&pdata->stats_lock);
|
|
+ stats->rx_errors += stats->rx_length_errors +
|
|
+ stats->rx_crc_errors +
|
|
+ stats->rx_frame_errors +
|
|
+ stats->rx_fifo_errors;
|
|
+ memcpy(storage, &pdata->stats, sizeof(struct rtnl_link_stats64));
|
|
+ spin_unlock(&pdata->stats_lock);
|
|
+
|
|
+ return storage;
|
|
+}
|
|
+
|
|
+static int xgene_enet_set_mac_address(struct net_device *ndev, void *addr)
|
|
+{
|
|
+ struct xgene_enet_pdata *pdata = netdev_priv(ndev);
|
|
+ int ret;
|
|
+
|
|
+ ret = eth_mac_addr(ndev, addr);
|
|
+ if (ret)
|
|
+ return ret;
|
|
+ xgene_gmac_set_mac_addr(pdata);
|
|
+
|
|
+ return ret;
|
|
+}
|
|
+
|
|
+static const struct net_device_ops xgene_ndev_ops = {
|
|
+ .ndo_open = xgene_enet_open,
|
|
+ .ndo_stop = xgene_enet_close,
|
|
+ .ndo_start_xmit = xgene_enet_start_xmit,
|
|
+ .ndo_tx_timeout = xgene_enet_timeout,
|
|
+ .ndo_get_stats64 = xgene_enet_get_stats64,
|
|
+ .ndo_change_mtu = eth_change_mtu,
|
|
+ .ndo_set_mac_address = xgene_enet_set_mac_address,
|
|
+};
|
|
+
|
|
+static int xgene_enet_get_resources(struct xgene_enet_pdata *pdata)
|
|
+{
|
|
+ struct platform_device *pdev;
|
|
+ struct net_device *ndev;
|
|
+ struct device *dev;
|
|
+ struct resource *res;
|
|
+ void *base_addr;
|
|
+ const char *mac;
|
|
+ int ret;
|
|
+
|
|
+ pdev = pdata->pdev;
|
|
+ dev = &pdev->dev;
|
|
+ ndev = pdata->ndev;
|
|
+
|
|
+ res = platform_get_resource_byname(pdev, IORESOURCE_MEM, "enet_csr");
|
|
+ if (!res) {
|
|
+ dev_err(dev, "Resource enet_csr not defined\n");
|
|
+ return -ENODEV;
|
|
+ }
|
|
+ pdata->base_addr = devm_ioremap_resource(dev, res);
|
|
+ if (IS_ERR(pdata->base_addr)) {
|
|
+ dev_err(dev, "Unable to retrieve ENET Port CSR region\n");
|
|
+ return PTR_ERR(pdata->base_addr);
|
|
+ }
|
|
+
|
|
+ res = platform_get_resource_byname(pdev, IORESOURCE_MEM, "ring_csr");
|
|
+ if (!res) {
|
|
+ dev_err(dev, "Resource ring_csr not defined\n");
|
|
+ return -ENODEV;
|
|
+ }
|
|
+ pdata->ring_csr_addr = devm_ioremap_resource(dev, res);
|
|
+ if (IS_ERR(pdata->ring_csr_addr)) {
|
|
+ dev_err(dev, "Unable to retrieve ENET Ring CSR region\n");
|
|
+ return PTR_ERR(pdata->ring_csr_addr);
|
|
+ }
|
|
+
|
|
+ res = platform_get_resource_byname(pdev, IORESOURCE_MEM, "ring_cmd");
|
|
+ if (!res) {
|
|
+ dev_err(dev, "Resource ring_cmd not defined\n");
|
|
+ return -ENODEV;
|
|
+ }
|
|
+ pdata->ring_cmd_addr = devm_ioremap_resource(dev, res);
|
|
+ if (IS_ERR(pdata->ring_cmd_addr)) {
|
|
+ dev_err(dev, "Unable to retrieve ENET Ring command region\n");
|
|
+ return PTR_ERR(pdata->ring_cmd_addr);
|
|
+ }
|
|
+
|
|
+ ret = platform_get_irq(pdev, 0);
|
|
+ if (ret <= 0) {
|
|
+ dev_err(dev, "Unable to get ENET Rx IRQ\n");
|
|
+ ret = ret ? : -ENXIO;
|
|
+ return ret;
|
|
+ }
|
|
+ pdata->rx_irq = ret;
|
|
+
|
|
+ mac = of_get_mac_address(dev->of_node);
|
|
+ if (mac)
|
|
+ memcpy(ndev->dev_addr, mac, ndev->addr_len);
|
|
+ else
|
|
+ eth_hw_addr_random(ndev);
|
|
+ memcpy(ndev->perm_addr, ndev->dev_addr, ndev->addr_len);
|
|
+
|
|
+ pdata->phy_mode = of_get_phy_mode(pdev->dev.of_node);
|
|
+ if (pdata->phy_mode < 0) {
|
|
+ dev_err(dev, "Incorrect phy-connection-type in DTS\n");
|
|
+ return -EINVAL;
|
|
+ }
|
|
+
|
|
+ pdata->clk = devm_clk_get(&pdev->dev, NULL);
|
|
+ ret = IS_ERR(pdata->clk);
|
|
+ if (IS_ERR(pdata->clk)) {
|
|
+ dev_err(&pdev->dev, "can't get clock\n");
|
|
+ ret = PTR_ERR(pdata->clk);
|
|
+ return ret;
|
|
+ }
|
|
+
|
|
+ base_addr = pdata->base_addr;
|
|
+ pdata->eth_csr_addr = base_addr + BLOCK_ETH_CSR_OFFSET;
|
|
+ pdata->eth_ring_if_addr = base_addr + BLOCK_ETH_RING_IF_OFFSET;
|
|
+ pdata->eth_diag_csr_addr = base_addr + BLOCK_ETH_DIAG_CSR_OFFSET;
|
|
+ pdata->mcx_mac_addr = base_addr + BLOCK_ETH_MAC_OFFSET;
|
|
+ pdata->mcx_stats_addr = base_addr + BLOCK_ETH_STATS_OFFSET;
|
|
+ pdata->mcx_mac_csr_addr = base_addr + BLOCK_ETH_MAC_CSR_OFFSET;
|
|
+ pdata->rx_buff_cnt = NUM_PKT_BUF;
|
|
+
|
|
+ return ret;
|
|
+}
|
|
+
|
|
+static int xgene_enet_init_hw(struct xgene_enet_pdata *pdata)
|
|
+{
|
|
+ struct net_device *ndev = pdata->ndev;
|
|
+ struct xgene_enet_desc_ring *buf_pool;
|
|
+ u16 dst_ring_num;
|
|
+ int ret;
|
|
+
|
|
+ xgene_gmac_tx_disable(pdata);
|
|
+ xgene_gmac_rx_disable(pdata);
|
|
+
|
|
+ ret = xgene_enet_create_desc_rings(ndev);
|
|
+ if (ret) {
|
|
+ netdev_err(ndev, "Error in ring configuration\n");
|
|
+ return ret;
|
|
+ }
|
|
+
|
|
+ /* setup buffer pool */
|
|
+ buf_pool = pdata->rx_ring->buf_pool;
|
|
+ xgene_enet_init_bufpool(buf_pool);
|
|
+ ret = xgene_enet_refill_bufpool(buf_pool, pdata->rx_buff_cnt);
|
|
+ if (ret)
|
|
+ return ret;
|
|
+
|
|
+ dst_ring_num = xgene_enet_dst_ring_num(pdata->rx_ring);
|
|
+ xgene_enet_cle_bypass(pdata, dst_ring_num, buf_pool->id);
|
|
+
|
|
+ return ret;
|
|
+}
|
|
+
|
|
+static int xgene_enet_probe(struct platform_device *pdev)
|
|
+{
|
|
+ struct net_device *ndev;
|
|
+ struct xgene_enet_pdata *pdata;
|
|
+ struct device *dev = &pdev->dev;
|
|
+ struct napi_struct *napi;
|
|
+ int ret;
|
|
+
|
|
+ ndev = alloc_etherdev(sizeof(struct xgene_enet_pdata));
|
|
+ if (!ndev)
|
|
+ return -ENOMEM;
|
|
+
|
|
+ pdata = netdev_priv(ndev);
|
|
+
|
|
+ pdata->pdev = pdev;
|
|
+ pdata->ndev = ndev;
|
|
+ SET_NETDEV_DEV(ndev, dev);
|
|
+ platform_set_drvdata(pdev, pdata);
|
|
+ ndev->netdev_ops = &xgene_ndev_ops;
|
|
+ xgene_enet_set_ethtool_ops(ndev);
|
|
+ ndev->features |= NETIF_F_IP_CSUM |
|
|
+ NETIF_F_GSO |
|
|
+ NETIF_F_GRO;
|
|
+
|
|
+ ret = xgene_enet_get_resources(pdata);
|
|
+ if (ret)
|
|
+ goto err;
|
|
+
|
|
+ xgene_enet_reset(pdata);
|
|
+ xgene_gmac_init(pdata, SPEED_1000);
|
|
+
|
|
+ spin_lock_init(&pdata->stats_lock);
|
|
+ ret = register_netdev(ndev);
|
|
+ if (ret) {
|
|
+ netdev_err(ndev, "Failed to register netdev\n");
|
|
+ goto err;
|
|
+ }
|
|
+
|
|
+ ret = dma_set_mask_and_coherent(dev, DMA_BIT_MASK(64));
|
|
+ if (ret) {
|
|
+ netdev_err(ndev, "No usable DMA configuration\n");
|
|
+ goto err;
|
|
+ }
|
|
+
|
|
+ ret = xgene_enet_init_hw(pdata);
|
|
+ if (ret)
|
|
+ goto err;
|
|
+
|
|
+ napi = &pdata->rx_ring->napi;
|
|
+ netif_napi_add(ndev, napi, xgene_enet_napi, NAPI_POLL_WEIGHT);
|
|
+ ret = xgene_enet_mdio_config(pdata);
|
|
+
|
|
+ return ret;
|
|
+err:
|
|
+ free_netdev(ndev);
|
|
+ return ret;
|
|
+}
|
|
+
|
|
+static int xgene_enet_remove(struct platform_device *pdev)
|
|
+{
|
|
+ struct xgene_enet_pdata *pdata;
|
|
+ struct net_device *ndev;
|
|
+
|
|
+ pdata = platform_get_drvdata(pdev);
|
|
+ ndev = pdata->ndev;
|
|
+
|
|
+ xgene_gmac_rx_disable(pdata);
|
|
+ xgene_gmac_tx_disable(pdata);
|
|
+
|
|
+ netif_napi_del(&pdata->rx_ring->napi);
|
|
+ xgene_enet_mdio_remove(pdata);
|
|
+ xgene_enet_delete_desc_rings(pdata);
|
|
+ unregister_netdev(ndev);
|
|
+ xgene_gport_shutdown(pdata);
|
|
+ free_netdev(ndev);
|
|
+
|
|
+ return 0;
|
|
+}
|
|
+
|
|
+static struct of_device_id xgene_enet_match[] = {
|
|
+ {.compatible = "apm,xgene-enet",},
|
|
+ {},
|
|
+};
|
|
+
|
|
+MODULE_DEVICE_TABLE(of, xgene_enet_match);
|
|
+
|
|
+static struct platform_driver xgene_enet_driver = {
|
|
+ .driver = {
|
|
+ .name = "xgene-enet",
|
|
+ .owner = THIS_MODULE,
|
|
+ .of_match_table = xgene_enet_match,
|
|
+ },
|
|
+ .probe = xgene_enet_probe,
|
|
+ .remove = xgene_enet_remove,
|
|
+};
|
|
+
|
|
+module_platform_driver(xgene_enet_driver);
|
|
+
|
|
+MODULE_DESCRIPTION("APM X-Gene SoC Ethernet driver");
|
|
+MODULE_VERSION(XGENE_DRV_VERSION);
|
|
+MODULE_AUTHOR("Keyur Chudgar <kchudgar@apm.com>");
|
|
+MODULE_LICENSE("GPL");
|
|
diff --git a/drivers/net/ethernet/apm/xgene/xgene_enet_main.h b/drivers/net/ethernet/apm/xgene/xgene_enet_main.h
|
|
new file mode 100644
|
|
index 0000000..f4f7e4a
|
|
--- /dev/null
|
|
+++ b/drivers/net/ethernet/apm/xgene/xgene_enet_main.h
|
|
@@ -0,0 +1,107 @@
|
|
+/* Applied Micro X-Gene SoC Ethernet Driver
|
|
+ *
|
|
+ * Copyright (c) 2014, Applied Micro Circuits Corporation
|
|
+ * Authors: Iyappan Subramanian <isubramanian@apm.com>
|
|
+ * Ravi Patel <rapatel@apm.com>
|
|
+ * Keyur Chudgar <kchudgar@apm.com>
|
|
+ *
|
|
+ * This program is free software; you can redistribute it and/or modify it
|
|
+ * under the terms of the GNU General Public License as published by the
|
|
+ * Free Software Foundation; either version 2 of the License, or (at your
|
|
+ * option) any later version.
|
|
+ *
|
|
+ * This program is distributed in the hope that it will be useful,
|
|
+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
+ * GNU General Public License for more details.
|
|
+ *
|
|
+ * You should have received a copy of the GNU General Public License
|
|
+ * along with this program. If not, see <http://www.gnu.org/licenses/>.
|
|
+ */
|
|
+
|
|
+#ifndef __XGENE_ENET_MAIN_H__
|
|
+#define __XGENE_ENET_MAIN_H__
|
|
+
|
|
+#include <linux/clk.h>
|
|
+#include <linux/of_platform.h>
|
|
+#include <linux/of_net.h>
|
|
+#include <linux/of_mdio.h>
|
|
+#include <linux/module.h>
|
|
+#include <net/ip.h>
|
|
+#include <linux/prefetch.h>
|
|
+#include <linux/if_vlan.h>
|
|
+#include <linux/phy.h>
|
|
+#include "xgene_enet_hw.h"
|
|
+
|
|
+#define XGENE_DRV_VERSION "v1.0"
|
|
+#define XGENE_ENET_MAX_MTU 1536
|
|
+#define SKB_BUFFER_SIZE (XGENE_ENET_MAX_MTU - NET_IP_ALIGN)
|
|
+#define NUM_PKT_BUF 64
|
|
+#define NUM_BUFPOOL 32
|
|
+
|
|
+/* software context of a descriptor ring */
|
|
+struct xgene_enet_desc_ring {
|
|
+ struct net_device *ndev;
|
|
+ u16 id;
|
|
+ u16 num;
|
|
+ u16 head;
|
|
+ u16 tail;
|
|
+ u16 slots;
|
|
+ u16 irq;
|
|
+ u32 size;
|
|
+ u32 state[NUM_RING_CONFIG];
|
|
+ void __iomem *cmd_base;
|
|
+ void __iomem *cmd;
|
|
+ dma_addr_t dma;
|
|
+ u16 dst_ring_num;
|
|
+ u8 nbufpool;
|
|
+ struct sk_buff *(*rx_skb);
|
|
+ struct sk_buff *(*cp_skb);
|
|
+ enum xgene_enet_ring_cfgsize cfgsize;
|
|
+ struct xgene_enet_desc_ring *cp_ring;
|
|
+ struct xgene_enet_desc_ring *buf_pool;
|
|
+ struct napi_struct napi;
|
|
+ union {
|
|
+ void *desc_addr;
|
|
+ struct xgene_enet_raw_desc *raw_desc;
|
|
+ struct xgene_enet_raw_desc16 *raw_desc16;
|
|
+ };
|
|
+};
|
|
+
|
|
+/* ethernet private data */
|
|
+struct xgene_enet_pdata {
|
|
+ struct net_device *ndev;
|
|
+ struct mii_bus *mdio_bus;
|
|
+ struct phy_device *phy_dev;
|
|
+ int phy_speed;
|
|
+ struct clk *clk;
|
|
+ struct platform_device *pdev;
|
|
+ struct xgene_enet_desc_ring *tx_ring;
|
|
+ struct xgene_enet_desc_ring *rx_ring;
|
|
+ char *dev_name;
|
|
+ u32 rx_buff_cnt;
|
|
+ u32 tx_qcnt_hi;
|
|
+ u32 cp_qcnt_hi;
|
|
+ u32 cp_qcnt_low;
|
|
+ u32 rx_irq;
|
|
+ void __iomem *eth_csr_addr;
|
|
+ void __iomem *eth_ring_if_addr;
|
|
+ void __iomem *eth_diag_csr_addr;
|
|
+ void __iomem *mcx_mac_addr;
|
|
+ void __iomem *mcx_stats_addr;
|
|
+ void __iomem *mcx_mac_csr_addr;
|
|
+ void __iomem *base_addr;
|
|
+ void __iomem *ring_csr_addr;
|
|
+ void __iomem *ring_cmd_addr;
|
|
+ u32 phy_addr;
|
|
+ int phy_mode;
|
|
+ u32 speed;
|
|
+ u16 rm;
|
|
+ struct rtnl_link_stats64 stats;
|
|
+ /* statistics lock */
|
|
+ spinlock_t stats_lock;
|
|
+};
|
|
+
|
|
+void xgene_enet_set_ethtool_ops(struct net_device *netdev);
|
|
+
|
|
+#endif /* __XGENE_ENET_MAIN_H__ */
|
|
diff --git a/drivers/of/address.c b/drivers/of/address.c
|
|
index 5edfcb0..cbbaed2 100644
|
|
--- a/drivers/of/address.c
|
|
+++ b/drivers/of/address.c
|
|
@@ -5,6 +5,7 @@
|
|
#include <linux/module.h>
|
|
#include <linux/of_address.h>
|
|
#include <linux/pci_regs.h>
|
|
+#include <linux/slab.h>
|
|
#include <linux/string.h>
|
|
|
|
/* Max address size we deal with */
|
|
@@ -601,12 +602,72 @@ const __be32 *of_get_address(struct device_node *dev, int index, u64 *size,
|
|
}
|
|
EXPORT_SYMBOL(of_get_address);
|
|
|
|
+struct io_range {
|
|
+ struct list_head list;
|
|
+ phys_addr_t start;
|
|
+ resource_size_t size;
|
|
+};
|
|
+
|
|
+static LIST_HEAD(io_range_list);
|
|
+
|
|
+/*
|
|
+ * Record the PCI IO range (expressed as CPU physical address + size).
|
|
+ * Return a negative value if an error has occured, zero otherwise
|
|
+ */
|
|
+int __weak pci_register_io_range(phys_addr_t addr, resource_size_t size)
|
|
+{
|
|
+#ifdef PCI_IOBASE
|
|
+ struct io_range *res;
|
|
+ resource_size_t allocated_size = 0;
|
|
+
|
|
+ /* check if the range hasn't been previously recorded */
|
|
+ list_for_each_entry(res, &io_range_list, list) {
|
|
+ if (addr >= res->start && addr + size <= res->start + size)
|
|
+ return 0;
|
|
+ allocated_size += res->size;
|
|
+ }
|
|
+
|
|
+ /* range not registed yet, check for available space */
|
|
+ if (allocated_size + size - 1 > IO_SPACE_LIMIT)
|
|
+ return -E2BIG;
|
|
+
|
|
+ /* add the range to the list */
|
|
+ res = kzalloc(sizeof(*res), GFP_KERNEL);
|
|
+ if (!res)
|
|
+ return -ENOMEM;
|
|
+
|
|
+ res->start = addr;
|
|
+ res->size = size;
|
|
+
|
|
+ list_add_tail(&res->list, &io_range_list);
|
|
+
|
|
+ return 0;
|
|
+#else
|
|
+ return -EINVAL;
|
|
+#endif
|
|
+}
|
|
+
|
|
unsigned long __weak pci_address_to_pio(phys_addr_t address)
|
|
{
|
|
+#ifdef PCI_IOBASE
|
|
+ struct io_range *res;
|
|
+ resource_size_t offset = 0;
|
|
+
|
|
+ list_for_each_entry(res, &io_range_list, list) {
|
|
+ if (address >= res->start &&
|
|
+ address < res->start + res->size) {
|
|
+ return res->start - address + offset;
|
|
+ }
|
|
+ offset += res->size;
|
|
+ }
|
|
+
|
|
+ return (unsigned long)-1;
|
|
+#else
|
|
if (address > IO_SPACE_LIMIT)
|
|
return (unsigned long)-1;
|
|
|
|
return (unsigned long) address;
|
|
+#endif
|
|
}
|
|
|
|
static int __of_address_to_resource(struct device_node *dev,
|
|
@@ -811,3 +872,50 @@ bool of_dma_is_coherent(struct device_node *np)
|
|
return false;
|
|
}
|
|
EXPORT_SYMBOL_GPL(of_dma_is_coherent);
|
|
+
|
|
+/*
|
|
+ * of_pci_range_to_resource - Create a resource from an of_pci_range
|
|
+ * @range: the PCI range that describes the resource
|
|
+ * @np: device node where the range belongs to
|
|
+ * @res: pointer to a valid resource that will be updated to
|
|
+ * reflect the values contained in the range.
|
|
+ *
|
|
+ * Returns EINVAL if the range cannot be converted to resource.
|
|
+ *
|
|
+ * Note that if the range is an IO range, the resource will be converted
|
|
+ * using pci_address_to_pio() which can fail if it is called too early or
|
|
+ * if the range cannot be matched to any host bridge IO space (our case here).
|
|
+ * To guard against that we try to register the IO range first.
|
|
+ * If that fails we know that pci_address_to_pio() will do too.
|
|
+ */
|
|
+int of_pci_range_to_resource(struct of_pci_range *range,
|
|
+ struct device_node *np, struct resource *res)
|
|
+{
|
|
+ int err;
|
|
+ res->flags = range->flags;
|
|
+ res->parent = res->child = res->sibling = NULL;
|
|
+ res->name = np->full_name;
|
|
+
|
|
+ if (res->flags & IORESOURCE_IO) {
|
|
+ unsigned long port = -1;
|
|
+ err = pci_register_io_range(range->cpu_addr, range->size);
|
|
+ if (err)
|
|
+ goto invalid_range;
|
|
+ port = pci_address_to_pio(range->cpu_addr);
|
|
+ if (port == (unsigned long)-1) {
|
|
+ err = -EINVAL;
|
|
+ goto invalid_range;
|
|
+ }
|
|
+ res->start = port;
|
|
+ } else {
|
|
+ res->start = range->cpu_addr;
|
|
+ }
|
|
+ res->end = res->start + range->size - 1;
|
|
+ return 0;
|
|
+
|
|
+invalid_range:
|
|
+ res->start = (resource_size_t)OF_BAD_ADDR;
|
|
+ res->end = (resource_size_t)OF_BAD_ADDR;
|
|
+ return err;
|
|
+}
|
|
+
|
|
diff --git a/drivers/of/of_pci.c b/drivers/of/of_pci.c
|
|
index 8481996..e81402a 100644
|
|
--- a/drivers/of/of_pci.c
|
|
+++ b/drivers/of/of_pci.c
|
|
@@ -1,6 +1,7 @@
|
|
#include <linux/kernel.h>
|
|
#include <linux/export.h>
|
|
#include <linux/of.h>
|
|
+#include <linux/of_address.h>
|
|
#include <linux/of_pci.h>
|
|
|
|
static inline int __of_pci_pci_compare(struct device_node *node,
|
|
@@ -89,6 +90,141 @@ int of_pci_parse_bus_range(struct device_node *node, struct resource *res)
|
|
}
|
|
EXPORT_SYMBOL_GPL(of_pci_parse_bus_range);
|
|
|
|
+/**
|
|
+ * pci_host_bridge_of_get_ranges - Parse PCI host bridge resources from DT
|
|
+ * @dev: device node of the host bridge having the range property
|
|
+ * @resources: list where the range of resources will be added after DT parsing
|
|
+ * @io_base: pointer to a variable that will contain the physical address for
|
|
+ * the start of the I/O range.
|
|
+ *
|
|
+ * It is the callers job to free the @resources list if an error is returned.
|
|
+ *
|
|
+ * This function will parse the "ranges" property of a PCI host bridge device
|
|
+ * node and setup the resource mapping based on its content. It is expected
|
|
+ * that the property conforms with the Power ePAPR document.
|
|
+ *
|
|
+ * Each architecture is then offered the chance of applying their own
|
|
+ * filtering of pci_host_bridge_windows based on their own restrictions by
|
|
+ * calling pcibios_fixup_bridge_ranges(). The filtered list of windows
|
|
+ * can then be used when creating a pci_host_bridge structure.
|
|
+ */
|
|
+static int pci_host_bridge_of_get_ranges(struct device_node *dev,
|
|
+ struct list_head *resources, resource_size_t *io_base)
|
|
+{
|
|
+ struct resource *res;
|
|
+ struct of_pci_range range;
|
|
+ struct of_pci_range_parser parser;
|
|
+ int err;
|
|
+
|
|
+ pr_info("PCI host bridge %s ranges:\n", dev->full_name);
|
|
+
|
|
+ /* Check for ranges property */
|
|
+ err = of_pci_range_parser_init(&parser, dev);
|
|
+ if (err)
|
|
+ return err;
|
|
+
|
|
+ pr_debug("Parsing ranges property...\n");
|
|
+ for_each_of_pci_range(&parser, &range) {
|
|
+ /* Read next ranges element */
|
|
+ pr_debug("pci_space: 0x%08x pci_addr:0x%016llx cpu_addr:0x%016llx size:0x%016llx\n",
|
|
+ range.pci_space, range.pci_addr, range.cpu_addr, range.size);
|
|
+
|
|
+ /*
|
|
+ * If we failed translation or got a zero-sized region
|
|
+ * then skip this range
|
|
+ */
|
|
+ if (range.cpu_addr == OF_BAD_ADDR || range.size == 0)
|
|
+ continue;
|
|
+
|
|
+ res = kzalloc(sizeof(struct resource), GFP_KERNEL);
|
|
+ if (!res)
|
|
+ return -ENOMEM;
|
|
+
|
|
+ err = of_pci_range_to_resource(&range, dev, res);
|
|
+ if (err)
|
|
+ return err;
|
|
+
|
|
+ if (resource_type(res) == IORESOURCE_IO)
|
|
+ *io_base = range.cpu_addr;
|
|
+
|
|
+ pci_add_resource_offset(resources, res,
|
|
+ res->start - range.pci_addr);
|
|
+ }
|
|
+
|
|
+ /* Apply architecture specific fixups for the ranges */
|
|
+ return pcibios_fixup_bridge_ranges(resources);
|
|
+}
|
|
+
|
|
+static atomic_t domain_nr = ATOMIC_INIT(-1);
|
|
+
|
|
+/**
|
|
+ * of_create_pci_host_bridge - Create a PCI host bridge structure using
|
|
+ * information passed in the DT.
|
|
+ * @parent: device owning this host bridge
|
|
+ * @ops: pci_ops associated with the host controller
|
|
+ * @host_data: opaque data structure used by the host controller.
|
|
+ *
|
|
+ * returns a pointer to the newly created pci_host_bridge structure, or
|
|
+ * NULL if the call failed.
|
|
+ *
|
|
+ * This function will try to obtain the host bridge domain number by
|
|
+ * using of_alias_get_id() call with "pci-domain" as a stem. If that
|
|
+ * fails, a local allocator will be used that will put each host bridge
|
|
+ * in a new domain.
|
|
+ */
|
|
+struct pci_host_bridge *
|
|
+of_create_pci_host_bridge(struct device *parent, struct pci_ops *ops, void *host_data)
|
|
+{
|
|
+ int err, domain, busno;
|
|
+ struct resource *bus_range;
|
|
+ struct pci_bus *root_bus;
|
|
+ struct pci_host_bridge *bridge;
|
|
+ resource_size_t io_base = 0;
|
|
+ LIST_HEAD(res);
|
|
+
|
|
+ bus_range = kzalloc(sizeof(*bus_range), GFP_KERNEL);
|
|
+ if (!bus_range)
|
|
+ return ERR_PTR(-ENOMEM);
|
|
+
|
|
+ domain = of_alias_get_id(parent->of_node, "pci-domain");
|
|
+ if (domain == -ENODEV)
|
|
+ domain = atomic_inc_return(&domain_nr);
|
|
+
|
|
+ err = of_pci_parse_bus_range(parent->of_node, bus_range);
|
|
+ if (err) {
|
|
+ dev_info(parent, "No bus range for %s, using default [0-255]\n",
|
|
+ parent->of_node->full_name);
|
|
+ bus_range->start = 0;
|
|
+ bus_range->end = 255;
|
|
+ bus_range->flags = IORESOURCE_BUS;
|
|
+ }
|
|
+ busno = bus_range->start;
|
|
+ pci_add_resource(&res, bus_range);
|
|
+
|
|
+ /* now parse the rest of host bridge bus ranges */
|
|
+ err = pci_host_bridge_of_get_ranges(parent->of_node, &res, &io_base);
|
|
+ if (err)
|
|
+ goto err_create;
|
|
+
|
|
+ /* then create the root bus */
|
|
+ root_bus = pci_create_root_bus_in_domain(parent, domain, busno,
|
|
+ ops, host_data, &res);
|
|
+ if (IS_ERR(root_bus)) {
|
|
+ err = PTR_ERR(root_bus);
|
|
+ goto err_create;
|
|
+ }
|
|
+
|
|
+ bridge = to_pci_host_bridge(root_bus->bridge);
|
|
+ bridge->io_base = io_base;
|
|
+
|
|
+ return bridge;
|
|
+
|
|
+err_create:
|
|
+ pci_free_resource_list(&res);
|
|
+ return ERR_PTR(err);
|
|
+}
|
|
+EXPORT_SYMBOL_GPL(of_create_pci_host_bridge);
|
|
+
|
|
#ifdef CONFIG_PCI_MSI
|
|
|
|
static LIST_HEAD(of_pci_msi_chip_list);
|
|
diff --git a/drivers/pci/host-bridge.c b/drivers/pci/host-bridge.c
|
|
index 0e5f3c9..54ceafd 100644
|
|
--- a/drivers/pci/host-bridge.c
|
|
+++ b/drivers/pci/host-bridge.c
|
|
@@ -16,12 +16,13 @@ static struct pci_bus *find_pci_root_bus(struct pci_bus *bus)
|
|
return bus;
|
|
}
|
|
|
|
-static struct pci_host_bridge *find_pci_host_bridge(struct pci_bus *bus)
|
|
+struct pci_host_bridge *find_pci_host_bridge(struct pci_bus *bus)
|
|
{
|
|
struct pci_bus *root_bus = find_pci_root_bus(bus);
|
|
|
|
return to_pci_host_bridge(root_bus->bridge);
|
|
}
|
|
+EXPORT_SYMBOL_GPL(find_pci_host_bridge);
|
|
|
|
void pci_set_host_bridge_release(struct pci_host_bridge *bridge,
|
|
void (*release_fn)(struct pci_host_bridge *),
|
|
@@ -82,3 +83,18 @@ void pcibios_bus_to_resource(struct pci_bus *bus, struct resource *res,
|
|
res->end = region->end + offset;
|
|
}
|
|
EXPORT_SYMBOL(pcibios_bus_to_resource);
|
|
+
|
|
+/**
|
|
+ * Simple version of the platform specific code for filtering the list
|
|
+ * of resources obtained from the ranges declaration in DT.
|
|
+ *
|
|
+ * Platforms can override this function in order to impose stronger
|
|
+ * constraints onto the list of resources that a host bridge can use.
|
|
+ * The filtered list will then be used to create a root bus and associate
|
|
+ * it with the host bridge.
|
|
+ *
|
|
+ */
|
|
+int __weak pcibios_fixup_bridge_ranges(struct list_head *resources)
|
|
+{
|
|
+ return 0;
|
|
+}
|
|
diff --git a/drivers/pci/host/Kconfig b/drivers/pci/host/Kconfig
|
|
index 21df477..3b988a2 100644
|
|
--- a/drivers/pci/host/Kconfig
|
|
+++ b/drivers/pci/host/Kconfig
|
|
@@ -46,4 +46,14 @@ config PCI_HOST_GENERIC
|
|
Say Y here if you want to support a simple generic PCI host
|
|
controller, such as the one emulated by kvmtool.
|
|
|
|
+config PCI_XGENE
|
|
+ bool "X-Gene PCIe controller"
|
|
+ depends on ARCH_XGENE
|
|
+ depends on OF
|
|
+ select PCIEPORTBUS
|
|
+ help
|
|
+ Say Y here if you want internal PCI support on APM X-Gene SoC.
|
|
+ There are 5 internal PCIe ports available. Each port is GEN3 capable
|
|
+ and have varied lanes from x1 to x8.
|
|
+
|
|
endmenu
|
|
diff --git a/drivers/pci/host/Makefile b/drivers/pci/host/Makefile
|
|
index 611ba4b..0801606 100644
|
|
--- a/drivers/pci/host/Makefile
|
|
+++ b/drivers/pci/host/Makefile
|
|
@@ -6,3 +6,4 @@ obj-$(CONFIG_PCI_TEGRA) += pci-tegra.o
|
|
obj-$(CONFIG_PCI_RCAR_GEN2) += pci-rcar-gen2.o
|
|
obj-$(CONFIG_PCI_RCAR_GEN2_PCIE) += pcie-rcar.o
|
|
obj-$(CONFIG_PCI_HOST_GENERIC) += pci-host-generic.o
|
|
+obj-$(CONFIG_PCI_XGENE) += pci-xgene.o
|
|
diff --git a/drivers/pci/host/pci-xgene.c b/drivers/pci/host/pci-xgene.c
|
|
new file mode 100644
|
|
index 0000000..7bf4ac7
|
|
--- /dev/null
|
|
+++ b/drivers/pci/host/pci-xgene.c
|
|
@@ -0,0 +1,725 @@
|
|
+/**
|
|
+ * APM X-Gene PCIe Driver
|
|
+ *
|
|
+ * Copyright (c) 2013 Applied Micro Circuits Corporation.
|
|
+ *
|
|
+ * Author: Tanmay Inamdar <tinamdar@apm.com>.
|
|
+ *
|
|
+ * This program is free software; you can redistribute it and/or modify it
|
|
+ * under the terms of the GNU General Public License as published by the
|
|
+ * Free Software Foundation; either version 2 of the License, or (at your
|
|
+ * option) any later version.
|
|
+ *
|
|
+ * This program is distributed in the hope that it will be useful,
|
|
+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
+ * GNU General Public License for more details.
|
|
+ *
|
|
+ */
|
|
+#include <linux/clk-private.h>
|
|
+#include <linux/delay.h>
|
|
+#include <linux/io.h>
|
|
+#include <linux/jiffies.h>
|
|
+#include <linux/memblock.h>
|
|
+#include <linux/module.h>
|
|
+#include <linux/of.h>
|
|
+#include <linux/of_address.h>
|
|
+#include <linux/of_irq.h>
|
|
+#include <linux/of_pci.h>
|
|
+#include <linux/pci.h>
|
|
+#include <linux/platform_device.h>
|
|
+#include <linux/slab.h>
|
|
+
|
|
+#define PCIECORE_LTSSM 0x4c
|
|
+#define PCIECORE_CTLANDSTATUS 0x50
|
|
+#define INTXSTATUSMASK 0x6c
|
|
+#define PIM1_1L 0x80
|
|
+#define IBAR2 0x98
|
|
+#define IR2MSK 0x9c
|
|
+#define PIM2_1L 0xa0
|
|
+#define IBAR3L 0xb4
|
|
+#define IR3MSKL 0xbc
|
|
+#define PIM3_1L 0xc4
|
|
+#define OMR1BARL 0x100
|
|
+#define OMR2BARL 0x118
|
|
+#define OMR3BARL 0x130
|
|
+#define CFGBARL 0x154
|
|
+#define CFGBARH 0x158
|
|
+#define CFGCTL 0x15c
|
|
+#define RTDID 0x160
|
|
+#define BRIDGE_CFG_0 0x2000
|
|
+#define BRIDGE_CFG_1 0x2004
|
|
+#define BRIDGE_CFG_4 0x2010
|
|
+#define BRIDGE_CFG_32 0x2030
|
|
+#define BRIDGE_CFG_14 0x2038
|
|
+#define BRIDGE_CTRL_1 0x2204
|
|
+#define BRIDGE_CTRL_2 0x2208
|
|
+#define BRIDGE_CTRL_5 0x2214
|
|
+#define BRIDGE_STATUS_0 0x2600
|
|
+#define MEM_RAM_SHUTDOWN 0xd070
|
|
+#define BLOCK_MEM_RDY 0xd074
|
|
+
|
|
+#define DEVICE_PORT_TYPE_MASK 0x03c00000
|
|
+#define PM_FORCE_RP_MODE_MASK 0x00000400
|
|
+#define SWITCH_PORT_MODE_MASK 0x00000800
|
|
+#define CLASS_CODE_MASK 0xffffff00
|
|
+#define LINK_UP_MASK 0x00000100
|
|
+#define AER_OPTIONAL_ERROR_EN 0xffc00000
|
|
+#define XGENE_PCIE_DEV_CTRL 0x2f0f
|
|
+#define AXI_EP_CFG_ACCESS 0x10000
|
|
+#define ENABLE_ASPM 0x08000000
|
|
+#define XGENE_PORT_TYPE_RC 0x05000000
|
|
+#define BLOCK_MEM_RDY_VAL 0xFFFFFFFF
|
|
+#define EN_COHERENCY 0xF0000000
|
|
+#define EN_REG 0x00000001
|
|
+#define OB_LO_IO 0x00000002
|
|
+#define XGENE_PCIE_VENDORID 0xE008
|
|
+#define XGENE_PCIE_DEVICEID 0xE004
|
|
+#define XGENE_PCIE_ECC_TIMEOUT 10 /* ms */
|
|
+#define XGENE_LTSSM_DETECT_WAIT 20 /* ms */
|
|
+#define XGENE_LTSSM_L0_WAIT 4 /* ms */
|
|
+#define SZ_1T (SZ_1G*1024ULL)
|
|
+#define PIPE_PHY_RATE_RD(src) ((0xc000 & (u32)(src)) >> 0xe)
|
|
+
|
|
+struct xgene_pcie_port {
|
|
+ struct device_node *node;
|
|
+ struct device *dev;
|
|
+ struct clk *clk;
|
|
+ void __iomem *csr_base;
|
|
+ void __iomem *cfg_base;
|
|
+ u8 link_up;
|
|
+};
|
|
+
|
|
+static inline u32 pcie_bar_low_val(u32 addr, u32 flags)
|
|
+{
|
|
+ return (addr & PCI_BASE_ADDRESS_MEM_MASK) | flags;
|
|
+}
|
|
+
|
|
+/* PCIE Configuration Out/In */
|
|
+static inline void xgene_pcie_cfg_out32(void __iomem *addr, int offset, u32 val)
|
|
+{
|
|
+ writel(val, addr + offset);
|
|
+}
|
|
+
|
|
+static inline void xgene_pcie_cfg_out16(void __iomem *addr, int offset, u16 val)
|
|
+{
|
|
+ u32 val32 = readl(addr + (offset & ~0x3));
|
|
+
|
|
+ switch (offset & 0x3) {
|
|
+ case 2:
|
|
+ val32 &= ~0xFFFF0000;
|
|
+ val32 |= (u32)val << 16;
|
|
+ break;
|
|
+ case 0:
|
|
+ default:
|
|
+ val32 &= ~0xFFFF;
|
|
+ val32 |= val;
|
|
+ break;
|
|
+ }
|
|
+ writel(val32, addr + (offset & ~0x3));
|
|
+}
|
|
+
|
|
+static inline void xgene_pcie_cfg_out8(void __iomem *addr, int offset, u8 val)
|
|
+{
|
|
+ u32 val32 = readl(addr + (offset & ~0x3));
|
|
+
|
|
+ switch (offset & 0x3) {
|
|
+ case 0:
|
|
+ val32 &= ~0xFF;
|
|
+ val32 |= val;
|
|
+ break;
|
|
+ case 1:
|
|
+ val32 &= ~0xFF00;
|
|
+ val32 |= (u32)val << 8;
|
|
+ break;
|
|
+ case 2:
|
|
+ val32 &= ~0xFF0000;
|
|
+ val32 |= (u32)val << 16;
|
|
+ break;
|
|
+ case 3:
|
|
+ default:
|
|
+ val32 &= ~0xFF000000;
|
|
+ val32 |= (u32)val << 24;
|
|
+ break;
|
|
+ }
|
|
+ writel(val32, addr + (offset & ~0x3));
|
|
+}
|
|
+
|
|
+static inline void xgene_pcie_cfg_in32(void __iomem *addr, int offset, u32 *val)
|
|
+{
|
|
+ *val = readl(addr + offset);
|
|
+}
|
|
+
|
|
+static inline void
|
|
+xgene_pcie_cfg_in16(void __iomem *addr, int offset, u32 *val)
|
|
+{
|
|
+ *val = readl(addr + (offset & ~0x3));
|
|
+
|
|
+ switch (offset & 0x3) {
|
|
+ case 2:
|
|
+ *val >>= 16;
|
|
+ break;
|
|
+ }
|
|
+
|
|
+ *val &= 0xFFFF;
|
|
+}
|
|
+
|
|
+static inline void
|
|
+xgene_pcie_cfg_in8(void __iomem *addr, int offset, u32 *val)
|
|
+{
|
|
+ *val = readl(addr + (offset & ~0x3));
|
|
+
|
|
+ switch (offset & 0x3) {
|
|
+ case 3:
|
|
+ *val = *val >> 24;
|
|
+ break;
|
|
+ case 2:
|
|
+ *val = *val >> 16;
|
|
+ break;
|
|
+ case 1:
|
|
+ *val = *val >> 8;
|
|
+ break;
|
|
+ }
|
|
+ *val &= 0xFF;
|
|
+}
|
|
+
|
|
+/* When the address bit [17:16] is 2'b01, the Configuration access will be
|
|
+ * treated as Type 1 and it will be forwarded to external PCIe device.
|
|
+ */
|
|
+static void __iomem *xgene_pcie_get_cfg_base(struct pci_bus *bus)
|
|
+{
|
|
+ struct xgene_pcie_port *port = bus->sysdata;
|
|
+
|
|
+ if (bus->number >= (bus->primary + 1))
|
|
+ return port->cfg_base + AXI_EP_CFG_ACCESS;
|
|
+
|
|
+ return port->cfg_base;
|
|
+}
|
|
+
|
|
+/* For Configuration request, RTDID register is used as Bus Number,
|
|
+ * Device Number and Function number of the header fields.
|
|
+ */
|
|
+static void xgene_pcie_set_rtdid_reg(struct pci_bus *bus, uint devfn)
|
|
+{
|
|
+ struct xgene_pcie_port *port = bus->sysdata;
|
|
+ unsigned int b, d, f;
|
|
+ u32 rtdid_val = 0;
|
|
+
|
|
+ b = bus->number;
|
|
+ d = PCI_SLOT(devfn);
|
|
+ f = PCI_FUNC(devfn);
|
|
+
|
|
+ if (!pci_is_root_bus(bus))
|
|
+ rtdid_val = (b << 8) | (d << 3) | f;
|
|
+
|
|
+ writel(rtdid_val, port->csr_base + RTDID);
|
|
+ /* read the register back to ensure flush */
|
|
+ readl(port->csr_base + RTDID);
|
|
+}
|
|
+
|
|
+static int xgene_pcie_read_config(struct pci_bus *bus, unsigned int devfn,
|
|
+ int offset, int len, u32 *val)
|
|
+{
|
|
+ struct xgene_pcie_port *port = bus->sysdata;
|
|
+ void __iomem *addr;
|
|
+
|
|
+ if ((pci_is_root_bus(bus) && devfn != 0) || !port->link_up)
|
|
+ return PCIBIOS_DEVICE_NOT_FOUND;
|
|
+
|
|
+ xgene_pcie_set_rtdid_reg(bus, devfn);
|
|
+ addr = xgene_pcie_get_cfg_base(bus);
|
|
+ switch (len) {
|
|
+ case 1:
|
|
+ xgene_pcie_cfg_in8(addr, offset, val);
|
|
+ break;
|
|
+ case 2:
|
|
+ xgene_pcie_cfg_in16(addr, offset, val);
|
|
+ break;
|
|
+ default:
|
|
+ xgene_pcie_cfg_in32(addr, offset, val);
|
|
+ break;
|
|
+ }
|
|
+ return PCIBIOS_SUCCESSFUL;
|
|
+}
|
|
+
|
|
+static int xgene_pcie_write_config(struct pci_bus *bus, unsigned int devfn,
|
|
+ int offset, int len, u32 val)
|
|
+{
|
|
+ struct xgene_pcie_port *port = bus->sysdata;
|
|
+ void __iomem *addr;
|
|
+
|
|
+ if ((pci_is_root_bus(bus) && devfn != 0) || !port->link_up)
|
|
+ return PCIBIOS_DEVICE_NOT_FOUND;
|
|
+
|
|
+ xgene_pcie_set_rtdid_reg(bus, devfn);
|
|
+ addr = xgene_pcie_get_cfg_base(bus);
|
|
+ switch (len) {
|
|
+ case 1:
|
|
+ xgene_pcie_cfg_out8(addr, offset, (u8)val);
|
|
+ break;
|
|
+ case 2:
|
|
+ xgene_pcie_cfg_out16(addr, offset, (u16)val);
|
|
+ break;
|
|
+ default:
|
|
+ xgene_pcie_cfg_out32(addr, offset, val);
|
|
+ break;
|
|
+ }
|
|
+ return PCIBIOS_SUCCESSFUL;
|
|
+}
|
|
+
|
|
+static struct pci_ops xgene_pcie_ops = {
|
|
+ .read = xgene_pcie_read_config,
|
|
+ .write = xgene_pcie_write_config
|
|
+};
|
|
+
|
|
+static void xgene_pcie_program_core(void __iomem *csr_base)
|
|
+{
|
|
+ u32 val;
|
|
+
|
|
+ val = readl(csr_base + BRIDGE_CFG_0);
|
|
+ val |= AER_OPTIONAL_ERROR_EN;
|
|
+ writel(val, csr_base + BRIDGE_CFG_0);
|
|
+ writel(0x0, csr_base + INTXSTATUSMASK);
|
|
+ val = readl(csr_base + BRIDGE_CTRL_1);
|
|
+ val = (val & ~0xffff) | XGENE_PCIE_DEV_CTRL;
|
|
+ writel(val, csr_base + BRIDGE_CTRL_1);
|
|
+}
|
|
+
|
|
+static u64 xgene_pcie_set_ib_mask(void __iomem *csr_base, u32 addr,
|
|
+ u32 flags, u64 size)
|
|
+{
|
|
+ u64 mask = (~(size - 1) & PCI_BASE_ADDRESS_MEM_MASK) | flags;
|
|
+ u32 val32 = 0;
|
|
+ u32 val;
|
|
+
|
|
+ val32 = readl(csr_base + addr);
|
|
+ val = (val32 & 0x0000ffff) | (lower_32_bits(mask) << 16);
|
|
+ writel(val, csr_base + addr);
|
|
+
|
|
+ val32 = readl(csr_base + addr + 0x04);
|
|
+ val = (val32 & 0xffff0000) | (lower_32_bits(mask) >> 16);
|
|
+ writel(val, csr_base + addr + 0x04);
|
|
+
|
|
+ val32 = readl(csr_base + addr + 0x04);
|
|
+ val = (val32 & 0x0000ffff) | (upper_32_bits(mask) << 16);
|
|
+ writel(val, csr_base + addr + 0x04);
|
|
+
|
|
+ val32 = readl(csr_base + addr + 0x08);
|
|
+ val = (val32 & 0xffff0000) | (upper_32_bits(mask) >> 16);
|
|
+ writel(val, csr_base + addr + 0x08);
|
|
+
|
|
+ return mask;
|
|
+}
|
|
+
|
|
+static void xgene_pcie_poll_linkup(struct xgene_pcie_port *port,
|
|
+ u32 *lanes, u32 *speed)
|
|
+{
|
|
+ void __iomem *csr_base = port->csr_base;
|
|
+ ulong timeout;
|
|
+ u32 val32;
|
|
+
|
|
+ /*
|
|
+ * A component enters the LTSSM Detect state within
|
|
+ * 20ms of the end of fundamental core reset.
|
|
+ */
|
|
+ msleep(XGENE_LTSSM_DETECT_WAIT);
|
|
+ port->link_up = 0;
|
|
+ timeout = jiffies + msecs_to_jiffies(XGENE_LTSSM_L0_WAIT);
|
|
+ while (time_before(jiffies, timeout)) {
|
|
+ val32 = readl(csr_base + PCIECORE_CTLANDSTATUS);
|
|
+ if (val32 & LINK_UP_MASK) {
|
|
+ port->link_up = 1;
|
|
+ *speed = PIPE_PHY_RATE_RD(val32);
|
|
+ val32 = readl(csr_base + BRIDGE_STATUS_0);
|
|
+ *lanes = val32 >> 26;
|
|
+ break;
|
|
+ }
|
|
+ msleep(1);
|
|
+ }
|
|
+}
|
|
+
|
|
+static void xgene_pcie_setup_root_complex(struct xgene_pcie_port *port)
|
|
+{
|
|
+ void __iomem *csr_base = port->csr_base;
|
|
+ u32 val;
|
|
+
|
|
+ val = (XGENE_PCIE_DEVICEID << 16) | XGENE_PCIE_VENDORID;
|
|
+ writel(val, csr_base + BRIDGE_CFG_0);
|
|
+
|
|
+ val = readl(csr_base + BRIDGE_CFG_1);
|
|
+ val &= ~CLASS_CODE_MASK;
|
|
+ val |= PCI_CLASS_BRIDGE_PCI << 16;
|
|
+ writel(val, csr_base + BRIDGE_CFG_1);
|
|
+
|
|
+ val = readl(csr_base + BRIDGE_CFG_14);
|
|
+ val |= SWITCH_PORT_MODE_MASK;
|
|
+ val &= ~PM_FORCE_RP_MODE_MASK;
|
|
+ writel(val, csr_base + BRIDGE_CFG_14);
|
|
+
|
|
+ val = readl(csr_base + BRIDGE_CTRL_5);
|
|
+ val &= ~DEVICE_PORT_TYPE_MASK;
|
|
+ val |= XGENE_PORT_TYPE_RC;
|
|
+ writel(val, csr_base + BRIDGE_CTRL_5);
|
|
+
|
|
+ val = readl(csr_base + BRIDGE_CTRL_2);
|
|
+ val |= ENABLE_ASPM;
|
|
+ writel(val, csr_base + BRIDGE_CTRL_2);
|
|
+
|
|
+ val = readl(csr_base + BRIDGE_CFG_32);
|
|
+ writel(val | (1 << 19), csr_base + BRIDGE_CFG_32);
|
|
+}
|
|
+
|
|
+/* Return 0 on success */
|
|
+static int xgene_pcie_init_ecc(struct xgene_pcie_port *port)
|
|
+{
|
|
+ void __iomem *csr_base = port->csr_base;
|
|
+ ulong timeout;
|
|
+ u32 val;
|
|
+
|
|
+ val = readl(csr_base + MEM_RAM_SHUTDOWN);
|
|
+ if (!val)
|
|
+ return 0;
|
|
+ writel(0x0, csr_base + MEM_RAM_SHUTDOWN);
|
|
+ timeout = jiffies + msecs_to_jiffies(XGENE_PCIE_ECC_TIMEOUT);
|
|
+ while (time_before(jiffies, timeout)) {
|
|
+ val = readl(csr_base + BLOCK_MEM_RDY);
|
|
+ if (val == BLOCK_MEM_RDY_VAL)
|
|
+ return 0;
|
|
+ msleep(1);
|
|
+ }
|
|
+
|
|
+ return 1;
|
|
+}
|
|
+
|
|
+static int xgene_pcie_init_port(struct xgene_pcie_port *port)
|
|
+{
|
|
+ int rc;
|
|
+
|
|
+ port->clk = clk_get(port->dev, NULL);
|
|
+ if (IS_ERR(port->clk)) {
|
|
+ dev_err(port->dev, "clock not available\n");
|
|
+ return -ENODEV;
|
|
+ }
|
|
+
|
|
+ rc = clk_prepare_enable(port->clk);
|
|
+ if (rc) {
|
|
+ dev_err(port->dev, "clock enable failed\n");
|
|
+ return rc;
|
|
+ }
|
|
+
|
|
+ rc = xgene_pcie_init_ecc(port);
|
|
+ if (rc) {
|
|
+ dev_err(port->dev, "memory init failed\n");
|
|
+ return rc;
|
|
+ }
|
|
+
|
|
+ return 0;
|
|
+}
|
|
+
|
|
+static void xgene_pcie_fixup_bridge(struct pci_dev *dev)
|
|
+{
|
|
+ int i;
|
|
+
|
|
+ /* Hide the PCI host BARs from the kernel as their content doesn't
|
|
+ * fit well in the resource management
|
|
+ */
|
|
+ for (i = 0; i < DEVICE_COUNT_RESOURCE; i++) {
|
|
+ dev->resource[i].start = dev->resource[i].end = 0;
|
|
+ dev->resource[i].flags = 0;
|
|
+ }
|
|
+ dev_info(&dev->dev, "Hiding X-Gene pci host bridge resources %s\n",
|
|
+ pci_name(dev));
|
|
+}
|
|
+DECLARE_PCI_FIXUP_HEADER(XGENE_PCIE_VENDORID, XGENE_PCIE_DEVICEID,
|
|
+ xgene_pcie_fixup_bridge);
|
|
+
|
|
+static int xgene_pcie_map_reg(struct xgene_pcie_port *port,
|
|
+ struct platform_device *pdev, u64 *cfg_addr)
|
|
+{
|
|
+ struct resource *res;
|
|
+
|
|
+ res = platform_get_resource_byname(pdev, IORESOURCE_MEM, "csr");
|
|
+ port->csr_base = devm_ioremap_resource(port->dev, res);
|
|
+ if (IS_ERR(port->csr_base))
|
|
+ return PTR_ERR(port->csr_base);
|
|
+
|
|
+ res = platform_get_resource_byname(pdev, IORESOURCE_MEM, "cfg");
|
|
+ port->cfg_base = devm_ioremap_resource(port->dev, res);
|
|
+ if (IS_ERR(port->cfg_base))
|
|
+ return PTR_ERR(port->cfg_base);
|
|
+ *cfg_addr = res->start;
|
|
+
|
|
+ return 0;
|
|
+}
|
|
+
|
|
+static void xgene_pcie_setup_ob_reg(struct xgene_pcie_port *port,
|
|
+ struct resource *res, u32 offset, u64 addr)
|
|
+{
|
|
+ void __iomem *base = port->csr_base + offset;
|
|
+ resource_size_t size = resource_size(res);
|
|
+ u64 restype = resource_type(res);
|
|
+ u64 cpu_addr, pci_addr;
|
|
+ u64 mask = 0;
|
|
+ u32 min_size;
|
|
+ u32 flag = EN_REG;
|
|
+
|
|
+ if (restype == IORESOURCE_MEM) {
|
|
+ cpu_addr = res->start;
|
|
+ pci_addr = addr;
|
|
+ min_size = SZ_128M;
|
|
+ } else {
|
|
+ cpu_addr = addr;
|
|
+ pci_addr = res->start;
|
|
+ min_size = 128;
|
|
+ flag |= OB_LO_IO;
|
|
+ }
|
|
+ if (size >= min_size)
|
|
+ mask = ~(size - 1) | flag;
|
|
+ else
|
|
+ dev_warn(port->dev, "res size 0x%llx less than minimum 0x%x\n",
|
|
+ (u64)size, min_size);
|
|
+ writel(lower_32_bits(cpu_addr), base);
|
|
+ writel(upper_32_bits(cpu_addr), base + 0x04);
|
|
+ writel(lower_32_bits(mask), base + 0x08);
|
|
+ writel(upper_32_bits(mask), base + 0x0c);
|
|
+ writel(lower_32_bits(pci_addr), base + 0x10);
|
|
+ writel(upper_32_bits(pci_addr), base + 0x14);
|
|
+}
|
|
+
|
|
+static void xgene_pcie_setup_cfg_reg(void __iomem *csr_base, u64 addr)
|
|
+{
|
|
+ writel(lower_32_bits(addr), csr_base + CFGBARL);
|
|
+ writel(upper_32_bits(addr), csr_base + CFGBARH);
|
|
+ writel(EN_REG, csr_base + CFGCTL);
|
|
+}
|
|
+
|
|
+static int xgene_pcie_map_ranges(struct xgene_pcie_port *port,
|
|
+ struct pci_host_bridge *bridge,
|
|
+ u64 cfg_addr)
|
|
+{
|
|
+ struct device *dev = port->dev;
|
|
+ struct pci_host_bridge_window *window;
|
|
+ int ret;
|
|
+
|
|
+ list_for_each_entry(window, &bridge->windows, list) {
|
|
+ struct resource *res = window->res;
|
|
+ u64 restype = resource_type(res);
|
|
+ dev_dbg(port->dev, "0x%08lx 0x%016llx...0x%016llx\n",
|
|
+ res->flags, res->start, res->end);
|
|
+
|
|
+ switch (restype) {
|
|
+ case IORESOURCE_IO:
|
|
+ xgene_pcie_setup_ob_reg(port, res, OMR2BARL,
|
|
+ bridge->io_base);
|
|
+ ret = pci_remap_iospace(res, bridge->io_base);
|
|
+ if (ret < 0)
|
|
+ return ret;
|
|
+ break;
|
|
+ case IORESOURCE_MEM:
|
|
+ xgene_pcie_setup_ob_reg(port, res, OMR3BARL,
|
|
+ res->start - window->offset);
|
|
+ break;
|
|
+ case IORESOURCE_BUS:
|
|
+ break;
|
|
+ default:
|
|
+ dev_err(dev, "invalid io resource!");
|
|
+ return -EINVAL;
|
|
+ }
|
|
+ }
|
|
+ xgene_pcie_setup_cfg_reg(port->csr_base, cfg_addr);
|
|
+ return 0;
|
|
+}
|
|
+
|
|
+static void xgene_pcie_setup_pims(void *addr, u64 pim, u64 size)
|
|
+{
|
|
+ writel(lower_32_bits(pim), addr);
|
|
+ writel(upper_32_bits(pim) | EN_COHERENCY, addr + 0x04);
|
|
+ writel(lower_32_bits(size), addr + 0x10);
|
|
+ writel(upper_32_bits(size), addr + 0x14);
|
|
+}
|
|
+
|
|
+/*
|
|
+ * X-Gene PCIe support maximum 3 inbound memory regions
|
|
+ * This function helps to select a region based on size of region
|
|
+ */
|
|
+static int xgene_pcie_select_ib_reg(u8 *ib_reg_mask, u64 size)
|
|
+{
|
|
+ if ((size > 4) && (size < SZ_16M) && !(*ib_reg_mask & (1 << 1))) {
|
|
+ *ib_reg_mask |= (1 << 1);
|
|
+ return 1;
|
|
+ }
|
|
+
|
|
+ if ((size > SZ_1K) && (size < SZ_1T) && !(*ib_reg_mask & (1 << 0))) {
|
|
+ *ib_reg_mask |= (1 << 0);
|
|
+ return 0;
|
|
+ }
|
|
+
|
|
+ if ((size > SZ_1M) && (size < SZ_1T) && !(*ib_reg_mask & (1 << 2))) {
|
|
+ *ib_reg_mask |= (1 << 2);
|
|
+ return 2;
|
|
+ }
|
|
+ return -EINVAL;
|
|
+}
|
|
+
|
|
+static void xgene_pcie_setup_ib_reg(struct xgene_pcie_port *port,
|
|
+ struct of_pci_range *range, u8 *ib_reg_mask)
|
|
+{
|
|
+ void __iomem *csr_base = port->csr_base;
|
|
+ void __iomem *cfg_base = port->cfg_base;
|
|
+ void *bar_addr;
|
|
+ void *pim_addr;
|
|
+ u64 restype = range->flags & IORESOURCE_TYPE_BITS;
|
|
+ u64 cpu_addr = range->cpu_addr;
|
|
+ u64 pci_addr = range->pci_addr;
|
|
+ u64 size = range->size;
|
|
+ u64 mask = ~(size - 1) | EN_REG;
|
|
+ u32 flags = PCI_BASE_ADDRESS_MEM_TYPE_64;
|
|
+ u32 bar_low;
|
|
+ int region;
|
|
+
|
|
+ region = xgene_pcie_select_ib_reg(ib_reg_mask, range->size);
|
|
+ if (region < 0) {
|
|
+ dev_warn(port->dev, "invalid pcie dma-range config\n");
|
|
+ return;
|
|
+ }
|
|
+
|
|
+ if (restype == PCI_BASE_ADDRESS_MEM_PREFETCH)
|
|
+ flags |= PCI_BASE_ADDRESS_MEM_PREFETCH;
|
|
+
|
|
+ bar_low = pcie_bar_low_val((u32)cpu_addr, flags);
|
|
+ switch (region) {
|
|
+ case 0:
|
|
+ xgene_pcie_set_ib_mask(csr_base, BRIDGE_CFG_4, flags, size);
|
|
+ bar_addr = cfg_base + PCI_BASE_ADDRESS_0;
|
|
+ writel(bar_low, bar_addr);
|
|
+ writel(upper_32_bits(cpu_addr), bar_addr + 0x4);
|
|
+ pim_addr = csr_base + PIM1_1L;
|
|
+ break;
|
|
+ case 1:
|
|
+ bar_addr = csr_base + IBAR2;
|
|
+ writel(bar_low, bar_addr);
|
|
+ writel(lower_32_bits(mask), csr_base + IR2MSK);
|
|
+ pim_addr = csr_base + PIM2_1L;
|
|
+ break;
|
|
+ case 2:
|
|
+ bar_addr = csr_base + IBAR3L;
|
|
+ writel(bar_low, bar_addr);
|
|
+ writel(upper_32_bits(cpu_addr), bar_addr + 0x4);
|
|
+ writel(lower_32_bits(mask), csr_base + IR3MSKL);
|
|
+ writel(upper_32_bits(mask), csr_base + IR3MSKL + 0x4);
|
|
+ pim_addr = csr_base + PIM3_1L;
|
|
+ break;
|
|
+ }
|
|
+
|
|
+ xgene_pcie_setup_pims(pim_addr, pci_addr, size);
|
|
+}
|
|
+
|
|
+static int pci_dma_range_parser_init(struct of_pci_range_parser *parser,
|
|
+ struct device_node *node)
|
|
+{
|
|
+ const int na = 3, ns = 2;
|
|
+ int rlen;
|
|
+
|
|
+ parser->node = node;
|
|
+ parser->pna = of_n_addr_cells(node);
|
|
+ parser->np = parser->pna + na + ns;
|
|
+
|
|
+ parser->range = of_get_property(node, "dma-ranges", &rlen);
|
|
+ if (!parser->range)
|
|
+ return -ENOENT;
|
|
+
|
|
+ parser->end = parser->range + rlen / sizeof(__be32);
|
|
+ return 0;
|
|
+}
|
|
+
|
|
+static int xgene_pcie_parse_map_dma_ranges(struct xgene_pcie_port *port)
|
|
+{
|
|
+ struct device_node *np = port->node;
|
|
+ struct of_pci_range range;
|
|
+ struct of_pci_range_parser parser;
|
|
+ struct device *dev = port->dev;
|
|
+ u8 ib_reg_mask = 0;
|
|
+
|
|
+ if (pci_dma_range_parser_init(&parser, np)) {
|
|
+ dev_err(dev, "missing dma-ranges property\n");
|
|
+ return -EINVAL;
|
|
+ }
|
|
+
|
|
+ /* Get the dma-ranges from DT */
|
|
+ for_each_of_pci_range(&parser, &range) {
|
|
+ u64 end = range.cpu_addr + range.size - 1;
|
|
+ dev_dbg(port->dev, "0x%08x 0x%016llx..0x%016llx -> 0x%016llx\n",
|
|
+ range.flags, range.cpu_addr, end, range.pci_addr);
|
|
+ xgene_pcie_setup_ib_reg(port, &range, &ib_reg_mask);
|
|
+ }
|
|
+ return 0;
|
|
+}
|
|
+
|
|
+static int xgene_pcie_probe_bridge(struct platform_device *pdev)
|
|
+{
|
|
+ struct device_node *np = of_node_get(pdev->dev.of_node);
|
|
+ struct xgene_pcie_port *port;
|
|
+ struct pci_host_bridge *bridge;
|
|
+ resource_size_t lastbus;
|
|
+ u32 lanes = 0, speed = 0;
|
|
+ u64 cfg_addr = 0;
|
|
+ int ret;
|
|
+
|
|
+ port = devm_kzalloc(&pdev->dev, sizeof(*port), GFP_KERNEL);
|
|
+ if (!port)
|
|
+ return -ENOMEM;
|
|
+ port->node = np;
|
|
+ port->dev = &pdev->dev;
|
|
+
|
|
+ ret = xgene_pcie_map_reg(port, pdev, &cfg_addr);
|
|
+ if (ret)
|
|
+ return ret;
|
|
+
|
|
+ ret = xgene_pcie_init_port(port);
|
|
+ if (ret)
|
|
+ return ret;
|
|
+ xgene_pcie_program_core(port->csr_base);
|
|
+ xgene_pcie_setup_root_complex(port);
|
|
+
|
|
+ bridge = of_create_pci_host_bridge(&pdev->dev, &xgene_pcie_ops, port);
|
|
+ if (IS_ERR_OR_NULL(bridge))
|
|
+ return PTR_ERR(bridge);
|
|
+
|
|
+ ret = xgene_pcie_map_ranges(port, bridge, cfg_addr);
|
|
+ if (ret)
|
|
+ return ret;
|
|
+
|
|
+ ret = xgene_pcie_parse_map_dma_ranges(port);
|
|
+ if (ret)
|
|
+ return ret;
|
|
+
|
|
+ xgene_pcie_poll_linkup(port, &lanes, &speed);
|
|
+ if (!port->link_up)
|
|
+ dev_info(port->dev, "(rc) link down\n");
|
|
+ else
|
|
+ dev_info(port->dev, "(rc) x%d gen-%d link up\n",
|
|
+ lanes, speed + 1);
|
|
+ platform_set_drvdata(pdev, port);
|
|
+ lastbus = pci_rescan_bus(bridge->bus);
|
|
+ pci_bus_update_busn_res_end(bridge->bus, lastbus);
|
|
+ return 0;
|
|
+}
|
|
+
|
|
+static const struct of_device_id xgene_pcie_match_table[] = {
|
|
+ {.compatible = "apm,xgene-pcie",},
|
|
+ {},
|
|
+};
|
|
+
|
|
+static struct platform_driver xgene_pcie_driver = {
|
|
+ .driver = {
|
|
+ .name = "xgene-pcie",
|
|
+ .owner = THIS_MODULE,
|
|
+ .of_match_table = of_match_ptr(xgene_pcie_match_table),
|
|
+ },
|
|
+ .probe = xgene_pcie_probe_bridge,
|
|
+};
|
|
+module_platform_driver(xgene_pcie_driver);
|
|
+
|
|
+MODULE_AUTHOR("Tanmay Inamdar <tinamdar@apm.com>");
|
|
+MODULE_DESCRIPTION("APM X-Gene PCIe driver");
|
|
+MODULE_LICENSE("GPL v2");
|
|
diff --git a/drivers/pci/pci.c b/drivers/pci/pci.c
|
|
index 2c9ac70..7bae0f9 100644
|
|
--- a/drivers/pci/pci.c
|
|
+++ b/drivers/pci/pci.c
|
|
@@ -17,6 +17,7 @@
|
|
#include <linux/spinlock.h>
|
|
#include <linux/string.h>
|
|
#include <linux/log2.h>
|
|
+#include <linux/of_pci.h>
|
|
#include <linux/pci-aspm.h>
|
|
#include <linux/pm_wakeup.h>
|
|
#include <linux/interrupt.h>
|
|
@@ -1453,6 +1454,9 @@ EXPORT_SYMBOL(pcim_pin_device);
|
|
*/
|
|
int __weak pcibios_add_device(struct pci_dev *dev)
|
|
{
|
|
+#ifdef CONFIG_OF
|
|
+ dev->irq = of_irq_parse_and_map_pci(dev, 0, 0);
|
|
+#endif
|
|
return 0;
|
|
}
|
|
|
|
@@ -2704,6 +2708,39 @@ int pci_request_regions_exclusive(struct pci_dev *pdev, const char *res_name)
|
|
}
|
|
EXPORT_SYMBOL(pci_request_regions_exclusive);
|
|
|
|
+/**
|
|
+ * pci_remap_iospace - Remap the memory mapped I/O space
|
|
+ * @res: Resource describing the I/O space
|
|
+ * @phys_addr: physical address where the range will be mapped.
|
|
+ *
|
|
+ * Remap the memory mapped I/O space described by the @res
|
|
+ * into the CPU physical address space. Only architectures
|
|
+ * that have memory mapped IO defined (and hence PCI_IOBASE)
|
|
+ * should call this function.
|
|
+ */
|
|
+int __weak pci_remap_iospace(const struct resource *res, phys_addr_t phys_addr)
|
|
+{
|
|
+ int err = -ENODEV;
|
|
+
|
|
+#ifdef PCI_IOBASE
|
|
+ if (!(res->flags & IORESOURCE_IO))
|
|
+ return -EINVAL;
|
|
+
|
|
+ if (res->end > IO_SPACE_LIMIT)
|
|
+ return -EINVAL;
|
|
+
|
|
+ err = ioremap_page_range(res->start + (unsigned long)PCI_IOBASE,
|
|
+ res->end + 1 + (unsigned long)PCI_IOBASE,
|
|
+ phys_addr, __pgprot(PROT_DEVICE_nGnRE));
|
|
+#else
|
|
+ /* this architecture does not have memory mapped I/O space,
|
|
+ so this function should never be called */
|
|
+ WARN_ON(1);
|
|
+#endif
|
|
+
|
|
+ return err;
|
|
+}
|
|
+
|
|
static void __pci_set_master(struct pci_dev *dev, bool enable)
|
|
{
|
|
u16 old_cmd, cmd;
|
|
diff --git a/drivers/pci/probe.c b/drivers/pci/probe.c
|
|
index e3cf8a2..abf5e82 100644
|
|
--- a/drivers/pci/probe.c
|
|
+++ b/drivers/pci/probe.c
|
|
@@ -515,7 +515,7 @@ static void pci_release_host_bridge_dev(struct device *dev)
|
|
kfree(bridge);
|
|
}
|
|
|
|
-static struct pci_host_bridge *pci_alloc_host_bridge(struct pci_bus *b)
|
|
+static struct pci_host_bridge *pci_alloc_host_bridge(void)
|
|
{
|
|
struct pci_host_bridge *bridge;
|
|
|
|
@@ -524,7 +524,6 @@ static struct pci_host_bridge *pci_alloc_host_bridge(struct pci_bus *b)
|
|
return NULL;
|
|
|
|
INIT_LIST_HEAD(&bridge->windows);
|
|
- bridge->bus = b;
|
|
return bridge;
|
|
}
|
|
|
|
@@ -1749,8 +1748,9 @@ void __weak pcibios_remove_bus(struct pci_bus *bus)
|
|
{
|
|
}
|
|
|
|
-struct pci_bus *pci_create_root_bus(struct device *parent, int bus,
|
|
- struct pci_ops *ops, void *sysdata, struct list_head *resources)
|
|
+struct pci_bus *pci_create_root_bus_in_domain(struct device *parent,
|
|
+ int domain, int bus, struct pci_ops *ops, void *sysdata,
|
|
+ struct list_head *resources)
|
|
{
|
|
int error;
|
|
struct pci_host_bridge *bridge;
|
|
@@ -1761,37 +1761,41 @@ struct pci_bus *pci_create_root_bus(struct device *parent, int bus,
|
|
char bus_addr[64];
|
|
char *fmt;
|
|
|
|
+ bridge = pci_alloc_host_bridge();
|
|
+ if (!bridge)
|
|
+ return ERR_PTR(-ENOMEM);
|
|
+
|
|
+ bridge->dev.parent = parent;
|
|
+ bridge->dev.release = pci_release_host_bridge_dev;
|
|
+ bridge->domain_nr = domain;
|
|
+
|
|
b = pci_alloc_bus();
|
|
- if (!b)
|
|
- return NULL;
|
|
+ if (!b) {
|
|
+ error = -ENOMEM;
|
|
+ goto err_out;
|
|
+ }
|
|
|
|
b->sysdata = sysdata;
|
|
b->ops = ops;
|
|
b->number = b->busn_res.start = bus;
|
|
- b2 = pci_find_bus(pci_domain_nr(b), bus);
|
|
+ b2 = pci_find_bus(bridge->domain_nr, bus);
|
|
if (b2) {
|
|
/* If we already got to this bus through a different bridge, ignore it */
|
|
dev_dbg(&b2->dev, "bus already known\n");
|
|
- goto err_out;
|
|
+ error = -EEXIST;
|
|
+ goto err_bus_out;
|
|
}
|
|
|
|
- bridge = pci_alloc_host_bridge(b);
|
|
- if (!bridge)
|
|
- goto err_out;
|
|
-
|
|
- bridge->dev.parent = parent;
|
|
- bridge->dev.release = pci_release_host_bridge_dev;
|
|
- dev_set_name(&bridge->dev, "pci%04x:%02x", pci_domain_nr(b), bus);
|
|
+ bridge->bus = b;
|
|
+ dev_set_name(&bridge->dev, "pci%04x:%02x", bridge->domain_nr, bus);
|
|
error = pcibios_root_bridge_prepare(bridge);
|
|
- if (error) {
|
|
- kfree(bridge);
|
|
+ if (error)
|
|
goto err_out;
|
|
- }
|
|
|
|
error = device_register(&bridge->dev);
|
|
if (error) {
|
|
put_device(&bridge->dev);
|
|
- goto err_out;
|
|
+ goto err_bus_out;
|
|
}
|
|
b->bridge = get_device(&bridge->dev);
|
|
device_enable_async_suspend(b->bridge);
|
|
@@ -1802,7 +1806,7 @@ struct pci_bus *pci_create_root_bus(struct device *parent, int bus,
|
|
|
|
b->dev.class = &pcibus_class;
|
|
b->dev.parent = b->bridge;
|
|
- dev_set_name(&b->dev, "%04x:%02x", pci_domain_nr(b), bus);
|
|
+ dev_set_name(&b->dev, "%04x:%02x", bridge->domain_nr, bus);
|
|
error = device_register(&b->dev);
|
|
if (error)
|
|
goto class_dev_reg_err;
|
|
@@ -1848,9 +1852,31 @@ struct pci_bus *pci_create_root_bus(struct device *parent, int bus,
|
|
class_dev_reg_err:
|
|
put_device(&bridge->dev);
|
|
device_unregister(&bridge->dev);
|
|
+err_bus_out:
|
|
+ kfree(b);
|
|
err_out:
|
|
+ kfree(bridge);
|
|
+ return ERR_PTR(error);
|
|
+}
|
|
+
|
|
+struct pci_bus *pci_create_root_bus(struct device *parent, int bus,
|
|
+ struct pci_ops *ops, void *sysdata, struct list_head *resources)
|
|
+{
|
|
+ int domain_nr;
|
|
+ struct pci_bus *b = pci_alloc_bus();
|
|
+ if (!b)
|
|
+ return NULL;
|
|
+
|
|
+ b->sysdata = sysdata;
|
|
+ domain_nr = pci_domain_nr(b);
|
|
kfree(b);
|
|
- return NULL;
|
|
+
|
|
+ b = pci_create_root_bus_in_domain(parent, domain_nr, bus,
|
|
+ ops, sysdata, resources);
|
|
+ if (IS_ERR(b))
|
|
+ return NULL;
|
|
+
|
|
+ return b;
|
|
}
|
|
|
|
int pci_bus_insert_busn_res(struct pci_bus *b, int bus, int bus_max)
|
|
diff --git a/drivers/rtc/Kconfig b/drivers/rtc/Kconfig
|
|
index 0754f5c..4478a59 100644
|
|
--- a/drivers/rtc/Kconfig
|
|
+++ b/drivers/rtc/Kconfig
|
|
@@ -789,7 +789,7 @@ config RTC_DRV_DA9063
|
|
|
|
config RTC_DRV_EFI
|
|
tristate "EFI RTC"
|
|
- depends on IA64
|
|
+ depends on EFI
|
|
help
|
|
If you say yes here you will get support for the EFI
|
|
Real Time Clock.
|
|
diff --git a/drivers/rtc/Makefile b/drivers/rtc/Makefile
|
|
index 70347d0..f1dfc36 100644
|
|
--- a/drivers/rtc/Makefile
|
|
+++ b/drivers/rtc/Makefile
|
|
@@ -10,6 +10,10 @@ obj-$(CONFIG_RTC_SYSTOHC) += systohc.o
|
|
obj-$(CONFIG_RTC_CLASS) += rtc-core.o
|
|
rtc-core-y := class.o interface.o
|
|
|
|
+ifdef CONFIG_RTC_DRV_EFI
|
|
+rtc-core-y += rtc-efi-platform.o
|
|
+endif
|
|
+
|
|
rtc-core-$(CONFIG_RTC_INTF_DEV) += rtc-dev.o
|
|
rtc-core-$(CONFIG_RTC_INTF_PROC) += rtc-proc.o
|
|
rtc-core-$(CONFIG_RTC_INTF_SYSFS) += rtc-sysfs.o
|
|
diff --git a/drivers/rtc/rtc-efi-platform.c b/drivers/rtc/rtc-efi-platform.c
|
|
new file mode 100644
|
|
index 0000000..b40fbe3
|
|
--- /dev/null
|
|
+++ b/drivers/rtc/rtc-efi-platform.c
|
|
@@ -0,0 +1,31 @@
|
|
+/*
|
|
+ * Moved from arch/ia64/kernel/time.c
|
|
+ *
|
|
+ * Copyright (C) 1998-2003 Hewlett-Packard Co
|
|
+ * Stephane Eranian <eranian@hpl.hp.com>
|
|
+ * David Mosberger <davidm@hpl.hp.com>
|
|
+ * Copyright (C) 1999 Don Dugger <don.dugger@intel.com>
|
|
+ * Copyright (C) 1999-2000 VA Linux Systems
|
|
+ * Copyright (C) 1999-2000 Walt Drummond <drummond@valinux.com>
|
|
+ */
|
|
+#include <linux/init.h>
|
|
+#include <linux/kernel.h>
|
|
+#include <linux/module.h>
|
|
+#include <linux/efi.h>
|
|
+#include <linux/platform_device.h>
|
|
+
|
|
+static struct platform_device rtc_efi_dev = {
|
|
+ .name = "rtc-efi",
|
|
+ .id = -1,
|
|
+};
|
|
+
|
|
+static int __init rtc_init(void)
|
|
+{
|
|
+ if (efi_enabled(EFI_RUNTIME_SERVICES))
|
|
+ if (platform_device_register(&rtc_efi_dev) < 0)
|
|
+ pr_err("unable to register rtc device...\n");
|
|
+
|
|
+ /* not necessarily an error */
|
|
+ return 0;
|
|
+}
|
|
+module_init(rtc_init);
|
|
diff --git a/include/asm-generic/io.h b/include/asm-generic/io.h
|
|
index 975e1cc..2e2161b 100644
|
|
--- a/include/asm-generic/io.h
|
|
+++ b/include/asm-generic/io.h
|
|
@@ -331,7 +331,7 @@ static inline void iounmap(void __iomem *addr)
|
|
#ifndef CONFIG_GENERIC_IOMAP
|
|
static inline void __iomem *ioport_map(unsigned long port, unsigned int nr)
|
|
{
|
|
- return (void __iomem *) port;
|
|
+ return (void __iomem *)(PCI_IOBASE + (port & IO_SPACE_LIMIT));
|
|
}
|
|
|
|
static inline void ioport_unmap(void __iomem *p)
|
|
diff --git a/include/kvm/arm_vgic.h b/include/kvm/arm_vgic.h
|
|
index f27000f..35b0c12 100644
|
|
--- a/include/kvm/arm_vgic.h
|
|
+++ b/include/kvm/arm_vgic.h
|
|
@@ -24,7 +24,6 @@
|
|
#include <linux/irqreturn.h>
|
|
#include <linux/spinlock.h>
|
|
#include <linux/types.h>
|
|
-#include <linux/irqchip/arm-gic.h>
|
|
|
|
#define VGIC_NR_IRQS 256
|
|
#define VGIC_NR_SGIS 16
|
|
@@ -32,7 +31,9 @@
|
|
#define VGIC_NR_PRIVATE_IRQS (VGIC_NR_SGIS + VGIC_NR_PPIS)
|
|
#define VGIC_NR_SHARED_IRQS (VGIC_NR_IRQS - VGIC_NR_PRIVATE_IRQS)
|
|
#define VGIC_MAX_CPUS KVM_MAX_VCPUS
|
|
-#define VGIC_MAX_LRS (1 << 6)
|
|
+
|
|
+#define VGIC_V2_MAX_LRS (1 << 6)
|
|
+#define VGIC_V3_MAX_LRS 16
|
|
|
|
/* Sanity checks... */
|
|
#if (VGIC_MAX_CPUS > 8)
|
|
@@ -68,9 +69,62 @@ struct vgic_bytemap {
|
|
u32 shared[VGIC_NR_SHARED_IRQS / 4];
|
|
};
|
|
|
|
+struct kvm_vcpu;
|
|
+
|
|
+enum vgic_type {
|
|
+ VGIC_V2, /* Good ol' GICv2 */
|
|
+ VGIC_V3, /* New fancy GICv3 */
|
|
+};
|
|
+
|
|
+#define LR_STATE_PENDING (1 << 0)
|
|
+#define LR_STATE_ACTIVE (1 << 1)
|
|
+#define LR_STATE_MASK (3 << 0)
|
|
+#define LR_EOI_INT (1 << 2)
|
|
+
|
|
+struct vgic_lr {
|
|
+ u16 irq;
|
|
+ u8 source;
|
|
+ u8 state;
|
|
+};
|
|
+
|
|
+struct vgic_vmcr {
|
|
+ u32 ctlr;
|
|
+ u32 abpr;
|
|
+ u32 bpr;
|
|
+ u32 pmr;
|
|
+};
|
|
+
|
|
+struct vgic_ops {
|
|
+ struct vgic_lr (*get_lr)(const struct kvm_vcpu *, int);
|
|
+ void (*set_lr)(struct kvm_vcpu *, int, struct vgic_lr);
|
|
+ void (*sync_lr_elrsr)(struct kvm_vcpu *, int, struct vgic_lr);
|
|
+ u64 (*get_elrsr)(const struct kvm_vcpu *vcpu);
|
|
+ u64 (*get_eisr)(const struct kvm_vcpu *vcpu);
|
|
+ u32 (*get_interrupt_status)(const struct kvm_vcpu *vcpu);
|
|
+ void (*enable_underflow)(struct kvm_vcpu *vcpu);
|
|
+ void (*disable_underflow)(struct kvm_vcpu *vcpu);
|
|
+ void (*get_vmcr)(struct kvm_vcpu *vcpu, struct vgic_vmcr *vmcr);
|
|
+ void (*set_vmcr)(struct kvm_vcpu *vcpu, struct vgic_vmcr *vmcr);
|
|
+ void (*enable)(struct kvm_vcpu *vcpu);
|
|
+};
|
|
+
|
|
+struct vgic_params {
|
|
+ /* vgic type */
|
|
+ enum vgic_type type;
|
|
+ /* Physical address of vgic virtual cpu interface */
|
|
+ phys_addr_t vcpu_base;
|
|
+ /* Number of list registers */
|
|
+ u32 nr_lr;
|
|
+ /* Interrupt number */
|
|
+ unsigned int maint_irq;
|
|
+ /* Virtual control interface base address */
|
|
+ void __iomem *vctrl_base;
|
|
+};
|
|
+
|
|
struct vgic_dist {
|
|
#ifdef CONFIG_KVM_ARM_VGIC
|
|
spinlock_t lock;
|
|
+ bool in_kernel;
|
|
bool ready;
|
|
|
|
/* Virtual control interface mapping */
|
|
@@ -110,6 +164,29 @@ struct vgic_dist {
|
|
#endif
|
|
};
|
|
|
|
+struct vgic_v2_cpu_if {
|
|
+ u32 vgic_hcr;
|
|
+ u32 vgic_vmcr;
|
|
+ u32 vgic_misr; /* Saved only */
|
|
+ u32 vgic_eisr[2]; /* Saved only */
|
|
+ u32 vgic_elrsr[2]; /* Saved only */
|
|
+ u32 vgic_apr;
|
|
+ u32 vgic_lr[VGIC_V2_MAX_LRS];
|
|
+};
|
|
+
|
|
+struct vgic_v3_cpu_if {
|
|
+#ifdef CONFIG_ARM_GIC_V3
|
|
+ u32 vgic_hcr;
|
|
+ u32 vgic_vmcr;
|
|
+ u32 vgic_misr; /* Saved only */
|
|
+ u32 vgic_eisr; /* Saved only */
|
|
+ u32 vgic_elrsr; /* Saved only */
|
|
+ u32 vgic_ap0r[4];
|
|
+ u32 vgic_ap1r[4];
|
|
+ u64 vgic_lr[VGIC_V3_MAX_LRS];
|
|
+#endif
|
|
+};
|
|
+
|
|
struct vgic_cpu {
|
|
#ifdef CONFIG_KVM_ARM_VGIC
|
|
/* per IRQ to LR mapping */
|
|
@@ -120,24 +197,24 @@ struct vgic_cpu {
|
|
DECLARE_BITMAP( pending_shared, VGIC_NR_SHARED_IRQS);
|
|
|
|
/* Bitmap of used/free list registers */
|
|
- DECLARE_BITMAP( lr_used, VGIC_MAX_LRS);
|
|
+ DECLARE_BITMAP( lr_used, VGIC_V2_MAX_LRS);
|
|
|
|
/* Number of list registers on this CPU */
|
|
int nr_lr;
|
|
|
|
/* CPU vif control registers for world switch */
|
|
- u32 vgic_hcr;
|
|
- u32 vgic_vmcr;
|
|
- u32 vgic_misr; /* Saved only */
|
|
- u32 vgic_eisr[2]; /* Saved only */
|
|
- u32 vgic_elrsr[2]; /* Saved only */
|
|
- u32 vgic_apr;
|
|
- u32 vgic_lr[VGIC_MAX_LRS];
|
|
+ union {
|
|
+ struct vgic_v2_cpu_if vgic_v2;
|
|
+ struct vgic_v3_cpu_if vgic_v3;
|
|
+ };
|
|
#endif
|
|
};
|
|
|
|
#define LR_EMPTY 0xff
|
|
|
|
+#define INT_STATUS_EOI (1 << 0)
|
|
+#define INT_STATUS_UNDERFLOW (1 << 1)
|
|
+
|
|
struct kvm;
|
|
struct kvm_vcpu;
|
|
struct kvm_run;
|
|
@@ -157,9 +234,25 @@ int kvm_vgic_vcpu_pending_irq(struct kvm_vcpu *vcpu);
|
|
bool vgic_handle_mmio(struct kvm_vcpu *vcpu, struct kvm_run *run,
|
|
struct kvm_exit_mmio *mmio);
|
|
|
|
-#define irqchip_in_kernel(k) (!!((k)->arch.vgic.vctrl_base))
|
|
+#define irqchip_in_kernel(k) (!!((k)->arch.vgic.in_kernel))
|
|
#define vgic_initialized(k) ((k)->arch.vgic.ready)
|
|
|
|
+int vgic_v2_probe(struct device_node *vgic_node,
|
|
+ const struct vgic_ops **ops,
|
|
+ const struct vgic_params **params);
|
|
+#ifdef CONFIG_ARM_GIC_V3
|
|
+int vgic_v3_probe(struct device_node *vgic_node,
|
|
+ const struct vgic_ops **ops,
|
|
+ const struct vgic_params **params);
|
|
+#else
|
|
+static inline int vgic_v3_probe(struct device_node *vgic_node,
|
|
+ const struct vgic_ops **ops,
|
|
+ const struct vgic_params **params)
|
|
+{
|
|
+ return -ENODEV;
|
|
+}
|
|
+#endif
|
|
+
|
|
#else
|
|
static inline int kvm_vgic_hyp_init(void)
|
|
{
|
|
diff --git a/include/linux/of_address.h b/include/linux/of_address.h
|
|
index c13b878..33c0420 100644
|
|
--- a/include/linux/of_address.h
|
|
+++ b/include/linux/of_address.h
|
|
@@ -23,17 +23,8 @@ struct of_pci_range {
|
|
#define for_each_of_pci_range(parser, range) \
|
|
for (; of_pci_range_parser_one(parser, range);)
|
|
|
|
-static inline void of_pci_range_to_resource(struct of_pci_range *range,
|
|
- struct device_node *np,
|
|
- struct resource *res)
|
|
-{
|
|
- res->flags = range->flags;
|
|
- res->start = range->cpu_addr;
|
|
- res->end = range->cpu_addr + range->size - 1;
|
|
- res->parent = res->child = res->sibling = NULL;
|
|
- res->name = np->full_name;
|
|
-}
|
|
-
|
|
+extern int of_pci_range_to_resource(struct of_pci_range *range,
|
|
+ struct device_node *np, struct resource *res);
|
|
/* Translate a DMA address from device space to CPU space */
|
|
extern u64 of_translate_dma_address(struct device_node *dev,
|
|
const __be32 *in_addr);
|
|
@@ -55,6 +46,7 @@ extern void __iomem *of_iomap(struct device_node *device, int index);
|
|
extern const __be32 *of_get_address(struct device_node *dev, int index,
|
|
u64 *size, unsigned int *flags);
|
|
|
|
+extern int pci_register_io_range(phys_addr_t addr, resource_size_t size);
|
|
extern unsigned long pci_address_to_pio(phys_addr_t addr);
|
|
|
|
extern int of_pci_range_parser_init(struct of_pci_range_parser *parser,
|
|
diff --git a/include/linux/of_pci.h b/include/linux/of_pci.h
|
|
index dde3a4a..71e36d0 100644
|
|
--- a/include/linux/of_pci.h
|
|
+++ b/include/linux/of_pci.h
|
|
@@ -15,6 +15,9 @@ struct device_node *of_pci_find_child_device(struct device_node *parent,
|
|
int of_pci_get_devfn(struct device_node *np);
|
|
int of_irq_parse_and_map_pci(const struct pci_dev *dev, u8 slot, u8 pin);
|
|
int of_pci_parse_bus_range(struct device_node *node, struct resource *res);
|
|
+struct pci_host_bridge *of_create_pci_host_bridge(struct device *parent,
|
|
+ struct pci_ops *ops, void *host_data);
|
|
+
|
|
#else
|
|
static inline int of_irq_parse_pci(const struct pci_dev *pdev, struct of_phandle_args *out_irq)
|
|
{
|
|
@@ -43,6 +46,13 @@ of_pci_parse_bus_range(struct device_node *node, struct resource *res)
|
|
{
|
|
return -EINVAL;
|
|
}
|
|
+
|
|
+static inline struct pci_host_bridge *
|
|
+of_create_pci_host_bridge(struct device *parent, struct pci_ops *ops,
|
|
+ void *host_data)
|
|
+{
|
|
+ return NULL;
|
|
+}
|
|
#endif
|
|
|
|
#if defined(CONFIG_OF) && defined(CONFIG_PCI_MSI)
|
|
diff --git a/include/linux/pci.h b/include/linux/pci.h
|
|
index 61978a4..f582746 100644
|
|
--- a/include/linux/pci.h
|
|
+++ b/include/linux/pci.h
|
|
@@ -401,6 +401,8 @@ struct pci_host_bridge_window {
|
|
struct pci_host_bridge {
|
|
struct device dev;
|
|
struct pci_bus *bus; /* root bus */
|
|
+ int domain_nr;
|
|
+ resource_size_t io_base; /* physical address for the start of I/O area */
|
|
struct list_head windows; /* pci_host_bridge_windows */
|
|
void (*release_fn)(struct pci_host_bridge *);
|
|
void *release_data;
|
|
@@ -769,6 +771,9 @@ struct pci_bus *pci_scan_bus(int bus, struct pci_ops *ops, void *sysdata);
|
|
struct pci_bus *pci_create_root_bus(struct device *parent, int bus,
|
|
struct pci_ops *ops, void *sysdata,
|
|
struct list_head *resources);
|
|
+struct pci_bus *pci_create_root_bus_in_domain(struct device *parent,
|
|
+ int domain, int bus, struct pci_ops *ops,
|
|
+ void *sysdata, struct list_head *resources);
|
|
int pci_bus_insert_busn_res(struct pci_bus *b, int bus, int busmax);
|
|
int pci_bus_update_busn_res_end(struct pci_bus *b, int busmax);
|
|
void pci_bus_release_busn_res(struct pci_bus *b);
|
|
@@ -1097,6 +1102,9 @@ int __must_check pci_bus_alloc_resource(struct pci_bus *bus,
|
|
resource_size_t),
|
|
void *alignf_data);
|
|
|
|
+
|
|
+int pci_remap_iospace(const struct resource *res, phys_addr_t phys_addr);
|
|
+
|
|
static inline dma_addr_t pci_bus_address(struct pci_dev *pdev, int bar)
|
|
{
|
|
struct pci_bus_region region;
|
|
@@ -1815,8 +1823,15 @@ static inline void pci_set_of_node(struct pci_dev *dev) { }
|
|
static inline void pci_release_of_node(struct pci_dev *dev) { }
|
|
static inline void pci_set_bus_of_node(struct pci_bus *bus) { }
|
|
static inline void pci_release_bus_of_node(struct pci_bus *bus) { }
|
|
+
|
|
#endif /* CONFIG_OF */
|
|
|
|
+/* Used by architecture code to apply any quirks to the list of
|
|
+ * pci_host_bridge resource ranges before they are being used
|
|
+ * by of_create_pci_host_bridge()
|
|
+ */
|
|
+extern int pcibios_fixup_bridge_ranges(struct list_head *resources);
|
|
+
|
|
#ifdef CONFIG_EEH
|
|
static inline struct eeh_dev *pci_dev_to_eeh_dev(struct pci_dev *pdev)
|
|
{
|
|
diff --git a/tools/perf/arch/arm64/include/perf_regs.h b/tools/perf/arch/arm64/include/perf_regs.h
|
|
index e9441b9..1d3f39c 100644
|
|
--- a/tools/perf/arch/arm64/include/perf_regs.h
|
|
+++ b/tools/perf/arch/arm64/include/perf_regs.h
|
|
@@ -6,6 +6,8 @@
|
|
#include <asm/perf_regs.h>
|
|
|
|
#define PERF_REGS_MASK ((1ULL << PERF_REG_ARM64_MAX) - 1)
|
|
+#define PERF_REGS_MAX PERF_REG_ARM64_MAX
|
|
+
|
|
#define PERF_REG_IP PERF_REG_ARM64_PC
|
|
#define PERF_REG_SP PERF_REG_ARM64_SP
|
|
|
|
diff --git a/virt/kvm/arm/vgic-v2.c b/virt/kvm/arm/vgic-v2.c
|
|
new file mode 100644
|
|
index 0000000..ff597d8
|
|
--- /dev/null
|
|
+++ b/virt/kvm/arm/vgic-v2.c
|
|
@@ -0,0 +1,259 @@
|
|
+/*
|
|
+ * Copyright (C) 2012,2013 ARM Limited, All Rights Reserved.
|
|
+ * Author: Marc Zyngier <marc.zyngier@arm.com>
|
|
+ *
|
|
+ * This program is free software; you can redistribute it and/or modify
|
|
+ * it under the terms of the GNU General Public License version 2 as
|
|
+ * published by the Free Software Foundation.
|
|
+ *
|
|
+ * This program is distributed in the hope that it will be useful,
|
|
+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
+ * GNU General Public License for more details.
|
|
+ *
|
|
+ * You should have received a copy of the GNU General Public License
|
|
+ * along with this program. If not, see <http://www.gnu.org/licenses/>.
|
|
+ */
|
|
+
|
|
+#include <linux/cpu.h>
|
|
+#include <linux/kvm.h>
|
|
+#include <linux/kvm_host.h>
|
|
+#include <linux/interrupt.h>
|
|
+#include <linux/io.h>
|
|
+#include <linux/of.h>
|
|
+#include <linux/of_address.h>
|
|
+#include <linux/of_irq.h>
|
|
+
|
|
+#include <linux/irqchip/arm-gic.h>
|
|
+
|
|
+#include <asm/kvm_emulate.h>
|
|
+#include <asm/kvm_arm.h>
|
|
+#include <asm/kvm_mmu.h>
|
|
+
|
|
+static struct vgic_lr vgic_v2_get_lr(const struct kvm_vcpu *vcpu, int lr)
|
|
+{
|
|
+ struct vgic_lr lr_desc;
|
|
+ u32 val = vcpu->arch.vgic_cpu.vgic_v2.vgic_lr[lr];
|
|
+
|
|
+ lr_desc.irq = val & GICH_LR_VIRTUALID;
|
|
+ if (lr_desc.irq <= 15)
|
|
+ lr_desc.source = (val >> GICH_LR_PHYSID_CPUID_SHIFT) & 0x7;
|
|
+ else
|
|
+ lr_desc.source = 0;
|
|
+ lr_desc.state = 0;
|
|
+
|
|
+ if (val & GICH_LR_PENDING_BIT)
|
|
+ lr_desc.state |= LR_STATE_PENDING;
|
|
+ if (val & GICH_LR_ACTIVE_BIT)
|
|
+ lr_desc.state |= LR_STATE_ACTIVE;
|
|
+ if (val & GICH_LR_EOI)
|
|
+ lr_desc.state |= LR_EOI_INT;
|
|
+
|
|
+ return lr_desc;
|
|
+}
|
|
+
|
|
+static void vgic_v2_set_lr(struct kvm_vcpu *vcpu, int lr,
|
|
+ struct vgic_lr lr_desc)
|
|
+{
|
|
+ u32 lr_val = (lr_desc.source << GICH_LR_PHYSID_CPUID_SHIFT) | lr_desc.irq;
|
|
+
|
|
+ if (lr_desc.state & LR_STATE_PENDING)
|
|
+ lr_val |= GICH_LR_PENDING_BIT;
|
|
+ if (lr_desc.state & LR_STATE_ACTIVE)
|
|
+ lr_val |= GICH_LR_ACTIVE_BIT;
|
|
+ if (lr_desc.state & LR_EOI_INT)
|
|
+ lr_val |= GICH_LR_EOI;
|
|
+
|
|
+ vcpu->arch.vgic_cpu.vgic_v2.vgic_lr[lr] = lr_val;
|
|
+}
|
|
+
|
|
+static void vgic_v2_sync_lr_elrsr(struct kvm_vcpu *vcpu, int lr,
|
|
+ struct vgic_lr lr_desc)
|
|
+{
|
|
+ if (!(lr_desc.state & LR_STATE_MASK))
|
|
+ set_bit(lr, (unsigned long *)vcpu->arch.vgic_cpu.vgic_v2.vgic_elrsr);
|
|
+}
|
|
+
|
|
+static u64 vgic_v2_get_elrsr(const struct kvm_vcpu *vcpu)
|
|
+{
|
|
+ u64 val;
|
|
+
|
|
+ val = vcpu->arch.vgic_cpu.vgic_v2.vgic_elrsr[1];
|
|
+ val <<= 32;
|
|
+ val |= vcpu->arch.vgic_cpu.vgic_v2.vgic_elrsr[0];
|
|
+
|
|
+ return val;
|
|
+}
|
|
+
|
|
+static u64 vgic_v2_get_eisr(const struct kvm_vcpu *vcpu)
|
|
+{
|
|
+ u64 val;
|
|
+
|
|
+ val = vcpu->arch.vgic_cpu.vgic_v2.vgic_eisr[1];
|
|
+ val <<= 32;
|
|
+ val |= vcpu->arch.vgic_cpu.vgic_v2.vgic_eisr[0];
|
|
+
|
|
+ return val;
|
|
+}
|
|
+
|
|
+static u32 vgic_v2_get_interrupt_status(const struct kvm_vcpu *vcpu)
|
|
+{
|
|
+ u32 misr = vcpu->arch.vgic_cpu.vgic_v2.vgic_misr;
|
|
+ u32 ret = 0;
|
|
+
|
|
+ if (misr & GICH_MISR_EOI)
|
|
+ ret |= INT_STATUS_EOI;
|
|
+ if (misr & GICH_MISR_U)
|
|
+ ret |= INT_STATUS_UNDERFLOW;
|
|
+
|
|
+ return ret;
|
|
+}
|
|
+
|
|
+static void vgic_v2_enable_underflow(struct kvm_vcpu *vcpu)
|
|
+{
|
|
+ vcpu->arch.vgic_cpu.vgic_v2.vgic_hcr |= GICH_HCR_UIE;
|
|
+}
|
|
+
|
|
+static void vgic_v2_disable_underflow(struct kvm_vcpu *vcpu)
|
|
+{
|
|
+ vcpu->arch.vgic_cpu.vgic_v2.vgic_hcr &= ~GICH_HCR_UIE;
|
|
+}
|
|
+
|
|
+static void vgic_v2_get_vmcr(struct kvm_vcpu *vcpu, struct vgic_vmcr *vmcrp)
|
|
+{
|
|
+ u32 vmcr = vcpu->arch.vgic_cpu.vgic_v2.vgic_vmcr;
|
|
+
|
|
+ vmcrp->ctlr = (vmcr & GICH_VMCR_CTRL_MASK) >> GICH_VMCR_CTRL_SHIFT;
|
|
+ vmcrp->abpr = (vmcr & GICH_VMCR_ALIAS_BINPOINT_MASK) >> GICH_VMCR_ALIAS_BINPOINT_SHIFT;
|
|
+ vmcrp->bpr = (vmcr & GICH_VMCR_BINPOINT_MASK) >> GICH_VMCR_BINPOINT_SHIFT;
|
|
+ vmcrp->pmr = (vmcr & GICH_VMCR_PRIMASK_MASK) >> GICH_VMCR_PRIMASK_SHIFT;
|
|
+}
|
|
+
|
|
+static void vgic_v2_set_vmcr(struct kvm_vcpu *vcpu, struct vgic_vmcr *vmcrp)
|
|
+{
|
|
+ u32 vmcr;
|
|
+
|
|
+ vmcr = (vmcrp->ctlr << GICH_VMCR_CTRL_SHIFT) & GICH_VMCR_CTRL_MASK;
|
|
+ vmcr |= (vmcrp->abpr << GICH_VMCR_ALIAS_BINPOINT_SHIFT) & GICH_VMCR_ALIAS_BINPOINT_MASK;
|
|
+ vmcr |= (vmcrp->bpr << GICH_VMCR_BINPOINT_SHIFT) & GICH_VMCR_BINPOINT_MASK;
|
|
+ vmcr |= (vmcrp->pmr << GICH_VMCR_PRIMASK_SHIFT) & GICH_VMCR_PRIMASK_MASK;
|
|
+
|
|
+ vcpu->arch.vgic_cpu.vgic_v2.vgic_vmcr = vmcr;
|
|
+}
|
|
+
|
|
+static void vgic_v2_enable(struct kvm_vcpu *vcpu)
|
|
+{
|
|
+ /*
|
|
+ * By forcing VMCR to zero, the GIC will restore the binary
|
|
+ * points to their reset values. Anything else resets to zero
|
|
+ * anyway.
|
|
+ */
|
|
+ vcpu->arch.vgic_cpu.vgic_v2.vgic_vmcr = 0;
|
|
+
|
|
+ /* Get the show on the road... */
|
|
+ vcpu->arch.vgic_cpu.vgic_v2.vgic_hcr = GICH_HCR_EN;
|
|
+}
|
|
+
|
|
+static const struct vgic_ops vgic_v2_ops = {
|
|
+ .get_lr = vgic_v2_get_lr,
|
|
+ .set_lr = vgic_v2_set_lr,
|
|
+ .sync_lr_elrsr = vgic_v2_sync_lr_elrsr,
|
|
+ .get_elrsr = vgic_v2_get_elrsr,
|
|
+ .get_eisr = vgic_v2_get_eisr,
|
|
+ .get_interrupt_status = vgic_v2_get_interrupt_status,
|
|
+ .enable_underflow = vgic_v2_enable_underflow,
|
|
+ .disable_underflow = vgic_v2_disable_underflow,
|
|
+ .get_vmcr = vgic_v2_get_vmcr,
|
|
+ .set_vmcr = vgic_v2_set_vmcr,
|
|
+ .enable = vgic_v2_enable,
|
|
+};
|
|
+
|
|
+static struct vgic_params vgic_v2_params;
|
|
+
|
|
+/**
|
|
+ * vgic_v2_probe - probe for a GICv2 compatible interrupt controller in DT
|
|
+ * @node: pointer to the DT node
|
|
+ * @ops: address of a pointer to the GICv2 operations
|
|
+ * @params: address of a pointer to HW-specific parameters
|
|
+ *
|
|
+ * Returns 0 if a GICv2 has been found, with the low level operations
|
|
+ * in *ops and the HW parameters in *params. Returns an error code
|
|
+ * otherwise.
|
|
+ */
|
|
+int vgic_v2_probe(struct device_node *vgic_node,
|
|
+ const struct vgic_ops **ops,
|
|
+ const struct vgic_params **params)
|
|
+{
|
|
+ int ret;
|
|
+ struct resource vctrl_res;
|
|
+ struct resource vcpu_res;
|
|
+ struct vgic_params *vgic = &vgic_v2_params;
|
|
+
|
|
+ vgic->maint_irq = irq_of_parse_and_map(vgic_node, 0);
|
|
+ if (!vgic->maint_irq) {
|
|
+ kvm_err("error getting vgic maintenance irq from DT\n");
|
|
+ ret = -ENXIO;
|
|
+ goto out;
|
|
+ }
|
|
+
|
|
+ ret = of_address_to_resource(vgic_node, 2, &vctrl_res);
|
|
+ if (ret) {
|
|
+ kvm_err("Cannot obtain GICH resource\n");
|
|
+ goto out;
|
|
+ }
|
|
+
|
|
+ vgic->vctrl_base = of_iomap(vgic_node, 2);
|
|
+ if (!vgic->vctrl_base) {
|
|
+ kvm_err("Cannot ioremap GICH\n");
|
|
+ ret = -ENOMEM;
|
|
+ goto out;
|
|
+ }
|
|
+
|
|
+ vgic->nr_lr = readl_relaxed(vgic->vctrl_base + GICH_VTR);
|
|
+ vgic->nr_lr = (vgic->nr_lr & 0x3f) + 1;
|
|
+
|
|
+ ret = create_hyp_io_mappings(vgic->vctrl_base,
|
|
+ vgic->vctrl_base + resource_size(&vctrl_res),
|
|
+ vctrl_res.start);
|
|
+ if (ret) {
|
|
+ kvm_err("Cannot map VCTRL into hyp\n");
|
|
+ goto out_unmap;
|
|
+ }
|
|
+
|
|
+ if (of_address_to_resource(vgic_node, 3, &vcpu_res)) {
|
|
+ kvm_err("Cannot obtain GICV resource\n");
|
|
+ ret = -ENXIO;
|
|
+ goto out_unmap;
|
|
+ }
|
|
+
|
|
+ if (!PAGE_ALIGNED(vcpu_res.start)) {
|
|
+ kvm_err("GICV physical address 0x%llx not page aligned\n",
|
|
+ (unsigned long long)vcpu_res.start);
|
|
+ ret = -ENXIO;
|
|
+ goto out_unmap;
|
|
+ }
|
|
+
|
|
+ if (!PAGE_ALIGNED(resource_size(&vcpu_res))) {
|
|
+ kvm_err("GICV size 0x%llx not a multiple of page size 0x%lx\n",
|
|
+ (unsigned long long)resource_size(&vcpu_res),
|
|
+ PAGE_SIZE);
|
|
+ ret = -ENXIO;
|
|
+ goto out_unmap;
|
|
+ }
|
|
+
|
|
+ vgic->vcpu_base = vcpu_res.start;
|
|
+
|
|
+ kvm_info("%s@%llx IRQ%d\n", vgic_node->name,
|
|
+ vgic->vctrl_base, vgic->maint_irq);
|
|
+
|
|
+ vgic->type = VGIC_V2;
|
|
+ *ops = &vgic_v2_ops;
|
|
+ *params = vgic;
|
|
+ goto out;
|
|
+
|
|
+out_unmap:
|
|
+ iounmap(vgic->vctrl_base);
|
|
+out:
|
|
+ of_node_put(vgic_node);
|
|
+ return ret;
|
|
+}
|
|
diff --git a/virt/kvm/arm/vgic-v3.c b/virt/kvm/arm/vgic-v3.c
|
|
new file mode 100644
|
|
index 0000000..f01d446
|
|
--- /dev/null
|
|
+++ b/virt/kvm/arm/vgic-v3.c
|
|
@@ -0,0 +1,231 @@
|
|
+/*
|
|
+ * Copyright (C) 2013 ARM Limited, All Rights Reserved.
|
|
+ * Author: Marc Zyngier <marc.zyngier@arm.com>
|
|
+ *
|
|
+ * This program is free software; you can redistribute it and/or modify
|
|
+ * it under the terms of the GNU General Public License version 2 as
|
|
+ * published by the Free Software Foundation.
|
|
+ *
|
|
+ * This program is distributed in the hope that it will be useful,
|
|
+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
+ * GNU General Public License for more details.
|
|
+ *
|
|
+ * You should have received a copy of the GNU General Public License
|
|
+ * along with this program. If not, see <http://www.gnu.org/licenses/>.
|
|
+ */
|
|
+
|
|
+#include <linux/cpu.h>
|
|
+#include <linux/kvm.h>
|
|
+#include <linux/kvm_host.h>
|
|
+#include <linux/interrupt.h>
|
|
+#include <linux/io.h>
|
|
+#include <linux/of.h>
|
|
+#include <linux/of_address.h>
|
|
+#include <linux/of_irq.h>
|
|
+
|
|
+#include <linux/irqchip/arm-gic-v3.h>
|
|
+
|
|
+#include <asm/kvm_emulate.h>
|
|
+#include <asm/kvm_arm.h>
|
|
+#include <asm/kvm_mmu.h>
|
|
+
|
|
+/* These are for GICv2 emulation only */
|
|
+#define GICH_LR_VIRTUALID (0x3ffUL << 0)
|
|
+#define GICH_LR_PHYSID_CPUID_SHIFT (10)
|
|
+#define GICH_LR_PHYSID_CPUID (7UL << GICH_LR_PHYSID_CPUID_SHIFT)
|
|
+
|
|
+/*
|
|
+ * LRs are stored in reverse order in memory. make sure we index them
|
|
+ * correctly.
|
|
+ */
|
|
+#define LR_INDEX(lr) (VGIC_V3_MAX_LRS - 1 - lr)
|
|
+
|
|
+static u32 ich_vtr_el2;
|
|
+
|
|
+static struct vgic_lr vgic_v3_get_lr(const struct kvm_vcpu *vcpu, int lr)
|
|
+{
|
|
+ struct vgic_lr lr_desc;
|
|
+ u64 val = vcpu->arch.vgic_cpu.vgic_v3.vgic_lr[LR_INDEX(lr)];
|
|
+
|
|
+ lr_desc.irq = val & GICH_LR_VIRTUALID;
|
|
+ if (lr_desc.irq <= 15)
|
|
+ lr_desc.source = (val >> GICH_LR_PHYSID_CPUID_SHIFT) & 0x7;
|
|
+ else
|
|
+ lr_desc.source = 0;
|
|
+ lr_desc.state = 0;
|
|
+
|
|
+ if (val & ICH_LR_PENDING_BIT)
|
|
+ lr_desc.state |= LR_STATE_PENDING;
|
|
+ if (val & ICH_LR_ACTIVE_BIT)
|
|
+ lr_desc.state |= LR_STATE_ACTIVE;
|
|
+ if (val & ICH_LR_EOI)
|
|
+ lr_desc.state |= LR_EOI_INT;
|
|
+
|
|
+ return lr_desc;
|
|
+}
|
|
+
|
|
+static void vgic_v3_set_lr(struct kvm_vcpu *vcpu, int lr,
|
|
+ struct vgic_lr lr_desc)
|
|
+{
|
|
+ u64 lr_val = (((u32)lr_desc.source << GICH_LR_PHYSID_CPUID_SHIFT) |
|
|
+ lr_desc.irq);
|
|
+
|
|
+ if (lr_desc.state & LR_STATE_PENDING)
|
|
+ lr_val |= ICH_LR_PENDING_BIT;
|
|
+ if (lr_desc.state & LR_STATE_ACTIVE)
|
|
+ lr_val |= ICH_LR_ACTIVE_BIT;
|
|
+ if (lr_desc.state & LR_EOI_INT)
|
|
+ lr_val |= ICH_LR_EOI;
|
|
+
|
|
+ vcpu->arch.vgic_cpu.vgic_v3.vgic_lr[LR_INDEX(lr)] = lr_val;
|
|
+}
|
|
+
|
|
+static void vgic_v3_sync_lr_elrsr(struct kvm_vcpu *vcpu, int lr,
|
|
+ struct vgic_lr lr_desc)
|
|
+{
|
|
+ if (!(lr_desc.state & LR_STATE_MASK))
|
|
+ vcpu->arch.vgic_cpu.vgic_v3.vgic_elrsr |= (1U << lr);
|
|
+}
|
|
+
|
|
+static u64 vgic_v3_get_elrsr(const struct kvm_vcpu *vcpu)
|
|
+{
|
|
+ return vcpu->arch.vgic_cpu.vgic_v3.vgic_elrsr;
|
|
+}
|
|
+
|
|
+static u64 vgic_v3_get_eisr(const struct kvm_vcpu *vcpu)
|
|
+{
|
|
+ return vcpu->arch.vgic_cpu.vgic_v3.vgic_eisr;
|
|
+}
|
|
+
|
|
+static u32 vgic_v3_get_interrupt_status(const struct kvm_vcpu *vcpu)
|
|
+{
|
|
+ u32 misr = vcpu->arch.vgic_cpu.vgic_v3.vgic_misr;
|
|
+ u32 ret = 0;
|
|
+
|
|
+ if (misr & ICH_MISR_EOI)
|
|
+ ret |= INT_STATUS_EOI;
|
|
+ if (misr & ICH_MISR_U)
|
|
+ ret |= INT_STATUS_UNDERFLOW;
|
|
+
|
|
+ return ret;
|
|
+}
|
|
+
|
|
+static void vgic_v3_get_vmcr(struct kvm_vcpu *vcpu, struct vgic_vmcr *vmcrp)
|
|
+{
|
|
+ u32 vmcr = vcpu->arch.vgic_cpu.vgic_v3.vgic_vmcr;
|
|
+
|
|
+ vmcrp->ctlr = (vmcr & ICH_VMCR_CTLR_MASK) >> ICH_VMCR_CTLR_SHIFT;
|
|
+ vmcrp->abpr = (vmcr & ICH_VMCR_BPR1_MASK) >> ICH_VMCR_BPR1_SHIFT;
|
|
+ vmcrp->bpr = (vmcr & ICH_VMCR_BPR0_MASK) >> ICH_VMCR_BPR0_SHIFT;
|
|
+ vmcrp->pmr = (vmcr & ICH_VMCR_PMR_MASK) >> ICH_VMCR_PMR_SHIFT;
|
|
+}
|
|
+
|
|
+static void vgic_v3_enable_underflow(struct kvm_vcpu *vcpu)
|
|
+{
|
|
+ vcpu->arch.vgic_cpu.vgic_v3.vgic_hcr |= ICH_HCR_UIE;
|
|
+}
|
|
+
|
|
+static void vgic_v3_disable_underflow(struct kvm_vcpu *vcpu)
|
|
+{
|
|
+ vcpu->arch.vgic_cpu.vgic_v3.vgic_hcr &= ~ICH_HCR_UIE;
|
|
+}
|
|
+
|
|
+static void vgic_v3_set_vmcr(struct kvm_vcpu *vcpu, struct vgic_vmcr *vmcrp)
|
|
+{
|
|
+ u32 vmcr;
|
|
+
|
|
+ vmcr = (vmcrp->ctlr << ICH_VMCR_CTLR_SHIFT) & ICH_VMCR_CTLR_MASK;
|
|
+ vmcr |= (vmcrp->abpr << ICH_VMCR_BPR1_SHIFT) & ICH_VMCR_BPR1_MASK;
|
|
+ vmcr |= (vmcrp->bpr << ICH_VMCR_BPR0_SHIFT) & ICH_VMCR_BPR0_MASK;
|
|
+ vmcr |= (vmcrp->pmr << ICH_VMCR_PMR_SHIFT) & ICH_VMCR_PMR_MASK;
|
|
+
|
|
+ vcpu->arch.vgic_cpu.vgic_v3.vgic_vmcr = vmcr;
|
|
+}
|
|
+
|
|
+static void vgic_v3_enable(struct kvm_vcpu *vcpu)
|
|
+{
|
|
+ /*
|
|
+ * By forcing VMCR to zero, the GIC will restore the binary
|
|
+ * points to their reset values. Anything else resets to zero
|
|
+ * anyway.
|
|
+ */
|
|
+ vcpu->arch.vgic_cpu.vgic_v3.vgic_vmcr = 0;
|
|
+
|
|
+ /* Get the show on the road... */
|
|
+ vcpu->arch.vgic_cpu.vgic_v3.vgic_hcr = ICH_HCR_EN;
|
|
+}
|
|
+
|
|
+static const struct vgic_ops vgic_v3_ops = {
|
|
+ .get_lr = vgic_v3_get_lr,
|
|
+ .set_lr = vgic_v3_set_lr,
|
|
+ .sync_lr_elrsr = vgic_v3_sync_lr_elrsr,
|
|
+ .get_elrsr = vgic_v3_get_elrsr,
|
|
+ .get_eisr = vgic_v3_get_eisr,
|
|
+ .get_interrupt_status = vgic_v3_get_interrupt_status,
|
|
+ .enable_underflow = vgic_v3_enable_underflow,
|
|
+ .disable_underflow = vgic_v3_disable_underflow,
|
|
+ .get_vmcr = vgic_v3_get_vmcr,
|
|
+ .set_vmcr = vgic_v3_set_vmcr,
|
|
+ .enable = vgic_v3_enable,
|
|
+};
|
|
+
|
|
+static struct vgic_params vgic_v3_params;
|
|
+
|
|
+/**
|
|
+ * vgic_v3_probe - probe for a GICv3 compatible interrupt controller in DT
|
|
+ * @node: pointer to the DT node
|
|
+ * @ops: address of a pointer to the GICv3 operations
|
|
+ * @params: address of a pointer to HW-specific parameters
|
|
+ *
|
|
+ * Returns 0 if a GICv3 has been found, with the low level operations
|
|
+ * in *ops and the HW parameters in *params. Returns an error code
|
|
+ * otherwise.
|
|
+ */
|
|
+int vgic_v3_probe(struct device_node *vgic_node,
|
|
+ const struct vgic_ops **ops,
|
|
+ const struct vgic_params **params)
|
|
+{
|
|
+ int ret = 0;
|
|
+ u32 gicv_idx;
|
|
+ struct resource vcpu_res;
|
|
+ struct vgic_params *vgic = &vgic_v3_params;
|
|
+
|
|
+ vgic->maint_irq = irq_of_parse_and_map(vgic_node, 0);
|
|
+ if (!vgic->maint_irq) {
|
|
+ kvm_err("error getting vgic maintenance irq from DT\n");
|
|
+ ret = -ENXIO;
|
|
+ goto out;
|
|
+ }
|
|
+
|
|
+ ich_vtr_el2 = kvm_call_hyp(__vgic_v3_get_ich_vtr_el2);
|
|
+
|
|
+ /*
|
|
+ * The ListRegs field is 5 bits, but there is a architectural
|
|
+ * maximum of 16 list registers. Just ignore bit 4...
|
|
+ */
|
|
+ vgic->nr_lr = (ich_vtr_el2 & 0xf) + 1;
|
|
+
|
|
+ if (of_property_read_u32(vgic_node, "#redistributor-regions", &gicv_idx))
|
|
+ gicv_idx = 1;
|
|
+
|
|
+ gicv_idx += 3; /* Also skip GICD, GICC, GICH */
|
|
+ if (of_address_to_resource(vgic_node, gicv_idx, &vcpu_res)) {
|
|
+ kvm_err("Cannot obtain GICV region\n");
|
|
+ ret = -ENXIO;
|
|
+ goto out;
|
|
+ }
|
|
+ vgic->vcpu_base = vcpu_res.start;
|
|
+ vgic->vctrl_base = NULL;
|
|
+ vgic->type = VGIC_V3;
|
|
+
|
|
+ kvm_info("%s@%llx IRQ%d\n", vgic_node->name,
|
|
+ vcpu_res.start, vgic->maint_irq);
|
|
+
|
|
+ *ops = &vgic_v3_ops;
|
|
+ *params = vgic;
|
|
+
|
|
+out:
|
|
+ of_node_put(vgic_node);
|
|
+ return ret;
|
|
+}
|
|
diff --git a/virt/kvm/arm/vgic.c b/virt/kvm/arm/vgic.c
|
|
index 476d3bf..795ab48 100644
|
|
--- a/virt/kvm/arm/vgic.c
|
|
+++ b/virt/kvm/arm/vgic.c
|
|
@@ -76,14 +76,6 @@
|
|
#define IMPLEMENTER_ARM 0x43b
|
|
#define GICC_ARCH_VERSION_V2 0x2
|
|
|
|
-/* Physical address of vgic virtual cpu interface */
|
|
-static phys_addr_t vgic_vcpu_base;
|
|
-
|
|
-/* Virtual control interface base address */
|
|
-static void __iomem *vgic_vctrl_base;
|
|
-
|
|
-static struct device_node *vgic_node;
|
|
-
|
|
#define ACCESS_READ_VALUE (1 << 0)
|
|
#define ACCESS_READ_RAZ (0 << 0)
|
|
#define ACCESS_READ_MASK(x) ((x) & (1 << 0))
|
|
@@ -94,12 +86,17 @@ static struct device_node *vgic_node;
|
|
#define ACCESS_WRITE_MASK(x) ((x) & (3 << 1))
|
|
|
|
static void vgic_retire_disabled_irqs(struct kvm_vcpu *vcpu);
|
|
+static void vgic_retire_lr(int lr_nr, int irq, struct kvm_vcpu *vcpu);
|
|
static void vgic_update_state(struct kvm *kvm);
|
|
static void vgic_kick_vcpus(struct kvm *kvm);
|
|
static void vgic_dispatch_sgi(struct kvm_vcpu *vcpu, u32 reg);
|
|
-static u32 vgic_nr_lr;
|
|
+static struct vgic_lr vgic_get_lr(const struct kvm_vcpu *vcpu, int lr);
|
|
+static void vgic_set_lr(struct kvm_vcpu *vcpu, int lr, struct vgic_lr lr_desc);
|
|
+static void vgic_get_vmcr(struct kvm_vcpu *vcpu, struct vgic_vmcr *vmcr);
|
|
+static void vgic_set_vmcr(struct kvm_vcpu *vcpu, struct vgic_vmcr *vmcr);
|
|
|
|
-static unsigned int vgic_maint_irq;
|
|
+static const struct vgic_ops *vgic_ops;
|
|
+static const struct vgic_params *vgic;
|
|
|
|
static u32 *vgic_bitmap_get_reg(struct vgic_bitmap *x,
|
|
int cpuid, u32 offset)
|
|
@@ -593,18 +590,6 @@ static bool handle_mmio_sgi_reg(struct kvm_vcpu *vcpu,
|
|
return false;
|
|
}
|
|
|
|
-#define LR_CPUID(lr) \
|
|
- (((lr) & GICH_LR_PHYSID_CPUID) >> GICH_LR_PHYSID_CPUID_SHIFT)
|
|
-#define LR_IRQID(lr) \
|
|
- ((lr) & GICH_LR_VIRTUALID)
|
|
-
|
|
-static void vgic_retire_lr(int lr_nr, int irq, struct vgic_cpu *vgic_cpu)
|
|
-{
|
|
- clear_bit(lr_nr, vgic_cpu->lr_used);
|
|
- vgic_cpu->vgic_lr[lr_nr] &= ~GICH_LR_STATE;
|
|
- vgic_cpu->vgic_irq_lr_map[irq] = LR_EMPTY;
|
|
-}
|
|
-
|
|
/**
|
|
* vgic_unqueue_irqs - move pending IRQs from LRs to the distributor
|
|
* @vgic_cpu: Pointer to the vgic_cpu struct holding the LRs
|
|
@@ -622,13 +607,10 @@ static void vgic_unqueue_irqs(struct kvm_vcpu *vcpu)
|
|
struct vgic_dist *dist = &vcpu->kvm->arch.vgic;
|
|
struct vgic_cpu *vgic_cpu = &vcpu->arch.vgic_cpu;
|
|
int vcpu_id = vcpu->vcpu_id;
|
|
- int i, irq, source_cpu;
|
|
- u32 *lr;
|
|
+ int i;
|
|
|
|
for_each_set_bit(i, vgic_cpu->lr_used, vgic_cpu->nr_lr) {
|
|
- lr = &vgic_cpu->vgic_lr[i];
|
|
- irq = LR_IRQID(*lr);
|
|
- source_cpu = LR_CPUID(*lr);
|
|
+ struct vgic_lr lr = vgic_get_lr(vcpu, i);
|
|
|
|
/*
|
|
* There are three options for the state bits:
|
|
@@ -640,7 +622,7 @@ static void vgic_unqueue_irqs(struct kvm_vcpu *vcpu)
|
|
* If the LR holds only an active interrupt (not pending) then
|
|
* just leave it alone.
|
|
*/
|
|
- if ((*lr & GICH_LR_STATE) == GICH_LR_ACTIVE_BIT)
|
|
+ if ((lr.state & LR_STATE_MASK) == LR_STATE_ACTIVE)
|
|
continue;
|
|
|
|
/*
|
|
@@ -649,18 +631,19 @@ static void vgic_unqueue_irqs(struct kvm_vcpu *vcpu)
|
|
* is fine, then we are only setting a few bits that were
|
|
* already set.
|
|
*/
|
|
- vgic_dist_irq_set(vcpu, irq);
|
|
- if (irq < VGIC_NR_SGIS)
|
|
- dist->irq_sgi_sources[vcpu_id][irq] |= 1 << source_cpu;
|
|
- *lr &= ~GICH_LR_PENDING_BIT;
|
|
+ vgic_dist_irq_set(vcpu, lr.irq);
|
|
+ if (lr.irq < VGIC_NR_SGIS)
|
|
+ dist->irq_sgi_sources[vcpu_id][lr.irq] |= 1 << lr.source;
|
|
+ lr.state &= ~LR_STATE_PENDING;
|
|
+ vgic_set_lr(vcpu, i, lr);
|
|
|
|
/*
|
|
* If there's no state left on the LR (it could still be
|
|
* active), then the LR does not hold any useful info and can
|
|
* be marked as free for other use.
|
|
*/
|
|
- if (!(*lr & GICH_LR_STATE))
|
|
- vgic_retire_lr(i, irq, vgic_cpu);
|
|
+ if (!(lr.state & LR_STATE_MASK))
|
|
+ vgic_retire_lr(i, lr.irq, vcpu);
|
|
|
|
/* Finally update the VGIC state. */
|
|
vgic_update_state(vcpu->kvm);
|
|
@@ -989,8 +972,73 @@ static void vgic_update_state(struct kvm *kvm)
|
|
}
|
|
}
|
|
|
|
-#define MK_LR_PEND(src, irq) \
|
|
- (GICH_LR_PENDING_BIT | ((src) << GICH_LR_PHYSID_CPUID_SHIFT) | (irq))
|
|
+static struct vgic_lr vgic_get_lr(const struct kvm_vcpu *vcpu, int lr)
|
|
+{
|
|
+ return vgic_ops->get_lr(vcpu, lr);
|
|
+}
|
|
+
|
|
+static void vgic_set_lr(struct kvm_vcpu *vcpu, int lr,
|
|
+ struct vgic_lr vlr)
|
|
+{
|
|
+ vgic_ops->set_lr(vcpu, lr, vlr);
|
|
+}
|
|
+
|
|
+static void vgic_sync_lr_elrsr(struct kvm_vcpu *vcpu, int lr,
|
|
+ struct vgic_lr vlr)
|
|
+{
|
|
+ vgic_ops->sync_lr_elrsr(vcpu, lr, vlr);
|
|
+}
|
|
+
|
|
+static inline u64 vgic_get_elrsr(struct kvm_vcpu *vcpu)
|
|
+{
|
|
+ return vgic_ops->get_elrsr(vcpu);
|
|
+}
|
|
+
|
|
+static inline u64 vgic_get_eisr(struct kvm_vcpu *vcpu)
|
|
+{
|
|
+ return vgic_ops->get_eisr(vcpu);
|
|
+}
|
|
+
|
|
+static inline u32 vgic_get_interrupt_status(struct kvm_vcpu *vcpu)
|
|
+{
|
|
+ return vgic_ops->get_interrupt_status(vcpu);
|
|
+}
|
|
+
|
|
+static inline void vgic_enable_underflow(struct kvm_vcpu *vcpu)
|
|
+{
|
|
+ vgic_ops->enable_underflow(vcpu);
|
|
+}
|
|
+
|
|
+static inline void vgic_disable_underflow(struct kvm_vcpu *vcpu)
|
|
+{
|
|
+ vgic_ops->disable_underflow(vcpu);
|
|
+}
|
|
+
|
|
+static inline void vgic_get_vmcr(struct kvm_vcpu *vcpu, struct vgic_vmcr *vmcr)
|
|
+{
|
|
+ vgic_ops->get_vmcr(vcpu, vmcr);
|
|
+}
|
|
+
|
|
+static void vgic_set_vmcr(struct kvm_vcpu *vcpu, struct vgic_vmcr *vmcr)
|
|
+{
|
|
+ vgic_ops->set_vmcr(vcpu, vmcr);
|
|
+}
|
|
+
|
|
+static inline void vgic_enable(struct kvm_vcpu *vcpu)
|
|
+{
|
|
+ vgic_ops->enable(vcpu);
|
|
+}
|
|
+
|
|
+static void vgic_retire_lr(int lr_nr, int irq, struct kvm_vcpu *vcpu)
|
|
+{
|
|
+ struct vgic_cpu *vgic_cpu = &vcpu->arch.vgic_cpu;
|
|
+ struct vgic_lr vlr = vgic_get_lr(vcpu, lr_nr);
|
|
+
|
|
+ vlr.state = 0;
|
|
+ vgic_set_lr(vcpu, lr_nr, vlr);
|
|
+ clear_bit(lr_nr, vgic_cpu->lr_used);
|
|
+ vgic_cpu->vgic_irq_lr_map[irq] = LR_EMPTY;
|
|
+}
|
|
|
|
/*
|
|
* An interrupt may have been disabled after being made pending on the
|
|
@@ -1006,13 +1054,13 @@ static void vgic_retire_disabled_irqs(struct kvm_vcpu *vcpu)
|
|
struct vgic_cpu *vgic_cpu = &vcpu->arch.vgic_cpu;
|
|
int lr;
|
|
|
|
- for_each_set_bit(lr, vgic_cpu->lr_used, vgic_cpu->nr_lr) {
|
|
- int irq = vgic_cpu->vgic_lr[lr] & GICH_LR_VIRTUALID;
|
|
+ for_each_set_bit(lr, vgic_cpu->lr_used, vgic->nr_lr) {
|
|
+ struct vgic_lr vlr = vgic_get_lr(vcpu, lr);
|
|
|
|
- if (!vgic_irq_is_enabled(vcpu, irq)) {
|
|
- vgic_retire_lr(lr, irq, vgic_cpu);
|
|
- if (vgic_irq_is_active(vcpu, irq))
|
|
- vgic_irq_clear_active(vcpu, irq);
|
|
+ if (!vgic_irq_is_enabled(vcpu, vlr.irq)) {
|
|
+ vgic_retire_lr(lr, vlr.irq, vcpu);
|
|
+ if (vgic_irq_is_active(vcpu, vlr.irq))
|
|
+ vgic_irq_clear_active(vcpu, vlr.irq);
|
|
}
|
|
}
|
|
}
|
|
@@ -1024,6 +1072,7 @@ static void vgic_retire_disabled_irqs(struct kvm_vcpu *vcpu)
|
|
static bool vgic_queue_irq(struct kvm_vcpu *vcpu, u8 sgi_source_id, int irq)
|
|
{
|
|
struct vgic_cpu *vgic_cpu = &vcpu->arch.vgic_cpu;
|
|
+ struct vgic_lr vlr;
|
|
int lr;
|
|
|
|
/* Sanitize the input... */
|
|
@@ -1036,28 +1085,34 @@ static bool vgic_queue_irq(struct kvm_vcpu *vcpu, u8 sgi_source_id, int irq)
|
|
lr = vgic_cpu->vgic_irq_lr_map[irq];
|
|
|
|
/* Do we have an active interrupt for the same CPUID? */
|
|
- if (lr != LR_EMPTY &&
|
|
- (LR_CPUID(vgic_cpu->vgic_lr[lr]) == sgi_source_id)) {
|
|
- kvm_debug("LR%d piggyback for IRQ%d %x\n",
|
|
- lr, irq, vgic_cpu->vgic_lr[lr]);
|
|
- BUG_ON(!test_bit(lr, vgic_cpu->lr_used));
|
|
- vgic_cpu->vgic_lr[lr] |= GICH_LR_PENDING_BIT;
|
|
- return true;
|
|
+ if (lr != LR_EMPTY) {
|
|
+ vlr = vgic_get_lr(vcpu, lr);
|
|
+ if (vlr.source == sgi_source_id) {
|
|
+ kvm_debug("LR%d piggyback for IRQ%d\n", lr, vlr.irq);
|
|
+ BUG_ON(!test_bit(lr, vgic_cpu->lr_used));
|
|
+ vlr.state |= LR_STATE_PENDING;
|
|
+ vgic_set_lr(vcpu, lr, vlr);
|
|
+ return true;
|
|
+ }
|
|
}
|
|
|
|
/* Try to use another LR for this interrupt */
|
|
lr = find_first_zero_bit((unsigned long *)vgic_cpu->lr_used,
|
|
- vgic_cpu->nr_lr);
|
|
- if (lr >= vgic_cpu->nr_lr)
|
|
+ vgic->nr_lr);
|
|
+ if (lr >= vgic->nr_lr)
|
|
return false;
|
|
|
|
kvm_debug("LR%d allocated for IRQ%d %x\n", lr, irq, sgi_source_id);
|
|
- vgic_cpu->vgic_lr[lr] = MK_LR_PEND(sgi_source_id, irq);
|
|
vgic_cpu->vgic_irq_lr_map[irq] = lr;
|
|
set_bit(lr, vgic_cpu->lr_used);
|
|
|
|
+ vlr.irq = irq;
|
|
+ vlr.source = sgi_source_id;
|
|
+ vlr.state = LR_STATE_PENDING;
|
|
if (!vgic_irq_is_edge(vcpu, irq))
|
|
- vgic_cpu->vgic_lr[lr] |= GICH_LR_EOI;
|
|
+ vlr.state |= LR_EOI_INT;
|
|
+
|
|
+ vgic_set_lr(vcpu, lr, vlr);
|
|
|
|
return true;
|
|
}
|
|
@@ -1155,9 +1210,9 @@ static void __kvm_vgic_flush_hwstate(struct kvm_vcpu *vcpu)
|
|
|
|
epilog:
|
|
if (overflow) {
|
|
- vgic_cpu->vgic_hcr |= GICH_HCR_UIE;
|
|
+ vgic_enable_underflow(vcpu);
|
|
} else {
|
|
- vgic_cpu->vgic_hcr &= ~GICH_HCR_UIE;
|
|
+ vgic_disable_underflow(vcpu);
|
|
/*
|
|
* We're about to run this VCPU, and we've consumed
|
|
* everything the distributor had in store for
|
|
@@ -1170,44 +1225,46 @@ epilog:
|
|
|
|
static bool vgic_process_maintenance(struct kvm_vcpu *vcpu)
|
|
{
|
|
- struct vgic_cpu *vgic_cpu = &vcpu->arch.vgic_cpu;
|
|
+ u32 status = vgic_get_interrupt_status(vcpu);
|
|
bool level_pending = false;
|
|
|
|
- kvm_debug("MISR = %08x\n", vgic_cpu->vgic_misr);
|
|
+ kvm_debug("STATUS = %08x\n", status);
|
|
|
|
- if (vgic_cpu->vgic_misr & GICH_MISR_EOI) {
|
|
+ if (status & INT_STATUS_EOI) {
|
|
/*
|
|
* Some level interrupts have been EOIed. Clear their
|
|
* active bit.
|
|
*/
|
|
- int lr, irq;
|
|
+ u64 eisr = vgic_get_eisr(vcpu);
|
|
+ unsigned long *eisr_ptr = (unsigned long *)&eisr;
|
|
+ int lr;
|
|
|
|
- for_each_set_bit(lr, (unsigned long *)vgic_cpu->vgic_eisr,
|
|
- vgic_cpu->nr_lr) {
|
|
- irq = vgic_cpu->vgic_lr[lr] & GICH_LR_VIRTUALID;
|
|
+ for_each_set_bit(lr, eisr_ptr, vgic->nr_lr) {
|
|
+ struct vgic_lr vlr = vgic_get_lr(vcpu, lr);
|
|
|
|
- vgic_irq_clear_active(vcpu, irq);
|
|
- vgic_cpu->vgic_lr[lr] &= ~GICH_LR_EOI;
|
|
+ vgic_irq_clear_active(vcpu, vlr.irq);
|
|
+ WARN_ON(vlr.state & LR_STATE_MASK);
|
|
+ vlr.state = 0;
|
|
+ vgic_set_lr(vcpu, lr, vlr);
|
|
|
|
/* Any additional pending interrupt? */
|
|
- if (vgic_dist_irq_is_pending(vcpu, irq)) {
|
|
- vgic_cpu_irq_set(vcpu, irq);
|
|
+ if (vgic_dist_irq_is_pending(vcpu, vlr.irq)) {
|
|
+ vgic_cpu_irq_set(vcpu, vlr.irq);
|
|
level_pending = true;
|
|
} else {
|
|
- vgic_cpu_irq_clear(vcpu, irq);
|
|
+ vgic_cpu_irq_clear(vcpu, vlr.irq);
|
|
}
|
|
|
|
/*
|
|
* Despite being EOIed, the LR may not have
|
|
* been marked as empty.
|
|
*/
|
|
- set_bit(lr, (unsigned long *)vgic_cpu->vgic_elrsr);
|
|
- vgic_cpu->vgic_lr[lr] &= ~GICH_LR_ACTIVE_BIT;
|
|
+ vgic_sync_lr_elrsr(vcpu, lr, vlr);
|
|
}
|
|
}
|
|
|
|
- if (vgic_cpu->vgic_misr & GICH_MISR_U)
|
|
- vgic_cpu->vgic_hcr &= ~GICH_HCR_UIE;
|
|
+ if (status & INT_STATUS_UNDERFLOW)
|
|
+ vgic_disable_underflow(vcpu);
|
|
|
|
return level_pending;
|
|
}
|
|
@@ -1220,29 +1277,31 @@ static void __kvm_vgic_sync_hwstate(struct kvm_vcpu *vcpu)
|
|
{
|
|
struct vgic_cpu *vgic_cpu = &vcpu->arch.vgic_cpu;
|
|
struct vgic_dist *dist = &vcpu->kvm->arch.vgic;
|
|
+ u64 elrsr;
|
|
+ unsigned long *elrsr_ptr;
|
|
int lr, pending;
|
|
bool level_pending;
|
|
|
|
level_pending = vgic_process_maintenance(vcpu);
|
|
+ elrsr = vgic_get_elrsr(vcpu);
|
|
+ elrsr_ptr = (unsigned long *)&elrsr;
|
|
|
|
/* Clear mappings for empty LRs */
|
|
- for_each_set_bit(lr, (unsigned long *)vgic_cpu->vgic_elrsr,
|
|
- vgic_cpu->nr_lr) {
|
|
- int irq;
|
|
+ for_each_set_bit(lr, elrsr_ptr, vgic->nr_lr) {
|
|
+ struct vgic_lr vlr;
|
|
|
|
if (!test_and_clear_bit(lr, vgic_cpu->lr_used))
|
|
continue;
|
|
|
|
- irq = vgic_cpu->vgic_lr[lr] & GICH_LR_VIRTUALID;
|
|
+ vlr = vgic_get_lr(vcpu, lr);
|
|
|
|
- BUG_ON(irq >= VGIC_NR_IRQS);
|
|
- vgic_cpu->vgic_irq_lr_map[irq] = LR_EMPTY;
|
|
+ BUG_ON(vlr.irq >= VGIC_NR_IRQS);
|
|
+ vgic_cpu->vgic_irq_lr_map[vlr.irq] = LR_EMPTY;
|
|
}
|
|
|
|
/* Check if we still have something up our sleeve... */
|
|
- pending = find_first_zero_bit((unsigned long *)vgic_cpu->vgic_elrsr,
|
|
- vgic_cpu->nr_lr);
|
|
- if (level_pending || pending < vgic_cpu->nr_lr)
|
|
+ pending = find_first_zero_bit(elrsr_ptr, vgic->nr_lr);
|
|
+ if (level_pending || pending < vgic->nr_lr)
|
|
set_bit(vcpu->vcpu_id, &dist->irq_pending_on_cpu);
|
|
}
|
|
|
|
@@ -1432,21 +1491,20 @@ int kvm_vgic_vcpu_init(struct kvm_vcpu *vcpu)
|
|
}
|
|
|
|
/*
|
|
- * By forcing VMCR to zero, the GIC will restore the binary
|
|
- * points to their reset values. Anything else resets to zero
|
|
- * anyway.
|
|
+ * Store the number of LRs per vcpu, so we don't have to go
|
|
+ * all the way to the distributor structure to find out. Only
|
|
+ * assembly code should use this one.
|
|
*/
|
|
- vgic_cpu->vgic_vmcr = 0;
|
|
+ vgic_cpu->nr_lr = vgic->nr_lr;
|
|
|
|
- vgic_cpu->nr_lr = vgic_nr_lr;
|
|
- vgic_cpu->vgic_hcr = GICH_HCR_EN; /* Get the show on the road... */
|
|
+ vgic_enable(vcpu);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void vgic_init_maintenance_interrupt(void *info)
|
|
{
|
|
- enable_percpu_irq(vgic_maint_irq, 0);
|
|
+ enable_percpu_irq(vgic->maint_irq, 0);
|
|
}
|
|
|
|
static int vgic_cpu_notify(struct notifier_block *self,
|
|
@@ -1459,7 +1517,7 @@ static int vgic_cpu_notify(struct notifier_block *self,
|
|
break;
|
|
case CPU_DYING:
|
|
case CPU_DYING_FROZEN:
|
|
- disable_percpu_irq(vgic_maint_irq);
|
|
+ disable_percpu_irq(vgic->maint_irq);
|
|
break;
|
|
}
|
|
|
|
@@ -1470,30 +1528,37 @@ static struct notifier_block vgic_cpu_nb = {
|
|
.notifier_call = vgic_cpu_notify,
|
|
};
|
|
|
|
+static const struct of_device_id vgic_ids[] = {
|
|
+ { .compatible = "arm,cortex-a15-gic", .data = vgic_v2_probe, },
|
|
+ { .compatible = "arm,gic-v3", .data = vgic_v3_probe, },
|
|
+ {},
|
|
+};
|
|
+
|
|
int kvm_vgic_hyp_init(void)
|
|
{
|
|
+ const struct of_device_id *matched_id;
|
|
+ int (*vgic_probe)(struct device_node *,const struct vgic_ops **,
|
|
+ const struct vgic_params **);
|
|
+ struct device_node *vgic_node;
|
|
int ret;
|
|
- struct resource vctrl_res;
|
|
- struct resource vcpu_res;
|
|
|
|
- vgic_node = of_find_compatible_node(NULL, NULL, "arm,cortex-a15-gic");
|
|
+ vgic_node = of_find_matching_node_and_match(NULL,
|
|
+ vgic_ids, &matched_id);
|
|
if (!vgic_node) {
|
|
- kvm_err("error: no compatible vgic node in DT\n");
|
|
+ kvm_err("error: no compatible GIC node found\n");
|
|
return -ENODEV;
|
|
}
|
|
|
|
- vgic_maint_irq = irq_of_parse_and_map(vgic_node, 0);
|
|
- if (!vgic_maint_irq) {
|
|
- kvm_err("error getting vgic maintenance irq from DT\n");
|
|
- ret = -ENXIO;
|
|
- goto out;
|
|
- }
|
|
+ vgic_probe = matched_id->data;
|
|
+ ret = vgic_probe(vgic_node, &vgic_ops, &vgic);
|
|
+ if (ret)
|
|
+ return ret;
|
|
|
|
- ret = request_percpu_irq(vgic_maint_irq, vgic_maintenance_handler,
|
|
+ ret = request_percpu_irq(vgic->maint_irq, vgic_maintenance_handler,
|
|
"vgic", kvm_get_running_vcpus());
|
|
if (ret) {
|
|
- kvm_err("Cannot register interrupt %d\n", vgic_maint_irq);
|
|
- goto out;
|
|
+ kvm_err("Cannot register interrupt %d\n", vgic->maint_irq);
|
|
+ return ret;
|
|
}
|
|
|
|
ret = __register_cpu_notifier(&vgic_cpu_nb);
|
|
@@ -1502,65 +1567,15 @@ int kvm_vgic_hyp_init(void)
|
|
goto out_free_irq;
|
|
}
|
|
|
|
- ret = of_address_to_resource(vgic_node, 2, &vctrl_res);
|
|
- if (ret) {
|
|
- kvm_err("Cannot obtain VCTRL resource\n");
|
|
- goto out_free_irq;
|
|
- }
|
|
-
|
|
- vgic_vctrl_base = of_iomap(vgic_node, 2);
|
|
- if (!vgic_vctrl_base) {
|
|
- kvm_err("Cannot ioremap VCTRL\n");
|
|
- ret = -ENOMEM;
|
|
- goto out_free_irq;
|
|
- }
|
|
-
|
|
- vgic_nr_lr = readl_relaxed(vgic_vctrl_base + GICH_VTR);
|
|
- vgic_nr_lr = (vgic_nr_lr & 0x3f) + 1;
|
|
-
|
|
- ret = create_hyp_io_mappings(vgic_vctrl_base,
|
|
- vgic_vctrl_base + resource_size(&vctrl_res),
|
|
- vctrl_res.start);
|
|
- if (ret) {
|
|
- kvm_err("Cannot map VCTRL into hyp\n");
|
|
- goto out_unmap;
|
|
- }
|
|
-
|
|
- if (of_address_to_resource(vgic_node, 3, &vcpu_res)) {
|
|
- kvm_err("Cannot obtain VCPU resource\n");
|
|
- ret = -ENXIO;
|
|
- goto out_unmap;
|
|
- }
|
|
-
|
|
- if (!PAGE_ALIGNED(vcpu_res.start)) {
|
|
- kvm_err("GICV physical address 0x%llx not page aligned\n",
|
|
- (unsigned long long)vcpu_res.start);
|
|
- ret = -ENXIO;
|
|
- goto out_unmap;
|
|
- }
|
|
-
|
|
- if (!PAGE_ALIGNED(resource_size(&vcpu_res))) {
|
|
- kvm_err("GICV size 0x%llx not a multiple of page size 0x%lx\n",
|
|
- (unsigned long long)resource_size(&vcpu_res),
|
|
- PAGE_SIZE);
|
|
- ret = -ENXIO;
|
|
- goto out_unmap;
|
|
- }
|
|
-
|
|
- vgic_vcpu_base = vcpu_res.start;
|
|
-
|
|
- kvm_info("%s@%llx IRQ%d\n", vgic_node->name,
|
|
- vctrl_res.start, vgic_maint_irq);
|
|
on_each_cpu(vgic_init_maintenance_interrupt, NULL, 1);
|
|
|
|
- goto out;
|
|
+ /* Callback into for arch code for setup */
|
|
+ vgic_arch_setup(vgic);
|
|
+
|
|
+ return 0;
|
|
|
|
-out_unmap:
|
|
- iounmap(vgic_vctrl_base);
|
|
out_free_irq:
|
|
- free_percpu_irq(vgic_maint_irq, kvm_get_running_vcpus());
|
|
-out:
|
|
- of_node_put(vgic_node);
|
|
+ free_percpu_irq(vgic->maint_irq, kvm_get_running_vcpus());
|
|
return ret;
|
|
}
|
|
|
|
@@ -1593,7 +1608,7 @@ int kvm_vgic_init(struct kvm *kvm)
|
|
}
|
|
|
|
ret = kvm_phys_addr_ioremap(kvm, kvm->arch.vgic.vgic_cpu_base,
|
|
- vgic_vcpu_base, KVM_VGIC_V2_CPU_SIZE);
|
|
+ vgic->vcpu_base, KVM_VGIC_V2_CPU_SIZE);
|
|
if (ret) {
|
|
kvm_err("Unable to remap VGIC CPU to VCPU\n");
|
|
goto out;
|
|
@@ -1639,7 +1654,8 @@ int kvm_vgic_create(struct kvm *kvm)
|
|
}
|
|
|
|
spin_lock_init(&kvm->arch.vgic.lock);
|
|
- kvm->arch.vgic.vctrl_base = vgic_vctrl_base;
|
|
+ kvm->arch.vgic.in_kernel = true;
|
|
+ kvm->arch.vgic.vctrl_base = vgic->vctrl_base;
|
|
kvm->arch.vgic.vgic_dist_base = VGIC_ADDR_UNDEF;
|
|
kvm->arch.vgic.vgic_cpu_base = VGIC_ADDR_UNDEF;
|
|
|
|
@@ -1738,39 +1754,40 @@ int kvm_vgic_addr(struct kvm *kvm, unsigned long type, u64 *addr, bool write)
|
|
static bool handle_cpu_mmio_misc(struct kvm_vcpu *vcpu,
|
|
struct kvm_exit_mmio *mmio, phys_addr_t offset)
|
|
{
|
|
- struct vgic_cpu *vgic_cpu = &vcpu->arch.vgic_cpu;
|
|
- u32 reg, mask = 0, shift = 0;
|
|
bool updated = false;
|
|
+ struct vgic_vmcr vmcr;
|
|
+ u32 *vmcr_field;
|
|
+ u32 reg;
|
|
+
|
|
+ vgic_get_vmcr(vcpu, &vmcr);
|
|
|
|
switch (offset & ~0x3) {
|
|
case GIC_CPU_CTRL:
|
|
- mask = GICH_VMCR_CTRL_MASK;
|
|
- shift = GICH_VMCR_CTRL_SHIFT;
|
|
+ vmcr_field = &vmcr.ctlr;
|
|
break;
|
|
case GIC_CPU_PRIMASK:
|
|
- mask = GICH_VMCR_PRIMASK_MASK;
|
|
- shift = GICH_VMCR_PRIMASK_SHIFT;
|
|
+ vmcr_field = &vmcr.pmr;
|
|
break;
|
|
case GIC_CPU_BINPOINT:
|
|
- mask = GICH_VMCR_BINPOINT_MASK;
|
|
- shift = GICH_VMCR_BINPOINT_SHIFT;
|
|
+ vmcr_field = &vmcr.bpr;
|
|
break;
|
|
case GIC_CPU_ALIAS_BINPOINT:
|
|
- mask = GICH_VMCR_ALIAS_BINPOINT_MASK;
|
|
- shift = GICH_VMCR_ALIAS_BINPOINT_SHIFT;
|
|
+ vmcr_field = &vmcr.abpr;
|
|
break;
|
|
+ default:
|
|
+ BUG();
|
|
}
|
|
|
|
if (!mmio->is_write) {
|
|
- reg = (vgic_cpu->vgic_vmcr & mask) >> shift;
|
|
+ reg = *vmcr_field;
|
|
mmio_data_write(mmio, ~0, reg);
|
|
} else {
|
|
reg = mmio_data_read(mmio, ~0);
|
|
- reg = (reg << shift) & mask;
|
|
- if (reg != (vgic_cpu->vgic_vmcr & mask))
|
|
+ if (reg != *vmcr_field) {
|
|
+ *vmcr_field = reg;
|
|
+ vgic_set_vmcr(vcpu, &vmcr);
|
|
updated = true;
|
|
- vgic_cpu->vgic_vmcr &= ~mask;
|
|
- vgic_cpu->vgic_vmcr |= reg;
|
|
+ }
|
|
}
|
|
return updated;
|
|
}
|