10137 lines
292 KiB
Diff
10137 lines
292 KiB
Diff
commit 649c9e3a45e81852daf80c126a332297b75cb109
|
|
Author: Ard Biesheuvel <ard.biesheuvel@linaro.org>
|
|
Date: Tue Jul 29 12:49:10 2014 +0200
|
|
|
|
arm64/efi: efistub: don't abort if base of DRAM is occupied
|
|
|
|
If we cannot relocate the kernel Image to its preferred offset of base of DRAM
|
|
plus TEXT_OFFSET, instead relocate it to the lowest available 2 MB boundary plus
|
|
TEXT_OFFSET. We may lose a bit of memory at the low end, but we can still
|
|
proceed normally otherwise.
|
|
|
|
Signed-off-by: Ard Biesheuvel <ard.biesheuvel@linaro.org>
|
|
|
|
commit 5102fd06b12467a0518537061805483a759bc856
|
|
Author: Ard Biesheuvel <ard.biesheuvel@linaro.org>
|
|
Date: Tue Jul 29 12:49:09 2014 +0200
|
|
|
|
arm64/efi: efistub: cover entire static mem footprint in PE/COFF .text
|
|
|
|
The static memory footprint of a kernel Image at boot is larger than the
|
|
Image file itself. Things like .bss data and initial page tables are allocated
|
|
statically but populated dynamically so their content is not contained in the
|
|
Image file.
|
|
|
|
However, if EFI (or GRUB) has loaded the Image at precisely the desired offset
|
|
of base of DRAM + TEXT_OFFSET, the Image will be booted in place, and we have
|
|
to make sure that the allocation done by the PE/COFF loader is large enough.
|
|
|
|
Fix this by growing the PE/COFF .text section to cover the entire static
|
|
memory footprint. The part of the section that is not covered by the payload
|
|
will be zero initialised by the PE/COFF loader.
|
|
|
|
Signed-off-by: Ard Biesheuvel <ard.biesheuvel@linaro.org>
|
|
|
|
commit 3b4dfb00a401b7fecf01d3c89b154907167dff52
|
|
Author: Mark Rutland <mark.rutland@arm.com>
|
|
Date: Tue Jul 29 12:49:08 2014 +0200
|
|
|
|
arm64: spin-table: handle unmapped cpu-release-addrs
|
|
|
|
In certain cases the cpu-release-addr of a CPU may not fall in the
|
|
linear mapping (e.g. when the kernel is loaded above this address due to
|
|
the presence of other images in memory). This is problematic for the
|
|
spin-table code as it assumes that it can trivially convert a
|
|
cpu-release-addr to a valid VA in the linear map.
|
|
|
|
This patch modifies the spin-table code to use a temporary cached
|
|
mapping to write to a given cpu-release-addr, enabling us to support
|
|
addresses regardless of whether they are covered by the linear mapping.
|
|
|
|
Signed-off-by: Mark Rutland <mark.rutland@arm.com>
|
|
|
|
commit a49ad891b2e91338587dc5576c9da73b249a9d13
|
|
Author: Mark Salter <msalter@redhat.com>
|
|
Date: Mon Jul 14 15:52:06 2014 -0400
|
|
|
|
pci/xgene: use pci_remap_iospace() instead of pci_ioremap_io()
|
|
|
|
Signed-off-by: Mark Salter <msalter@redhat.com>
|
|
|
|
commit 1a958ec71db226d35ed51756b7164142bb0a60a3
|
|
Author: Tanmay Inamdar <tinamdar@apm.com>
|
|
Date: Wed Mar 19 16:12:42 2014 -0700
|
|
|
|
MAINTAINERS: entry for APM X-Gene PCIe host driver
|
|
|
|
Add entry for AppliedMicro X-Gene PCIe host driver.
|
|
|
|
Signed-off-by: Tanmay Inamdar <tinamdar@apm.com>
|
|
|
|
commit 03dc92f35b8f8be898ca12e3dc7c15961f414907
|
|
Author: Tanmay Inamdar <tinamdar@apm.com>
|
|
Date: Wed Mar 19 16:12:41 2014 -0700
|
|
|
|
dt-bindings: pci: xgene pcie device tree bindings
|
|
|
|
This patch adds the bindings for X-Gene PCIe driver. The driver resides
|
|
under 'drivers/pci/host/pci-xgene.c' file.
|
|
|
|
Signed-off-by: Tanmay Inamdar <tinamdar@apm.com>
|
|
|
|
commit 1f8d894f2066d9db2b251d512f6f6f772ae7147f
|
|
Author: Tanmay Inamdar <tinamdar@apm.com>
|
|
Date: Wed Mar 19 16:12:40 2014 -0700
|
|
|
|
arm64: dts: APM X-Gene PCIe device tree nodes
|
|
|
|
This patch adds the device tree nodes for APM X-Gene PCIe controller and
|
|
PCIe clock interface. Since X-Gene SOC supports maximum 5 ports, 5 dts
|
|
nodes are added.
|
|
|
|
Signed-off-by: Tanmay Inamdar <tinamdar@apm.com>
|
|
|
|
commit c0855fcf0cc9adcb1ba5e6b1318536c56244796d
|
|
Author: Tanmay Inamdar <tinamdar@apm.com>
|
|
Date: Wed Mar 19 16:12:39 2014 -0700
|
|
|
|
pci: APM X-Gene PCIe controller driver
|
|
|
|
This patch adds the AppliedMicro X-Gene SOC PCIe controller driver.
|
|
X-Gene PCIe controller supports maximum up to 8 lanes and GEN3 speed.
|
|
X-Gene SOC supports maximum 5 PCIe ports.
|
|
|
|
Signed-off-by: Tanmay Inamdar <tinamdar@apm.com>
|
|
|
|
commit 1cc0c322237d3b58b08fe39e79e6c2e2f90a8c98
|
|
Author: Liviu Dudau <Liviu.Dudau@arm.com>
|
|
Date: Tue Jul 1 19:44:00 2014 +0100
|
|
|
|
arm64: Add architectural support for PCI
|
|
|
|
Use the generic host bridge functions to provide support for
|
|
PCI Express on arm64.
|
|
|
|
Signed-off-by: Liviu Dudau <Liviu.Dudau@arm.com>
|
|
|
|
commit aba1eca911a87959eb4be515110f7a6b8692e9a4
|
|
Author: Liviu Dudau <Liviu.Dudau@arm.com>
|
|
Date: Tue Jul 1 19:43:34 2014 +0100
|
|
|
|
pci: Remap I/O bus resources into CPU space with pci_remap_iospace()
|
|
|
|
Introduce a default implementation for remapping PCI bus I/O resources
|
|
onto the CPU address space. Architectures with special needs may
|
|
provide their own version, but most should be able to use this one.
|
|
|
|
Signed-off-by: Liviu Dudau <Liviu.Dudau@arm.com>
|
|
|
|
commit e1eacc3da10ca19eff1f88fb342a13586092b613
|
|
Author: Liviu Dudau <liviu@dudau.co.uk>
|
|
Date: Tue Jul 1 21:50:50 2014 +0100
|
|
|
|
pci: Add support for creating a generic host_bridge from device tree
|
|
|
|
Several platforms use a rather generic version of parsing
|
|
the device tree to find the host bridge ranges. Move the common code
|
|
into the generic PCI code and use it to create a pci_host_bridge
|
|
structure that can be used by arch code.
|
|
|
|
Based on early attempts by Andrew Murray to unify the code.
|
|
Used powerpc and microblaze PCI code as starting point.
|
|
|
|
Signed-off-by: Liviu Dudau <Liviu.Dudau@arm.com>
|
|
Tested-by: Tanmay Inamdar <tinamdar@apm.com>
|
|
|
|
commit aee55d507eb451223b51b52e6617b06b8e518ea6
|
|
Author: Liviu Dudau <Liviu.Dudau@arm.com>
|
|
Date: Tue Jul 1 19:43:32 2014 +0100
|
|
|
|
pci: of: Parse and map the IRQ when adding the PCI device.
|
|
|
|
Enhance the default implementation of pcibios_add_device() to
|
|
parse and map the IRQ of the device if a DT binding is available.
|
|
|
|
Signed-off-by: Liviu Dudau <Liviu.Dudau@arm.com>
|
|
|
|
commit 0cf0470962f0498b598ff44e0c671407df54b22e
|
|
Author: Liviu Dudau <Liviu.Dudau@arm.com>
|
|
Date: Tue Jul 1 19:43:31 2014 +0100
|
|
|
|
pci: Introduce a domain number for pci_host_bridge.
|
|
|
|
Make it easier to discover the domain number of a bus by storing
|
|
the number in pci_host_bridge for the root bus. Several architectures
|
|
have their own way of storing this information, so it makes sense
|
|
to try to unify the code. While at this, add a new function that
|
|
creates a root bus in a given domain and make pci_create_root_bus()
|
|
a wrapper around this function.
|
|
|
|
Signed-off-by: Liviu Dudau <Liviu.Dudau@arm.com>
|
|
Tested-by: Tanmay Inamdar <tinamdar@apm.com>
|
|
|
|
commit 1bc8fcf01469c202b4aea5f1d0a3a75c9302f3e2
|
|
Author: Liviu Dudau <Liviu.Dudau@arm.com>
|
|
Date: Tue Jul 1 19:43:30 2014 +0100
|
|
|
|
pci: Create pci_host_bridge before its associated bus in pci_create_root_bus.
|
|
|
|
Before commit 7b5436635800 the pci_host_bridge was created before the root bus.
|
|
As that commit has added a needless dependency on the bus for pci_alloc_host_bridge()
|
|
the creation order has been changed for no good reason. Revert the order of
|
|
creation as we are going to depend on the pci_host_bridge structure to retrieve the
|
|
domain number of the root bus.
|
|
|
|
Signed-off-by: Liviu Dudau <Liviu.Dudau@arm.com>
|
|
Acked-by: Grant Likely <grant.likely@linaro.org>
|
|
Tested-by: Tanmay Inamdar <tinamdar@apm.com>
|
|
|
|
commit 55353327169dc08c1047d994e00b8a38630f5a8d
|
|
Author: Liviu Dudau <Liviu.Dudau@arm.com>
|
|
Date: Tue Jul 1 19:43:29 2014 +0100
|
|
|
|
pci: OF: Fix the conversion of IO ranges into IO resources.
|
|
|
|
The ranges property for a host bridge controller in DT describes
|
|
the mapping between the PCI bus address and the CPU physical address.
|
|
The resources framework however expects that the IO resources start
|
|
at a pseudo "port" address 0 (zero) and have a maximum size of IO_SPACE_LIMIT.
|
|
The conversion from pci ranges to resources failed to take that into account.
|
|
|
|
In the process move the function into drivers/of/address.c as it now
|
|
depends on pci_address_to_pio() code and make it return an error message.
|
|
|
|
Signed-off-by: Liviu Dudau <Liviu.Dudau@arm.com>
|
|
Tested-by: Tanmay Inamdar <tinamdar@apm.com>
|
|
|
|
commit 34079a20796d5c74e2984c37a99baef4871709a6
|
|
Author: Liviu Dudau <Liviu.Dudau@arm.com>
|
|
Date: Tue Jul 1 19:43:28 2014 +0100
|
|
|
|
pci: Introduce pci_register_io_range() helper function.
|
|
|
|
Some architectures do not have a simple view of the PCI I/O space
|
|
and instead use a range of CPU addresses that map to bus addresses. For
|
|
some architectures these ranges will be expressed by OF bindings
|
|
in a device tree file.
|
|
|
|
Introduce a pci_register_io_range() helper function with a generic
|
|
implementation that can be used by such architectures to keep track
|
|
of the I/O ranges described by the PCI bindings. If the PCI_IOBASE
|
|
macro is not defined that signals lack of support for PCI and we
|
|
return an error.
|
|
|
|
Signed-off-by: Liviu Dudau <Liviu.Dudau@arm.com>
|
|
|
|
commit a81abc095ab4b9b90e446ddbd59247e23df9d4ad
|
|
Author: Liviu Dudau <Liviu.Dudau@arm.com>
|
|
Date: Tue Jul 1 19:43:27 2014 +0100
|
|
|
|
pci: Export find_pci_host_bridge() function.
|
|
|
|
This is a useful function and we should make it visible outside the
|
|
generic PCI code. Export it as a GPL symbol.
|
|
|
|
Signed-off-by: Liviu Dudau <Liviu.Dudau@arm.com>
|
|
Tested-by: Tanmay Inamdar <tinamdar@apm.com>
|
|
|
|
commit 78361698444e81bedbf30ec2b7aae1afd110d11f
|
|
Author: Liviu Dudau <Liviu.Dudau@arm.com>
|
|
Date: Tue Jul 1 19:43:26 2014 +0100
|
|
|
|
Fix ioport_map() for !CONFIG_GENERIC_IOMAP cases.
|
|
|
|
The inline version of ioport_map() that gets used when !CONFIG_GENERIC_IOMAP
|
|
is wrong. It returns a mapped (i.e. virtual) address that can start from
|
|
zero and completely ignores the PCI_IOBASE and IO_SPACE_LIMIT that most
|
|
architectures that use !CONFIG_GENERIC_MAP define.
|
|
|
|
Signed-off-by: Liviu Dudau <Liviu.Dudau@arm.com>
|
|
Acked-by: Arnd Bergmann <arnd@arndb.de>
|
|
Tested-by: Tanmay Inamdar <tinamdar@apm.com>
|
|
|
|
commit 5540bbe2eca72e37be2e6a1c18e2fc7e73f0eab6
|
|
Author: Marc Zyngier <marc.zyngier@arm.com>
|
|
Date: Mon Jun 30 16:01:50 2014 +0100
|
|
|
|
arm64: KVM: vgic: enable GICv2 emulation on top on GICv3 hardware
|
|
|
|
Add the last missing bits that enable GICv2 emulation on top of
|
|
GICv3 hardware.
|
|
|
|
Signed-off-by: Marc Zyngier <marc.zyngier@arm.com>
|
|
|
|
commit f057aaf81a7df641bcaa992965a257e3260ad36e
|
|
Author: Marc Zyngier <marc.zyngier@arm.com>
|
|
Date: Mon Jun 30 16:01:49 2014 +0100
|
|
|
|
arm64: KVM: vgic: add GICv3 world switch
|
|
|
|
Introduce the GICv3 world switch code used to save/restore the
|
|
GICv3 context.
|
|
|
|
Acked-by: Catalin Marinas <catalin.marinas@arm.com>
|
|
Reviewed-by: Christoffer Dall <christoffer.dall@linaro.org>
|
|
Signed-off-by: Marc Zyngier <marc.zyngier@arm.com>
|
|
|
|
commit 7f1b8a791bb375933fdc8420cd08f127d07dd259
|
|
Author: Marc Zyngier <marc.zyngier@arm.com>
|
|
Date: Mon Jun 30 16:01:48 2014 +0100
|
|
|
|
KVM: ARM: vgic: add the GICv3 backend
|
|
|
|
Introduce the support code for emulating a GICv2 on top of GICv3
|
|
hardware.
|
|
|
|
Acked-by: Catalin Marinas <catalin.marinas@arm.com>
|
|
Signed-off-by: Marc Zyngier <marc.zyngier@arm.com>
|
|
|
|
commit 74428905c3b450eab53a21bee74236501629a443
|
|
Author: Marc Zyngier <marc.zyngier@arm.com>
|
|
Date: Mon Jun 30 16:01:47 2014 +0100
|
|
|
|
arm64: KVM: move HCR_EL2.{IMO, FMO} manipulation into the vgic switch code
|
|
|
|
GICv3 requires the IMO and FMO bits to be tightly coupled with some
|
|
of the interrupt controller's register switch.
|
|
|
|
In order to have similar code paths, move the manipulation of these
|
|
bits to the GICv2 switch code.
|
|
|
|
Acked-by: Catalin Marinas <catalin.marinas@arm.com>
|
|
Reviewed-by: Christoffer Dall <christoffer.dall@linaro.org>
|
|
Signed-off-by: Marc Zyngier <marc.zyngier@arm.com>
|
|
|
|
commit b691c1f97f1fb5b29c3ae4cc836fdbefe61a11ff
|
|
Author: Marc Zyngier <marc.zyngier@arm.com>
|
|
Date: Mon Jun 30 16:01:46 2014 +0100
|
|
|
|
arm64: KVM: split GICv2 world switch from hyp code
|
|
|
|
Move the GICv2 world switch code into its own file, and add the
|
|
necessary indirection to the arm64 switch code.
|
|
|
|
Also introduce a new type field to the vgic_params structure.
|
|
|
|
Acked-by: Catalin Marinas <catalin.marinas@arm.com>
|
|
Reviewed-by: Christoffer Dall <christoffer.dall@linaro.org>
|
|
Signed-off-by: Marc Zyngier <marc.zyngier@arm.com>
|
|
|
|
commit d3f4563f1bebed7f60f714bdab640e477a081c4b
|
|
Author: Marc Zyngier <marc.zyngier@arm.com>
|
|
Date: Mon Jun 30 16:01:45 2014 +0100
|
|
|
|
arm64: KVM: remove __kvm_hyp_code_{start, end} from hyp.S
|
|
|
|
We already have __hyp_text_{start,end} to express the boundaries
|
|
of the HYP text section, and __kvm_hyp_code_{start,end} are getting
|
|
in the way of a more modular world switch code.
|
|
|
|
Just turn __kvm_hyp_code_{start,end} into #defines mapping the
|
|
linker-emited symbols.
|
|
|
|
Acked-by: Catalin Marinas <catalin.marinas@arm.com>
|
|
Reviewed-by: Christoffer Dall <christoffer.dall@linaro.org>
|
|
Signed-off-by: Marc Zyngier <marc.zyngier@arm.com>
|
|
|
|
commit 8ede261e09ad25ab8229d5efb476d5b4f6dc6434
|
|
Author: Marc Zyngier <marc.zyngier@arm.com>
|
|
Date: Mon Jun 30 16:01:44 2014 +0100
|
|
|
|
KVM: ARM: vgic: revisit implementation of irqchip_in_kernel
|
|
|
|
So far, irqchip_in_kernel() was implemented by testing the value of
|
|
vctrl_base, which worked fine with GICv2.
|
|
|
|
With GICv3, this field is useless, as we're using system registers
|
|
instead of a emmory mapped interface. To solve this, add a boolean
|
|
flag indicating if the we're using a vgic or not.
|
|
|
|
Reviewed-by: Christoffer Dall <christoffer.dall@linaro.org>
|
|
Signed-off-by: Marc Zyngier <marc.zyngier@arm.com>
|
|
|
|
commit ff7faf70feb47e5f1cf1e0f0d02e0f2807da11f5
|
|
Author: Marc Zyngier <marc.zyngier@arm.com>
|
|
Date: Mon Jun 30 16:01:43 2014 +0100
|
|
|
|
KVM: ARM: vgic: split GICv2 backend from the main vgic code
|
|
|
|
Brutally hack the innocent vgic code, and move the GICv2 specific code
|
|
to its own file, using vgic_ops and vgic_params as a way to pass
|
|
information between the two blocks.
|
|
|
|
Acked-by: Catalin Marinas <catalin.marinas@arm.com>
|
|
Reviewed-by: Christoffer Dall <christoffer.dall@linaro.org>
|
|
Signed-off-by: Marc Zyngier <marc.zyngier@arm.com>
|
|
|
|
commit 63480283c0e1dc92f506e5e5306be0ac9b239499
|
|
Author: Marc Zyngier <marc.zyngier@arm.com>
|
|
Date: Mon Jun 30 16:01:42 2014 +0100
|
|
|
|
KVM: ARM: introduce vgic_params structure
|
|
|
|
Move all the data specific to a given GIC implementation into its own
|
|
little structure.
|
|
|
|
Acked-by: Catalin Marinas <catalin.marinas@arm.com>
|
|
Reviewed-by: Christoffer Dall <christoffer.dall@linaro.org>
|
|
Signed-off-by: Marc Zyngier <marc.zyngier@arm.com>
|
|
|
|
commit dcb20f9b66d5615a3e7e492424fa7953c1fe9f01
|
|
Author: Marc Zyngier <marc.zyngier@arm.com>
|
|
Date: Mon Jun 30 16:01:41 2014 +0100
|
|
|
|
KVM: ARM: vgic: introduce vgic_enable
|
|
|
|
Move the code dealing with enabling the VGIC on to vgic_ops.
|
|
|
|
Acked-by: Catalin Marinas <catalin.marinas@arm.com>
|
|
Reviewed-by: Christoffer Dall <christoffer.dall@linaro.org>
|
|
Signed-off-by: Marc Zyngier <marc.zyngier@arm.com>
|
|
|
|
commit 67e3bcf8dd823eec79c74bc993fa62fb08b1acea
|
|
Author: Marc Zyngier <marc.zyngier@arm.com>
|
|
Date: Mon Jun 30 16:01:40 2014 +0100
|
|
|
|
KVM: ARM: vgic: abstract VMCR access
|
|
|
|
Instead of directly messing with with the GICH_VMCR bits for the CPU
|
|
interface save/restore code, add accessors that encode/decode the
|
|
entire set of registers exposed by VMCR.
|
|
|
|
Not the most efficient thing, but given that this code is only used
|
|
by the save/restore code, performance is far from being critical.
|
|
|
|
Reviewed-by: Christoffer Dall <christoffer.dall@linaro.org>
|
|
Signed-off-by: Marc Zyngier <marc.zyngier@arm.com>
|
|
|
|
commit 67caf34f138a4f2516a2afcd5657add2eaaf0ab4
|
|
Author: Marc Zyngier <marc.zyngier@arm.com>
|
|
Date: Mon Jun 30 16:01:39 2014 +0100
|
|
|
|
KVM: ARM: vgic: move underflow handling to vgic_ops
|
|
|
|
Move the code dealing with LR underflow handling to its own functions,
|
|
and make them accessible through vgic_ops.
|
|
|
|
Acked-by: Catalin Marinas <catalin.marinas@arm.com>
|
|
Reviewed-by: Christoffer Dall <christoffer.dall@linaro.org>
|
|
Signed-off-by: Marc Zyngier <marc.zyngier@arm.com>
|
|
|
|
commit 30acb0a340d388135ae89bed7e248bad203ec876
|
|
Author: Marc Zyngier <marc.zyngier@arm.com>
|
|
Date: Mon Jun 30 16:01:38 2014 +0100
|
|
|
|
KVM: ARM: vgic: abstract MISR decoding
|
|
|
|
Instead of directly dealing with the GICH_MISR bits, move the code to
|
|
its own function and use a couple of public flags to represent the
|
|
actual state.
|
|
|
|
Acked-by: Catalin Marinas <catalin.marinas@arm.com>
|
|
Reviewed-by: Christoffer Dall <christoffer.dall@linaro.org>
|
|
Signed-off-by: Marc Zyngier <marc.zyngier@arm.com>
|
|
|
|
commit 464cbe47011e07b654d161ab4d4bdd05b4d025b3
|
|
Author: Marc Zyngier <marc.zyngier@arm.com>
|
|
Date: Mon Jun 30 16:01:37 2014 +0100
|
|
|
|
KVM: ARM: vgic: abstract EISR bitmap access
|
|
|
|
Move the GICH_EISR access to its own function.
|
|
|
|
Acked-by: Catalin Marinas <catalin.marinas@arm.com>
|
|
Reviewed-by: Christoffer Dall <christoffer.dall@linaro.org>
|
|
Signed-off-by: Marc Zyngier <marc.zyngier@arm.com>
|
|
|
|
commit 33930dc5f401e9fc2268c2f128853eb5275e7ab1
|
|
Author: Marc Zyngier <marc.zyngier@arm.com>
|
|
Date: Mon Jun 30 16:01:36 2014 +0100
|
|
|
|
KVM: ARM: vgic: abstract access to the ELRSR bitmap
|
|
|
|
Move the GICH_ELRSR access to its own functions, and add them to
|
|
the vgic_ops structure.
|
|
|
|
Acked-by: Catalin Marinas <catalin.marinas@arm.com>
|
|
Reviewed-by: Christoffer Dall <christoffer.dall@linaro.org>
|
|
Signed-off-by: Marc Zyngier <marc.zyngier@arm.com>
|
|
|
|
commit 0dca962ccd5b96fb7174880c1bc25085dcc09927
|
|
Author: Marc Zyngier <marc.zyngier@arm.com>
|
|
Date: Mon Jun 30 16:01:35 2014 +0100
|
|
|
|
KVM: ARM: vgic: introduce vgic_ops and LR manipulation primitives
|
|
|
|
In order to split the various register manipulation from the main vgic
|
|
code, introduce a vgic_ops structure, and start by abstracting the
|
|
LR manipulation code with a couple of accessors.
|
|
|
|
Reviewed-by: Christoffer Dall <christoffer.dall@linaro.org>
|
|
Signed-off-by: Marc Zyngier <marc.zyngier@arm.com>
|
|
|
|
commit e66aa6cbdaa25764c58ba8d21da8d1b7d75e7570
|
|
Author: Marc Zyngier <marc.zyngier@arm.com>
|
|
Date: Mon Jun 30 16:01:34 2014 +0100
|
|
|
|
KVM: arm/arm64: vgic: move GICv2 registers to their own structure
|
|
|
|
In order to make way for the GICv3 registers, move the v2-specific
|
|
registers to their own structure.
|
|
|
|
Acked-by: Catalin Marinas <catalin.marinas@arm.com>
|
|
Reviewed-by: Christoffer Dall <christoffer.dall@linaro.org>
|
|
Signed-off-by: Marc Zyngier <marc.zyngier@arm.com>
|
|
|
|
commit b04259737c6bb38592c02c93e939fd53909dad04
|
|
Author: Marc Zyngier <marc.zyngier@arm.com>
|
|
Date: Mon Jun 30 16:01:33 2014 +0100
|
|
|
|
arm64: boot protocol documentation update for GICv3
|
|
|
|
Linux has some requirements that must be satisfied in order to boot
|
|
on a system built with a GICv3.
|
|
|
|
Acked-by: Christoffer Dall <christoffer.dall@linaro.org>
|
|
Signed-off-by: Marc Zyngier <marc.zyngier@arm.com>
|
|
|
|
commit a864693311bd4305214d966b7ca0e0015216c2c4
|
|
Author: Marc Zyngier <marc.zyngier@arm.com>
|
|
Date: Mon Jun 30 16:01:32 2014 +0100
|
|
|
|
arm64: GICv3 device tree binding documentation
|
|
|
|
Add the necessary documentation to support GICv3.
|
|
|
|
Cc: Thomas Gleixner <tglx@linutronix.de>
|
|
Cc: Mark Rutland <mark.rutland@arm.com>
|
|
Acked-by: Catalin Marinas <catalin.marinas@arm.com>
|
|
Acked-by: Rob Herring <robh@kernel.org>
|
|
Acked-by: Christoffer Dall <christoffer.dall@linaro.org>
|
|
Signed-off-by: Marc Zyngier <marc.zyngier@arm.com>
|
|
|
|
commit af3035d3ddddb6e19bac9295f0c785bb2c8e718c
|
|
Author: Marc Zyngier <marc.zyngier@arm.com>
|
|
Date: Mon Jun 30 16:01:31 2014 +0100
|
|
|
|
irqchip: arm64: Initial support for GICv3
|
|
|
|
The Generic Interrupt Controller (version 3) offers services that are
|
|
similar to GICv2, with a number of additional features:
|
|
- Affinity routing based on the CPU MPIDR (ARE)
|
|
- System register for the CPU interfaces (SRE)
|
|
- Support for more that 8 CPUs
|
|
- Locality-specific Peripheral Interrupts (LPIs)
|
|
- Interrupt Translation Services (ITS)
|
|
|
|
This patch adds preliminary support for GICv3 with ARE and SRE,
|
|
non-secure mode only. It relies on higher exception levels to grant ARE
|
|
and SRE access.
|
|
|
|
Support for LPI and ITS will be added at a later time.
|
|
|
|
Cc: Thomas Gleixner <tglx@linutronix.de>
|
|
Cc: Jason Cooper <jason@lakedaemon.net>
|
|
Reviewed-by: Zi Shen Lim <zlim@broadcom.com>
|
|
Reviewed-by: Christoffer Dall <christoffer.dall@linaro.org>
|
|
Reviewed-by: Tirumalesh Chalamarla <tchalamarla@cavium.com>
|
|
Reviewed-by: Yun Wu <wuyun.wu@huawei.com>
|
|
Reviewed-by: Zhen Lei <thunder.leizhen@huawei.com>
|
|
Tested-by: Tirumalesh Chalamarla<tchalamarla@cavium.com>
|
|
Tested-by: Radha Mohan Chintakuntla <rchintakuntla@cavium.com>
|
|
Acked-by: Radha Mohan Chintakuntla <rchintakuntla@cavium.com>
|
|
Acked-by: Catalin Marinas <catalin.marinas@arm.com>
|
|
Signed-off-by: Marc Zyngier <marc.zyngier@arm.com>
|
|
|
|
commit c50b02761c3ad5d37ce8fffe7c0bf6b46d23109e
|
|
Author: Marc Zyngier <marc.zyngier@arm.com>
|
|
Date: Mon Jun 30 16:01:30 2014 +0100
|
|
|
|
irqchip: ARM: GIC: Move some bits of GICv2 to a library-type file
|
|
|
|
A few GICv2 low-level function are actually very useful to GICv3,
|
|
and it makes some sense to share them across the two drivers.
|
|
They end-up in their own file, with an additional parameter used
|
|
to ensure an optional synchronization (unused on GICv2).
|
|
|
|
Cc: Thomas Gleixner <tglx@linutronix.de>
|
|
Cc: Jason Cooper <jason@lakedaemon.net>
|
|
Acked-by: Christoffer Dall <christoffer.dall@linaro.org>
|
|
Signed-off-by: Marc Zyngier <marc.zyngier@arm.com>
|
|
|
|
commit 6a6033cdcbf6022c2848e2181ed6d8b7545af02e
|
|
Author: Mark Salter <msalter@redhat.com>
|
|
Date: Tue Jun 24 23:16:45 2014 -0400
|
|
|
|
perf: fix arm64 build error
|
|
|
|
I'm seeing the following build error on arm64:
|
|
|
|
In file included from util/event.c:3:0:
|
|
util/event.h:95:17: error: 'PERF_REGS_MAX' undeclared here (not in a function)
|
|
u64 cache_regs[PERF_REGS_MAX];
|
|
^
|
|
|
|
This patch adds a PEFF_REGS_MAX definition for arm64.
|
|
|
|
Signed-off-by: Mark Salter <msalter@redhat.com>
|
|
|
|
commit 1de8987fa549d421576c1b61282c4041c8c78a5f
|
|
Author: Mark Salter <msalter@redhat.com>
|
|
Date: Tue Jun 24 09:50:28 2014 -0400
|
|
|
|
arm64: use EFI as last resort for reboot and poweroff
|
|
|
|
Wire in support for EFI reboot and poweroff functions. We use these
|
|
only if no other mechanism has been registered with arm_pm_reboot
|
|
and/or pm_power_off respectively.
|
|
|
|
Signed-off-by: Mark Salter <msalter@redhat.com>
|
|
|
|
commit 1c973051e86625be7ffb3db90d4a70b9ca4199c6
|
|
Author: Matt Fleming <matt.fleming@intel.com>
|
|
Date: Fri Jun 13 12:39:55 2014 +0100
|
|
|
|
x86/reboot: Add EFI reboot quirk for ACPI Hardware Reduced flag
|
|
|
|
It appears that the BayTrail-T class of hardware requires EFI in order
|
|
to powerdown and reboot and no other reliable method exists.
|
|
|
|
This quirk is generally applicable to all hardware that has the ACPI
|
|
Hardware Reduced bit set, since usually ACPI would be the preferred
|
|
method.
|
|
|
|
Cc: Len Brown <len.brown@intel.com>
|
|
Cc: Mark Salter <msalter@redhat.com>
|
|
Cc: "Rafael J. Wysocki" <rafael.j.wysocki@intel.com>
|
|
Signed-off-by: Matt Fleming <matt.fleming@intel.com>
|
|
|
|
commit 621b2a0f1df86bd2f147799303b94575f3acee95
|
|
Author: Matt Fleming <matt.fleming@intel.com>
|
|
Date: Fri Jun 13 12:35:21 2014 +0100
|
|
|
|
efi/reboot: Allow powering off machines using EFI
|
|
|
|
Not only can EfiResetSystem() be used to reboot, it can also be used to
|
|
power down machines.
|
|
|
|
By and large, this functionality doesn't work very well across the range
|
|
of EFI machines in the wild, so it should definitely only be used as a
|
|
last resort. In an ideal world, this wouldn't be needed at all.
|
|
|
|
Unfortunately, we're starting to see machines where EFI is the *only*
|
|
reliable way to power down, and nothing else, not PCI, not ACPI, works.
|
|
|
|
efi_poweroff_required() should be implemented on a per-architecture
|
|
basis, since exactly when we should be using EFI runtime services is a
|
|
platform-specific decision. There's no analogue for reboot because each
|
|
architecture handles reboot very differently - the x86 code in
|
|
particular is pretty complex.
|
|
|
|
Patches to enable this for specific classes of hardware will be
|
|
submitted separately.
|
|
|
|
Cc: Mark Salter <msalter@redhat.com>
|
|
Signed-off-by: Matt Fleming <matt.fleming@intel.com>
|
|
|
|
commit 6b2e219b20933cad5d5ba34f7af4efc5317c0fb9
|
|
Author: Matt Fleming <matt.fleming@intel.com>
|
|
Date: Fri Jun 13 12:22:22 2014 +0100
|
|
|
|
efi/reboot: Add generic wrapper around EfiResetSystem()
|
|
|
|
Implement efi_reboot(), which is really just a wrapper around the
|
|
EfiResetSystem() EFI runtime service, but it does at least allow us to
|
|
funnel all callers through a single location.
|
|
|
|
It also simplifies the callsites since users no longer need to check to
|
|
see whether EFI_RUNTIME_SERVICES are enabled.
|
|
|
|
Cc: Tony Luck <tony.luck@intel.com>
|
|
Cc: Mark Salter <msalter@redhat.com>
|
|
Signed-off-by: Matt Fleming <matt.fleming@intel.com>
|
|
|
|
commit d0d41b99122d97f81ad05868dff38ccf0a3ffd33
|
|
Author: Saurabh Tangri <saurabh.tangri@intel.com>
|
|
Date: Mon Jun 2 05:18:35 2014 -0700
|
|
|
|
x86/efi: Move all workarounds to a separate file quirks.c
|
|
|
|
Currently, it's difficult to find all the workarounds that are
|
|
applied when running on EFI, because they're littered throughout
|
|
various code paths. This change moves all of them into a separate
|
|
file with the hope that it will be come the single location for all
|
|
our well documented quirks.
|
|
|
|
Signed-off-by: Saurabh Tangri <saurabh.tangri@intel.com>
|
|
Signed-off-by: Matt Fleming <matt.fleming@intel.com>
|
|
|
|
commit b5e3a1e8825abb0406ead0e85436df4df20ddcdb
|
|
Author: Don Dutile <ddutile@redhat.com>
|
|
Date: Tue Mar 25 20:22:26 2014 -0400
|
|
|
|
pmu: Adding support for Xgene PMUs
|
|
|
|
Message-id: <1395778948-47814-2-git-send-email-ddutile@redhat.com>
|
|
Patchwork-id: 78602
|
|
O-Subject: [PATCH 1/3] pmu: Adding support for Xgene PMUs
|
|
Bugzilla: 1079110
|
|
|
|
Backport of these two posted (but not upstream) patches.
|
|
Combined into single patch due to gic-patch dependency.
|
|
|
|
Signed-off-by: Donald Dutile <ddutile@redhat.com>
|
|
|
|
commit 9f4c27693bb120a3134e3e7e8d452fb02d023e2b
|
|
Author: Mark Salter <msalter@redhat.com>
|
|
Date: Sun Jun 15 09:06:55 2014 -0400
|
|
|
|
arm64: fix up APM Mustang devicetree
|
|
|
|
These are changes needed when loading device tree blob built with
|
|
kernel. i.e. with grub. These are not needed when using devicetree
|
|
from Tianocore which will be fixed up at tianocore runtime.
|
|
|
|
Signed-off-by: Mark Salter <msalter@redhat.com>
|
|
|
|
commit 1f3a5b228be88be3f734d7a43db3b3f81e160443
|
|
Author: Iyappan Subramanian <isubramanian@apm.com>
|
|
Date: Mon Jul 14 15:18:05 2014 -0700
|
|
|
|
drivers: net: Add APM X-Gene SoC ethernet driver support.
|
|
|
|
This patch adds network driver for APM X-Gene SoC ethernet.
|
|
|
|
Signed-off-by: Iyappan Subramanian <isubramanian@apm.com>
|
|
Signed-off-by: Ravi Patel <rapatel@apm.com>
|
|
Signed-off-by: Keyur Chudgar <kchudgar@apm.com>
|
|
|
|
commit 54b3fe04c4a953eeb6907ffe9f57aae282f59457
|
|
Author: Iyappan Subramanian <isubramanian@apm.com>
|
|
Date: Mon Jul 14 15:18:04 2014 -0700
|
|
|
|
dts: Add bindings for APM X-Gene SoC ethernet driver
|
|
|
|
This patch adds bindings for APM X-Gene SoC ethernet driver.
|
|
|
|
Signed-off-by: Iyappan Subramanian <isubramanian@apm.com>
|
|
Signed-off-by: Ravi Patel <rapatel@apm.com>
|
|
Signed-off-by: Keyur Chudgar <kchudgar@apm.com>
|
|
|
|
commit 85125c4e1c1b1ef53d6cb77966efa89062540f43
|
|
Author: Iyappan Subramanian <isubramanian@apm.com>
|
|
Date: Mon Jul 14 15:18:03 2014 -0700
|
|
|
|
Documentation: dts: Add bindings for APM X-Gene SoC ethernet driver
|
|
|
|
This patch adds documentation for APM X-Gene SoC ethernet DTS binding.
|
|
|
|
Signed-off-by: Iyappan Subramanian <isubramanian@apm.com>
|
|
Signed-off-by: Ravi Patel <rapatel@apm.com>
|
|
Signed-off-by: Keyur Chudgar <kchudgar@apm.com>
|
|
|
|
commit fe8ec437eedc45384c23e1e12a09baa82d24fa16
|
|
Author: Iyappan Subramanian <isubramanian@apm.com>
|
|
Date: Mon Jul 14 15:18:02 2014 -0700
|
|
|
|
MAINTAINERS: Add entry for APM X-Gene SoC ethernet driver
|
|
|
|
This patch adds a MAINTAINERS entry for APM X-Gene SoC
|
|
ethernet driver.
|
|
|
|
Signed-off-by: Iyappan Subramanian <isubramanian@apm.com>
|
|
Signed-off-by: Ravi Patel <rapatel@apm.com>
|
|
Signed-off-by: Keyur Chudgar <kchudgar@apm.com>
|
|
|
|
commit 237639e43c3d6587985a736f33264e129123d7a5
|
|
Author: Mark Salter <msalter@redhat.com>
|
|
Date: Fri Jul 25 15:14:32 2014 -0400
|
|
|
|
arm/kvm: WIP fix for stage2 pgd memory leak
|
|
|
|
Signed-off-by: Mark Salter <msalter@redhat.com>
|
|
|
|
commit 0794e2900e5e4be4b7aa0b389e6b0bf8b55c5cd7
|
|
Author: Mark Salter <msalter@redhat.com>
|
|
Date: Thu Jul 17 13:34:50 2014 -0400
|
|
|
|
ahci_xgene: add errata workaround for ATA_CMD_SMART
|
|
|
|
commit 2a0bdff6b958d1b2:
|
|
|
|
ahci_xgene: fix the dma state machine lockup for the IDENTIFY DEVICE PIO mode command.
|
|
|
|
added a workaround for X-Gene AHCI controller errata. This was done
|
|
for all ATA_CMD_ID_ATA commands. The errata also appears to affect
|
|
ATA_CMD_SMART commands as well. This was discovered when running
|
|
smartd or just smartctl -x. This patch adds a dma engine restart for
|
|
ATA_CMD_SMART commands which clears up the issues seen with smartd.
|
|
|
|
Signed-off-by: Mark Salter <msalter@redhat.com>
|
|
|
|
commit 7adf85b63608b8bea1148f2faa84f475252a9e43
|
|
Author: Mark Salter <msalter@redhat.com>
|
|
Date: Fri Jul 25 15:32:05 2014 -0400
|
|
|
|
rtc: ia64: allow other architectures to use EFI RTC
|
|
|
|
Currently, the rtc-efi driver is restricted to ia64 only. Newer
|
|
architectures with EFI support may want to also use that driver. This
|
|
patch moves the platform device setup from ia64 into drivers/rtc and allow
|
|
any architecture with CONFIG_EFI=y to use the rtc-efi driver.
|
|
|
|
Signed-off-by: Mark Salter <msalter@redhat.com>
|
|
Cc: Alessandro Zummo <a.zummo@towertech.it>
|
|
Cc: Tony Luck <tony.luck@intel.com>
|
|
Cc: Fenghua Yu <fenghua.yu@intel.com>
|
|
Cc: Andrew Morton <akpm@linux-foundation.org>
|
|
|
|
commit 3b2f96c7a6bfbd46e7dee1d7000081422a7983ce
|
|
Author: Kyle McMartin <kmcmarti@redhat.com>
|
|
Date: Tue May 13 22:25:26 2014 -0400
|
|
|
|
arm64: don't set READ_IMPLIES_EXEC for EM_AARCH64 ELF objects
|
|
|
|
Message-id: <20140513222526.GC26038@redacted.bos.redhat.com>
|
|
Patchwork-id: 79789
|
|
O-Subject: [ACADIA PATCH] arm64: don't set READ_IMPLIES_EXEC for EM_AARCH64 ELF objects
|
|
Bugzilla: 1085528
|
|
|
|
BZ: https://bugzilla.redhat.com/show_bug.cgi?id=1085528
|
|
Upstream: submitted soon
|
|
|
|
[Sadly this isn't (yet) sufficient... but it fixes at least one issue
|
|
here... cat /proc/$$/personality shows READ_IMPLIES_EXEC before. I'll
|
|
try to figure the rest out tomorrow.]
|
|
|
|
Currently, we're accidentally ending up with executable stacks on
|
|
AArch64 when the ABI says we shouldn't be, and relying on glibc to fix
|
|
things up for us when we're loaded. However, SELinux will deny us
|
|
mucking with the stack, and hit us with execmem AVCs.
|
|
|
|
The reason this is happening is somewhat complex:
|
|
|
|
fs/binfmt_elf.c:load_elf_binary()
|
|
- initializes executable_stack = EXSTACK_DEFAULT implying the
|
|
architecture should make up its mind.
|
|
- does a pile of loading goo
|
|
- runs through the program headers, looking for PT_GNU_STACK
|
|
and setting (or unsetting) executable_stack if it finds it.
|
|
|
|
This is our first problem, we won't generate these unless an
|
|
executable stack is explicitly requested.
|
|
|
|
- more ELF loading goo
|
|
- sets whether we're a compat task or not (TIF_32BIT) based on compat.h
|
|
- for compat reasons (pre-GNU_STACK) checks if the READ_IMPLIES_EXEC
|
|
flag should be set for ancient toolchains
|
|
|
|
Here's our second problem, we test if read_implies_exec based on
|
|
stk != EXSTACK_DISABLE_X, which is true since stk == EXSTACK_DEFAULT.
|
|
|
|
So we set current->personality |= READ_IMPLIES_EXEC like a broken
|
|
legacy toolchain would want.
|
|
|
|
- Now we call setup_arg_pages to set up the stack...
|
|
|
|
fs/exec.c:setup_arg_pages()
|
|
- lots of magic happens here
|
|
- vm_flags gets initialized to VM_STACK_FLAGS
|
|
|
|
Here's our third problem, VM_STACK_FLAGS on arm64 is
|
|
VM_DEFAULT_DATA_FLAG which tests READ_IMPLIES_EXEC and sets VM_EXEC
|
|
if it's true. So we end up with an executable stack mapping, since we
|
|
don't have executable_stack set (it's still EXSTACK_DEFAULT at this
|
|
point) to unset it anywhere.
|
|
|
|
Bang. execstack AVC when the program starts running.
|
|
|
|
The easiest way I can see to fix this is to test if we're a legacy task
|
|
and fix it up there. But that's not as simple as it sounds, because
|
|
the 32-bit ABI depends on what revision of the CPU we've enabled (not
|
|
that it matters since we're ARMv8...) Regardless, in the compat case,
|
|
set READ_IMPLIES_EXEC if we've found a GNU_STACK header which explicitly
|
|
requested it as in arch/arm/kernel/elf.c:arm_elf_read_implies_exec().
|
|
|
|
Signed-off-by: Kyle McMartin <kmcmarti@redhat.com>
|
|
Signed-off-by: Donald Dutile <ddutile@redhat.com>
|
|
|
|
commit 4b866971e92b925a44da8d876cb57864942a90b8
|
|
Author: Mark Salter <msalter@redhat.com>
|
|
Date: Thu Jul 24 15:56:15 2014 +0100
|
|
|
|
arm64: fix soft lockup due to large tlb flush range
|
|
|
|
Under certain loads, this soft lockup has been observed:
|
|
|
|
BUG: soft lockup - CPU#2 stuck for 22s! [ip6tables:1016]
|
|
Modules linked in: ip6t_rpfilter ip6t_REJECT cfg80211 rfkill xt_conntrack ebtable_nat ebtable_broute bridge stp llc ebtable_filter ebtables ip6table_nat nf_conntrack_ipv6 nf_defrag_ipv6 nf_nat_ipv6 ip6table_mangle ip6table_security ip6table_raw ip6table_filter ip6_tables iptable_nat nf_conntrack_ipv4 nf_defrag_ipv4 nf_nat_ipv4 nf_nat nf_conntrack iptable_mangle iptable_security iptable_raw vfat fat efivarfs xfs libcrc32c
|
|
|
|
CPU: 2 PID: 1016 Comm: ip6tables Not tainted 3.13.0-0.rc7.30.sa2.aarch64 #1
|
|
task: fffffe03e81d1400 ti: fffffe03f01f8000 task.ti: fffffe03f01f8000
|
|
PC is at __cpu_flush_kern_tlb_range+0xc/0x40
|
|
LR is at __purge_vmap_area_lazy+0x28c/0x3ac
|
|
pc : [<fffffe000009c5cc>] lr : [<fffffe0000182710>] pstate: 80000145
|
|
sp : fffffe03f01fbb70
|
|
x29: fffffe03f01fbb70 x28: fffffe03f01f8000
|
|
x27: fffffe0000b19000 x26: 00000000000000d0
|
|
x25: 000000000000001c x24: fffffe03f01fbc50
|
|
x23: fffffe03f01fbc58 x22: fffffe03f01fbc10
|
|
x21: fffffe0000b2a3f8 x20: 0000000000000802
|
|
x19: fffffe0000b2a3c8 x18: 000003fffdf52710
|
|
x17: 000003ff9d8bb910 x16: fffffe000050fbfc
|
|
x15: 0000000000005735 x14: 000003ff9d7e1a5c
|
|
x13: 0000000000000000 x12: 000003ff9d7e1a5c
|
|
x11: 0000000000000007 x10: fffffe0000c09af0
|
|
x9 : fffffe0000ad1000 x8 : 000000000000005c
|
|
x7 : fffffe03e8624000 x6 : 0000000000000000
|
|
x5 : 0000000000000000 x4 : 0000000000000000
|
|
x3 : fffffe0000c09cc8 x2 : 0000000000000000
|
|
x1 : 000fffffdfffca80 x0 : 000fffffcd742150
|
|
|
|
The __cpu_flush_kern_tlb_range() function looks like:
|
|
|
|
ENTRY(__cpu_flush_kern_tlb_range)
|
|
dsb sy
|
|
lsr x0, x0, #12
|
|
lsr x1, x1, #12
|
|
1: tlbi vaae1is, x0
|
|
add x0, x0, #1
|
|
cmp x0, x1
|
|
b.lo 1b
|
|
dsb sy
|
|
isb
|
|
ret
|
|
ENDPROC(__cpu_flush_kern_tlb_range)
|
|
|
|
The above soft lockup shows the PC at tlbi insn with:
|
|
|
|
x0 = 0x000fffffcd742150
|
|
x1 = 0x000fffffdfffca80
|
|
|
|
So __cpu_flush_kern_tlb_range has 0x128ba930 tlbi flushes left
|
|
after it has already been looping for 23 seconds!.
|
|
|
|
Looking up one frame at __purge_vmap_area_lazy(), there is:
|
|
|
|
...
|
|
list_for_each_entry_rcu(va, &vmap_area_list, list) {
|
|
if (va->flags & VM_LAZY_FREE) {
|
|
if (va->va_start < *start)
|
|
*start = va->va_start;
|
|
if (va->va_end > *end)
|
|
*end = va->va_end;
|
|
nr += (va->va_end - va->va_start) >> PAGE_SHIFT;
|
|
list_add_tail(&va->purge_list, &valist);
|
|
va->flags |= VM_LAZY_FREEING;
|
|
va->flags &= ~VM_LAZY_FREE;
|
|
}
|
|
}
|
|
...
|
|
if (nr || force_flush)
|
|
flush_tlb_kernel_range(*start, *end);
|
|
|
|
So if two areas are being freed, the range passed to
|
|
flush_tlb_kernel_range() may be as large as the vmalloc
|
|
space. For arm64, this is ~240GB for 4k pagesize and ~2TB
|
|
for 64kpage size.
|
|
|
|
This patch works around this problem by adding a loop limit.
|
|
If the range is larger than the limit, use flush_tlb_all()
|
|
rather than flushing based on individual pages. The limit
|
|
chosen is arbitrary as the TLB size is implementation
|
|
specific and not accessible in an architected way. The aim
|
|
of the arbitrary limit is to avoid soft lockup.
|
|
|
|
Signed-off-by: Mark Salter <msalter@redhat.com>
|
|
[catalin.marinas@arm.com: commit log update]
|
|
[catalin.marinas@arm.com: marginal optimisation]
|
|
[catalin.marinas@arm.com: changed to MAX_TLB_RANGE and added comment]
|
|
Signed-off-by: Catalin Marinas <catalin.marinas@arm.com>
|
|
|
|
diff --git a/Documentation/arm64/booting.txt b/Documentation/arm64/booting.txt
|
|
index 37fc4f6..da1d4bf 100644
|
|
--- a/Documentation/arm64/booting.txt
|
|
+++ b/Documentation/arm64/booting.txt
|
|
@@ -141,6 +141,14 @@ Before jumping into the kernel, the following conditions must be met:
|
|
the kernel image will be entered must be initialised by software at a
|
|
higher exception level to prevent execution in an UNKNOWN state.
|
|
|
|
+ For systems with a GICv3 interrupt controller:
|
|
+ - If EL3 is present:
|
|
+ ICC_SRE_EL3.Enable (bit 3) must be initialiased to 0b1.
|
|
+ ICC_SRE_EL3.SRE (bit 0) must be initialised to 0b1.
|
|
+ - If the kernel is entered at EL1:
|
|
+ ICC.SRE_EL2.Enable (bit 3) must be initialised to 0b1
|
|
+ ICC_SRE_EL2.SRE (bit 0) must be initialised to 0b1.
|
|
+
|
|
The requirements described above for CPU mode, caches, MMUs, architected
|
|
timers, coherency and system registers apply to all CPUs. All CPUs must
|
|
enter the kernel in the same exception level.
|
|
diff --git a/Documentation/devicetree/bindings/arm/gic-v3.txt b/Documentation/devicetree/bindings/arm/gic-v3.txt
|
|
new file mode 100644
|
|
index 0000000..33cd05e
|
|
--- /dev/null
|
|
+++ b/Documentation/devicetree/bindings/arm/gic-v3.txt
|
|
@@ -0,0 +1,79 @@
|
|
+* ARM Generic Interrupt Controller, version 3
|
|
+
|
|
+AArch64 SMP cores are often associated with a GICv3, providing Private
|
|
+Peripheral Interrupts (PPI), Shared Peripheral Interrupts (SPI),
|
|
+Software Generated Interrupts (SGI), and Locality-specific Peripheral
|
|
+Interrupts (LPI).
|
|
+
|
|
+Main node required properties:
|
|
+
|
|
+- compatible : should at least contain "arm,gic-v3".
|
|
+- interrupt-controller : Identifies the node as an interrupt controller
|
|
+- #interrupt-cells : Specifies the number of cells needed to encode an
|
|
+ interrupt source. Must be a single cell with a value of at least 3.
|
|
+
|
|
+ The 1st cell is the interrupt type; 0 for SPI interrupts, 1 for PPI
|
|
+ interrupts. Other values are reserved for future use.
|
|
+
|
|
+ The 2nd cell contains the interrupt number for the interrupt type.
|
|
+ SPI interrupts are in the range [0-987]. PPI interrupts are in the
|
|
+ range [0-15].
|
|
+
|
|
+ The 3rd cell is the flags, encoded as follows:
|
|
+ bits[3:0] trigger type and level flags.
|
|
+ 1 = edge triggered
|
|
+ 4 = level triggered
|
|
+
|
|
+ Cells 4 and beyond are reserved for future use. When the 1st cell
|
|
+ has a value of 0 or 1, cells 4 and beyond act as padding, and may be
|
|
+ ignored. It is recommended that padding cells have a value of 0.
|
|
+
|
|
+- reg : Specifies base physical address(s) and size of the GIC
|
|
+ registers, in the following order:
|
|
+ - GIC Distributor interface (GICD)
|
|
+ - GIC Redistributors (GICR), one range per redistributor region
|
|
+ - GIC CPU interface (GICC)
|
|
+ - GIC Hypervisor interface (GICH)
|
|
+ - GIC Virtual CPU interface (GICV)
|
|
+
|
|
+ GICC, GICH and GICV are optional.
|
|
+
|
|
+- interrupts : Interrupt source of the VGIC maintenance interrupt.
|
|
+
|
|
+Optional
|
|
+
|
|
+- redistributor-stride : If using padding pages, specifies the stride
|
|
+ of consecutive redistributors. Must be a multiple of 64kB.
|
|
+
|
|
+- #redistributor-regions: The number of independent contiguous regions
|
|
+ occupied by the redistributors. Required if more than one such
|
|
+ region is present.
|
|
+
|
|
+Examples:
|
|
+
|
|
+ gic: interrupt-controller@2cf00000 {
|
|
+ compatible = "arm,gic-v3";
|
|
+ #interrupt-cells = <3>;
|
|
+ interrupt-controller;
|
|
+ reg = <0x0 0x2f000000 0 0x10000>, // GICD
|
|
+ <0x0 0x2f100000 0 0x200000>, // GICR
|
|
+ <0x0 0x2c000000 0 0x2000>, // GICC
|
|
+ <0x0 0x2c010000 0 0x2000>, // GICH
|
|
+ <0x0 0x2c020000 0 0x2000>; // GICV
|
|
+ interrupts = <1 9 4>;
|
|
+ };
|
|
+
|
|
+ gic: interrupt-controller@2c010000 {
|
|
+ compatible = "arm,gic-v3";
|
|
+ #interrupt-cells = <3>;
|
|
+ interrupt-controller;
|
|
+ redistributor-stride = <0x0 0x40000>; // 256kB stride
|
|
+ #redistributor-regions = <2>;
|
|
+ reg = <0x0 0x2c010000 0 0x10000>, // GICD
|
|
+ <0x0 0x2d000000 0 0x800000>, // GICR 1: CPUs 0-31
|
|
+ <0x0 0x2e000000 0 0x800000>; // GICR 2: CPUs 32-63
|
|
+ <0x0 0x2c040000 0 0x2000>, // GICC
|
|
+ <0x0 0x2c060000 0 0x2000>, // GICH
|
|
+ <0x0 0x2c080000 0 0x2000>; // GICV
|
|
+ interrupts = <1 9 4>;
|
|
+ };
|
|
diff --git a/Documentation/devicetree/bindings/net/apm-xgene-enet.txt b/Documentation/devicetree/bindings/net/apm-xgene-enet.txt
|
|
new file mode 100644
|
|
index 0000000..3e2a295
|
|
--- /dev/null
|
|
+++ b/Documentation/devicetree/bindings/net/apm-xgene-enet.txt
|
|
@@ -0,0 +1,72 @@
|
|
+APM X-Gene SoC Ethernet nodes
|
|
+
|
|
+Ethernet nodes are defined to describe on-chip ethernet interfaces in
|
|
+APM X-Gene SoC.
|
|
+
|
|
+Required properties:
|
|
+- compatible: Should be "apm,xgene-enet"
|
|
+- reg: Address and length of the register set for the device. It contains the
|
|
+ information of registers in the same order as described by reg-names
|
|
+- reg-names: Should contain the register set names
|
|
+ "enet_csr": Ethernet control and status register address space
|
|
+ "ring_csr": Descriptor ring control and status register address space
|
|
+ "ring_cmd": Descriptor ring command register address space
|
|
+- interrupts: Ethernet main interrupt
|
|
+- clocks: Reference to the clock entry.
|
|
+- local-mac-address: MAC address assigned to this device
|
|
+- phy-connection-type: Interface type between ethernet device and PHY device
|
|
+- phy-handle: Reference to a PHY node connected to this device
|
|
+
|
|
+- mdio: Device tree subnode with the following required
|
|
+ properties:
|
|
+
|
|
+ - compatible: Must be "apm,xgene-mdio".
|
|
+ - #address-cells: Must be <1>.
|
|
+ - #size-cells: Must be <0>.
|
|
+
|
|
+ For the phy on the mdio bus, there must be a node with the following
|
|
+ fields:
|
|
+
|
|
+ - compatible: PHY identifier. Please refer ./phy.txt for the format.
|
|
+ - reg: The ID number for the phy.
|
|
+
|
|
+Optional properties:
|
|
+- status : Should be "ok" or "disabled" for enabled/disabled.
|
|
+ Default is "ok".
|
|
+
|
|
+
|
|
+Example:
|
|
+ menetclk: menetclk {
|
|
+ compatible = "apm,xgene-device-clock";
|
|
+ clock-output-names = "menetclk";
|
|
+ status = "ok";
|
|
+ };
|
|
+
|
|
+ menet: ethernet@17020000 {
|
|
+ compatible = "apm,xgene-enet";
|
|
+ status = "disabled";
|
|
+ reg = <0x0 0x17020000 0x0 0xd100>,
|
|
+ <0x0 0X17030000 0x0 0X400>,
|
|
+ <0x0 0X10000000 0x0 0X200>;
|
|
+ reg-names = "enet_csr", "ring_csr", "ring_cmd";
|
|
+ interrupts = <0x0 0x3c 0x4>;
|
|
+ clocks = <&menetclk 0>;
|
|
+ local-mac-address = [00 01 73 00 00 01];
|
|
+ phy-connection-type = "rgmii";
|
|
+ phy-handle = <&menetphy>;
|
|
+ mdio {
|
|
+ compatible = "apm,xgene-mdio";
|
|
+ #address-cells = <1>;
|
|
+ #size-cells = <0>;
|
|
+ menetphy: menetphy@3 {
|
|
+ compatible = "ethernet-phy-id001c.c915";
|
|
+ reg = <0x3>;
|
|
+ };
|
|
+
|
|
+ };
|
|
+ };
|
|
+
|
|
+/* Board-specific peripheral configurations */
|
|
+&menet {
|
|
+ status = "ok";
|
|
+};
|
|
diff --git a/Documentation/devicetree/bindings/pci/xgene-pci.txt b/Documentation/devicetree/bindings/pci/xgene-pci.txt
|
|
new file mode 100644
|
|
index 0000000..e19fdb8
|
|
--- /dev/null
|
|
+++ b/Documentation/devicetree/bindings/pci/xgene-pci.txt
|
|
@@ -0,0 +1,52 @@
|
|
+* AppliedMicro X-Gene PCIe interface
|
|
+
|
|
+Required properties:
|
|
+- device_type: set to "pci"
|
|
+- compatible: should contain "apm,xgene-pcie" to identify the core.
|
|
+- reg: A list of physical base address and length for each set of controller
|
|
+ registers. Must contain an entry for each entry in the reg-names
|
|
+ property.
|
|
+- reg-names: Must include the following entries:
|
|
+ "csr": controller configuration registers.
|
|
+ "cfg": pcie configuration space registers.
|
|
+- #address-cells: set to <3>
|
|
+- #size-cells: set to <2>
|
|
+- ranges: ranges for the outbound memory, I/O regions.
|
|
+- dma-ranges: ranges for the inbound memory regions.
|
|
+- #interrupt-cells: set to <1>
|
|
+- interrupt-map-mask and interrupt-map: standard PCI properties
|
|
+ to define the mapping of the PCIe interface to interrupt
|
|
+ numbers.
|
|
+- clocks: from common clock binding: handle to pci clock.
|
|
+
|
|
+Optional properties:
|
|
+- status: Either "ok" or "disabled".
|
|
+
|
|
+Example:
|
|
+
|
|
+SoC specific DT Entry:
|
|
+ pcie0: pcie@1f2b0000 {
|
|
+ status = "disabled";
|
|
+ device_type = "pci";
|
|
+ compatible = "apm,xgene-storm-pcie", "apm,xgene-pcie";
|
|
+ #interrupt-cells = <1>;
|
|
+ #size-cells = <2>;
|
|
+ #address-cells = <3>;
|
|
+ reg = < 0x00 0x1f2b0000 0x0 0x00010000 /* Controller registers */
|
|
+ 0xe0 0xd0000000 0x0 0x00200000>; /* PCI config space */
|
|
+ reg-names = "csr", "cfg";
|
|
+ ranges = <0x01000000 0x00 0x00000000 0xe0 0x00000000 0x00 0x00010000 /* io */
|
|
+ 0x02000000 0x00 0x10000000 0xe0 0x10000000 0x00 0x80000000>; /* mem */
|
|
+ dma-ranges = <0x42000000 0x40 0x00000000 0x40 0x00000000 0x40 0x00000000>;
|
|
+ interrupt-map-mask = <0x0 0x0 0x0 0x7>;
|
|
+ interrupt-map = <0x0 0x0 0x0 0x1 &gic 0x0 0xc2 0x1
|
|
+ 0x0 0x0 0x0 0x2 &gic 0x0 0xc3 0x1
|
|
+ 0x0 0x0 0x0 0x3 &gic 0x0 0xc4 0x1
|
|
+ 0x0 0x0 0x0 0x4 &gic 0x0 0xc5 0x1>;
|
|
+ clocks = <&pcie0clk 0>;
|
|
+ };
|
|
+
|
|
+Board specific DT Entry:
|
|
+ &pcie0 {
|
|
+ status = "ok";
|
|
+ };
|
|
diff --git a/MAINTAINERS b/MAINTAINERS
|
|
index 86efa7e..14a3ef1 100644
|
|
--- a/MAINTAINERS
|
|
+++ b/MAINTAINERS
|
|
@@ -699,6 +699,14 @@ S: Maintained
|
|
F: drivers/net/appletalk/
|
|
F: net/appletalk/
|
|
|
|
+APPLIED MICRO (APM) X-GENE SOC ETHERNET DRIVER
|
|
+M: Iyappan Subramanian <isubramanian@apm.com>
|
|
+M: Keyur Chudgar <kchudgar@apm.com>
|
|
+M: Ravi Patel <rapatel@apm.com>
|
|
+S: Supported
|
|
+F: drivers/net/ethernet/apm/xgene/
|
|
+F: Documentation/devicetree/bindings/net/apm-xgene-enet.txt
|
|
+
|
|
APTINA CAMERA SENSOR PLL
|
|
M: Laurent Pinchart <Laurent.pinchart@ideasonboard.com>
|
|
L: linux-media@vger.kernel.org
|
|
@@ -6851,6 +6859,13 @@ S: Maintained
|
|
F: Documentation/devicetree/bindings/pci/host-generic-pci.txt
|
|
F: drivers/pci/host/pci-host-generic.c
|
|
|
|
+PCI DRIVER FOR APPLIEDMICRO XGENE
|
|
+M: Tanmay Inamdar <tinamdar@apm.com>
|
|
+L: linux-pci@vger.kernel.org
|
|
+L: linux-arm-kernel@lists.infradead.org
|
|
+S: Maintained
|
|
+F: drivers/pci/host/pci-xgene.c
|
|
+
|
|
PCMCIA SUBSYSTEM
|
|
P: Linux PCMCIA Team
|
|
L: linux-pcmcia@lists.infradead.org
|
|
diff --git a/arch/arm/include/asm/kvm_host.h b/arch/arm/include/asm/kvm_host.h
|
|
index 193ceaf..d6d5227 100644
|
|
--- a/arch/arm/include/asm/kvm_host.h
|
|
+++ b/arch/arm/include/asm/kvm_host.h
|
|
@@ -225,6 +225,11 @@ static inline int kvm_arch_dev_ioctl_check_extension(long ext)
|
|
return 0;
|
|
}
|
|
|
|
+static inline void vgic_arch_setup(const struct vgic_params *vgic)
|
|
+{
|
|
+ BUG_ON(vgic->type != VGIC_V2);
|
|
+}
|
|
+
|
|
int kvm_perf_init(void);
|
|
int kvm_perf_teardown(void);
|
|
|
|
diff --git a/arch/arm/kernel/asm-offsets.c b/arch/arm/kernel/asm-offsets.c
|
|
index 85598b5..713e807 100644
|
|
--- a/arch/arm/kernel/asm-offsets.c
|
|
+++ b/arch/arm/kernel/asm-offsets.c
|
|
@@ -182,13 +182,13 @@ int main(void)
|
|
DEFINE(VCPU_HYP_PC, offsetof(struct kvm_vcpu, arch.fault.hyp_pc));
|
|
#ifdef CONFIG_KVM_ARM_VGIC
|
|
DEFINE(VCPU_VGIC_CPU, offsetof(struct kvm_vcpu, arch.vgic_cpu));
|
|
- DEFINE(VGIC_CPU_HCR, offsetof(struct vgic_cpu, vgic_hcr));
|
|
- DEFINE(VGIC_CPU_VMCR, offsetof(struct vgic_cpu, vgic_vmcr));
|
|
- DEFINE(VGIC_CPU_MISR, offsetof(struct vgic_cpu, vgic_misr));
|
|
- DEFINE(VGIC_CPU_EISR, offsetof(struct vgic_cpu, vgic_eisr));
|
|
- DEFINE(VGIC_CPU_ELRSR, offsetof(struct vgic_cpu, vgic_elrsr));
|
|
- DEFINE(VGIC_CPU_APR, offsetof(struct vgic_cpu, vgic_apr));
|
|
- DEFINE(VGIC_CPU_LR, offsetof(struct vgic_cpu, vgic_lr));
|
|
+ DEFINE(VGIC_V2_CPU_HCR, offsetof(struct vgic_cpu, vgic_v2.vgic_hcr));
|
|
+ DEFINE(VGIC_V2_CPU_VMCR, offsetof(struct vgic_cpu, vgic_v2.vgic_vmcr));
|
|
+ DEFINE(VGIC_V2_CPU_MISR, offsetof(struct vgic_cpu, vgic_v2.vgic_misr));
|
|
+ DEFINE(VGIC_V2_CPU_EISR, offsetof(struct vgic_cpu, vgic_v2.vgic_eisr));
|
|
+ DEFINE(VGIC_V2_CPU_ELRSR, offsetof(struct vgic_cpu, vgic_v2.vgic_elrsr));
|
|
+ DEFINE(VGIC_V2_CPU_APR, offsetof(struct vgic_cpu, vgic_v2.vgic_apr));
|
|
+ DEFINE(VGIC_V2_CPU_LR, offsetof(struct vgic_cpu, vgic_v2.vgic_lr));
|
|
DEFINE(VGIC_CPU_NR_LR, offsetof(struct vgic_cpu, nr_lr));
|
|
#ifdef CONFIG_KVM_ARM_TIMER
|
|
DEFINE(VCPU_TIMER_CNTV_CTL, offsetof(struct kvm_vcpu, arch.timer_cpu.cntv_ctl));
|
|
diff --git a/arch/arm/kvm/Makefile b/arch/arm/kvm/Makefile
|
|
index 789bca9..f7057ed 100644
|
|
--- a/arch/arm/kvm/Makefile
|
|
+++ b/arch/arm/kvm/Makefile
|
|
@@ -21,4 +21,5 @@ obj-y += kvm-arm.o init.o interrupts.o
|
|
obj-y += arm.o handle_exit.o guest.o mmu.o emulate.o reset.o
|
|
obj-y += coproc.o coproc_a15.o coproc_a7.o mmio.o psci.o perf.o
|
|
obj-$(CONFIG_KVM_ARM_VGIC) += $(KVM)/arm/vgic.o
|
|
+obj-$(CONFIG_KVM_ARM_VGIC) += $(KVM)/arm/vgic-v2.o
|
|
obj-$(CONFIG_KVM_ARM_TIMER) += $(KVM)/arm/arch_timer.o
|
|
diff --git a/arch/arm/kvm/interrupts_head.S b/arch/arm/kvm/interrupts_head.S
|
|
index 76af9302..e4eaf30 100644
|
|
--- a/arch/arm/kvm/interrupts_head.S
|
|
+++ b/arch/arm/kvm/interrupts_head.S
|
|
@@ -421,14 +421,14 @@ vcpu .req r0 @ vcpu pointer always in r0
|
|
ldr r9, [r2, #GICH_ELRSR1]
|
|
ldr r10, [r2, #GICH_APR]
|
|
|
|
- str r3, [r11, #VGIC_CPU_HCR]
|
|
- str r4, [r11, #VGIC_CPU_VMCR]
|
|
- str r5, [r11, #VGIC_CPU_MISR]
|
|
- str r6, [r11, #VGIC_CPU_EISR]
|
|
- str r7, [r11, #(VGIC_CPU_EISR + 4)]
|
|
- str r8, [r11, #VGIC_CPU_ELRSR]
|
|
- str r9, [r11, #(VGIC_CPU_ELRSR + 4)]
|
|
- str r10, [r11, #VGIC_CPU_APR]
|
|
+ str r3, [r11, #VGIC_V2_CPU_HCR]
|
|
+ str r4, [r11, #VGIC_V2_CPU_VMCR]
|
|
+ str r5, [r11, #VGIC_V2_CPU_MISR]
|
|
+ str r6, [r11, #VGIC_V2_CPU_EISR]
|
|
+ str r7, [r11, #(VGIC_V2_CPU_EISR + 4)]
|
|
+ str r8, [r11, #VGIC_V2_CPU_ELRSR]
|
|
+ str r9, [r11, #(VGIC_V2_CPU_ELRSR + 4)]
|
|
+ str r10, [r11, #VGIC_V2_CPU_APR]
|
|
|
|
/* Clear GICH_HCR */
|
|
mov r5, #0
|
|
@@ -436,7 +436,7 @@ vcpu .req r0 @ vcpu pointer always in r0
|
|
|
|
/* Save list registers */
|
|
add r2, r2, #GICH_LR0
|
|
- add r3, r11, #VGIC_CPU_LR
|
|
+ add r3, r11, #VGIC_V2_CPU_LR
|
|
ldr r4, [r11, #VGIC_CPU_NR_LR]
|
|
1: ldr r6, [r2], #4
|
|
str r6, [r3], #4
|
|
@@ -463,9 +463,9 @@ vcpu .req r0 @ vcpu pointer always in r0
|
|
add r11, vcpu, #VCPU_VGIC_CPU
|
|
|
|
/* We only restore a minimal set of registers */
|
|
- ldr r3, [r11, #VGIC_CPU_HCR]
|
|
- ldr r4, [r11, #VGIC_CPU_VMCR]
|
|
- ldr r8, [r11, #VGIC_CPU_APR]
|
|
+ ldr r3, [r11, #VGIC_V2_CPU_HCR]
|
|
+ ldr r4, [r11, #VGIC_V2_CPU_VMCR]
|
|
+ ldr r8, [r11, #VGIC_V2_CPU_APR]
|
|
|
|
str r3, [r2, #GICH_HCR]
|
|
str r4, [r2, #GICH_VMCR]
|
|
@@ -473,7 +473,7 @@ vcpu .req r0 @ vcpu pointer always in r0
|
|
|
|
/* Restore list registers */
|
|
add r2, r2, #GICH_LR0
|
|
- add r3, r11, #VGIC_CPU_LR
|
|
+ add r3, r11, #VGIC_V2_CPU_LR
|
|
ldr r4, [r11, #VGIC_CPU_NR_LR]
|
|
1: ldr r6, [r3], #4
|
|
str r6, [r2], #4
|
|
diff --git a/arch/arm64/Kconfig b/arch/arm64/Kconfig
|
|
index 839f48c..23871dd 100644
|
|
--- a/arch/arm64/Kconfig
|
|
+++ b/arch/arm64/Kconfig
|
|
@@ -11,6 +11,7 @@ config ARM64
|
|
select ARM_AMBA
|
|
select ARM_ARCH_TIMER
|
|
select ARM_GIC
|
|
+ select ARM_GIC_V3
|
|
select BUILDTIME_EXTABLE_SORT
|
|
select CLONE_BACKWARDS
|
|
select COMMON_CLK
|
|
@@ -76,7 +77,7 @@ config MMU
|
|
def_bool y
|
|
|
|
config NO_IOPORT_MAP
|
|
- def_bool y
|
|
+ def_bool y if !PCI
|
|
|
|
config STACKTRACE_SUPPORT
|
|
def_bool y
|
|
@@ -151,6 +152,23 @@ menu "Bus support"
|
|
config ARM_AMBA
|
|
bool
|
|
|
|
+config PCI
|
|
+ bool "PCI support"
|
|
+ help
|
|
+ This feature enables support for PCIe bus system. If you say Y
|
|
+ here, the kernel will include drivers and infrastructure code
|
|
+ to support PCIe bus devices.
|
|
+
|
|
+config PCI_DOMAINS
|
|
+ def_bool PCI
|
|
+
|
|
+config PCI_SYSCALL
|
|
+ def_bool PCI
|
|
+
|
|
+source "drivers/pci/Kconfig"
|
|
+source "drivers/pci/pcie/Kconfig"
|
|
+source "drivers/pci/hotplug/Kconfig"
|
|
+
|
|
endmenu
|
|
|
|
menu "Kernel Features"
|
|
diff --git a/arch/arm64/boot/dts/apm-mustang.dts b/arch/arm64/boot/dts/apm-mustang.dts
|
|
index 6541962..0cb67fc 100644
|
|
--- a/arch/arm64/boot/dts/apm-mustang.dts
|
|
+++ b/arch/arm64/boot/dts/apm-mustang.dts
|
|
@@ -28,3 +28,15 @@
|
|
&serial0 {
|
|
status = "ok";
|
|
};
|
|
+
|
|
+&menet {
|
|
+ status = "ok";
|
|
+};
|
|
+
|
|
+&pcie0clk {
|
|
+ status = "ok";
|
|
+};
|
|
+
|
|
+&pcie0 {
|
|
+ status = "ok";
|
|
+};
|
|
diff --git a/arch/arm64/boot/dts/apm-storm.dtsi b/arch/arm64/boot/dts/apm-storm.dtsi
|
|
index 40aa96c..fb2ee54 100644
|
|
--- a/arch/arm64/boot/dts/apm-storm.dtsi
|
|
+++ b/arch/arm64/boot/dts/apm-storm.dtsi
|
|
@@ -24,56 +24,56 @@
|
|
compatible = "apm,potenza", "arm,armv8";
|
|
reg = <0x0 0x000>;
|
|
enable-method = "spin-table";
|
|
- cpu-release-addr = <0x1 0x0000fff8>;
|
|
+ cpu-release-addr = <0x40 0x0000f000>;
|
|
};
|
|
cpu@001 {
|
|
device_type = "cpu";
|
|
compatible = "apm,potenza", "arm,armv8";
|
|
reg = <0x0 0x001>;
|
|
enable-method = "spin-table";
|
|
- cpu-release-addr = <0x1 0x0000fff8>;
|
|
+ cpu-release-addr = <0x40 0x0000f000>;
|
|
};
|
|
cpu@100 {
|
|
device_type = "cpu";
|
|
compatible = "apm,potenza", "arm,armv8";
|
|
reg = <0x0 0x100>;
|
|
enable-method = "spin-table";
|
|
- cpu-release-addr = <0x1 0x0000fff8>;
|
|
+ cpu-release-addr = <0x40 0x0000f000>;
|
|
};
|
|
cpu@101 {
|
|
device_type = "cpu";
|
|
compatible = "apm,potenza", "arm,armv8";
|
|
reg = <0x0 0x101>;
|
|
enable-method = "spin-table";
|
|
- cpu-release-addr = <0x1 0x0000fff8>;
|
|
+ cpu-release-addr = <0x40 0x0000f000>;
|
|
};
|
|
cpu@200 {
|
|
device_type = "cpu";
|
|
compatible = "apm,potenza", "arm,armv8";
|
|
reg = <0x0 0x200>;
|
|
enable-method = "spin-table";
|
|
- cpu-release-addr = <0x1 0x0000fff8>;
|
|
+ cpu-release-addr = <0x40 0x0000f000>;
|
|
};
|
|
cpu@201 {
|
|
device_type = "cpu";
|
|
compatible = "apm,potenza", "arm,armv8";
|
|
reg = <0x0 0x201>;
|
|
enable-method = "spin-table";
|
|
- cpu-release-addr = <0x1 0x0000fff8>;
|
|
+ cpu-release-addr = <0x40 0x0000f000>;
|
|
};
|
|
cpu@300 {
|
|
device_type = "cpu";
|
|
compatible = "apm,potenza", "arm,armv8";
|
|
reg = <0x0 0x300>;
|
|
enable-method = "spin-table";
|
|
- cpu-release-addr = <0x1 0x0000fff8>;
|
|
+ cpu-release-addr = <0x40 0x0000f000>;
|
|
};
|
|
cpu@301 {
|
|
device_type = "cpu";
|
|
compatible = "apm,potenza", "arm,armv8";
|
|
reg = <0x0 0x301>;
|
|
enable-method = "spin-table";
|
|
- cpu-release-addr = <0x1 0x0000fff8>;
|
|
+ cpu-release-addr = <0x40 0x0000f000>;
|
|
};
|
|
};
|
|
|
|
@@ -97,6 +97,11 @@
|
|
clock-frequency = <50000000>;
|
|
};
|
|
|
|
+ pmu {
|
|
+ compatible = "arm,armv8-pmuv3";
|
|
+ interrupts = <1 12 0xff04>;
|
|
+ };
|
|
+
|
|
soc {
|
|
compatible = "simple-bus";
|
|
#address-cells = <2>;
|
|
@@ -167,14 +172,13 @@
|
|
clock-output-names = "ethclk";
|
|
};
|
|
|
|
- eth8clk: eth8clk {
|
|
+ menetclk: menetclk {
|
|
compatible = "apm,xgene-device-clock";
|
|
#clock-cells = <1>;
|
|
clocks = <ðclk 0>;
|
|
- clock-names = "eth8clk";
|
|
reg = <0x0 0x1702C000 0x0 0x1000>;
|
|
reg-names = "csr-reg";
|
|
- clock-output-names = "eth8clk";
|
|
+ clock-output-names = "menetclk";
|
|
};
|
|
|
|
sataphy1clk: sataphy1clk@1f21c000 {
|
|
@@ -270,6 +274,161 @@
|
|
enable-mask = <0x2>;
|
|
clock-output-names = "rtcclk";
|
|
};
|
|
+
|
|
+ pcie0clk: pcie0clk@1f2bc000 {
|
|
+ status = "disabled";
|
|
+ compatible = "apm,xgene-device-clock";
|
|
+ #clock-cells = <1>;
|
|
+ clocks = <&socplldiv2 0>;
|
|
+ reg = <0x0 0x1f2bc000 0x0 0x1000>;
|
|
+ reg-names = "csr-reg";
|
|
+ clock-output-names = "pcie0clk";
|
|
+ };
|
|
+
|
|
+ pcie1clk: pcie1clk@1f2cc000 {
|
|
+ status = "disabled";
|
|
+ compatible = "apm,xgene-device-clock";
|
|
+ #clock-cells = <1>;
|
|
+ clocks = <&socplldiv2 0>;
|
|
+ reg = <0x0 0x1f2cc000 0x0 0x1000>;
|
|
+ reg-names = "csr-reg";
|
|
+ clock-output-names = "pcie1clk";
|
|
+ };
|
|
+
|
|
+ pcie2clk: pcie2clk@1f2dc000 {
|
|
+ status = "disabled";
|
|
+ compatible = "apm,xgene-device-clock";
|
|
+ #clock-cells = <1>;
|
|
+ clocks = <&socplldiv2 0>;
|
|
+ reg = <0x0 0x1f2dc000 0x0 0x1000>;
|
|
+ reg-names = "csr-reg";
|
|
+ clock-output-names = "pcie2clk";
|
|
+ };
|
|
+
|
|
+ pcie3clk: pcie3clk@1f50c000 {
|
|
+ status = "disabled";
|
|
+ compatible = "apm,xgene-device-clock";
|
|
+ #clock-cells = <1>;
|
|
+ clocks = <&socplldiv2 0>;
|
|
+ reg = <0x0 0x1f50c000 0x0 0x1000>;
|
|
+ reg-names = "csr-reg";
|
|
+ clock-output-names = "pcie3clk";
|
|
+ };
|
|
+
|
|
+ pcie4clk: pcie4clk@1f51c000 {
|
|
+ status = "disabled";
|
|
+ compatible = "apm,xgene-device-clock";
|
|
+ #clock-cells = <1>;
|
|
+ clocks = <&socplldiv2 0>;
|
|
+ reg = <0x0 0x1f51c000 0x0 0x1000>;
|
|
+ reg-names = "csr-reg";
|
|
+ clock-output-names = "pcie4clk";
|
|
+ };
|
|
+ };
|
|
+
|
|
+ pcie0: pcie@1f2b0000 {
|
|
+ status = "disabled";
|
|
+ device_type = "pci";
|
|
+ compatible = "apm,xgene-storm-pcie", "apm,xgene-pcie";
|
|
+ #interrupt-cells = <1>;
|
|
+ #size-cells = <2>;
|
|
+ #address-cells = <3>;
|
|
+ reg = < 0x00 0x1f2b0000 0x0 0x00010000 /* Controller registers */
|
|
+ 0xe0 0xd0000000 0x0 0x00200000>; /* PCI config space */
|
|
+ reg-names = "csr", "cfg";
|
|
+ ranges = <0x01000000 0x00 0x00000000 0xe0 0x00000000 0x00 0x00010000 /* io */
|
|
+ 0x02000000 0x00 0x10000000 0xe0 0x10000000 0x00 0x80000000>; /* mem */
|
|
+ dma-ranges = <0x42000000 0x40 0x00000000 0x40 0x00000000 0x40 0x00000000>;
|
|
+ interrupt-map-mask = <0x0 0x0 0x0 0x7>;
|
|
+ interrupt-map = <0x0 0x0 0x0 0x1 &gic 0x0 0xc2 0x1
|
|
+ 0x0 0x0 0x0 0x2 &gic 0x0 0xc3 0x1
|
|
+ 0x0 0x0 0x0 0x3 &gic 0x0 0xc4 0x1
|
|
+ 0x0 0x0 0x0 0x4 &gic 0x0 0xc5 0x1>;
|
|
+ clocks = <&pcie0clk 0>;
|
|
+ };
|
|
+
|
|
+ pcie1: pcie@1f2c0000 {
|
|
+ status = "disabled";
|
|
+ device_type = "pci";
|
|
+ compatible = "apm,xgene-storm-pcie", "apm,xgene-pcie";
|
|
+ #interrupt-cells = <1>;
|
|
+ #size-cells = <2>;
|
|
+ #address-cells = <3>;
|
|
+ reg = < 0x00 0x1f2c0000 0x0 0x00010000 /* Controller registers */
|
|
+ 0xd0 0xd0000000 0x0 0x00200000>; /* PCI config space */
|
|
+ reg-names = "csr", "cfg";
|
|
+ ranges = <0x01000000 0x0 0x00000000 0xd0 0x00000000 0x00 0x00010000 /* io */
|
|
+ 0x02000000 0x0 0x10000000 0xd0 0x10000000 0x00 0x80000000>; /* mem */
|
|
+ dma-ranges = <0x42000000 0x40 0x00000000 0x40 0x00000000 0x40 0x00000000>;
|
|
+ interrupt-map-mask = <0x0 0x0 0x0 0x7>;
|
|
+ interrupt-map = <0x0 0x0 0x0 0x1 &gic 0x0 0xc8 0x1
|
|
+ 0x0 0x0 0x0 0x2 &gic 0x0 0xc9 0x1
|
|
+ 0x0 0x0 0x0 0x3 &gic 0x0 0xca 0x1
|
|
+ 0x0 0x0 0x0 0x4 &gic 0x0 0xcb 0x1>;
|
|
+ clocks = <&pcie1clk 0>;
|
|
+ };
|
|
+
|
|
+ pcie2: pcie@1f2d0000 {
|
|
+ status = "disabled";
|
|
+ device_type = "pci";
|
|
+ compatible = "apm,xgene-storm-pcie", "apm,xgene-pcie";
|
|
+ #interrupt-cells = <1>;
|
|
+ #size-cells = <2>;
|
|
+ #address-cells = <3>;
|
|
+ reg = < 0x00 0x1f2d0000 0x0 0x00010000 /* Controller registers */
|
|
+ 0x90 0xd0000000 0x0 0x00200000>; /* PCI config space */
|
|
+ reg-names = "csr", "cfg";
|
|
+ ranges = <0x01000000 0x0 0x00000000 0x90 0x00000000 0x0 0x00010000 /* io */
|
|
+ 0x02000000 0x0 0x10000000 0x90 0x10000000 0x0 0x80000000>; /* mem */
|
|
+ dma-ranges = <0x42000000 0x40 0x00000000 0x40 0x00000000 0x40 0x00000000>;
|
|
+ interrupt-map-mask = <0x0 0x0 0x0 0x7>;
|
|
+ interrupt-map = <0x0 0x0 0x0 0x1 &gic 0x0 0xce 0x1
|
|
+ 0x0 0x0 0x0 0x2 &gic 0x0 0xcf 0x1
|
|
+ 0x0 0x0 0x0 0x3 &gic 0x0 0xd0 0x1
|
|
+ 0x0 0x0 0x0 0x4 &gic 0x0 0xd1 0x1>;
|
|
+ clocks = <&pcie2clk 0>;
|
|
+ };
|
|
+
|
|
+ pcie3: pcie@1f500000 {
|
|
+ status = "disabled";
|
|
+ device_type = "pci";
|
|
+ compatible = "apm,xgene-storm-pcie", "apm,xgene-pcie";
|
|
+ #interrupt-cells = <1>;
|
|
+ #size-cells = <2>;
|
|
+ #address-cells = <3>;
|
|
+ reg = < 0x00 0x1f500000 0x0 0x00010000 /* Controller registers */
|
|
+ 0xa0 0xd0000000 0x0 0x00200000>; /* PCI config space */
|
|
+ reg-names = "csr", "cfg";
|
|
+ ranges = <0x01000000 0x0 0x00000000 0xa0 0x00000000 0x0 0x00010000 /* io */
|
|
+ 0x02000000 0x0 0x10000000 0xa0 0x10000000 0x0 0x80000000>; /* mem */
|
|
+ dma-ranges = <0x42000000 0x40 0x00000000 0x40 0x00000000 0x40 0x00000000>;
|
|
+ interrupt-map-mask = <0x0 0x0 0x0 0x7>;
|
|
+ interrupt-map = <0x0 0x0 0x0 0x1 &gic 0x0 0xd4 0x1
|
|
+ 0x0 0x0 0x0 0x2 &gic 0x0 0xd5 0x1
|
|
+ 0x0 0x0 0x0 0x3 &gic 0x0 0xd6 0x1
|
|
+ 0x0 0x0 0x0 0x4 &gic 0x0 0xd7 0x1>;
|
|
+ clocks = <&pcie3clk 0>;
|
|
+ };
|
|
+
|
|
+ pcie4: pcie@1f510000 {
|
|
+ status = "disabled";
|
|
+ device_type = "pci";
|
|
+ compatible = "apm,xgene-storm-pcie", "apm,xgene-pcie";
|
|
+ #interrupt-cells = <1>;
|
|
+ #size-cells = <2>;
|
|
+ #address-cells = <3>;
|
|
+ reg = < 0x00 0x1f510000 0x0 0x00010000 /* Controller registers */
|
|
+ 0xc0 0xd0000000 0x0 0x00200000>; /* PCI config space */
|
|
+ reg-names = "csr", "cfg";
|
|
+ ranges = <0x01000000 0x0 0x00000000 0xc0 0x00000000 0x0 0x00010000 /* io */
|
|
+ 0x02000000 0x0 0x10000000 0xc0 0x10000000 0x0 0x80000000>; /* mem */
|
|
+ dma-ranges = <0x42000000 0x40 0x00000000 0x40 0x00000000 0x40 0x00000000>;
|
|
+ interrupt-map-mask = <0x0 0x0 0x0 0x7>;
|
|
+ interrupt-map = <0x0 0x0 0x0 0x1 &gic 0x0 0xda 0x1
|
|
+ 0x0 0x0 0x0 0x2 &gic 0x0 0xdb 0x1
|
|
+ 0x0 0x0 0x0 0x3 &gic 0x0 0xdc 0x1
|
|
+ 0x0 0x0 0x0 0x4 &gic 0x0 0xdd 0x1>;
|
|
+ clocks = <&pcie4clk 0>;
|
|
};
|
|
|
|
serial0: serial@1c020000 {
|
|
@@ -278,7 +437,7 @@
|
|
compatible = "ns16550a";
|
|
reg = <0 0x1c020000 0x0 0x1000>;
|
|
reg-shift = <2>;
|
|
- clock-frequency = <10000000>; /* Updated by bootloader */
|
|
+ clock-frequency = <50000000>; /* Updated by bootloader */
|
|
interrupt-parent = <&gic>;
|
|
interrupts = <0x0 0x4c 0x4>;
|
|
};
|
|
@@ -397,5 +556,30 @@
|
|
#clock-cells = <1>;
|
|
clocks = <&rtcclk 0>;
|
|
};
|
|
+
|
|
+ menet: ethernet@17020000 {
|
|
+ compatible = "apm,xgene-enet";
|
|
+ status = "disabled";
|
|
+ reg = <0x0 0x17020000 0x0 0xd100>,
|
|
+ <0x0 0X17030000 0x0 0X400>,
|
|
+ <0x0 0X10000000 0x0 0X200>;
|
|
+ reg-names = "enet_csr", "ring_csr", "ring_cmd";
|
|
+ interrupts = <0x0 0x3c 0x4>;
|
|
+ dma-coherent;
|
|
+ clocks = <&menetclk 0>;
|
|
+ local-mac-address = [00 00 00 00 00 00];
|
|
+ phy-connection-type = "rgmii";
|
|
+ phy-handle = <&menetphy>;
|
|
+ mdio {
|
|
+ compatible = "apm,xgene-mdio";
|
|
+ #address-cells = <1>;
|
|
+ #size-cells = <0>;
|
|
+ menetphy: menetphy@3 {
|
|
+ compatible = "ethernet-phy-id001c.c915";
|
|
+ reg = <0x3>;
|
|
+ };
|
|
+
|
|
+ };
|
|
+ };
|
|
};
|
|
};
|
|
diff --git a/arch/arm64/include/asm/Kbuild b/arch/arm64/include/asm/Kbuild
|
|
index 0b3fcf8..07cb417 100644
|
|
--- a/arch/arm64/include/asm/Kbuild
|
|
+++ b/arch/arm64/include/asm/Kbuild
|
|
@@ -29,6 +29,7 @@ generic-y += mman.h
|
|
generic-y += msgbuf.h
|
|
generic-y += mutex.h
|
|
generic-y += pci.h
|
|
+generic-y += pci-bridge.h
|
|
generic-y += poll.h
|
|
generic-y += preempt.h
|
|
generic-y += resource.h
|
|
diff --git a/arch/arm64/include/asm/elf.h b/arch/arm64/include/asm/elf.h
|
|
index 01d3aab..8186df6 100644
|
|
--- a/arch/arm64/include/asm/elf.h
|
|
+++ b/arch/arm64/include/asm/elf.h
|
|
@@ -114,7 +114,8 @@ typedef struct user_fpsimd_state elf_fpregset_t;
|
|
*/
|
|
#define elf_check_arch(x) ((x)->e_machine == EM_AARCH64)
|
|
|
|
-#define elf_read_implies_exec(ex,stk) (stk != EXSTACK_DISABLE_X)
|
|
+#define elf_read_implies_exec(ex,stk) (test_thread_flag(TIF_32BIT) \
|
|
+ ? (stk == EXSTACK_ENABLE_X) : 0)
|
|
|
|
#define CORE_DUMP_USE_REGSET
|
|
#define ELF_EXEC_PAGESIZE PAGE_SIZE
|
|
diff --git a/arch/arm64/include/asm/io.h b/arch/arm64/include/asm/io.h
|
|
index e0ecdcf..dc34039 100644
|
|
--- a/arch/arm64/include/asm/io.h
|
|
+++ b/arch/arm64/include/asm/io.h
|
|
@@ -121,7 +121,8 @@ static inline u64 __raw_readq(const volatile void __iomem *addr)
|
|
/*
|
|
* I/O port access primitives.
|
|
*/
|
|
-#define IO_SPACE_LIMIT 0xffff
|
|
+#define arch_has_dev_port() (1)
|
|
+#define IO_SPACE_LIMIT 0x1ffffff
|
|
#define PCI_IOBASE ((void __iomem *)(MODULES_VADDR - SZ_32M))
|
|
|
|
static inline u8 inb(unsigned long addr)
|
|
diff --git a/arch/arm64/include/asm/kvm_arm.h b/arch/arm64/include/asm/kvm_arm.h
|
|
index 3d69030..cc83520 100644
|
|
--- a/arch/arm64/include/asm/kvm_arm.h
|
|
+++ b/arch/arm64/include/asm/kvm_arm.h
|
|
@@ -76,9 +76,10 @@
|
|
*/
|
|
#define HCR_GUEST_FLAGS (HCR_TSC | HCR_TSW | HCR_TWE | HCR_TWI | HCR_VM | \
|
|
HCR_TVM | HCR_BSU_IS | HCR_FB | HCR_TAC | \
|
|
- HCR_AMO | HCR_IMO | HCR_FMO | \
|
|
- HCR_SWIO | HCR_TIDCP | HCR_RW)
|
|
+ HCR_AMO | HCR_SWIO | HCR_TIDCP | HCR_RW)
|
|
#define HCR_VIRT_EXCP_MASK (HCR_VA | HCR_VI | HCR_VF)
|
|
+#define HCR_INT_OVERRIDE (HCR_FMO | HCR_IMO)
|
|
+
|
|
|
|
/* Hyp System Control Register (SCTLR_EL2) bits */
|
|
#define SCTLR_EL2_EE (1 << 25)
|
|
diff --git a/arch/arm64/include/asm/kvm_asm.h b/arch/arm64/include/asm/kvm_asm.h
|
|
index 9fcd54b..a28c35b 100644
|
|
--- a/arch/arm64/include/asm/kvm_asm.h
|
|
+++ b/arch/arm64/include/asm/kvm_asm.h
|
|
@@ -18,6 +18,8 @@
|
|
#ifndef __ARM_KVM_ASM_H__
|
|
#define __ARM_KVM_ASM_H__
|
|
|
|
+#include <asm/virt.h>
|
|
+
|
|
/*
|
|
* 0 is reserved as an invalid value.
|
|
* Order *must* be kept in sync with the hyp switch code.
|
|
@@ -96,13 +98,21 @@ extern char __kvm_hyp_init_end[];
|
|
|
|
extern char __kvm_hyp_vector[];
|
|
|
|
-extern char __kvm_hyp_code_start[];
|
|
-extern char __kvm_hyp_code_end[];
|
|
+#define __kvm_hyp_code_start __hyp_text_start
|
|
+#define __kvm_hyp_code_end __hyp_text_end
|
|
|
|
extern void __kvm_flush_vm_context(void);
|
|
extern void __kvm_tlb_flush_vmid_ipa(struct kvm *kvm, phys_addr_t ipa);
|
|
|
|
extern int __kvm_vcpu_run(struct kvm_vcpu *vcpu);
|
|
+
|
|
+extern u64 __vgic_v3_get_ich_vtr_el2(void);
|
|
+
|
|
+extern char __save_vgic_v2_state[];
|
|
+extern char __restore_vgic_v2_state[];
|
|
+extern char __save_vgic_v3_state[];
|
|
+extern char __restore_vgic_v3_state[];
|
|
+
|
|
#endif
|
|
|
|
#endif /* __ARM_KVM_ASM_H__ */
|
|
diff --git a/arch/arm64/include/asm/kvm_host.h b/arch/arm64/include/asm/kvm_host.h
|
|
index 92242ce..4ae9213 100644
|
|
--- a/arch/arm64/include/asm/kvm_host.h
|
|
+++ b/arch/arm64/include/asm/kvm_host.h
|
|
@@ -200,4 +200,32 @@ static inline void __cpu_init_hyp_mode(phys_addr_t boot_pgd_ptr,
|
|
hyp_stack_ptr, vector_ptr);
|
|
}
|
|
|
|
+struct vgic_sr_vectors {
|
|
+ void *save_vgic;
|
|
+ void *restore_vgic;
|
|
+};
|
|
+
|
|
+static inline void vgic_arch_setup(const struct vgic_params *vgic)
|
|
+{
|
|
+ extern struct vgic_sr_vectors __vgic_sr_vectors;
|
|
+
|
|
+ switch(vgic->type)
|
|
+ {
|
|
+ case VGIC_V2:
|
|
+ __vgic_sr_vectors.save_vgic = __save_vgic_v2_state;
|
|
+ __vgic_sr_vectors.restore_vgic = __restore_vgic_v2_state;
|
|
+ break;
|
|
+
|
|
+#ifdef CONFIG_ARM_GIC_V3
|
|
+ case VGIC_V3:
|
|
+ __vgic_sr_vectors.save_vgic = __save_vgic_v3_state;
|
|
+ __vgic_sr_vectors.restore_vgic = __restore_vgic_v3_state;
|
|
+ break;
|
|
+#endif
|
|
+
|
|
+ default:
|
|
+ BUG();
|
|
+ }
|
|
+}
|
|
+
|
|
#endif /* __ARM64_KVM_HOST_H__ */
|
|
diff --git a/arch/arm64/include/asm/kvm_mmu.h b/arch/arm64/include/asm/kvm_mmu.h
|
|
index 7d29847..d7f77ff 100644
|
|
--- a/arch/arm64/include/asm/kvm_mmu.h
|
|
+++ b/arch/arm64/include/asm/kvm_mmu.h
|
|
@@ -122,8 +122,16 @@ static inline void kvm_set_s2pmd_writable(pmd_t *pmd)
|
|
}
|
|
|
|
#define kvm_pgd_addr_end(addr, end) pgd_addr_end(addr, end)
|
|
-#define kvm_pud_addr_end(addr, end) pud_addr_end(addr, end)
|
|
-#define kvm_pmd_addr_end(addr, end) pmd_addr_end(addr, end)
|
|
+
|
|
+#define kvm_pud_addr_end(addr, end) \
|
|
+({ unsigned long __boundary = ((addr) + PUD_SIZE) & PUD_MASK; \
|
|
+ (__boundary - 1 < (end) - 1)? __boundary: (end); \
|
|
+})
|
|
+
|
|
+#define kvm_pmd_addr_end(addr, end) \
|
|
+({ unsigned long __boundary = ((addr) + PMD_SIZE) & PMD_MASK; \
|
|
+ (__boundary - 1 < (end) - 1)? __boundary: (end); \
|
|
+})
|
|
|
|
struct kvm;
|
|
|
|
diff --git a/arch/arm64/include/asm/pci.h b/arch/arm64/include/asm/pci.h
|
|
new file mode 100644
|
|
index 0000000..3f7856e
|
|
--- /dev/null
|
|
+++ b/arch/arm64/include/asm/pci.h
|
|
@@ -0,0 +1,49 @@
|
|
+#ifndef __ASM_PCI_H
|
|
+#define __ASM_PCI_H
|
|
+#ifdef __KERNEL__
|
|
+
|
|
+#include <linux/types.h>
|
|
+#include <linux/slab.h>
|
|
+#include <linux/dma-mapping.h>
|
|
+
|
|
+#include <asm/io.h>
|
|
+#include <asm-generic/pci-bridge.h>
|
|
+#include <asm-generic/pci-dma-compat.h>
|
|
+
|
|
+#define PCIBIOS_MIN_IO 0x1000
|
|
+#define PCIBIOS_MIN_MEM 0
|
|
+
|
|
+struct pci_host_bridge *find_pci_host_bridge(struct pci_bus *bus);
|
|
+
|
|
+/*
|
|
+ * Set to 1 if the kernel should re-assign all PCI bus numbers
|
|
+ */
|
|
+#define pcibios_assign_all_busses() \
|
|
+ (pci_has_flag(PCI_REASSIGN_ALL_BUS))
|
|
+
|
|
+/*
|
|
+ * PCI address space differs from physical memory address space
|
|
+ */
|
|
+#define PCI_DMA_BUS_IS_PHYS (0)
|
|
+
|
|
+extern int isa_dma_bridge_buggy;
|
|
+
|
|
+#ifdef CONFIG_PCI
|
|
+static inline int pci_domain_nr(struct pci_bus *bus)
|
|
+{
|
|
+ struct pci_host_bridge *bridge = find_pci_host_bridge(bus);
|
|
+
|
|
+ if (bridge)
|
|
+ return bridge->domain_nr;
|
|
+
|
|
+ return 0;
|
|
+}
|
|
+
|
|
+static inline int pci_proc_domain(struct pci_bus *bus)
|
|
+{
|
|
+ return 1;
|
|
+}
|
|
+#endif /* CONFIG_PCI */
|
|
+
|
|
+#endif /* __KERNEL__ */
|
|
+#endif /* __ASM_PCI_H */
|
|
diff --git a/arch/arm64/include/asm/tlbflush.h b/arch/arm64/include/asm/tlbflush.h
|
|
index b9349c4..ecbd081 100644
|
|
--- a/arch/arm64/include/asm/tlbflush.h
|
|
+++ b/arch/arm64/include/asm/tlbflush.h
|
|
@@ -98,8 +98,8 @@ static inline void flush_tlb_page(struct vm_area_struct *vma,
|
|
dsb(ish);
|
|
}
|
|
|
|
-static inline void flush_tlb_range(struct vm_area_struct *vma,
|
|
- unsigned long start, unsigned long end)
|
|
+static inline void __flush_tlb_range(struct vm_area_struct *vma,
|
|
+ unsigned long start, unsigned long end)
|
|
{
|
|
unsigned long asid = (unsigned long)ASID(vma->vm_mm) << 48;
|
|
unsigned long addr;
|
|
@@ -112,7 +112,7 @@ static inline void flush_tlb_range(struct vm_area_struct *vma,
|
|
dsb(ish);
|
|
}
|
|
|
|
-static inline void flush_tlb_kernel_range(unsigned long start, unsigned long end)
|
|
+static inline void __flush_tlb_kernel_range(unsigned long start, unsigned long end)
|
|
{
|
|
unsigned long addr;
|
|
start >>= 12;
|
|
@@ -125,6 +125,29 @@ static inline void flush_tlb_kernel_range(unsigned long start, unsigned long end
|
|
}
|
|
|
|
/*
|
|
+ * This is meant to avoid soft lock-ups on large TLB flushing ranges and not
|
|
+ * necessarily a performance improvement.
|
|
+ */
|
|
+#define MAX_TLB_RANGE (1024UL << PAGE_SHIFT)
|
|
+
|
|
+static inline void flush_tlb_range(struct vm_area_struct *vma,
|
|
+ unsigned long start, unsigned long end)
|
|
+{
|
|
+ if ((end - start) <= MAX_TLB_RANGE)
|
|
+ __flush_tlb_range(vma, start, end);
|
|
+ else
|
|
+ flush_tlb_mm(vma->vm_mm);
|
|
+}
|
|
+
|
|
+static inline void flush_tlb_kernel_range(unsigned long start, unsigned long end)
|
|
+{
|
|
+ if ((end - start) <= MAX_TLB_RANGE)
|
|
+ __flush_tlb_kernel_range(start, end);
|
|
+ else
|
|
+ flush_tlb_all();
|
|
+}
|
|
+
|
|
+/*
|
|
* On AArch64, the cache coherency is handled via the set_pte_at() function.
|
|
*/
|
|
static inline void update_mmu_cache(struct vm_area_struct *vma,
|
|
diff --git a/arch/arm64/include/asm/virt.h b/arch/arm64/include/asm/virt.h
|
|
index 215ad46..7a5df52 100644
|
|
--- a/arch/arm64/include/asm/virt.h
|
|
+++ b/arch/arm64/include/asm/virt.h
|
|
@@ -50,6 +50,10 @@ static inline bool is_hyp_mode_mismatched(void)
|
|
return __boot_cpu_mode[0] != __boot_cpu_mode[1];
|
|
}
|
|
|
|
+/* The section containing the hypervisor text */
|
|
+extern char __hyp_text_start[];
|
|
+extern char __hyp_text_end[];
|
|
+
|
|
#endif /* __ASSEMBLY__ */
|
|
|
|
#endif /* ! __ASM__VIRT_H */
|
|
diff --git a/arch/arm64/kernel/Makefile b/arch/arm64/kernel/Makefile
|
|
index cdaedad..36b117a 100644
|
|
--- a/arch/arm64/kernel/Makefile
|
|
+++ b/arch/arm64/kernel/Makefile
|
|
@@ -29,6 +29,7 @@ arm64-obj-$(CONFIG_ARM64_CPU_SUSPEND) += sleep.o suspend.o
|
|
arm64-obj-$(CONFIG_JUMP_LABEL) += jump_label.o
|
|
arm64-obj-$(CONFIG_KGDB) += kgdb.o
|
|
arm64-obj-$(CONFIG_EFI) += efi.o efi-stub.o efi-entry.o
|
|
+arm64-obj-$(CONFIG_PCI) += pci.o
|
|
|
|
obj-y += $(arm64-obj-y) vdso/
|
|
obj-m += $(arm64-obj-m)
|
|
diff --git a/arch/arm64/kernel/asm-offsets.c b/arch/arm64/kernel/asm-offsets.c
|
|
index 646f888..e74654c 100644
|
|
--- a/arch/arm64/kernel/asm-offsets.c
|
|
+++ b/arch/arm64/kernel/asm-offsets.c
|
|
@@ -129,13 +129,24 @@ int main(void)
|
|
DEFINE(KVM_TIMER_ENABLED, offsetof(struct kvm, arch.timer.enabled));
|
|
DEFINE(VCPU_KVM, offsetof(struct kvm_vcpu, kvm));
|
|
DEFINE(VCPU_VGIC_CPU, offsetof(struct kvm_vcpu, arch.vgic_cpu));
|
|
- DEFINE(VGIC_CPU_HCR, offsetof(struct vgic_cpu, vgic_hcr));
|
|
- DEFINE(VGIC_CPU_VMCR, offsetof(struct vgic_cpu, vgic_vmcr));
|
|
- DEFINE(VGIC_CPU_MISR, offsetof(struct vgic_cpu, vgic_misr));
|
|
- DEFINE(VGIC_CPU_EISR, offsetof(struct vgic_cpu, vgic_eisr));
|
|
- DEFINE(VGIC_CPU_ELRSR, offsetof(struct vgic_cpu, vgic_elrsr));
|
|
- DEFINE(VGIC_CPU_APR, offsetof(struct vgic_cpu, vgic_apr));
|
|
- DEFINE(VGIC_CPU_LR, offsetof(struct vgic_cpu, vgic_lr));
|
|
+ DEFINE(VGIC_SAVE_FN, offsetof(struct vgic_sr_vectors, save_vgic));
|
|
+ DEFINE(VGIC_RESTORE_FN, offsetof(struct vgic_sr_vectors, restore_vgic));
|
|
+ DEFINE(VGIC_SR_VECTOR_SZ, sizeof(struct vgic_sr_vectors));
|
|
+ DEFINE(VGIC_V2_CPU_HCR, offsetof(struct vgic_cpu, vgic_v2.vgic_hcr));
|
|
+ DEFINE(VGIC_V2_CPU_VMCR, offsetof(struct vgic_cpu, vgic_v2.vgic_vmcr));
|
|
+ DEFINE(VGIC_V2_CPU_MISR, offsetof(struct vgic_cpu, vgic_v2.vgic_misr));
|
|
+ DEFINE(VGIC_V2_CPU_EISR, offsetof(struct vgic_cpu, vgic_v2.vgic_eisr));
|
|
+ DEFINE(VGIC_V2_CPU_ELRSR, offsetof(struct vgic_cpu, vgic_v2.vgic_elrsr));
|
|
+ DEFINE(VGIC_V2_CPU_APR, offsetof(struct vgic_cpu, vgic_v2.vgic_apr));
|
|
+ DEFINE(VGIC_V2_CPU_LR, offsetof(struct vgic_cpu, vgic_v2.vgic_lr));
|
|
+ DEFINE(VGIC_V3_CPU_HCR, offsetof(struct vgic_cpu, vgic_v3.vgic_hcr));
|
|
+ DEFINE(VGIC_V3_CPU_VMCR, offsetof(struct vgic_cpu, vgic_v3.vgic_vmcr));
|
|
+ DEFINE(VGIC_V3_CPU_MISR, offsetof(struct vgic_cpu, vgic_v3.vgic_misr));
|
|
+ DEFINE(VGIC_V3_CPU_EISR, offsetof(struct vgic_cpu, vgic_v3.vgic_eisr));
|
|
+ DEFINE(VGIC_V3_CPU_ELRSR, offsetof(struct vgic_cpu, vgic_v3.vgic_elrsr));
|
|
+ DEFINE(VGIC_V3_CPU_AP0R, offsetof(struct vgic_cpu, vgic_v3.vgic_ap0r));
|
|
+ DEFINE(VGIC_V3_CPU_AP1R, offsetof(struct vgic_cpu, vgic_v3.vgic_ap1r));
|
|
+ DEFINE(VGIC_V3_CPU_LR, offsetof(struct vgic_cpu, vgic_v3.vgic_lr));
|
|
DEFINE(VGIC_CPU_NR_LR, offsetof(struct vgic_cpu, nr_lr));
|
|
DEFINE(KVM_VTTBR, offsetof(struct kvm, arch.vttbr));
|
|
DEFINE(KVM_VGIC_VCTRL, offsetof(struct kvm, arch.vgic.vctrl_base));
|
|
diff --git a/arch/arm64/kernel/efi-stub.c b/arch/arm64/kernel/efi-stub.c
|
|
index e786e6c..320ef48 100644
|
|
--- a/arch/arm64/kernel/efi-stub.c
|
|
+++ b/arch/arm64/kernel/efi-stub.c
|
|
@@ -58,20 +58,16 @@ static efi_status_t handle_kernel_image(efi_system_table_t *sys_table,
|
|
kernel_size = _edata - _text;
|
|
if (*image_addr != (dram_base + TEXT_OFFSET)) {
|
|
kernel_memsize = kernel_size + (_end - _edata);
|
|
- status = efi_relocate_kernel(sys_table, image_addr,
|
|
- kernel_size, kernel_memsize,
|
|
- dram_base + TEXT_OFFSET,
|
|
- PAGE_SIZE);
|
|
+ status = efi_low_alloc(sys_table, kernel_memsize + TEXT_OFFSET,
|
|
+ SZ_2M, reserve_addr);
|
|
if (status != EFI_SUCCESS) {
|
|
pr_efi_err(sys_table, "Failed to relocate kernel\n");
|
|
return status;
|
|
}
|
|
- if (*image_addr != (dram_base + TEXT_OFFSET)) {
|
|
- pr_efi_err(sys_table, "Failed to alloc kernel memory\n");
|
|
- efi_free(sys_table, kernel_memsize, *image_addr);
|
|
- return EFI_ERROR;
|
|
- }
|
|
- *image_size = kernel_memsize;
|
|
+ memcpy((void *)*reserve_addr + TEXT_OFFSET, (void *)*image_addr,
|
|
+ kernel_size);
|
|
+ *image_addr = *reserve_addr + TEXT_OFFSET;
|
|
+ *reserve_size = kernel_memsize + TEXT_OFFSET;
|
|
}
|
|
|
|
|
|
diff --git a/arch/arm64/kernel/efi.c b/arch/arm64/kernel/efi.c
|
|
index 14db1f6..453b7f8 100644
|
|
--- a/arch/arm64/kernel/efi.c
|
|
+++ b/arch/arm64/kernel/efi.c
|
|
@@ -467,3 +467,14 @@ static int __init arm64_enter_virtual_mode(void)
|
|
return 0;
|
|
}
|
|
early_initcall(arm64_enter_virtual_mode);
|
|
+
|
|
+/*
|
|
+ * If nothing else is handling pm_power_off, use EFI
|
|
+ *
|
|
+ * This is called from a late_initcall after other mechanisms
|
|
+ * have had a chance to register a handler.
|
|
+ */
|
|
+bool efi_poweroff_required(void)
|
|
+{
|
|
+ return pm_power_off == NULL;
|
|
+}
|
|
diff --git a/arch/arm64/kernel/head.S b/arch/arm64/kernel/head.S
|
|
index a2c1195..8df59be 100644
|
|
--- a/arch/arm64/kernel/head.S
|
|
+++ b/arch/arm64/kernel/head.S
|
|
@@ -22,6 +22,7 @@
|
|
|
|
#include <linux/linkage.h>
|
|
#include <linux/init.h>
|
|
+#include <linux/irqchip/arm-gic-v3.h>
|
|
|
|
#include <asm/assembler.h>
|
|
#include <asm/ptrace.h>
|
|
@@ -156,7 +157,7 @@ optional_header:
|
|
.short 0x20b // PE32+ format
|
|
.byte 0x02 // MajorLinkerVersion
|
|
.byte 0x14 // MinorLinkerVersion
|
|
- .long _edata - stext // SizeOfCode
|
|
+ .long _end - stext // SizeOfCode
|
|
.long 0 // SizeOfInitializedData
|
|
.long 0 // SizeOfUninitializedData
|
|
.long efi_stub_entry - efi_head // AddressOfEntryPoint
|
|
@@ -174,7 +175,7 @@ extra_header_fields:
|
|
.short 0 // MinorSubsystemVersion
|
|
.long 0 // Win32VersionValue
|
|
|
|
- .long _edata - efi_head // SizeOfImage
|
|
+ .long _end - efi_head // SizeOfImage
|
|
|
|
// Everything before the kernel image is considered part of the header
|
|
.long stext - efi_head // SizeOfHeaders
|
|
@@ -221,7 +222,7 @@ section_table:
|
|
.byte 0
|
|
.byte 0
|
|
.byte 0 // end of 0 padding of section name
|
|
- .long _edata - stext // VirtualSize
|
|
+ .long _end - stext // VirtualSize
|
|
.long stext - efi_head // VirtualAddress
|
|
.long _edata - stext // SizeOfRawData
|
|
.long stext - efi_head // PointerToRawData
|
|
@@ -295,6 +296,23 @@ CPU_LE( bic x0, x0, #(3 << 24) ) // Clear the EE and E0E bits for EL1
|
|
msr cnthctl_el2, x0
|
|
msr cntvoff_el2, xzr // Clear virtual offset
|
|
|
|
+#ifdef CONFIG_ARM_GIC_V3
|
|
+ /* GICv3 system register access */
|
|
+ mrs x0, id_aa64pfr0_el1
|
|
+ ubfx x0, x0, #24, #4
|
|
+ cmp x0, #1
|
|
+ b.ne 3f
|
|
+
|
|
+ mrs x0, ICC_SRE_EL2
|
|
+ orr x0, x0, #ICC_SRE_EL2_SRE // Set ICC_SRE_EL2.SRE==1
|
|
+ orr x0, x0, #ICC_SRE_EL2_ENABLE // Set ICC_SRE_EL2.Enable==1
|
|
+ msr ICC_SRE_EL2, x0
|
|
+ isb // Make sure SRE is now set
|
|
+ msr ICH_HCR_EL2, xzr // Reset ICC_HCR_EL2 to defaults
|
|
+
|
|
+3:
|
|
+#endif
|
|
+
|
|
/* Populate ID registers. */
|
|
mrs x0, midr_el1
|
|
mrs x1, mpidr_el1
|
|
diff --git a/arch/arm64/kernel/hyp-stub.S b/arch/arm64/kernel/hyp-stub.S
|
|
index 0959611..a272f33 100644
|
|
--- a/arch/arm64/kernel/hyp-stub.S
|
|
+++ b/arch/arm64/kernel/hyp-stub.S
|
|
@@ -19,6 +19,7 @@
|
|
|
|
#include <linux/init.h>
|
|
#include <linux/linkage.h>
|
|
+#include <linux/irqchip/arm-gic-v3.h>
|
|
|
|
#include <asm/assembler.h>
|
|
#include <asm/ptrace.h>
|
|
diff --git a/arch/arm64/kernel/pci.c b/arch/arm64/kernel/pci.c
|
|
new file mode 100644
|
|
index 0000000..955d6d1
|
|
--- /dev/null
|
|
+++ b/arch/arm64/kernel/pci.c
|
|
@@ -0,0 +1,38 @@
|
|
+/*
|
|
+ * Code borrowed from powerpc/kernel/pci-common.c
|
|
+ *
|
|
+ * Copyright (C) 2003 Anton Blanchard <anton@au.ibm.com>, IBM
|
|
+ * Copyright (C) 2014 ARM Ltd.
|
|
+ *
|
|
+ * This program is free software; you can redistribute it and/or
|
|
+ * modify it under the terms of the GNU General Public License
|
|
+ * version 2 as published by the Free Software Foundation.
|
|
+ *
|
|
+ */
|
|
+
|
|
+#include <linux/init.h>
|
|
+#include <linux/io.h>
|
|
+#include <linux/kernel.h>
|
|
+#include <linux/mm.h>
|
|
+#include <linux/of_pci.h>
|
|
+#include <linux/of_platform.h>
|
|
+#include <linux/slab.h>
|
|
+
|
|
+#include <asm/pci-bridge.h>
|
|
+
|
|
+/*
|
|
+ * Called after each bus is probed, but before its children are examined
|
|
+ */
|
|
+void pcibios_fixup_bus(struct pci_bus *bus)
|
|
+{
|
|
+ /* nothing to do, expected to be removed in the future */
|
|
+}
|
|
+
|
|
+/*
|
|
+ * We don't have to worry about legacy ISA devices, so nothing to do here
|
|
+ */
|
|
+resource_size_t pcibios_align_resource(void *data, const struct resource *res,
|
|
+ resource_size_t size, resource_size_t align)
|
|
+{
|
|
+ return res->start;
|
|
+}
|
|
diff --git a/arch/arm64/kernel/process.c b/arch/arm64/kernel/process.c
|
|
index 43b7c34..ec5cbbe 100644
|
|
--- a/arch/arm64/kernel/process.c
|
|
+++ b/arch/arm64/kernel/process.c
|
|
@@ -43,6 +43,7 @@
|
|
#include <linux/hw_breakpoint.h>
|
|
#include <linux/personality.h>
|
|
#include <linux/notifier.h>
|
|
+#include <linux/efi.h>
|
|
|
|
#include <asm/compat.h>
|
|
#include <asm/cacheflush.h>
|
|
@@ -176,6 +177,11 @@ void machine_restart(char *cmd)
|
|
arm_pm_restart(reboot_mode, cmd);
|
|
|
|
/*
|
|
+ * If all else fails, try EFI
|
|
+ */
|
|
+ efi_reboot(reboot_mode, cmd);
|
|
+
|
|
+ /*
|
|
* Whoops - the architecture was unable to reboot.
|
|
*/
|
|
printk("Reboot failed -- System halted\n");
|
|
diff --git a/arch/arm64/kernel/smp_spin_table.c b/arch/arm64/kernel/smp_spin_table.c
|
|
index 0347d38..70181c1 100644
|
|
--- a/arch/arm64/kernel/smp_spin_table.c
|
|
+++ b/arch/arm64/kernel/smp_spin_table.c
|
|
@@ -20,6 +20,7 @@
|
|
#include <linux/init.h>
|
|
#include <linux/of.h>
|
|
#include <linux/smp.h>
|
|
+#include <linux/types.h>
|
|
|
|
#include <asm/cacheflush.h>
|
|
#include <asm/cpu_ops.h>
|
|
@@ -65,12 +66,21 @@ static int smp_spin_table_cpu_init(struct device_node *dn, unsigned int cpu)
|
|
|
|
static int smp_spin_table_cpu_prepare(unsigned int cpu)
|
|
{
|
|
- void **release_addr;
|
|
+ __le64 __iomem *release_addr;
|
|
|
|
if (!cpu_release_addr[cpu])
|
|
return -ENODEV;
|
|
|
|
- release_addr = __va(cpu_release_addr[cpu]);
|
|
+ /*
|
|
+ * The cpu-release-addr may or may not be inside the linear mapping.
|
|
+ * As ioremap_cache will either give us a new mapping or reuse the
|
|
+ * existing linear mapping, we can use it to cover both cases. In
|
|
+ * either case the memory will be MT_NORMAL.
|
|
+ */
|
|
+ release_addr = ioremap_cache(cpu_release_addr[cpu],
|
|
+ sizeof(*release_addr));
|
|
+ if (!release_addr)
|
|
+ return -ENOMEM;
|
|
|
|
/*
|
|
* We write the release address as LE regardless of the native
|
|
@@ -79,15 +89,16 @@ static int smp_spin_table_cpu_prepare(unsigned int cpu)
|
|
* boot-loader's endianess before jumping. This is mandated by
|
|
* the boot protocol.
|
|
*/
|
|
- release_addr[0] = (void *) cpu_to_le64(__pa(secondary_holding_pen));
|
|
-
|
|
- __flush_dcache_area(release_addr, sizeof(release_addr[0]));
|
|
+ writeq_relaxed(__pa(secondary_holding_pen), release_addr);
|
|
+ __flush_dcache_area(release_addr, sizeof(*release_addr));
|
|
|
|
/*
|
|
* Send an event to wake up the secondary CPU.
|
|
*/
|
|
sev();
|
|
|
|
+ iounmap(release_addr);
|
|
+
|
|
return 0;
|
|
}
|
|
|
|
diff --git a/arch/arm64/kvm/Makefile b/arch/arm64/kvm/Makefile
|
|
index 72a9fd5..32a0961 100644
|
|
--- a/arch/arm64/kvm/Makefile
|
|
+++ b/arch/arm64/kvm/Makefile
|
|
@@ -20,4 +20,8 @@ kvm-$(CONFIG_KVM_ARM_HOST) += hyp.o hyp-init.o handle_exit.o
|
|
kvm-$(CONFIG_KVM_ARM_HOST) += guest.o reset.o sys_regs.o sys_regs_generic_v8.o
|
|
|
|
kvm-$(CONFIG_KVM_ARM_VGIC) += $(KVM)/arm/vgic.o
|
|
+kvm-$(CONFIG_KVM_ARM_VGIC) += $(KVM)/arm/vgic-v2.o
|
|
+kvm-$(CONFIG_KVM_ARM_VGIC) += vgic-v2-switch.o
|
|
+kvm-$(CONFIG_KVM_ARM_VGIC) += $(KVM)/arm/vgic-v3.o
|
|
+kvm-$(CONFIG_KVM_ARM_VGIC) += vgic-v3-switch.o
|
|
kvm-$(CONFIG_KVM_ARM_TIMER) += $(KVM)/arm/arch_timer.o
|
|
diff --git a/arch/arm64/kvm/hyp.S b/arch/arm64/kvm/hyp.S
|
|
index b0d1512..5945f3b 100644
|
|
--- a/arch/arm64/kvm/hyp.S
|
|
+++ b/arch/arm64/kvm/hyp.S
|
|
@@ -16,7 +16,6 @@
|
|
*/
|
|
|
|
#include <linux/linkage.h>
|
|
-#include <linux/irqchip/arm-gic.h>
|
|
|
|
#include <asm/assembler.h>
|
|
#include <asm/memory.h>
|
|
@@ -36,9 +35,6 @@
|
|
.pushsection .hyp.text, "ax"
|
|
.align PAGE_SHIFT
|
|
|
|
-__kvm_hyp_code_start:
|
|
- .globl __kvm_hyp_code_start
|
|
-
|
|
.macro save_common_regs
|
|
// x2: base address for cpu context
|
|
// x3: tmp register
|
|
@@ -339,11 +335,8 @@ __kvm_hyp_code_start:
|
|
.endm
|
|
|
|
.macro activate_traps
|
|
- ldr x2, [x0, #VCPU_IRQ_LINES]
|
|
- ldr x1, [x0, #VCPU_HCR_EL2]
|
|
- orr x2, x2, x1
|
|
- msr hcr_el2, x2
|
|
-
|
|
+ ldr x2, [x0, #VCPU_HCR_EL2]
|
|
+ msr hcr_el2, x2
|
|
ldr x2, =(CPTR_EL2_TTA)
|
|
msr cptr_el2, x2
|
|
|
|
@@ -379,100 +372,33 @@ __kvm_hyp_code_start:
|
|
.endm
|
|
|
|
/*
|
|
- * Save the VGIC CPU state into memory
|
|
- * x0: Register pointing to VCPU struct
|
|
- * Do not corrupt x1!!!
|
|
+ * Call into the vgic backend for state saving
|
|
*/
|
|
.macro save_vgic_state
|
|
- /* Get VGIC VCTRL base into x2 */
|
|
- ldr x2, [x0, #VCPU_KVM]
|
|
- kern_hyp_va x2
|
|
- ldr x2, [x2, #KVM_VGIC_VCTRL]
|
|
- kern_hyp_va x2
|
|
- cbz x2, 2f // disabled
|
|
-
|
|
- /* Compute the address of struct vgic_cpu */
|
|
- add x3, x0, #VCPU_VGIC_CPU
|
|
-
|
|
- /* Save all interesting registers */
|
|
- ldr w4, [x2, #GICH_HCR]
|
|
- ldr w5, [x2, #GICH_VMCR]
|
|
- ldr w6, [x2, #GICH_MISR]
|
|
- ldr w7, [x2, #GICH_EISR0]
|
|
- ldr w8, [x2, #GICH_EISR1]
|
|
- ldr w9, [x2, #GICH_ELRSR0]
|
|
- ldr w10, [x2, #GICH_ELRSR1]
|
|
- ldr w11, [x2, #GICH_APR]
|
|
-CPU_BE( rev w4, w4 )
|
|
-CPU_BE( rev w5, w5 )
|
|
-CPU_BE( rev w6, w6 )
|
|
-CPU_BE( rev w7, w7 )
|
|
-CPU_BE( rev w8, w8 )
|
|
-CPU_BE( rev w9, w9 )
|
|
-CPU_BE( rev w10, w10 )
|
|
-CPU_BE( rev w11, w11 )
|
|
-
|
|
- str w4, [x3, #VGIC_CPU_HCR]
|
|
- str w5, [x3, #VGIC_CPU_VMCR]
|
|
- str w6, [x3, #VGIC_CPU_MISR]
|
|
- str w7, [x3, #VGIC_CPU_EISR]
|
|
- str w8, [x3, #(VGIC_CPU_EISR + 4)]
|
|
- str w9, [x3, #VGIC_CPU_ELRSR]
|
|
- str w10, [x3, #(VGIC_CPU_ELRSR + 4)]
|
|
- str w11, [x3, #VGIC_CPU_APR]
|
|
-
|
|
- /* Clear GICH_HCR */
|
|
- str wzr, [x2, #GICH_HCR]
|
|
-
|
|
- /* Save list registers */
|
|
- add x2, x2, #GICH_LR0
|
|
- ldr w4, [x3, #VGIC_CPU_NR_LR]
|
|
- add x3, x3, #VGIC_CPU_LR
|
|
-1: ldr w5, [x2], #4
|
|
-CPU_BE( rev w5, w5 )
|
|
- str w5, [x3], #4
|
|
- sub w4, w4, #1
|
|
- cbnz w4, 1b
|
|
-2:
|
|
+ adr x24, __vgic_sr_vectors
|
|
+ ldr x24, [x24, VGIC_SAVE_FN]
|
|
+ kern_hyp_va x24
|
|
+ blr x24
|
|
+ mrs x24, hcr_el2
|
|
+ mov x25, #HCR_INT_OVERRIDE
|
|
+ neg x25, x25
|
|
+ and x24, x24, x25
|
|
+ msr hcr_el2, x24
|
|
.endm
|
|
|
|
/*
|
|
- * Restore the VGIC CPU state from memory
|
|
- * x0: Register pointing to VCPU struct
|
|
+ * Call into the vgic backend for state restoring
|
|
*/
|
|
.macro restore_vgic_state
|
|
- /* Get VGIC VCTRL base into x2 */
|
|
- ldr x2, [x0, #VCPU_KVM]
|
|
- kern_hyp_va x2
|
|
- ldr x2, [x2, #KVM_VGIC_VCTRL]
|
|
- kern_hyp_va x2
|
|
- cbz x2, 2f // disabled
|
|
-
|
|
- /* Compute the address of struct vgic_cpu */
|
|
- add x3, x0, #VCPU_VGIC_CPU
|
|
-
|
|
- /* We only restore a minimal set of registers */
|
|
- ldr w4, [x3, #VGIC_CPU_HCR]
|
|
- ldr w5, [x3, #VGIC_CPU_VMCR]
|
|
- ldr w6, [x3, #VGIC_CPU_APR]
|
|
-CPU_BE( rev w4, w4 )
|
|
-CPU_BE( rev w5, w5 )
|
|
-CPU_BE( rev w6, w6 )
|
|
-
|
|
- str w4, [x2, #GICH_HCR]
|
|
- str w5, [x2, #GICH_VMCR]
|
|
- str w6, [x2, #GICH_APR]
|
|
-
|
|
- /* Restore list registers */
|
|
- add x2, x2, #GICH_LR0
|
|
- ldr w4, [x3, #VGIC_CPU_NR_LR]
|
|
- add x3, x3, #VGIC_CPU_LR
|
|
-1: ldr w5, [x3], #4
|
|
-CPU_BE( rev w5, w5 )
|
|
- str w5, [x2], #4
|
|
- sub w4, w4, #1
|
|
- cbnz w4, 1b
|
|
-2:
|
|
+ mrs x24, hcr_el2
|
|
+ ldr x25, [x0, #VCPU_IRQ_LINES]
|
|
+ orr x24, x24, #HCR_INT_OVERRIDE
|
|
+ orr x24, x24, x25
|
|
+ msr hcr_el2, x24
|
|
+ adr x24, __vgic_sr_vectors
|
|
+ ldr x24, [x24, #VGIC_RESTORE_FN]
|
|
+ kern_hyp_va x24
|
|
+ blr x24
|
|
.endm
|
|
|
|
.macro save_timer_state
|
|
@@ -653,6 +579,12 @@ ENTRY(__kvm_flush_vm_context)
|
|
ret
|
|
ENDPROC(__kvm_flush_vm_context)
|
|
|
|
+ // struct vgic_sr_vectors __vgi_sr_vectors;
|
|
+ .align 3
|
|
+ENTRY(__vgic_sr_vectors)
|
|
+ .skip VGIC_SR_VECTOR_SZ
|
|
+ENDPROC(__vgic_sr_vectors)
|
|
+
|
|
__kvm_hyp_panic:
|
|
// Guess the context by looking at VTTBR:
|
|
// If zero, then we're already a host.
|
|
@@ -880,7 +812,4 @@ ENTRY(__kvm_hyp_vector)
|
|
ventry el1_error_invalid // Error 32-bit EL1
|
|
ENDPROC(__kvm_hyp_vector)
|
|
|
|
-__kvm_hyp_code_end:
|
|
- .globl __kvm_hyp_code_end
|
|
-
|
|
.popsection
|
|
diff --git a/arch/arm64/kvm/vgic-v2-switch.S b/arch/arm64/kvm/vgic-v2-switch.S
|
|
new file mode 100644
|
|
index 0000000..ae21177
|
|
--- /dev/null
|
|
+++ b/arch/arm64/kvm/vgic-v2-switch.S
|
|
@@ -0,0 +1,133 @@
|
|
+/*
|
|
+ * Copyright (C) 2012,2013 - ARM Ltd
|
|
+ * Author: Marc Zyngier <marc.zyngier@arm.com>
|
|
+ *
|
|
+ * This program is free software; you can redistribute it and/or modify
|
|
+ * it under the terms of the GNU General Public License version 2 as
|
|
+ * published by the Free Software Foundation.
|
|
+ *
|
|
+ * This program is distributed in the hope that it will be useful,
|
|
+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
+ * GNU General Public License for more details.
|
|
+ *
|
|
+ * You should have received a copy of the GNU General Public License
|
|
+ * along with this program. If not, see <http://www.gnu.org/licenses/>.
|
|
+ */
|
|
+
|
|
+#include <linux/linkage.h>
|
|
+#include <linux/irqchip/arm-gic.h>
|
|
+
|
|
+#include <asm/assembler.h>
|
|
+#include <asm/memory.h>
|
|
+#include <asm/asm-offsets.h>
|
|
+#include <asm/kvm.h>
|
|
+#include <asm/kvm_asm.h>
|
|
+#include <asm/kvm_arm.h>
|
|
+#include <asm/kvm_mmu.h>
|
|
+
|
|
+ .text
|
|
+ .pushsection .hyp.text, "ax"
|
|
+
|
|
+/*
|
|
+ * Save the VGIC CPU state into memory
|
|
+ * x0: Register pointing to VCPU struct
|
|
+ * Do not corrupt x1!!!
|
|
+ */
|
|
+ENTRY(__save_vgic_v2_state)
|
|
+__save_vgic_v2_state:
|
|
+ /* Get VGIC VCTRL base into x2 */
|
|
+ ldr x2, [x0, #VCPU_KVM]
|
|
+ kern_hyp_va x2
|
|
+ ldr x2, [x2, #KVM_VGIC_VCTRL]
|
|
+ kern_hyp_va x2
|
|
+ cbz x2, 2f // disabled
|
|
+
|
|
+ /* Compute the address of struct vgic_cpu */
|
|
+ add x3, x0, #VCPU_VGIC_CPU
|
|
+
|
|
+ /* Save all interesting registers */
|
|
+ ldr w4, [x2, #GICH_HCR]
|
|
+ ldr w5, [x2, #GICH_VMCR]
|
|
+ ldr w6, [x2, #GICH_MISR]
|
|
+ ldr w7, [x2, #GICH_EISR0]
|
|
+ ldr w8, [x2, #GICH_EISR1]
|
|
+ ldr w9, [x2, #GICH_ELRSR0]
|
|
+ ldr w10, [x2, #GICH_ELRSR1]
|
|
+ ldr w11, [x2, #GICH_APR]
|
|
+CPU_BE( rev w4, w4 )
|
|
+CPU_BE( rev w5, w5 )
|
|
+CPU_BE( rev w6, w6 )
|
|
+CPU_BE( rev w7, w7 )
|
|
+CPU_BE( rev w8, w8 )
|
|
+CPU_BE( rev w9, w9 )
|
|
+CPU_BE( rev w10, w10 )
|
|
+CPU_BE( rev w11, w11 )
|
|
+
|
|
+ str w4, [x3, #VGIC_V2_CPU_HCR]
|
|
+ str w5, [x3, #VGIC_V2_CPU_VMCR]
|
|
+ str w6, [x3, #VGIC_V2_CPU_MISR]
|
|
+ str w7, [x3, #VGIC_V2_CPU_EISR]
|
|
+ str w8, [x3, #(VGIC_V2_CPU_EISR + 4)]
|
|
+ str w9, [x3, #VGIC_V2_CPU_ELRSR]
|
|
+ str w10, [x3, #(VGIC_V2_CPU_ELRSR + 4)]
|
|
+ str w11, [x3, #VGIC_V2_CPU_APR]
|
|
+
|
|
+ /* Clear GICH_HCR */
|
|
+ str wzr, [x2, #GICH_HCR]
|
|
+
|
|
+ /* Save list registers */
|
|
+ add x2, x2, #GICH_LR0
|
|
+ ldr w4, [x3, #VGIC_CPU_NR_LR]
|
|
+ add x3, x3, #VGIC_V2_CPU_LR
|
|
+1: ldr w5, [x2], #4
|
|
+CPU_BE( rev w5, w5 )
|
|
+ str w5, [x3], #4
|
|
+ sub w4, w4, #1
|
|
+ cbnz w4, 1b
|
|
+2:
|
|
+ ret
|
|
+ENDPROC(__save_vgic_v2_state)
|
|
+
|
|
+/*
|
|
+ * Restore the VGIC CPU state from memory
|
|
+ * x0: Register pointing to VCPU struct
|
|
+ */
|
|
+ENTRY(__restore_vgic_v2_state)
|
|
+__restore_vgic_v2_state:
|
|
+ /* Get VGIC VCTRL base into x2 */
|
|
+ ldr x2, [x0, #VCPU_KVM]
|
|
+ kern_hyp_va x2
|
|
+ ldr x2, [x2, #KVM_VGIC_VCTRL]
|
|
+ kern_hyp_va x2
|
|
+ cbz x2, 2f // disabled
|
|
+
|
|
+ /* Compute the address of struct vgic_cpu */
|
|
+ add x3, x0, #VCPU_VGIC_CPU
|
|
+
|
|
+ /* We only restore a minimal set of registers */
|
|
+ ldr w4, [x3, #VGIC_V2_CPU_HCR]
|
|
+ ldr w5, [x3, #VGIC_V2_CPU_VMCR]
|
|
+ ldr w6, [x3, #VGIC_V2_CPU_APR]
|
|
+CPU_BE( rev w4, w4 )
|
|
+CPU_BE( rev w5, w5 )
|
|
+CPU_BE( rev w6, w6 )
|
|
+
|
|
+ str w4, [x2, #GICH_HCR]
|
|
+ str w5, [x2, #GICH_VMCR]
|
|
+ str w6, [x2, #GICH_APR]
|
|
+
|
|
+ /* Restore list registers */
|
|
+ add x2, x2, #GICH_LR0
|
|
+ ldr w4, [x3, #VGIC_CPU_NR_LR]
|
|
+ add x3, x3, #VGIC_V2_CPU_LR
|
|
+1: ldr w5, [x3], #4
|
|
+CPU_BE( rev w5, w5 )
|
|
+ str w5, [x2], #4
|
|
+ sub w4, w4, #1
|
|
+ cbnz w4, 1b
|
|
+2:
|
|
+ ret
|
|
+ENDPROC(__restore_vgic_v2_state)
|
|
+
|
|
+ .popsection
|
|
diff --git a/arch/arm64/kvm/vgic-v3-switch.S b/arch/arm64/kvm/vgic-v3-switch.S
|
|
new file mode 100644
|
|
index 0000000..21e68f6
|
|
--- /dev/null
|
|
+++ b/arch/arm64/kvm/vgic-v3-switch.S
|
|
@@ -0,0 +1,267 @@
|
|
+/*
|
|
+ * Copyright (C) 2012,2013 - ARM Ltd
|
|
+ * Author: Marc Zyngier <marc.zyngier@arm.com>
|
|
+ *
|
|
+ * This program is free software; you can redistribute it and/or modify
|
|
+ * it under the terms of the GNU General Public License version 2 as
|
|
+ * published by the Free Software Foundation.
|
|
+ *
|
|
+ * This program is distributed in the hope that it will be useful,
|
|
+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
+ * GNU General Public License for more details.
|
|
+ *
|
|
+ * You should have received a copy of the GNU General Public License
|
|
+ * along with this program. If not, see <http://www.gnu.org/licenses/>.
|
|
+ */
|
|
+
|
|
+#include <linux/linkage.h>
|
|
+#include <linux/irqchip/arm-gic-v3.h>
|
|
+
|
|
+#include <asm/assembler.h>
|
|
+#include <asm/memory.h>
|
|
+#include <asm/asm-offsets.h>
|
|
+#include <asm/kvm.h>
|
|
+#include <asm/kvm_asm.h>
|
|
+#include <asm/kvm_arm.h>
|
|
+
|
|
+ .text
|
|
+ .pushsection .hyp.text, "ax"
|
|
+
|
|
+/*
|
|
+ * We store LRs in reverse order to let the CPU deal with streaming
|
|
+ * access. Use this macro to make it look saner...
|
|
+ */
|
|
+#define LR_OFFSET(n) (VGIC_V3_CPU_LR + (15 - n) * 8)
|
|
+
|
|
+/*
|
|
+ * Save the VGIC CPU state into memory
|
|
+ * x0: Register pointing to VCPU struct
|
|
+ * Do not corrupt x1!!!
|
|
+ */
|
|
+.macro save_vgic_v3_state
|
|
+ // Compute the address of struct vgic_cpu
|
|
+ add x3, x0, #VCPU_VGIC_CPU
|
|
+
|
|
+ // Make sure stores to the GIC via the memory mapped interface
|
|
+ // are now visible to the system register interface
|
|
+ dsb st
|
|
+
|
|
+ // Save all interesting registers
|
|
+ mrs x4, ICH_HCR_EL2
|
|
+ mrs x5, ICH_VMCR_EL2
|
|
+ mrs x6, ICH_MISR_EL2
|
|
+ mrs x7, ICH_EISR_EL2
|
|
+ mrs x8, ICH_ELSR_EL2
|
|
+
|
|
+ str w4, [x3, #VGIC_V3_CPU_HCR]
|
|
+ str w5, [x3, #VGIC_V3_CPU_VMCR]
|
|
+ str w6, [x3, #VGIC_V3_CPU_MISR]
|
|
+ str w7, [x3, #VGIC_V3_CPU_EISR]
|
|
+ str w8, [x3, #VGIC_V3_CPU_ELRSR]
|
|
+
|
|
+ msr ICH_HCR_EL2, xzr
|
|
+
|
|
+ mrs x21, ICH_VTR_EL2
|
|
+ mvn w22, w21
|
|
+ ubfiz w23, w22, 2, 4 // w23 = (15 - ListRegs) * 4
|
|
+
|
|
+ adr x24, 1f
|
|
+ add x24, x24, x23
|
|
+ br x24
|
|
+
|
|
+1:
|
|
+ mrs x20, ICH_LR15_EL2
|
|
+ mrs x19, ICH_LR14_EL2
|
|
+ mrs x18, ICH_LR13_EL2
|
|
+ mrs x17, ICH_LR12_EL2
|
|
+ mrs x16, ICH_LR11_EL2
|
|
+ mrs x15, ICH_LR10_EL2
|
|
+ mrs x14, ICH_LR9_EL2
|
|
+ mrs x13, ICH_LR8_EL2
|
|
+ mrs x12, ICH_LR7_EL2
|
|
+ mrs x11, ICH_LR6_EL2
|
|
+ mrs x10, ICH_LR5_EL2
|
|
+ mrs x9, ICH_LR4_EL2
|
|
+ mrs x8, ICH_LR3_EL2
|
|
+ mrs x7, ICH_LR2_EL2
|
|
+ mrs x6, ICH_LR1_EL2
|
|
+ mrs x5, ICH_LR0_EL2
|
|
+
|
|
+ adr x24, 1f
|
|
+ add x24, x24, x23
|
|
+ br x24
|
|
+
|
|
+1:
|
|
+ str x20, [x3, #LR_OFFSET(15)]
|
|
+ str x19, [x3, #LR_OFFSET(14)]
|
|
+ str x18, [x3, #LR_OFFSET(13)]
|
|
+ str x17, [x3, #LR_OFFSET(12)]
|
|
+ str x16, [x3, #LR_OFFSET(11)]
|
|
+ str x15, [x3, #LR_OFFSET(10)]
|
|
+ str x14, [x3, #LR_OFFSET(9)]
|
|
+ str x13, [x3, #LR_OFFSET(8)]
|
|
+ str x12, [x3, #LR_OFFSET(7)]
|
|
+ str x11, [x3, #LR_OFFSET(6)]
|
|
+ str x10, [x3, #LR_OFFSET(5)]
|
|
+ str x9, [x3, #LR_OFFSET(4)]
|
|
+ str x8, [x3, #LR_OFFSET(3)]
|
|
+ str x7, [x3, #LR_OFFSET(2)]
|
|
+ str x6, [x3, #LR_OFFSET(1)]
|
|
+ str x5, [x3, #LR_OFFSET(0)]
|
|
+
|
|
+ tbnz w21, #29, 6f // 6 bits
|
|
+ tbz w21, #30, 5f // 5 bits
|
|
+ // 7 bits
|
|
+ mrs x20, ICH_AP0R3_EL2
|
|
+ str w20, [x3, #(VGIC_V3_CPU_AP0R + 3*4)]
|
|
+ mrs x19, ICH_AP0R2_EL2
|
|
+ str w19, [x3, #(VGIC_V3_CPU_AP0R + 2*4)]
|
|
+6: mrs x18, ICH_AP0R1_EL2
|
|
+ str w18, [x3, #(VGIC_V3_CPU_AP0R + 1*4)]
|
|
+5: mrs x17, ICH_AP0R0_EL2
|
|
+ str w17, [x3, #VGIC_V3_CPU_AP0R]
|
|
+
|
|
+ tbnz w21, #29, 6f // 6 bits
|
|
+ tbz w21, #30, 5f // 5 bits
|
|
+ // 7 bits
|
|
+ mrs x20, ICH_AP1R3_EL2
|
|
+ str w20, [x3, #(VGIC_V3_CPU_AP1R + 3*4)]
|
|
+ mrs x19, ICH_AP1R2_EL2
|
|
+ str w19, [x3, #(VGIC_V3_CPU_AP1R + 2*4)]
|
|
+6: mrs x18, ICH_AP1R1_EL2
|
|
+ str w18, [x3, #(VGIC_V3_CPU_AP1R + 1*4)]
|
|
+5: mrs x17, ICH_AP1R0_EL2
|
|
+ str w17, [x3, #VGIC_V3_CPU_AP1R]
|
|
+
|
|
+ // Restore SRE_EL1 access and re-enable SRE at EL1.
|
|
+ mrs x5, ICC_SRE_EL2
|
|
+ orr x5, x5, #ICC_SRE_EL2_ENABLE
|
|
+ msr ICC_SRE_EL2, x5
|
|
+ isb
|
|
+ mov x5, #1
|
|
+ msr ICC_SRE_EL1, x5
|
|
+.endm
|
|
+
|
|
+/*
|
|
+ * Restore the VGIC CPU state from memory
|
|
+ * x0: Register pointing to VCPU struct
|
|
+ */
|
|
+.macro restore_vgic_v3_state
|
|
+ // Disable SRE_EL1 access. Necessary, otherwise
|
|
+ // ICH_VMCR_EL2.VFIQEn becomes one, and FIQ happens...
|
|
+ msr ICC_SRE_EL1, xzr
|
|
+ isb
|
|
+
|
|
+ // Compute the address of struct vgic_cpu
|
|
+ add x3, x0, #VCPU_VGIC_CPU
|
|
+
|
|
+ // Restore all interesting registers
|
|
+ ldr w4, [x3, #VGIC_V3_CPU_HCR]
|
|
+ ldr w5, [x3, #VGIC_V3_CPU_VMCR]
|
|
+
|
|
+ msr ICH_HCR_EL2, x4
|
|
+ msr ICH_VMCR_EL2, x5
|
|
+
|
|
+ mrs x21, ICH_VTR_EL2
|
|
+
|
|
+ tbnz w21, #29, 6f // 6 bits
|
|
+ tbz w21, #30, 5f // 5 bits
|
|
+ // 7 bits
|
|
+ ldr w20, [x3, #(VGIC_V3_CPU_AP1R + 3*4)]
|
|
+ msr ICH_AP1R3_EL2, x20
|
|
+ ldr w19, [x3, #(VGIC_V3_CPU_AP1R + 2*4)]
|
|
+ msr ICH_AP1R2_EL2, x19
|
|
+6: ldr w18, [x3, #(VGIC_V3_CPU_AP1R + 1*4)]
|
|
+ msr ICH_AP1R1_EL2, x18
|
|
+5: ldr w17, [x3, #VGIC_V3_CPU_AP1R]
|
|
+ msr ICH_AP1R0_EL2, x17
|
|
+
|
|
+ tbnz w21, #29, 6f // 6 bits
|
|
+ tbz w21, #30, 5f // 5 bits
|
|
+ // 7 bits
|
|
+ ldr w20, [x3, #(VGIC_V3_CPU_AP0R + 3*4)]
|
|
+ msr ICH_AP0R3_EL2, x20
|
|
+ ldr w19, [x3, #(VGIC_V3_CPU_AP0R + 2*4)]
|
|
+ msr ICH_AP0R2_EL2, x19
|
|
+6: ldr w18, [x3, #(VGIC_V3_CPU_AP0R + 1*4)]
|
|
+ msr ICH_AP0R1_EL2, x18
|
|
+5: ldr w17, [x3, #VGIC_V3_CPU_AP0R]
|
|
+ msr ICH_AP0R0_EL2, x17
|
|
+
|
|
+ and w22, w21, #0xf
|
|
+ mvn w22, w21
|
|
+ ubfiz w23, w22, 2, 4 // w23 = (15 - ListRegs) * 4
|
|
+
|
|
+ adr x24, 1f
|
|
+ add x24, x24, x23
|
|
+ br x24
|
|
+
|
|
+1:
|
|
+ ldr x20, [x3, #LR_OFFSET(15)]
|
|
+ ldr x19, [x3, #LR_OFFSET(14)]
|
|
+ ldr x18, [x3, #LR_OFFSET(13)]
|
|
+ ldr x17, [x3, #LR_OFFSET(12)]
|
|
+ ldr x16, [x3, #LR_OFFSET(11)]
|
|
+ ldr x15, [x3, #LR_OFFSET(10)]
|
|
+ ldr x14, [x3, #LR_OFFSET(9)]
|
|
+ ldr x13, [x3, #LR_OFFSET(8)]
|
|
+ ldr x12, [x3, #LR_OFFSET(7)]
|
|
+ ldr x11, [x3, #LR_OFFSET(6)]
|
|
+ ldr x10, [x3, #LR_OFFSET(5)]
|
|
+ ldr x9, [x3, #LR_OFFSET(4)]
|
|
+ ldr x8, [x3, #LR_OFFSET(3)]
|
|
+ ldr x7, [x3, #LR_OFFSET(2)]
|
|
+ ldr x6, [x3, #LR_OFFSET(1)]
|
|
+ ldr x5, [x3, #LR_OFFSET(0)]
|
|
+
|
|
+ adr x24, 1f
|
|
+ add x24, x24, x23
|
|
+ br x24
|
|
+
|
|
+1:
|
|
+ msr ICH_LR15_EL2, x20
|
|
+ msr ICH_LR14_EL2, x19
|
|
+ msr ICH_LR13_EL2, x18
|
|
+ msr ICH_LR12_EL2, x17
|
|
+ msr ICH_LR11_EL2, x16
|
|
+ msr ICH_LR10_EL2, x15
|
|
+ msr ICH_LR9_EL2, x14
|
|
+ msr ICH_LR8_EL2, x13
|
|
+ msr ICH_LR7_EL2, x12
|
|
+ msr ICH_LR6_EL2, x11
|
|
+ msr ICH_LR5_EL2, x10
|
|
+ msr ICH_LR4_EL2, x9
|
|
+ msr ICH_LR3_EL2, x8
|
|
+ msr ICH_LR2_EL2, x7
|
|
+ msr ICH_LR1_EL2, x6
|
|
+ msr ICH_LR0_EL2, x5
|
|
+
|
|
+ // Ensure that the above will have reached the
|
|
+ // (re)distributors. This ensure the guest will read
|
|
+ // the correct values from the memory-mapped interface.
|
|
+ isb
|
|
+ dsb sy
|
|
+
|
|
+ // Prevent the guest from touching the GIC system registers
|
|
+ mrs x5, ICC_SRE_EL2
|
|
+ and x5, x5, #~ICC_SRE_EL2_ENABLE
|
|
+ msr ICC_SRE_EL2, x5
|
|
+.endm
|
|
+
|
|
+ENTRY(__save_vgic_v3_state)
|
|
+ save_vgic_v3_state
|
|
+ ret
|
|
+ENDPROC(__save_vgic_v3_state)
|
|
+
|
|
+ENTRY(__restore_vgic_v3_state)
|
|
+ restore_vgic_v3_state
|
|
+ ret
|
|
+ENDPROC(__restore_vgic_v3_state)
|
|
+
|
|
+ENTRY(__vgic_v3_get_ich_vtr_el2)
|
|
+ mrs x0, ICH_VTR_EL2
|
|
+ ret
|
|
+ENDPROC(__vgic_v3_get_ich_vtr_el2)
|
|
+
|
|
+ .popsection
|
|
diff --git a/arch/ia64/kernel/process.c b/arch/ia64/kernel/process.c
|
|
index 55d4ba4..deed6fa 100644
|
|
--- a/arch/ia64/kernel/process.c
|
|
+++ b/arch/ia64/kernel/process.c
|
|
@@ -662,7 +662,7 @@ void
|
|
machine_restart (char *restart_cmd)
|
|
{
|
|
(void) notify_die(DIE_MACHINE_RESTART, restart_cmd, NULL, 0, 0, 0);
|
|
- (*efi.reset_system)(EFI_RESET_WARM, 0, 0, NULL);
|
|
+ efi_reboot(REBOOT_WARM, NULL);
|
|
}
|
|
|
|
void
|
|
diff --git a/arch/ia64/kernel/time.c b/arch/ia64/kernel/time.c
|
|
index 71c52bc..a149c67 100644
|
|
--- a/arch/ia64/kernel/time.c
|
|
+++ b/arch/ia64/kernel/time.c
|
|
@@ -384,21 +384,6 @@ static struct irqaction timer_irqaction = {
|
|
.name = "timer"
|
|
};
|
|
|
|
-static struct platform_device rtc_efi_dev = {
|
|
- .name = "rtc-efi",
|
|
- .id = -1,
|
|
-};
|
|
-
|
|
-static int __init rtc_init(void)
|
|
-{
|
|
- if (platform_device_register(&rtc_efi_dev) < 0)
|
|
- printk(KERN_ERR "unable to register rtc device...\n");
|
|
-
|
|
- /* not necessarily an error */
|
|
- return 0;
|
|
-}
|
|
-module_init(rtc_init);
|
|
-
|
|
void read_persistent_clock(struct timespec *ts)
|
|
{
|
|
efi_gettimeofday(ts);
|
|
diff --git a/arch/x86/include/asm/efi.h b/arch/x86/include/asm/efi.h
|
|
index 1eb5f64..5d71d0e 100644
|
|
--- a/arch/x86/include/asm/efi.h
|
|
+++ b/arch/x86/include/asm/efi.h
|
|
@@ -104,6 +104,8 @@ extern void __init runtime_code_page_mkexec(void);
|
|
extern void __init efi_runtime_mkexec(void);
|
|
extern void __init efi_dump_pagetable(void);
|
|
extern void __init efi_apply_memmap_quirks(void);
|
|
+extern int __init efi_reuse_config(u64 tables, int nr_tables);
|
|
+extern void efi_delete_dummy_variable(void);
|
|
|
|
struct efi_setup_data {
|
|
u64 fw_vendor;
|
|
@@ -156,6 +158,8 @@ static inline efi_status_t efi_thunk_set_virtual_address_map(
|
|
return EFI_SUCCESS;
|
|
}
|
|
#endif /* CONFIG_EFI_MIXED */
|
|
+
|
|
+extern bool efi_reboot_required(void);
|
|
#else
|
|
/*
|
|
* IF EFI is not configured, have the EFI calls return -ENOSYS.
|
|
@@ -168,6 +172,10 @@ static inline efi_status_t efi_thunk_set_virtual_address_map(
|
|
#define efi_call5(_f, _a1, _a2, _a3, _a4, _a5) (-ENOSYS)
|
|
#define efi_call6(_f, _a1, _a2, _a3, _a4, _a5, _a6) (-ENOSYS)
|
|
static inline void parse_efi_setup(u64 phys_addr, u32 data_len) {}
|
|
+static inline bool efi_reboot_required(void)
|
|
+{
|
|
+ return false;
|
|
+}
|
|
#endif /* CONFIG_EFI */
|
|
|
|
#endif /* _ASM_X86_EFI_H */
|
|
diff --git a/arch/x86/kernel/reboot.c b/arch/x86/kernel/reboot.c
|
|
index 52b1157..17962e6 100644
|
|
--- a/arch/x86/kernel/reboot.c
|
|
+++ b/arch/x86/kernel/reboot.c
|
|
@@ -28,6 +28,7 @@
|
|
#include <linux/mc146818rtc.h>
|
|
#include <asm/realmode.h>
|
|
#include <asm/x86_init.h>
|
|
+#include <asm/efi.h>
|
|
|
|
/*
|
|
* Power off function, if any
|
|
@@ -401,12 +402,25 @@ static struct dmi_system_id __initdata reboot_dmi_table[] = {
|
|
|
|
static int __init reboot_init(void)
|
|
{
|
|
+ int rv;
|
|
+
|
|
/*
|
|
* Only do the DMI check if reboot_type hasn't been overridden
|
|
* on the command line
|
|
*/
|
|
- if (reboot_default)
|
|
- dmi_check_system(reboot_dmi_table);
|
|
+ if (!reboot_default)
|
|
+ return 0;
|
|
+
|
|
+ /*
|
|
+ * The DMI quirks table takes precedence. If no quirks entry
|
|
+ * matches and the ACPI Hardware Reduced bit is set, force EFI
|
|
+ * reboot.
|
|
+ */
|
|
+ rv = dmi_check_system(reboot_dmi_table);
|
|
+
|
|
+ if (!rv && efi_reboot_required())
|
|
+ reboot_type = BOOT_EFI;
|
|
+
|
|
return 0;
|
|
}
|
|
core_initcall(reboot_init);
|
|
@@ -528,11 +542,7 @@ static void native_machine_emergency_restart(void)
|
|
break;
|
|
|
|
case BOOT_EFI:
|
|
- if (efi_enabled(EFI_RUNTIME_SERVICES))
|
|
- efi.reset_system(reboot_mode == REBOOT_WARM ?
|
|
- EFI_RESET_WARM :
|
|
- EFI_RESET_COLD,
|
|
- EFI_SUCCESS, 0, NULL);
|
|
+ efi_reboot(reboot_mode, NULL);
|
|
reboot_type = BOOT_BIOS;
|
|
break;
|
|
|
|
diff --git a/arch/x86/platform/efi/Makefile b/arch/x86/platform/efi/Makefile
|
|
index d51045a..2846aaa 100644
|
|
--- a/arch/x86/platform/efi/Makefile
|
|
+++ b/arch/x86/platform/efi/Makefile
|
|
@@ -1,4 +1,4 @@
|
|
-obj-$(CONFIG_EFI) += efi.o efi_$(BITS).o efi_stub_$(BITS).o
|
|
+obj-$(CONFIG_EFI) += quirks.o efi.o efi_$(BITS).o efi_stub_$(BITS).o
|
|
obj-$(CONFIG_ACPI_BGRT) += efi-bgrt.o
|
|
obj-$(CONFIG_EARLY_PRINTK_EFI) += early_printk.o
|
|
obj-$(CONFIG_EFI_MIXED) += efi_thunk_$(BITS).o
|
|
diff --git a/arch/x86/platform/efi/efi.c b/arch/x86/platform/efi/efi.c
|
|
index 87fc96b..f852443 100644
|
|
--- a/arch/x86/platform/efi/efi.c
|
|
+++ b/arch/x86/platform/efi/efi.c
|
|
@@ -56,13 +56,6 @@
|
|
|
|
#define EFI_DEBUG
|
|
|
|
-#define EFI_MIN_RESERVE 5120
|
|
-
|
|
-#define EFI_DUMMY_GUID \
|
|
- EFI_GUID(0x4424ac57, 0xbe4b, 0x47dd, 0x9e, 0x97, 0xed, 0x50, 0xf0, 0x9f, 0x92, 0xa9)
|
|
-
|
|
-static efi_char16_t efi_dummy_name[6] = { 'D', 'U', 'M', 'M', 'Y', 0 };
|
|
-
|
|
struct efi_memory_map memmap;
|
|
|
|
static struct efi efi_phys __initdata;
|
|
@@ -95,15 +88,6 @@ static int __init setup_add_efi_memmap(char *arg)
|
|
}
|
|
early_param("add_efi_memmap", setup_add_efi_memmap);
|
|
|
|
-static bool efi_no_storage_paranoia;
|
|
-
|
|
-static int __init setup_storage_paranoia(char *arg)
|
|
-{
|
|
- efi_no_storage_paranoia = true;
|
|
- return 0;
|
|
-}
|
|
-early_param("efi_no_storage_paranoia", setup_storage_paranoia);
|
|
-
|
|
static efi_status_t virt_efi_get_time(efi_time_t *tm, efi_time_cap_t *tc)
|
|
{
|
|
unsigned long flags;
|
|
@@ -392,37 +376,6 @@ static void __init print_efi_memmap(void)
|
|
#endif /* EFI_DEBUG */
|
|
}
|
|
|
|
-void __init efi_reserve_boot_services(void)
|
|
-{
|
|
- void *p;
|
|
-
|
|
- for (p = memmap.map; p < memmap.map_end; p += memmap.desc_size) {
|
|
- efi_memory_desc_t *md = p;
|
|
- u64 start = md->phys_addr;
|
|
- u64 size = md->num_pages << EFI_PAGE_SHIFT;
|
|
-
|
|
- if (md->type != EFI_BOOT_SERVICES_CODE &&
|
|
- md->type != EFI_BOOT_SERVICES_DATA)
|
|
- continue;
|
|
- /* Only reserve where possible:
|
|
- * - Not within any already allocated areas
|
|
- * - Not over any memory area (really needed, if above?)
|
|
- * - Not within any part of the kernel
|
|
- * - Not the bios reserved area
|
|
- */
|
|
- if ((start + size > __pa_symbol(_text)
|
|
- && start <= __pa_symbol(_end)) ||
|
|
- !e820_all_mapped(start, start+size, E820_RAM) ||
|
|
- memblock_is_region_reserved(start, size)) {
|
|
- /* Could not reserve, skip it */
|
|
- md->num_pages = 0;
|
|
- memblock_dbg("Could not reserve boot range [0x%010llx-0x%010llx]\n",
|
|
- start, start+size-1);
|
|
- } else
|
|
- memblock_reserve(start, size);
|
|
- }
|
|
-}
|
|
-
|
|
void __init efi_unmap_memmap(void)
|
|
{
|
|
clear_bit(EFI_MEMMAP, &efi.flags);
|
|
@@ -432,29 +385,6 @@ void __init efi_unmap_memmap(void)
|
|
}
|
|
}
|
|
|
|
-void __init efi_free_boot_services(void)
|
|
-{
|
|
- void *p;
|
|
-
|
|
- for (p = memmap.map; p < memmap.map_end; p += memmap.desc_size) {
|
|
- efi_memory_desc_t *md = p;
|
|
- unsigned long long start = md->phys_addr;
|
|
- unsigned long long size = md->num_pages << EFI_PAGE_SHIFT;
|
|
-
|
|
- if (md->type != EFI_BOOT_SERVICES_CODE &&
|
|
- md->type != EFI_BOOT_SERVICES_DATA)
|
|
- continue;
|
|
-
|
|
- /* Could not reserve boot area */
|
|
- if (!size)
|
|
- continue;
|
|
-
|
|
- free_bootmem_late(start, size);
|
|
- }
|
|
-
|
|
- efi_unmap_memmap();
|
|
-}
|
|
-
|
|
static int __init efi_systab_init(void *phys)
|
|
{
|
|
if (efi_enabled(EFI_64BIT)) {
|
|
@@ -649,62 +579,6 @@ static int __init efi_memmap_init(void)
|
|
return 0;
|
|
}
|
|
|
|
-/*
|
|
- * A number of config table entries get remapped to virtual addresses
|
|
- * after entering EFI virtual mode. However, the kexec kernel requires
|
|
- * their physical addresses therefore we pass them via setup_data and
|
|
- * correct those entries to their respective physical addresses here.
|
|
- *
|
|
- * Currently only handles smbios which is necessary for some firmware
|
|
- * implementation.
|
|
- */
|
|
-static int __init efi_reuse_config(u64 tables, int nr_tables)
|
|
-{
|
|
- int i, sz, ret = 0;
|
|
- void *p, *tablep;
|
|
- struct efi_setup_data *data;
|
|
-
|
|
- if (!efi_setup)
|
|
- return 0;
|
|
-
|
|
- if (!efi_enabled(EFI_64BIT))
|
|
- return 0;
|
|
-
|
|
- data = early_memremap(efi_setup, sizeof(*data));
|
|
- if (!data) {
|
|
- ret = -ENOMEM;
|
|
- goto out;
|
|
- }
|
|
-
|
|
- if (!data->smbios)
|
|
- goto out_memremap;
|
|
-
|
|
- sz = sizeof(efi_config_table_64_t);
|
|
-
|
|
- p = tablep = early_memremap(tables, nr_tables * sz);
|
|
- if (!p) {
|
|
- pr_err("Could not map Configuration table!\n");
|
|
- ret = -ENOMEM;
|
|
- goto out_memremap;
|
|
- }
|
|
-
|
|
- for (i = 0; i < efi.systab->nr_tables; i++) {
|
|
- efi_guid_t guid;
|
|
-
|
|
- guid = ((efi_config_table_64_t *)p)->guid;
|
|
-
|
|
- if (!efi_guidcmp(guid, SMBIOS_TABLE_GUID))
|
|
- ((efi_config_table_64_t *)p)->table = data->smbios;
|
|
- p += sz;
|
|
- }
|
|
- early_iounmap(tablep, nr_tables * sz);
|
|
-
|
|
-out_memremap:
|
|
- early_iounmap(data, sizeof(*data));
|
|
-out:
|
|
- return ret;
|
|
-}
|
|
-
|
|
void __init efi_init(void)
|
|
{
|
|
efi_char16_t *c16;
|
|
@@ -1057,11 +931,7 @@ static void __init kexec_enter_virtual_mode(void)
|
|
runtime_code_page_mkexec();
|
|
|
|
/* clean DUMMY object */
|
|
- efi.set_variable(efi_dummy_name, &EFI_DUMMY_GUID,
|
|
- EFI_VARIABLE_NON_VOLATILE |
|
|
- EFI_VARIABLE_BOOTSERVICE_ACCESS |
|
|
- EFI_VARIABLE_RUNTIME_ACCESS,
|
|
- 0, NULL);
|
|
+ efi_delete_dummy_variable();
|
|
#endif
|
|
}
|
|
|
|
@@ -1179,11 +1049,7 @@ static void __init __efi_enter_virtual_mode(void)
|
|
free_pages((unsigned long)new_memmap, pg_shift);
|
|
|
|
/* clean DUMMY object */
|
|
- efi.set_variable(efi_dummy_name, &EFI_DUMMY_GUID,
|
|
- EFI_VARIABLE_NON_VOLATILE |
|
|
- EFI_VARIABLE_BOOTSERVICE_ACCESS |
|
|
- EFI_VARIABLE_RUNTIME_ACCESS,
|
|
- 0, NULL);
|
|
+ efi_delete_dummy_variable();
|
|
}
|
|
|
|
void __init efi_enter_virtual_mode(void)
|
|
@@ -1230,86 +1096,6 @@ u64 efi_mem_attributes(unsigned long phys_addr)
|
|
return 0;
|
|
}
|
|
|
|
-/*
|
|
- * Some firmware implementations refuse to boot if there's insufficient space
|
|
- * in the variable store. Ensure that we never use more than a safe limit.
|
|
- *
|
|
- * Return EFI_SUCCESS if it is safe to write 'size' bytes to the variable
|
|
- * store.
|
|
- */
|
|
-efi_status_t efi_query_variable_store(u32 attributes, unsigned long size)
|
|
-{
|
|
- efi_status_t status;
|
|
- u64 storage_size, remaining_size, max_size;
|
|
-
|
|
- if (!(attributes & EFI_VARIABLE_NON_VOLATILE))
|
|
- return 0;
|
|
-
|
|
- status = efi.query_variable_info(attributes, &storage_size,
|
|
- &remaining_size, &max_size);
|
|
- if (status != EFI_SUCCESS)
|
|
- return status;
|
|
-
|
|
- /*
|
|
- * We account for that by refusing the write if permitting it would
|
|
- * reduce the available space to under 5KB. This figure was provided by
|
|
- * Samsung, so should be safe.
|
|
- */
|
|
- if ((remaining_size - size < EFI_MIN_RESERVE) &&
|
|
- !efi_no_storage_paranoia) {
|
|
-
|
|
- /*
|
|
- * Triggering garbage collection may require that the firmware
|
|
- * generate a real EFI_OUT_OF_RESOURCES error. We can force
|
|
- * that by attempting to use more space than is available.
|
|
- */
|
|
- unsigned long dummy_size = remaining_size + 1024;
|
|
- void *dummy = kzalloc(dummy_size, GFP_ATOMIC);
|
|
-
|
|
- if (!dummy)
|
|
- return EFI_OUT_OF_RESOURCES;
|
|
-
|
|
- status = efi.set_variable(efi_dummy_name, &EFI_DUMMY_GUID,
|
|
- EFI_VARIABLE_NON_VOLATILE |
|
|
- EFI_VARIABLE_BOOTSERVICE_ACCESS |
|
|
- EFI_VARIABLE_RUNTIME_ACCESS,
|
|
- dummy_size, dummy);
|
|
-
|
|
- if (status == EFI_SUCCESS) {
|
|
- /*
|
|
- * This should have failed, so if it didn't make sure
|
|
- * that we delete it...
|
|
- */
|
|
- efi.set_variable(efi_dummy_name, &EFI_DUMMY_GUID,
|
|
- EFI_VARIABLE_NON_VOLATILE |
|
|
- EFI_VARIABLE_BOOTSERVICE_ACCESS |
|
|
- EFI_VARIABLE_RUNTIME_ACCESS,
|
|
- 0, dummy);
|
|
- }
|
|
-
|
|
- kfree(dummy);
|
|
-
|
|
- /*
|
|
- * The runtime code may now have triggered a garbage collection
|
|
- * run, so check the variable info again
|
|
- */
|
|
- status = efi.query_variable_info(attributes, &storage_size,
|
|
- &remaining_size, &max_size);
|
|
-
|
|
- if (status != EFI_SUCCESS)
|
|
- return status;
|
|
-
|
|
- /*
|
|
- * There still isn't enough room, so return an error
|
|
- */
|
|
- if (remaining_size - size < EFI_MIN_RESERVE)
|
|
- return EFI_OUT_OF_RESOURCES;
|
|
- }
|
|
-
|
|
- return EFI_SUCCESS;
|
|
-}
|
|
-EXPORT_SYMBOL_GPL(efi_query_variable_store);
|
|
-
|
|
static int __init parse_efi_cmdline(char *str)
|
|
{
|
|
if (*str == '=')
|
|
@@ -1321,22 +1107,3 @@ static int __init parse_efi_cmdline(char *str)
|
|
return 0;
|
|
}
|
|
early_param("efi", parse_efi_cmdline);
|
|
-
|
|
-void __init efi_apply_memmap_quirks(void)
|
|
-{
|
|
- /*
|
|
- * Once setup is done earlier, unmap the EFI memory map on mismatched
|
|
- * firmware/kernel architectures since there is no support for runtime
|
|
- * services.
|
|
- */
|
|
- if (!efi_runtime_supported()) {
|
|
- pr_info("efi: Setup done, disabling due to 32/64-bit mismatch\n");
|
|
- efi_unmap_memmap();
|
|
- }
|
|
-
|
|
- /*
|
|
- * UV doesn't support the new EFI pagetable mapping yet.
|
|
- */
|
|
- if (is_uv_system())
|
|
- set_bit(EFI_OLD_MEMMAP, &efi.flags);
|
|
-}
|
|
diff --git a/arch/x86/platform/efi/quirks.c b/arch/x86/platform/efi/quirks.c
|
|
new file mode 100644
|
|
index 0000000..1b9c4c3
|
|
--- /dev/null
|
|
+++ b/arch/x86/platform/efi/quirks.c
|
|
@@ -0,0 +1,290 @@
|
|
+#include <linux/init.h>
|
|
+#include <linux/kernel.h>
|
|
+#include <linux/string.h>
|
|
+#include <linux/time.h>
|
|
+#include <linux/types.h>
|
|
+#include <linux/efi.h>
|
|
+#include <linux/slab.h>
|
|
+#include <linux/memblock.h>
|
|
+#include <linux/bootmem.h>
|
|
+#include <linux/acpi.h>
|
|
+#include <asm/efi.h>
|
|
+#include <asm/uv/uv.h>
|
|
+
|
|
+#define EFI_MIN_RESERVE 5120
|
|
+
|
|
+#define EFI_DUMMY_GUID \
|
|
+ EFI_GUID(0x4424ac57, 0xbe4b, 0x47dd, 0x9e, 0x97, 0xed, 0x50, 0xf0, 0x9f, 0x92, 0xa9)
|
|
+
|
|
+static efi_char16_t efi_dummy_name[6] = { 'D', 'U', 'M', 'M', 'Y', 0 };
|
|
+
|
|
+static bool efi_no_storage_paranoia;
|
|
+
|
|
+/*
|
|
+ * Some firmware implementations refuse to boot if there's insufficient
|
|
+ * space in the variable store. The implementation of garbage collection
|
|
+ * in some FW versions causes stale (deleted) variables to take up space
|
|
+ * longer than intended and space is only freed once the store becomes
|
|
+ * almost completely full.
|
|
+ *
|
|
+ * Enabling this option disables the space checks in
|
|
+ * efi_query_variable_store() and forces garbage collection.
|
|
+ *
|
|
+ * Only enable this option if deleting EFI variables does not free up
|
|
+ * space in your variable store, e.g. if despite deleting variables
|
|
+ * you're unable to create new ones.
|
|
+ */
|
|
+static int __init setup_storage_paranoia(char *arg)
|
|
+{
|
|
+ efi_no_storage_paranoia = true;
|
|
+ return 0;
|
|
+}
|
|
+early_param("efi_no_storage_paranoia", setup_storage_paranoia);
|
|
+
|
|
+/*
|
|
+ * Deleting the dummy variable which kicks off garbage collection
|
|
+*/
|
|
+void efi_delete_dummy_variable(void)
|
|
+{
|
|
+ efi.set_variable(efi_dummy_name, &EFI_DUMMY_GUID,
|
|
+ EFI_VARIABLE_NON_VOLATILE |
|
|
+ EFI_VARIABLE_BOOTSERVICE_ACCESS |
|
|
+ EFI_VARIABLE_RUNTIME_ACCESS,
|
|
+ 0, NULL);
|
|
+}
|
|
+
|
|
+/*
|
|
+ * Some firmware implementations refuse to boot if there's insufficient space
|
|
+ * in the variable store. Ensure that we never use more than a safe limit.
|
|
+ *
|
|
+ * Return EFI_SUCCESS if it is safe to write 'size' bytes to the variable
|
|
+ * store.
|
|
+ */
|
|
+efi_status_t efi_query_variable_store(u32 attributes, unsigned long size)
|
|
+{
|
|
+ efi_status_t status;
|
|
+ u64 storage_size, remaining_size, max_size;
|
|
+
|
|
+ if (!(attributes & EFI_VARIABLE_NON_VOLATILE))
|
|
+ return 0;
|
|
+
|
|
+ status = efi.query_variable_info(attributes, &storage_size,
|
|
+ &remaining_size, &max_size);
|
|
+ if (status != EFI_SUCCESS)
|
|
+ return status;
|
|
+
|
|
+ /*
|
|
+ * We account for that by refusing the write if permitting it would
|
|
+ * reduce the available space to under 5KB. This figure was provided by
|
|
+ * Samsung, so should be safe.
|
|
+ */
|
|
+ if ((remaining_size - size < EFI_MIN_RESERVE) &&
|
|
+ !efi_no_storage_paranoia) {
|
|
+
|
|
+ /*
|
|
+ * Triggering garbage collection may require that the firmware
|
|
+ * generate a real EFI_OUT_OF_RESOURCES error. We can force
|
|
+ * that by attempting to use more space than is available.
|
|
+ */
|
|
+ unsigned long dummy_size = remaining_size + 1024;
|
|
+ void *dummy = kzalloc(dummy_size, GFP_ATOMIC);
|
|
+
|
|
+ if (!dummy)
|
|
+ return EFI_OUT_OF_RESOURCES;
|
|
+
|
|
+ status = efi.set_variable(efi_dummy_name, &EFI_DUMMY_GUID,
|
|
+ EFI_VARIABLE_NON_VOLATILE |
|
|
+ EFI_VARIABLE_BOOTSERVICE_ACCESS |
|
|
+ EFI_VARIABLE_RUNTIME_ACCESS,
|
|
+ dummy_size, dummy);
|
|
+
|
|
+ if (status == EFI_SUCCESS) {
|
|
+ /*
|
|
+ * This should have failed, so if it didn't make sure
|
|
+ * that we delete it...
|
|
+ */
|
|
+ efi_delete_dummy_variable();
|
|
+ }
|
|
+
|
|
+ kfree(dummy);
|
|
+
|
|
+ /*
|
|
+ * The runtime code may now have triggered a garbage collection
|
|
+ * run, so check the variable info again
|
|
+ */
|
|
+ status = efi.query_variable_info(attributes, &storage_size,
|
|
+ &remaining_size, &max_size);
|
|
+
|
|
+ if (status != EFI_SUCCESS)
|
|
+ return status;
|
|
+
|
|
+ /*
|
|
+ * There still isn't enough room, so return an error
|
|
+ */
|
|
+ if (remaining_size - size < EFI_MIN_RESERVE)
|
|
+ return EFI_OUT_OF_RESOURCES;
|
|
+ }
|
|
+
|
|
+ return EFI_SUCCESS;
|
|
+}
|
|
+EXPORT_SYMBOL_GPL(efi_query_variable_store);
|
|
+
|
|
+/*
|
|
+ * The UEFI specification makes it clear that the operating system is free to do
|
|
+ * whatever it wants with boot services code after ExitBootServices() has been
|
|
+ * called. Ignoring this recommendation a significant bunch of EFI implementations
|
|
+ * continue calling into boot services code (SetVirtualAddressMap). In order to
|
|
+ * work around such buggy implementations we reserve boot services region during
|
|
+ * EFI init and make sure it stays executable. Then, after SetVirtualAddressMap(), it
|
|
+* is discarded.
|
|
+*/
|
|
+void __init efi_reserve_boot_services(void)
|
|
+{
|
|
+ void *p;
|
|
+
|
|
+ for (p = memmap.map; p < memmap.map_end; p += memmap.desc_size) {
|
|
+ efi_memory_desc_t *md = p;
|
|
+ u64 start = md->phys_addr;
|
|
+ u64 size = md->num_pages << EFI_PAGE_SHIFT;
|
|
+
|
|
+ if (md->type != EFI_BOOT_SERVICES_CODE &&
|
|
+ md->type != EFI_BOOT_SERVICES_DATA)
|
|
+ continue;
|
|
+ /* Only reserve where possible:
|
|
+ * - Not within any already allocated areas
|
|
+ * - Not over any memory area (really needed, if above?)
|
|
+ * - Not within any part of the kernel
|
|
+ * - Not the bios reserved area
|
|
+ */
|
|
+ if ((start + size > __pa_symbol(_text)
|
|
+ && start <= __pa_symbol(_end)) ||
|
|
+ !e820_all_mapped(start, start+size, E820_RAM) ||
|
|
+ memblock_is_region_reserved(start, size)) {
|
|
+ /* Could not reserve, skip it */
|
|
+ md->num_pages = 0;
|
|
+ memblock_dbg("Could not reserve boot range [0x%010llx-0x%010llx]\n",
|
|
+ start, start+size-1);
|
|
+ } else
|
|
+ memblock_reserve(start, size);
|
|
+ }
|
|
+}
|
|
+
|
|
+void __init efi_free_boot_services(void)
|
|
+{
|
|
+ void *p;
|
|
+
|
|
+ for (p = memmap.map; p < memmap.map_end; p += memmap.desc_size) {
|
|
+ efi_memory_desc_t *md = p;
|
|
+ unsigned long long start = md->phys_addr;
|
|
+ unsigned long long size = md->num_pages << EFI_PAGE_SHIFT;
|
|
+
|
|
+ if (md->type != EFI_BOOT_SERVICES_CODE &&
|
|
+ md->type != EFI_BOOT_SERVICES_DATA)
|
|
+ continue;
|
|
+
|
|
+ /* Could not reserve boot area */
|
|
+ if (!size)
|
|
+ continue;
|
|
+
|
|
+ free_bootmem_late(start, size);
|
|
+ }
|
|
+
|
|
+ efi_unmap_memmap();
|
|
+}
|
|
+
|
|
+/*
|
|
+ * A number of config table entries get remapped to virtual addresses
|
|
+ * after entering EFI virtual mode. However, the kexec kernel requires
|
|
+ * their physical addresses therefore we pass them via setup_data and
|
|
+ * correct those entries to their respective physical addresses here.
|
|
+ *
|
|
+ * Currently only handles smbios which is necessary for some firmware
|
|
+ * implementation.
|
|
+ */
|
|
+int __init efi_reuse_config(u64 tables, int nr_tables)
|
|
+{
|
|
+ int i, sz, ret = 0;
|
|
+ void *p, *tablep;
|
|
+ struct efi_setup_data *data;
|
|
+
|
|
+ if (!efi_setup)
|
|
+ return 0;
|
|
+
|
|
+ if (!efi_enabled(EFI_64BIT))
|
|
+ return 0;
|
|
+
|
|
+ data = early_memremap(efi_setup, sizeof(*data));
|
|
+ if (!data) {
|
|
+ ret = -ENOMEM;
|
|
+ goto out;
|
|
+ }
|
|
+
|
|
+ if (!data->smbios)
|
|
+ goto out_memremap;
|
|
+
|
|
+ sz = sizeof(efi_config_table_64_t);
|
|
+
|
|
+ p = tablep = early_memremap(tables, nr_tables * sz);
|
|
+ if (!p) {
|
|
+ pr_err("Could not map Configuration table!\n");
|
|
+ ret = -ENOMEM;
|
|
+ goto out_memremap;
|
|
+ }
|
|
+
|
|
+ for (i = 0; i < efi.systab->nr_tables; i++) {
|
|
+ efi_guid_t guid;
|
|
+
|
|
+ guid = ((efi_config_table_64_t *)p)->guid;
|
|
+
|
|
+ if (!efi_guidcmp(guid, SMBIOS_TABLE_GUID))
|
|
+ ((efi_config_table_64_t *)p)->table = data->smbios;
|
|
+ p += sz;
|
|
+ }
|
|
+ early_iounmap(tablep, nr_tables * sz);
|
|
+
|
|
+out_memremap:
|
|
+ early_iounmap(data, sizeof(*data));
|
|
+out:
|
|
+ return ret;
|
|
+}
|
|
+
|
|
+void __init efi_apply_memmap_quirks(void)
|
|
+{
|
|
+ /*
|
|
+ * Once setup is done earlier, unmap the EFI memory map on mismatched
|
|
+ * firmware/kernel architectures since there is no support for runtime
|
|
+ * services.
|
|
+ */
|
|
+ if (!efi_runtime_supported()) {
|
|
+ pr_info("efi: Setup done, disabling due to 32/64-bit mismatch\n");
|
|
+ efi_unmap_memmap();
|
|
+ }
|
|
+
|
|
+ /*
|
|
+ * UV doesn't support the new EFI pagetable mapping yet.
|
|
+ */
|
|
+ if (is_uv_system())
|
|
+ set_bit(EFI_OLD_MEMMAP, &efi.flags);
|
|
+}
|
|
+
|
|
+/*
|
|
+ * For most modern platforms the preferred method of powering off is via
|
|
+ * ACPI. However, there are some that are known to require the use of
|
|
+ * EFI runtime services and for which ACPI does not work at all.
|
|
+ *
|
|
+ * Using EFI is a last resort, to be used only if no other option
|
|
+ * exists.
|
|
+ */
|
|
+bool efi_reboot_required(void)
|
|
+{
|
|
+ if (!acpi_gbl_reduced_hardware)
|
|
+ return false;
|
|
+
|
|
+ efi_reboot_quirk_mode = EFI_RESET_WARM;
|
|
+ return true;
|
|
+}
|
|
+
|
|
+bool efi_poweroff_required(void)
|
|
+{
|
|
+ return !!acpi_gbl_reduced_hardware;
|
|
+}
|
|
diff --git a/drivers/ata/ahci_xgene.c b/drivers/ata/ahci_xgene.c
|
|
index ee3a365..f9431b4 100644
|
|
--- a/drivers/ata/ahci_xgene.c
|
|
+++ b/drivers/ata/ahci_xgene.c
|
|
@@ -131,7 +131,8 @@ static unsigned int xgene_ahci_qc_issue(struct ata_queued_cmd *qc)
|
|
struct xgene_ahci_context *ctx = hpriv->plat_data;
|
|
int rc = 0;
|
|
|
|
- if (unlikely(ctx->last_cmd[ap->port_no] == ATA_CMD_ID_ATA))
|
|
+ if (unlikely(ctx->last_cmd[ap->port_no] == ATA_CMD_ID_ATA ||
|
|
+ ctx->last_cmd[ap->port_no] == ATA_CMD_SMART))
|
|
xgene_ahci_restart_engine(ap);
|
|
|
|
rc = ahci_qc_issue(qc);
|
|
diff --git a/drivers/firmware/efi/Makefile b/drivers/firmware/efi/Makefile
|
|
index 9553496..c135154 100644
|
|
--- a/drivers/firmware/efi/Makefile
|
|
+++ b/drivers/firmware/efi/Makefile
|
|
@@ -1,7 +1,7 @@
|
|
#
|
|
# Makefile for linux kernel
|
|
#
|
|
-obj-$(CONFIG_EFI) += efi.o vars.o
|
|
+obj-$(CONFIG_EFI) += efi.o vars.o reboot.o
|
|
obj-$(CONFIG_EFI_VARS) += efivars.o
|
|
obj-$(CONFIG_EFI_VARS_PSTORE) += efi-pstore.o
|
|
obj-$(CONFIG_UEFI_CPER) += cper.o
|
|
diff --git a/drivers/firmware/efi/reboot.c b/drivers/firmware/efi/reboot.c
|
|
new file mode 100644
|
|
index 0000000..f94fb95
|
|
--- /dev/null
|
|
+++ b/drivers/firmware/efi/reboot.c
|
|
@@ -0,0 +1,55 @@
|
|
+/*
|
|
+ * Copyright (C) 2014 Intel Corporation; author Matt Fleming
|
|
+ */
|
|
+#include <linux/efi.h>
|
|
+#include <linux/reboot.h>
|
|
+
|
|
+int efi_reboot_quirk_mode = -1;
|
|
+
|
|
+void efi_reboot(enum reboot_mode reboot_mode, const char *__unused)
|
|
+{
|
|
+ int efi_mode;
|
|
+
|
|
+ if (!efi_enabled(EFI_RUNTIME_SERVICES))
|
|
+ return;
|
|
+
|
|
+ switch (reboot_mode) {
|
|
+ case REBOOT_WARM:
|
|
+ case REBOOT_SOFT:
|
|
+ efi_mode = EFI_RESET_WARM;
|
|
+ break;
|
|
+ default:
|
|
+ efi_mode = EFI_RESET_COLD;
|
|
+ break;
|
|
+ }
|
|
+
|
|
+ /*
|
|
+ * If a quirk forced an EFI reset mode, always use that.
|
|
+ */
|
|
+ if (efi_reboot_quirk_mode != -1)
|
|
+ efi_mode = efi_reboot_quirk_mode;
|
|
+
|
|
+ efi.reset_system(efi_mode, EFI_SUCCESS, 0, NULL);
|
|
+}
|
|
+
|
|
+bool __weak efi_poweroff_required(void)
|
|
+{
|
|
+ return false;
|
|
+}
|
|
+
|
|
+static void efi_power_off(void)
|
|
+{
|
|
+ efi.reset_system(EFI_RESET_SHUTDOWN, EFI_SUCCESS, 0, NULL);
|
|
+}
|
|
+
|
|
+static int __init efi_shutdown_init(void)
|
|
+{
|
|
+ if (!efi_enabled(EFI_RUNTIME_SERVICES))
|
|
+ return -ENODEV;
|
|
+
|
|
+ if (efi_poweroff_required())
|
|
+ pm_power_off = efi_power_off;
|
|
+
|
|
+ return 0;
|
|
+}
|
|
+late_initcall(efi_shutdown_init);
|
|
diff --git a/drivers/irqchip/Kconfig b/drivers/irqchip/Kconfig
|
|
index bbb746e..7f0c2a3 100644
|
|
--- a/drivers/irqchip/Kconfig
|
|
+++ b/drivers/irqchip/Kconfig
|
|
@@ -10,6 +10,11 @@ config ARM_GIC
|
|
config GIC_NON_BANKED
|
|
bool
|
|
|
|
+config ARM_GIC_V3
|
|
+ bool
|
|
+ select IRQ_DOMAIN
|
|
+ select MULTI_IRQ_HANDLER
|
|
+
|
|
config ARM_NVIC
|
|
bool
|
|
select IRQ_DOMAIN
|
|
diff --git a/drivers/irqchip/Makefile b/drivers/irqchip/Makefile
|
|
index 62a13e5..c57e642 100644
|
|
--- a/drivers/irqchip/Makefile
|
|
+++ b/drivers/irqchip/Makefile
|
|
@@ -15,7 +15,8 @@ obj-$(CONFIG_ORION_IRQCHIP) += irq-orion.o
|
|
obj-$(CONFIG_ARCH_SUNXI) += irq-sun4i.o
|
|
obj-$(CONFIG_ARCH_SUNXI) += irq-sunxi-nmi.o
|
|
obj-$(CONFIG_ARCH_SPEAR3XX) += spear-shirq.o
|
|
-obj-$(CONFIG_ARM_GIC) += irq-gic.o
|
|
+obj-$(CONFIG_ARM_GIC) += irq-gic.o irq-gic-common.o
|
|
+obj-$(CONFIG_ARM_GIC_V3) += irq-gic-v3.o irq-gic-common.o
|
|
obj-$(CONFIG_ARM_NVIC) += irq-nvic.o
|
|
obj-$(CONFIG_ARM_VIC) += irq-vic.o
|
|
obj-$(CONFIG_IMGPDC_IRQ) += irq-imgpdc.o
|
|
diff --git a/drivers/irqchip/irq-gic-common.c b/drivers/irqchip/irq-gic-common.c
|
|
new file mode 100644
|
|
index 0000000..60ac704
|
|
--- /dev/null
|
|
+++ b/drivers/irqchip/irq-gic-common.c
|
|
@@ -0,0 +1,115 @@
|
|
+/*
|
|
+ * Copyright (C) 2002 ARM Limited, All Rights Reserved.
|
|
+ *
|
|
+ * This program is free software; you can redistribute it and/or modify
|
|
+ * it under the terms of the GNU General Public License version 2 as
|
|
+ * published by the Free Software Foundation.
|
|
+ *
|
|
+ * This program is distributed in the hope that it will be useful,
|
|
+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
+ * GNU General Public License for more details.
|
|
+ *
|
|
+ * You should have received a copy of the GNU General Public License
|
|
+ * along with this program. If not, see <http://www.gnu.org/licenses/>.
|
|
+ */
|
|
+
|
|
+#include <linux/interrupt.h>
|
|
+#include <linux/io.h>
|
|
+#include <linux/irq.h>
|
|
+#include <linux/irqchip/arm-gic.h>
|
|
+
|
|
+#include "irq-gic-common.h"
|
|
+
|
|
+void gic_configure_irq(unsigned int irq, unsigned int type,
|
|
+ void __iomem *base, void (*sync_access)(void))
|
|
+{
|
|
+ u32 enablemask = 1 << (irq % 32);
|
|
+ u32 enableoff = (irq / 32) * 4;
|
|
+ u32 confmask = 0x2 << ((irq % 16) * 2);
|
|
+ u32 confoff = (irq / 16) * 4;
|
|
+ bool enabled = false;
|
|
+ u32 val;
|
|
+
|
|
+ /*
|
|
+ * Read current configuration register, and insert the config
|
|
+ * for "irq", depending on "type".
|
|
+ */
|
|
+ val = readl_relaxed(base + GIC_DIST_CONFIG + confoff);
|
|
+ if (type == IRQ_TYPE_LEVEL_HIGH)
|
|
+ val &= ~confmask;
|
|
+ else if (type == IRQ_TYPE_EDGE_RISING)
|
|
+ val |= confmask;
|
|
+
|
|
+ /*
|
|
+ * As recommended by the spec, disable the interrupt before changing
|
|
+ * the configuration
|
|
+ */
|
|
+ if (readl_relaxed(base + GIC_DIST_ENABLE_SET + enableoff) & enablemask) {
|
|
+ writel_relaxed(enablemask, base + GIC_DIST_ENABLE_CLEAR + enableoff);
|
|
+ if (sync_access)
|
|
+ sync_access();
|
|
+ enabled = true;
|
|
+ }
|
|
+
|
|
+ /*
|
|
+ * Write back the new configuration, and possibly re-enable
|
|
+ * the interrupt.
|
|
+ */
|
|
+ writel_relaxed(val, base + GIC_DIST_CONFIG + confoff);
|
|
+
|
|
+ if (enabled)
|
|
+ writel_relaxed(enablemask, base + GIC_DIST_ENABLE_SET + enableoff);
|
|
+
|
|
+ if (sync_access)
|
|
+ sync_access();
|
|
+}
|
|
+
|
|
+void __init gic_dist_config(void __iomem *base, int gic_irqs,
|
|
+ void (*sync_access)(void))
|
|
+{
|
|
+ unsigned int i;
|
|
+
|
|
+ /*
|
|
+ * Set all global interrupts to be level triggered, active low.
|
|
+ */
|
|
+ for (i = 32; i < gic_irqs; i += 16)
|
|
+ writel_relaxed(0, base + GIC_DIST_CONFIG + i / 4);
|
|
+
|
|
+ /*
|
|
+ * Set priority on all global interrupts.
|
|
+ */
|
|
+ for (i = 32; i < gic_irqs; i += 4)
|
|
+ writel_relaxed(0xa0a0a0a0, base + GIC_DIST_PRI + i);
|
|
+
|
|
+ /*
|
|
+ * Disable all interrupts. Leave the PPI and SGIs alone
|
|
+ * as they are enabled by redistributor registers.
|
|
+ */
|
|
+ for (i = 32; i < gic_irqs; i += 32)
|
|
+ writel_relaxed(0xffffffff, base + GIC_DIST_ENABLE_CLEAR + i / 8);
|
|
+
|
|
+ if (sync_access)
|
|
+ sync_access();
|
|
+}
|
|
+
|
|
+void gic_cpu_config(void __iomem *base, void (*sync_access)(void))
|
|
+{
|
|
+ int i;
|
|
+
|
|
+ /*
|
|
+ * Deal with the banked PPI and SGI interrupts - disable all
|
|
+ * PPI interrupts, ensure all SGI interrupts are enabled.
|
|
+ */
|
|
+ writel_relaxed(0xffff0000, base + GIC_DIST_ENABLE_CLEAR);
|
|
+ writel_relaxed(0x0000ffff, base + GIC_DIST_ENABLE_SET);
|
|
+
|
|
+ /*
|
|
+ * Set priority on PPI and SGI interrupts
|
|
+ */
|
|
+ for (i = 0; i < 32; i += 4)
|
|
+ writel_relaxed(0xa0a0a0a0, base + GIC_DIST_PRI + i * 4 / 4);
|
|
+
|
|
+ if (sync_access)
|
|
+ sync_access();
|
|
+}
|
|
diff --git a/drivers/irqchip/irq-gic-common.h b/drivers/irqchip/irq-gic-common.h
|
|
new file mode 100644
|
|
index 0000000..b41f024
|
|
--- /dev/null
|
|
+++ b/drivers/irqchip/irq-gic-common.h
|
|
@@ -0,0 +1,29 @@
|
|
+/*
|
|
+ * Copyright (C) 2002 ARM Limited, All Rights Reserved.
|
|
+ *
|
|
+ * This program is free software; you can redistribute it and/or modify
|
|
+ * it under the terms of the GNU General Public License version 2 as
|
|
+ * published by the Free Software Foundation.
|
|
+ *
|
|
+ * This program is distributed in the hope that it will be useful,
|
|
+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
+ * GNU General Public License for more details.
|
|
+ *
|
|
+ * You should have received a copy of the GNU General Public License
|
|
+ * along with this program. If not, see <http://www.gnu.org/licenses/>.
|
|
+ */
|
|
+
|
|
+#ifndef _IRQ_GIC_COMMON_H
|
|
+#define _IRQ_GIC_COMMON_H
|
|
+
|
|
+#include <linux/of.h>
|
|
+#include <linux/irqdomain.h>
|
|
+
|
|
+void gic_configure_irq(unsigned int irq, unsigned int type,
|
|
+ void __iomem *base, void (*sync_access)(void));
|
|
+void gic_dist_config(void __iomem *base, int gic_irqs,
|
|
+ void (*sync_access)(void));
|
|
+void gic_cpu_config(void __iomem *base, void (*sync_access)(void));
|
|
+
|
|
+#endif /* _IRQ_GIC_COMMON_H */
|
|
diff --git a/drivers/irqchip/irq-gic-v3.c b/drivers/irqchip/irq-gic-v3.c
|
|
new file mode 100644
|
|
index 0000000..81519ba
|
|
--- /dev/null
|
|
+++ b/drivers/irqchip/irq-gic-v3.c
|
|
@@ -0,0 +1,692 @@
|
|
+/*
|
|
+ * Copyright (C) 2013, 2014 ARM Limited, All Rights Reserved.
|
|
+ * Author: Marc Zyngier <marc.zyngier@arm.com>
|
|
+ *
|
|
+ * This program is free software; you can redistribute it and/or modify
|
|
+ * it under the terms of the GNU General Public License version 2 as
|
|
+ * published by the Free Software Foundation.
|
|
+ *
|
|
+ * This program is distributed in the hope that it will be useful,
|
|
+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
+ * GNU General Public License for more details.
|
|
+ *
|
|
+ * You should have received a copy of the GNU General Public License
|
|
+ * along with this program. If not, see <http://www.gnu.org/licenses/>.
|
|
+ */
|
|
+
|
|
+#include <linux/cpu.h>
|
|
+#include <linux/delay.h>
|
|
+#include <linux/interrupt.h>
|
|
+#include <linux/of.h>
|
|
+#include <linux/of_address.h>
|
|
+#include <linux/of_irq.h>
|
|
+#include <linux/percpu.h>
|
|
+#include <linux/slab.h>
|
|
+
|
|
+#include <linux/irqchip/arm-gic-v3.h>
|
|
+
|
|
+#include <asm/cputype.h>
|
|
+#include <asm/exception.h>
|
|
+#include <asm/smp_plat.h>
|
|
+
|
|
+#include "irq-gic-common.h"
|
|
+#include "irqchip.h"
|
|
+
|
|
+struct gic_chip_data {
|
|
+ void __iomem *dist_base;
|
|
+ void __iomem **redist_base;
|
|
+ void __percpu __iomem **rdist;
|
|
+ struct irq_domain *domain;
|
|
+ u64 redist_stride;
|
|
+ u32 redist_regions;
|
|
+ unsigned int irq_nr;
|
|
+};
|
|
+
|
|
+static struct gic_chip_data gic_data __read_mostly;
|
|
+
|
|
+#define gic_data_rdist() (this_cpu_ptr(gic_data.rdist))
|
|
+#define gic_data_rdist_rd_base() (*gic_data_rdist())
|
|
+#define gic_data_rdist_sgi_base() (gic_data_rdist_rd_base() + SZ_64K)
|
|
+
|
|
+/* Our default, arbitrary priority value. Linux only uses one anyway. */
|
|
+#define DEFAULT_PMR_VALUE 0xf0
|
|
+
|
|
+static inline unsigned int gic_irq(struct irq_data *d)
|
|
+{
|
|
+ return d->hwirq;
|
|
+}
|
|
+
|
|
+static inline int gic_irq_in_rdist(struct irq_data *d)
|
|
+{
|
|
+ return gic_irq(d) < 32;
|
|
+}
|
|
+
|
|
+static inline void __iomem *gic_dist_base(struct irq_data *d)
|
|
+{
|
|
+ if (gic_irq_in_rdist(d)) /* SGI+PPI -> SGI_base for this CPU */
|
|
+ return gic_data_rdist_sgi_base();
|
|
+
|
|
+ if (d->hwirq <= 1023) /* SPI -> dist_base */
|
|
+ return gic_data.dist_base;
|
|
+
|
|
+ if (d->hwirq >= 8192)
|
|
+ BUG(); /* LPI Detected!!! */
|
|
+
|
|
+ return NULL;
|
|
+}
|
|
+
|
|
+static void gic_do_wait_for_rwp(void __iomem *base)
|
|
+{
|
|
+ u32 count = 1000000; /* 1s! */
|
|
+
|
|
+ while (readl_relaxed(base + GICD_CTLR) & GICD_CTLR_RWP) {
|
|
+ count--;
|
|
+ if (!count) {
|
|
+ pr_err_ratelimited("RWP timeout, gone fishing\n");
|
|
+ return;
|
|
+ }
|
|
+ cpu_relax();
|
|
+ udelay(1);
|
|
+ };
|
|
+}
|
|
+
|
|
+/* Wait for completion of a distributor change */
|
|
+static void gic_dist_wait_for_rwp(void)
|
|
+{
|
|
+ gic_do_wait_for_rwp(gic_data.dist_base);
|
|
+}
|
|
+
|
|
+/* Wait for completion of a redistributor change */
|
|
+static void gic_redist_wait_for_rwp(void)
|
|
+{
|
|
+ gic_do_wait_for_rwp(gic_data_rdist_rd_base());
|
|
+}
|
|
+
|
|
+/* Low level accessors */
|
|
+static u64 gic_read_iar(void)
|
|
+{
|
|
+ u64 irqstat;
|
|
+
|
|
+ asm volatile("mrs %0, " __stringify(ICC_IAR1_EL1) : "=r" (irqstat));
|
|
+ return irqstat;
|
|
+}
|
|
+
|
|
+static void gic_write_pmr(u64 val)
|
|
+{
|
|
+ asm volatile("msr " __stringify(ICC_PMR_EL1) ", %0" : : "r" (val));
|
|
+}
|
|
+
|
|
+static void gic_write_ctlr(u64 val)
|
|
+{
|
|
+ asm volatile("msr " __stringify(ICC_CTLR_EL1) ", %0" : : "r" (val));
|
|
+ isb();
|
|
+}
|
|
+
|
|
+static void gic_write_grpen1(u64 val)
|
|
+{
|
|
+ asm volatile("msr " __stringify(ICC_GRPEN1_EL1) ", %0" : : "r" (val));
|
|
+ isb();
|
|
+}
|
|
+
|
|
+static void gic_write_sgi1r(u64 val)
|
|
+{
|
|
+ asm volatile("msr " __stringify(ICC_SGI1R_EL1) ", %0" : : "r" (val));
|
|
+}
|
|
+
|
|
+static void gic_enable_sre(void)
|
|
+{
|
|
+ u64 val;
|
|
+
|
|
+ asm volatile("mrs %0, " __stringify(ICC_SRE_EL1) : "=r" (val));
|
|
+ val |= ICC_SRE_EL1_SRE;
|
|
+ asm volatile("msr " __stringify(ICC_SRE_EL1) ", %0" : : "r" (val));
|
|
+ isb();
|
|
+
|
|
+ /*
|
|
+ * Need to check that the SRE bit has actually been set. If
|
|
+ * not, it means that SRE is disabled at EL2. We're going to
|
|
+ * die painfully, and there is nothing we can do about it.
|
|
+ *
|
|
+ * Kindly inform the luser.
|
|
+ */
|
|
+ asm volatile("mrs %0, " __stringify(ICC_SRE_EL1) : "=r" (val));
|
|
+ if (!(val & ICC_SRE_EL1_SRE))
|
|
+ pr_err("GIC: unable to set SRE (disabled at EL2), panic ahead\n");
|
|
+}
|
|
+
|
|
+static void gic_enable_redist(void)
|
|
+{
|
|
+ void __iomem *rbase;
|
|
+ u32 count = 1000000; /* 1s! */
|
|
+ u32 val;
|
|
+
|
|
+ rbase = gic_data_rdist_rd_base();
|
|
+
|
|
+ /* Wake up this CPU redistributor */
|
|
+ val = readl_relaxed(rbase + GICR_WAKER);
|
|
+ val &= ~GICR_WAKER_ProcessorSleep;
|
|
+ writel_relaxed(val, rbase + GICR_WAKER);
|
|
+
|
|
+ while (readl_relaxed(rbase + GICR_WAKER) & GICR_WAKER_ChildrenAsleep) {
|
|
+ count--;
|
|
+ if (!count) {
|
|
+ pr_err_ratelimited("redist didn't wake up...\n");
|
|
+ return;
|
|
+ }
|
|
+ cpu_relax();
|
|
+ udelay(1);
|
|
+ };
|
|
+}
|
|
+
|
|
+/*
|
|
+ * Routines to disable, enable, EOI and route interrupts
|
|
+ */
|
|
+static void gic_poke_irq(struct irq_data *d, u32 offset)
|
|
+{
|
|
+ u32 mask = 1 << (gic_irq(d) % 32);
|
|
+ void (*rwp_wait)(void);
|
|
+ void __iomem *base;
|
|
+
|
|
+ if (gic_irq_in_rdist(d)) {
|
|
+ base = gic_data_rdist_sgi_base();
|
|
+ rwp_wait = gic_redist_wait_for_rwp;
|
|
+ } else {
|
|
+ base = gic_data.dist_base;
|
|
+ rwp_wait = gic_dist_wait_for_rwp;
|
|
+ }
|
|
+
|
|
+ writel_relaxed(mask, base + offset + (gic_irq(d) / 32) * 4);
|
|
+ rwp_wait();
|
|
+}
|
|
+
|
|
+static int gic_peek_irq(struct irq_data *d, u32 offset)
|
|
+{
|
|
+ u32 mask = 1 << (gic_irq(d) % 32);
|
|
+ void __iomem *base;
|
|
+
|
|
+ if (gic_irq_in_rdist(d))
|
|
+ base = gic_data_rdist_sgi_base();
|
|
+ else
|
|
+ base = gic_data.dist_base;
|
|
+
|
|
+ return !!(readl_relaxed(base + offset + (gic_irq(d) / 32) * 4) & mask);
|
|
+}
|
|
+
|
|
+static void gic_mask_irq(struct irq_data *d)
|
|
+{
|
|
+ gic_poke_irq(d, GICD_ICENABLER);
|
|
+}
|
|
+
|
|
+static void gic_unmask_irq(struct irq_data *d)
|
|
+{
|
|
+ gic_poke_irq(d, GICD_ISENABLER);
|
|
+}
|
|
+
|
|
+static void gic_eoi_irq(struct irq_data *d)
|
|
+{
|
|
+ gic_write_eoir(gic_irq(d));
|
|
+}
|
|
+
|
|
+static int gic_set_type(struct irq_data *d, unsigned int type)
|
|
+{
|
|
+ unsigned int irq = gic_irq(d);
|
|
+ void (*rwp_wait)(void);
|
|
+ void __iomem *base;
|
|
+
|
|
+ /* Interrupt configuration for SGIs can't be changed */
|
|
+ if (irq < 16)
|
|
+ return -EINVAL;
|
|
+
|
|
+ if (type != IRQ_TYPE_LEVEL_HIGH && type != IRQ_TYPE_EDGE_RISING)
|
|
+ return -EINVAL;
|
|
+
|
|
+ if (gic_irq_in_rdist(d)) {
|
|
+ base = gic_data_rdist_sgi_base();
|
|
+ rwp_wait = gic_redist_wait_for_rwp;
|
|
+ } else {
|
|
+ base = gic_data.dist_base;
|
|
+ rwp_wait = gic_dist_wait_for_rwp;
|
|
+ }
|
|
+
|
|
+ gic_configure_irq(irq, type, base, rwp_wait);
|
|
+
|
|
+ return 0;
|
|
+}
|
|
+
|
|
+static u64 gic_mpidr_to_affinity(u64 mpidr)
|
|
+{
|
|
+ u64 aff;
|
|
+
|
|
+ aff = (MPIDR_AFFINITY_LEVEL(mpidr, 3) << 32 |
|
|
+ MPIDR_AFFINITY_LEVEL(mpidr, 2) << 16 |
|
|
+ MPIDR_AFFINITY_LEVEL(mpidr, 1) << 8 |
|
|
+ MPIDR_AFFINITY_LEVEL(mpidr, 0));
|
|
+
|
|
+ return aff;
|
|
+}
|
|
+
|
|
+static asmlinkage void __exception_irq_entry gic_handle_irq(struct pt_regs *regs)
|
|
+{
|
|
+ u64 irqnr;
|
|
+
|
|
+ do {
|
|
+ irqnr = gic_read_iar();
|
|
+
|
|
+ if (likely(irqnr > 15 && irqnr < 1020)) {
|
|
+ u64 irq = irq_find_mapping(gic_data.domain, irqnr);
|
|
+ if (likely(irq)) {
|
|
+ handle_IRQ(irq, regs);
|
|
+ continue;
|
|
+ }
|
|
+
|
|
+ WARN_ONCE(true, "Unexpected SPI received!\n");
|
|
+ gic_write_eoir(irqnr);
|
|
+ }
|
|
+ if (irqnr < 16) {
|
|
+ gic_write_eoir(irqnr);
|
|
+#ifdef CONFIG_SMP
|
|
+ handle_IPI(irqnr, regs);
|
|
+#else
|
|
+ WARN_ONCE(true, "Unexpected SGI received!\n");
|
|
+#endif
|
|
+ continue;
|
|
+ }
|
|
+ } while (irqnr != ICC_IAR1_EL1_SPURIOUS);
|
|
+}
|
|
+
|
|
+static void __init gic_dist_init(void)
|
|
+{
|
|
+ unsigned int i;
|
|
+ u64 affinity;
|
|
+ void __iomem *base = gic_data.dist_base;
|
|
+
|
|
+ /* Disable the distributor */
|
|
+ writel_relaxed(0, base + GICD_CTLR);
|
|
+ gic_dist_wait_for_rwp();
|
|
+
|
|
+ gic_dist_config(base, gic_data.irq_nr, gic_dist_wait_for_rwp);
|
|
+
|
|
+ /* Enable distributor with ARE, Group1 */
|
|
+ writel_relaxed(GICD_CTLR_ARE_NS | GICD_CTLR_ENABLE_G1A | GICD_CTLR_ENABLE_G1,
|
|
+ base + GICD_CTLR);
|
|
+
|
|
+ /*
|
|
+ * Set all global interrupts to the boot CPU only. ARE must be
|
|
+ * enabled.
|
|
+ */
|
|
+ affinity = gic_mpidr_to_affinity(cpu_logical_map(smp_processor_id()));
|
|
+ for (i = 32; i < gic_data.irq_nr; i++)
|
|
+ writeq_relaxed(affinity, base + GICD_IROUTER + i * 8);
|
|
+}
|
|
+
|
|
+static int gic_populate_rdist(void)
|
|
+{
|
|
+ u64 mpidr = cpu_logical_map(smp_processor_id());
|
|
+ u64 typer;
|
|
+ u32 aff;
|
|
+ int i;
|
|
+
|
|
+ /*
|
|
+ * Convert affinity to a 32bit value that can be matched to
|
|
+ * GICR_TYPER bits [63:32].
|
|
+ */
|
|
+ aff = (MPIDR_AFFINITY_LEVEL(mpidr, 3) << 24 |
|
|
+ MPIDR_AFFINITY_LEVEL(mpidr, 2) << 16 |
|
|
+ MPIDR_AFFINITY_LEVEL(mpidr, 1) << 8 |
|
|
+ MPIDR_AFFINITY_LEVEL(mpidr, 0));
|
|
+
|
|
+ for (i = 0; i < gic_data.redist_regions; i++) {
|
|
+ void __iomem *ptr = gic_data.redist_base[i];
|
|
+ u32 reg;
|
|
+
|
|
+ reg = readl_relaxed(ptr + GICR_PIDR2) & GIC_PIDR2_ARCH_MASK;
|
|
+ if (reg != GIC_PIDR2_ARCH_GICv3 &&
|
|
+ reg != GIC_PIDR2_ARCH_GICv4) { /* We're in trouble... */
|
|
+ pr_warn("No redistributor present @%p\n", ptr);
|
|
+ break;
|
|
+ }
|
|
+
|
|
+ do {
|
|
+ typer = readq_relaxed(ptr + GICR_TYPER);
|
|
+ if ((typer >> 32) == aff) {
|
|
+ gic_data_rdist_rd_base() = ptr;
|
|
+ pr_info("CPU%d: found redistributor %llx @%p\n",
|
|
+ smp_processor_id(),
|
|
+ (unsigned long long)mpidr, ptr);
|
|
+ return 0;
|
|
+ }
|
|
+
|
|
+ if (gic_data.redist_stride) {
|
|
+ ptr += gic_data.redist_stride;
|
|
+ } else {
|
|
+ ptr += SZ_64K * 2; /* Skip RD_base + SGI_base */
|
|
+ if (typer & GICR_TYPER_VLPIS)
|
|
+ ptr += SZ_64K * 2; /* Skip VLPI_base + reserved page */
|
|
+ }
|
|
+ } while (!(typer & GICR_TYPER_LAST));
|
|
+ }
|
|
+
|
|
+ /* We couldn't even deal with ourselves... */
|
|
+ WARN(true, "CPU%d: mpidr %llx has no re-distributor!\n",
|
|
+ smp_processor_id(), (unsigned long long)mpidr);
|
|
+ return -ENODEV;
|
|
+}
|
|
+
|
|
+static void gic_cpu_init(void)
|
|
+{
|
|
+ void __iomem *rbase;
|
|
+
|
|
+ /* Register ourselves with the rest of the world */
|
|
+ if (gic_populate_rdist())
|
|
+ return;
|
|
+
|
|
+ gic_enable_redist();
|
|
+
|
|
+ rbase = gic_data_rdist_sgi_base();
|
|
+
|
|
+ gic_cpu_config(rbase, gic_redist_wait_for_rwp);
|
|
+
|
|
+ /* Enable system registers */
|
|
+ gic_enable_sre();
|
|
+
|
|
+ /* Set priority mask register */
|
|
+ gic_write_pmr(DEFAULT_PMR_VALUE);
|
|
+
|
|
+ /* EOI deactivates interrupt too (mode 0) */
|
|
+ gic_write_ctlr(ICC_CTLR_EL1_EOImode_drop_dir);
|
|
+
|
|
+ /* ... and let's hit the road... */
|
|
+ gic_write_grpen1(1);
|
|
+}
|
|
+
|
|
+#ifdef CONFIG_SMP
|
|
+static int gic_secondary_init(struct notifier_block *nfb,
|
|
+ unsigned long action, void *hcpu)
|
|
+{
|
|
+ if (action == CPU_STARTING || action == CPU_STARTING_FROZEN)
|
|
+ gic_cpu_init();
|
|
+ return NOTIFY_OK;
|
|
+}
|
|
+
|
|
+/*
|
|
+ * Notifier for enabling the GIC CPU interface. Set an arbitrarily high
|
|
+ * priority because the GIC needs to be up before the ARM generic timers.
|
|
+ */
|
|
+static struct notifier_block gic_cpu_notifier = {
|
|
+ .notifier_call = gic_secondary_init,
|
|
+ .priority = 100,
|
|
+};
|
|
+
|
|
+static u16 gic_compute_target_list(int *base_cpu, const struct cpumask *mask,
|
|
+ u64 cluster_id)
|
|
+{
|
|
+ int cpu = *base_cpu;
|
|
+ u64 mpidr = cpu_logical_map(cpu);
|
|
+ u16 tlist = 0;
|
|
+
|
|
+ while (cpu < nr_cpu_ids) {
|
|
+ /*
|
|
+ * If we ever get a cluster of more than 16 CPUs, just
|
|
+ * scream and skip that CPU.
|
|
+ */
|
|
+ if (WARN_ON((mpidr & 0xff) >= 16))
|
|
+ goto out;
|
|
+
|
|
+ tlist |= 1 << (mpidr & 0xf);
|
|
+
|
|
+ cpu = cpumask_next(cpu, mask);
|
|
+ if (cpu == nr_cpu_ids)
|
|
+ goto out;
|
|
+
|
|
+ mpidr = cpu_logical_map(cpu);
|
|
+
|
|
+ if (cluster_id != (mpidr & ~0xffUL)) {
|
|
+ cpu--;
|
|
+ goto out;
|
|
+ }
|
|
+ }
|
|
+out:
|
|
+ *base_cpu = cpu;
|
|
+ return tlist;
|
|
+}
|
|
+
|
|
+static void gic_send_sgi(u64 cluster_id, u16 tlist, unsigned int irq)
|
|
+{
|
|
+ u64 val;
|
|
+
|
|
+ val = (MPIDR_AFFINITY_LEVEL(cluster_id, 3) << 48 |
|
|
+ MPIDR_AFFINITY_LEVEL(cluster_id, 2) << 32 |
|
|
+ irq << 24 |
|
|
+ MPIDR_AFFINITY_LEVEL(cluster_id, 1) << 16 |
|
|
+ tlist);
|
|
+
|
|
+ pr_debug("CPU%d: ICC_SGI1R_EL1 %llx\n", smp_processor_id(), val);
|
|
+ gic_write_sgi1r(val);
|
|
+}
|
|
+
|
|
+static void gic_raise_softirq(const struct cpumask *mask, unsigned int irq)
|
|
+{
|
|
+ int cpu;
|
|
+
|
|
+ if (WARN_ON(irq >= 16))
|
|
+ return;
|
|
+
|
|
+ /*
|
|
+ * Ensure that stores to Normal memory are visible to the
|
|
+ * other CPUs before issuing the IPI.
|
|
+ */
|
|
+ smp_wmb();
|
|
+
|
|
+ for_each_cpu_mask(cpu, *mask) {
|
|
+ u64 cluster_id = cpu_logical_map(cpu) & ~0xffUL;
|
|
+ u16 tlist;
|
|
+
|
|
+ tlist = gic_compute_target_list(&cpu, mask, cluster_id);
|
|
+ gic_send_sgi(cluster_id, tlist, irq);
|
|
+ }
|
|
+
|
|
+ /* Force the above writes to ICC_SGI1R_EL1 to be executed */
|
|
+ isb();
|
|
+}
|
|
+
|
|
+static void gic_smp_init(void)
|
|
+{
|
|
+ set_smp_cross_call(gic_raise_softirq);
|
|
+ register_cpu_notifier(&gic_cpu_notifier);
|
|
+}
|
|
+
|
|
+static int gic_set_affinity(struct irq_data *d, const struct cpumask *mask_val,
|
|
+ bool force)
|
|
+{
|
|
+ unsigned int cpu = cpumask_any_and(mask_val, cpu_online_mask);
|
|
+ void __iomem *reg;
|
|
+ int enabled;
|
|
+ u64 val;
|
|
+
|
|
+ if (gic_irq_in_rdist(d))
|
|
+ return -EINVAL;
|
|
+
|
|
+ /* If interrupt was enabled, disable it first */
|
|
+ enabled = gic_peek_irq(d, GICD_ISENABLER);
|
|
+ if (enabled)
|
|
+ gic_mask_irq(d);
|
|
+
|
|
+ reg = gic_dist_base(d) + GICD_IROUTER + (gic_irq(d) * 8);
|
|
+ val = gic_mpidr_to_affinity(cpu_logical_map(cpu));
|
|
+
|
|
+ writeq_relaxed(val, reg);
|
|
+
|
|
+ /*
|
|
+ * If the interrupt was enabled, enabled it again. Otherwise,
|
|
+ * just wait for the distributor to have digested our changes.
|
|
+ */
|
|
+ if (enabled)
|
|
+ gic_unmask_irq(d);
|
|
+ else
|
|
+ gic_dist_wait_for_rwp();
|
|
+
|
|
+ return IRQ_SET_MASK_OK;
|
|
+}
|
|
+#else
|
|
+#define gic_set_affinity NULL
|
|
+#define gic_smp_init() do { } while(0)
|
|
+#endif
|
|
+
|
|
+static struct irq_chip gic_chip = {
|
|
+ .name = "GICv3",
|
|
+ .irq_mask = gic_mask_irq,
|
|
+ .irq_unmask = gic_unmask_irq,
|
|
+ .irq_eoi = gic_eoi_irq,
|
|
+ .irq_set_type = gic_set_type,
|
|
+ .irq_set_affinity = gic_set_affinity,
|
|
+};
|
|
+
|
|
+static int gic_irq_domain_map(struct irq_domain *d, unsigned int irq,
|
|
+ irq_hw_number_t hw)
|
|
+{
|
|
+ /* SGIs are private to the core kernel */
|
|
+ if (hw < 16)
|
|
+ return -EPERM;
|
|
+ /* PPIs */
|
|
+ if (hw < 32) {
|
|
+ irq_set_percpu_devid(irq);
|
|
+ irq_set_chip_and_handler(irq, &gic_chip,
|
|
+ handle_percpu_devid_irq);
|
|
+ set_irq_flags(irq, IRQF_VALID | IRQF_NOAUTOEN);
|
|
+ }
|
|
+ /* SPIs */
|
|
+ if (hw >= 32 && hw < gic_data.irq_nr) {
|
|
+ irq_set_chip_and_handler(irq, &gic_chip,
|
|
+ handle_fasteoi_irq);
|
|
+ set_irq_flags(irq, IRQF_VALID | IRQF_PROBE);
|
|
+ }
|
|
+ irq_set_chip_data(irq, d->host_data);
|
|
+ return 0;
|
|
+}
|
|
+
|
|
+static int gic_irq_domain_xlate(struct irq_domain *d,
|
|
+ struct device_node *controller,
|
|
+ const u32 *intspec, unsigned int intsize,
|
|
+ unsigned long *out_hwirq, unsigned int *out_type)
|
|
+{
|
|
+ if (d->of_node != controller)
|
|
+ return -EINVAL;
|
|
+ if (intsize < 3)
|
|
+ return -EINVAL;
|
|
+
|
|
+ switch(intspec[0]) {
|
|
+ case 0: /* SPI */
|
|
+ *out_hwirq = intspec[1] + 32;
|
|
+ break;
|
|
+ case 1: /* PPI */
|
|
+ *out_hwirq = intspec[1] + 16;
|
|
+ break;
|
|
+ default:
|
|
+ return -EINVAL;
|
|
+ }
|
|
+
|
|
+ *out_type = intspec[2] & IRQ_TYPE_SENSE_MASK;
|
|
+ return 0;
|
|
+}
|
|
+
|
|
+static const struct irq_domain_ops gic_irq_domain_ops = {
|
|
+ .map = gic_irq_domain_map,
|
|
+ .xlate = gic_irq_domain_xlate,
|
|
+};
|
|
+
|
|
+static int __init gic_of_init(struct device_node *node, struct device_node *parent)
|
|
+{
|
|
+ void __iomem *dist_base;
|
|
+ void __iomem **redist_base;
|
|
+ u64 redist_stride;
|
|
+ u32 redist_regions;
|
|
+ u32 reg;
|
|
+ int gic_irqs;
|
|
+ int err;
|
|
+ int i;
|
|
+
|
|
+ dist_base = of_iomap(node, 0);
|
|
+ if (!dist_base) {
|
|
+ pr_err("%s: unable to map gic dist registers\n",
|
|
+ node->full_name);
|
|
+ return -ENXIO;
|
|
+ }
|
|
+
|
|
+ reg = readl_relaxed(dist_base + GICD_PIDR2) & GIC_PIDR2_ARCH_MASK;
|
|
+ if (reg != GIC_PIDR2_ARCH_GICv3 && reg != GIC_PIDR2_ARCH_GICv4) {
|
|
+ pr_err("%s: no distributor detected, giving up\n",
|
|
+ node->full_name);
|
|
+ err = -ENODEV;
|
|
+ goto out_unmap_dist;
|
|
+ }
|
|
+
|
|
+ if (of_property_read_u32(node, "#redistributor-regions", &redist_regions))
|
|
+ redist_regions = 1;
|
|
+
|
|
+ redist_base = kzalloc(sizeof(*redist_base) * redist_regions, GFP_KERNEL);
|
|
+ if (!redist_base) {
|
|
+ err = -ENOMEM;
|
|
+ goto out_unmap_dist;
|
|
+ }
|
|
+
|
|
+ for (i = 0; i < redist_regions; i++) {
|
|
+ redist_base[i] = of_iomap(node, 1 + i);
|
|
+ if (!redist_base[i]) {
|
|
+ pr_err("%s: couldn't map region %d\n",
|
|
+ node->full_name, i);
|
|
+ err = -ENODEV;
|
|
+ goto out_unmap_rdist;
|
|
+ }
|
|
+ }
|
|
+
|
|
+ if (of_property_read_u64(node, "redistributor-stride", &redist_stride))
|
|
+ redist_stride = 0;
|
|
+
|
|
+ gic_data.dist_base = dist_base;
|
|
+ gic_data.redist_base = redist_base;
|
|
+ gic_data.redist_regions = redist_regions;
|
|
+ gic_data.redist_stride = redist_stride;
|
|
+
|
|
+ /*
|
|
+ * Find out how many interrupts are supported.
|
|
+ * The GIC only supports up to 1020 interrupt sources (SGI+PPI+SPI)
|
|
+ */
|
|
+ gic_irqs = readl_relaxed(gic_data.dist_base + GICD_TYPER) & 0x1f;
|
|
+ gic_irqs = (gic_irqs + 1) * 32;
|
|
+ if (gic_irqs > 1020)
|
|
+ gic_irqs = 1020;
|
|
+ gic_data.irq_nr = gic_irqs;
|
|
+
|
|
+ gic_data.domain = irq_domain_add_tree(node, &gic_irq_domain_ops,
|
|
+ &gic_data);
|
|
+ gic_data.rdist = alloc_percpu(typeof(*gic_data.rdist));
|
|
+
|
|
+ if (WARN_ON(!gic_data.domain) || WARN_ON(!gic_data.rdist)) {
|
|
+ err = -ENOMEM;
|
|
+ goto out_free;
|
|
+ }
|
|
+
|
|
+ set_handle_irq(gic_handle_irq);
|
|
+
|
|
+ gic_smp_init();
|
|
+ gic_dist_init();
|
|
+ gic_cpu_init();
|
|
+
|
|
+ return 0;
|
|
+
|
|
+out_free:
|
|
+ if (gic_data.domain)
|
|
+ irq_domain_remove(gic_data.domain);
|
|
+ free_percpu(gic_data.rdist);
|
|
+out_unmap_rdist:
|
|
+ for (i = 0; i < redist_regions; i++)
|
|
+ if (redist_base[i])
|
|
+ iounmap(redist_base[i]);
|
|
+ kfree(redist_base);
|
|
+out_unmap_dist:
|
|
+ iounmap(dist_base);
|
|
+ return err;
|
|
+}
|
|
+
|
|
+IRQCHIP_DECLARE(gic_v3, "arm,gic-v3", gic_of_init);
|
|
diff --git a/drivers/irqchip/irq-gic.c b/drivers/irqchip/irq-gic.c
|
|
index 7c131cf..1ddfdde 100644
|
|
--- a/drivers/irqchip/irq-gic.c
|
|
+++ b/drivers/irqchip/irq-gic.c
|
|
@@ -47,6 +47,7 @@
|
|
#include <asm/exception.h>
|
|
#include <asm/smp_plat.h>
|
|
|
|
+#include "irq-gic-common.h"
|
|
#include "irqchip.h"
|
|
|
|
union gic_base {
|
|
@@ -189,12 +190,6 @@ static int gic_set_type(struct irq_data *d, unsigned int type)
|
|
{
|
|
void __iomem *base = gic_dist_base(d);
|
|
unsigned int gicirq = gic_irq(d);
|
|
- u32 enablemask = 1 << (gicirq % 32);
|
|
- u32 enableoff = (gicirq / 32) * 4;
|
|
- u32 confmask = 0x2 << ((gicirq % 16) * 2);
|
|
- u32 confoff = (gicirq / 16) * 4;
|
|
- bool enabled = false;
|
|
- u32 val;
|
|
|
|
/* Interrupt configuration for SGIs can't be changed */
|
|
if (gicirq < 16)
|
|
@@ -208,25 +203,7 @@ static int gic_set_type(struct irq_data *d, unsigned int type)
|
|
if (gic_arch_extn.irq_set_type)
|
|
gic_arch_extn.irq_set_type(d, type);
|
|
|
|
- val = readl_relaxed(base + GIC_DIST_CONFIG + confoff);
|
|
- if (type == IRQ_TYPE_LEVEL_HIGH)
|
|
- val &= ~confmask;
|
|
- else if (type == IRQ_TYPE_EDGE_RISING)
|
|
- val |= confmask;
|
|
-
|
|
- /*
|
|
- * As recommended by the spec, disable the interrupt before changing
|
|
- * the configuration
|
|
- */
|
|
- if (readl_relaxed(base + GIC_DIST_ENABLE_SET + enableoff) & enablemask) {
|
|
- writel_relaxed(enablemask, base + GIC_DIST_ENABLE_CLEAR + enableoff);
|
|
- enabled = true;
|
|
- }
|
|
-
|
|
- writel_relaxed(val, base + GIC_DIST_CONFIG + confoff);
|
|
-
|
|
- if (enabled)
|
|
- writel_relaxed(enablemask, base + GIC_DIST_ENABLE_SET + enableoff);
|
|
+ gic_configure_irq(gicirq, type, base, NULL);
|
|
|
|
raw_spin_unlock(&irq_controller_lock);
|
|
|
|
@@ -388,12 +365,6 @@ static void __init gic_dist_init(struct gic_chip_data *gic)
|
|
writel_relaxed(0, base + GIC_DIST_CTRL);
|
|
|
|
/*
|
|
- * Set all global interrupts to be level triggered, active low.
|
|
- */
|
|
- for (i = 32; i < gic_irqs; i += 16)
|
|
- writel_relaxed(0, base + GIC_DIST_CONFIG + i * 4 / 16);
|
|
-
|
|
- /*
|
|
* Set all global interrupts to this CPU only.
|
|
*/
|
|
cpumask = gic_get_cpumask(gic);
|
|
@@ -402,18 +373,7 @@ static void __init gic_dist_init(struct gic_chip_data *gic)
|
|
for (i = 32; i < gic_irqs; i += 4)
|
|
writel_relaxed(cpumask, base + GIC_DIST_TARGET + i * 4 / 4);
|
|
|
|
- /*
|
|
- * Set priority on all global interrupts.
|
|
- */
|
|
- for (i = 32; i < gic_irqs; i += 4)
|
|
- writel_relaxed(0xa0a0a0a0, base + GIC_DIST_PRI + i * 4 / 4);
|
|
-
|
|
- /*
|
|
- * Disable all interrupts. Leave the PPI and SGIs alone
|
|
- * as these enables are banked registers.
|
|
- */
|
|
- for (i = 32; i < gic_irqs; i += 32)
|
|
- writel_relaxed(0xffffffff, base + GIC_DIST_ENABLE_CLEAR + i * 4 / 32);
|
|
+ gic_dist_config(base, gic_irqs, NULL);
|
|
|
|
writel_relaxed(1, base + GIC_DIST_CTRL);
|
|
}
|
|
@@ -423,6 +383,7 @@ static void gic_cpu_init(struct gic_chip_data *gic)
|
|
void __iomem *dist_base = gic_data_dist_base(gic);
|
|
void __iomem *base = gic_data_cpu_base(gic);
|
|
unsigned int cpu_mask, cpu = smp_processor_id();
|
|
+ unsigned int ctrl_mask;
|
|
int i;
|
|
|
|
/*
|
|
@@ -440,27 +401,32 @@ static void gic_cpu_init(struct gic_chip_data *gic)
|
|
if (i != cpu)
|
|
gic_cpu_map[i] &= ~cpu_mask;
|
|
|
|
- /*
|
|
- * Deal with the banked PPI and SGI interrupts - disable all
|
|
- * PPI interrupts, ensure all SGI interrupts are enabled.
|
|
- */
|
|
- writel_relaxed(0xffff0000, dist_base + GIC_DIST_ENABLE_CLEAR);
|
|
- writel_relaxed(0x0000ffff, dist_base + GIC_DIST_ENABLE_SET);
|
|
-
|
|
- /*
|
|
- * Set priority on PPI and SGI interrupts
|
|
- */
|
|
- for (i = 0; i < 32; i += 4)
|
|
- writel_relaxed(0xa0a0a0a0, dist_base + GIC_DIST_PRI + i * 4 / 4);
|
|
+ gic_cpu_config(dist_base, NULL);
|
|
|
|
writel_relaxed(0xf0, base + GIC_CPU_PRIMASK);
|
|
- writel_relaxed(1, base + GIC_CPU_CTRL);
|
|
+
|
|
+ ctrl_mask = readl(base + GIC_CPU_CTRL);
|
|
+
|
|
+ /* Mask out the gic v2 bypass bits */
|
|
+ ctrl_mask &= 0x1e0;
|
|
+
|
|
+ /* Enable group 0 */
|
|
+ ctrl_mask |= 0x1;
|
|
+ writel_relaxed(ctrl_mask, base + GIC_CPU_CTRL);
|
|
}
|
|
|
|
void gic_cpu_if_down(void)
|
|
{
|
|
+ unsigned int ctrl_mask;
|
|
void __iomem *cpu_base = gic_data_cpu_base(&gic_data[0]);
|
|
- writel_relaxed(0, cpu_base + GIC_CPU_CTRL);
|
|
+
|
|
+ ctrl_mask = readl(cpu_base + GIC_CPU_CTRL);
|
|
+ /*
|
|
+ * Disable grp enable bit, leave the bypass bits alone as changing
|
|
+ * them could leave the system unstable
|
|
+ */
|
|
+ ctrl_mask &= 0x1e0;
|
|
+ writel_relaxed(ctrl_mask, cpu_base + GIC_CPU_CTRL);
|
|
}
|
|
|
|
#ifdef CONFIG_CPU_PM
|
|
@@ -571,6 +537,7 @@ static void gic_cpu_restore(unsigned int gic_nr)
|
|
{
|
|
int i;
|
|
u32 *ptr;
|
|
+ unsigned int ctrl_mask;
|
|
void __iomem *dist_base;
|
|
void __iomem *cpu_base;
|
|
|
|
@@ -595,7 +562,15 @@ static void gic_cpu_restore(unsigned int gic_nr)
|
|
writel_relaxed(0xa0a0a0a0, dist_base + GIC_DIST_PRI + i * 4);
|
|
|
|
writel_relaxed(0xf0, cpu_base + GIC_CPU_PRIMASK);
|
|
- writel_relaxed(1, cpu_base + GIC_CPU_CTRL);
|
|
+
|
|
+ ctrl_mask = readl(cpu_base + GIC_CPU_CTRL);
|
|
+
|
|
+ /* Mask out the gic v2 bypass bits */
|
|
+ ctrl_mask &= 0x1e0;
|
|
+
|
|
+ /* Enable group 0 */
|
|
+ ctrl_mask |= 0x1;
|
|
+ writel_relaxed(ctrl_mask, cpu_base + GIC_CPU_CTRL);
|
|
}
|
|
|
|
static int gic_notifier(struct notifier_block *self, unsigned long cmd, void *v)
|
|
diff --git a/drivers/net/ethernet/Kconfig b/drivers/net/ethernet/Kconfig
|
|
index edb7186..dc7406c 100644
|
|
--- a/drivers/net/ethernet/Kconfig
|
|
+++ b/drivers/net/ethernet/Kconfig
|
|
@@ -24,6 +24,7 @@ source "drivers/net/ethernet/allwinner/Kconfig"
|
|
source "drivers/net/ethernet/alteon/Kconfig"
|
|
source "drivers/net/ethernet/altera/Kconfig"
|
|
source "drivers/net/ethernet/amd/Kconfig"
|
|
+source "drivers/net/ethernet/apm/Kconfig"
|
|
source "drivers/net/ethernet/apple/Kconfig"
|
|
source "drivers/net/ethernet/arc/Kconfig"
|
|
source "drivers/net/ethernet/atheros/Kconfig"
|
|
diff --git a/drivers/net/ethernet/Makefile b/drivers/net/ethernet/Makefile
|
|
index 58de333..224a018 100644
|
|
--- a/drivers/net/ethernet/Makefile
|
|
+++ b/drivers/net/ethernet/Makefile
|
|
@@ -10,6 +10,7 @@ obj-$(CONFIG_NET_VENDOR_ALLWINNER) += allwinner/
|
|
obj-$(CONFIG_NET_VENDOR_ALTEON) += alteon/
|
|
obj-$(CONFIG_ALTERA_TSE) += altera/
|
|
obj-$(CONFIG_NET_VENDOR_AMD) += amd/
|
|
+obj-$(CONFIG_NET_XGENE) += apm/
|
|
obj-$(CONFIG_NET_VENDOR_APPLE) += apple/
|
|
obj-$(CONFIG_NET_VENDOR_ARC) += arc/
|
|
obj-$(CONFIG_NET_VENDOR_ATHEROS) += atheros/
|
|
diff --git a/drivers/net/ethernet/apm/Kconfig b/drivers/net/ethernet/apm/Kconfig
|
|
new file mode 100644
|
|
index 0000000..ec63d70
|
|
--- /dev/null
|
|
+++ b/drivers/net/ethernet/apm/Kconfig
|
|
@@ -0,0 +1 @@
|
|
+source "drivers/net/ethernet/apm/xgene/Kconfig"
|
|
diff --git a/drivers/net/ethernet/apm/Makefile b/drivers/net/ethernet/apm/Makefile
|
|
new file mode 100644
|
|
index 0000000..65ce32a
|
|
--- /dev/null
|
|
+++ b/drivers/net/ethernet/apm/Makefile
|
|
@@ -0,0 +1,5 @@
|
|
+#
|
|
+# Makefile for APM X-GENE Ethernet driver.
|
|
+#
|
|
+
|
|
+obj-$(CONFIG_NET_XGENE) += xgene/
|
|
diff --git a/drivers/net/ethernet/apm/xgene/Kconfig b/drivers/net/ethernet/apm/xgene/Kconfig
|
|
new file mode 100644
|
|
index 0000000..616dff6
|
|
--- /dev/null
|
|
+++ b/drivers/net/ethernet/apm/xgene/Kconfig
|
|
@@ -0,0 +1,9 @@
|
|
+config NET_XGENE
|
|
+ tristate "APM X-Gene SoC Ethernet Driver"
|
|
+ select PHYLIB
|
|
+ help
|
|
+ This is the Ethernet driver for the on-chip ethernet interface on the
|
|
+ APM X-Gene SoC.
|
|
+
|
|
+ To compile this driver as a module, choose M here. This module will
|
|
+ be called xgene_enet.
|
|
diff --git a/drivers/net/ethernet/apm/xgene/Makefile b/drivers/net/ethernet/apm/xgene/Makefile
|
|
new file mode 100644
|
|
index 0000000..c643e8a
|
|
--- /dev/null
|
|
+++ b/drivers/net/ethernet/apm/xgene/Makefile
|
|
@@ -0,0 +1,6 @@
|
|
+#
|
|
+# Makefile for APM X-Gene Ethernet Driver.
|
|
+#
|
|
+
|
|
+xgene-enet-objs := xgene_enet_hw.o xgene_enet_main.o xgene_enet_ethtool.o
|
|
+obj-$(CONFIG_NET_XGENE) += xgene-enet.o
|
|
diff --git a/drivers/net/ethernet/apm/xgene/xgene_enet_ethtool.c b/drivers/net/ethernet/apm/xgene/xgene_enet_ethtool.c
|
|
new file mode 100644
|
|
index 0000000..63f2aa5
|
|
--- /dev/null
|
|
+++ b/drivers/net/ethernet/apm/xgene/xgene_enet_ethtool.c
|
|
@@ -0,0 +1,125 @@
|
|
+/* Applied Micro X-Gene SoC Ethernet Driver
|
|
+ *
|
|
+ * Copyright (c) 2014, Applied Micro Circuits Corporation
|
|
+ * Authors: Iyappan Subramanian <isubramanian@apm.com>
|
|
+ *
|
|
+ * This program is free software; you can redistribute it and/or modify it
|
|
+ * under the terms of the GNU General Public License as published by the
|
|
+ * Free Software Foundation; either version 2 of the License, or (at your
|
|
+ * option) any later version.
|
|
+ *
|
|
+ * This program is distributed in the hope that it will be useful,
|
|
+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
+ * GNU General Public License for more details.
|
|
+ *
|
|
+ * You should have received a copy of the GNU General Public License
|
|
+ * along with this program. If not, see <http://www.gnu.org/licenses/>.
|
|
+ */
|
|
+
|
|
+#include <linux/ethtool.h>
|
|
+#include "xgene_enet_main.h"
|
|
+
|
|
+struct xgene_gstrings_stats {
|
|
+ char name[ETH_GSTRING_LEN];
|
|
+ int offset;
|
|
+};
|
|
+
|
|
+#define XGENE_STAT(m) { #m, offsetof(struct xgene_enet_pdata, stats.m) }
|
|
+
|
|
+static const struct xgene_gstrings_stats gstrings_stats[] = {
|
|
+ XGENE_STAT(rx_packets),
|
|
+ XGENE_STAT(tx_packets),
|
|
+ XGENE_STAT(rx_bytes),
|
|
+ XGENE_STAT(tx_bytes),
|
|
+ XGENE_STAT(rx_errors),
|
|
+ XGENE_STAT(tx_errors),
|
|
+ XGENE_STAT(rx_length_errors),
|
|
+ XGENE_STAT(rx_crc_errors),
|
|
+ XGENE_STAT(rx_frame_errors),
|
|
+ XGENE_STAT(rx_fifo_errors)
|
|
+};
|
|
+
|
|
+#define XGENE_STATS_LEN ARRAY_SIZE(gstrings_stats)
|
|
+
|
|
+static void xgene_get_drvinfo(struct net_device *ndev,
|
|
+ struct ethtool_drvinfo *info)
|
|
+{
|
|
+ struct xgene_enet_pdata *pdata = netdev_priv(ndev);
|
|
+ struct platform_device *pdev = pdata->pdev;
|
|
+
|
|
+ strcpy(info->driver, "xgene_enet");
|
|
+ strcpy(info->version, XGENE_DRV_VERSION);
|
|
+ snprintf(info->fw_version, ETHTOOL_FWVERS_LEN, "N/A");
|
|
+ sprintf(info->bus_info, "%s", pdev->name);
|
|
+}
|
|
+
|
|
+static int xgene_get_settings(struct net_device *ndev, struct ethtool_cmd *cmd)
|
|
+{
|
|
+ struct xgene_enet_pdata *pdata = netdev_priv(ndev);
|
|
+ struct phy_device *phydev = pdata->phy_dev;
|
|
+
|
|
+ if (phydev == NULL)
|
|
+ return -ENODEV;
|
|
+
|
|
+ return phy_ethtool_gset(phydev, cmd);
|
|
+}
|
|
+
|
|
+static int xgene_set_settings(struct net_device *ndev, struct ethtool_cmd *cmd)
|
|
+{
|
|
+ struct xgene_enet_pdata *pdata = netdev_priv(ndev);
|
|
+ struct phy_device *phydev = pdata->phy_dev;
|
|
+
|
|
+ if (phydev == NULL)
|
|
+ return -ENODEV;
|
|
+
|
|
+ return phy_ethtool_sset(phydev, cmd);
|
|
+}
|
|
+
|
|
+static void xgene_get_strings(struct net_device *ndev, u32 stringset, u8 *data)
|
|
+{
|
|
+ int i;
|
|
+ u8 *p = data;
|
|
+
|
|
+ if (stringset != ETH_SS_STATS)
|
|
+ return;
|
|
+
|
|
+ for (i = 0; i < XGENE_STATS_LEN; i++) {
|
|
+ memcpy(p, gstrings_stats[i].name, ETH_GSTRING_LEN);
|
|
+ p += ETH_GSTRING_LEN;
|
|
+ }
|
|
+}
|
|
+
|
|
+static int xgene_get_sset_count(struct net_device *ndev, int sset)
|
|
+{
|
|
+ if (sset != ETH_SS_STATS)
|
|
+ return -EINVAL;
|
|
+
|
|
+ return XGENE_STATS_LEN;
|
|
+}
|
|
+
|
|
+static void xgene_get_ethtool_stats(struct net_device *ndev,
|
|
+ struct ethtool_stats *dummy,
|
|
+ u64 *data)
|
|
+{
|
|
+ void *pdata = netdev_priv(ndev);
|
|
+ int i;
|
|
+
|
|
+ for (i = 0; i < XGENE_STATS_LEN; i++)
|
|
+ *data++ = *(u64 *)(pdata + gstrings_stats[i].offset);
|
|
+}
|
|
+
|
|
+static const struct ethtool_ops xgene_ethtool_ops = {
|
|
+ .get_drvinfo = xgene_get_drvinfo,
|
|
+ .get_settings = xgene_get_settings,
|
|
+ .set_settings = xgene_set_settings,
|
|
+ .get_link = ethtool_op_get_link,
|
|
+ .get_strings = xgene_get_strings,
|
|
+ .get_sset_count = xgene_get_sset_count,
|
|
+ .get_ethtool_stats = xgene_get_ethtool_stats
|
|
+};
|
|
+
|
|
+void xgene_enet_set_ethtool_ops(struct net_device *ndev)
|
|
+{
|
|
+ ndev->ethtool_ops = &xgene_ethtool_ops;
|
|
+}
|
|
diff --git a/drivers/net/ethernet/apm/xgene/xgene_enet_hw.c b/drivers/net/ethernet/apm/xgene/xgene_enet_hw.c
|
|
new file mode 100644
|
|
index 0000000..e52af60
|
|
--- /dev/null
|
|
+++ b/drivers/net/ethernet/apm/xgene/xgene_enet_hw.c
|
|
@@ -0,0 +1,747 @@
|
|
+/* Applied Micro X-Gene SoC Ethernet Driver
|
|
+ *
|
|
+ * Copyright (c) 2014, Applied Micro Circuits Corporation
|
|
+ * Authors: Iyappan Subramanian <isubramanian@apm.com>
|
|
+ * Ravi Patel <rapatel@apm.com>
|
|
+ * Keyur Chudgar <kchudgar@apm.com>
|
|
+ *
|
|
+ * This program is free software; you can redistribute it and/or modify it
|
|
+ * under the terms of the GNU General Public License as published by the
|
|
+ * Free Software Foundation; either version 2 of the License, or (at your
|
|
+ * option) any later version.
|
|
+ *
|
|
+ * This program is distributed in the hope that it will be useful,
|
|
+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
+ * GNU General Public License for more details.
|
|
+ *
|
|
+ * You should have received a copy of the GNU General Public License
|
|
+ * along with this program. If not, see <http://www.gnu.org/licenses/>.
|
|
+ */
|
|
+
|
|
+#include "xgene_enet_main.h"
|
|
+#include "xgene_enet_hw.h"
|
|
+
|
|
+static void xgene_enet_ring_init(struct xgene_enet_desc_ring *ring)
|
|
+{
|
|
+ u32 *ring_cfg = ring->state;
|
|
+ u64 addr = ring->dma;
|
|
+ enum xgene_enet_ring_cfgsize cfgsize = ring->cfgsize;
|
|
+
|
|
+ ring_cfg[4] |= (1 << SELTHRSH_POS) &
|
|
+ CREATE_MASK(SELTHRSH_POS, SELTHRSH_LEN);
|
|
+ ring_cfg[3] |= ACCEPTLERR;
|
|
+ ring_cfg[2] |= QCOHERENT;
|
|
+
|
|
+ addr >>= 8;
|
|
+ ring_cfg[2] |= (addr << RINGADDRL_POS) &
|
|
+ CREATE_MASK_ULL(RINGADDRL_POS, RINGADDRL_LEN);
|
|
+ addr >>= RINGADDRL_LEN;
|
|
+ ring_cfg[3] |= addr & CREATE_MASK_ULL(RINGADDRH_POS, RINGADDRH_LEN);
|
|
+ ring_cfg[3] |= ((u32) cfgsize << RINGSIZE_POS) &
|
|
+ CREATE_MASK(RINGSIZE_POS, RINGSIZE_LEN);
|
|
+}
|
|
+
|
|
+static void xgene_enet_ring_set_type(struct xgene_enet_desc_ring *ring)
|
|
+{
|
|
+ u32 *ring_cfg = ring->state;
|
|
+ bool is_bufpool;
|
|
+ u32 val;
|
|
+
|
|
+ is_bufpool = xgene_enet_is_bufpool(ring->id);
|
|
+ val = (is_bufpool) ? RING_BUFPOOL : RING_REGULAR;
|
|
+ ring_cfg[4] |= (val << RINGTYPE_POS) &
|
|
+ CREATE_MASK(RINGTYPE_POS, RINGTYPE_LEN);
|
|
+
|
|
+ if (is_bufpool) {
|
|
+ ring_cfg[3] |= (BUFPOOL_MODE << RINGMODE_POS) &
|
|
+ CREATE_MASK(RINGMODE_POS, RINGMODE_LEN);
|
|
+ }
|
|
+}
|
|
+
|
|
+static void xgene_enet_ring_set_recombbuf(struct xgene_enet_desc_ring *ring)
|
|
+{
|
|
+ u32 *ring_cfg = ring->state;
|
|
+
|
|
+ ring_cfg[3] |= RECOMBBUF;
|
|
+ ring_cfg[3] |= (0xf << RECOMTIMEOUTL_POS) &
|
|
+ CREATE_MASK(RECOMTIMEOUTL_POS, RECOMTIMEOUTL_LEN);
|
|
+ ring_cfg[4] |= 0x7 & CREATE_MASK(RECOMTIMEOUTH_POS, RECOMTIMEOUTH_LEN);
|
|
+}
|
|
+
|
|
+static void xgene_enet_ring_wr32(struct xgene_enet_desc_ring *ring,
|
|
+ u32 offset, u32 data)
|
|
+{
|
|
+ struct xgene_enet_pdata *pdata = netdev_priv(ring->ndev);
|
|
+
|
|
+ iowrite32(data, pdata->ring_csr_addr + offset);
|
|
+}
|
|
+
|
|
+static void xgene_enet_ring_rd32(struct xgene_enet_desc_ring *ring,
|
|
+ u32 offset, u32 *data)
|
|
+{
|
|
+ struct xgene_enet_pdata *pdata = netdev_priv(ring->ndev);
|
|
+
|
|
+ *data = ioread32(pdata->ring_csr_addr + offset);
|
|
+}
|
|
+
|
|
+static void xgene_enet_write_ring_state(struct xgene_enet_desc_ring *ring)
|
|
+{
|
|
+ int i;
|
|
+
|
|
+ xgene_enet_ring_wr32(ring, CSR_RING_CONFIG, ring->num);
|
|
+ for (i = 0; i < NUM_RING_CONFIG; i++) {
|
|
+ xgene_enet_ring_wr32(ring, CSR_RING_WR_BASE + (i * 4),
|
|
+ ring->state[i]);
|
|
+ }
|
|
+}
|
|
+
|
|
+static void xgene_enet_clr_ring_state(struct xgene_enet_desc_ring *ring)
|
|
+{
|
|
+ memset(ring->state, 0, sizeof(u32) * NUM_RING_CONFIG);
|
|
+ xgene_enet_write_ring_state(ring);
|
|
+}
|
|
+
|
|
+static void xgene_enet_set_ring_state(struct xgene_enet_desc_ring *ring)
|
|
+{
|
|
+ xgene_enet_ring_set_type(ring);
|
|
+
|
|
+ if (xgene_enet_ring_owner(ring->id) == RING_OWNER_ETH0)
|
|
+ xgene_enet_ring_set_recombbuf(ring);
|
|
+
|
|
+ xgene_enet_ring_init(ring);
|
|
+ xgene_enet_write_ring_state(ring);
|
|
+}
|
|
+
|
|
+static void xgene_enet_set_ring_id(struct xgene_enet_desc_ring *ring)
|
|
+{
|
|
+ u32 ring_id_val, ring_id_buf;
|
|
+ bool is_bufpool;
|
|
+
|
|
+ is_bufpool = xgene_enet_is_bufpool(ring->id);
|
|
+
|
|
+ ring_id_val = ring->id & GENMASK(9, 0);
|
|
+ ring_id_val |= OVERWRITE;
|
|
+
|
|
+ ring_id_buf = (ring->num << 9) & GENMASK(18, 9);
|
|
+ ring_id_buf |= PREFETCH_BUF_EN;
|
|
+ if (is_bufpool)
|
|
+ ring_id_buf |= IS_BUFFER_POOL;
|
|
+
|
|
+ xgene_enet_ring_wr32(ring, CSR_RING_ID, ring_id_val);
|
|
+ xgene_enet_ring_wr32(ring, CSR_RING_ID_BUF, ring_id_buf);
|
|
+}
|
|
+
|
|
+static void xgene_enet_clr_desc_ring_id(struct xgene_enet_desc_ring *ring)
|
|
+{
|
|
+ u32 ring_id;
|
|
+
|
|
+ ring_id = ring->id | OVERWRITE;
|
|
+ xgene_enet_ring_wr32(ring, CSR_RING_ID, ring_id);
|
|
+ xgene_enet_ring_wr32(ring, CSR_RING_ID_BUF, 0);
|
|
+}
|
|
+
|
|
+struct xgene_enet_desc_ring *xgene_enet_setup_ring(
|
|
+ struct xgene_enet_desc_ring *ring)
|
|
+{
|
|
+ u32 size = ring->size;
|
|
+ u32 i, data;
|
|
+ u64 *desc;
|
|
+ bool is_bufpool;
|
|
+
|
|
+ xgene_enet_clr_ring_state(ring);
|
|
+ xgene_enet_set_ring_state(ring);
|
|
+ xgene_enet_set_ring_id(ring);
|
|
+
|
|
+ ring->slots = xgene_enet_get_numslots(ring->id, size);
|
|
+
|
|
+ is_bufpool = xgene_enet_is_bufpool(ring->id);
|
|
+ if (is_bufpool || xgene_enet_ring_owner(ring->id) != RING_OWNER_CPU)
|
|
+ return ring;
|
|
+
|
|
+ for (i = 0; i < ring->slots; i++) {
|
|
+ desc = (u64 *)&ring->raw_desc[i];
|
|
+ desc[EMPTY_SLOT_INDEX] = EMPTY_SLOT;
|
|
+ }
|
|
+
|
|
+ xgene_enet_ring_rd32(ring, CSR_RING_NE_INT_MODE, &data);
|
|
+ data |= BIT(31 - xgene_enet_ring_bufnum(ring->id));
|
|
+ xgene_enet_ring_wr32(ring, CSR_RING_NE_INT_MODE, data);
|
|
+
|
|
+ return ring;
|
|
+}
|
|
+
|
|
+void xgene_enet_clear_ring(struct xgene_enet_desc_ring *ring)
|
|
+{
|
|
+ u32 data;
|
|
+ bool is_bufpool;
|
|
+
|
|
+ is_bufpool = xgene_enet_is_bufpool(ring->id);
|
|
+ if (is_bufpool || xgene_enet_ring_owner(ring->id) != RING_OWNER_CPU)
|
|
+ goto out;
|
|
+
|
|
+ xgene_enet_ring_rd32(ring, CSR_RING_NE_INT_MODE, &data);
|
|
+ data &= ~BIT(31 - xgene_enet_ring_bufnum(ring->id));
|
|
+ xgene_enet_ring_wr32(ring, CSR_RING_NE_INT_MODE, data);
|
|
+
|
|
+out:
|
|
+ xgene_enet_clr_desc_ring_id(ring);
|
|
+ xgene_enet_clr_ring_state(ring);
|
|
+}
|
|
+
|
|
+void xgene_enet_parse_error(struct xgene_enet_desc_ring *ring,
|
|
+ struct xgene_enet_pdata *pdata,
|
|
+ enum xgene_enet_err_code status)
|
|
+{
|
|
+ struct rtnl_link_stats64 *stats = &pdata->stats;
|
|
+
|
|
+ switch (status) {
|
|
+ case INGRESS_CRC:
|
|
+ stats->rx_crc_errors++;
|
|
+ break;
|
|
+ case INGRESS_CHECKSUM:
|
|
+ case INGRESS_CHECKSUM_COMPUTE:
|
|
+ stats->rx_errors++;
|
|
+ break;
|
|
+ case INGRESS_TRUNC_FRAME:
|
|
+ stats->rx_frame_errors++;
|
|
+ break;
|
|
+ case INGRESS_PKT_LEN:
|
|
+ stats->rx_length_errors++;
|
|
+ break;
|
|
+ case INGRESS_PKT_UNDER:
|
|
+ stats->rx_frame_errors++;
|
|
+ break;
|
|
+ case INGRESS_FIFO_OVERRUN:
|
|
+ stats->rx_fifo_errors++;
|
|
+ break;
|
|
+ default:
|
|
+ break;
|
|
+ }
|
|
+}
|
|
+
|
|
+static void xgene_enet_wr_csr(struct xgene_enet_pdata *pdata,
|
|
+ u32 offset, u32 val)
|
|
+{
|
|
+ void __iomem *addr = pdata->eth_csr_addr + offset;
|
|
+
|
|
+ iowrite32(val, addr);
|
|
+}
|
|
+
|
|
+static void xgene_enet_wr_ring_if(struct xgene_enet_pdata *pdata,
|
|
+ u32 offset, u32 val)
|
|
+{
|
|
+ void __iomem *addr = pdata->eth_ring_if_addr + offset;
|
|
+
|
|
+ iowrite32(val, addr);
|
|
+}
|
|
+
|
|
+static void xgene_enet_wr_diag_csr(struct xgene_enet_pdata *pdata,
|
|
+ u32 offset, u32 val)
|
|
+{
|
|
+ void __iomem *addr = pdata->eth_diag_csr_addr + offset;
|
|
+
|
|
+ iowrite32(val, addr);
|
|
+}
|
|
+
|
|
+static void xgene_enet_wr_mcx_csr(struct xgene_enet_pdata *pdata,
|
|
+ u32 offset, u32 val)
|
|
+{
|
|
+ void __iomem *addr = pdata->mcx_mac_csr_addr + offset;
|
|
+
|
|
+ iowrite32(val, addr);
|
|
+}
|
|
+
|
|
+static bool xgene_enet_wr_indirect(void __iomem *addr, void __iomem *wr,
|
|
+ void __iomem *cmd, void __iomem *cmd_done,
|
|
+ u32 wr_addr, u32 wr_data)
|
|
+{
|
|
+ u32 done;
|
|
+ u8 wait = 10;
|
|
+
|
|
+ iowrite32(wr_addr, addr);
|
|
+ iowrite32(wr_data, wr);
|
|
+ iowrite32(XGENE_ENET_WR_CMD, cmd);
|
|
+
|
|
+ /* wait for write command to complete */
|
|
+ while (!(done = ioread32(cmd_done)) && wait--)
|
|
+ udelay(1);
|
|
+
|
|
+ if (!done)
|
|
+ return false;
|
|
+
|
|
+ iowrite32(0, cmd);
|
|
+
|
|
+ return true;
|
|
+}
|
|
+
|
|
+static void xgene_enet_wr_mcx_mac(struct xgene_enet_pdata *pdata,
|
|
+ u32 wr_addr, u32 wr_data)
|
|
+{
|
|
+ void __iomem *addr, *wr, *cmd, *cmd_done;
|
|
+ bool ret;
|
|
+
|
|
+ addr = pdata->mcx_mac_addr + MAC_ADDR_REG_OFFSET;
|
|
+ wr = pdata->mcx_mac_addr + MAC_WRITE_REG_OFFSET;
|
|
+ cmd = pdata->mcx_mac_addr + MAC_COMMAND_REG_OFFSET;
|
|
+ cmd_done = pdata->mcx_mac_addr + MAC_COMMAND_DONE_REG_OFFSET;
|
|
+
|
|
+ ret = xgene_enet_wr_indirect(addr, wr, cmd, cmd_done, wr_addr, wr_data);
|
|
+ if (!ret)
|
|
+ netdev_err(pdata->ndev, "MCX mac write failed, addr: %04x\n",
|
|
+ wr_addr);
|
|
+}
|
|
+
|
|
+static void xgene_enet_rd_csr(struct xgene_enet_pdata *pdata,
|
|
+ u32 offset, u32 *val)
|
|
+{
|
|
+ void __iomem *addr = pdata->eth_csr_addr + offset;
|
|
+
|
|
+ *val = ioread32(addr);
|
|
+}
|
|
+
|
|
+static void xgene_enet_rd_diag_csr(struct xgene_enet_pdata *pdata,
|
|
+ u32 offset, u32 *val)
|
|
+{
|
|
+ void __iomem *addr = pdata->eth_diag_csr_addr + offset;
|
|
+
|
|
+ *val = ioread32(addr);
|
|
+}
|
|
+
|
|
+static void xgene_enet_rd_mcx_csr(struct xgene_enet_pdata *pdata,
|
|
+ u32 offset, u32 *val)
|
|
+{
|
|
+ void __iomem *addr = pdata->mcx_mac_csr_addr + offset;
|
|
+
|
|
+ *val = ioread32(addr);
|
|
+}
|
|
+
|
|
+static bool xgene_enet_rd_indirect(void __iomem *addr, void __iomem *rd,
|
|
+ void __iomem *cmd, void __iomem *cmd_done,
|
|
+ u32 rd_addr, u32 *rd_data)
|
|
+{
|
|
+ u32 done;
|
|
+ u8 wait = 10;
|
|
+
|
|
+ iowrite32(rd_addr, addr);
|
|
+ iowrite32(XGENE_ENET_RD_CMD, cmd);
|
|
+
|
|
+ /* wait for read command to complete */
|
|
+ while (!(done = ioread32(cmd_done)) && wait--)
|
|
+ udelay(1);
|
|
+
|
|
+ if (!done)
|
|
+ return false;
|
|
+
|
|
+ *rd_data = ioread32(rd);
|
|
+ iowrite32(0, cmd);
|
|
+
|
|
+ return true;
|
|
+}
|
|
+
|
|
+static void xgene_enet_rd_mcx_mac(struct xgene_enet_pdata *pdata,
|
|
+ u32 rd_addr, u32 *rd_data)
|
|
+{
|
|
+ void __iomem *addr, *rd, *cmd, *cmd_done;
|
|
+ bool ret;
|
|
+
|
|
+ addr = pdata->mcx_mac_addr + MAC_ADDR_REG_OFFSET;
|
|
+ rd = pdata->mcx_mac_addr + MAC_READ_REG_OFFSET;
|
|
+ cmd = pdata->mcx_mac_addr + MAC_COMMAND_REG_OFFSET;
|
|
+ cmd_done = pdata->mcx_mac_addr + MAC_COMMAND_DONE_REG_OFFSET;
|
|
+
|
|
+ ret = xgene_enet_rd_indirect(addr, rd, cmd, cmd_done, rd_addr, rd_data);
|
|
+ if (!ret)
|
|
+ netdev_err(pdata->ndev, "MCX mac read failed, addr: %04x\n",
|
|
+ rd_addr);
|
|
+}
|
|
+
|
|
+static int xgene_mii_phy_write(struct xgene_enet_pdata *pdata, int phy_id,
|
|
+ u32 reg, u16 data)
|
|
+{
|
|
+ u32 addr = 0, wr_data = 0;
|
|
+ u32 done;
|
|
+ u8 wait = 10;
|
|
+
|
|
+ PHY_ADDR_SET(&addr, phy_id);
|
|
+ REG_ADDR_SET(&addr, reg);
|
|
+ xgene_enet_wr_mcx_mac(pdata, MII_MGMT_ADDRESS_ADDR, addr);
|
|
+
|
|
+ PHY_CONTROL_SET(&wr_data, data);
|
|
+ xgene_enet_wr_mcx_mac(pdata, MII_MGMT_CONTROL_ADDR, wr_data);
|
|
+ do {
|
|
+ usleep_range(5, 10);
|
|
+ xgene_enet_rd_mcx_mac(pdata, MII_MGMT_INDICATORS_ADDR, &done);
|
|
+ } while ((done & BUSY_MASK) && wait--);
|
|
+
|
|
+ if (done & BUSY_MASK) {
|
|
+ netdev_err(pdata->ndev, "MII_MGMT write failed\n");
|
|
+ return -1;
|
|
+ }
|
|
+
|
|
+ return 0;
|
|
+}
|
|
+
|
|
+static int xgene_mii_phy_read(struct xgene_enet_pdata *pdata,
|
|
+ u8 phy_id, u32 reg)
|
|
+{
|
|
+ u32 addr = 0;
|
|
+ u32 data, done;
|
|
+ u8 wait = 10;
|
|
+
|
|
+ PHY_ADDR_SET(&addr, phy_id);
|
|
+ REG_ADDR_SET(&addr, reg);
|
|
+ xgene_enet_wr_mcx_mac(pdata, MII_MGMT_ADDRESS_ADDR, addr);
|
|
+ xgene_enet_wr_mcx_mac(pdata, MII_MGMT_COMMAND_ADDR, READ_CYCLE_MASK);
|
|
+ do {
|
|
+ usleep_range(5, 10);
|
|
+ xgene_enet_rd_mcx_mac(pdata, MII_MGMT_INDICATORS_ADDR, &done);
|
|
+ } while ((done & BUSY_MASK) && wait--);
|
|
+
|
|
+ if (done & BUSY_MASK) {
|
|
+ netdev_err(pdata->ndev, "MII_MGMT read failed\n");
|
|
+ return -1;
|
|
+ }
|
|
+
|
|
+ xgene_enet_rd_mcx_mac(pdata, MII_MGMT_STATUS_ADDR, &data);
|
|
+ xgene_enet_wr_mcx_mac(pdata, MII_MGMT_COMMAND_ADDR, 0);
|
|
+
|
|
+ return data;
|
|
+}
|
|
+
|
|
+void xgene_gmac_set_mac_addr(struct xgene_enet_pdata *pdata)
|
|
+{
|
|
+ u32 addr0, addr1;
|
|
+ u8 *dev_addr = pdata->ndev->dev_addr;
|
|
+
|
|
+ addr0 = (dev_addr[3] << 24) | (dev_addr[2] << 16) |
|
|
+ (dev_addr[1] << 8) | dev_addr[0];
|
|
+ addr1 = (dev_addr[5] << 24) | (dev_addr[4] << 16);
|
|
+ addr1 |= pdata->phy_addr & 0xFFFF;
|
|
+
|
|
+ xgene_enet_wr_mcx_mac(pdata, STATION_ADDR0_ADDR, addr0);
|
|
+ xgene_enet_wr_mcx_mac(pdata, STATION_ADDR1_ADDR, addr1);
|
|
+}
|
|
+
|
|
+static int xgene_enet_ecc_init(struct xgene_enet_pdata *pdata)
|
|
+{
|
|
+ struct net_device *ndev = pdata->ndev;
|
|
+ u32 data;
|
|
+ u8 wait = 10;
|
|
+
|
|
+ xgene_enet_wr_diag_csr(pdata, ENET_CFG_MEM_RAM_SHUTDOWN_ADDR, 0x0);
|
|
+ do {
|
|
+ usleep_range(100, 110);
|
|
+ xgene_enet_rd_diag_csr(pdata, ENET_BLOCK_MEM_RDY_ADDR, &data);
|
|
+ } while ((data != 0xffffffff) && wait--);
|
|
+
|
|
+ if (data != 0xffffffff) {
|
|
+ netdev_err(ndev, "Failed to release memory from shutdown\n");
|
|
+ return -ENODEV;
|
|
+ }
|
|
+
|
|
+ return 0;
|
|
+}
|
|
+
|
|
+void xgene_gmac_reset(struct xgene_enet_pdata *pdata)
|
|
+{
|
|
+ xgene_enet_wr_mcx_mac(pdata, MAC_CONFIG_1_ADDR, SOFT_RESET1);
|
|
+ xgene_enet_wr_mcx_mac(pdata, MAC_CONFIG_1_ADDR, 0);
|
|
+}
|
|
+
|
|
+void xgene_gmac_init(struct xgene_enet_pdata *pdata, int speed)
|
|
+{
|
|
+ u32 value, mc2;
|
|
+ u32 intf_ctl, rgmii;
|
|
+ u32 icm0, icm2;
|
|
+
|
|
+ xgene_gmac_reset(pdata);
|
|
+
|
|
+ xgene_enet_rd_mcx_csr(pdata, ICM_CONFIG0_REG_0_ADDR, &icm0);
|
|
+ xgene_enet_rd_mcx_csr(pdata, ICM_CONFIG2_REG_0_ADDR, &icm2);
|
|
+ xgene_enet_rd_mcx_mac(pdata, MAC_CONFIG_2_ADDR, &mc2);
|
|
+ xgene_enet_rd_mcx_mac(pdata, INTERFACE_CONTROL_ADDR, &intf_ctl);
|
|
+ xgene_enet_rd_csr(pdata, RGMII_REG_0_ADDR, &rgmii);
|
|
+
|
|
+ switch (speed) {
|
|
+ case SPEED_10:
|
|
+ ENET_INTERFACE_MODE2_SET(&mc2, 1);
|
|
+ CFG_MACMODE_SET(&icm0, 0);
|
|
+ CFG_WAITASYNCRD_SET(&icm2, 500);
|
|
+ rgmii &= ~CFG_SPEED_1250;
|
|
+ break;
|
|
+ case SPEED_100:
|
|
+ ENET_INTERFACE_MODE2_SET(&mc2, 1);
|
|
+ intf_ctl |= ENET_LHD_MODE;
|
|
+ CFG_MACMODE_SET(&icm0, 1);
|
|
+ CFG_WAITASYNCRD_SET(&icm2, 80);
|
|
+ rgmii &= ~CFG_SPEED_1250;
|
|
+ break;
|
|
+ default:
|
|
+ ENET_INTERFACE_MODE2_SET(&mc2, 2);
|
|
+ intf_ctl |= ENET_GHD_MODE;
|
|
+ CFG_TXCLK_MUXSEL0_SET(&rgmii, 4);
|
|
+ xgene_enet_rd_csr(pdata, DEBUG_REG_ADDR, &value);
|
|
+ value |= CFG_BYPASS_UNISEC_TX | CFG_BYPASS_UNISEC_RX;
|
|
+ xgene_enet_wr_csr(pdata, DEBUG_REG_ADDR, value);
|
|
+ break;
|
|
+ }
|
|
+
|
|
+ mc2 |= FULL_DUPLEX2;
|
|
+ xgene_enet_wr_mcx_mac(pdata, MAC_CONFIG_2_ADDR, mc2);
|
|
+ xgene_enet_wr_mcx_mac(pdata, INTERFACE_CONTROL_ADDR, intf_ctl);
|
|
+
|
|
+ xgene_gmac_set_mac_addr(pdata);
|
|
+
|
|
+ /* Adjust MDC clock frequency */
|
|
+ xgene_enet_rd_mcx_mac(pdata, MII_MGMT_CONFIG_ADDR, &value);
|
|
+ MGMT_CLOCK_SEL_SET(&value, 7);
|
|
+ xgene_enet_wr_mcx_mac(pdata, MII_MGMT_CONFIG_ADDR, value);
|
|
+
|
|
+ /* Enable drop if bufpool not available */
|
|
+ xgene_enet_rd_csr(pdata, RSIF_CONFIG_REG_ADDR, &value);
|
|
+ value |= CFG_RSIF_FPBUFF_TIMEOUT_EN;
|
|
+ xgene_enet_wr_csr(pdata, RSIF_CONFIG_REG_ADDR, value);
|
|
+
|
|
+ /* Rtype should be copied from FP */
|
|
+ xgene_enet_wr_csr(pdata, RSIF_RAM_DBG_REG0_ADDR, 0);
|
|
+ xgene_enet_wr_csr(pdata, RGMII_REG_0_ADDR, rgmii);
|
|
+
|
|
+ /* Rx-Tx traffic resume */
|
|
+ xgene_enet_wr_csr(pdata, CFG_LINK_AGGR_RESUME_0_ADDR, TX_PORT0);
|
|
+
|
|
+ xgene_enet_wr_mcx_csr(pdata, ICM_CONFIG0_REG_0_ADDR, icm0);
|
|
+ xgene_enet_wr_mcx_csr(pdata, ICM_CONFIG2_REG_0_ADDR, icm2);
|
|
+
|
|
+ xgene_enet_rd_mcx_csr(pdata, RX_DV_GATE_REG_0_ADDR, &value);
|
|
+ value &= ~TX_DV_GATE_EN0;
|
|
+ value &= ~RX_DV_GATE_EN0;
|
|
+ value |= RESUME_RX0;
|
|
+ xgene_enet_wr_mcx_csr(pdata, RX_DV_GATE_REG_0_ADDR, value);
|
|
+
|
|
+ xgene_enet_wr_csr(pdata, CFG_BYPASS_ADDR, RESUME_TX);
|
|
+}
|
|
+
|
|
+static void xgene_enet_config_ring_if_assoc(struct xgene_enet_pdata *pdata)
|
|
+{
|
|
+ u32 val = 0xffffffff;
|
|
+
|
|
+ xgene_enet_wr_ring_if(pdata, ENET_CFGSSQMIWQASSOC_ADDR, val);
|
|
+ xgene_enet_wr_ring_if(pdata, ENET_CFGSSQMIFPQASSOC_ADDR, val);
|
|
+ xgene_enet_wr_ring_if(pdata, ENET_CFGSSQMIQMLITEWQASSOC_ADDR, val);
|
|
+ xgene_enet_wr_ring_if(pdata, ENET_CFGSSQMIQMLITEFPQASSOC_ADDR, val);
|
|
+}
|
|
+
|
|
+void xgene_enet_cle_bypass(struct xgene_enet_pdata *pdata,
|
|
+ u32 dst_ring_num, u16 bufpool_id)
|
|
+{
|
|
+ u32 cb;
|
|
+ u32 fpsel;
|
|
+
|
|
+ fpsel = xgene_enet_ring_bufnum(bufpool_id) - 0x20;
|
|
+
|
|
+ xgene_enet_rd_csr(pdata, CLE_BYPASS_REG0_0_ADDR, &cb);
|
|
+ cb |= CFG_CLE_BYPASS_EN0;
|
|
+ CFG_CLE_IP_PROTOCOL0_SET(&cb, 3);
|
|
+ xgene_enet_wr_csr(pdata, CLE_BYPASS_REG0_0_ADDR, cb);
|
|
+
|
|
+ xgene_enet_rd_csr(pdata, CLE_BYPASS_REG1_0_ADDR, &cb);
|
|
+ CFG_CLE_DSTQID0_SET(&cb, dst_ring_num);
|
|
+ CFG_CLE_FPSEL0_SET(&cb, fpsel);
|
|
+ xgene_enet_wr_csr(pdata, CLE_BYPASS_REG1_0_ADDR, cb);
|
|
+}
|
|
+
|
|
+void xgene_gmac_rx_enable(struct xgene_enet_pdata *pdata)
|
|
+{
|
|
+ u32 data;
|
|
+
|
|
+ xgene_enet_rd_mcx_mac(pdata, MAC_CONFIG_1_ADDR, &data);
|
|
+ xgene_enet_wr_mcx_mac(pdata, MAC_CONFIG_1_ADDR, data | RX_EN);
|
|
+}
|
|
+
|
|
+void xgene_gmac_tx_enable(struct xgene_enet_pdata *pdata)
|
|
+{
|
|
+ u32 data;
|
|
+
|
|
+ xgene_enet_rd_mcx_mac(pdata, MAC_CONFIG_1_ADDR, &data);
|
|
+ xgene_enet_wr_mcx_mac(pdata, MAC_CONFIG_1_ADDR, data | TX_EN);
|
|
+}
|
|
+
|
|
+void xgene_gmac_rx_disable(struct xgene_enet_pdata *pdata)
|
|
+{
|
|
+ u32 data;
|
|
+
|
|
+ xgene_enet_rd_mcx_mac(pdata, MAC_CONFIG_1_ADDR, &data);
|
|
+ xgene_enet_wr_mcx_mac(pdata, MAC_CONFIG_1_ADDR, data & ~RX_EN);
|
|
+}
|
|
+
|
|
+void xgene_gmac_tx_disable(struct xgene_enet_pdata *pdata)
|
|
+{
|
|
+ u32 data;
|
|
+
|
|
+ xgene_enet_rd_mcx_mac(pdata, MAC_CONFIG_1_ADDR, &data);
|
|
+ xgene_enet_wr_mcx_mac(pdata, MAC_CONFIG_1_ADDR, data & ~TX_EN);
|
|
+}
|
|
+
|
|
+void xgene_enet_reset(struct xgene_enet_pdata *pdata)
|
|
+{
|
|
+ u32 val;
|
|
+
|
|
+ clk_prepare_enable(pdata->clk);
|
|
+ clk_disable_unprepare(pdata->clk);
|
|
+ clk_prepare_enable(pdata->clk);
|
|
+ xgene_enet_ecc_init(pdata);
|
|
+ xgene_enet_config_ring_if_assoc(pdata);
|
|
+
|
|
+ /* Enable auto-incr for scanning */
|
|
+ xgene_enet_rd_mcx_mac(pdata, MII_MGMT_CONFIG_ADDR, &val);
|
|
+ val |= SCAN_AUTO_INCR;
|
|
+ MGMT_CLOCK_SEL_SET(&val, 1);
|
|
+ xgene_enet_wr_mcx_mac(pdata, MII_MGMT_CONFIG_ADDR, val);
|
|
+}
|
|
+
|
|
+void xgene_gport_shutdown(struct xgene_enet_pdata *pdata)
|
|
+{
|
|
+ clk_disable_unprepare(pdata->clk);
|
|
+}
|
|
+
|
|
+static int xgene_enet_mdio_read(struct mii_bus *bus, int mii_id, int regnum)
|
|
+{
|
|
+ struct xgene_enet_pdata *pdata = bus->priv;
|
|
+ u32 val;
|
|
+
|
|
+ val = xgene_mii_phy_read(pdata, mii_id, regnum);
|
|
+ netdev_dbg(pdata->ndev, "mdio_rd: bus=%d reg=%d val=%x\n",
|
|
+ mii_id, regnum, val);
|
|
+
|
|
+ return val;
|
|
+}
|
|
+
|
|
+static int xgene_enet_mdio_write(struct mii_bus *bus, int mii_id, int regnum,
|
|
+ u16 val)
|
|
+{
|
|
+ struct xgene_enet_pdata *pdata = bus->priv;
|
|
+ int ret;
|
|
+
|
|
+ netdev_dbg(pdata->ndev, "mdio_wr: bus=%d reg=%d val=%x\n",
|
|
+ mii_id, regnum, val);
|
|
+ ret = xgene_mii_phy_write(pdata, mii_id, regnum, val);
|
|
+
|
|
+ return ret;
|
|
+}
|
|
+
|
|
+static void xgene_enet_adjust_link(struct net_device *ndev)
|
|
+{
|
|
+ struct xgene_enet_pdata *pdata = netdev_priv(ndev);
|
|
+ struct phy_device *phydev = pdata->phy_dev;
|
|
+
|
|
+ if (phydev->link) {
|
|
+ if (pdata->phy_speed != phydev->speed) {
|
|
+ xgene_gmac_init(pdata, phydev->speed);
|
|
+ xgene_gmac_rx_enable(pdata);
|
|
+ xgene_gmac_tx_enable(pdata);
|
|
+ pdata->phy_speed = phydev->speed;
|
|
+ phy_print_status(phydev);
|
|
+ }
|
|
+ } else {
|
|
+ xgene_gmac_rx_disable(pdata);
|
|
+ xgene_gmac_tx_disable(pdata);
|
|
+ pdata->phy_speed = SPEED_UNKNOWN;
|
|
+ phy_print_status(phydev);
|
|
+ }
|
|
+}
|
|
+
|
|
+static int xgene_enet_phy_connect(struct net_device *ndev)
|
|
+{
|
|
+ struct xgene_enet_pdata *pdata = netdev_priv(ndev);
|
|
+ struct device_node *phy_np;
|
|
+ struct phy_device *phy_dev;
|
|
+ struct device *dev = &pdata->pdev->dev;
|
|
+
|
|
+ phy_np = of_parse_phandle(dev->of_node, "phy-handle", 0);
|
|
+ if (!phy_np) {
|
|
+ netdev_dbg(ndev, "No phy-handle found\n");
|
|
+ return -ENODEV;
|
|
+ }
|
|
+
|
|
+ phy_dev = of_phy_connect(ndev, phy_np, &xgene_enet_adjust_link,
|
|
+ 0, pdata->phy_mode);
|
|
+ if (!phy_dev) {
|
|
+ netdev_err(ndev, "Could not connect to PHY\n");
|
|
+ return -ENODEV;
|
|
+ }
|
|
+
|
|
+ pdata->phy_speed = SPEED_UNKNOWN;
|
|
+ phy_dev->supported &= ~SUPPORTED_10baseT_Half &
|
|
+ ~SUPPORTED_100baseT_Half &
|
|
+ ~SUPPORTED_1000baseT_Half;
|
|
+ phy_dev->advertising = phy_dev->supported;
|
|
+ pdata->phy_dev = phy_dev;
|
|
+
|
|
+ return 0;
|
|
+}
|
|
+
|
|
+int xgene_enet_mdio_config(struct xgene_enet_pdata *pdata)
|
|
+{
|
|
+ struct net_device *ndev = pdata->ndev;
|
|
+ struct device *dev = &pdata->pdev->dev;
|
|
+ struct device_node *child_np;
|
|
+ struct device_node *mdio_np = NULL;
|
|
+ struct mii_bus *mdio_bus;
|
|
+ int ret;
|
|
+
|
|
+ for_each_child_of_node(dev->of_node, child_np) {
|
|
+ if (of_device_is_compatible(child_np, "apm,xgene-mdio")) {
|
|
+ mdio_np = child_np;
|
|
+ break;
|
|
+ }
|
|
+ }
|
|
+
|
|
+ if (!mdio_np) {
|
|
+ netdev_dbg(ndev, "No mdio node in the dts\n");
|
|
+ return -1;
|
|
+ }
|
|
+
|
|
+ mdio_bus = mdiobus_alloc();
|
|
+ if (!mdio_bus)
|
|
+ return -ENOMEM;
|
|
+
|
|
+ mdio_bus->name = "APM X-Gene MDIO bus";
|
|
+ mdio_bus->read = xgene_enet_mdio_read;
|
|
+ mdio_bus->write = xgene_enet_mdio_write;
|
|
+ snprintf(mdio_bus->id, MII_BUS_ID_SIZE, "%s-%s", "xgene-mii",
|
|
+ ndev->name);
|
|
+
|
|
+ mdio_bus->priv = pdata;
|
|
+ mdio_bus->parent = &ndev->dev;
|
|
+
|
|
+ ret = of_mdiobus_register(mdio_bus, mdio_np);
|
|
+ if (ret) {
|
|
+ netdev_err(ndev, "Failed to register MDIO bus\n");
|
|
+ goto err;
|
|
+ }
|
|
+ pdata->mdio_bus = mdio_bus;
|
|
+
|
|
+ ret = xgene_enet_phy_connect(ndev);
|
|
+ if (ret)
|
|
+ goto err;
|
|
+
|
|
+ return ret;
|
|
+
|
|
+err:
|
|
+ mdiobus_free(mdio_bus);
|
|
+
|
|
+ return ret;
|
|
+}
|
|
+
|
|
+int xgene_enet_mdio_remove(struct xgene_enet_pdata *pdata)
|
|
+{
|
|
+ struct mii_bus *mdio_bus;
|
|
+
|
|
+ mdio_bus = pdata->mdio_bus;
|
|
+ mdiobus_unregister(mdio_bus);
|
|
+ mdiobus_free(mdio_bus);
|
|
+ pdata->mdio_bus = NULL;
|
|
+
|
|
+ return 0;
|
|
+}
|
|
diff --git a/drivers/net/ethernet/apm/xgene/xgene_enet_hw.h b/drivers/net/ethernet/apm/xgene/xgene_enet_hw.h
|
|
new file mode 100644
|
|
index 0000000..2041313
|
|
--- /dev/null
|
|
+++ b/drivers/net/ethernet/apm/xgene/xgene_enet_hw.h
|
|
@@ -0,0 +1,375 @@
|
|
+/* Applied Micro X-Gene SoC Ethernet Driver
|
|
+ *
|
|
+ * Copyright (c) 2014, Applied Micro Circuits Corporation
|
|
+ * Authors: Iyappan Subramanian <isubramanian@apm.com>
|
|
+ * Ravi Patel <rapatel@apm.com>
|
|
+ * Keyur Chudgar <kchudgar@apm.com>
|
|
+ *
|
|
+ * This program is free software; you can redistribute it and/or modify it
|
|
+ * under the terms of the GNU General Public License as published by the
|
|
+ * Free Software Foundation; either version 2 of the License, or (at your
|
|
+ * option) any later version.
|
|
+ *
|
|
+ * This program is distributed in the hope that it will be useful,
|
|
+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
+ * GNU General Public License for more details.
|
|
+ *
|
|
+ * You should have received a copy of the GNU General Public License
|
|
+ * along with this program. If not, see <http://www.gnu.org/licenses/>.
|
|
+ */
|
|
+
|
|
+#ifndef __XGENE_ENET_HW_H__
|
|
+#define __XGENE_ENET_HW_H__
|
|
+
|
|
+#include "xgene_enet_main.h"
|
|
+
|
|
+struct xgene_enet_pdata;
|
|
+struct xgene_enet_stats;
|
|
+
|
|
+/* clears and then set bits */
|
|
+static inline void xgene_set_bits(u32 *dst, u32 val, u32 start, u32 len)
|
|
+{
|
|
+ u32 end = start + len - 1;
|
|
+ u32 mask = GENMASK(end, start);
|
|
+
|
|
+ *dst &= ~mask;
|
|
+ *dst |= (val << start) & mask;
|
|
+}
|
|
+
|
|
+static inline u32 xgene_get_bits(u32 val, u32 start, u32 end)
|
|
+{
|
|
+ return (val & GENMASK(end, start)) >> start;
|
|
+}
|
|
+
|
|
+#define CSR_RING_ID 0x0008
|
|
+#define OVERWRITE BIT(31)
|
|
+#define IS_BUFFER_POOL BIT(20)
|
|
+#define PREFETCH_BUF_EN BIT(21)
|
|
+#define CSR_RING_ID_BUF 0x000c
|
|
+#define CSR_RING_NE_INT_MODE 0x017c
|
|
+#define CSR_RING_CONFIG 0x006c
|
|
+#define CSR_RING_WR_BASE 0x0070
|
|
+#define NUM_RING_CONFIG 5
|
|
+#define BUFPOOL_MODE 3
|
|
+#define RM3 3
|
|
+#define INC_DEC_CMD_ADDR 0x002c
|
|
+#define UDP_HDR_SIZE 2
|
|
+#define BUF_LEN_CODE_2K 0x5000
|
|
+
|
|
+#define CREATE_MASK(pos, len) GENMASK((pos)+(len)-1, (pos))
|
|
+#define CREATE_MASK_ULL(pos, len) GENMASK_ULL((pos)+(len)-1, (pos))
|
|
+
|
|
+/* Empty slot soft signature */
|
|
+#define EMPTY_SLOT_INDEX 1
|
|
+#define EMPTY_SLOT ~0ULL
|
|
+
|
|
+#define WORK_DESC_SIZE 32
|
|
+#define BUFPOOL_DESC_SIZE 16
|
|
+
|
|
+#define RING_OWNER_MASK GENMASK(9, 6)
|
|
+#define RING_BUFNUM_MASK GENMASK(5, 0)
|
|
+
|
|
+#define SELTHRSH_POS 3
|
|
+#define SELTHRSH_LEN 3
|
|
+#define RINGADDRL_POS 5
|
|
+#define RINGADDRL_LEN 27
|
|
+#define RINGADDRH_POS 0
|
|
+#define RINGADDRH_LEN 6
|
|
+#define RINGSIZE_POS 23
|
|
+#define RINGSIZE_LEN 3
|
|
+#define RINGTYPE_POS 19
|
|
+#define RINGTYPE_LEN 2
|
|
+#define RINGMODE_POS 20
|
|
+#define RINGMODE_LEN 3
|
|
+#define RECOMTIMEOUTL_POS 28
|
|
+#define RECOMTIMEOUTL_LEN 3
|
|
+#define RECOMTIMEOUTH_POS 0
|
|
+#define RECOMTIMEOUTH_LEN 2
|
|
+#define NUMMSGSINQ_POS 1
|
|
+#define NUMMSGSINQ_LEN 16
|
|
+#define ACCEPTLERR BIT(19)
|
|
+#define QCOHERENT BIT(4)
|
|
+#define RECOMBBUF BIT(27)
|
|
+
|
|
+#define BLOCK_ETH_CSR_OFFSET 0x2000
|
|
+#define BLOCK_ETH_RING_IF_OFFSET 0x9000
|
|
+#define BLOCK_ETH_CLKRST_CSR_OFFSET 0xC000
|
|
+#define BLOCK_ETH_DIAG_CSR_OFFSET 0xD000
|
|
+
|
|
+#define BLOCK_ETH_MAC_OFFSET 0x0000
|
|
+#define BLOCK_ETH_STATS_OFFSET 0x0014
|
|
+#define BLOCK_ETH_MAC_CSR_OFFSET 0x2800
|
|
+
|
|
+#define MAC_ADDR_REG_OFFSET 0x00
|
|
+#define MAC_COMMAND_REG_OFFSET 0x04
|
|
+#define MAC_WRITE_REG_OFFSET 0x08
|
|
+#define MAC_READ_REG_OFFSET 0x0c
|
|
+#define MAC_COMMAND_DONE_REG_OFFSET 0x10
|
|
+
|
|
+#define STAT_ADDR_REG_OFFSET 0x00
|
|
+#define STAT_COMMAND_REG_OFFSET 0x04
|
|
+#define STAT_WRITE_REG_OFFSET 0x08
|
|
+#define STAT_READ_REG_OFFSET 0x0c
|
|
+#define STAT_COMMAND_DONE_REG_OFFSET 0x10
|
|
+
|
|
+#define MII_MGMT_CONFIG_ADDR 0x20
|
|
+#define MII_MGMT_COMMAND_ADDR 0x24
|
|
+#define MII_MGMT_ADDRESS_ADDR 0x28
|
|
+#define MII_MGMT_CONTROL_ADDR 0x2c
|
|
+#define MII_MGMT_STATUS_ADDR 0x30
|
|
+#define MII_MGMT_INDICATORS_ADDR 0x34
|
|
+
|
|
+#define BUSY_MASK BIT(0)
|
|
+#define READ_CYCLE_MASK BIT(0)
|
|
+#define PHY_CONTROL_SET(dst, val) xgene_set_bits(dst, val, 0, 16)
|
|
+
|
|
+#define ENET_SPARE_CFG_REG_ADDR 0x0750
|
|
+#define RSIF_CONFIG_REG_ADDR 0x0010
|
|
+#define RSIF_RAM_DBG_REG0_ADDR 0x0048
|
|
+#define RGMII_REG_0_ADDR 0x07e0
|
|
+#define CFG_LINK_AGGR_RESUME_0_ADDR 0x07c8
|
|
+#define DEBUG_REG_ADDR 0x0700
|
|
+#define CFG_BYPASS_ADDR 0x0294
|
|
+#define CLE_BYPASS_REG0_0_ADDR 0x0490
|
|
+#define CLE_BYPASS_REG1_0_ADDR 0x0494
|
|
+#define CFG_RSIF_FPBUFF_TIMEOUT_EN BIT(31)
|
|
+#define RESUME_TX BIT(0)
|
|
+#define CFG_SPEED_1250 BIT(24)
|
|
+#define TX_PORT0 BIT(0)
|
|
+#define CFG_BYPASS_UNISEC_TX BIT(2)
|
|
+#define CFG_BYPASS_UNISEC_RX BIT(1)
|
|
+#define CFG_CLE_BYPASS_EN0 BIT(31)
|
|
+#define CFG_TXCLK_MUXSEL0_SET(dst, val) xgene_set_bits(dst, val, 29, 3)
|
|
+
|
|
+#define CFG_CLE_IP_PROTOCOL0_SET(dst, val) xgene_set_bits(dst, val, 16, 2)
|
|
+#define CFG_CLE_DSTQID0_SET(dst, val) xgene_set_bits(dst, val, 0, 12)
|
|
+#define CFG_CLE_FPSEL0_SET(dst, val) xgene_set_bits(dst, val, 16, 4)
|
|
+#define CFG_MACMODE_SET(dst, val) xgene_set_bits(dst, val, 18, 2)
|
|
+#define CFG_WAITASYNCRD_SET(dst, val) xgene_set_bits(dst, val, 0, 16)
|
|
+#define ICM_CONFIG0_REG_0_ADDR 0x0400
|
|
+#define ICM_CONFIG2_REG_0_ADDR 0x0410
|
|
+#define RX_DV_GATE_REG_0_ADDR 0x05fc
|
|
+#define TX_DV_GATE_EN0 BIT(2)
|
|
+#define RX_DV_GATE_EN0 BIT(1)
|
|
+#define RESUME_RX0 BIT(0)
|
|
+#define ENET_CFGSSQMIWQASSOC_ADDR 0xe0
|
|
+#define ENET_CFGSSQMIFPQASSOC_ADDR 0xdc
|
|
+#define ENET_CFGSSQMIQMLITEFPQASSOC_ADDR 0xf0
|
|
+#define ENET_CFGSSQMIQMLITEWQASSOC_ADDR 0xf4
|
|
+#define ENET_CFG_MEM_RAM_SHUTDOWN_ADDR 0x70
|
|
+#define ENET_BLOCK_MEM_RDY_ADDR 0x74
|
|
+#define MAC_CONFIG_1_ADDR 0x00
|
|
+#define MAC_CONFIG_2_ADDR 0x04
|
|
+#define MAX_FRAME_LEN_ADDR 0x10
|
|
+#define INTERFACE_CONTROL_ADDR 0x38
|
|
+#define STATION_ADDR0_ADDR 0x40
|
|
+#define STATION_ADDR1_ADDR 0x44
|
|
+#define PHY_ADDR_SET(dst, val) xgene_set_bits(dst, val, 8, 5)
|
|
+#define REG_ADDR_SET(dst, val) xgene_set_bits(dst, val, 0, 5)
|
|
+#define ENET_INTERFACE_MODE2_SET(dst, val) xgene_set_bits(dst, val, 8, 2)
|
|
+#define MGMT_CLOCK_SEL_SET(dst, val) xgene_set_bits(dst, val, 0, 3)
|
|
+#define SOFT_RESET1 BIT(31)
|
|
+#define TX_EN BIT(0)
|
|
+#define RX_EN BIT(2)
|
|
+#define ENET_LHD_MODE BIT(25)
|
|
+#define ENET_GHD_MODE BIT(26)
|
|
+#define FULL_DUPLEX2 BIT(0)
|
|
+#define SCAN_AUTO_INCR BIT(5)
|
|
+#define TBYT_ADDR 0x38
|
|
+#define TPKT_ADDR 0x39
|
|
+#define TDRP_ADDR 0x45
|
|
+#define TFCS_ADDR 0x47
|
|
+#define TUND_ADDR 0x4a
|
|
+
|
|
+#define TSO_IPPROTO_TCP 1
|
|
+#define FULL_DUPLEX 2
|
|
+
|
|
+#define USERINFO_POS 0
|
|
+#define USERINFO_LEN 32
|
|
+#define FPQNUM_POS 32
|
|
+#define FPQNUM_LEN 12
|
|
+#define LERR_POS 60
|
|
+#define LERR_LEN 3
|
|
+#define STASH_POS 52
|
|
+#define STASH_LEN 2
|
|
+#define BUFDATALEN_POS 48
|
|
+#define BUFDATALEN_LEN 12
|
|
+#define DATAADDR_POS 0
|
|
+#define DATAADDR_LEN 42
|
|
+#define COHERENT_POS 63
|
|
+#define HENQNUM_POS 48
|
|
+#define HENQNUM_LEN 12
|
|
+#define TYPESEL_POS 44
|
|
+#define TYPESEL_LEN 4
|
|
+#define ETHHDR_POS 12
|
|
+#define IC_POS 35 /* Insert CRC */
|
|
+#define TCPHDR_POS 0
|
|
+#define TCPHDR_LEN 6
|
|
+#define IPHDR_POS 6
|
|
+#define IPHDR_LEN 6
|
|
+#define EC_POS 22 /* Enable checksum */
|
|
+#define IS_POS 24 /* IP protocol select */
|
|
+
|
|
+#define DATAADDR_MASK CREATE_MASK_ULL(DATAADDR_POS, DATAADDR_LEN)
|
|
+#define BUFDATALEN_MASK CREATE_MASK_ULL(BUFDATALEN_POS, BUFDATALEN_LEN)
|
|
+#define USERINFO_MASK CREATE_MASK_ULL(USERINFO_POS, USERINFO_LEN)
|
|
+#define FPQNUM_MASK CREATE_MASK_ULL(FPQNUM_POS, FPQNUM_LEN)
|
|
+#define LERR_MASK CREATE_MASK_ULL(LERR_POS, LERR_LEN)
|
|
+#define STASHING_MASK CREATE_MASK_ULL(STASH_POS, STASH_LEN)
|
|
+#define COHERENT_MASK BIT_ULL(COHERENT_POS)
|
|
+#define HENQNUM_MASK CREATE_MASK_ULL(HENQNUM_POS, HENQNUM_LEN)
|
|
+#define TCPHDR_MASK CREATE_MASK(TCPHDR_POS, TCPHDR_LEN)
|
|
+#define IPHDR_MASK CREATE_MASK(IPHDR_POS, IPHDR_LEN)
|
|
+#define EC_MASK BIT(EC_POS)
|
|
+#define IS_MASK BIT(IS_POS)
|
|
+#define INSERT_CRC BIT_ULL(IC_POS)
|
|
+#define TYPE_ETH_WORK_MESSAGE BIT_ULL(44)
|
|
+
|
|
+struct xgene_enet_raw_desc {
|
|
+ u64 m0;
|
|
+ u64 m1;
|
|
+ u64 m2;
|
|
+ u64 m3;
|
|
+};
|
|
+
|
|
+struct xgene_enet_raw_desc16 {
|
|
+ u64 m0;
|
|
+ u64 m1;
|
|
+};
|
|
+
|
|
+static inline void xgene_enet_cpu_to_le64(void *desc_ptr, int count)
|
|
+{
|
|
+ u64 *desc = desc_ptr;
|
|
+ int i;
|
|
+
|
|
+ for (i = 0; i < count; i++)
|
|
+ desc[i] = cpu_to_le64(desc[i]);
|
|
+}
|
|
+
|
|
+static inline void xgene_enet_le64_to_cpu(void *desc_ptr, int count)
|
|
+{
|
|
+ u64 *desc = desc_ptr;
|
|
+ int i;
|
|
+
|
|
+ for (i = 0; i < count; i++)
|
|
+ desc[i] = le64_to_cpu(desc[i]);
|
|
+}
|
|
+
|
|
+static inline void xgene_enet_desc16_to_le64(void *desc_ptr)
|
|
+{
|
|
+ u64 *desc;
|
|
+
|
|
+ desc = desc_ptr;
|
|
+ desc[1] = cpu_to_le64(desc[1]);
|
|
+}
|
|
+
|
|
+static inline void xgene_enet_le64_to_desc16(void *desc_ptr)
|
|
+{
|
|
+ u64 *desc;
|
|
+
|
|
+ desc = desc_ptr;
|
|
+ desc[1] = le64_to_cpu(desc[1]);
|
|
+}
|
|
+
|
|
+enum xgene_enet_ring_cfgsize {
|
|
+ RING_CFGSIZE_512B,
|
|
+ RING_CFGSIZE_2KB,
|
|
+ RING_CFGSIZE_16KB,
|
|
+ RING_CFGSIZE_64KB,
|
|
+ RING_CFGSIZE_512KB,
|
|
+ RING_CFGSIZE_INVALID
|
|
+};
|
|
+
|
|
+enum xgene_enet_ring_type {
|
|
+ RING_DISABLED,
|
|
+ RING_REGULAR,
|
|
+ RING_BUFPOOL
|
|
+};
|
|
+
|
|
+enum xgene_ring_owner {
|
|
+ RING_OWNER_ETH0,
|
|
+ RING_OWNER_CPU = 15,
|
|
+ RING_OWNER_INVALID
|
|
+};
|
|
+
|
|
+enum xgene_enet_ring_bufnum {
|
|
+ RING_BUFNUM_REGULAR = 0x0,
|
|
+ RING_BUFNUM_BUFPOOL = 0x20,
|
|
+ RING_BUFNUM_INVALID
|
|
+};
|
|
+
|
|
+enum xgene_enet_cmd {
|
|
+ XGENE_ENET_WR_CMD = BIT(31),
|
|
+ XGENE_ENET_RD_CMD = BIT(30)
|
|
+};
|
|
+
|
|
+enum xgene_enet_err_code {
|
|
+ HBF_READ_DATA = 3,
|
|
+ HBF_LL_READ = 4,
|
|
+ BAD_WORK_MSG = 6,
|
|
+ BUFPOOL_TIMEOUT = 15,
|
|
+ INGRESS_CRC = 16,
|
|
+ INGRESS_CHECKSUM = 17,
|
|
+ INGRESS_TRUNC_FRAME = 18,
|
|
+ INGRESS_PKT_LEN = 19,
|
|
+ INGRESS_PKT_UNDER = 20,
|
|
+ INGRESS_FIFO_OVERRUN = 21,
|
|
+ INGRESS_CHECKSUM_COMPUTE = 26,
|
|
+ ERR_CODE_INVALID
|
|
+};
|
|
+
|
|
+static inline enum xgene_ring_owner xgene_enet_ring_owner(u16 id)
|
|
+{
|
|
+ return (id & RING_OWNER_MASK) >> 6;
|
|
+}
|
|
+
|
|
+static inline u8 xgene_enet_ring_bufnum(u16 id)
|
|
+{
|
|
+ return id & RING_BUFNUM_MASK;
|
|
+}
|
|
+
|
|
+static inline bool xgene_enet_is_bufpool(u16 id)
|
|
+{
|
|
+ return ((id & RING_BUFNUM_MASK) >= 0x20) ? true : false;
|
|
+}
|
|
+
|
|
+static inline u16 xgene_enet_get_numslots(u16 id, u32 size)
|
|
+{
|
|
+ bool is_bufpool = xgene_enet_is_bufpool(id);
|
|
+
|
|
+ return (is_bufpool) ? size / BUFPOOL_DESC_SIZE :
|
|
+ size / WORK_DESC_SIZE;
|
|
+}
|
|
+
|
|
+struct xgene_enet_desc_ring *xgene_enet_setup_ring(
|
|
+ struct xgene_enet_desc_ring *ring);
|
|
+void xgene_enet_clear_ring(struct xgene_enet_desc_ring *ring);
|
|
+
|
|
+void xgene_set_tx_desc(struct xgene_enet_desc_ring *ring,
|
|
+ struct xgene_enet_raw_desc *raw_desc);
|
|
+void xgene_get_desc(struct xgene_enet_desc_ring *ring,
|
|
+ struct xgene_enet_raw_desc *raw_desc);
|
|
+void xgene_get_bufpool_desc(struct xgene_enet_desc_ring *ring,
|
|
+ struct xgene_enet_raw_desc16 *raw_desc);
|
|
+void xgene_enet_parse_error(struct xgene_enet_desc_ring *ring,
|
|
+ struct xgene_enet_pdata *pdata,
|
|
+ enum xgene_enet_err_code status);
|
|
+
|
|
+void xgene_enet_reset(struct xgene_enet_pdata *priv);
|
|
+void xgene_gmac_reset(struct xgene_enet_pdata *priv);
|
|
+void xgene_gmac_init(struct xgene_enet_pdata *priv, int speed);
|
|
+void xgene_gmac_tx_enable(struct xgene_enet_pdata *priv);
|
|
+void xgene_gmac_rx_enable(struct xgene_enet_pdata *priv);
|
|
+void xgene_gmac_tx_disable(struct xgene_enet_pdata *priv);
|
|
+void xgene_gmac_rx_disable(struct xgene_enet_pdata *priv);
|
|
+void xgene_gmac_set_mac_addr(struct xgene_enet_pdata *pdata);
|
|
+void xgene_enet_cle_bypass(struct xgene_enet_pdata *pdata,
|
|
+ u32 dst_ring_num, u16 bufpool_id);
|
|
+void xgene_gport_shutdown(struct xgene_enet_pdata *priv);
|
|
+void xgene_gmac_get_tx_stats(struct xgene_enet_pdata *pdata);
|
|
+
|
|
+int xgene_enet_mdio_config(struct xgene_enet_pdata *pdata);
|
|
+int xgene_enet_mdio_remove(struct xgene_enet_pdata *pdata);
|
|
+
|
|
+#endif /* __XGENE_ENET_HW_H__ */
|
|
diff --git a/drivers/net/ethernet/apm/xgene/xgene_enet_main.c b/drivers/net/ethernet/apm/xgene/xgene_enet_main.c
|
|
new file mode 100644
|
|
index 0000000..756523a
|
|
--- /dev/null
|
|
+++ b/drivers/net/ethernet/apm/xgene/xgene_enet_main.c
|
|
@@ -0,0 +1,962 @@
|
|
+/* Applied Micro X-Gene SoC Ethernet Driver
|
|
+ *
|
|
+ * Copyright (c) 2014, Applied Micro Circuits Corporation
|
|
+ * Authors: Iyappan Subramanian <isubramanian@apm.com>
|
|
+ * Ravi Patel <rapatel@apm.com>
|
|
+ * Keyur Chudgar <kchudgar@apm.com>
|
|
+ *
|
|
+ * This program is free software; you can redistribute it and/or modify it
|
|
+ * under the terms of the GNU General Public License as published by the
|
|
+ * Free Software Foundation; either version 2 of the License, or (at your
|
|
+ * option) any later version.
|
|
+ *
|
|
+ * This program is distributed in the hope that it will be useful,
|
|
+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
+ * GNU General Public License for more details.
|
|
+ *
|
|
+ * You should have received a copy of the GNU General Public License
|
|
+ * along with this program. If not, see <http://www.gnu.org/licenses/>.
|
|
+ */
|
|
+
|
|
+#include "xgene_enet_main.h"
|
|
+#include "xgene_enet_hw.h"
|
|
+
|
|
+static void xgene_enet_init_bufpool(struct xgene_enet_desc_ring *buf_pool)
|
|
+{
|
|
+ struct xgene_enet_raw_desc16 *raw_desc;
|
|
+ int i;
|
|
+
|
|
+ for (i = 0; i < buf_pool->slots; i++) {
|
|
+ raw_desc = &buf_pool->raw_desc16[i];
|
|
+
|
|
+ /* Hardware expects descriptor in little endian format */
|
|
+ raw_desc->m0 = cpu_to_le64(i |
|
|
+ (((u64)buf_pool->dst_ring_num << FPQNUM_POS) &
|
|
+ FPQNUM_MASK) | STASHING_MASK);
|
|
+ }
|
|
+}
|
|
+
|
|
+static struct device *ndev_to_dev(struct net_device *ndev)
|
|
+{
|
|
+ return ndev->dev.parent;
|
|
+}
|
|
+
|
|
+static int xgene_enet_refill_bufpool(struct xgene_enet_desc_ring *buf_pool,
|
|
+ u32 nbuf)
|
|
+{
|
|
+ struct sk_buff *skb;
|
|
+ struct xgene_enet_raw_desc16 *raw_desc;
|
|
+ struct net_device *ndev;
|
|
+ struct device *dev;
|
|
+ dma_addr_t dma_addr;
|
|
+ u32 tail = buf_pool->tail;
|
|
+ u32 slots = buf_pool->slots - 1;
|
|
+ u16 bufdatalen, len;
|
|
+ int i;
|
|
+
|
|
+ ndev = buf_pool->ndev;
|
|
+ dev = ndev_to_dev(buf_pool->ndev);
|
|
+ bufdatalen = BUF_LEN_CODE_2K | (SKB_BUFFER_SIZE & GENMASK(11, 0));
|
|
+ len = XGENE_ENET_MAX_MTU;
|
|
+
|
|
+ for (i = 0; i < nbuf; i++) {
|
|
+ raw_desc = &buf_pool->raw_desc16[tail];
|
|
+
|
|
+ skb = netdev_alloc_skb_ip_align(ndev, len);
|
|
+ if (unlikely(!skb))
|
|
+ return -ENOMEM;
|
|
+ buf_pool->rx_skb[tail] = skb;
|
|
+
|
|
+ dma_addr = dma_map_single(dev, skb->data, len, DMA_FROM_DEVICE);
|
|
+ if (dma_mapping_error(dev, dma_addr)) {
|
|
+ netdev_err(ndev, "DMA mapping error\n");
|
|
+ dev_kfree_skb_any(skb);
|
|
+ return -EINVAL;
|
|
+ }
|
|
+
|
|
+ raw_desc->m1 = cpu_to_le64((dma_addr & DATAADDR_MASK) |
|
|
+ (((u64)bufdatalen << BUFDATALEN_POS) &
|
|
+ BUFDATALEN_MASK) | COHERENT_MASK);
|
|
+ tail = (tail + 1) & slots;
|
|
+ }
|
|
+
|
|
+ iowrite32(nbuf, buf_pool->cmd);
|
|
+ buf_pool->tail = tail;
|
|
+
|
|
+ return 0;
|
|
+}
|
|
+
|
|
+static u16 xgene_enet_dst_ring_num(struct xgene_enet_desc_ring *ring)
|
|
+{
|
|
+ struct xgene_enet_pdata *pdata = netdev_priv(ring->ndev);
|
|
+
|
|
+ return ((u16)pdata->rm << 10) | ring->num;
|
|
+}
|
|
+
|
|
+static u8 xgene_enet_hdr_len(const void *data)
|
|
+{
|
|
+ const struct ethhdr *eth = data;
|
|
+
|
|
+ return (eth->h_proto == htons(ETH_P_8021Q)) ? VLAN_ETH_HLEN : ETH_HLEN;
|
|
+}
|
|
+
|
|
+static u32 xgene_enet_ring_len(struct xgene_enet_desc_ring *ring)
|
|
+{
|
|
+ u32 *cmd_base = ring->cmd_base;
|
|
+ u32 ring_state, num_msgs;
|
|
+
|
|
+ ring_state = ioread32(&cmd_base[1]);
|
|
+ num_msgs = ring_state & CREATE_MASK(NUMMSGSINQ_POS, NUMMSGSINQ_LEN);
|
|
+
|
|
+ return num_msgs >> NUMMSGSINQ_POS;
|
|
+}
|
|
+
|
|
+static void xgene_enet_delete_bufpool(struct xgene_enet_desc_ring *buf_pool)
|
|
+{
|
|
+ struct xgene_enet_raw_desc16 *raw_desc;
|
|
+ u32 slots = buf_pool->slots - 1;
|
|
+ u32 tail = buf_pool->tail;
|
|
+ u32 userinfo;
|
|
+ int i, len;
|
|
+
|
|
+ len = xgene_enet_ring_len(buf_pool);
|
|
+ for (i = 0; i < len; i++) {
|
|
+ tail = (tail - 1) & slots;
|
|
+ raw_desc = &buf_pool->raw_desc16[tail];
|
|
+
|
|
+ /* Hardware stores descriptor in little endian format */
|
|
+ userinfo = le64_to_cpu(raw_desc->m0) & USERINFO_MASK;
|
|
+ dev_kfree_skb_any(buf_pool->rx_skb[userinfo]);
|
|
+ }
|
|
+
|
|
+ iowrite32(-len, buf_pool->cmd);
|
|
+ buf_pool->tail = tail;
|
|
+}
|
|
+
|
|
+static irqreturn_t xgene_enet_rx_irq(const int irq, void *data)
|
|
+{
|
|
+ struct xgene_enet_desc_ring *rx_ring = data;
|
|
+
|
|
+ if (napi_schedule_prep(&rx_ring->napi)) {
|
|
+ disable_irq_nosync(irq);
|
|
+ __napi_schedule(&rx_ring->napi);
|
|
+ }
|
|
+
|
|
+ return IRQ_HANDLED;
|
|
+}
|
|
+
|
|
+static int xgene_enet_tx_completion(struct xgene_enet_desc_ring *cp_ring,
|
|
+ struct xgene_enet_raw_desc *raw_desc)
|
|
+{
|
|
+ struct sk_buff *skb;
|
|
+ struct device *dev;
|
|
+ u16 skb_index;
|
|
+ u8 status;
|
|
+ int ret = 0;
|
|
+
|
|
+ skb_index = raw_desc->m0 & USERINFO_MASK;
|
|
+ skb = cp_ring->cp_skb[skb_index];
|
|
+
|
|
+ dev = ndev_to_dev(cp_ring->ndev);
|
|
+ dma_unmap_single(dev, raw_desc->m1 & DATAADDR_MASK,
|
|
+ (raw_desc->m1 & BUFDATALEN_MASK) >> BUFDATALEN_POS,
|
|
+ DMA_TO_DEVICE);
|
|
+
|
|
+ /* Checking for error */
|
|
+ status = (raw_desc->m0 & LERR_MASK) >> LERR_POS;
|
|
+ if (unlikely(status > 2)) {
|
|
+ xgene_enet_parse_error(cp_ring, netdev_priv(cp_ring->ndev),
|
|
+ status);
|
|
+ ret = -1;
|
|
+ }
|
|
+
|
|
+ if (likely(skb)) {
|
|
+ dev_kfree_skb_any(skb);
|
|
+ } else {
|
|
+ netdev_err(cp_ring->ndev, "completion skb is NULL\n");
|
|
+ ret = -1;
|
|
+ }
|
|
+
|
|
+ return ret;
|
|
+}
|
|
+
|
|
+static u64 xgene_enet_work_msg(struct sk_buff *skb)
|
|
+{
|
|
+ struct iphdr *iph;
|
|
+ u8 l3hlen, l4hlen = 0;
|
|
+ u8 csum_enable = 0;
|
|
+ u8 proto = 0;
|
|
+ u8 ethhdr;
|
|
+ u64 hopinfo;
|
|
+
|
|
+ if (unlikely(skb->protocol != htons(ETH_P_IP)) &&
|
|
+ unlikely(skb->protocol != htons(ETH_P_8021Q)))
|
|
+ goto out;
|
|
+
|
|
+ if (unlikely(!(skb->dev->features & NETIF_F_IP_CSUM)))
|
|
+ goto out;
|
|
+
|
|
+ iph = ip_hdr(skb);
|
|
+ if (unlikely(ip_is_fragment(iph)))
|
|
+ goto out;
|
|
+
|
|
+ if (likely(iph->protocol == IPPROTO_TCP)) {
|
|
+ l4hlen = tcp_hdrlen(skb) >> 2;
|
|
+ csum_enable = 1;
|
|
+ proto = TSO_IPPROTO_TCP;
|
|
+ } else if (iph->protocol == IPPROTO_UDP) {
|
|
+ l4hlen = UDP_HDR_SIZE;
|
|
+ csum_enable = 1;
|
|
+ }
|
|
+out:
|
|
+ l3hlen = ip_hdrlen(skb) >> 2;
|
|
+ ethhdr = xgene_enet_hdr_len(skb->data);
|
|
+ hopinfo = (l4hlen & TCPHDR_MASK) |
|
|
+ ((l3hlen << IPHDR_POS) & IPHDR_MASK) |
|
|
+ (ethhdr << ETHHDR_POS) |
|
|
+ (csum_enable << EC_POS) |
|
|
+ (proto << IS_POS) |
|
|
+ INSERT_CRC |
|
|
+ TYPE_ETH_WORK_MESSAGE;
|
|
+
|
|
+ return hopinfo;
|
|
+}
|
|
+
|
|
+static int xgene_enet_setup_tx_desc(struct xgene_enet_desc_ring *tx_ring,
|
|
+ struct sk_buff *skb)
|
|
+{
|
|
+ struct device *dev = ndev_to_dev(tx_ring->ndev);
|
|
+ struct xgene_enet_raw_desc *raw_desc;
|
|
+ dma_addr_t dma_addr;
|
|
+ u16 tail = tx_ring->tail;
|
|
+ u64 hopinfo;
|
|
+
|
|
+ raw_desc = &tx_ring->raw_desc[tail];
|
|
+ memset(raw_desc, 0, sizeof(struct xgene_enet_raw_desc));
|
|
+
|
|
+ dma_addr = dma_map_single(dev, skb->data, skb->len, DMA_TO_DEVICE);
|
|
+ if (dma_mapping_error(dev, dma_addr)) {
|
|
+ netdev_err(tx_ring->ndev, "DMA mapping error\n");
|
|
+ return -EINVAL;
|
|
+ }
|
|
+
|
|
+ /* Hardware expects descriptor in little endian format */
|
|
+ raw_desc->m0 = cpu_to_le64(tail);
|
|
+ raw_desc->m1 = cpu_to_le64((dma_addr & DATAADDR_MASK) |
|
|
+ (((u64)skb->len << BUFDATALEN_POS) & BUFDATALEN_MASK) |
|
|
+ COHERENT_MASK);
|
|
+ hopinfo = xgene_enet_work_msg(skb);
|
|
+ raw_desc->m3 = cpu_to_le64(
|
|
+ (((u64)tx_ring->dst_ring_num << HENQNUM_POS) &
|
|
+ HENQNUM_MASK) | hopinfo);
|
|
+ tx_ring->cp_ring->cp_skb[tail] = skb;
|
|
+
|
|
+ return 0;
|
|
+}
|
|
+
|
|
+static netdev_tx_t xgene_enet_start_xmit(struct sk_buff *skb,
|
|
+ struct net_device *ndev)
|
|
+{
|
|
+ struct xgene_enet_pdata *pdata = netdev_priv(ndev);
|
|
+ struct xgene_enet_desc_ring *tx_ring = pdata->tx_ring;
|
|
+ struct xgene_enet_desc_ring *cp_ring = tx_ring->cp_ring;
|
|
+ u32 tx_level, cq_level;
|
|
+
|
|
+ tx_level = xgene_enet_ring_len(tx_ring);
|
|
+ cq_level = xgene_enet_ring_len(cp_ring);
|
|
+ if (unlikely(tx_level > pdata->tx_qcnt_hi ||
|
|
+ cq_level > pdata->cp_qcnt_hi)) {
|
|
+ netif_stop_queue(ndev);
|
|
+ return NETDEV_TX_BUSY;
|
|
+ }
|
|
+
|
|
+ if (xgene_enet_setup_tx_desc(tx_ring, skb)) {
|
|
+ dev_kfree_skb_any(skb);
|
|
+ return NETDEV_TX_OK;
|
|
+ }
|
|
+
|
|
+ iowrite32(1, tx_ring->cmd);
|
|
+ skb_tx_timestamp(skb);
|
|
+ tx_ring->tail = (tx_ring->tail + 1) & (tx_ring->slots - 1);
|
|
+
|
|
+ pdata->stats.tx_packets++;
|
|
+ pdata->stats.tx_bytes += skb->len;
|
|
+
|
|
+ return NETDEV_TX_OK;
|
|
+}
|
|
+
|
|
+static void xgene_enet_skip_csum(struct sk_buff *skb)
|
|
+{
|
|
+ struct iphdr *iph = ip_hdr(skb);
|
|
+
|
|
+ if (!ip_is_fragment(iph) ||
|
|
+ (iph->protocol != IPPROTO_TCP && iph->protocol != IPPROTO_UDP)) {
|
|
+ skb->ip_summed = CHECKSUM_UNNECESSARY;
|
|
+ }
|
|
+}
|
|
+
|
|
+static int xgene_enet_rx_frame(struct xgene_enet_desc_ring *rx_ring,
|
|
+ struct xgene_enet_raw_desc *raw_desc)
|
|
+{
|
|
+ struct net_device *ndev;
|
|
+ struct xgene_enet_pdata *pdata;
|
|
+ struct device *dev;
|
|
+ struct xgene_enet_desc_ring *buf_pool;
|
|
+ u32 datalen, skb_index;
|
|
+ struct sk_buff *skb;
|
|
+ u8 status;
|
|
+ int ret = 0;
|
|
+
|
|
+ ndev = rx_ring->ndev;
|
|
+ pdata = netdev_priv(ndev);
|
|
+ dev = ndev_to_dev(rx_ring->ndev);
|
|
+ buf_pool = rx_ring->buf_pool;
|
|
+
|
|
+ dma_unmap_single(dev, raw_desc->m1 & DATAADDR_MASK, XGENE_ENET_MAX_MTU,
|
|
+ DMA_FROM_DEVICE);
|
|
+ skb_index = raw_desc->m0 & USERINFO_MASK;
|
|
+ skb = buf_pool->rx_skb[skb_index];
|
|
+
|
|
+ /* checking for error */
|
|
+ status = (raw_desc->m0 & LERR_MASK) >> LERR_POS;
|
|
+ if (unlikely(status > 2)) {
|
|
+ dev_kfree_skb_any(skb);
|
|
+ xgene_enet_parse_error(rx_ring, netdev_priv(rx_ring->ndev),
|
|
+ status);
|
|
+ pdata->stats.rx_dropped++;
|
|
+ ret = -1;
|
|
+ goto out;
|
|
+ }
|
|
+
|
|
+ /* strip off CRC as HW isn't doing this */
|
|
+ datalen = (raw_desc->m1 & BUFDATALEN_MASK) >> BUFDATALEN_POS;
|
|
+ datalen -= 4;
|
|
+ prefetch(skb->data - NET_IP_ALIGN);
|
|
+ skb_put(skb, datalen);
|
|
+
|
|
+ skb_checksum_none_assert(skb);
|
|
+ skb->protocol = eth_type_trans(skb, ndev);
|
|
+ if (likely((ndev->features & NETIF_F_IP_CSUM) &&
|
|
+ skb->protocol == htons(ETH_P_IP))) {
|
|
+ xgene_enet_skip_csum(skb);
|
|
+ }
|
|
+
|
|
+ pdata->stats.rx_packets++;
|
|
+ pdata->stats.rx_bytes += datalen;
|
|
+ napi_gro_receive(&rx_ring->napi, skb);
|
|
+out:
|
|
+ if (--rx_ring->nbufpool == 0) {
|
|
+ ret = xgene_enet_refill_bufpool(buf_pool, NUM_BUFPOOL);
|
|
+ rx_ring->nbufpool = NUM_BUFPOOL;
|
|
+ }
|
|
+
|
|
+ return ret;
|
|
+}
|
|
+
|
|
+static bool is_rx_desc(struct xgene_enet_raw_desc *raw_desc)
|
|
+{
|
|
+ /* Hardware stores descriptor in little endian format */
|
|
+ raw_desc->m0 = le64_to_cpu(raw_desc->m0);
|
|
+ raw_desc->m1 = le64_to_cpu(raw_desc->m1);
|
|
+ return ((raw_desc->m0 & FPQNUM_MASK) >> FPQNUM_POS) ? true : false;
|
|
+}
|
|
+
|
|
+static int xgene_enet_process_ring(struct xgene_enet_desc_ring *ring,
|
|
+ int budget)
|
|
+{
|
|
+ struct xgene_enet_pdata *pdata = netdev_priv(ring->ndev);
|
|
+ struct xgene_enet_raw_desc *raw_desc;
|
|
+ u16 head = ring->head;
|
|
+ u16 slots = ring->slots - 1;
|
|
+ int ret, count = 0;
|
|
+
|
|
+ do {
|
|
+ raw_desc = &ring->raw_desc[head];
|
|
+ if (unlikely(((u64 *)raw_desc)[EMPTY_SLOT_INDEX] == EMPTY_SLOT))
|
|
+ break;
|
|
+
|
|
+ if (is_rx_desc(raw_desc))
|
|
+ ret = xgene_enet_rx_frame(ring, raw_desc);
|
|
+ else
|
|
+ ret = xgene_enet_tx_completion(ring, raw_desc);
|
|
+ ((u64 *)raw_desc)[EMPTY_SLOT_INDEX] = EMPTY_SLOT;
|
|
+
|
|
+ head = (head + 1) & slots;
|
|
+ count++;
|
|
+
|
|
+ if (ret)
|
|
+ break;
|
|
+ } while (--budget);
|
|
+
|
|
+ if (likely(count)) {
|
|
+ iowrite32(-count, ring->cmd);
|
|
+ ring->head = head;
|
|
+
|
|
+ if (netif_queue_stopped(ring->ndev)) {
|
|
+ if (xgene_enet_ring_len(ring) < pdata->cp_qcnt_low)
|
|
+ netif_wake_queue(ring->ndev);
|
|
+ }
|
|
+ }
|
|
+
|
|
+ return budget;
|
|
+}
|
|
+
|
|
+static int xgene_enet_napi(struct napi_struct *napi, const int budget)
|
|
+{
|
|
+ struct xgene_enet_desc_ring *ring;
|
|
+ int processed;
|
|
+
|
|
+ ring = container_of(napi, struct xgene_enet_desc_ring, napi);
|
|
+ processed = xgene_enet_process_ring(ring, budget);
|
|
+
|
|
+ if (processed != budget) {
|
|
+ napi_complete(napi);
|
|
+ enable_irq(ring->irq);
|
|
+ }
|
|
+
|
|
+ return processed;
|
|
+}
|
|
+
|
|
+static void xgene_enet_timeout(struct net_device *ndev)
|
|
+{
|
|
+ struct xgene_enet_pdata *pdata = netdev_priv(ndev);
|
|
+
|
|
+ xgene_gmac_reset(pdata);
|
|
+}
|
|
+
|
|
+static int xgene_enet_register_irq(struct net_device *ndev)
|
|
+{
|
|
+ struct xgene_enet_pdata *pdata = netdev_priv(ndev);
|
|
+ struct device *dev = ndev_to_dev(ndev);
|
|
+ int ret;
|
|
+
|
|
+ ret = devm_request_irq(dev, pdata->rx_ring->irq, xgene_enet_rx_irq,
|
|
+ IRQF_SHARED, ndev->name, pdata->rx_ring);
|
|
+ if (ret) {
|
|
+ netdev_err(ndev, "rx%d interrupt request failed\n",
|
|
+ pdata->rx_ring->irq);
|
|
+ }
|
|
+
|
|
+ return ret;
|
|
+}
|
|
+
|
|
+static void xgene_enet_free_irq(struct net_device *ndev)
|
|
+{
|
|
+ struct xgene_enet_pdata *pdata;
|
|
+ struct device *dev;
|
|
+
|
|
+ pdata = netdev_priv(ndev);
|
|
+ dev = ndev_to_dev(ndev);
|
|
+ devm_free_irq(dev, pdata->rx_ring->irq, pdata->rx_ring);
|
|
+}
|
|
+
|
|
+static int xgene_enet_open(struct net_device *ndev)
|
|
+{
|
|
+ struct xgene_enet_pdata *pdata = netdev_priv(ndev);
|
|
+ int ret;
|
|
+
|
|
+ xgene_gmac_tx_enable(pdata);
|
|
+ xgene_gmac_rx_enable(pdata);
|
|
+
|
|
+ ret = xgene_enet_register_irq(ndev);
|
|
+ if (ret)
|
|
+ return ret;
|
|
+ napi_enable(&pdata->rx_ring->napi);
|
|
+
|
|
+ if (pdata->phy_dev)
|
|
+ phy_start(pdata->phy_dev);
|
|
+
|
|
+ netif_start_queue(ndev);
|
|
+
|
|
+ return ret;
|
|
+}
|
|
+
|
|
+static int xgene_enet_close(struct net_device *ndev)
|
|
+{
|
|
+ struct xgene_enet_pdata *pdata = netdev_priv(ndev);
|
|
+
|
|
+ netif_stop_queue(ndev);
|
|
+
|
|
+ if (pdata->phy_dev)
|
|
+ phy_stop(pdata->phy_dev);
|
|
+
|
|
+ napi_disable(&pdata->rx_ring->napi);
|
|
+ xgene_enet_free_irq(ndev);
|
|
+ xgene_enet_process_ring(pdata->rx_ring, -1);
|
|
+
|
|
+ xgene_gmac_tx_disable(pdata);
|
|
+ xgene_gmac_rx_disable(pdata);
|
|
+
|
|
+ return 0;
|
|
+}
|
|
+
|
|
+static void xgene_enet_delete_ring(struct xgene_enet_desc_ring *ring)
|
|
+{
|
|
+ struct xgene_enet_pdata *pdata;
|
|
+ struct device *dev;
|
|
+
|
|
+ pdata = netdev_priv(ring->ndev);
|
|
+ dev = ndev_to_dev(ring->ndev);
|
|
+
|
|
+ xgene_enet_clear_ring(ring);
|
|
+ dma_free_coherent(dev, ring->size, ring->desc_addr, ring->dma);
|
|
+}
|
|
+
|
|
+static void xgene_enet_delete_desc_rings(struct xgene_enet_pdata *pdata)
|
|
+{
|
|
+ struct xgene_enet_desc_ring *buf_pool;
|
|
+
|
|
+ if (pdata->tx_ring) {
|
|
+ xgene_enet_delete_ring(pdata->tx_ring);
|
|
+ pdata->tx_ring = NULL;
|
|
+ }
|
|
+
|
|
+ if (pdata->rx_ring) {
|
|
+ buf_pool = pdata->rx_ring->buf_pool;
|
|
+ xgene_enet_delete_bufpool(buf_pool);
|
|
+ xgene_enet_delete_ring(buf_pool);
|
|
+ xgene_enet_delete_ring(pdata->rx_ring);
|
|
+ pdata->rx_ring = NULL;
|
|
+ }
|
|
+}
|
|
+
|
|
+static int xgene_enet_get_ring_size(struct device *dev,
|
|
+ enum xgene_enet_ring_cfgsize cfgsize)
|
|
+{
|
|
+ int size = -EINVAL;
|
|
+
|
|
+ switch (cfgsize) {
|
|
+ case RING_CFGSIZE_512B:
|
|
+ size = 0x200;
|
|
+ break;
|
|
+ case RING_CFGSIZE_2KB:
|
|
+ size = 0x800;
|
|
+ break;
|
|
+ case RING_CFGSIZE_16KB:
|
|
+ size = 0x4000;
|
|
+ break;
|
|
+ case RING_CFGSIZE_64KB:
|
|
+ size = 0x10000;
|
|
+ break;
|
|
+ case RING_CFGSIZE_512KB:
|
|
+ size = 0x80000;
|
|
+ break;
|
|
+ default:
|
|
+ dev_err(dev, "Unsupported cfg ring size %d\n", cfgsize);
|
|
+ break;
|
|
+ }
|
|
+
|
|
+ return size;
|
|
+}
|
|
+
|
|
+static void xgene_enet_free_desc_ring(struct xgene_enet_desc_ring *ring)
|
|
+{
|
|
+ struct device *dev;
|
|
+
|
|
+ if (!ring)
|
|
+ return;
|
|
+
|
|
+ dev = ndev_to_dev(ring->ndev);
|
|
+
|
|
+ if (ring->desc_addr) {
|
|
+ xgene_enet_clear_ring(ring);
|
|
+ dma_free_coherent(dev, ring->size, ring->desc_addr, ring->dma);
|
|
+ }
|
|
+ devm_kfree(dev, ring);
|
|
+}
|
|
+
|
|
+static void xgene_enet_free_desc_rings(struct xgene_enet_pdata *pdata)
|
|
+{
|
|
+ struct device *dev = &pdata->pdev->dev;
|
|
+ struct xgene_enet_desc_ring *ring;
|
|
+
|
|
+ ring = pdata->tx_ring;
|
|
+ if (ring && ring->cp_ring && ring->cp_ring->cp_skb)
|
|
+ devm_kfree(dev, ring->cp_ring->cp_skb);
|
|
+ xgene_enet_free_desc_ring(ring);
|
|
+
|
|
+ ring = pdata->rx_ring;
|
|
+ if (ring && ring->buf_pool && ring->buf_pool->rx_skb)
|
|
+ devm_kfree(dev, ring->buf_pool->rx_skb);
|
|
+ xgene_enet_free_desc_ring(ring->buf_pool);
|
|
+ xgene_enet_free_desc_ring(ring);
|
|
+}
|
|
+
|
|
+static struct xgene_enet_desc_ring *xgene_enet_create_desc_ring(
|
|
+ struct net_device *ndev, u32 ring_num,
|
|
+ enum xgene_enet_ring_cfgsize cfgsize, u32 ring_id)
|
|
+{
|
|
+ struct xgene_enet_desc_ring *ring;
|
|
+ struct xgene_enet_pdata *pdata = netdev_priv(ndev);
|
|
+ struct device *dev = ndev_to_dev(ndev);
|
|
+ u32 size;
|
|
+
|
|
+ ring = devm_kzalloc(dev, sizeof(struct xgene_enet_desc_ring),
|
|
+ GFP_KERNEL);
|
|
+ if (!ring)
|
|
+ return NULL;
|
|
+
|
|
+ ring->ndev = ndev;
|
|
+ ring->num = ring_num;
|
|
+ ring->cfgsize = cfgsize;
|
|
+ ring->id = ring_id;
|
|
+
|
|
+ size = xgene_enet_get_ring_size(dev, cfgsize);
|
|
+ ring->desc_addr = dma_zalloc_coherent(dev, size, &ring->dma,
|
|
+ GFP_KERNEL);
|
|
+ if (!ring->desc_addr) {
|
|
+ devm_kfree(dev, ring);
|
|
+ return NULL;
|
|
+ }
|
|
+ ring->size = size;
|
|
+
|
|
+ ring->cmd_base = pdata->ring_cmd_addr + (ring->num << 6);
|
|
+ ring->cmd = ring->cmd_base + INC_DEC_CMD_ADDR;
|
|
+ pdata->rm = RM3;
|
|
+ ring = xgene_enet_setup_ring(ring);
|
|
+ netdev_dbg(ndev, "ring info: num=%d size=%d id=%d slots=%d\n",
|
|
+ ring->num, ring->size, ring->id, ring->slots);
|
|
+
|
|
+ return ring;
|
|
+}
|
|
+
|
|
+static u16 xgene_enet_get_ring_id(enum xgene_ring_owner owner, u8 bufnum)
|
|
+{
|
|
+ return (owner << 6) | (bufnum & GENMASK(5, 0));
|
|
+}
|
|
+
|
|
+static int xgene_enet_create_desc_rings(struct net_device *ndev)
|
|
+{
|
|
+ struct xgene_enet_pdata *pdata = netdev_priv(ndev);
|
|
+ struct device *dev = ndev_to_dev(ndev);
|
|
+ struct xgene_enet_desc_ring *rx_ring, *tx_ring, *cp_ring;
|
|
+ struct xgene_enet_desc_ring *buf_pool = NULL;
|
|
+ u8 cpu_bufnum = 0, eth_bufnum = 0;
|
|
+ u8 bp_bufnum = 0x20;
|
|
+ u16 ring_id, ring_num = 0;
|
|
+ int ret;
|
|
+
|
|
+ /* allocate rx descriptor ring */
|
|
+ ring_id = xgene_enet_get_ring_id(RING_OWNER_CPU, cpu_bufnum++);
|
|
+ rx_ring = xgene_enet_create_desc_ring(ndev, ring_num++,
|
|
+ RING_CFGSIZE_16KB, ring_id);
|
|
+ if (!rx_ring) {
|
|
+ ret = -ENOMEM;
|
|
+ goto err;
|
|
+ }
|
|
+
|
|
+ /* allocate buffer pool for receiving packets */
|
|
+ ring_id = xgene_enet_get_ring_id(RING_OWNER_ETH0, bp_bufnum++);
|
|
+ buf_pool = xgene_enet_create_desc_ring(ndev, ring_num++,
|
|
+ RING_CFGSIZE_2KB, ring_id);
|
|
+ if (!buf_pool) {
|
|
+ ret = -ENOMEM;
|
|
+ goto err;
|
|
+ }
|
|
+
|
|
+ rx_ring->nbufpool = NUM_BUFPOOL;
|
|
+ rx_ring->buf_pool = buf_pool;
|
|
+ rx_ring->irq = pdata->rx_irq;
|
|
+ buf_pool->rx_skb = devm_kcalloc(dev, buf_pool->slots,
|
|
+ sizeof(struct sk_buff *), GFP_KERNEL);
|
|
+ if (!buf_pool->rx_skb) {
|
|
+ ret = -ENOMEM;
|
|
+ goto err;
|
|
+ }
|
|
+
|
|
+ buf_pool->dst_ring_num = xgene_enet_dst_ring_num(buf_pool);
|
|
+ rx_ring->buf_pool = buf_pool;
|
|
+ pdata->rx_ring = rx_ring;
|
|
+
|
|
+ /* allocate tx descriptor ring */
|
|
+ ring_id = xgene_enet_get_ring_id(RING_OWNER_ETH0, eth_bufnum++);
|
|
+ tx_ring = xgene_enet_create_desc_ring(ndev, ring_num++,
|
|
+ RING_CFGSIZE_16KB, ring_id);
|
|
+ if (!tx_ring) {
|
|
+ ret = -ENOMEM;
|
|
+ goto err;
|
|
+ }
|
|
+ pdata->tx_ring = tx_ring;
|
|
+
|
|
+ cp_ring = pdata->rx_ring;
|
|
+ cp_ring->cp_skb = devm_kcalloc(dev, tx_ring->slots,
|
|
+ sizeof(struct sk_buff *), GFP_KERNEL);
|
|
+ if (!cp_ring->cp_skb) {
|
|
+ ret = -ENOMEM;
|
|
+ goto err;
|
|
+ }
|
|
+ pdata->tx_ring->cp_ring = cp_ring;
|
|
+ pdata->tx_ring->dst_ring_num = xgene_enet_dst_ring_num(cp_ring);
|
|
+
|
|
+ pdata->tx_qcnt_hi = pdata->tx_ring->slots / 2;
|
|
+ pdata->cp_qcnt_hi = pdata->rx_ring->slots / 2;
|
|
+ pdata->cp_qcnt_low = pdata->cp_qcnt_hi / 2;
|
|
+
|
|
+ return 0;
|
|
+
|
|
+err:
|
|
+ xgene_enet_free_desc_rings(pdata);
|
|
+ return ret;
|
|
+}
|
|
+
|
|
+static struct rtnl_link_stats64 *xgene_enet_get_stats64(
|
|
+ struct net_device *ndev,
|
|
+ struct rtnl_link_stats64 *storage)
|
|
+{
|
|
+ struct xgene_enet_pdata *pdata = netdev_priv(ndev);
|
|
+ struct rtnl_link_stats64 *stats = &pdata->stats;
|
|
+
|
|
+ spin_lock(&pdata->stats_lock);
|
|
+ stats->rx_errors += stats->rx_length_errors +
|
|
+ stats->rx_crc_errors +
|
|
+ stats->rx_frame_errors +
|
|
+ stats->rx_fifo_errors;
|
|
+ memcpy(storage, &pdata->stats, sizeof(struct rtnl_link_stats64));
|
|
+ spin_unlock(&pdata->stats_lock);
|
|
+
|
|
+ return storage;
|
|
+}
|
|
+
|
|
+static int xgene_enet_set_mac_address(struct net_device *ndev, void *addr)
|
|
+{
|
|
+ struct xgene_enet_pdata *pdata = netdev_priv(ndev);
|
|
+ int ret;
|
|
+
|
|
+ ret = eth_mac_addr(ndev, addr);
|
|
+ if (ret)
|
|
+ return ret;
|
|
+ xgene_gmac_set_mac_addr(pdata);
|
|
+
|
|
+ return ret;
|
|
+}
|
|
+
|
|
+static const struct net_device_ops xgene_ndev_ops = {
|
|
+ .ndo_open = xgene_enet_open,
|
|
+ .ndo_stop = xgene_enet_close,
|
|
+ .ndo_start_xmit = xgene_enet_start_xmit,
|
|
+ .ndo_tx_timeout = xgene_enet_timeout,
|
|
+ .ndo_get_stats64 = xgene_enet_get_stats64,
|
|
+ .ndo_change_mtu = eth_change_mtu,
|
|
+ .ndo_set_mac_address = xgene_enet_set_mac_address,
|
|
+};
|
|
+
|
|
+static int xgene_enet_get_resources(struct xgene_enet_pdata *pdata)
|
|
+{
|
|
+ struct platform_device *pdev;
|
|
+ struct net_device *ndev;
|
|
+ struct device *dev;
|
|
+ struct resource *res;
|
|
+ void *base_addr;
|
|
+ const char *mac;
|
|
+ int ret;
|
|
+
|
|
+ pdev = pdata->pdev;
|
|
+ dev = &pdev->dev;
|
|
+ ndev = pdata->ndev;
|
|
+
|
|
+ res = platform_get_resource_byname(pdev, IORESOURCE_MEM, "enet_csr");
|
|
+ if (!res) {
|
|
+ dev_err(dev, "Resource enet_csr not defined\n");
|
|
+ return -ENODEV;
|
|
+ }
|
|
+ pdata->base_addr = devm_ioremap_resource(dev, res);
|
|
+ if (IS_ERR(pdata->base_addr)) {
|
|
+ dev_err(dev, "Unable to retrieve ENET Port CSR region\n");
|
|
+ return PTR_ERR(pdata->base_addr);
|
|
+ }
|
|
+
|
|
+ res = platform_get_resource_byname(pdev, IORESOURCE_MEM, "ring_csr");
|
|
+ if (!res) {
|
|
+ dev_err(dev, "Resource ring_csr not defined\n");
|
|
+ return -ENODEV;
|
|
+ }
|
|
+ pdata->ring_csr_addr = devm_ioremap_resource(dev, res);
|
|
+ if (IS_ERR(pdata->ring_csr_addr)) {
|
|
+ dev_err(dev, "Unable to retrieve ENET Ring CSR region\n");
|
|
+ return PTR_ERR(pdata->ring_csr_addr);
|
|
+ }
|
|
+
|
|
+ res = platform_get_resource_byname(pdev, IORESOURCE_MEM, "ring_cmd");
|
|
+ if (!res) {
|
|
+ dev_err(dev, "Resource ring_cmd not defined\n");
|
|
+ return -ENODEV;
|
|
+ }
|
|
+ pdata->ring_cmd_addr = devm_ioremap_resource(dev, res);
|
|
+ if (IS_ERR(pdata->ring_cmd_addr)) {
|
|
+ dev_err(dev, "Unable to retrieve ENET Ring command region\n");
|
|
+ return PTR_ERR(pdata->ring_cmd_addr);
|
|
+ }
|
|
+
|
|
+ ret = platform_get_irq(pdev, 0);
|
|
+ if (ret <= 0) {
|
|
+ dev_err(dev, "Unable to get ENET Rx IRQ\n");
|
|
+ ret = ret ? : -ENXIO;
|
|
+ return ret;
|
|
+ }
|
|
+ pdata->rx_irq = ret;
|
|
+
|
|
+ mac = of_get_mac_address(dev->of_node);
|
|
+ if (mac)
|
|
+ memcpy(ndev->dev_addr, mac, ndev->addr_len);
|
|
+ else
|
|
+ eth_hw_addr_random(ndev);
|
|
+ memcpy(ndev->perm_addr, ndev->dev_addr, ndev->addr_len);
|
|
+
|
|
+ pdata->phy_mode = of_get_phy_mode(pdev->dev.of_node);
|
|
+ if (pdata->phy_mode < 0) {
|
|
+ dev_err(dev, "Incorrect phy-connection-type in DTS\n");
|
|
+ return -EINVAL;
|
|
+ }
|
|
+
|
|
+ pdata->clk = devm_clk_get(&pdev->dev, NULL);
|
|
+ ret = IS_ERR(pdata->clk);
|
|
+ if (IS_ERR(pdata->clk)) {
|
|
+ dev_err(&pdev->dev, "can't get clock\n");
|
|
+ ret = PTR_ERR(pdata->clk);
|
|
+ return ret;
|
|
+ }
|
|
+
|
|
+ base_addr = pdata->base_addr;
|
|
+ pdata->eth_csr_addr = base_addr + BLOCK_ETH_CSR_OFFSET;
|
|
+ pdata->eth_ring_if_addr = base_addr + BLOCK_ETH_RING_IF_OFFSET;
|
|
+ pdata->eth_diag_csr_addr = base_addr + BLOCK_ETH_DIAG_CSR_OFFSET;
|
|
+ pdata->mcx_mac_addr = base_addr + BLOCK_ETH_MAC_OFFSET;
|
|
+ pdata->mcx_stats_addr = base_addr + BLOCK_ETH_STATS_OFFSET;
|
|
+ pdata->mcx_mac_csr_addr = base_addr + BLOCK_ETH_MAC_CSR_OFFSET;
|
|
+ pdata->rx_buff_cnt = NUM_PKT_BUF;
|
|
+
|
|
+ return ret;
|
|
+}
|
|
+
|
|
+static int xgene_enet_init_hw(struct xgene_enet_pdata *pdata)
|
|
+{
|
|
+ struct net_device *ndev = pdata->ndev;
|
|
+ struct xgene_enet_desc_ring *buf_pool;
|
|
+ u16 dst_ring_num;
|
|
+ int ret;
|
|
+
|
|
+ xgene_gmac_tx_disable(pdata);
|
|
+ xgene_gmac_rx_disable(pdata);
|
|
+
|
|
+ ret = xgene_enet_create_desc_rings(ndev);
|
|
+ if (ret) {
|
|
+ netdev_err(ndev, "Error in ring configuration\n");
|
|
+ return ret;
|
|
+ }
|
|
+
|
|
+ /* setup buffer pool */
|
|
+ buf_pool = pdata->rx_ring->buf_pool;
|
|
+ xgene_enet_init_bufpool(buf_pool);
|
|
+ ret = xgene_enet_refill_bufpool(buf_pool, pdata->rx_buff_cnt);
|
|
+ if (ret)
|
|
+ return ret;
|
|
+
|
|
+ dst_ring_num = xgene_enet_dst_ring_num(pdata->rx_ring);
|
|
+ xgene_enet_cle_bypass(pdata, dst_ring_num, buf_pool->id);
|
|
+
|
|
+ return ret;
|
|
+}
|
|
+
|
|
+static int xgene_enet_probe(struct platform_device *pdev)
|
|
+{
|
|
+ struct net_device *ndev;
|
|
+ struct xgene_enet_pdata *pdata;
|
|
+ struct device *dev = &pdev->dev;
|
|
+ struct napi_struct *napi;
|
|
+ int ret;
|
|
+
|
|
+ ndev = alloc_etherdev(sizeof(struct xgene_enet_pdata));
|
|
+ if (!ndev)
|
|
+ return -ENOMEM;
|
|
+
|
|
+ pdata = netdev_priv(ndev);
|
|
+
|
|
+ pdata->pdev = pdev;
|
|
+ pdata->ndev = ndev;
|
|
+ SET_NETDEV_DEV(ndev, dev);
|
|
+ platform_set_drvdata(pdev, pdata);
|
|
+ ndev->netdev_ops = &xgene_ndev_ops;
|
|
+ xgene_enet_set_ethtool_ops(ndev);
|
|
+ ndev->features |= NETIF_F_IP_CSUM |
|
|
+ NETIF_F_GSO |
|
|
+ NETIF_F_GRO;
|
|
+
|
|
+ ret = xgene_enet_get_resources(pdata);
|
|
+ if (ret)
|
|
+ goto err;
|
|
+
|
|
+ xgene_enet_reset(pdata);
|
|
+ xgene_gmac_init(pdata, SPEED_1000);
|
|
+
|
|
+ spin_lock_init(&pdata->stats_lock);
|
|
+ ret = register_netdev(ndev);
|
|
+ if (ret) {
|
|
+ netdev_err(ndev, "Failed to register netdev\n");
|
|
+ goto err;
|
|
+ }
|
|
+
|
|
+ ret = dma_set_mask_and_coherent(dev, DMA_BIT_MASK(64));
|
|
+ if (ret) {
|
|
+ netdev_err(ndev, "No usable DMA configuration\n");
|
|
+ goto err;
|
|
+ }
|
|
+
|
|
+ ret = xgene_enet_init_hw(pdata);
|
|
+ if (ret)
|
|
+ goto err;
|
|
+
|
|
+ napi = &pdata->rx_ring->napi;
|
|
+ netif_napi_add(ndev, napi, xgene_enet_napi, NAPI_POLL_WEIGHT);
|
|
+ ret = xgene_enet_mdio_config(pdata);
|
|
+
|
|
+ return ret;
|
|
+err:
|
|
+ free_netdev(ndev);
|
|
+ return ret;
|
|
+}
|
|
+
|
|
+static int xgene_enet_remove(struct platform_device *pdev)
|
|
+{
|
|
+ struct xgene_enet_pdata *pdata;
|
|
+ struct net_device *ndev;
|
|
+
|
|
+ pdata = platform_get_drvdata(pdev);
|
|
+ ndev = pdata->ndev;
|
|
+
|
|
+ xgene_gmac_rx_disable(pdata);
|
|
+ xgene_gmac_tx_disable(pdata);
|
|
+
|
|
+ netif_napi_del(&pdata->rx_ring->napi);
|
|
+ xgene_enet_mdio_remove(pdata);
|
|
+ xgene_enet_delete_desc_rings(pdata);
|
|
+ unregister_netdev(ndev);
|
|
+ xgene_gport_shutdown(pdata);
|
|
+ free_netdev(ndev);
|
|
+
|
|
+ return 0;
|
|
+}
|
|
+
|
|
+static struct of_device_id xgene_enet_match[] = {
|
|
+ {.compatible = "apm,xgene-enet",},
|
|
+ {},
|
|
+};
|
|
+
|
|
+MODULE_DEVICE_TABLE(of, xgene_enet_match);
|
|
+
|
|
+static struct platform_driver xgene_enet_driver = {
|
|
+ .driver = {
|
|
+ .name = "xgene-enet",
|
|
+ .owner = THIS_MODULE,
|
|
+ .of_match_table = xgene_enet_match,
|
|
+ },
|
|
+ .probe = xgene_enet_probe,
|
|
+ .remove = xgene_enet_remove,
|
|
+};
|
|
+
|
|
+module_platform_driver(xgene_enet_driver);
|
|
+
|
|
+MODULE_DESCRIPTION("APM X-Gene SoC Ethernet driver");
|
|
+MODULE_VERSION(XGENE_DRV_VERSION);
|
|
+MODULE_AUTHOR("Keyur Chudgar <kchudgar@apm.com>");
|
|
+MODULE_LICENSE("GPL");
|
|
diff --git a/drivers/net/ethernet/apm/xgene/xgene_enet_main.h b/drivers/net/ethernet/apm/xgene/xgene_enet_main.h
|
|
new file mode 100644
|
|
index 0000000..f4f7e4a
|
|
--- /dev/null
|
|
+++ b/drivers/net/ethernet/apm/xgene/xgene_enet_main.h
|
|
@@ -0,0 +1,107 @@
|
|
+/* Applied Micro X-Gene SoC Ethernet Driver
|
|
+ *
|
|
+ * Copyright (c) 2014, Applied Micro Circuits Corporation
|
|
+ * Authors: Iyappan Subramanian <isubramanian@apm.com>
|
|
+ * Ravi Patel <rapatel@apm.com>
|
|
+ * Keyur Chudgar <kchudgar@apm.com>
|
|
+ *
|
|
+ * This program is free software; you can redistribute it and/or modify it
|
|
+ * under the terms of the GNU General Public License as published by the
|
|
+ * Free Software Foundation; either version 2 of the License, or (at your
|
|
+ * option) any later version.
|
|
+ *
|
|
+ * This program is distributed in the hope that it will be useful,
|
|
+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
+ * GNU General Public License for more details.
|
|
+ *
|
|
+ * You should have received a copy of the GNU General Public License
|
|
+ * along with this program. If not, see <http://www.gnu.org/licenses/>.
|
|
+ */
|
|
+
|
|
+#ifndef __XGENE_ENET_MAIN_H__
|
|
+#define __XGENE_ENET_MAIN_H__
|
|
+
|
|
+#include <linux/clk.h>
|
|
+#include <linux/of_platform.h>
|
|
+#include <linux/of_net.h>
|
|
+#include <linux/of_mdio.h>
|
|
+#include <linux/module.h>
|
|
+#include <net/ip.h>
|
|
+#include <linux/prefetch.h>
|
|
+#include <linux/if_vlan.h>
|
|
+#include <linux/phy.h>
|
|
+#include "xgene_enet_hw.h"
|
|
+
|
|
+#define XGENE_DRV_VERSION "v1.0"
|
|
+#define XGENE_ENET_MAX_MTU 1536
|
|
+#define SKB_BUFFER_SIZE (XGENE_ENET_MAX_MTU - NET_IP_ALIGN)
|
|
+#define NUM_PKT_BUF 64
|
|
+#define NUM_BUFPOOL 32
|
|
+
|
|
+/* software context of a descriptor ring */
|
|
+struct xgene_enet_desc_ring {
|
|
+ struct net_device *ndev;
|
|
+ u16 id;
|
|
+ u16 num;
|
|
+ u16 head;
|
|
+ u16 tail;
|
|
+ u16 slots;
|
|
+ u16 irq;
|
|
+ u32 size;
|
|
+ u32 state[NUM_RING_CONFIG];
|
|
+ void __iomem *cmd_base;
|
|
+ void __iomem *cmd;
|
|
+ dma_addr_t dma;
|
|
+ u16 dst_ring_num;
|
|
+ u8 nbufpool;
|
|
+ struct sk_buff *(*rx_skb);
|
|
+ struct sk_buff *(*cp_skb);
|
|
+ enum xgene_enet_ring_cfgsize cfgsize;
|
|
+ struct xgene_enet_desc_ring *cp_ring;
|
|
+ struct xgene_enet_desc_ring *buf_pool;
|
|
+ struct napi_struct napi;
|
|
+ union {
|
|
+ void *desc_addr;
|
|
+ struct xgene_enet_raw_desc *raw_desc;
|
|
+ struct xgene_enet_raw_desc16 *raw_desc16;
|
|
+ };
|
|
+};
|
|
+
|
|
+/* ethernet private data */
|
|
+struct xgene_enet_pdata {
|
|
+ struct net_device *ndev;
|
|
+ struct mii_bus *mdio_bus;
|
|
+ struct phy_device *phy_dev;
|
|
+ int phy_speed;
|
|
+ struct clk *clk;
|
|
+ struct platform_device *pdev;
|
|
+ struct xgene_enet_desc_ring *tx_ring;
|
|
+ struct xgene_enet_desc_ring *rx_ring;
|
|
+ char *dev_name;
|
|
+ u32 rx_buff_cnt;
|
|
+ u32 tx_qcnt_hi;
|
|
+ u32 cp_qcnt_hi;
|
|
+ u32 cp_qcnt_low;
|
|
+ u32 rx_irq;
|
|
+ void __iomem *eth_csr_addr;
|
|
+ void __iomem *eth_ring_if_addr;
|
|
+ void __iomem *eth_diag_csr_addr;
|
|
+ void __iomem *mcx_mac_addr;
|
|
+ void __iomem *mcx_stats_addr;
|
|
+ void __iomem *mcx_mac_csr_addr;
|
|
+ void __iomem *base_addr;
|
|
+ void __iomem *ring_csr_addr;
|
|
+ void __iomem *ring_cmd_addr;
|
|
+ u32 phy_addr;
|
|
+ int phy_mode;
|
|
+ u32 speed;
|
|
+ u16 rm;
|
|
+ struct rtnl_link_stats64 stats;
|
|
+ /* statistics lock */
|
|
+ spinlock_t stats_lock;
|
|
+};
|
|
+
|
|
+void xgene_enet_set_ethtool_ops(struct net_device *netdev);
|
|
+
|
|
+#endif /* __XGENE_ENET_MAIN_H__ */
|
|
diff --git a/drivers/of/address.c b/drivers/of/address.c
|
|
index 5edfcb0..cbbaed2 100644
|
|
--- a/drivers/of/address.c
|
|
+++ b/drivers/of/address.c
|
|
@@ -5,6 +5,7 @@
|
|
#include <linux/module.h>
|
|
#include <linux/of_address.h>
|
|
#include <linux/pci_regs.h>
|
|
+#include <linux/slab.h>
|
|
#include <linux/string.h>
|
|
|
|
/* Max address size we deal with */
|
|
@@ -601,12 +602,72 @@ const __be32 *of_get_address(struct device_node *dev, int index, u64 *size,
|
|
}
|
|
EXPORT_SYMBOL(of_get_address);
|
|
|
|
+struct io_range {
|
|
+ struct list_head list;
|
|
+ phys_addr_t start;
|
|
+ resource_size_t size;
|
|
+};
|
|
+
|
|
+static LIST_HEAD(io_range_list);
|
|
+
|
|
+/*
|
|
+ * Record the PCI IO range (expressed as CPU physical address + size).
|
|
+ * Return a negative value if an error has occured, zero otherwise
|
|
+ */
|
|
+int __weak pci_register_io_range(phys_addr_t addr, resource_size_t size)
|
|
+{
|
|
+#ifdef PCI_IOBASE
|
|
+ struct io_range *res;
|
|
+ resource_size_t allocated_size = 0;
|
|
+
|
|
+ /* check if the range hasn't been previously recorded */
|
|
+ list_for_each_entry(res, &io_range_list, list) {
|
|
+ if (addr >= res->start && addr + size <= res->start + size)
|
|
+ return 0;
|
|
+ allocated_size += res->size;
|
|
+ }
|
|
+
|
|
+ /* range not registed yet, check for available space */
|
|
+ if (allocated_size + size - 1 > IO_SPACE_LIMIT)
|
|
+ return -E2BIG;
|
|
+
|
|
+ /* add the range to the list */
|
|
+ res = kzalloc(sizeof(*res), GFP_KERNEL);
|
|
+ if (!res)
|
|
+ return -ENOMEM;
|
|
+
|
|
+ res->start = addr;
|
|
+ res->size = size;
|
|
+
|
|
+ list_add_tail(&res->list, &io_range_list);
|
|
+
|
|
+ return 0;
|
|
+#else
|
|
+ return -EINVAL;
|
|
+#endif
|
|
+}
|
|
+
|
|
unsigned long __weak pci_address_to_pio(phys_addr_t address)
|
|
{
|
|
+#ifdef PCI_IOBASE
|
|
+ struct io_range *res;
|
|
+ resource_size_t offset = 0;
|
|
+
|
|
+ list_for_each_entry(res, &io_range_list, list) {
|
|
+ if (address >= res->start &&
|
|
+ address < res->start + res->size) {
|
|
+ return res->start - address + offset;
|
|
+ }
|
|
+ offset += res->size;
|
|
+ }
|
|
+
|
|
+ return (unsigned long)-1;
|
|
+#else
|
|
if (address > IO_SPACE_LIMIT)
|
|
return (unsigned long)-1;
|
|
|
|
return (unsigned long) address;
|
|
+#endif
|
|
}
|
|
|
|
static int __of_address_to_resource(struct device_node *dev,
|
|
@@ -811,3 +872,50 @@ bool of_dma_is_coherent(struct device_node *np)
|
|
return false;
|
|
}
|
|
EXPORT_SYMBOL_GPL(of_dma_is_coherent);
|
|
+
|
|
+/*
|
|
+ * of_pci_range_to_resource - Create a resource from an of_pci_range
|
|
+ * @range: the PCI range that describes the resource
|
|
+ * @np: device node where the range belongs to
|
|
+ * @res: pointer to a valid resource that will be updated to
|
|
+ * reflect the values contained in the range.
|
|
+ *
|
|
+ * Returns EINVAL if the range cannot be converted to resource.
|
|
+ *
|
|
+ * Note that if the range is an IO range, the resource will be converted
|
|
+ * using pci_address_to_pio() which can fail if it is called too early or
|
|
+ * if the range cannot be matched to any host bridge IO space (our case here).
|
|
+ * To guard against that we try to register the IO range first.
|
|
+ * If that fails we know that pci_address_to_pio() will do too.
|
|
+ */
|
|
+int of_pci_range_to_resource(struct of_pci_range *range,
|
|
+ struct device_node *np, struct resource *res)
|
|
+{
|
|
+ int err;
|
|
+ res->flags = range->flags;
|
|
+ res->parent = res->child = res->sibling = NULL;
|
|
+ res->name = np->full_name;
|
|
+
|
|
+ if (res->flags & IORESOURCE_IO) {
|
|
+ unsigned long port = -1;
|
|
+ err = pci_register_io_range(range->cpu_addr, range->size);
|
|
+ if (err)
|
|
+ goto invalid_range;
|
|
+ port = pci_address_to_pio(range->cpu_addr);
|
|
+ if (port == (unsigned long)-1) {
|
|
+ err = -EINVAL;
|
|
+ goto invalid_range;
|
|
+ }
|
|
+ res->start = port;
|
|
+ } else {
|
|
+ res->start = range->cpu_addr;
|
|
+ }
|
|
+ res->end = res->start + range->size - 1;
|
|
+ return 0;
|
|
+
|
|
+invalid_range:
|
|
+ res->start = (resource_size_t)OF_BAD_ADDR;
|
|
+ res->end = (resource_size_t)OF_BAD_ADDR;
|
|
+ return err;
|
|
+}
|
|
+
|
|
diff --git a/drivers/of/of_pci.c b/drivers/of/of_pci.c
|
|
index 8481996..e81402a 100644
|
|
--- a/drivers/of/of_pci.c
|
|
+++ b/drivers/of/of_pci.c
|
|
@@ -1,6 +1,7 @@
|
|
#include <linux/kernel.h>
|
|
#include <linux/export.h>
|
|
#include <linux/of.h>
|
|
+#include <linux/of_address.h>
|
|
#include <linux/of_pci.h>
|
|
|
|
static inline int __of_pci_pci_compare(struct device_node *node,
|
|
@@ -89,6 +90,141 @@ int of_pci_parse_bus_range(struct device_node *node, struct resource *res)
|
|
}
|
|
EXPORT_SYMBOL_GPL(of_pci_parse_bus_range);
|
|
|
|
+/**
|
|
+ * pci_host_bridge_of_get_ranges - Parse PCI host bridge resources from DT
|
|
+ * @dev: device node of the host bridge having the range property
|
|
+ * @resources: list where the range of resources will be added after DT parsing
|
|
+ * @io_base: pointer to a variable that will contain the physical address for
|
|
+ * the start of the I/O range.
|
|
+ *
|
|
+ * It is the callers job to free the @resources list if an error is returned.
|
|
+ *
|
|
+ * This function will parse the "ranges" property of a PCI host bridge device
|
|
+ * node and setup the resource mapping based on its content. It is expected
|
|
+ * that the property conforms with the Power ePAPR document.
|
|
+ *
|
|
+ * Each architecture is then offered the chance of applying their own
|
|
+ * filtering of pci_host_bridge_windows based on their own restrictions by
|
|
+ * calling pcibios_fixup_bridge_ranges(). The filtered list of windows
|
|
+ * can then be used when creating a pci_host_bridge structure.
|
|
+ */
|
|
+static int pci_host_bridge_of_get_ranges(struct device_node *dev,
|
|
+ struct list_head *resources, resource_size_t *io_base)
|
|
+{
|
|
+ struct resource *res;
|
|
+ struct of_pci_range range;
|
|
+ struct of_pci_range_parser parser;
|
|
+ int err;
|
|
+
|
|
+ pr_info("PCI host bridge %s ranges:\n", dev->full_name);
|
|
+
|
|
+ /* Check for ranges property */
|
|
+ err = of_pci_range_parser_init(&parser, dev);
|
|
+ if (err)
|
|
+ return err;
|
|
+
|
|
+ pr_debug("Parsing ranges property...\n");
|
|
+ for_each_of_pci_range(&parser, &range) {
|
|
+ /* Read next ranges element */
|
|
+ pr_debug("pci_space: 0x%08x pci_addr:0x%016llx cpu_addr:0x%016llx size:0x%016llx\n",
|
|
+ range.pci_space, range.pci_addr, range.cpu_addr, range.size);
|
|
+
|
|
+ /*
|
|
+ * If we failed translation or got a zero-sized region
|
|
+ * then skip this range
|
|
+ */
|
|
+ if (range.cpu_addr == OF_BAD_ADDR || range.size == 0)
|
|
+ continue;
|
|
+
|
|
+ res = kzalloc(sizeof(struct resource), GFP_KERNEL);
|
|
+ if (!res)
|
|
+ return -ENOMEM;
|
|
+
|
|
+ err = of_pci_range_to_resource(&range, dev, res);
|
|
+ if (err)
|
|
+ return err;
|
|
+
|
|
+ if (resource_type(res) == IORESOURCE_IO)
|
|
+ *io_base = range.cpu_addr;
|
|
+
|
|
+ pci_add_resource_offset(resources, res,
|
|
+ res->start - range.pci_addr);
|
|
+ }
|
|
+
|
|
+ /* Apply architecture specific fixups for the ranges */
|
|
+ return pcibios_fixup_bridge_ranges(resources);
|
|
+}
|
|
+
|
|
+static atomic_t domain_nr = ATOMIC_INIT(-1);
|
|
+
|
|
+/**
|
|
+ * of_create_pci_host_bridge - Create a PCI host bridge structure using
|
|
+ * information passed in the DT.
|
|
+ * @parent: device owning this host bridge
|
|
+ * @ops: pci_ops associated with the host controller
|
|
+ * @host_data: opaque data structure used by the host controller.
|
|
+ *
|
|
+ * returns a pointer to the newly created pci_host_bridge structure, or
|
|
+ * NULL if the call failed.
|
|
+ *
|
|
+ * This function will try to obtain the host bridge domain number by
|
|
+ * using of_alias_get_id() call with "pci-domain" as a stem. If that
|
|
+ * fails, a local allocator will be used that will put each host bridge
|
|
+ * in a new domain.
|
|
+ */
|
|
+struct pci_host_bridge *
|
|
+of_create_pci_host_bridge(struct device *parent, struct pci_ops *ops, void *host_data)
|
|
+{
|
|
+ int err, domain, busno;
|
|
+ struct resource *bus_range;
|
|
+ struct pci_bus *root_bus;
|
|
+ struct pci_host_bridge *bridge;
|
|
+ resource_size_t io_base = 0;
|
|
+ LIST_HEAD(res);
|
|
+
|
|
+ bus_range = kzalloc(sizeof(*bus_range), GFP_KERNEL);
|
|
+ if (!bus_range)
|
|
+ return ERR_PTR(-ENOMEM);
|
|
+
|
|
+ domain = of_alias_get_id(parent->of_node, "pci-domain");
|
|
+ if (domain == -ENODEV)
|
|
+ domain = atomic_inc_return(&domain_nr);
|
|
+
|
|
+ err = of_pci_parse_bus_range(parent->of_node, bus_range);
|
|
+ if (err) {
|
|
+ dev_info(parent, "No bus range for %s, using default [0-255]\n",
|
|
+ parent->of_node->full_name);
|
|
+ bus_range->start = 0;
|
|
+ bus_range->end = 255;
|
|
+ bus_range->flags = IORESOURCE_BUS;
|
|
+ }
|
|
+ busno = bus_range->start;
|
|
+ pci_add_resource(&res, bus_range);
|
|
+
|
|
+ /* now parse the rest of host bridge bus ranges */
|
|
+ err = pci_host_bridge_of_get_ranges(parent->of_node, &res, &io_base);
|
|
+ if (err)
|
|
+ goto err_create;
|
|
+
|
|
+ /* then create the root bus */
|
|
+ root_bus = pci_create_root_bus_in_domain(parent, domain, busno,
|
|
+ ops, host_data, &res);
|
|
+ if (IS_ERR(root_bus)) {
|
|
+ err = PTR_ERR(root_bus);
|
|
+ goto err_create;
|
|
+ }
|
|
+
|
|
+ bridge = to_pci_host_bridge(root_bus->bridge);
|
|
+ bridge->io_base = io_base;
|
|
+
|
|
+ return bridge;
|
|
+
|
|
+err_create:
|
|
+ pci_free_resource_list(&res);
|
|
+ return ERR_PTR(err);
|
|
+}
|
|
+EXPORT_SYMBOL_GPL(of_create_pci_host_bridge);
|
|
+
|
|
#ifdef CONFIG_PCI_MSI
|
|
|
|
static LIST_HEAD(of_pci_msi_chip_list);
|
|
diff --git a/drivers/pci/host-bridge.c b/drivers/pci/host-bridge.c
|
|
index 0e5f3c9..54ceafd 100644
|
|
--- a/drivers/pci/host-bridge.c
|
|
+++ b/drivers/pci/host-bridge.c
|
|
@@ -16,12 +16,13 @@ static struct pci_bus *find_pci_root_bus(struct pci_bus *bus)
|
|
return bus;
|
|
}
|
|
|
|
-static struct pci_host_bridge *find_pci_host_bridge(struct pci_bus *bus)
|
|
+struct pci_host_bridge *find_pci_host_bridge(struct pci_bus *bus)
|
|
{
|
|
struct pci_bus *root_bus = find_pci_root_bus(bus);
|
|
|
|
return to_pci_host_bridge(root_bus->bridge);
|
|
}
|
|
+EXPORT_SYMBOL_GPL(find_pci_host_bridge);
|
|
|
|
void pci_set_host_bridge_release(struct pci_host_bridge *bridge,
|
|
void (*release_fn)(struct pci_host_bridge *),
|
|
@@ -82,3 +83,18 @@ void pcibios_bus_to_resource(struct pci_bus *bus, struct resource *res,
|
|
res->end = region->end + offset;
|
|
}
|
|
EXPORT_SYMBOL(pcibios_bus_to_resource);
|
|
+
|
|
+/**
|
|
+ * Simple version of the platform specific code for filtering the list
|
|
+ * of resources obtained from the ranges declaration in DT.
|
|
+ *
|
|
+ * Platforms can override this function in order to impose stronger
|
|
+ * constraints onto the list of resources that a host bridge can use.
|
|
+ * The filtered list will then be used to create a root bus and associate
|
|
+ * it with the host bridge.
|
|
+ *
|
|
+ */
|
|
+int __weak pcibios_fixup_bridge_ranges(struct list_head *resources)
|
|
+{
|
|
+ return 0;
|
|
+}
|
|
diff --git a/drivers/pci/host/Kconfig b/drivers/pci/host/Kconfig
|
|
index 21df477..3b988a2 100644
|
|
--- a/drivers/pci/host/Kconfig
|
|
+++ b/drivers/pci/host/Kconfig
|
|
@@ -46,4 +46,14 @@ config PCI_HOST_GENERIC
|
|
Say Y here if you want to support a simple generic PCI host
|
|
controller, such as the one emulated by kvmtool.
|
|
|
|
+config PCI_XGENE
|
|
+ bool "X-Gene PCIe controller"
|
|
+ depends on ARCH_XGENE
|
|
+ depends on OF
|
|
+ select PCIEPORTBUS
|
|
+ help
|
|
+ Say Y here if you want internal PCI support on APM X-Gene SoC.
|
|
+ There are 5 internal PCIe ports available. Each port is GEN3 capable
|
|
+ and have varied lanes from x1 to x8.
|
|
+
|
|
endmenu
|
|
diff --git a/drivers/pci/host/Makefile b/drivers/pci/host/Makefile
|
|
index 611ba4b..0801606 100644
|
|
--- a/drivers/pci/host/Makefile
|
|
+++ b/drivers/pci/host/Makefile
|
|
@@ -6,3 +6,4 @@ obj-$(CONFIG_PCI_TEGRA) += pci-tegra.o
|
|
obj-$(CONFIG_PCI_RCAR_GEN2) += pci-rcar-gen2.o
|
|
obj-$(CONFIG_PCI_RCAR_GEN2_PCIE) += pcie-rcar.o
|
|
obj-$(CONFIG_PCI_HOST_GENERIC) += pci-host-generic.o
|
|
+obj-$(CONFIG_PCI_XGENE) += pci-xgene.o
|
|
diff --git a/drivers/pci/host/pci-xgene.c b/drivers/pci/host/pci-xgene.c
|
|
new file mode 100644
|
|
index 0000000..7bf4ac7
|
|
--- /dev/null
|
|
+++ b/drivers/pci/host/pci-xgene.c
|
|
@@ -0,0 +1,725 @@
|
|
+/**
|
|
+ * APM X-Gene PCIe Driver
|
|
+ *
|
|
+ * Copyright (c) 2013 Applied Micro Circuits Corporation.
|
|
+ *
|
|
+ * Author: Tanmay Inamdar <tinamdar@apm.com>.
|
|
+ *
|
|
+ * This program is free software; you can redistribute it and/or modify it
|
|
+ * under the terms of the GNU General Public License as published by the
|
|
+ * Free Software Foundation; either version 2 of the License, or (at your
|
|
+ * option) any later version.
|
|
+ *
|
|
+ * This program is distributed in the hope that it will be useful,
|
|
+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
+ * GNU General Public License for more details.
|
|
+ *
|
|
+ */
|
|
+#include <linux/clk-private.h>
|
|
+#include <linux/delay.h>
|
|
+#include <linux/io.h>
|
|
+#include <linux/jiffies.h>
|
|
+#include <linux/memblock.h>
|
|
+#include <linux/module.h>
|
|
+#include <linux/of.h>
|
|
+#include <linux/of_address.h>
|
|
+#include <linux/of_irq.h>
|
|
+#include <linux/of_pci.h>
|
|
+#include <linux/pci.h>
|
|
+#include <linux/platform_device.h>
|
|
+#include <linux/slab.h>
|
|
+
|
|
+#define PCIECORE_LTSSM 0x4c
|
|
+#define PCIECORE_CTLANDSTATUS 0x50
|
|
+#define INTXSTATUSMASK 0x6c
|
|
+#define PIM1_1L 0x80
|
|
+#define IBAR2 0x98
|
|
+#define IR2MSK 0x9c
|
|
+#define PIM2_1L 0xa0
|
|
+#define IBAR3L 0xb4
|
|
+#define IR3MSKL 0xbc
|
|
+#define PIM3_1L 0xc4
|
|
+#define OMR1BARL 0x100
|
|
+#define OMR2BARL 0x118
|
|
+#define OMR3BARL 0x130
|
|
+#define CFGBARL 0x154
|
|
+#define CFGBARH 0x158
|
|
+#define CFGCTL 0x15c
|
|
+#define RTDID 0x160
|
|
+#define BRIDGE_CFG_0 0x2000
|
|
+#define BRIDGE_CFG_1 0x2004
|
|
+#define BRIDGE_CFG_4 0x2010
|
|
+#define BRIDGE_CFG_32 0x2030
|
|
+#define BRIDGE_CFG_14 0x2038
|
|
+#define BRIDGE_CTRL_1 0x2204
|
|
+#define BRIDGE_CTRL_2 0x2208
|
|
+#define BRIDGE_CTRL_5 0x2214
|
|
+#define BRIDGE_STATUS_0 0x2600
|
|
+#define MEM_RAM_SHUTDOWN 0xd070
|
|
+#define BLOCK_MEM_RDY 0xd074
|
|
+
|
|
+#define DEVICE_PORT_TYPE_MASK 0x03c00000
|
|
+#define PM_FORCE_RP_MODE_MASK 0x00000400
|
|
+#define SWITCH_PORT_MODE_MASK 0x00000800
|
|
+#define CLASS_CODE_MASK 0xffffff00
|
|
+#define LINK_UP_MASK 0x00000100
|
|
+#define AER_OPTIONAL_ERROR_EN 0xffc00000
|
|
+#define XGENE_PCIE_DEV_CTRL 0x2f0f
|
|
+#define AXI_EP_CFG_ACCESS 0x10000
|
|
+#define ENABLE_ASPM 0x08000000
|
|
+#define XGENE_PORT_TYPE_RC 0x05000000
|
|
+#define BLOCK_MEM_RDY_VAL 0xFFFFFFFF
|
|
+#define EN_COHERENCY 0xF0000000
|
|
+#define EN_REG 0x00000001
|
|
+#define OB_LO_IO 0x00000002
|
|
+#define XGENE_PCIE_VENDORID 0xE008
|
|
+#define XGENE_PCIE_DEVICEID 0xE004
|
|
+#define XGENE_PCIE_ECC_TIMEOUT 10 /* ms */
|
|
+#define XGENE_LTSSM_DETECT_WAIT 20 /* ms */
|
|
+#define XGENE_LTSSM_L0_WAIT 4 /* ms */
|
|
+#define SZ_1T (SZ_1G*1024ULL)
|
|
+#define PIPE_PHY_RATE_RD(src) ((0xc000 & (u32)(src)) >> 0xe)
|
|
+
|
|
+struct xgene_pcie_port {
|
|
+ struct device_node *node;
|
|
+ struct device *dev;
|
|
+ struct clk *clk;
|
|
+ void __iomem *csr_base;
|
|
+ void __iomem *cfg_base;
|
|
+ u8 link_up;
|
|
+};
|
|
+
|
|
+static inline u32 pcie_bar_low_val(u32 addr, u32 flags)
|
|
+{
|
|
+ return (addr & PCI_BASE_ADDRESS_MEM_MASK) | flags;
|
|
+}
|
|
+
|
|
+/* PCIE Configuration Out/In */
|
|
+static inline void xgene_pcie_cfg_out32(void __iomem *addr, int offset, u32 val)
|
|
+{
|
|
+ writel(val, addr + offset);
|
|
+}
|
|
+
|
|
+static inline void xgene_pcie_cfg_out16(void __iomem *addr, int offset, u16 val)
|
|
+{
|
|
+ u32 val32 = readl(addr + (offset & ~0x3));
|
|
+
|
|
+ switch (offset & 0x3) {
|
|
+ case 2:
|
|
+ val32 &= ~0xFFFF0000;
|
|
+ val32 |= (u32)val << 16;
|
|
+ break;
|
|
+ case 0:
|
|
+ default:
|
|
+ val32 &= ~0xFFFF;
|
|
+ val32 |= val;
|
|
+ break;
|
|
+ }
|
|
+ writel(val32, addr + (offset & ~0x3));
|
|
+}
|
|
+
|
|
+static inline void xgene_pcie_cfg_out8(void __iomem *addr, int offset, u8 val)
|
|
+{
|
|
+ u32 val32 = readl(addr + (offset & ~0x3));
|
|
+
|
|
+ switch (offset & 0x3) {
|
|
+ case 0:
|
|
+ val32 &= ~0xFF;
|
|
+ val32 |= val;
|
|
+ break;
|
|
+ case 1:
|
|
+ val32 &= ~0xFF00;
|
|
+ val32 |= (u32)val << 8;
|
|
+ break;
|
|
+ case 2:
|
|
+ val32 &= ~0xFF0000;
|
|
+ val32 |= (u32)val << 16;
|
|
+ break;
|
|
+ case 3:
|
|
+ default:
|
|
+ val32 &= ~0xFF000000;
|
|
+ val32 |= (u32)val << 24;
|
|
+ break;
|
|
+ }
|
|
+ writel(val32, addr + (offset & ~0x3));
|
|
+}
|
|
+
|
|
+static inline void xgene_pcie_cfg_in32(void __iomem *addr, int offset, u32 *val)
|
|
+{
|
|
+ *val = readl(addr + offset);
|
|
+}
|
|
+
|
|
+static inline void
|
|
+xgene_pcie_cfg_in16(void __iomem *addr, int offset, u32 *val)
|
|
+{
|
|
+ *val = readl(addr + (offset & ~0x3));
|
|
+
|
|
+ switch (offset & 0x3) {
|
|
+ case 2:
|
|
+ *val >>= 16;
|
|
+ break;
|
|
+ }
|
|
+
|
|
+ *val &= 0xFFFF;
|
|
+}
|
|
+
|
|
+static inline void
|
|
+xgene_pcie_cfg_in8(void __iomem *addr, int offset, u32 *val)
|
|
+{
|
|
+ *val = readl(addr + (offset & ~0x3));
|
|
+
|
|
+ switch (offset & 0x3) {
|
|
+ case 3:
|
|
+ *val = *val >> 24;
|
|
+ break;
|
|
+ case 2:
|
|
+ *val = *val >> 16;
|
|
+ break;
|
|
+ case 1:
|
|
+ *val = *val >> 8;
|
|
+ break;
|
|
+ }
|
|
+ *val &= 0xFF;
|
|
+}
|
|
+
|
|
+/* When the address bit [17:16] is 2'b01, the Configuration access will be
|
|
+ * treated as Type 1 and it will be forwarded to external PCIe device.
|
|
+ */
|
|
+static void __iomem *xgene_pcie_get_cfg_base(struct pci_bus *bus)
|
|
+{
|
|
+ struct xgene_pcie_port *port = bus->sysdata;
|
|
+
|
|
+ if (bus->number >= (bus->primary + 1))
|
|
+ return port->cfg_base + AXI_EP_CFG_ACCESS;
|
|
+
|
|
+ return port->cfg_base;
|
|
+}
|
|
+
|
|
+/* For Configuration request, RTDID register is used as Bus Number,
|
|
+ * Device Number and Function number of the header fields.
|
|
+ */
|
|
+static void xgene_pcie_set_rtdid_reg(struct pci_bus *bus, uint devfn)
|
|
+{
|
|
+ struct xgene_pcie_port *port = bus->sysdata;
|
|
+ unsigned int b, d, f;
|
|
+ u32 rtdid_val = 0;
|
|
+
|
|
+ b = bus->number;
|
|
+ d = PCI_SLOT(devfn);
|
|
+ f = PCI_FUNC(devfn);
|
|
+
|
|
+ if (!pci_is_root_bus(bus))
|
|
+ rtdid_val = (b << 8) | (d << 3) | f;
|
|
+
|
|
+ writel(rtdid_val, port->csr_base + RTDID);
|
|
+ /* read the register back to ensure flush */
|
|
+ readl(port->csr_base + RTDID);
|
|
+}
|
|
+
|
|
+static int xgene_pcie_read_config(struct pci_bus *bus, unsigned int devfn,
|
|
+ int offset, int len, u32 *val)
|
|
+{
|
|
+ struct xgene_pcie_port *port = bus->sysdata;
|
|
+ void __iomem *addr;
|
|
+
|
|
+ if ((pci_is_root_bus(bus) && devfn != 0) || !port->link_up)
|
|
+ return PCIBIOS_DEVICE_NOT_FOUND;
|
|
+
|
|
+ xgene_pcie_set_rtdid_reg(bus, devfn);
|
|
+ addr = xgene_pcie_get_cfg_base(bus);
|
|
+ switch (len) {
|
|
+ case 1:
|
|
+ xgene_pcie_cfg_in8(addr, offset, val);
|
|
+ break;
|
|
+ case 2:
|
|
+ xgene_pcie_cfg_in16(addr, offset, val);
|
|
+ break;
|
|
+ default:
|
|
+ xgene_pcie_cfg_in32(addr, offset, val);
|
|
+ break;
|
|
+ }
|
|
+ return PCIBIOS_SUCCESSFUL;
|
|
+}
|
|
+
|
|
+static int xgene_pcie_write_config(struct pci_bus *bus, unsigned int devfn,
|
|
+ int offset, int len, u32 val)
|
|
+{
|
|
+ struct xgene_pcie_port *port = bus->sysdata;
|
|
+ void __iomem *addr;
|
|
+
|
|
+ if ((pci_is_root_bus(bus) && devfn != 0) || !port->link_up)
|
|
+ return PCIBIOS_DEVICE_NOT_FOUND;
|
|
+
|
|
+ xgene_pcie_set_rtdid_reg(bus, devfn);
|
|
+ addr = xgene_pcie_get_cfg_base(bus);
|
|
+ switch (len) {
|
|
+ case 1:
|
|
+ xgene_pcie_cfg_out8(addr, offset, (u8)val);
|
|
+ break;
|
|
+ case 2:
|
|
+ xgene_pcie_cfg_out16(addr, offset, (u16)val);
|
|
+ break;
|
|
+ default:
|
|
+ xgene_pcie_cfg_out32(addr, offset, val);
|
|
+ break;
|
|
+ }
|
|
+ return PCIBIOS_SUCCESSFUL;
|
|
+}
|
|
+
|
|
+static struct pci_ops xgene_pcie_ops = {
|
|
+ .read = xgene_pcie_read_config,
|
|
+ .write = xgene_pcie_write_config
|
|
+};
|
|
+
|
|
+static void xgene_pcie_program_core(void __iomem *csr_base)
|
|
+{
|
|
+ u32 val;
|
|
+
|
|
+ val = readl(csr_base + BRIDGE_CFG_0);
|
|
+ val |= AER_OPTIONAL_ERROR_EN;
|
|
+ writel(val, csr_base + BRIDGE_CFG_0);
|
|
+ writel(0x0, csr_base + INTXSTATUSMASK);
|
|
+ val = readl(csr_base + BRIDGE_CTRL_1);
|
|
+ val = (val & ~0xffff) | XGENE_PCIE_DEV_CTRL;
|
|
+ writel(val, csr_base + BRIDGE_CTRL_1);
|
|
+}
|
|
+
|
|
+static u64 xgene_pcie_set_ib_mask(void __iomem *csr_base, u32 addr,
|
|
+ u32 flags, u64 size)
|
|
+{
|
|
+ u64 mask = (~(size - 1) & PCI_BASE_ADDRESS_MEM_MASK) | flags;
|
|
+ u32 val32 = 0;
|
|
+ u32 val;
|
|
+
|
|
+ val32 = readl(csr_base + addr);
|
|
+ val = (val32 & 0x0000ffff) | (lower_32_bits(mask) << 16);
|
|
+ writel(val, csr_base + addr);
|
|
+
|
|
+ val32 = readl(csr_base + addr + 0x04);
|
|
+ val = (val32 & 0xffff0000) | (lower_32_bits(mask) >> 16);
|
|
+ writel(val, csr_base + addr + 0x04);
|
|
+
|
|
+ val32 = readl(csr_base + addr + 0x04);
|
|
+ val = (val32 & 0x0000ffff) | (upper_32_bits(mask) << 16);
|
|
+ writel(val, csr_base + addr + 0x04);
|
|
+
|
|
+ val32 = readl(csr_base + addr + 0x08);
|
|
+ val = (val32 & 0xffff0000) | (upper_32_bits(mask) >> 16);
|
|
+ writel(val, csr_base + addr + 0x08);
|
|
+
|
|
+ return mask;
|
|
+}
|
|
+
|
|
+static void xgene_pcie_poll_linkup(struct xgene_pcie_port *port,
|
|
+ u32 *lanes, u32 *speed)
|
|
+{
|
|
+ void __iomem *csr_base = port->csr_base;
|
|
+ ulong timeout;
|
|
+ u32 val32;
|
|
+
|
|
+ /*
|
|
+ * A component enters the LTSSM Detect state within
|
|
+ * 20ms of the end of fundamental core reset.
|
|
+ */
|
|
+ msleep(XGENE_LTSSM_DETECT_WAIT);
|
|
+ port->link_up = 0;
|
|
+ timeout = jiffies + msecs_to_jiffies(XGENE_LTSSM_L0_WAIT);
|
|
+ while (time_before(jiffies, timeout)) {
|
|
+ val32 = readl(csr_base + PCIECORE_CTLANDSTATUS);
|
|
+ if (val32 & LINK_UP_MASK) {
|
|
+ port->link_up = 1;
|
|
+ *speed = PIPE_PHY_RATE_RD(val32);
|
|
+ val32 = readl(csr_base + BRIDGE_STATUS_0);
|
|
+ *lanes = val32 >> 26;
|
|
+ break;
|
|
+ }
|
|
+ msleep(1);
|
|
+ }
|
|
+}
|
|
+
|
|
+static void xgene_pcie_setup_root_complex(struct xgene_pcie_port *port)
|
|
+{
|
|
+ void __iomem *csr_base = port->csr_base;
|
|
+ u32 val;
|
|
+
|
|
+ val = (XGENE_PCIE_DEVICEID << 16) | XGENE_PCIE_VENDORID;
|
|
+ writel(val, csr_base + BRIDGE_CFG_0);
|
|
+
|
|
+ val = readl(csr_base + BRIDGE_CFG_1);
|
|
+ val &= ~CLASS_CODE_MASK;
|
|
+ val |= PCI_CLASS_BRIDGE_PCI << 16;
|
|
+ writel(val, csr_base + BRIDGE_CFG_1);
|
|
+
|
|
+ val = readl(csr_base + BRIDGE_CFG_14);
|
|
+ val |= SWITCH_PORT_MODE_MASK;
|
|
+ val &= ~PM_FORCE_RP_MODE_MASK;
|
|
+ writel(val, csr_base + BRIDGE_CFG_14);
|
|
+
|
|
+ val = readl(csr_base + BRIDGE_CTRL_5);
|
|
+ val &= ~DEVICE_PORT_TYPE_MASK;
|
|
+ val |= XGENE_PORT_TYPE_RC;
|
|
+ writel(val, csr_base + BRIDGE_CTRL_5);
|
|
+
|
|
+ val = readl(csr_base + BRIDGE_CTRL_2);
|
|
+ val |= ENABLE_ASPM;
|
|
+ writel(val, csr_base + BRIDGE_CTRL_2);
|
|
+
|
|
+ val = readl(csr_base + BRIDGE_CFG_32);
|
|
+ writel(val | (1 << 19), csr_base + BRIDGE_CFG_32);
|
|
+}
|
|
+
|
|
+/* Return 0 on success */
|
|
+static int xgene_pcie_init_ecc(struct xgene_pcie_port *port)
|
|
+{
|
|
+ void __iomem *csr_base = port->csr_base;
|
|
+ ulong timeout;
|
|
+ u32 val;
|
|
+
|
|
+ val = readl(csr_base + MEM_RAM_SHUTDOWN);
|
|
+ if (!val)
|
|
+ return 0;
|
|
+ writel(0x0, csr_base + MEM_RAM_SHUTDOWN);
|
|
+ timeout = jiffies + msecs_to_jiffies(XGENE_PCIE_ECC_TIMEOUT);
|
|
+ while (time_before(jiffies, timeout)) {
|
|
+ val = readl(csr_base + BLOCK_MEM_RDY);
|
|
+ if (val == BLOCK_MEM_RDY_VAL)
|
|
+ return 0;
|
|
+ msleep(1);
|
|
+ }
|
|
+
|
|
+ return 1;
|
|
+}
|
|
+
|
|
+static int xgene_pcie_init_port(struct xgene_pcie_port *port)
|
|
+{
|
|
+ int rc;
|
|
+
|
|
+ port->clk = clk_get(port->dev, NULL);
|
|
+ if (IS_ERR(port->clk)) {
|
|
+ dev_err(port->dev, "clock not available\n");
|
|
+ return -ENODEV;
|
|
+ }
|
|
+
|
|
+ rc = clk_prepare_enable(port->clk);
|
|
+ if (rc) {
|
|
+ dev_err(port->dev, "clock enable failed\n");
|
|
+ return rc;
|
|
+ }
|
|
+
|
|
+ rc = xgene_pcie_init_ecc(port);
|
|
+ if (rc) {
|
|
+ dev_err(port->dev, "memory init failed\n");
|
|
+ return rc;
|
|
+ }
|
|
+
|
|
+ return 0;
|
|
+}
|
|
+
|
|
+static void xgene_pcie_fixup_bridge(struct pci_dev *dev)
|
|
+{
|
|
+ int i;
|
|
+
|
|
+ /* Hide the PCI host BARs from the kernel as their content doesn't
|
|
+ * fit well in the resource management
|
|
+ */
|
|
+ for (i = 0; i < DEVICE_COUNT_RESOURCE; i++) {
|
|
+ dev->resource[i].start = dev->resource[i].end = 0;
|
|
+ dev->resource[i].flags = 0;
|
|
+ }
|
|
+ dev_info(&dev->dev, "Hiding X-Gene pci host bridge resources %s\n",
|
|
+ pci_name(dev));
|
|
+}
|
|
+DECLARE_PCI_FIXUP_HEADER(XGENE_PCIE_VENDORID, XGENE_PCIE_DEVICEID,
|
|
+ xgene_pcie_fixup_bridge);
|
|
+
|
|
+static int xgene_pcie_map_reg(struct xgene_pcie_port *port,
|
|
+ struct platform_device *pdev, u64 *cfg_addr)
|
|
+{
|
|
+ struct resource *res;
|
|
+
|
|
+ res = platform_get_resource_byname(pdev, IORESOURCE_MEM, "csr");
|
|
+ port->csr_base = devm_ioremap_resource(port->dev, res);
|
|
+ if (IS_ERR(port->csr_base))
|
|
+ return PTR_ERR(port->csr_base);
|
|
+
|
|
+ res = platform_get_resource_byname(pdev, IORESOURCE_MEM, "cfg");
|
|
+ port->cfg_base = devm_ioremap_resource(port->dev, res);
|
|
+ if (IS_ERR(port->cfg_base))
|
|
+ return PTR_ERR(port->cfg_base);
|
|
+ *cfg_addr = res->start;
|
|
+
|
|
+ return 0;
|
|
+}
|
|
+
|
|
+static void xgene_pcie_setup_ob_reg(struct xgene_pcie_port *port,
|
|
+ struct resource *res, u32 offset, u64 addr)
|
|
+{
|
|
+ void __iomem *base = port->csr_base + offset;
|
|
+ resource_size_t size = resource_size(res);
|
|
+ u64 restype = resource_type(res);
|
|
+ u64 cpu_addr, pci_addr;
|
|
+ u64 mask = 0;
|
|
+ u32 min_size;
|
|
+ u32 flag = EN_REG;
|
|
+
|
|
+ if (restype == IORESOURCE_MEM) {
|
|
+ cpu_addr = res->start;
|
|
+ pci_addr = addr;
|
|
+ min_size = SZ_128M;
|
|
+ } else {
|
|
+ cpu_addr = addr;
|
|
+ pci_addr = res->start;
|
|
+ min_size = 128;
|
|
+ flag |= OB_LO_IO;
|
|
+ }
|
|
+ if (size >= min_size)
|
|
+ mask = ~(size - 1) | flag;
|
|
+ else
|
|
+ dev_warn(port->dev, "res size 0x%llx less than minimum 0x%x\n",
|
|
+ (u64)size, min_size);
|
|
+ writel(lower_32_bits(cpu_addr), base);
|
|
+ writel(upper_32_bits(cpu_addr), base + 0x04);
|
|
+ writel(lower_32_bits(mask), base + 0x08);
|
|
+ writel(upper_32_bits(mask), base + 0x0c);
|
|
+ writel(lower_32_bits(pci_addr), base + 0x10);
|
|
+ writel(upper_32_bits(pci_addr), base + 0x14);
|
|
+}
|
|
+
|
|
+static void xgene_pcie_setup_cfg_reg(void __iomem *csr_base, u64 addr)
|
|
+{
|
|
+ writel(lower_32_bits(addr), csr_base + CFGBARL);
|
|
+ writel(upper_32_bits(addr), csr_base + CFGBARH);
|
|
+ writel(EN_REG, csr_base + CFGCTL);
|
|
+}
|
|
+
|
|
+static int xgene_pcie_map_ranges(struct xgene_pcie_port *port,
|
|
+ struct pci_host_bridge *bridge,
|
|
+ u64 cfg_addr)
|
|
+{
|
|
+ struct device *dev = port->dev;
|
|
+ struct pci_host_bridge_window *window;
|
|
+ int ret;
|
|
+
|
|
+ list_for_each_entry(window, &bridge->windows, list) {
|
|
+ struct resource *res = window->res;
|
|
+ u64 restype = resource_type(res);
|
|
+ dev_dbg(port->dev, "0x%08lx 0x%016llx...0x%016llx\n",
|
|
+ res->flags, res->start, res->end);
|
|
+
|
|
+ switch (restype) {
|
|
+ case IORESOURCE_IO:
|
|
+ xgene_pcie_setup_ob_reg(port, res, OMR2BARL,
|
|
+ bridge->io_base);
|
|
+ ret = pci_remap_iospace(res, bridge->io_base);
|
|
+ if (ret < 0)
|
|
+ return ret;
|
|
+ break;
|
|
+ case IORESOURCE_MEM:
|
|
+ xgene_pcie_setup_ob_reg(port, res, OMR3BARL,
|
|
+ res->start - window->offset);
|
|
+ break;
|
|
+ case IORESOURCE_BUS:
|
|
+ break;
|
|
+ default:
|
|
+ dev_err(dev, "invalid io resource!");
|
|
+ return -EINVAL;
|
|
+ }
|
|
+ }
|
|
+ xgene_pcie_setup_cfg_reg(port->csr_base, cfg_addr);
|
|
+ return 0;
|
|
+}
|
|
+
|
|
+static void xgene_pcie_setup_pims(void *addr, u64 pim, u64 size)
|
|
+{
|
|
+ writel(lower_32_bits(pim), addr);
|
|
+ writel(upper_32_bits(pim) | EN_COHERENCY, addr + 0x04);
|
|
+ writel(lower_32_bits(size), addr + 0x10);
|
|
+ writel(upper_32_bits(size), addr + 0x14);
|
|
+}
|
|
+
|
|
+/*
|
|
+ * X-Gene PCIe support maximum 3 inbound memory regions
|
|
+ * This function helps to select a region based on size of region
|
|
+ */
|
|
+static int xgene_pcie_select_ib_reg(u8 *ib_reg_mask, u64 size)
|
|
+{
|
|
+ if ((size > 4) && (size < SZ_16M) && !(*ib_reg_mask & (1 << 1))) {
|
|
+ *ib_reg_mask |= (1 << 1);
|
|
+ return 1;
|
|
+ }
|
|
+
|
|
+ if ((size > SZ_1K) && (size < SZ_1T) && !(*ib_reg_mask & (1 << 0))) {
|
|
+ *ib_reg_mask |= (1 << 0);
|
|
+ return 0;
|
|
+ }
|
|
+
|
|
+ if ((size > SZ_1M) && (size < SZ_1T) && !(*ib_reg_mask & (1 << 2))) {
|
|
+ *ib_reg_mask |= (1 << 2);
|
|
+ return 2;
|
|
+ }
|
|
+ return -EINVAL;
|
|
+}
|
|
+
|
|
+static void xgene_pcie_setup_ib_reg(struct xgene_pcie_port *port,
|
|
+ struct of_pci_range *range, u8 *ib_reg_mask)
|
|
+{
|
|
+ void __iomem *csr_base = port->csr_base;
|
|
+ void __iomem *cfg_base = port->cfg_base;
|
|
+ void *bar_addr;
|
|
+ void *pim_addr;
|
|
+ u64 restype = range->flags & IORESOURCE_TYPE_BITS;
|
|
+ u64 cpu_addr = range->cpu_addr;
|
|
+ u64 pci_addr = range->pci_addr;
|
|
+ u64 size = range->size;
|
|
+ u64 mask = ~(size - 1) | EN_REG;
|
|
+ u32 flags = PCI_BASE_ADDRESS_MEM_TYPE_64;
|
|
+ u32 bar_low;
|
|
+ int region;
|
|
+
|
|
+ region = xgene_pcie_select_ib_reg(ib_reg_mask, range->size);
|
|
+ if (region < 0) {
|
|
+ dev_warn(port->dev, "invalid pcie dma-range config\n");
|
|
+ return;
|
|
+ }
|
|
+
|
|
+ if (restype == PCI_BASE_ADDRESS_MEM_PREFETCH)
|
|
+ flags |= PCI_BASE_ADDRESS_MEM_PREFETCH;
|
|
+
|
|
+ bar_low = pcie_bar_low_val((u32)cpu_addr, flags);
|
|
+ switch (region) {
|
|
+ case 0:
|
|
+ xgene_pcie_set_ib_mask(csr_base, BRIDGE_CFG_4, flags, size);
|
|
+ bar_addr = cfg_base + PCI_BASE_ADDRESS_0;
|
|
+ writel(bar_low, bar_addr);
|
|
+ writel(upper_32_bits(cpu_addr), bar_addr + 0x4);
|
|
+ pim_addr = csr_base + PIM1_1L;
|
|
+ break;
|
|
+ case 1:
|
|
+ bar_addr = csr_base + IBAR2;
|
|
+ writel(bar_low, bar_addr);
|
|
+ writel(lower_32_bits(mask), csr_base + IR2MSK);
|
|
+ pim_addr = csr_base + PIM2_1L;
|
|
+ break;
|
|
+ case 2:
|
|
+ bar_addr = csr_base + IBAR3L;
|
|
+ writel(bar_low, bar_addr);
|
|
+ writel(upper_32_bits(cpu_addr), bar_addr + 0x4);
|
|
+ writel(lower_32_bits(mask), csr_base + IR3MSKL);
|
|
+ writel(upper_32_bits(mask), csr_base + IR3MSKL + 0x4);
|
|
+ pim_addr = csr_base + PIM3_1L;
|
|
+ break;
|
|
+ }
|
|
+
|
|
+ xgene_pcie_setup_pims(pim_addr, pci_addr, size);
|
|
+}
|
|
+
|
|
+static int pci_dma_range_parser_init(struct of_pci_range_parser *parser,
|
|
+ struct device_node *node)
|
|
+{
|
|
+ const int na = 3, ns = 2;
|
|
+ int rlen;
|
|
+
|
|
+ parser->node = node;
|
|
+ parser->pna = of_n_addr_cells(node);
|
|
+ parser->np = parser->pna + na + ns;
|
|
+
|
|
+ parser->range = of_get_property(node, "dma-ranges", &rlen);
|
|
+ if (!parser->range)
|
|
+ return -ENOENT;
|
|
+
|
|
+ parser->end = parser->range + rlen / sizeof(__be32);
|
|
+ return 0;
|
|
+}
|
|
+
|
|
+static int xgene_pcie_parse_map_dma_ranges(struct xgene_pcie_port *port)
|
|
+{
|
|
+ struct device_node *np = port->node;
|
|
+ struct of_pci_range range;
|
|
+ struct of_pci_range_parser parser;
|
|
+ struct device *dev = port->dev;
|
|
+ u8 ib_reg_mask = 0;
|
|
+
|
|
+ if (pci_dma_range_parser_init(&parser, np)) {
|
|
+ dev_err(dev, "missing dma-ranges property\n");
|
|
+ return -EINVAL;
|
|
+ }
|
|
+
|
|
+ /* Get the dma-ranges from DT */
|
|
+ for_each_of_pci_range(&parser, &range) {
|
|
+ u64 end = range.cpu_addr + range.size - 1;
|
|
+ dev_dbg(port->dev, "0x%08x 0x%016llx..0x%016llx -> 0x%016llx\n",
|
|
+ range.flags, range.cpu_addr, end, range.pci_addr);
|
|
+ xgene_pcie_setup_ib_reg(port, &range, &ib_reg_mask);
|
|
+ }
|
|
+ return 0;
|
|
+}
|
|
+
|
|
+static int xgene_pcie_probe_bridge(struct platform_device *pdev)
|
|
+{
|
|
+ struct device_node *np = of_node_get(pdev->dev.of_node);
|
|
+ struct xgene_pcie_port *port;
|
|
+ struct pci_host_bridge *bridge;
|
|
+ resource_size_t lastbus;
|
|
+ u32 lanes = 0, speed = 0;
|
|
+ u64 cfg_addr = 0;
|
|
+ int ret;
|
|
+
|
|
+ port = devm_kzalloc(&pdev->dev, sizeof(*port), GFP_KERNEL);
|
|
+ if (!port)
|
|
+ return -ENOMEM;
|
|
+ port->node = np;
|
|
+ port->dev = &pdev->dev;
|
|
+
|
|
+ ret = xgene_pcie_map_reg(port, pdev, &cfg_addr);
|
|
+ if (ret)
|
|
+ return ret;
|
|
+
|
|
+ ret = xgene_pcie_init_port(port);
|
|
+ if (ret)
|
|
+ return ret;
|
|
+ xgene_pcie_program_core(port->csr_base);
|
|
+ xgene_pcie_setup_root_complex(port);
|
|
+
|
|
+ bridge = of_create_pci_host_bridge(&pdev->dev, &xgene_pcie_ops, port);
|
|
+ if (IS_ERR_OR_NULL(bridge))
|
|
+ return PTR_ERR(bridge);
|
|
+
|
|
+ ret = xgene_pcie_map_ranges(port, bridge, cfg_addr);
|
|
+ if (ret)
|
|
+ return ret;
|
|
+
|
|
+ ret = xgene_pcie_parse_map_dma_ranges(port);
|
|
+ if (ret)
|
|
+ return ret;
|
|
+
|
|
+ xgene_pcie_poll_linkup(port, &lanes, &speed);
|
|
+ if (!port->link_up)
|
|
+ dev_info(port->dev, "(rc) link down\n");
|
|
+ else
|
|
+ dev_info(port->dev, "(rc) x%d gen-%d link up\n",
|
|
+ lanes, speed + 1);
|
|
+ platform_set_drvdata(pdev, port);
|
|
+ lastbus = pci_rescan_bus(bridge->bus);
|
|
+ pci_bus_update_busn_res_end(bridge->bus, lastbus);
|
|
+ return 0;
|
|
+}
|
|
+
|
|
+static const struct of_device_id xgene_pcie_match_table[] = {
|
|
+ {.compatible = "apm,xgene-pcie",},
|
|
+ {},
|
|
+};
|
|
+
|
|
+static struct platform_driver xgene_pcie_driver = {
|
|
+ .driver = {
|
|
+ .name = "xgene-pcie",
|
|
+ .owner = THIS_MODULE,
|
|
+ .of_match_table = of_match_ptr(xgene_pcie_match_table),
|
|
+ },
|
|
+ .probe = xgene_pcie_probe_bridge,
|
|
+};
|
|
+module_platform_driver(xgene_pcie_driver);
|
|
+
|
|
+MODULE_AUTHOR("Tanmay Inamdar <tinamdar@apm.com>");
|
|
+MODULE_DESCRIPTION("APM X-Gene PCIe driver");
|
|
+MODULE_LICENSE("GPL v2");
|
|
diff --git a/drivers/pci/pci.c b/drivers/pci/pci.c
|
|
index 1c8592b..b81dc68 100644
|
|
--- a/drivers/pci/pci.c
|
|
+++ b/drivers/pci/pci.c
|
|
@@ -17,6 +17,7 @@
|
|
#include <linux/spinlock.h>
|
|
#include <linux/string.h>
|
|
#include <linux/log2.h>
|
|
+#include <linux/of_pci.h>
|
|
#include <linux/pci-aspm.h>
|
|
#include <linux/pm_wakeup.h>
|
|
#include <linux/interrupt.h>
|
|
@@ -1453,6 +1454,9 @@ EXPORT_SYMBOL(pcim_pin_device);
|
|
*/
|
|
int __weak pcibios_add_device(struct pci_dev *dev)
|
|
{
|
|
+#ifdef CONFIG_OF
|
|
+ dev->irq = of_irq_parse_and_map_pci(dev, 0, 0);
|
|
+#endif
|
|
return 0;
|
|
}
|
|
|
|
@@ -2704,6 +2708,39 @@ int pci_request_regions_exclusive(struct pci_dev *pdev, const char *res_name)
|
|
}
|
|
EXPORT_SYMBOL(pci_request_regions_exclusive);
|
|
|
|
+/**
|
|
+ * pci_remap_iospace - Remap the memory mapped I/O space
|
|
+ * @res: Resource describing the I/O space
|
|
+ * @phys_addr: physical address where the range will be mapped.
|
|
+ *
|
|
+ * Remap the memory mapped I/O space described by the @res
|
|
+ * into the CPU physical address space. Only architectures
|
|
+ * that have memory mapped IO defined (and hence PCI_IOBASE)
|
|
+ * should call this function.
|
|
+ */
|
|
+int __weak pci_remap_iospace(const struct resource *res, phys_addr_t phys_addr)
|
|
+{
|
|
+ int err = -ENODEV;
|
|
+
|
|
+#ifdef PCI_IOBASE
|
|
+ if (!(res->flags & IORESOURCE_IO))
|
|
+ return -EINVAL;
|
|
+
|
|
+ if (res->end > IO_SPACE_LIMIT)
|
|
+ return -EINVAL;
|
|
+
|
|
+ err = ioremap_page_range(res->start + (unsigned long)PCI_IOBASE,
|
|
+ res->end + 1 + (unsigned long)PCI_IOBASE,
|
|
+ phys_addr, __pgprot(PROT_DEVICE_nGnRE));
|
|
+#else
|
|
+ /* this architecture does not have memory mapped I/O space,
|
|
+ so this function should never be called */
|
|
+ WARN_ON(1);
|
|
+#endif
|
|
+
|
|
+ return err;
|
|
+}
|
|
+
|
|
static void __pci_set_master(struct pci_dev *dev, bool enable)
|
|
{
|
|
u16 old_cmd, cmd;
|
|
diff --git a/drivers/pci/probe.c b/drivers/pci/probe.c
|
|
index e3cf8a2..abf5e82 100644
|
|
--- a/drivers/pci/probe.c
|
|
+++ b/drivers/pci/probe.c
|
|
@@ -515,7 +515,7 @@ static void pci_release_host_bridge_dev(struct device *dev)
|
|
kfree(bridge);
|
|
}
|
|
|
|
-static struct pci_host_bridge *pci_alloc_host_bridge(struct pci_bus *b)
|
|
+static struct pci_host_bridge *pci_alloc_host_bridge(void)
|
|
{
|
|
struct pci_host_bridge *bridge;
|
|
|
|
@@ -524,7 +524,6 @@ static struct pci_host_bridge *pci_alloc_host_bridge(struct pci_bus *b)
|
|
return NULL;
|
|
|
|
INIT_LIST_HEAD(&bridge->windows);
|
|
- bridge->bus = b;
|
|
return bridge;
|
|
}
|
|
|
|
@@ -1749,8 +1748,9 @@ void __weak pcibios_remove_bus(struct pci_bus *bus)
|
|
{
|
|
}
|
|
|
|
-struct pci_bus *pci_create_root_bus(struct device *parent, int bus,
|
|
- struct pci_ops *ops, void *sysdata, struct list_head *resources)
|
|
+struct pci_bus *pci_create_root_bus_in_domain(struct device *parent,
|
|
+ int domain, int bus, struct pci_ops *ops, void *sysdata,
|
|
+ struct list_head *resources)
|
|
{
|
|
int error;
|
|
struct pci_host_bridge *bridge;
|
|
@@ -1761,37 +1761,41 @@ struct pci_bus *pci_create_root_bus(struct device *parent, int bus,
|
|
char bus_addr[64];
|
|
char *fmt;
|
|
|
|
+ bridge = pci_alloc_host_bridge();
|
|
+ if (!bridge)
|
|
+ return ERR_PTR(-ENOMEM);
|
|
+
|
|
+ bridge->dev.parent = parent;
|
|
+ bridge->dev.release = pci_release_host_bridge_dev;
|
|
+ bridge->domain_nr = domain;
|
|
+
|
|
b = pci_alloc_bus();
|
|
- if (!b)
|
|
- return NULL;
|
|
+ if (!b) {
|
|
+ error = -ENOMEM;
|
|
+ goto err_out;
|
|
+ }
|
|
|
|
b->sysdata = sysdata;
|
|
b->ops = ops;
|
|
b->number = b->busn_res.start = bus;
|
|
- b2 = pci_find_bus(pci_domain_nr(b), bus);
|
|
+ b2 = pci_find_bus(bridge->domain_nr, bus);
|
|
if (b2) {
|
|
/* If we already got to this bus through a different bridge, ignore it */
|
|
dev_dbg(&b2->dev, "bus already known\n");
|
|
- goto err_out;
|
|
+ error = -EEXIST;
|
|
+ goto err_bus_out;
|
|
}
|
|
|
|
- bridge = pci_alloc_host_bridge(b);
|
|
- if (!bridge)
|
|
- goto err_out;
|
|
-
|
|
- bridge->dev.parent = parent;
|
|
- bridge->dev.release = pci_release_host_bridge_dev;
|
|
- dev_set_name(&bridge->dev, "pci%04x:%02x", pci_domain_nr(b), bus);
|
|
+ bridge->bus = b;
|
|
+ dev_set_name(&bridge->dev, "pci%04x:%02x", bridge->domain_nr, bus);
|
|
error = pcibios_root_bridge_prepare(bridge);
|
|
- if (error) {
|
|
- kfree(bridge);
|
|
+ if (error)
|
|
goto err_out;
|
|
- }
|
|
|
|
error = device_register(&bridge->dev);
|
|
if (error) {
|
|
put_device(&bridge->dev);
|
|
- goto err_out;
|
|
+ goto err_bus_out;
|
|
}
|
|
b->bridge = get_device(&bridge->dev);
|
|
device_enable_async_suspend(b->bridge);
|
|
@@ -1802,7 +1806,7 @@ struct pci_bus *pci_create_root_bus(struct device *parent, int bus,
|
|
|
|
b->dev.class = &pcibus_class;
|
|
b->dev.parent = b->bridge;
|
|
- dev_set_name(&b->dev, "%04x:%02x", pci_domain_nr(b), bus);
|
|
+ dev_set_name(&b->dev, "%04x:%02x", bridge->domain_nr, bus);
|
|
error = device_register(&b->dev);
|
|
if (error)
|
|
goto class_dev_reg_err;
|
|
@@ -1848,9 +1852,31 @@ struct pci_bus *pci_create_root_bus(struct device *parent, int bus,
|
|
class_dev_reg_err:
|
|
put_device(&bridge->dev);
|
|
device_unregister(&bridge->dev);
|
|
+err_bus_out:
|
|
+ kfree(b);
|
|
err_out:
|
|
+ kfree(bridge);
|
|
+ return ERR_PTR(error);
|
|
+}
|
|
+
|
|
+struct pci_bus *pci_create_root_bus(struct device *parent, int bus,
|
|
+ struct pci_ops *ops, void *sysdata, struct list_head *resources)
|
|
+{
|
|
+ int domain_nr;
|
|
+ struct pci_bus *b = pci_alloc_bus();
|
|
+ if (!b)
|
|
+ return NULL;
|
|
+
|
|
+ b->sysdata = sysdata;
|
|
+ domain_nr = pci_domain_nr(b);
|
|
kfree(b);
|
|
- return NULL;
|
|
+
|
|
+ b = pci_create_root_bus_in_domain(parent, domain_nr, bus,
|
|
+ ops, sysdata, resources);
|
|
+ if (IS_ERR(b))
|
|
+ return NULL;
|
|
+
|
|
+ return b;
|
|
}
|
|
|
|
int pci_bus_insert_busn_res(struct pci_bus *b, int bus, int bus_max)
|
|
diff --git a/drivers/rtc/Kconfig b/drivers/rtc/Kconfig
|
|
index 0754f5c..4478a59 100644
|
|
--- a/drivers/rtc/Kconfig
|
|
+++ b/drivers/rtc/Kconfig
|
|
@@ -789,7 +789,7 @@ config RTC_DRV_DA9063
|
|
|
|
config RTC_DRV_EFI
|
|
tristate "EFI RTC"
|
|
- depends on IA64
|
|
+ depends on EFI
|
|
help
|
|
If you say yes here you will get support for the EFI
|
|
Real Time Clock.
|
|
diff --git a/drivers/rtc/Makefile b/drivers/rtc/Makefile
|
|
index 70347d0..f1dfc36 100644
|
|
--- a/drivers/rtc/Makefile
|
|
+++ b/drivers/rtc/Makefile
|
|
@@ -10,6 +10,10 @@ obj-$(CONFIG_RTC_SYSTOHC) += systohc.o
|
|
obj-$(CONFIG_RTC_CLASS) += rtc-core.o
|
|
rtc-core-y := class.o interface.o
|
|
|
|
+ifdef CONFIG_RTC_DRV_EFI
|
|
+rtc-core-y += rtc-efi-platform.o
|
|
+endif
|
|
+
|
|
rtc-core-$(CONFIG_RTC_INTF_DEV) += rtc-dev.o
|
|
rtc-core-$(CONFIG_RTC_INTF_PROC) += rtc-proc.o
|
|
rtc-core-$(CONFIG_RTC_INTF_SYSFS) += rtc-sysfs.o
|
|
diff --git a/drivers/rtc/rtc-efi-platform.c b/drivers/rtc/rtc-efi-platform.c
|
|
new file mode 100644
|
|
index 0000000..b40fbe3
|
|
--- /dev/null
|
|
+++ b/drivers/rtc/rtc-efi-platform.c
|
|
@@ -0,0 +1,31 @@
|
|
+/*
|
|
+ * Moved from arch/ia64/kernel/time.c
|
|
+ *
|
|
+ * Copyright (C) 1998-2003 Hewlett-Packard Co
|
|
+ * Stephane Eranian <eranian@hpl.hp.com>
|
|
+ * David Mosberger <davidm@hpl.hp.com>
|
|
+ * Copyright (C) 1999 Don Dugger <don.dugger@intel.com>
|
|
+ * Copyright (C) 1999-2000 VA Linux Systems
|
|
+ * Copyright (C) 1999-2000 Walt Drummond <drummond@valinux.com>
|
|
+ */
|
|
+#include <linux/init.h>
|
|
+#include <linux/kernel.h>
|
|
+#include <linux/module.h>
|
|
+#include <linux/efi.h>
|
|
+#include <linux/platform_device.h>
|
|
+
|
|
+static struct platform_device rtc_efi_dev = {
|
|
+ .name = "rtc-efi",
|
|
+ .id = -1,
|
|
+};
|
|
+
|
|
+static int __init rtc_init(void)
|
|
+{
|
|
+ if (efi_enabled(EFI_RUNTIME_SERVICES))
|
|
+ if (platform_device_register(&rtc_efi_dev) < 0)
|
|
+ pr_err("unable to register rtc device...\n");
|
|
+
|
|
+ /* not necessarily an error */
|
|
+ return 0;
|
|
+}
|
|
+module_init(rtc_init);
|
|
diff --git a/include/asm-generic/io.h b/include/asm-generic/io.h
|
|
index 975e1cc..2e2161b 100644
|
|
--- a/include/asm-generic/io.h
|
|
+++ b/include/asm-generic/io.h
|
|
@@ -331,7 +331,7 @@ static inline void iounmap(void __iomem *addr)
|
|
#ifndef CONFIG_GENERIC_IOMAP
|
|
static inline void __iomem *ioport_map(unsigned long port, unsigned int nr)
|
|
{
|
|
- return (void __iomem *) port;
|
|
+ return (void __iomem *)(PCI_IOBASE + (port & IO_SPACE_LIMIT));
|
|
}
|
|
|
|
static inline void ioport_unmap(void __iomem *p)
|
|
diff --git a/include/kvm/arm_vgic.h b/include/kvm/arm_vgic.h
|
|
index f27000f..35b0c12 100644
|
|
--- a/include/kvm/arm_vgic.h
|
|
+++ b/include/kvm/arm_vgic.h
|
|
@@ -24,7 +24,6 @@
|
|
#include <linux/irqreturn.h>
|
|
#include <linux/spinlock.h>
|
|
#include <linux/types.h>
|
|
-#include <linux/irqchip/arm-gic.h>
|
|
|
|
#define VGIC_NR_IRQS 256
|
|
#define VGIC_NR_SGIS 16
|
|
@@ -32,7 +31,9 @@
|
|
#define VGIC_NR_PRIVATE_IRQS (VGIC_NR_SGIS + VGIC_NR_PPIS)
|
|
#define VGIC_NR_SHARED_IRQS (VGIC_NR_IRQS - VGIC_NR_PRIVATE_IRQS)
|
|
#define VGIC_MAX_CPUS KVM_MAX_VCPUS
|
|
-#define VGIC_MAX_LRS (1 << 6)
|
|
+
|
|
+#define VGIC_V2_MAX_LRS (1 << 6)
|
|
+#define VGIC_V3_MAX_LRS 16
|
|
|
|
/* Sanity checks... */
|
|
#if (VGIC_MAX_CPUS > 8)
|
|
@@ -68,9 +69,62 @@ struct vgic_bytemap {
|
|
u32 shared[VGIC_NR_SHARED_IRQS / 4];
|
|
};
|
|
|
|
+struct kvm_vcpu;
|
|
+
|
|
+enum vgic_type {
|
|
+ VGIC_V2, /* Good ol' GICv2 */
|
|
+ VGIC_V3, /* New fancy GICv3 */
|
|
+};
|
|
+
|
|
+#define LR_STATE_PENDING (1 << 0)
|
|
+#define LR_STATE_ACTIVE (1 << 1)
|
|
+#define LR_STATE_MASK (3 << 0)
|
|
+#define LR_EOI_INT (1 << 2)
|
|
+
|
|
+struct vgic_lr {
|
|
+ u16 irq;
|
|
+ u8 source;
|
|
+ u8 state;
|
|
+};
|
|
+
|
|
+struct vgic_vmcr {
|
|
+ u32 ctlr;
|
|
+ u32 abpr;
|
|
+ u32 bpr;
|
|
+ u32 pmr;
|
|
+};
|
|
+
|
|
+struct vgic_ops {
|
|
+ struct vgic_lr (*get_lr)(const struct kvm_vcpu *, int);
|
|
+ void (*set_lr)(struct kvm_vcpu *, int, struct vgic_lr);
|
|
+ void (*sync_lr_elrsr)(struct kvm_vcpu *, int, struct vgic_lr);
|
|
+ u64 (*get_elrsr)(const struct kvm_vcpu *vcpu);
|
|
+ u64 (*get_eisr)(const struct kvm_vcpu *vcpu);
|
|
+ u32 (*get_interrupt_status)(const struct kvm_vcpu *vcpu);
|
|
+ void (*enable_underflow)(struct kvm_vcpu *vcpu);
|
|
+ void (*disable_underflow)(struct kvm_vcpu *vcpu);
|
|
+ void (*get_vmcr)(struct kvm_vcpu *vcpu, struct vgic_vmcr *vmcr);
|
|
+ void (*set_vmcr)(struct kvm_vcpu *vcpu, struct vgic_vmcr *vmcr);
|
|
+ void (*enable)(struct kvm_vcpu *vcpu);
|
|
+};
|
|
+
|
|
+struct vgic_params {
|
|
+ /* vgic type */
|
|
+ enum vgic_type type;
|
|
+ /* Physical address of vgic virtual cpu interface */
|
|
+ phys_addr_t vcpu_base;
|
|
+ /* Number of list registers */
|
|
+ u32 nr_lr;
|
|
+ /* Interrupt number */
|
|
+ unsigned int maint_irq;
|
|
+ /* Virtual control interface base address */
|
|
+ void __iomem *vctrl_base;
|
|
+};
|
|
+
|
|
struct vgic_dist {
|
|
#ifdef CONFIG_KVM_ARM_VGIC
|
|
spinlock_t lock;
|
|
+ bool in_kernel;
|
|
bool ready;
|
|
|
|
/* Virtual control interface mapping */
|
|
@@ -110,6 +164,29 @@ struct vgic_dist {
|
|
#endif
|
|
};
|
|
|
|
+struct vgic_v2_cpu_if {
|
|
+ u32 vgic_hcr;
|
|
+ u32 vgic_vmcr;
|
|
+ u32 vgic_misr; /* Saved only */
|
|
+ u32 vgic_eisr[2]; /* Saved only */
|
|
+ u32 vgic_elrsr[2]; /* Saved only */
|
|
+ u32 vgic_apr;
|
|
+ u32 vgic_lr[VGIC_V2_MAX_LRS];
|
|
+};
|
|
+
|
|
+struct vgic_v3_cpu_if {
|
|
+#ifdef CONFIG_ARM_GIC_V3
|
|
+ u32 vgic_hcr;
|
|
+ u32 vgic_vmcr;
|
|
+ u32 vgic_misr; /* Saved only */
|
|
+ u32 vgic_eisr; /* Saved only */
|
|
+ u32 vgic_elrsr; /* Saved only */
|
|
+ u32 vgic_ap0r[4];
|
|
+ u32 vgic_ap1r[4];
|
|
+ u64 vgic_lr[VGIC_V3_MAX_LRS];
|
|
+#endif
|
|
+};
|
|
+
|
|
struct vgic_cpu {
|
|
#ifdef CONFIG_KVM_ARM_VGIC
|
|
/* per IRQ to LR mapping */
|
|
@@ -120,24 +197,24 @@ struct vgic_cpu {
|
|
DECLARE_BITMAP( pending_shared, VGIC_NR_SHARED_IRQS);
|
|
|
|
/* Bitmap of used/free list registers */
|
|
- DECLARE_BITMAP( lr_used, VGIC_MAX_LRS);
|
|
+ DECLARE_BITMAP( lr_used, VGIC_V2_MAX_LRS);
|
|
|
|
/* Number of list registers on this CPU */
|
|
int nr_lr;
|
|
|
|
/* CPU vif control registers for world switch */
|
|
- u32 vgic_hcr;
|
|
- u32 vgic_vmcr;
|
|
- u32 vgic_misr; /* Saved only */
|
|
- u32 vgic_eisr[2]; /* Saved only */
|
|
- u32 vgic_elrsr[2]; /* Saved only */
|
|
- u32 vgic_apr;
|
|
- u32 vgic_lr[VGIC_MAX_LRS];
|
|
+ union {
|
|
+ struct vgic_v2_cpu_if vgic_v2;
|
|
+ struct vgic_v3_cpu_if vgic_v3;
|
|
+ };
|
|
#endif
|
|
};
|
|
|
|
#define LR_EMPTY 0xff
|
|
|
|
+#define INT_STATUS_EOI (1 << 0)
|
|
+#define INT_STATUS_UNDERFLOW (1 << 1)
|
|
+
|
|
struct kvm;
|
|
struct kvm_vcpu;
|
|
struct kvm_run;
|
|
@@ -157,9 +234,25 @@ int kvm_vgic_vcpu_pending_irq(struct kvm_vcpu *vcpu);
|
|
bool vgic_handle_mmio(struct kvm_vcpu *vcpu, struct kvm_run *run,
|
|
struct kvm_exit_mmio *mmio);
|
|
|
|
-#define irqchip_in_kernel(k) (!!((k)->arch.vgic.vctrl_base))
|
|
+#define irqchip_in_kernel(k) (!!((k)->arch.vgic.in_kernel))
|
|
#define vgic_initialized(k) ((k)->arch.vgic.ready)
|
|
|
|
+int vgic_v2_probe(struct device_node *vgic_node,
|
|
+ const struct vgic_ops **ops,
|
|
+ const struct vgic_params **params);
|
|
+#ifdef CONFIG_ARM_GIC_V3
|
|
+int vgic_v3_probe(struct device_node *vgic_node,
|
|
+ const struct vgic_ops **ops,
|
|
+ const struct vgic_params **params);
|
|
+#else
|
|
+static inline int vgic_v3_probe(struct device_node *vgic_node,
|
|
+ const struct vgic_ops **ops,
|
|
+ const struct vgic_params **params)
|
|
+{
|
|
+ return -ENODEV;
|
|
+}
|
|
+#endif
|
|
+
|
|
#else
|
|
static inline int kvm_vgic_hyp_init(void)
|
|
{
|
|
diff --git a/include/linux/efi.h b/include/linux/efi.h
|
|
index 41bbf8b..b3fac7c 100644
|
|
--- a/include/linux/efi.h
|
|
+++ b/include/linux/efi.h
|
|
@@ -20,6 +20,7 @@
|
|
#include <linux/ioport.h>
|
|
#include <linux/pfn.h>
|
|
#include <linux/pstore.h>
|
|
+#include <linux/reboot.h>
|
|
|
|
#include <asm/page.h>
|
|
|
|
@@ -875,6 +876,9 @@ extern void efi_reserve_boot_services(void);
|
|
extern int efi_get_fdt_params(struct efi_fdt_params *params, int verbose);
|
|
extern struct efi_memory_map memmap;
|
|
|
|
+extern int efi_reboot_quirk_mode;
|
|
+extern bool efi_poweroff_required(void);
|
|
+
|
|
/* Iterate through an efi_memory_map */
|
|
#define for_each_efi_memory_desc(m, md) \
|
|
for ((md) = (m)->map; \
|
|
@@ -926,11 +930,14 @@ static inline bool efi_enabled(int feature)
|
|
{
|
|
return test_bit(feature, &efi.flags) != 0;
|
|
}
|
|
+extern void efi_reboot(enum reboot_mode reboot_mode, const char *__unused);
|
|
#else
|
|
static inline bool efi_enabled(int feature)
|
|
{
|
|
return false;
|
|
}
|
|
+static inline void
|
|
+efi_reboot(enum reboot_mode reboot_mode, const char *__unused) {}
|
|
#endif
|
|
|
|
/*
|
|
diff --git a/include/linux/irqchip/arm-gic-v3.h b/include/linux/irqchip/arm-gic-v3.h
|
|
new file mode 100644
|
|
index 0000000..30cb755
|
|
--- /dev/null
|
|
+++ b/include/linux/irqchip/arm-gic-v3.h
|
|
@@ -0,0 +1,198 @@
|
|
+/*
|
|
+ * Copyright (C) 2013, 2014 ARM Limited, All Rights Reserved.
|
|
+ * Author: Marc Zyngier <marc.zyngier@arm.com>
|
|
+ *
|
|
+ *
|
|
+ * This program is free software; you can redistribute it and/or modify
|
|
+ * it under the terms of the GNU General Public License version 2 as
|
|
+ * published by the Free Software Foundation.
|
|
+ *
|
|
+ * This program is distributed in the hope that it will be useful,
|
|
+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
+ * GNU General Public License for more details.
|
|
+ *
|
|
+ * You should have received a copy of the GNU General Public License
|
|
+ * along with this program. If not, see <http://www.gnu.org/licenses/>.
|
|
+ */
|
|
+#ifndef __LINUX_IRQCHIP_ARM_GIC_V3_H
|
|
+#define __LINUX_IRQCHIP_ARM_GIC_V3_H
|
|
+
|
|
+/*
|
|
+ * Distributor registers. We assume we're running non-secure, with ARE
|
|
+ * being set. Secure-only and non-ARE registers are not described.
|
|
+ */
|
|
+#define GICD_CTLR 0x0000
|
|
+#define GICD_TYPER 0x0004
|
|
+#define GICD_IIDR 0x0008
|
|
+#define GICD_STATUSR 0x0010
|
|
+#define GICD_SETSPI_NSR 0x0040
|
|
+#define GICD_CLRSPI_NSR 0x0048
|
|
+#define GICD_SETSPI_SR 0x0050
|
|
+#define GICD_CLRSPI_SR 0x0058
|
|
+#define GICD_SEIR 0x0068
|
|
+#define GICD_ISENABLER 0x0100
|
|
+#define GICD_ICENABLER 0x0180
|
|
+#define GICD_ISPENDR 0x0200
|
|
+#define GICD_ICPENDR 0x0280
|
|
+#define GICD_ISACTIVER 0x0300
|
|
+#define GICD_ICACTIVER 0x0380
|
|
+#define GICD_IPRIORITYR 0x0400
|
|
+#define GICD_ICFGR 0x0C00
|
|
+#define GICD_IROUTER 0x6000
|
|
+#define GICD_PIDR2 0xFFE8
|
|
+
|
|
+#define GICD_CTLR_RWP (1U << 31)
|
|
+#define GICD_CTLR_ARE_NS (1U << 4)
|
|
+#define GICD_CTLR_ENABLE_G1A (1U << 1)
|
|
+#define GICD_CTLR_ENABLE_G1 (1U << 0)
|
|
+
|
|
+#define GICD_IROUTER_SPI_MODE_ONE (0U << 31)
|
|
+#define GICD_IROUTER_SPI_MODE_ANY (1U << 31)
|
|
+
|
|
+#define GIC_PIDR2_ARCH_MASK 0xf0
|
|
+#define GIC_PIDR2_ARCH_GICv3 0x30
|
|
+#define GIC_PIDR2_ARCH_GICv4 0x40
|
|
+
|
|
+/*
|
|
+ * Re-Distributor registers, offsets from RD_base
|
|
+ */
|
|
+#define GICR_CTLR GICD_CTLR
|
|
+#define GICR_IIDR 0x0004
|
|
+#define GICR_TYPER 0x0008
|
|
+#define GICR_STATUSR GICD_STATUSR
|
|
+#define GICR_WAKER 0x0014
|
|
+#define GICR_SETLPIR 0x0040
|
|
+#define GICR_CLRLPIR 0x0048
|
|
+#define GICR_SEIR GICD_SEIR
|
|
+#define GICR_PROPBASER 0x0070
|
|
+#define GICR_PENDBASER 0x0078
|
|
+#define GICR_INVLPIR 0x00A0
|
|
+#define GICR_INVALLR 0x00B0
|
|
+#define GICR_SYNCR 0x00C0
|
|
+#define GICR_MOVLPIR 0x0100
|
|
+#define GICR_MOVALLR 0x0110
|
|
+#define GICR_PIDR2 GICD_PIDR2
|
|
+
|
|
+#define GICR_WAKER_ProcessorSleep (1U << 1)
|
|
+#define GICR_WAKER_ChildrenAsleep (1U << 2)
|
|
+
|
|
+/*
|
|
+ * Re-Distributor registers, offsets from SGI_base
|
|
+ */
|
|
+#define GICR_ISENABLER0 GICD_ISENABLER
|
|
+#define GICR_ICENABLER0 GICD_ICENABLER
|
|
+#define GICR_ISPENDR0 GICD_ISPENDR
|
|
+#define GICR_ICPENDR0 GICD_ICPENDR
|
|
+#define GICR_ISACTIVER0 GICD_ISACTIVER
|
|
+#define GICR_ICACTIVER0 GICD_ICACTIVER
|
|
+#define GICR_IPRIORITYR0 GICD_IPRIORITYR
|
|
+#define GICR_ICFGR0 GICD_ICFGR
|
|
+
|
|
+#define GICR_TYPER_VLPIS (1U << 1)
|
|
+#define GICR_TYPER_LAST (1U << 4)
|
|
+
|
|
+/*
|
|
+ * CPU interface registers
|
|
+ */
|
|
+#define ICC_CTLR_EL1_EOImode_drop_dir (0U << 1)
|
|
+#define ICC_CTLR_EL1_EOImode_drop (1U << 1)
|
|
+#define ICC_SRE_EL1_SRE (1U << 0)
|
|
+
|
|
+/*
|
|
+ * Hypervisor interface registers (SRE only)
|
|
+ */
|
|
+#define ICH_LR_VIRTUAL_ID_MASK ((1UL << 32) - 1)
|
|
+
|
|
+#define ICH_LR_EOI (1UL << 41)
|
|
+#define ICH_LR_GROUP (1UL << 60)
|
|
+#define ICH_LR_STATE (3UL << 62)
|
|
+#define ICH_LR_PENDING_BIT (1UL << 62)
|
|
+#define ICH_LR_ACTIVE_BIT (1UL << 63)
|
|
+
|
|
+#define ICH_MISR_EOI (1 << 0)
|
|
+#define ICH_MISR_U (1 << 1)
|
|
+
|
|
+#define ICH_HCR_EN (1 << 0)
|
|
+#define ICH_HCR_UIE (1 << 1)
|
|
+
|
|
+#define ICH_VMCR_CTLR_SHIFT 0
|
|
+#define ICH_VMCR_CTLR_MASK (0x21f << ICH_VMCR_CTLR_SHIFT)
|
|
+#define ICH_VMCR_BPR1_SHIFT 18
|
|
+#define ICH_VMCR_BPR1_MASK (7 << ICH_VMCR_BPR1_SHIFT)
|
|
+#define ICH_VMCR_BPR0_SHIFT 21
|
|
+#define ICH_VMCR_BPR0_MASK (7 << ICH_VMCR_BPR0_SHIFT)
|
|
+#define ICH_VMCR_PMR_SHIFT 24
|
|
+#define ICH_VMCR_PMR_MASK (0xffUL << ICH_VMCR_PMR_SHIFT)
|
|
+
|
|
+#define ICC_EOIR1_EL1 S3_0_C12_C12_1
|
|
+#define ICC_IAR1_EL1 S3_0_C12_C12_0
|
|
+#define ICC_SGI1R_EL1 S3_0_C12_C11_5
|
|
+#define ICC_PMR_EL1 S3_0_C4_C6_0
|
|
+#define ICC_CTLR_EL1 S3_0_C12_C12_4
|
|
+#define ICC_SRE_EL1 S3_0_C12_C12_5
|
|
+#define ICC_GRPEN1_EL1 S3_0_C12_C12_7
|
|
+
|
|
+#define ICC_IAR1_EL1_SPURIOUS 0x3ff
|
|
+
|
|
+#define ICC_SRE_EL2 S3_4_C12_C9_5
|
|
+
|
|
+#define ICC_SRE_EL2_SRE (1 << 0)
|
|
+#define ICC_SRE_EL2_ENABLE (1 << 3)
|
|
+
|
|
+/*
|
|
+ * System register definitions
|
|
+ */
|
|
+#define ICH_VSEIR_EL2 S3_4_C12_C9_4
|
|
+#define ICH_HCR_EL2 S3_4_C12_C11_0
|
|
+#define ICH_VTR_EL2 S3_4_C12_C11_1
|
|
+#define ICH_MISR_EL2 S3_4_C12_C11_2
|
|
+#define ICH_EISR_EL2 S3_4_C12_C11_3
|
|
+#define ICH_ELSR_EL2 S3_4_C12_C11_5
|
|
+#define ICH_VMCR_EL2 S3_4_C12_C11_7
|
|
+
|
|
+#define __LR0_EL2(x) S3_4_C12_C12_ ## x
|
|
+#define __LR8_EL2(x) S3_4_C12_C13_ ## x
|
|
+
|
|
+#define ICH_LR0_EL2 __LR0_EL2(0)
|
|
+#define ICH_LR1_EL2 __LR0_EL2(1)
|
|
+#define ICH_LR2_EL2 __LR0_EL2(2)
|
|
+#define ICH_LR3_EL2 __LR0_EL2(3)
|
|
+#define ICH_LR4_EL2 __LR0_EL2(4)
|
|
+#define ICH_LR5_EL2 __LR0_EL2(5)
|
|
+#define ICH_LR6_EL2 __LR0_EL2(6)
|
|
+#define ICH_LR7_EL2 __LR0_EL2(7)
|
|
+#define ICH_LR8_EL2 __LR8_EL2(0)
|
|
+#define ICH_LR9_EL2 __LR8_EL2(1)
|
|
+#define ICH_LR10_EL2 __LR8_EL2(2)
|
|
+#define ICH_LR11_EL2 __LR8_EL2(3)
|
|
+#define ICH_LR12_EL2 __LR8_EL2(4)
|
|
+#define ICH_LR13_EL2 __LR8_EL2(5)
|
|
+#define ICH_LR14_EL2 __LR8_EL2(6)
|
|
+#define ICH_LR15_EL2 __LR8_EL2(7)
|
|
+
|
|
+#define __AP0Rx_EL2(x) S3_4_C12_C8_ ## x
|
|
+#define ICH_AP0R0_EL2 __AP0Rx_EL2(0)
|
|
+#define ICH_AP0R1_EL2 __AP0Rx_EL2(1)
|
|
+#define ICH_AP0R2_EL2 __AP0Rx_EL2(2)
|
|
+#define ICH_AP0R3_EL2 __AP0Rx_EL2(3)
|
|
+
|
|
+#define __AP1Rx_EL2(x) S3_4_C12_C9_ ## x
|
|
+#define ICH_AP1R0_EL2 __AP1Rx_EL2(0)
|
|
+#define ICH_AP1R1_EL2 __AP1Rx_EL2(1)
|
|
+#define ICH_AP1R2_EL2 __AP1Rx_EL2(2)
|
|
+#define ICH_AP1R3_EL2 __AP1Rx_EL2(3)
|
|
+
|
|
+#ifndef __ASSEMBLY__
|
|
+
|
|
+#include <linux/stringify.h>
|
|
+
|
|
+static inline void gic_write_eoir(u64 irq)
|
|
+{
|
|
+ asm volatile("msr " __stringify(ICC_EOIR1_EL1) ", %0" : : "r" (irq));
|
|
+ isb();
|
|
+}
|
|
+
|
|
+#endif
|
|
+
|
|
+#endif
|
|
diff --git a/include/linux/of_address.h b/include/linux/of_address.h
|
|
index c13b878..33c0420 100644
|
|
--- a/include/linux/of_address.h
|
|
+++ b/include/linux/of_address.h
|
|
@@ -23,17 +23,8 @@ struct of_pci_range {
|
|
#define for_each_of_pci_range(parser, range) \
|
|
for (; of_pci_range_parser_one(parser, range);)
|
|
|
|
-static inline void of_pci_range_to_resource(struct of_pci_range *range,
|
|
- struct device_node *np,
|
|
- struct resource *res)
|
|
-{
|
|
- res->flags = range->flags;
|
|
- res->start = range->cpu_addr;
|
|
- res->end = range->cpu_addr + range->size - 1;
|
|
- res->parent = res->child = res->sibling = NULL;
|
|
- res->name = np->full_name;
|
|
-}
|
|
-
|
|
+extern int of_pci_range_to_resource(struct of_pci_range *range,
|
|
+ struct device_node *np, struct resource *res);
|
|
/* Translate a DMA address from device space to CPU space */
|
|
extern u64 of_translate_dma_address(struct device_node *dev,
|
|
const __be32 *in_addr);
|
|
@@ -55,6 +46,7 @@ extern void __iomem *of_iomap(struct device_node *device, int index);
|
|
extern const __be32 *of_get_address(struct device_node *dev, int index,
|
|
u64 *size, unsigned int *flags);
|
|
|
|
+extern int pci_register_io_range(phys_addr_t addr, resource_size_t size);
|
|
extern unsigned long pci_address_to_pio(phys_addr_t addr);
|
|
|
|
extern int of_pci_range_parser_init(struct of_pci_range_parser *parser,
|
|
diff --git a/include/linux/of_pci.h b/include/linux/of_pci.h
|
|
index dde3a4a..71e36d0 100644
|
|
--- a/include/linux/of_pci.h
|
|
+++ b/include/linux/of_pci.h
|
|
@@ -15,6 +15,9 @@ struct device_node *of_pci_find_child_device(struct device_node *parent,
|
|
int of_pci_get_devfn(struct device_node *np);
|
|
int of_irq_parse_and_map_pci(const struct pci_dev *dev, u8 slot, u8 pin);
|
|
int of_pci_parse_bus_range(struct device_node *node, struct resource *res);
|
|
+struct pci_host_bridge *of_create_pci_host_bridge(struct device *parent,
|
|
+ struct pci_ops *ops, void *host_data);
|
|
+
|
|
#else
|
|
static inline int of_irq_parse_pci(const struct pci_dev *pdev, struct of_phandle_args *out_irq)
|
|
{
|
|
@@ -43,6 +46,13 @@ of_pci_parse_bus_range(struct device_node *node, struct resource *res)
|
|
{
|
|
return -EINVAL;
|
|
}
|
|
+
|
|
+static inline struct pci_host_bridge *
|
|
+of_create_pci_host_bridge(struct device *parent, struct pci_ops *ops,
|
|
+ void *host_data)
|
|
+{
|
|
+ return NULL;
|
|
+}
|
|
#endif
|
|
|
|
#if defined(CONFIG_OF) && defined(CONFIG_PCI_MSI)
|
|
diff --git a/include/linux/pci.h b/include/linux/pci.h
|
|
index 466bcd1..65fb1fc 100644
|
|
--- a/include/linux/pci.h
|
|
+++ b/include/linux/pci.h
|
|
@@ -401,6 +401,8 @@ struct pci_host_bridge_window {
|
|
struct pci_host_bridge {
|
|
struct device dev;
|
|
struct pci_bus *bus; /* root bus */
|
|
+ int domain_nr;
|
|
+ resource_size_t io_base; /* physical address for the start of I/O area */
|
|
struct list_head windows; /* pci_host_bridge_windows */
|
|
void (*release_fn)(struct pci_host_bridge *);
|
|
void *release_data;
|
|
@@ -769,6 +771,9 @@ struct pci_bus *pci_scan_bus(int bus, struct pci_ops *ops, void *sysdata);
|
|
struct pci_bus *pci_create_root_bus(struct device *parent, int bus,
|
|
struct pci_ops *ops, void *sysdata,
|
|
struct list_head *resources);
|
|
+struct pci_bus *pci_create_root_bus_in_domain(struct device *parent,
|
|
+ int domain, int bus, struct pci_ops *ops,
|
|
+ void *sysdata, struct list_head *resources);
|
|
int pci_bus_insert_busn_res(struct pci_bus *b, int bus, int busmax);
|
|
int pci_bus_update_busn_res_end(struct pci_bus *b, int busmax);
|
|
void pci_bus_release_busn_res(struct pci_bus *b);
|
|
@@ -1095,6 +1100,9 @@ int __must_check pci_bus_alloc_resource(struct pci_bus *bus,
|
|
resource_size_t),
|
|
void *alignf_data);
|
|
|
|
+
|
|
+int pci_remap_iospace(const struct resource *res, phys_addr_t phys_addr);
|
|
+
|
|
static inline dma_addr_t pci_bus_address(struct pci_dev *pdev, int bar)
|
|
{
|
|
struct pci_bus_region region;
|
|
@@ -1805,8 +1813,15 @@ static inline void pci_set_of_node(struct pci_dev *dev) { }
|
|
static inline void pci_release_of_node(struct pci_dev *dev) { }
|
|
static inline void pci_set_bus_of_node(struct pci_bus *bus) { }
|
|
static inline void pci_release_bus_of_node(struct pci_bus *bus) { }
|
|
+
|
|
#endif /* CONFIG_OF */
|
|
|
|
+/* Used by architecture code to apply any quirks to the list of
|
|
+ * pci_host_bridge resource ranges before they are being used
|
|
+ * by of_create_pci_host_bridge()
|
|
+ */
|
|
+extern int pcibios_fixup_bridge_ranges(struct list_head *resources);
|
|
+
|
|
#ifdef CONFIG_EEH
|
|
static inline struct eeh_dev *pci_dev_to_eeh_dev(struct pci_dev *pdev)
|
|
{
|
|
diff --git a/tools/perf/arch/arm64/include/perf_regs.h b/tools/perf/arch/arm64/include/perf_regs.h
|
|
index e9441b9..1d3f39c 100644
|
|
--- a/tools/perf/arch/arm64/include/perf_regs.h
|
|
+++ b/tools/perf/arch/arm64/include/perf_regs.h
|
|
@@ -6,6 +6,8 @@
|
|
#include <asm/perf_regs.h>
|
|
|
|
#define PERF_REGS_MASK ((1ULL << PERF_REG_ARM64_MAX) - 1)
|
|
+#define PERF_REGS_MAX PERF_REG_ARM64_MAX
|
|
+
|
|
#define PERF_REG_IP PERF_REG_ARM64_PC
|
|
#define PERF_REG_SP PERF_REG_ARM64_SP
|
|
|
|
diff --git a/virt/kvm/arm/vgic-v2.c b/virt/kvm/arm/vgic-v2.c
|
|
new file mode 100644
|
|
index 0000000..5fd2b75
|
|
--- /dev/null
|
|
+++ b/virt/kvm/arm/vgic-v2.c
|
|
@@ -0,0 +1,243 @@
|
|
+/*
|
|
+ * Copyright (C) 2012,2013 ARM Limited, All Rights Reserved.
|
|
+ * Author: Marc Zyngier <marc.zyngier@arm.com>
|
|
+ *
|
|
+ * This program is free software; you can redistribute it and/or modify
|
|
+ * it under the terms of the GNU General Public License version 2 as
|
|
+ * published by the Free Software Foundation.
|
|
+ *
|
|
+ * This program is distributed in the hope that it will be useful,
|
|
+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
+ * GNU General Public License for more details.
|
|
+ *
|
|
+ * You should have received a copy of the GNU General Public License
|
|
+ * along with this program. If not, see <http://www.gnu.org/licenses/>.
|
|
+ */
|
|
+
|
|
+#include <linux/cpu.h>
|
|
+#include <linux/kvm.h>
|
|
+#include <linux/kvm_host.h>
|
|
+#include <linux/interrupt.h>
|
|
+#include <linux/io.h>
|
|
+#include <linux/of.h>
|
|
+#include <linux/of_address.h>
|
|
+#include <linux/of_irq.h>
|
|
+
|
|
+#include <linux/irqchip/arm-gic.h>
|
|
+
|
|
+#include <asm/kvm_emulate.h>
|
|
+#include <asm/kvm_arm.h>
|
|
+#include <asm/kvm_mmu.h>
|
|
+
|
|
+static struct vgic_lr vgic_v2_get_lr(const struct kvm_vcpu *vcpu, int lr)
|
|
+{
|
|
+ struct vgic_lr lr_desc;
|
|
+ u32 val = vcpu->arch.vgic_cpu.vgic_v2.vgic_lr[lr];
|
|
+
|
|
+ lr_desc.irq = val & GICH_LR_VIRTUALID;
|
|
+ if (lr_desc.irq <= 15)
|
|
+ lr_desc.source = (val >> GICH_LR_PHYSID_CPUID_SHIFT) & 0x7;
|
|
+ else
|
|
+ lr_desc.source = 0;
|
|
+ lr_desc.state = 0;
|
|
+
|
|
+ if (val & GICH_LR_PENDING_BIT)
|
|
+ lr_desc.state |= LR_STATE_PENDING;
|
|
+ if (val & GICH_LR_ACTIVE_BIT)
|
|
+ lr_desc.state |= LR_STATE_ACTIVE;
|
|
+ if (val & GICH_LR_EOI)
|
|
+ lr_desc.state |= LR_EOI_INT;
|
|
+
|
|
+ return lr_desc;
|
|
+}
|
|
+
|
|
+static void vgic_v2_set_lr(struct kvm_vcpu *vcpu, int lr,
|
|
+ struct vgic_lr lr_desc)
|
|
+{
|
|
+ u32 lr_val = (lr_desc.source << GICH_LR_PHYSID_CPUID_SHIFT) | lr_desc.irq;
|
|
+
|
|
+ if (lr_desc.state & LR_STATE_PENDING)
|
|
+ lr_val |= GICH_LR_PENDING_BIT;
|
|
+ if (lr_desc.state & LR_STATE_ACTIVE)
|
|
+ lr_val |= GICH_LR_ACTIVE_BIT;
|
|
+ if (lr_desc.state & LR_EOI_INT)
|
|
+ lr_val |= GICH_LR_EOI;
|
|
+
|
|
+ vcpu->arch.vgic_cpu.vgic_v2.vgic_lr[lr] = lr_val;
|
|
+}
|
|
+
|
|
+static void vgic_v2_sync_lr_elrsr(struct kvm_vcpu *vcpu, int lr,
|
|
+ struct vgic_lr lr_desc)
|
|
+{
|
|
+ if (!(lr_desc.state & LR_STATE_MASK))
|
|
+ set_bit(lr, (unsigned long *)vcpu->arch.vgic_cpu.vgic_v2.vgic_elrsr);
|
|
+}
|
|
+
|
|
+static u64 vgic_v2_get_elrsr(const struct kvm_vcpu *vcpu)
|
|
+{
|
|
+ u64 val;
|
|
+
|
|
+ val = vcpu->arch.vgic_cpu.vgic_v2.vgic_elrsr[1];
|
|
+ val <<= 32;
|
|
+ val |= vcpu->arch.vgic_cpu.vgic_v2.vgic_elrsr[0];
|
|
+
|
|
+ return val;
|
|
+}
|
|
+
|
|
+static u64 vgic_v2_get_eisr(const struct kvm_vcpu *vcpu)
|
|
+{
|
|
+ u64 val;
|
|
+
|
|
+ val = vcpu->arch.vgic_cpu.vgic_v2.vgic_eisr[1];
|
|
+ val <<= 32;
|
|
+ val |= vcpu->arch.vgic_cpu.vgic_v2.vgic_eisr[0];
|
|
+
|
|
+ return val;
|
|
+}
|
|
+
|
|
+static u32 vgic_v2_get_interrupt_status(const struct kvm_vcpu *vcpu)
|
|
+{
|
|
+ u32 misr = vcpu->arch.vgic_cpu.vgic_v2.vgic_misr;
|
|
+ u32 ret = 0;
|
|
+
|
|
+ if (misr & GICH_MISR_EOI)
|
|
+ ret |= INT_STATUS_EOI;
|
|
+ if (misr & GICH_MISR_U)
|
|
+ ret |= INT_STATUS_UNDERFLOW;
|
|
+
|
|
+ return ret;
|
|
+}
|
|
+
|
|
+static void vgic_v2_enable_underflow(struct kvm_vcpu *vcpu)
|
|
+{
|
|
+ vcpu->arch.vgic_cpu.vgic_v2.vgic_hcr |= GICH_HCR_UIE;
|
|
+}
|
|
+
|
|
+static void vgic_v2_disable_underflow(struct kvm_vcpu *vcpu)
|
|
+{
|
|
+ vcpu->arch.vgic_cpu.vgic_v2.vgic_hcr &= ~GICH_HCR_UIE;
|
|
+}
|
|
+
|
|
+static void vgic_v2_get_vmcr(struct kvm_vcpu *vcpu, struct vgic_vmcr *vmcrp)
|
|
+{
|
|
+ u32 vmcr = vcpu->arch.vgic_cpu.vgic_v2.vgic_vmcr;
|
|
+
|
|
+ vmcrp->ctlr = (vmcr & GICH_VMCR_CTRL_MASK) >> GICH_VMCR_CTRL_SHIFT;
|
|
+ vmcrp->abpr = (vmcr & GICH_VMCR_ALIAS_BINPOINT_MASK) >> GICH_VMCR_ALIAS_BINPOINT_SHIFT;
|
|
+ vmcrp->bpr = (vmcr & GICH_VMCR_BINPOINT_MASK) >> GICH_VMCR_BINPOINT_SHIFT;
|
|
+ vmcrp->pmr = (vmcr & GICH_VMCR_PRIMASK_MASK) >> GICH_VMCR_PRIMASK_SHIFT;
|
|
+}
|
|
+
|
|
+static void vgic_v2_set_vmcr(struct kvm_vcpu *vcpu, struct vgic_vmcr *vmcrp)
|
|
+{
|
|
+ u32 vmcr;
|
|
+
|
|
+ vmcr = (vmcrp->ctlr << GICH_VMCR_CTRL_SHIFT) & GICH_VMCR_CTRL_MASK;
|
|
+ vmcr |= (vmcrp->abpr << GICH_VMCR_ALIAS_BINPOINT_SHIFT) & GICH_VMCR_ALIAS_BINPOINT_MASK;
|
|
+ vmcr |= (vmcrp->bpr << GICH_VMCR_BINPOINT_SHIFT) & GICH_VMCR_BINPOINT_MASK;
|
|
+ vmcr |= (vmcrp->pmr << GICH_VMCR_PRIMASK_SHIFT) & GICH_VMCR_PRIMASK_MASK;
|
|
+
|
|
+ vcpu->arch.vgic_cpu.vgic_v2.vgic_vmcr = vmcr;
|
|
+}
|
|
+
|
|
+static void vgic_v2_enable(struct kvm_vcpu *vcpu)
|
|
+{
|
|
+ /*
|
|
+ * By forcing VMCR to zero, the GIC will restore the binary
|
|
+ * points to their reset values. Anything else resets to zero
|
|
+ * anyway.
|
|
+ */
|
|
+ vcpu->arch.vgic_cpu.vgic_v2.vgic_vmcr = 0;
|
|
+
|
|
+ /* Get the show on the road... */
|
|
+ vcpu->arch.vgic_cpu.vgic_v2.vgic_hcr = GICH_HCR_EN;
|
|
+}
|
|
+
|
|
+static const struct vgic_ops vgic_v2_ops = {
|
|
+ .get_lr = vgic_v2_get_lr,
|
|
+ .set_lr = vgic_v2_set_lr,
|
|
+ .sync_lr_elrsr = vgic_v2_sync_lr_elrsr,
|
|
+ .get_elrsr = vgic_v2_get_elrsr,
|
|
+ .get_eisr = vgic_v2_get_eisr,
|
|
+ .get_interrupt_status = vgic_v2_get_interrupt_status,
|
|
+ .enable_underflow = vgic_v2_enable_underflow,
|
|
+ .disable_underflow = vgic_v2_disable_underflow,
|
|
+ .get_vmcr = vgic_v2_get_vmcr,
|
|
+ .set_vmcr = vgic_v2_set_vmcr,
|
|
+ .enable = vgic_v2_enable,
|
|
+};
|
|
+
|
|
+static struct vgic_params vgic_v2_params;
|
|
+
|
|
+/**
|
|
+ * vgic_v2_probe - probe for a GICv2 compatible interrupt controller in DT
|
|
+ * @node: pointer to the DT node
|
|
+ * @ops: address of a pointer to the GICv2 operations
|
|
+ * @params: address of a pointer to HW-specific parameters
|
|
+ *
|
|
+ * Returns 0 if a GICv2 has been found, with the low level operations
|
|
+ * in *ops and the HW parameters in *params. Returns an error code
|
|
+ * otherwise.
|
|
+ */
|
|
+int vgic_v2_probe(struct device_node *vgic_node,
|
|
+ const struct vgic_ops **ops,
|
|
+ const struct vgic_params **params)
|
|
+{
|
|
+ int ret;
|
|
+ struct resource vctrl_res;
|
|
+ struct resource vcpu_res;
|
|
+ struct vgic_params *vgic = &vgic_v2_params;
|
|
+
|
|
+ vgic->maint_irq = irq_of_parse_and_map(vgic_node, 0);
|
|
+ if (!vgic->maint_irq) {
|
|
+ kvm_err("error getting vgic maintenance irq from DT\n");
|
|
+ ret = -ENXIO;
|
|
+ goto out;
|
|
+ }
|
|
+
|
|
+ ret = of_address_to_resource(vgic_node, 2, &vctrl_res);
|
|
+ if (ret) {
|
|
+ kvm_err("Cannot obtain GICH resource\n");
|
|
+ goto out;
|
|
+ }
|
|
+
|
|
+ vgic->vctrl_base = of_iomap(vgic_node, 2);
|
|
+ if (!vgic->vctrl_base) {
|
|
+ kvm_err("Cannot ioremap GICH\n");
|
|
+ ret = -ENOMEM;
|
|
+ goto out;
|
|
+ }
|
|
+
|
|
+ vgic->nr_lr = readl_relaxed(vgic->vctrl_base + GICH_VTR);
|
|
+ vgic->nr_lr = (vgic->nr_lr & 0x3f) + 1;
|
|
+
|
|
+ ret = create_hyp_io_mappings(vgic->vctrl_base,
|
|
+ vgic->vctrl_base + resource_size(&vctrl_res),
|
|
+ vctrl_res.start);
|
|
+ if (ret) {
|
|
+ kvm_err("Cannot map VCTRL into hyp\n");
|
|
+ goto out_unmap;
|
|
+ }
|
|
+
|
|
+ if (of_address_to_resource(vgic_node, 3, &vcpu_res)) {
|
|
+ kvm_err("Cannot obtain GICV resource\n");
|
|
+ ret = -ENXIO;
|
|
+ goto out_unmap;
|
|
+ }
|
|
+ vgic->vcpu_base = vcpu_res.start;
|
|
+
|
|
+ kvm_info("%s@%llx IRQ%d\n", vgic_node->name,
|
|
+ vctrl_res.start, vgic->maint_irq);
|
|
+
|
|
+ vgic->type = VGIC_V2;
|
|
+ *ops = &vgic_v2_ops;
|
|
+ *params = vgic;
|
|
+ goto out;
|
|
+
|
|
+out_unmap:
|
|
+ iounmap(vgic->vctrl_base);
|
|
+out:
|
|
+ of_node_put(vgic_node);
|
|
+ return ret;
|
|
+}
|
|
diff --git a/virt/kvm/arm/vgic-v3.c b/virt/kvm/arm/vgic-v3.c
|
|
new file mode 100644
|
|
index 0000000..f01d446
|
|
--- /dev/null
|
|
+++ b/virt/kvm/arm/vgic-v3.c
|
|
@@ -0,0 +1,231 @@
|
|
+/*
|
|
+ * Copyright (C) 2013 ARM Limited, All Rights Reserved.
|
|
+ * Author: Marc Zyngier <marc.zyngier@arm.com>
|
|
+ *
|
|
+ * This program is free software; you can redistribute it and/or modify
|
|
+ * it under the terms of the GNU General Public License version 2 as
|
|
+ * published by the Free Software Foundation.
|
|
+ *
|
|
+ * This program is distributed in the hope that it will be useful,
|
|
+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
+ * GNU General Public License for more details.
|
|
+ *
|
|
+ * You should have received a copy of the GNU General Public License
|
|
+ * along with this program. If not, see <http://www.gnu.org/licenses/>.
|
|
+ */
|
|
+
|
|
+#include <linux/cpu.h>
|
|
+#include <linux/kvm.h>
|
|
+#include <linux/kvm_host.h>
|
|
+#include <linux/interrupt.h>
|
|
+#include <linux/io.h>
|
|
+#include <linux/of.h>
|
|
+#include <linux/of_address.h>
|
|
+#include <linux/of_irq.h>
|
|
+
|
|
+#include <linux/irqchip/arm-gic-v3.h>
|
|
+
|
|
+#include <asm/kvm_emulate.h>
|
|
+#include <asm/kvm_arm.h>
|
|
+#include <asm/kvm_mmu.h>
|
|
+
|
|
+/* These are for GICv2 emulation only */
|
|
+#define GICH_LR_VIRTUALID (0x3ffUL << 0)
|
|
+#define GICH_LR_PHYSID_CPUID_SHIFT (10)
|
|
+#define GICH_LR_PHYSID_CPUID (7UL << GICH_LR_PHYSID_CPUID_SHIFT)
|
|
+
|
|
+/*
|
|
+ * LRs are stored in reverse order in memory. make sure we index them
|
|
+ * correctly.
|
|
+ */
|
|
+#define LR_INDEX(lr) (VGIC_V3_MAX_LRS - 1 - lr)
|
|
+
|
|
+static u32 ich_vtr_el2;
|
|
+
|
|
+static struct vgic_lr vgic_v3_get_lr(const struct kvm_vcpu *vcpu, int lr)
|
|
+{
|
|
+ struct vgic_lr lr_desc;
|
|
+ u64 val = vcpu->arch.vgic_cpu.vgic_v3.vgic_lr[LR_INDEX(lr)];
|
|
+
|
|
+ lr_desc.irq = val & GICH_LR_VIRTUALID;
|
|
+ if (lr_desc.irq <= 15)
|
|
+ lr_desc.source = (val >> GICH_LR_PHYSID_CPUID_SHIFT) & 0x7;
|
|
+ else
|
|
+ lr_desc.source = 0;
|
|
+ lr_desc.state = 0;
|
|
+
|
|
+ if (val & ICH_LR_PENDING_BIT)
|
|
+ lr_desc.state |= LR_STATE_PENDING;
|
|
+ if (val & ICH_LR_ACTIVE_BIT)
|
|
+ lr_desc.state |= LR_STATE_ACTIVE;
|
|
+ if (val & ICH_LR_EOI)
|
|
+ lr_desc.state |= LR_EOI_INT;
|
|
+
|
|
+ return lr_desc;
|
|
+}
|
|
+
|
|
+static void vgic_v3_set_lr(struct kvm_vcpu *vcpu, int lr,
|
|
+ struct vgic_lr lr_desc)
|
|
+{
|
|
+ u64 lr_val = (((u32)lr_desc.source << GICH_LR_PHYSID_CPUID_SHIFT) |
|
|
+ lr_desc.irq);
|
|
+
|
|
+ if (lr_desc.state & LR_STATE_PENDING)
|
|
+ lr_val |= ICH_LR_PENDING_BIT;
|
|
+ if (lr_desc.state & LR_STATE_ACTIVE)
|
|
+ lr_val |= ICH_LR_ACTIVE_BIT;
|
|
+ if (lr_desc.state & LR_EOI_INT)
|
|
+ lr_val |= ICH_LR_EOI;
|
|
+
|
|
+ vcpu->arch.vgic_cpu.vgic_v3.vgic_lr[LR_INDEX(lr)] = lr_val;
|
|
+}
|
|
+
|
|
+static void vgic_v3_sync_lr_elrsr(struct kvm_vcpu *vcpu, int lr,
|
|
+ struct vgic_lr lr_desc)
|
|
+{
|
|
+ if (!(lr_desc.state & LR_STATE_MASK))
|
|
+ vcpu->arch.vgic_cpu.vgic_v3.vgic_elrsr |= (1U << lr);
|
|
+}
|
|
+
|
|
+static u64 vgic_v3_get_elrsr(const struct kvm_vcpu *vcpu)
|
|
+{
|
|
+ return vcpu->arch.vgic_cpu.vgic_v3.vgic_elrsr;
|
|
+}
|
|
+
|
|
+static u64 vgic_v3_get_eisr(const struct kvm_vcpu *vcpu)
|
|
+{
|
|
+ return vcpu->arch.vgic_cpu.vgic_v3.vgic_eisr;
|
|
+}
|
|
+
|
|
+static u32 vgic_v3_get_interrupt_status(const struct kvm_vcpu *vcpu)
|
|
+{
|
|
+ u32 misr = vcpu->arch.vgic_cpu.vgic_v3.vgic_misr;
|
|
+ u32 ret = 0;
|
|
+
|
|
+ if (misr & ICH_MISR_EOI)
|
|
+ ret |= INT_STATUS_EOI;
|
|
+ if (misr & ICH_MISR_U)
|
|
+ ret |= INT_STATUS_UNDERFLOW;
|
|
+
|
|
+ return ret;
|
|
+}
|
|
+
|
|
+static void vgic_v3_get_vmcr(struct kvm_vcpu *vcpu, struct vgic_vmcr *vmcrp)
|
|
+{
|
|
+ u32 vmcr = vcpu->arch.vgic_cpu.vgic_v3.vgic_vmcr;
|
|
+
|
|
+ vmcrp->ctlr = (vmcr & ICH_VMCR_CTLR_MASK) >> ICH_VMCR_CTLR_SHIFT;
|
|
+ vmcrp->abpr = (vmcr & ICH_VMCR_BPR1_MASK) >> ICH_VMCR_BPR1_SHIFT;
|
|
+ vmcrp->bpr = (vmcr & ICH_VMCR_BPR0_MASK) >> ICH_VMCR_BPR0_SHIFT;
|
|
+ vmcrp->pmr = (vmcr & ICH_VMCR_PMR_MASK) >> ICH_VMCR_PMR_SHIFT;
|
|
+}
|
|
+
|
|
+static void vgic_v3_enable_underflow(struct kvm_vcpu *vcpu)
|
|
+{
|
|
+ vcpu->arch.vgic_cpu.vgic_v3.vgic_hcr |= ICH_HCR_UIE;
|
|
+}
|
|
+
|
|
+static void vgic_v3_disable_underflow(struct kvm_vcpu *vcpu)
|
|
+{
|
|
+ vcpu->arch.vgic_cpu.vgic_v3.vgic_hcr &= ~ICH_HCR_UIE;
|
|
+}
|
|
+
|
|
+static void vgic_v3_set_vmcr(struct kvm_vcpu *vcpu, struct vgic_vmcr *vmcrp)
|
|
+{
|
|
+ u32 vmcr;
|
|
+
|
|
+ vmcr = (vmcrp->ctlr << ICH_VMCR_CTLR_SHIFT) & ICH_VMCR_CTLR_MASK;
|
|
+ vmcr |= (vmcrp->abpr << ICH_VMCR_BPR1_SHIFT) & ICH_VMCR_BPR1_MASK;
|
|
+ vmcr |= (vmcrp->bpr << ICH_VMCR_BPR0_SHIFT) & ICH_VMCR_BPR0_MASK;
|
|
+ vmcr |= (vmcrp->pmr << ICH_VMCR_PMR_SHIFT) & ICH_VMCR_PMR_MASK;
|
|
+
|
|
+ vcpu->arch.vgic_cpu.vgic_v3.vgic_vmcr = vmcr;
|
|
+}
|
|
+
|
|
+static void vgic_v3_enable(struct kvm_vcpu *vcpu)
|
|
+{
|
|
+ /*
|
|
+ * By forcing VMCR to zero, the GIC will restore the binary
|
|
+ * points to their reset values. Anything else resets to zero
|
|
+ * anyway.
|
|
+ */
|
|
+ vcpu->arch.vgic_cpu.vgic_v3.vgic_vmcr = 0;
|
|
+
|
|
+ /* Get the show on the road... */
|
|
+ vcpu->arch.vgic_cpu.vgic_v3.vgic_hcr = ICH_HCR_EN;
|
|
+}
|
|
+
|
|
+static const struct vgic_ops vgic_v3_ops = {
|
|
+ .get_lr = vgic_v3_get_lr,
|
|
+ .set_lr = vgic_v3_set_lr,
|
|
+ .sync_lr_elrsr = vgic_v3_sync_lr_elrsr,
|
|
+ .get_elrsr = vgic_v3_get_elrsr,
|
|
+ .get_eisr = vgic_v3_get_eisr,
|
|
+ .get_interrupt_status = vgic_v3_get_interrupt_status,
|
|
+ .enable_underflow = vgic_v3_enable_underflow,
|
|
+ .disable_underflow = vgic_v3_disable_underflow,
|
|
+ .get_vmcr = vgic_v3_get_vmcr,
|
|
+ .set_vmcr = vgic_v3_set_vmcr,
|
|
+ .enable = vgic_v3_enable,
|
|
+};
|
|
+
|
|
+static struct vgic_params vgic_v3_params;
|
|
+
|
|
+/**
|
|
+ * vgic_v3_probe - probe for a GICv3 compatible interrupt controller in DT
|
|
+ * @node: pointer to the DT node
|
|
+ * @ops: address of a pointer to the GICv3 operations
|
|
+ * @params: address of a pointer to HW-specific parameters
|
|
+ *
|
|
+ * Returns 0 if a GICv3 has been found, with the low level operations
|
|
+ * in *ops and the HW parameters in *params. Returns an error code
|
|
+ * otherwise.
|
|
+ */
|
|
+int vgic_v3_probe(struct device_node *vgic_node,
|
|
+ const struct vgic_ops **ops,
|
|
+ const struct vgic_params **params)
|
|
+{
|
|
+ int ret = 0;
|
|
+ u32 gicv_idx;
|
|
+ struct resource vcpu_res;
|
|
+ struct vgic_params *vgic = &vgic_v3_params;
|
|
+
|
|
+ vgic->maint_irq = irq_of_parse_and_map(vgic_node, 0);
|
|
+ if (!vgic->maint_irq) {
|
|
+ kvm_err("error getting vgic maintenance irq from DT\n");
|
|
+ ret = -ENXIO;
|
|
+ goto out;
|
|
+ }
|
|
+
|
|
+ ich_vtr_el2 = kvm_call_hyp(__vgic_v3_get_ich_vtr_el2);
|
|
+
|
|
+ /*
|
|
+ * The ListRegs field is 5 bits, but there is a architectural
|
|
+ * maximum of 16 list registers. Just ignore bit 4...
|
|
+ */
|
|
+ vgic->nr_lr = (ich_vtr_el2 & 0xf) + 1;
|
|
+
|
|
+ if (of_property_read_u32(vgic_node, "#redistributor-regions", &gicv_idx))
|
|
+ gicv_idx = 1;
|
|
+
|
|
+ gicv_idx += 3; /* Also skip GICD, GICC, GICH */
|
|
+ if (of_address_to_resource(vgic_node, gicv_idx, &vcpu_res)) {
|
|
+ kvm_err("Cannot obtain GICV region\n");
|
|
+ ret = -ENXIO;
|
|
+ goto out;
|
|
+ }
|
|
+ vgic->vcpu_base = vcpu_res.start;
|
|
+ vgic->vctrl_base = NULL;
|
|
+ vgic->type = VGIC_V3;
|
|
+
|
|
+ kvm_info("%s@%llx IRQ%d\n", vgic_node->name,
|
|
+ vcpu_res.start, vgic->maint_irq);
|
|
+
|
|
+ *ops = &vgic_v3_ops;
|
|
+ *params = vgic;
|
|
+
|
|
+out:
|
|
+ of_node_put(vgic_node);
|
|
+ return ret;
|
|
+}
|
|
diff --git a/virt/kvm/arm/vgic.c b/virt/kvm/arm/vgic.c
|
|
index 56ff9be..795ab48 100644
|
|
--- a/virt/kvm/arm/vgic.c
|
|
+++ b/virt/kvm/arm/vgic.c
|
|
@@ -76,14 +76,6 @@
|
|
#define IMPLEMENTER_ARM 0x43b
|
|
#define GICC_ARCH_VERSION_V2 0x2
|
|
|
|
-/* Physical address of vgic virtual cpu interface */
|
|
-static phys_addr_t vgic_vcpu_base;
|
|
-
|
|
-/* Virtual control interface base address */
|
|
-static void __iomem *vgic_vctrl_base;
|
|
-
|
|
-static struct device_node *vgic_node;
|
|
-
|
|
#define ACCESS_READ_VALUE (1 << 0)
|
|
#define ACCESS_READ_RAZ (0 << 0)
|
|
#define ACCESS_READ_MASK(x) ((x) & (1 << 0))
|
|
@@ -94,12 +86,17 @@ static struct device_node *vgic_node;
|
|
#define ACCESS_WRITE_MASK(x) ((x) & (3 << 1))
|
|
|
|
static void vgic_retire_disabled_irqs(struct kvm_vcpu *vcpu);
|
|
+static void vgic_retire_lr(int lr_nr, int irq, struct kvm_vcpu *vcpu);
|
|
static void vgic_update_state(struct kvm *kvm);
|
|
static void vgic_kick_vcpus(struct kvm *kvm);
|
|
static void vgic_dispatch_sgi(struct kvm_vcpu *vcpu, u32 reg);
|
|
-static u32 vgic_nr_lr;
|
|
+static struct vgic_lr vgic_get_lr(const struct kvm_vcpu *vcpu, int lr);
|
|
+static void vgic_set_lr(struct kvm_vcpu *vcpu, int lr, struct vgic_lr lr_desc);
|
|
+static void vgic_get_vmcr(struct kvm_vcpu *vcpu, struct vgic_vmcr *vmcr);
|
|
+static void vgic_set_vmcr(struct kvm_vcpu *vcpu, struct vgic_vmcr *vmcr);
|
|
|
|
-static unsigned int vgic_maint_irq;
|
|
+static const struct vgic_ops *vgic_ops;
|
|
+static const struct vgic_params *vgic;
|
|
|
|
static u32 *vgic_bitmap_get_reg(struct vgic_bitmap *x,
|
|
int cpuid, u32 offset)
|
|
@@ -593,18 +590,6 @@ static bool handle_mmio_sgi_reg(struct kvm_vcpu *vcpu,
|
|
return false;
|
|
}
|
|
|
|
-#define LR_CPUID(lr) \
|
|
- (((lr) & GICH_LR_PHYSID_CPUID) >> GICH_LR_PHYSID_CPUID_SHIFT)
|
|
-#define LR_IRQID(lr) \
|
|
- ((lr) & GICH_LR_VIRTUALID)
|
|
-
|
|
-static void vgic_retire_lr(int lr_nr, int irq, struct vgic_cpu *vgic_cpu)
|
|
-{
|
|
- clear_bit(lr_nr, vgic_cpu->lr_used);
|
|
- vgic_cpu->vgic_lr[lr_nr] &= ~GICH_LR_STATE;
|
|
- vgic_cpu->vgic_irq_lr_map[irq] = LR_EMPTY;
|
|
-}
|
|
-
|
|
/**
|
|
* vgic_unqueue_irqs - move pending IRQs from LRs to the distributor
|
|
* @vgic_cpu: Pointer to the vgic_cpu struct holding the LRs
|
|
@@ -622,13 +607,10 @@ static void vgic_unqueue_irqs(struct kvm_vcpu *vcpu)
|
|
struct vgic_dist *dist = &vcpu->kvm->arch.vgic;
|
|
struct vgic_cpu *vgic_cpu = &vcpu->arch.vgic_cpu;
|
|
int vcpu_id = vcpu->vcpu_id;
|
|
- int i, irq, source_cpu;
|
|
- u32 *lr;
|
|
+ int i;
|
|
|
|
for_each_set_bit(i, vgic_cpu->lr_used, vgic_cpu->nr_lr) {
|
|
- lr = &vgic_cpu->vgic_lr[i];
|
|
- irq = LR_IRQID(*lr);
|
|
- source_cpu = LR_CPUID(*lr);
|
|
+ struct vgic_lr lr = vgic_get_lr(vcpu, i);
|
|
|
|
/*
|
|
* There are three options for the state bits:
|
|
@@ -640,7 +622,7 @@ static void vgic_unqueue_irqs(struct kvm_vcpu *vcpu)
|
|
* If the LR holds only an active interrupt (not pending) then
|
|
* just leave it alone.
|
|
*/
|
|
- if ((*lr & GICH_LR_STATE) == GICH_LR_ACTIVE_BIT)
|
|
+ if ((lr.state & LR_STATE_MASK) == LR_STATE_ACTIVE)
|
|
continue;
|
|
|
|
/*
|
|
@@ -649,18 +631,19 @@ static void vgic_unqueue_irqs(struct kvm_vcpu *vcpu)
|
|
* is fine, then we are only setting a few bits that were
|
|
* already set.
|
|
*/
|
|
- vgic_dist_irq_set(vcpu, irq);
|
|
- if (irq < VGIC_NR_SGIS)
|
|
- dist->irq_sgi_sources[vcpu_id][irq] |= 1 << source_cpu;
|
|
- *lr &= ~GICH_LR_PENDING_BIT;
|
|
+ vgic_dist_irq_set(vcpu, lr.irq);
|
|
+ if (lr.irq < VGIC_NR_SGIS)
|
|
+ dist->irq_sgi_sources[vcpu_id][lr.irq] |= 1 << lr.source;
|
|
+ lr.state &= ~LR_STATE_PENDING;
|
|
+ vgic_set_lr(vcpu, i, lr);
|
|
|
|
/*
|
|
* If there's no state left on the LR (it could still be
|
|
* active), then the LR does not hold any useful info and can
|
|
* be marked as free for other use.
|
|
*/
|
|
- if (!(*lr & GICH_LR_STATE))
|
|
- vgic_retire_lr(i, irq, vgic_cpu);
|
|
+ if (!(lr.state & LR_STATE_MASK))
|
|
+ vgic_retire_lr(i, lr.irq, vcpu);
|
|
|
|
/* Finally update the VGIC state. */
|
|
vgic_update_state(vcpu->kvm);
|
|
@@ -989,8 +972,73 @@ static void vgic_update_state(struct kvm *kvm)
|
|
}
|
|
}
|
|
|
|
-#define MK_LR_PEND(src, irq) \
|
|
- (GICH_LR_PENDING_BIT | ((src) << GICH_LR_PHYSID_CPUID_SHIFT) | (irq))
|
|
+static struct vgic_lr vgic_get_lr(const struct kvm_vcpu *vcpu, int lr)
|
|
+{
|
|
+ return vgic_ops->get_lr(vcpu, lr);
|
|
+}
|
|
+
|
|
+static void vgic_set_lr(struct kvm_vcpu *vcpu, int lr,
|
|
+ struct vgic_lr vlr)
|
|
+{
|
|
+ vgic_ops->set_lr(vcpu, lr, vlr);
|
|
+}
|
|
+
|
|
+static void vgic_sync_lr_elrsr(struct kvm_vcpu *vcpu, int lr,
|
|
+ struct vgic_lr vlr)
|
|
+{
|
|
+ vgic_ops->sync_lr_elrsr(vcpu, lr, vlr);
|
|
+}
|
|
+
|
|
+static inline u64 vgic_get_elrsr(struct kvm_vcpu *vcpu)
|
|
+{
|
|
+ return vgic_ops->get_elrsr(vcpu);
|
|
+}
|
|
+
|
|
+static inline u64 vgic_get_eisr(struct kvm_vcpu *vcpu)
|
|
+{
|
|
+ return vgic_ops->get_eisr(vcpu);
|
|
+}
|
|
+
|
|
+static inline u32 vgic_get_interrupt_status(struct kvm_vcpu *vcpu)
|
|
+{
|
|
+ return vgic_ops->get_interrupt_status(vcpu);
|
|
+}
|
|
+
|
|
+static inline void vgic_enable_underflow(struct kvm_vcpu *vcpu)
|
|
+{
|
|
+ vgic_ops->enable_underflow(vcpu);
|
|
+}
|
|
+
|
|
+static inline void vgic_disable_underflow(struct kvm_vcpu *vcpu)
|
|
+{
|
|
+ vgic_ops->disable_underflow(vcpu);
|
|
+}
|
|
+
|
|
+static inline void vgic_get_vmcr(struct kvm_vcpu *vcpu, struct vgic_vmcr *vmcr)
|
|
+{
|
|
+ vgic_ops->get_vmcr(vcpu, vmcr);
|
|
+}
|
|
+
|
|
+static void vgic_set_vmcr(struct kvm_vcpu *vcpu, struct vgic_vmcr *vmcr)
|
|
+{
|
|
+ vgic_ops->set_vmcr(vcpu, vmcr);
|
|
+}
|
|
+
|
|
+static inline void vgic_enable(struct kvm_vcpu *vcpu)
|
|
+{
|
|
+ vgic_ops->enable(vcpu);
|
|
+}
|
|
+
|
|
+static void vgic_retire_lr(int lr_nr, int irq, struct kvm_vcpu *vcpu)
|
|
+{
|
|
+ struct vgic_cpu *vgic_cpu = &vcpu->arch.vgic_cpu;
|
|
+ struct vgic_lr vlr = vgic_get_lr(vcpu, lr_nr);
|
|
+
|
|
+ vlr.state = 0;
|
|
+ vgic_set_lr(vcpu, lr_nr, vlr);
|
|
+ clear_bit(lr_nr, vgic_cpu->lr_used);
|
|
+ vgic_cpu->vgic_irq_lr_map[irq] = LR_EMPTY;
|
|
+}
|
|
|
|
/*
|
|
* An interrupt may have been disabled after being made pending on the
|
|
@@ -1006,13 +1054,13 @@ static void vgic_retire_disabled_irqs(struct kvm_vcpu *vcpu)
|
|
struct vgic_cpu *vgic_cpu = &vcpu->arch.vgic_cpu;
|
|
int lr;
|
|
|
|
- for_each_set_bit(lr, vgic_cpu->lr_used, vgic_cpu->nr_lr) {
|
|
- int irq = vgic_cpu->vgic_lr[lr] & GICH_LR_VIRTUALID;
|
|
+ for_each_set_bit(lr, vgic_cpu->lr_used, vgic->nr_lr) {
|
|
+ struct vgic_lr vlr = vgic_get_lr(vcpu, lr);
|
|
|
|
- if (!vgic_irq_is_enabled(vcpu, irq)) {
|
|
- vgic_retire_lr(lr, irq, vgic_cpu);
|
|
- if (vgic_irq_is_active(vcpu, irq))
|
|
- vgic_irq_clear_active(vcpu, irq);
|
|
+ if (!vgic_irq_is_enabled(vcpu, vlr.irq)) {
|
|
+ vgic_retire_lr(lr, vlr.irq, vcpu);
|
|
+ if (vgic_irq_is_active(vcpu, vlr.irq))
|
|
+ vgic_irq_clear_active(vcpu, vlr.irq);
|
|
}
|
|
}
|
|
}
|
|
@@ -1024,6 +1072,7 @@ static void vgic_retire_disabled_irqs(struct kvm_vcpu *vcpu)
|
|
static bool vgic_queue_irq(struct kvm_vcpu *vcpu, u8 sgi_source_id, int irq)
|
|
{
|
|
struct vgic_cpu *vgic_cpu = &vcpu->arch.vgic_cpu;
|
|
+ struct vgic_lr vlr;
|
|
int lr;
|
|
|
|
/* Sanitize the input... */
|
|
@@ -1036,28 +1085,34 @@ static bool vgic_queue_irq(struct kvm_vcpu *vcpu, u8 sgi_source_id, int irq)
|
|
lr = vgic_cpu->vgic_irq_lr_map[irq];
|
|
|
|
/* Do we have an active interrupt for the same CPUID? */
|
|
- if (lr != LR_EMPTY &&
|
|
- (LR_CPUID(vgic_cpu->vgic_lr[lr]) == sgi_source_id)) {
|
|
- kvm_debug("LR%d piggyback for IRQ%d %x\n",
|
|
- lr, irq, vgic_cpu->vgic_lr[lr]);
|
|
- BUG_ON(!test_bit(lr, vgic_cpu->lr_used));
|
|
- vgic_cpu->vgic_lr[lr] |= GICH_LR_PENDING_BIT;
|
|
- return true;
|
|
+ if (lr != LR_EMPTY) {
|
|
+ vlr = vgic_get_lr(vcpu, lr);
|
|
+ if (vlr.source == sgi_source_id) {
|
|
+ kvm_debug("LR%d piggyback for IRQ%d\n", lr, vlr.irq);
|
|
+ BUG_ON(!test_bit(lr, vgic_cpu->lr_used));
|
|
+ vlr.state |= LR_STATE_PENDING;
|
|
+ vgic_set_lr(vcpu, lr, vlr);
|
|
+ return true;
|
|
+ }
|
|
}
|
|
|
|
/* Try to use another LR for this interrupt */
|
|
lr = find_first_zero_bit((unsigned long *)vgic_cpu->lr_used,
|
|
- vgic_cpu->nr_lr);
|
|
- if (lr >= vgic_cpu->nr_lr)
|
|
+ vgic->nr_lr);
|
|
+ if (lr >= vgic->nr_lr)
|
|
return false;
|
|
|
|
kvm_debug("LR%d allocated for IRQ%d %x\n", lr, irq, sgi_source_id);
|
|
- vgic_cpu->vgic_lr[lr] = MK_LR_PEND(sgi_source_id, irq);
|
|
vgic_cpu->vgic_irq_lr_map[irq] = lr;
|
|
set_bit(lr, vgic_cpu->lr_used);
|
|
|
|
+ vlr.irq = irq;
|
|
+ vlr.source = sgi_source_id;
|
|
+ vlr.state = LR_STATE_PENDING;
|
|
if (!vgic_irq_is_edge(vcpu, irq))
|
|
- vgic_cpu->vgic_lr[lr] |= GICH_LR_EOI;
|
|
+ vlr.state |= LR_EOI_INT;
|
|
+
|
|
+ vgic_set_lr(vcpu, lr, vlr);
|
|
|
|
return true;
|
|
}
|
|
@@ -1155,9 +1210,9 @@ static void __kvm_vgic_flush_hwstate(struct kvm_vcpu *vcpu)
|
|
|
|
epilog:
|
|
if (overflow) {
|
|
- vgic_cpu->vgic_hcr |= GICH_HCR_UIE;
|
|
+ vgic_enable_underflow(vcpu);
|
|
} else {
|
|
- vgic_cpu->vgic_hcr &= ~GICH_HCR_UIE;
|
|
+ vgic_disable_underflow(vcpu);
|
|
/*
|
|
* We're about to run this VCPU, and we've consumed
|
|
* everything the distributor had in store for
|
|
@@ -1170,44 +1225,46 @@ epilog:
|
|
|
|
static bool vgic_process_maintenance(struct kvm_vcpu *vcpu)
|
|
{
|
|
- struct vgic_cpu *vgic_cpu = &vcpu->arch.vgic_cpu;
|
|
+ u32 status = vgic_get_interrupt_status(vcpu);
|
|
bool level_pending = false;
|
|
|
|
- kvm_debug("MISR = %08x\n", vgic_cpu->vgic_misr);
|
|
+ kvm_debug("STATUS = %08x\n", status);
|
|
|
|
- if (vgic_cpu->vgic_misr & GICH_MISR_EOI) {
|
|
+ if (status & INT_STATUS_EOI) {
|
|
/*
|
|
* Some level interrupts have been EOIed. Clear their
|
|
* active bit.
|
|
*/
|
|
- int lr, irq;
|
|
+ u64 eisr = vgic_get_eisr(vcpu);
|
|
+ unsigned long *eisr_ptr = (unsigned long *)&eisr;
|
|
+ int lr;
|
|
|
|
- for_each_set_bit(lr, (unsigned long *)vgic_cpu->vgic_eisr,
|
|
- vgic_cpu->nr_lr) {
|
|
- irq = vgic_cpu->vgic_lr[lr] & GICH_LR_VIRTUALID;
|
|
+ for_each_set_bit(lr, eisr_ptr, vgic->nr_lr) {
|
|
+ struct vgic_lr vlr = vgic_get_lr(vcpu, lr);
|
|
|
|
- vgic_irq_clear_active(vcpu, irq);
|
|
- vgic_cpu->vgic_lr[lr] &= ~GICH_LR_EOI;
|
|
+ vgic_irq_clear_active(vcpu, vlr.irq);
|
|
+ WARN_ON(vlr.state & LR_STATE_MASK);
|
|
+ vlr.state = 0;
|
|
+ vgic_set_lr(vcpu, lr, vlr);
|
|
|
|
/* Any additional pending interrupt? */
|
|
- if (vgic_dist_irq_is_pending(vcpu, irq)) {
|
|
- vgic_cpu_irq_set(vcpu, irq);
|
|
+ if (vgic_dist_irq_is_pending(vcpu, vlr.irq)) {
|
|
+ vgic_cpu_irq_set(vcpu, vlr.irq);
|
|
level_pending = true;
|
|
} else {
|
|
- vgic_cpu_irq_clear(vcpu, irq);
|
|
+ vgic_cpu_irq_clear(vcpu, vlr.irq);
|
|
}
|
|
|
|
/*
|
|
* Despite being EOIed, the LR may not have
|
|
* been marked as empty.
|
|
*/
|
|
- set_bit(lr, (unsigned long *)vgic_cpu->vgic_elrsr);
|
|
- vgic_cpu->vgic_lr[lr] &= ~GICH_LR_ACTIVE_BIT;
|
|
+ vgic_sync_lr_elrsr(vcpu, lr, vlr);
|
|
}
|
|
}
|
|
|
|
- if (vgic_cpu->vgic_misr & GICH_MISR_U)
|
|
- vgic_cpu->vgic_hcr &= ~GICH_HCR_UIE;
|
|
+ if (status & INT_STATUS_UNDERFLOW)
|
|
+ vgic_disable_underflow(vcpu);
|
|
|
|
return level_pending;
|
|
}
|
|
@@ -1220,29 +1277,31 @@ static void __kvm_vgic_sync_hwstate(struct kvm_vcpu *vcpu)
|
|
{
|
|
struct vgic_cpu *vgic_cpu = &vcpu->arch.vgic_cpu;
|
|
struct vgic_dist *dist = &vcpu->kvm->arch.vgic;
|
|
+ u64 elrsr;
|
|
+ unsigned long *elrsr_ptr;
|
|
int lr, pending;
|
|
bool level_pending;
|
|
|
|
level_pending = vgic_process_maintenance(vcpu);
|
|
+ elrsr = vgic_get_elrsr(vcpu);
|
|
+ elrsr_ptr = (unsigned long *)&elrsr;
|
|
|
|
/* Clear mappings for empty LRs */
|
|
- for_each_set_bit(lr, (unsigned long *)vgic_cpu->vgic_elrsr,
|
|
- vgic_cpu->nr_lr) {
|
|
- int irq;
|
|
+ for_each_set_bit(lr, elrsr_ptr, vgic->nr_lr) {
|
|
+ struct vgic_lr vlr;
|
|
|
|
if (!test_and_clear_bit(lr, vgic_cpu->lr_used))
|
|
continue;
|
|
|
|
- irq = vgic_cpu->vgic_lr[lr] & GICH_LR_VIRTUALID;
|
|
+ vlr = vgic_get_lr(vcpu, lr);
|
|
|
|
- BUG_ON(irq >= VGIC_NR_IRQS);
|
|
- vgic_cpu->vgic_irq_lr_map[irq] = LR_EMPTY;
|
|
+ BUG_ON(vlr.irq >= VGIC_NR_IRQS);
|
|
+ vgic_cpu->vgic_irq_lr_map[vlr.irq] = LR_EMPTY;
|
|
}
|
|
|
|
/* Check if we still have something up our sleeve... */
|
|
- pending = find_first_zero_bit((unsigned long *)vgic_cpu->vgic_elrsr,
|
|
- vgic_cpu->nr_lr);
|
|
- if (level_pending || pending < vgic_cpu->nr_lr)
|
|
+ pending = find_first_zero_bit(elrsr_ptr, vgic->nr_lr);
|
|
+ if (level_pending || pending < vgic->nr_lr)
|
|
set_bit(vcpu->vcpu_id, &dist->irq_pending_on_cpu);
|
|
}
|
|
|
|
@@ -1432,21 +1491,20 @@ int kvm_vgic_vcpu_init(struct kvm_vcpu *vcpu)
|
|
}
|
|
|
|
/*
|
|
- * By forcing VMCR to zero, the GIC will restore the binary
|
|
- * points to their reset values. Anything else resets to zero
|
|
- * anyway.
|
|
+ * Store the number of LRs per vcpu, so we don't have to go
|
|
+ * all the way to the distributor structure to find out. Only
|
|
+ * assembly code should use this one.
|
|
*/
|
|
- vgic_cpu->vgic_vmcr = 0;
|
|
+ vgic_cpu->nr_lr = vgic->nr_lr;
|
|
|
|
- vgic_cpu->nr_lr = vgic_nr_lr;
|
|
- vgic_cpu->vgic_hcr = GICH_HCR_EN; /* Get the show on the road... */
|
|
+ vgic_enable(vcpu);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void vgic_init_maintenance_interrupt(void *info)
|
|
{
|
|
- enable_percpu_irq(vgic_maint_irq, 0);
|
|
+ enable_percpu_irq(vgic->maint_irq, 0);
|
|
}
|
|
|
|
static int vgic_cpu_notify(struct notifier_block *self,
|
|
@@ -1459,7 +1517,7 @@ static int vgic_cpu_notify(struct notifier_block *self,
|
|
break;
|
|
case CPU_DYING:
|
|
case CPU_DYING_FROZEN:
|
|
- disable_percpu_irq(vgic_maint_irq);
|
|
+ disable_percpu_irq(vgic->maint_irq);
|
|
break;
|
|
}
|
|
|
|
@@ -1470,30 +1528,37 @@ static struct notifier_block vgic_cpu_nb = {
|
|
.notifier_call = vgic_cpu_notify,
|
|
};
|
|
|
|
+static const struct of_device_id vgic_ids[] = {
|
|
+ { .compatible = "arm,cortex-a15-gic", .data = vgic_v2_probe, },
|
|
+ { .compatible = "arm,gic-v3", .data = vgic_v3_probe, },
|
|
+ {},
|
|
+};
|
|
+
|
|
int kvm_vgic_hyp_init(void)
|
|
{
|
|
+ const struct of_device_id *matched_id;
|
|
+ int (*vgic_probe)(struct device_node *,const struct vgic_ops **,
|
|
+ const struct vgic_params **);
|
|
+ struct device_node *vgic_node;
|
|
int ret;
|
|
- struct resource vctrl_res;
|
|
- struct resource vcpu_res;
|
|
|
|
- vgic_node = of_find_compatible_node(NULL, NULL, "arm,cortex-a15-gic");
|
|
+ vgic_node = of_find_matching_node_and_match(NULL,
|
|
+ vgic_ids, &matched_id);
|
|
if (!vgic_node) {
|
|
- kvm_err("error: no compatible vgic node in DT\n");
|
|
+ kvm_err("error: no compatible GIC node found\n");
|
|
return -ENODEV;
|
|
}
|
|
|
|
- vgic_maint_irq = irq_of_parse_and_map(vgic_node, 0);
|
|
- if (!vgic_maint_irq) {
|
|
- kvm_err("error getting vgic maintenance irq from DT\n");
|
|
- ret = -ENXIO;
|
|
- goto out;
|
|
- }
|
|
+ vgic_probe = matched_id->data;
|
|
+ ret = vgic_probe(vgic_node, &vgic_ops, &vgic);
|
|
+ if (ret)
|
|
+ return ret;
|
|
|
|
- ret = request_percpu_irq(vgic_maint_irq, vgic_maintenance_handler,
|
|
+ ret = request_percpu_irq(vgic->maint_irq, vgic_maintenance_handler,
|
|
"vgic", kvm_get_running_vcpus());
|
|
if (ret) {
|
|
- kvm_err("Cannot register interrupt %d\n", vgic_maint_irq);
|
|
- goto out;
|
|
+ kvm_err("Cannot register interrupt %d\n", vgic->maint_irq);
|
|
+ return ret;
|
|
}
|
|
|
|
ret = __register_cpu_notifier(&vgic_cpu_nb);
|
|
@@ -1502,49 +1567,15 @@ int kvm_vgic_hyp_init(void)
|
|
goto out_free_irq;
|
|
}
|
|
|
|
- ret = of_address_to_resource(vgic_node, 2, &vctrl_res);
|
|
- if (ret) {
|
|
- kvm_err("Cannot obtain VCTRL resource\n");
|
|
- goto out_free_irq;
|
|
- }
|
|
-
|
|
- vgic_vctrl_base = of_iomap(vgic_node, 2);
|
|
- if (!vgic_vctrl_base) {
|
|
- kvm_err("Cannot ioremap VCTRL\n");
|
|
- ret = -ENOMEM;
|
|
- goto out_free_irq;
|
|
- }
|
|
-
|
|
- vgic_nr_lr = readl_relaxed(vgic_vctrl_base + GICH_VTR);
|
|
- vgic_nr_lr = (vgic_nr_lr & 0x3f) + 1;
|
|
-
|
|
- ret = create_hyp_io_mappings(vgic_vctrl_base,
|
|
- vgic_vctrl_base + resource_size(&vctrl_res),
|
|
- vctrl_res.start);
|
|
- if (ret) {
|
|
- kvm_err("Cannot map VCTRL into hyp\n");
|
|
- goto out_unmap;
|
|
- }
|
|
-
|
|
- kvm_info("%s@%llx IRQ%d\n", vgic_node->name,
|
|
- vctrl_res.start, vgic_maint_irq);
|
|
on_each_cpu(vgic_init_maintenance_interrupt, NULL, 1);
|
|
|
|
- if (of_address_to_resource(vgic_node, 3, &vcpu_res)) {
|
|
- kvm_err("Cannot obtain VCPU resource\n");
|
|
- ret = -ENXIO;
|
|
- goto out_unmap;
|
|
- }
|
|
- vgic_vcpu_base = vcpu_res.start;
|
|
+ /* Callback into for arch code for setup */
|
|
+ vgic_arch_setup(vgic);
|
|
|
|
- goto out;
|
|
+ return 0;
|
|
|
|
-out_unmap:
|
|
- iounmap(vgic_vctrl_base);
|
|
out_free_irq:
|
|
- free_percpu_irq(vgic_maint_irq, kvm_get_running_vcpus());
|
|
-out:
|
|
- of_node_put(vgic_node);
|
|
+ free_percpu_irq(vgic->maint_irq, kvm_get_running_vcpus());
|
|
return ret;
|
|
}
|
|
|
|
@@ -1577,7 +1608,7 @@ int kvm_vgic_init(struct kvm *kvm)
|
|
}
|
|
|
|
ret = kvm_phys_addr_ioremap(kvm, kvm->arch.vgic.vgic_cpu_base,
|
|
- vgic_vcpu_base, KVM_VGIC_V2_CPU_SIZE);
|
|
+ vgic->vcpu_base, KVM_VGIC_V2_CPU_SIZE);
|
|
if (ret) {
|
|
kvm_err("Unable to remap VGIC CPU to VCPU\n");
|
|
goto out;
|
|
@@ -1623,7 +1654,8 @@ int kvm_vgic_create(struct kvm *kvm)
|
|
}
|
|
|
|
spin_lock_init(&kvm->arch.vgic.lock);
|
|
- kvm->arch.vgic.vctrl_base = vgic_vctrl_base;
|
|
+ kvm->arch.vgic.in_kernel = true;
|
|
+ kvm->arch.vgic.vctrl_base = vgic->vctrl_base;
|
|
kvm->arch.vgic.vgic_dist_base = VGIC_ADDR_UNDEF;
|
|
kvm->arch.vgic.vgic_cpu_base = VGIC_ADDR_UNDEF;
|
|
|
|
@@ -1722,39 +1754,40 @@ int kvm_vgic_addr(struct kvm *kvm, unsigned long type, u64 *addr, bool write)
|
|
static bool handle_cpu_mmio_misc(struct kvm_vcpu *vcpu,
|
|
struct kvm_exit_mmio *mmio, phys_addr_t offset)
|
|
{
|
|
- struct vgic_cpu *vgic_cpu = &vcpu->arch.vgic_cpu;
|
|
- u32 reg, mask = 0, shift = 0;
|
|
bool updated = false;
|
|
+ struct vgic_vmcr vmcr;
|
|
+ u32 *vmcr_field;
|
|
+ u32 reg;
|
|
+
|
|
+ vgic_get_vmcr(vcpu, &vmcr);
|
|
|
|
switch (offset & ~0x3) {
|
|
case GIC_CPU_CTRL:
|
|
- mask = GICH_VMCR_CTRL_MASK;
|
|
- shift = GICH_VMCR_CTRL_SHIFT;
|
|
+ vmcr_field = &vmcr.ctlr;
|
|
break;
|
|
case GIC_CPU_PRIMASK:
|
|
- mask = GICH_VMCR_PRIMASK_MASK;
|
|
- shift = GICH_VMCR_PRIMASK_SHIFT;
|
|
+ vmcr_field = &vmcr.pmr;
|
|
break;
|
|
case GIC_CPU_BINPOINT:
|
|
- mask = GICH_VMCR_BINPOINT_MASK;
|
|
- shift = GICH_VMCR_BINPOINT_SHIFT;
|
|
+ vmcr_field = &vmcr.bpr;
|
|
break;
|
|
case GIC_CPU_ALIAS_BINPOINT:
|
|
- mask = GICH_VMCR_ALIAS_BINPOINT_MASK;
|
|
- shift = GICH_VMCR_ALIAS_BINPOINT_SHIFT;
|
|
+ vmcr_field = &vmcr.abpr;
|
|
break;
|
|
+ default:
|
|
+ BUG();
|
|
}
|
|
|
|
if (!mmio->is_write) {
|
|
- reg = (vgic_cpu->vgic_vmcr & mask) >> shift;
|
|
+ reg = *vmcr_field;
|
|
mmio_data_write(mmio, ~0, reg);
|
|
} else {
|
|
reg = mmio_data_read(mmio, ~0);
|
|
- reg = (reg << shift) & mask;
|
|
- if (reg != (vgic_cpu->vgic_vmcr & mask))
|
|
+ if (reg != *vmcr_field) {
|
|
+ *vmcr_field = reg;
|
|
+ vgic_set_vmcr(vcpu, &vmcr);
|
|
updated = true;
|
|
- vgic_cpu->vgic_vmcr &= ~mask;
|
|
- vgic_cpu->vgic_vmcr |= reg;
|
|
+ }
|
|
}
|
|
return updated;
|
|
}
|