1. 03 6月, 2015 1 次提交
  2. 01 4月, 2015 1 次提交
    • Y
      powerpc/mpc85xx: call k(un)map_atomic rather than k(un)map · 88b7936d
      Yanjiang Jin 提交于
      The k(un)map function may be called in atomic context in the
      function map_and_flush(), so use k(un)map_atomic to replace it,
      else we would get the below warning during kdump:
      
      BUG: sleeping function called from invalid context at include/linux/highmem.h:58
      in_atomic(): 1, irqs_disabled(): 1, pid: 736, name: sh
      INFO: lockdep is turned off.
      irq event stamp: 0
      hardirqs last  enabled at (0): [<          (null)>]           (null)
      hardirqs last disabled at (0): [<c000000000066d1c>] .copy_process.part.44+0x50c/0x1360
      softirqs last  enabled at (0): [<c000000000066d1c>] .copy_process.part.44+0x50c/0x1360
      softirqs last disabled at (0): [<          (null)>]           (null)
      CPU: 1 PID: 736 Comm: sh Tainted: G      D W    3.10.62-ltsi-WR6.0.0.0_standard #2
      Call Trace:
      [c0000000f47cf120] [c00000000000b150] .show_stack+0x170/0x290 (unreliable)
      [c0000000f47cf210] [c000000000b71334] .dump_stack+0x28/0x3c
      [c0000000f47cf280] [c0000000000bb5d8] .__might_sleep+0x1a8/0x270
      [c0000000f47cf310] [c0000000000440cc] .map_and_flush+0x4c/0xc0
      [c0000000f47cf390] [c0000000000441cc] .mpc85xx_smp_machine_kexec+0x8c/0xec0
      [c0000000f47cf420] [c00000000002ae00] .machine_kexec+0x60/0x90
      [c0000000f47cf4b0] [c00000000010957c] .crash_kexec+0x8c/0x100
      [c0000000f47cf6a0] [c000000000015df8] .die+0x348/0x450
      [c0000000f47cf740] [c00000000002f3a0] .bad_page_fault+0xe0/0x130
      [c0000000f47cf7c0] [c00000000001f3e4] storage_fault_common+0x40/0x44
      Signed-off-by: NYanjiang Jin <yanjiang.jin@windriver.com>
      [scottwood@freescale.com: fix subject line]
      Signed-off-by: NScott Wood <scottwood@freescale.com>
      88b7936d
  3. 30 7月, 2014 1 次提交
    • A
      powerpc/e6500: Add support for hardware threads · e16c8765
      Andy Fleming 提交于
      The general idea is that each core will release all of its
      threads into the secondary thread startup code, which will
      eventually wait in the secondary core holding area, for the
      appropriate bit in the PACA to be set. The kick_cpu function
      pointer will set that bit in the PACA, and thus "release"
      the core/thread to boot. We also need to do a few things that
      U-Boot normally does for CPUs (like enable branch prediction).
      Signed-off-by: NAndy Fleming <afleming@freescale.com>
      [scottwood@freescale.com: various changes, including only enabling
       threads if Linux wants to kick them]
      Signed-off-by: NScott Wood <scottwood@freescale.com>
      e16c8765
  4. 23 4月, 2014 1 次提交
  5. 08 1月, 2014 1 次提交
  6. 11 11月, 2013 1 次提交
    • R
      powerpc: add missing explicit OF includes for ppc · c11eede6
      Rob Herring 提交于
      Commit b5b4bb3f (of: only include prom.h on sparc) removed implicit
      includes of of_*.h headers by powerpc's prom.h. Some components were
      missed in initial clean-up patch, so add the necessary includes to fix
      powerpc builds.
      Signed-off-by: NRob Herring <rob.herring@calxeda.com>
      Cc: Benjamin Herrenschmidt <benh@kernel.crashing.org>
      Cc: Paul Mackerras <paulus@samba.org>
      Cc: Tejun Heo <tj@kernel.org>
      Cc: Matt Mackall <mpm@selenic.com>
      Cc: Herbert Xu <herbert@gondor.apana.org.au>
      Cc: "David S. Miller" <davem@davemloft.net>
      Cc: Vinod Koul <vinod.koul@intel.com>
      Cc: Dan Williams <dan.j.williams@intel.com>
      Cc: linuxppc-dev@lists.ozlabs.org
      Cc: linux-ide@vger.kernel.org
      Cc: linux-crypto@vger.kernel.org
      c11eede6
  7. 21 8月, 2013 2 次提交
    • S
      powerpc: Convert some mftb/mftbu into mfspr · beb2dc0a
      Scott Wood 提交于
      Some CPUs (such as e500v1/v2) don't implement mftb and will take a
      trap.  mfspr should work on everything that has a timebase, and is the
      preferred instruction according to ISA v2.06.
      
      Currently we get away with mftb on 85xx because the assembler converts
      it to mfspr due to -Wa,-me500.  However, that flag has other effects
      that are undesireable for certain targets (e.g.  lwsync is converted to
      sync), and is hostile to multiplatform kernels.  Thus we would like to
      stop setting it for all e500-family builds.
      
      mftb/mftbu instances which are in 85xx code or common code are
      converted.  Instances which will never run on 85xx are left alone.
      Signed-off-by: NScott Wood <scottwood@freescale.com>
      beb2dc0a
    • S
      powerpc/fsl-booke: Work around erratum A-006958 · d52459ca
      Scott Wood 提交于
      Erratum A-006598 says that 64-bit mftb is not atomic -- it's subject
      to a similar race condition as doing mftbu/mftbl on 32-bit.  The lower
      half of timebase is updated before the upper half; thus, we can share
      the workaround for a similar bug on Cell.  This workaround involves
      looping if the lower half of timebase is zero, thus avoiding the need
      for a scratch register (other than CR0).  This workaround must be
      avoided when the timebase is frozen, such as during the timebase sync
      code.
      
      This deals with kernel and vdso accesses, but other userspace accesses
      will of course need to be fixed elsewhere.
      Signed-off-by: NScott Wood <scottwood@freescale.com>
      d52459ca
  8. 14 8月, 2013 1 次提交
  9. 01 7月, 2013 1 次提交
    • P
      powerpc: Delete __cpuinit usage from all users · 061d19f2
      Paul Gortmaker 提交于
      The __cpuinit type of throwaway sections might have made sense
      some time ago when RAM was more constrained, but now the savings
      do not offset the cost and complications.  For example, the fix in
      commit 5e427ec2 ("x86: Fix bit corruption at CPU resume time")
      is a good example of the nasty type of bugs that can be created
      with improper use of the various __init prefixes.
      
      After a discussion on LKML[1] it was decided that cpuinit should go
      the way of devinit and be phased out.  Once all the users are gone,
      we can then finally remove the macros themselves from linux/init.h.
      
      This removes all the powerpc uses of the __cpuinit macros.  There
      are no __CPUINIT users in assembly files in powerpc.
      
      [1] https://lkml.org/lkml/2013/5/20/589
      
      Cc: Benjamin Herrenschmidt <benh@kernel.crashing.org>
      Cc: Paul Mackerras <paulus@samba.org>
      Cc: Josh Boyer <jwboyer@gmail.com>
      Cc: Matt Porter <mporter@kernel.crashing.org>
      Cc: Kumar Gala <galak@kernel.crashing.org>
      Cc: linuxppc-dev@lists.ozlabs.org
      Signed-off-by: NPaul Gortmaker <paul.gortmaker@windriver.com>
      Signed-off-by: NBenjamin Herrenschmidt <benh@kernel.crashing.org>
      061d19f2
  10. 04 4月, 2013 1 次提交
  11. 25 11月, 2012 1 次提交
  12. 13 9月, 2012 3 次提交
  13. 07 12月, 2011 1 次提交
  14. 04 11月, 2011 1 次提交
  15. 13 10月, 2011 1 次提交
    • K
      powerpc/85xx: Setup secondary cores PIR with hard SMP id · 45116806
      Kumar Gala 提交于
      Normally logical and hard cpu ID are the same, however in same cases like
      on the P3060 they may differ.  Where the logical is 0..5, the hard id
      goes 0,1,4..7.  This can causes issues for places we utilize PIR to index
      into array like in debug exception handlers for finding the exception
      stack.
      
      Move to setting up PIR with hard_smp_processor_id fixes the issue.
      Signed-off-by: NKumar Gala <galak@kernel.crashing.org>
      45116806
  16. 12 10月, 2011 1 次提交
  17. 08 7月, 2011 1 次提交
  18. 23 6月, 2011 1 次提交
  19. 20 6月, 2011 1 次提交
  20. 19 5月, 2011 1 次提交
    • M
      powerpc: Consolidate ipi message mux and demux · 23d72bfd
      Milton Miller 提交于
      Consolidate the mux and demux of ipi messages into smp.c and call
      a new smp_ops callback to actually trigger the ipi.
      
      The powerpc architecture code is optimised for having 4 distinct
      ipi triggers, which are mapped to 4 distinct messages (ipi many, ipi
      single, scheduler ipi, and enter debugger).  However, several interrupt
      controllers only provide a single software triggered interrupt that
      can be delivered to each cpu.  To resolve this limitation, each smp_ops
      implementation created a per-cpu variable that is manipulated with atomic
      bitops.  Since these lines will be contended they are optimialy marked as
      shared_aligned and take a full cache line for each cpu.  Distro kernels
      may have 2 or 3 of these in their config, each taking per-cpu space
      even though at most one will be in use.
      
      This consolidation removes smp_message_recv and replaces the single call
      actions cases with direct calls from the common message recognition loop.
      The complicated debugger ipi case with its muxed crash handling code is
      moved to debug_ipi_action which is now called from the demux code (instead
      of the multi-message action calling smp_message_recv).
      
      I put a call to reschedule_action to increase the likelyhood of correctly
      merging the anticipated scheduler_ipi() hook coming from the scheduler
      tree; that single required call can be inlined later.
      
      The actual message decode is a copy of the old pseries xics code with its
      memory barriers and cache line spacing, augmented with a per-cpu unsigned
      long based on the book-e doorbell code.  The optional data is set via a
      callback from the implementation and is passed to the new cause-ipi hook
      along with the logical cpu number.  While currently only the doorbell
      implemntation uses this data it should be almost zero cost to retrieve and
      pass it -- it adds a single register load for the argument from the same
      cache line to which we just completed a store and the register is dead
      on return from the call.  I extended the data element from unsigned int
      to unsigned long in case some other code wanted to associate a pointer.
      
      The doorbell check_self is replaced by a call to smp_muxed_ipi_resend,
      conditioned on the CPU_DBELL feature.  The ifdef guard could be relaxed
      to CONFIG_SMP but I left it with BOOKE for now.
      
      Also, the doorbell interrupt vector for book-e was not calling irq_enter
      and irq_exit, which throws off cpu accounting and causes code to not
      realize it is running in interrupt context.  Add the missing calls.
      Signed-off-by: NMilton Miller <miltonm@bga.com>
      Signed-off-by: NBenjamin Herrenschmidt <benh@kernel.crashing.org>
      23d72bfd
  21. 20 4月, 2011 1 次提交
  22. 15 3月, 2011 1 次提交
    • K
      powerpc/85xx: Fix writing to spin table 'cpu-release-addr' on ppc64e · decbb280
      Kumar Gala 提交于
      If the spin table is located in the linear mapping (which can happen if
      we have 4G or more of memory) we need to access the spin table via a
      cacheable coherent mapping like we do on ppc32 (and do explicit cache
      flush).
      
      See the following commit for the ppc32 version of this issue:
      
      commit d1d47ec6
      Author: Peter Tyser <ptyser@xes-inc.com>
      Date:   Fri Dec 18 16:50:37 2009 -0600
      
          powerpc/85xx: Fix SMP when "cpu-release-addr" is in lowmem
      Signed-off-by: NKumar Gala <galak@kernel.crashing.org>
      decbb280
  23. 14 10月, 2010 3 次提交
  24. 08 10月, 2010 1 次提交
  25. 03 8月, 2010 1 次提交
  26. 09 7月, 2010 1 次提交
    • B
      powerpc/book3e: More doorbell cleanups. Sample the PIR register · b9f1cd71
      Benjamin Herrenschmidt 提交于
      The doorbells use the content of the PIR register to match messages
      from other CPUs. This may or may not be the same as our linux CPU
      number, so using that as the "target" is no right.
      
      Instead, we sample the PIR register at boot on every processor
      and use that value subsequently when sending IPIs.
      
      We also use a per-cpu message mask rather than a global array which
      should limit cache line contention.
      
      Note: We could use the CPU number in the device-tree instead of
      the PIR register, as they are supposed to be equivalent. This
      might prove useful if doorbells are to be used to kick CPUs out
      of FW at boot time, thus before we can sample the PIR. This is
      however not the case now and using the PIR just works.
      Signed-off-by: NBenjamin Herrenschmidt <benh@kernel.crashing.org>
      b9f1cd71
  27. 14 2月, 2010 1 次提交
    • P
      powerpc/85xx: Fix SMP when "cpu-release-addr" is in lowmem · d1d47ec6
      Peter Tyser 提交于
      Recent U-Boot commit 5ccd29c3679b3669b0bde5c501c1aa0f325a7acb caused
      the "cpu-release-addr" device tree property to contain the physical RAM
      location that secondary cores were spinning at.  Previously, the
      "cpu-release-addr" property contained a value referencing the boot page
      translation address range of 0xfffffxxx, which then indirectly accessed
      RAM.
      
      The "cpu-release-addr" is currently ioremapped and the secondary cores
      kicked.  However, due to the recent change in "cpu-release-addr", it
      sometimes points to a memory location in low memory that cannot be
      ioremapped.  For example on a P2020-based board with 512MB of RAM the
      following error occurs on bootup:
      
        <...>
        mpic: requesting IPIs ...
        __ioremap(): phys addr 0x1ffff000 is RAM lr c05df9a0
        Unable to handle kernel paging request for data at address 0x00000014
        Faulting instruction address: 0xc05df9b0
        Oops: Kernel access of bad area, sig: 11 [#1]
        SMP NR_CPUS=2 P2020 RDB
        Modules linked in:
        <... eventual kernel panic>
      
      Adding logic to conditionally ioremap or access memory directly resolves
      the issue.
      Signed-off-by: NPeter Tyser <ptyser@xes-inc.com>
      Signed-off-by: NNate Case <ncase@xes-inc.com>
      Reported-by: NDipen Dudhat <B09055@freescale.com>
      Tested-by: NDipen Dudhat <B09055@freescale.com>
      Signed-off-by: NKumar Gala <galak@kernel.crashing.org>
      d1d47ec6
  28. 03 2月, 2010 1 次提交
    • P
      powerpc/85xx: Fix SMP when "cpu-release-addr" is in lowmem · 7b62922a
      Peter Tyser 提交于
      Recent U-Boot commit 5ccd29c3679b3669b0bde5c501c1aa0f325a7acb caused
      the "cpu-release-addr" device tree property to contain the physical RAM
      location that secondary cores were spinning at.  Previously, the
      "cpu-release-addr" property contained a value referencing the boot page
      translation address range of 0xfffffxxx, which then indirectly accessed
      RAM.
      
      The "cpu-release-addr" is currently ioremapped and the secondary cores
      kicked.  However, due to the recent change in "cpu-release-addr", it
      sometimes points to a memory location in low memory that cannot be
      ioremapped.  For example on a P2020-based board with 512MB of RAM the
      following error occurs on bootup:
      
        <...>
        mpic: requesting IPIs ...
        __ioremap(): phys addr 0x1ffff000 is RAM lr c05df9a0
        Unable to handle kernel paging request for data at address 0x00000014
        Faulting instruction address: 0xc05df9b0
        Oops: Kernel access of bad area, sig: 11 [#1]
        SMP NR_CPUS=2 P2020 RDB
        Modules linked in:
        <... eventual kernel panic>
      
      Adding logic to conditionally ioremap or access memory directly resolves
      the issue.
      Signed-off-by: NPeter Tyser <ptyser@xes-inc.com>
      Signed-off-by: NNate Case <ncase@xes-inc.com>
      Reported-by: NDipen Dudhat <B09055@freescale.com>
      Tested-by: NDipen Dudhat <B09055@freescale.com>
      Signed-off-by: NBenjamin Herrenschmidt <benh@kernel.crashing.org>
      7b62922a
  29. 11 9月, 2009 1 次提交
  30. 28 8月, 2009 1 次提交
  31. 20 8月, 2009 1 次提交
  32. 23 6月, 2009 1 次提交
    • K
      powerpc/85xx: Fix issue found by lockdep trace in smp_85xx_kick_cpu · cb1ffb62
      Kumar Gala 提交于
      lockdep trace found the following:
      
      ------------[ cut here ]------------
      Badness at c007baf0 [verbose debug info unavailable]
      NIP: c007baf0 LR: c007bad8 CTR: 00000000
      REGS: ef855e00 TRAP: 0700   Tainted: G        W
      (2.6.30-06736-g12a31df-dirty)
      MSR: 00021000 <ME,CE>  CR: 24044022  XER: 20000000
      TASK = ef858000[1] 'swapper' THREAD: ef854000 CPU: 0
      GPR00: 00000000 ef855eb0 ef858000 00000001 000000d0 f1000000 ffbc8000 ffffffff
      GPR08: 000000d0 c0760000 c0710000 00000007 2fffffff 1004a388 7ffd9400 00000000
      GPR16: 00000000 7ffcd100 7ffcd100 7ffcd100 c059cd78 c075c498 c057da7c ffffffff
      GPR24: ffbc8000 f1000000 00000001 c00bf8b0 c07595d4 000000d0 00021000 000000d0
      NIP [c007baf0] lockdep_trace_alloc+0xc0/0xf0
      LR [c007bad8] lockdep_trace_alloc+0xa8/0xf0
      Call Trace:
      [ef855eb0] [c007ba60] lockdep_trace_alloc+0x30/0xf0 (unreliable)
      [ef855ec0] [c00cb3ac] kmem_cache_alloc+0x2c/0xf0
      [ef855ee0] [c00bf8b0] __get_vm_area_node+0x80/0x1c0
      [ef855f10] [c0017580] __ioremap_caller+0x1d0/0x1e0
      [ef855f40] [c057da7c] smp_85xx_kick_cpu+0x64/0x124
      [ef855f60] [c0599180] __cpu_up+0xd0/0x1a4
      [ef855f80] [c05997c4] cpu_up+0x14c/0x1e0
      [ef855fc0] [c05732a0] kernel_init+0x100/0x1c4
      [ef855ff0] [c0011524] kernel_thread+0x4c/0x68
      Instruction dump:
      8009c174 2f800000 409e0048 73c08000 40820040 4818980d 2f830000 419effa0
      3d20c076 8009c388 2f800000 409eff90 <0fe00000> 4bffff88 60000000 60000000
      
      We were calling ioremap after we local_irq_restore(flags).  A simple
      reorder fixes the problem.
      Signed-off-by: NKumar Gala <galak@kernel.crashing.org>
      cb1ffb62
  33. 11 3月, 2009 1 次提交
  34. 31 12月, 2008 1 次提交
    • J
      powerpc/85xx: Add local_irq_restore in error handling code · 870029a6
      Julia Lawall 提交于
      There is a call to local_irq_restore in the normal exit case, so it would
      seem that there should be one on an error return as well.
      
      The semantic patch that makes this change is as follows:
      (http://www.emn.fr/x-info/coccinelle/)
      
      // <smpl>
      @@
      expression l;
      expression E,E1,E2;
      @@
      
      local_irq_save(l);
      ... when != local_irq_restore(l)
          when != spin_unlock_irqrestore(E,l)
          when any
          when strict
      (
      if (...) { ... when != local_irq_restore(l)
                     when != spin_unlock_irqrestore(E1,l)
      +   local_irq_restore(l);
          return ...;
      }
      |
      if (...)
      +   {local_irq_restore(l);
          return ...;
      +   }
      |
      spin_unlock_irqrestore(E2,l);
      |
      local_irq_restore(l);
      )
      // </smpl>
      Signed-off-by: NJulia Lawall <julia@diku.dk>
      Signed-off-by: NKumar Gala <galak@kernel.crashing.org>
      870029a6
  35. 03 12月, 2008 1 次提交