You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@nuttx.apache.org by xi...@apache.org on 2021/07/24 16:00:52 UTC

[incubator-nuttx] branch master updated: sched/*/*spinlock.c: Fix some typos.

This is an automated email from the ASF dual-hosted git repository.

xiaoxiang pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/incubator-nuttx.git


The following commit(s) were added to refs/heads/master by this push:
     new de68507  sched/*/*spinlock.c: Fix some typos.
de68507 is described below

commit de68507f84f27e142e63dc4d17c8b44e2ac8b5f5
Author: Abdelatif Guettouche <ab...@espressif.com>
AuthorDate: Sat Jul 24 15:17:56 2021 +0200

    sched/*/*spinlock.c: Fix some typos.
    
    Signed-off-by: Abdelatif Guettouche <ab...@espressif.com>
---
 sched/irq/irq_spinlock.c   | 10 +++++-----
 sched/semaphore/spinlock.c |  2 +-
 2 files changed, 6 insertions(+), 6 deletions(-)

diff --git a/sched/irq/irq_spinlock.c b/sched/irq/irq_spinlock.c
index 0e41983..41638c8 100644
--- a/sched/irq/irq_spinlock.c
+++ b/sched/irq/irq_spinlock.c
@@ -57,12 +57,12 @@ static volatile uint8_t g_irq_spin_count[CONFIG_SMP_NCPUS];
  *     If the argument lock is not specified (i.e. NULL),
  *     disable local interrupts and take the global spinlock (g_irq_spin)
  *     if the call counter (g_irq_spin_count[cpu]) equals to 0. Then the
- *     counter on the CPU is increment to allow nested call and return
+ *     counter on the CPU is incremented to allow nested call and return
  *     the interrupt state.
  *
  *     If the argument lock is specified,
- *     disable local interrupts and take the lock spinlock and return
- *     the interrupt state.
+ *     disable local interrupts and take the given lock and return the
+ *     interrupt state.
  *
  *     NOTE: This API is very simple to protect data (e.g. H/W register
  *     or internal data structure) in SMP mode. But do not use this API
@@ -117,8 +117,8 @@ irqstate_t spin_lock_irqsave(spinlock_t *lock)
  *     restore the interrupt state as it was prior to the previous call to
  *     spin_lock_irqsave(NULL).
  *
- *     If the argument lock is specified, release the the lock and
- *     restore the interrupt state as it was prior to the previous call to
+ *     If the argument lock is specified, release the lock and restore
+ *     the interrupt state as it was prior to the previous call to
  *     spin_lock_irqsave(lock).
  *
  *   If SMP is not enabled:
diff --git a/sched/semaphore/spinlock.c b/sched/semaphore/spinlock.c
index d6a8bb4..5e87d8b 100644
--- a/sched/semaphore/spinlock.c
+++ b/sched/semaphore/spinlock.c
@@ -49,7 +49,7 @@
  *
  *   This implementation is non-reentrant and is prone to deadlocks in
  *   the case that any logic on the same CPU attempts to take the lock
- *   more than one
+ *   more than once.
  *
  * Input Parameters:
  *   lock - A reference to the spinlock object to lock.