You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@nuttx.apache.org by xi...@apache.org on 2021/11/19 01:27:48 UTC
[incubator-nuttx] branch master updated: sched/sched: Fix some typos in comments.
This is an automated email from the ASF dual-hosted git repository.
xiaoxiang pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/incubator-nuttx.git
The following commit(s) were added to refs/heads/master by this push:
new bdc157f sched/sched: Fix some typos in comments.
bdc157f is described below
commit bdc157f443ae42f46dc75461d9331ca86dfc8dca
Author: Abdelatif Guettouche <ab...@espressif.com>
AuthorDate: Thu Nov 18 18:45:00 2021 +0100
sched/sched: Fix some typos in comments.
Signed-off-by: Abdelatif Guettouche <ab...@espressif.com>
---
sched/sched/sched_addreadytorun.c | 4 ++--
sched/sched/sched_cpuselect.c | 8 ++------
sched/sched/sched_lock.c | 4 ++--
sched/sched/sched_mergeprioritized.c | 2 +-
sched/sched/sched_unlock.c | 6 +++---
5 files changed, 10 insertions(+), 14 deletions(-)
diff --git a/sched/sched/sched_addreadytorun.c b/sched/sched/sched_addreadytorun.c
index 1b2abd4..8bd5705 100644
--- a/sched/sched/sched_addreadytorun.c
+++ b/sched/sched/sched_addreadytorun.c
@@ -143,7 +143,7 @@ bool nxsched_add_readytorun(FAR struct tcb_s *btcb)
* - The caller has already removed the input rtcb from whatever list it
* was in.
* - The caller handles the condition that occurs if the head of the
- * ready-to-run list is changed.
+ * ready-to-run list has changed.
*
****************************************************************************/
@@ -162,7 +162,7 @@ bool nxsched_add_readytorun(FAR struct tcb_s *btcb)
if ((btcb->flags & TCB_FLAG_CPU_LOCKED) != 0)
{
- /* Yes.. that that is the CPU we must use */
+ /* Yes.. that is the CPU we must use */
cpu = btcb->cpu;
}
diff --git a/sched/sched/sched_cpuselect.c b/sched/sched/sched_cpuselect.c
index e93b4ec..ec9b47d 100644
--- a/sched/sched/sched_cpuselect.c
+++ b/sched/sched/sched_cpuselect.c
@@ -67,23 +67,19 @@ int nxsched_select_cpu(cpu_set_t affinity)
int cpu;
int i;
- /* Otherwise, find the CPU that is executing the lowest priority task
- * (possibly its IDLE task).
- */
-
minprio = SCHED_PRIORITY_MAX;
cpu = IMPOSSIBLE_CPU;
for (i = 0; i < CONFIG_SMP_NCPUS; i++)
{
- /* If the thread permitted to run on this CPU? */
+ /* Is the thread permitted to run on this CPU? */
if ((affinity & (1 << i)) != 0)
{
FAR struct tcb_s *rtcb = (FAR struct tcb_s *)
g_assignedtasks[i].head;
- /* If this thread is executing its IDLE task, the use it. The
+ /* If this CPU is executing its IDLE task, then use it. The
* IDLE task is always the last task in the assigned task list.
*/
diff --git a/sched/sched/sched_lock.c b/sched/sched/sched_lock.c
index 4e47832..ed1d07b 100644
--- a/sched/sched/sched_lock.c
+++ b/sched/sched/sched_lock.c
@@ -85,10 +85,10 @@
* This assures that the scheduler does enforce the critical section.
* NOTE: Because of this spinlock, there should never be more than one
* bit set in 'g_cpu_lockset'; attempts to set additional bits should
- * be cause the CPU to block on the spinlock. However, additional bits
+ * cause the CPU to block on the spinlock. However, additional bits
* could get set in 'g_cpu_lockset' due to the context switches on the
* various CPUs.
- * 5. Each the time the head of a g_assignedtasks[] list changes and the
+ * 5. Each time the head of a g_assignedtasks[] list changes and the
* scheduler modifies 'g_cpu_lockset', it must also set 'g_cpu_schedlock'
* depending on the new state of 'g_cpu_lockset'.
* 5. Logic that currently uses the currently running tasks lockcount
diff --git a/sched/sched/sched_mergeprioritized.c b/sched/sched/sched_mergeprioritized.c
index 3afb2b2..e2080c8 100644
--- a/sched/sched/sched_mergeprioritized.c
+++ b/sched/sched/sched_mergeprioritized.c
@@ -39,7 +39,7 @@
* Name: nxsched_merge_prioritized
*
* Description:
- * This function merges the content of the prioritized task list '1ist1'
+ * This function merges the content of the prioritized task list 'list1'
* into the prioritized task list, 'list2'. On return 'list2' will contain
* the prioritized content of both lists; 'list1' will be empty.
*
diff --git a/sched/sched/sched_unlock.c b/sched/sched/sched_unlock.c
index a3b25f6..d0736c5 100644
--- a/sched/sched/sched_unlock.c
+++ b/sched/sched/sched_unlock.c
@@ -263,7 +263,7 @@ int sched_unlock(void)
* NOTE: This operation has a very high likelihood of causing
* this task to be switched out!
*
- * In the single CPU case, decrementing irqcount to zero is
+ * In the single CPU case, decrementing lockcount to zero is
* sufficient to release the pending tasks. Further, in that
* configuration, critical sections and pre-emption can operate
* fully independently.
@@ -277,7 +277,7 @@ int sched_unlock(void)
#if CONFIG_RR_INTERVAL > 0
/* If (1) the task that was running supported round-robin
* scheduling and (2) if its time slice has already expired, but
- * (3) it could not slice out because pre-emption was disabled,
+ * (3) it could not be sliced out because pre-emption was disabled,
* then we need to swap the task out now and reassess the interval
* timer for the next time slice.
*/
@@ -285,7 +285,7 @@ int sched_unlock(void)
if ((rtcb->flags & TCB_FLAG_POLICY_MASK) == TCB_FLAG_SCHED_RR &&
rtcb->timeslice == 0)
{
- /* Yes.. that is the situation. But one more thing. The call
+ /* Yes.. that is the situation. But one more thing: The call
* to up_release_pending() above may have actually replaced
* the task at the head of the ready-to-run list. In that
* case, we need only to reset the timeslice value back to the