linux-kernel.vger.kernel.org archive mirror
 help / color / mirror / Atom feed
* [patch] CFS scheduler, -v18
@ 2007-06-22 22:02 Ingo Molnar
  2007-06-22 22:09 ` S.Çağlar Onur
                   ` (6 more replies)
  0 siblings, 7 replies; 43+ messages in thread
From: Ingo Molnar @ 2007-06-22 22:02 UTC (permalink / raw)
  To: linux-kernel
  Cc: Linus Torvalds, Andrew Morton, Mike Galbraith, Arjan van de Ven,
	Thomas Gleixner, Dmitry Adamushko, Srivatsa Vaddagiri


i'm pleased to announce release -v18 of the CFS scheduler patchset.

The rolled-up CFS patch against today's -git kernel, v2.6.22-rc5, 
v2.6.22-rc4-mm2, v2.6.21.5 or v2.6.20.14 can be downloaded from the 
usual place:

   http://people.redhat.com/mingo/cfs-scheduler/

The biggest change in -v18 are various performance related improvements. 
Thomas Gleixner has eliminated expensive 64-bit divisions by converting 
the arithmetics to scaled math (without impacting the quality of 
calculations). Srivatsa Vaddagiri and Dmitry Adamushko have continued 
the abstraction and cleanup work. Srivatsa Vaddagiri and Christoph 
Lameter fixed the NUMA balancing bug reported by Paul McKenney. There 
were also a good number of other refinements to the CFS code. (No 
reproducible behavioral regressions were reported against -v17 so far, 
so the 'behavioral' bits are mostly unchanged.)
 
Changes since -v17:

 - implement scaled math speedups for CFS. (Thomas Gleixner)

 - lots of core code updates, cleanups and streamlining.
   (Srivatsa Vaddagiri, Dmitry Adamushko, me.)

 - bugfix: fix NUMA balancing. (Srivatsa Vaddagiri, Christoph Lameter,
   Paul E. McKenney)

 - feature: SCHED_IDLE now also implies block-scheduler (CFQ)
   idle-IO-priority. (suggested by Thomas Sattler, picked up from -ck)

 - build fix for ppc32. (reported, tested and confirmed fixed by
   Art Haas)

 - ARM fix. (reported and debugged by Thomas Gleixner)

 - cleanup: implemented idle_sched_class in kernel/sched_idletask.c as a
   way to separate out rq->idle handling out of the core scheduler. This
   made a good deal of idle-task related special-cases go away.

 - debug: make the sysctls safer by introducing high and low limits.

 - cleanup: move some of the debug counters to under CONFIG_SCHEDSTATS.

 - speedup: various micro-optimizations

 - various other small updates.

As usual, any sort of feedback, bugreport, fix and suggestion is more 
than welcome!
   
	Ingo

^ permalink raw reply	[flat|nested] 43+ messages in thread

* Re: [patch] CFS scheduler, -v18
  2007-06-22 22:02 [patch] CFS scheduler, -v18 Ingo Molnar
@ 2007-06-22 22:09 ` S.Çağlar Onur
  2007-06-22 22:16   ` S.Çağlar Onur
  2007-06-22 23:08 ` Gene Heskett
                   ` (5 subsequent siblings)
  6 siblings, 1 reply; 43+ messages in thread
From: S.Çağlar Onur @ 2007-06-22 22:09 UTC (permalink / raw)
  To: Ingo Molnar
  Cc: linux-kernel, Linus Torvalds, Andrew Morton, Mike Galbraith,
	Arjan van de Ven, Thomas Gleixner, Dmitry Adamushko,
	Srivatsa Vaddagiri

[-- Attachment #1: Type: text/plain, Size: 1244 bytes --]

Hi Ingo;

23 Haz 2007 Cts tarihinde, Ingo Molnar şunları yazmıştı: 
> As usual, any sort of feedback, bugreport, fix and suggestion is more
> than welcome!

caglar@zangetsu linux-2.6 $ LC_ALL=C make
  CHK     include/linux/version.h
  CHK     include/linux/utsrelease.h
  CALL    scripts/checksyscalls.sh
  CHK     include/linux/compile.h
  CC      kernel/sched.o
kernel/sched.c:745:28: sched_idletask.c: No such file or directory
kernel/sched.c: In function `init_idle_bootup_task':
kernel/sched.c:4659: error: `idle_sched_class' undeclared (first use in this 
function)
kernel/sched.c:4659: error: (Each undeclared identifier is reported only once
kernel/sched.c:4659: error: for each function it appears in.)
kernel/sched.c: In function `init_idle':
kernel/sched.c:4698: error: `idle_sched_class' undeclared (first use in this 
function)
kernel/sched.c: In function `sched_init':
kernel/sched.c:6196: error: `idle_sched_class' undeclared (first use in this 
function)
make[1]: *** [kernel/sched.o] Error 1
make: *** [kernel] Error 2

Cheers
-- 
S.Çağlar Onur <caglar@pardus.org.tr>
http://cekirdek.pardus.org.tr/~caglar/

Linux is like living in a teepee. No Windows, no Gates and an Apache in house!

[-- Attachment #2: This is a digitally signed message part. --]
[-- Type: application/pgp-signature, Size: 189 bytes --]

^ permalink raw reply	[flat|nested] 43+ messages in thread

* Re: [patch] CFS scheduler, -v18
  2007-06-22 22:09 ` S.Çağlar Onur
@ 2007-06-22 22:16   ` S.Çağlar Onur
  2007-06-22 22:20     ` Ingo Molnar
  0 siblings, 1 reply; 43+ messages in thread
From: S.Çağlar Onur @ 2007-06-22 22:16 UTC (permalink / raw)
  To: Ingo Molnar
  Cc: linux-kernel, Linus Torvalds, Andrew Morton, Mike Galbraith,
	Arjan van de Ven, Thomas Gleixner, Dmitry Adamushko,
	Srivatsa Vaddagiri

[-- Attachment #1: Type: text/plain, Size: 2780 bytes --]

23 Haz 2007 Cts tarihinde, S.Çağlar Onur şunları yazmıştı: 
> Hi Ingo;
>
> 23 Haz 2007 Cts tarihinde, Ingo Molnar şunları yazmıştı:
> > As usual, any sort of feedback, bugreport, fix and suggestion is more
> > than welcome!
>
> caglar@zangetsu linux-2.6 $ LC_ALL=C make
>   CHK     include/linux/version.h
>   CHK     include/linux/utsrelease.h
>   CALL    scripts/checksyscalls.sh
>   CHK     include/linux/compile.h
>   CC      kernel/sched.o
> kernel/sched.c:745:28: sched_idletask.c: No such file or directory

Ahh and this happens with [1], grabbing sched_idletask.c from .18 one solves 
the problem...

Index: linux/kernel/sched_idletask.c
===================================================================
--- /dev/null
+++ linux/kernel/sched_idletask.c
@@ -0,0 +1,68 @@
+/*
+ * idle-task scheduling class.
+ *
+ * (NOTE: these are not related to SCHED_IDLE tasks which are
+ *  handled in sched_fair.c)
+ */
+
+/*
+ * Idle tasks are unconditionally rescheduled:
+ */
+static void check_preempt_curr_idle(struct rq *rq, struct task_struct *p)
+{
+	resched_task(rq->idle);
+}
+
+static struct task_struct *pick_next_task_idle(struct rq *rq, u64 now)
+{
+	schedstat_inc(rq, sched_goidle);
+
+	return rq->idle;
+}
+
+/*
+ * It is not legal to sleep in the idle task - print a warning
+ * message if some code attempts to do it:
+ */
+static void
+dequeue_task_idle(struct rq *rq, struct task_struct *p, int sleep, u64 now)
+{
+	spin_unlock_irq(&rq->lock);
+	printk(KERN_ERR "bad: scheduling from the idle thread!\n");
+	dump_stack();
+	spin_lock_irq(&rq->lock);
+}
+
+static void put_prev_task_idle(struct rq *rq, struct task_struct *prev, u64 
now)
+{
+}
+
+static struct task_struct *load_balance_start_idle(struct rq *rq)
+{
+	return NULL;
+}
+
+static void task_tick_idle(struct rq *rq, struct task_struct *curr)
+{
+}
+
+/*
+ * Simple, special scheduling class for the per-CPU idle tasks:
+ */
+struct sched_class idle_sched_class __read_mostly = {
+	/* no enqueue/yield_task for idle tasks */
+
+	/* dequeue is not valid, we print a debug message there: */
+	.dequeue_task		= dequeue_task_idle,
+
+	.check_preempt_curr	= check_preempt_curr_idle,
+
+	.pick_next_task		= pick_next_task_idle,
+	.put_prev_task		= put_prev_task_idle,
+
+	.load_balance_start	= load_balance_start_idle,
+	/* no .load_balance_next for idle tasks */
+
+	.task_tick		= task_tick_idle,
+	/* no .task_new for idle tasks */
+};


[1] 
http://people.redhat.com/mingo/cfs-scheduler/sched-cfs-v2.6.22-git-v18.patch

Cheers
-- 
S.Çağlar Onur <caglar@pardus.org.tr>
http://cekirdek.pardus.org.tr/~caglar/

Linux is like living in a teepee. No Windows, no Gates and an Apache in house!

[-- Attachment #2: This is a digitally signed message part. --]
[-- Type: application/pgp-signature, Size: 189 bytes --]

^ permalink raw reply	[flat|nested] 43+ messages in thread

* Re: [patch] CFS scheduler, -v18
  2007-06-22 22:16   ` S.Çağlar Onur
@ 2007-06-22 22:20     ` Ingo Molnar
  2007-06-26  3:02       ` Andrew Morton
  0 siblings, 1 reply; 43+ messages in thread
From: Ingo Molnar @ 2007-06-22 22:20 UTC (permalink / raw)
  To: S.Çağlar Onur
  Cc: linux-kernel, Linus Torvalds, Andrew Morton, Mike Galbraith,
	Arjan van de Ven, Thomas Gleixner, Dmitry Adamushko,
	Srivatsa Vaddagiri


* S.Çağlar Onur <caglar@pardus.org.tr> wrote:

> > kernel/sched.c:745:28: sched_idletask.c: No such file or directory
> 
> Ahh and this happens with [1], grabbing sched_idletask.c from .18 one solves 
> the problem...

oops, indeed - i've fixed up the -git patch:

  http://people.redhat.com/mingo/cfs-scheduler/sched-cfs-v2.6.22-git-v18.patch

	Ingo

^ permalink raw reply	[flat|nested] 43+ messages in thread

* Re: [patch] CFS scheduler, -v18
  2007-06-22 22:02 [patch] CFS scheduler, -v18 Ingo Molnar
  2007-06-22 22:09 ` S.Çağlar Onur
@ 2007-06-22 23:08 ` Gene Heskett
  2007-06-23  7:11   ` Ingo Molnar
  2007-06-23 10:22 ` Antonino Ingargiola
                   ` (4 subsequent siblings)
  6 siblings, 1 reply; 43+ messages in thread
From: Gene Heskett @ 2007-06-22 23:08 UTC (permalink / raw)
  To: Ingo Molnar
  Cc: linux-kernel, Linus Torvalds, Andrew Morton, Mike Galbraith,
	Arjan van de Ven, Thomas Gleixner, Dmitry Adamushko,
	Srivatsa Vaddagiri

On Friday 22 June 2007, Ingo Molnar wrote:
>i'm pleased to announce release -v18 of the CFS scheduler patchset.
>
>The rolled-up CFS patch against today's -git kernel, v2.6.22-rc5,
>v2.6.22-rc4-mm2, v2.6.21.5 or v2.6.20.14 can be downloaded from the
>usual place:
>
>   http://people.redhat.com/mingo/cfs-scheduler/
>
>The biggest change in -v18 are various performance related improvements.
>Thomas Gleixner has eliminated expensive 64-bit divisions by converting
>the arithmetics to scaled math (without impacting the quality of
>calculations). Srivatsa Vaddagiri and Dmitry Adamushko have continued
>the abstraction and cleanup work. Srivatsa Vaddagiri and Christoph
>Lameter fixed the NUMA balancing bug reported by Paul McKenney. There
>were also a good number of other refinements to the CFS code. (No
>reproducible behavioral regressions were reported against -v17 so far,
>so the 'behavioral' bits are mostly unchanged.)
>
>Changes since -v17:
>
> - implement scaled math speedups for CFS. (Thomas Gleixner)
>
> - lots of core code updates, cleanups and streamlining.
>   (Srivatsa Vaddagiri, Dmitry Adamushko, me.)
>
> - bugfix: fix NUMA balancing. (Srivatsa Vaddagiri, Christoph Lameter,
>   Paul E. McKenney)
>
> - feature: SCHED_IDLE now also implies block-scheduler (CFQ)
>   idle-IO-priority. (suggested by Thomas Sattler, picked up from -ck)
>
> - build fix for ppc32. (reported, tested and confirmed fixed by
>   Art Haas)
>
> - ARM fix. (reported and debugged by Thomas Gleixner)
>
> - cleanup: implemented idle_sched_class in kernel/sched_idletask.c as a
>   way to separate out rq->idle handling out of the core scheduler. This
>   made a good deal of idle-task related special-cases go away.
>
> - debug: make the sysctls safer by introducing high and low limits.
>
> - cleanup: move some of the debug counters to under CONFIG_SCHEDSTATS.
>
> - speedup: various micro-optimizations
>
> - various other small updates.
>
>As usual, any sort of feedback, bugreport, fix and suggestion is more
>than welcome!
>
Humm, problem methinks.  Applying the patch, with 2.6.22-rc5 applied to 2.6.21 
completed, from my script:

now applying patch sched-cfs-v2.6.22-rc5-v18.patch

patching file Documentation/kernel-parameters.txt
patching file Documentation/sched-design-CFS.txt
patching file Makefile
patching file arch/i386/kernel/smpboot.c
patching file arch/i386/kernel/tsc.c
patching file arch/ia64/kernel/setup.c
patching file arch/mips/kernel/smp.c
patching file arch/sparc/kernel/smp.c
patching file arch/sparc64/kernel/smp.c
patching file block/cfq-iosched.c
patching file fs/proc/array.c
patching file fs/proc/base.c
patching file include/asm-generic/bitops/sched.h
patching file include/linux/hardirq.h
patching file include/linux/sched.h
patching file include/linux/topology.h
patching file init/main.c
patching file kernel/delayacct.c
patching file kernel/exit.c
patching file kernel/fork.c
patching file kernel/posix-cpu-timers.c
patching file kernel/sched.c
patching file kernel/sched_debug.c
patching file kernel/sched_fair.c
patching file kernel/sched_idletask.c
patching file kernel/sched_rt.c
patching file kernel/sched_stats.h
patching file kernel/softirq.c
patching file kernel/sysctl.c
The next patch would delete the file l/kernel/sched.c,
which does not exist!  Assume -R? [n]

How to proceed?

-- 
Cheers, Gene
"There are four boxes to be used in defense of liberty:
 soap, ballot, jury, and ammo. Please use in that order."
-Ed Howdershelt (Author)
Even more amazing was the realization that God has Internet access.  I
wonder if He has a full newsfeed?
	-- Matt Welsh

^ permalink raw reply	[flat|nested] 43+ messages in thread

* Re: [patch] CFS scheduler, -v18
  2007-06-22 23:08 ` Gene Heskett
@ 2007-06-23  7:11   ` Ingo Molnar
  2007-06-23  9:55     ` Gene Heskett
  0 siblings, 1 reply; 43+ messages in thread
From: Ingo Molnar @ 2007-06-23  7:11 UTC (permalink / raw)
  To: Gene Heskett
  Cc: linux-kernel, Linus Torvalds, Andrew Morton, Mike Galbraith,
	Arjan van de Ven, Thomas Gleixner, Dmitry Adamushko,
	Srivatsa Vaddagiri


* Gene Heskett <gene.heskett@gmail.com> wrote:

> patching file kernel/softirq.c
> patching file kernel/sysctl.c
> The next patch would delete the file l/kernel/sched.c,
> which does not exist!  Assume -R? [n]
> 
> How to proceed?

oops - just ignore it, or re-download the patch (i fixed it).

	Ingo

^ permalink raw reply	[flat|nested] 43+ messages in thread

* Re: [patch] CFS scheduler, -v18
  2007-06-23  7:11   ` Ingo Molnar
@ 2007-06-23  9:55     ` Gene Heskett
  0 siblings, 0 replies; 43+ messages in thread
From: Gene Heskett @ 2007-06-23  9:55 UTC (permalink / raw)
  To: linux-kernel

On Saturday 23 June 2007, Ingo Molnar wrote:
>* Gene Heskett <gene.heskett@gmail.com> wrote:
>> patching file kernel/softirq.c
>> patching file kernel/sysctl.c
>> The next patch would delete the file l/kernel/sched.c,
>> which does not exist!  Assume -R? [n]
>>
>> How to proceed?
>
>oops - just ignore it, or re-download the patch (i fixed it).
>
>	Ingo

answering n for all that, I note the build, at the end of the make bzImage,
spits out this:
  MODPOST vmlinux
WARNING: arch/i386/kernel/built-in.o(.text+0x845d): Section mismatch: 
reference to .init.text:amd_init_mtrr (between 'mtrr_bp_init' 
and 'mtrr_save_state')
WARNING: arch/i386/kernel/built-in.o(.text+0x8462): Section mismatch: 
reference to .init.text:cyrix_init_mtrr (between 'mtrr_bp_init' 
and 'mtrr_save_state')
WARNING: arch/i386/kernel/built-in.o(.text+0x8467): Section mismatch: 
reference to .init.text:centaur_init_mtrr (between 'mtrr_bp_init' 
and 'mtrr_save_state')
WARNING: arch/i386/kernel/built-in.o(.text+0x9284): Section mismatch: 
reference to .init.text: (between 'get_mtrr_state' and 'generic_get_mtrr')
WARNING: arch/i386/kernel/built-in.o(.text+0x9298): Section mismatch: 
reference to .init.text: (between 'get_mtrr_state' and 'generic_get_mtrr')
WARNING: arch/i386/kernel/built-in.o(.text+0x92bc): Section mismatch: 
reference to .init.text: (between 'get_mtrr_state' and 'generic_get_mtrr')

But then proceeds with the make modules stage.  I believe I've seen references
to this in other threads.  Is It Serious?

-- 
Cheers, Gene
"There are four boxes to be used in defense of liberty:
 soap, ballot, jury, and ammo. Please use in that order."
-Ed Howdershelt (Author)
Debug is human, de-fix divine.

^ permalink raw reply	[flat|nested] 43+ messages in thread

* Re: [patch] CFS scheduler, -v18
  2007-06-22 22:02 [patch] CFS scheduler, -v18 Ingo Molnar
  2007-06-22 22:09 ` S.Çağlar Onur
  2007-06-22 23:08 ` Gene Heskett
@ 2007-06-23 10:22 ` Antonino Ingargiola
  2007-06-23 17:25   ` Ingo Molnar
  2007-06-23 13:24 ` Willy Tarreau
                   ` (3 subsequent siblings)
  6 siblings, 1 reply; 43+ messages in thread
From: Antonino Ingargiola @ 2007-06-23 10:22 UTC (permalink / raw)
  To: Ingo Molnar; +Cc: linux-kernel

Hi,

2007/6/23, Ingo Molnar <mingo@elte.hu>:
>
> i'm pleased to announce release -v18 of the CFS scheduler patchset.

I'm running -v18 on 2.6.22-rc5, no problems so far. How can I change a
task to SCHED_IDLE or SCHED_BATCH priority under CFS?


Thanks,

    ~ Antonio

^ permalink raw reply	[flat|nested] 43+ messages in thread

* Re: [patch] CFS scheduler, -v18
  2007-06-22 22:02 [patch] CFS scheduler, -v18 Ingo Molnar
                   ` (2 preceding siblings ...)
  2007-06-23 10:22 ` Antonino Ingargiola
@ 2007-06-23 13:24 ` Willy Tarreau
  2007-06-24 15:52   ` Ingo Molnar
  2007-06-26 20:17 ` Fortier,Vincent [Montreal]
                   ` (2 subsequent siblings)
  6 siblings, 1 reply; 43+ messages in thread
From: Willy Tarreau @ 2007-06-23 13:24 UTC (permalink / raw)
  To: Ingo Molnar
  Cc: linux-kernel, Linus Torvalds, Andrew Morton, Mike Galbraith,
	Arjan van de Ven, Thomas Gleixner, Dmitry Adamushko,
	Srivatsa Vaddagiri

Hi Ingo,

On Sat, Jun 23, 2007 at 12:02:02AM +0200, Ingo Molnar wrote:
> 
> i'm pleased to announce release -v18 of the CFS scheduler patchset.
> 
> The rolled-up CFS patch against today's -git kernel, v2.6.22-rc5, 
> v2.6.22-rc4-mm2, v2.6.21.5 or v2.6.20.14 can be downloaded from the 
> usual place:
> 
>    http://people.redhat.com/mingo/cfs-scheduler/
> 
> The biggest change in -v18 are various performance related improvements. 
> Thomas Gleixner has eliminated expensive 64-bit divisions by converting 
> the arithmetics to scaled math (without impacting the quality of 
> calculations). Srivatsa Vaddagiri and Dmitry Adamushko have continued 
> the abstraction and cleanup work. Srivatsa Vaddagiri and Christoph 
> Lameter fixed the NUMA balancing bug reported by Paul McKenney. There 
> were also a good number of other refinements to the CFS code. (No 
> reproducible behavioral regressions were reported against -v17 so far, 
> so the 'behavioral' bits are mostly unchanged.)

Today I had a little time to try CFS again (last time it was -v9!). I ran it
on top of 2.6.20.14, and simply tried ocbench again. You remember ? With -v9,
I ran 64 processes which all progressed very smoothly. With -v18, it's not
the case anymore. When I run 64 processes, only 7 of them show smooth rounds,
while all the other ones are only updated once a second. Sometimes they only
progress by one iteration, sometimes by a full round. Some are even updated
once ever 2 seconds, because if I drag an xterm above them and quickly remove
it, the xterm leaves a trace there for up to 2 seconds.

Also, only one of my 2 CPUs is used. I see the rq vary between 1 and 5, with 
a permanent 50% idle... :

   procs                      memory      swap          io     system      cpu
 r  b  w   swpd   free   buff  cache   si   so    bi    bo   in    cs us sy id
 1  0  0      0 874400   7864  90436    0    0     0     0  279  2204 50  0 50
 3  0  0      0 874408   7864  90436    0    0     0     0  273  2122 50  1 50
 1  0  0      0 874408   7864  90436    0    0     0     0  253  1660 49  1 50
 3  0  0      0 874408   7864  90436    0    0     0     0  252  1977 50  0 50
 2  0  0      0 874408   7864  90436    0    0     0     0  253  2274 49  1 50
 3  0  0      0 874408   7864  90436    0    0     0     0  252  1846 49  1 50
 1  0  0      0 874408   7864  90436    0    0     0     0  339  1782 49  1 50

I have no idea about what version brought that unexpected behaviour, but it's
clearly something which needs to be tracked down.

My scheduler is at 250 Hz, and here are the values I found in /proc/sys/kernel:
root@pcw:kernel# grep '' sched_*
sched_batch_wakeup_granularity_ns:40000000
sched_child_runs_first:1
sched_features:14
sched_granularity_ns:10000000
sched_runtime_limit_ns:40000000
sched_stat_granularity_ns:0
sched_wakeup_granularity_ns:4000000

I have tried to change each of them, with absolutely no effect. Seems really
strange. Unfortunately, I have to leave right now. Maybe you can try on your
side with this simple command :

  $ ./ocbench -x 8 -y 8

Sorry for not giving more information right now.

Regards,
Willy


^ permalink raw reply	[flat|nested] 43+ messages in thread

* Re: [patch] CFS scheduler, -v18
  2007-06-23 10:22 ` Antonino Ingargiola
@ 2007-06-23 17:25   ` Ingo Molnar
  2007-06-24 10:02     ` Antonino Ingargiola
  0 siblings, 1 reply; 43+ messages in thread
From: Ingo Molnar @ 2007-06-23 17:25 UTC (permalink / raw)
  To: Antonino Ingargiola; +Cc: linux-kernel


* Antonino Ingargiola <tritemio@gmail.com> wrote:

> 2007/6/23, Ingo Molnar <mingo@elte.hu>:
> >
> >i'm pleased to announce release -v18 of the CFS scheduler patchset.
> 
> I'm running -v18 on 2.6.22-rc5, no problems so far. How can I change a 
> task to SCHED_IDLE or SCHED_BATCH priority under CFS?

pick up schedtool, and these are the choices it gives:

    -N                    for SCHED_NORMAL
    -F -p PRIO            for SCHED_FIFO       only as root
    -R -p PRIO            for SCHED_RR         only as root
    -B                    for SCHED_BATCH
    -I -p PRIO            for SCHED_ISO
    -D                    for SCHED_IDLEPRIO

then for example to start up something as SCHED_IDLE:

  schedtool -D -e ./somecommand.sh

	Ingo

^ permalink raw reply	[flat|nested] 43+ messages in thread

* Re: [patch] CFS scheduler, -v18
  2007-06-23 17:25   ` Ingo Molnar
@ 2007-06-24 10:02     ` Antonino Ingargiola
  2007-06-24 11:07       ` Ingo Molnar
  0 siblings, 1 reply; 43+ messages in thread
From: Antonino Ingargiola @ 2007-06-24 10:02 UTC (permalink / raw)
  To: Ingo Molnar; +Cc: linux-kernel

2007/6/23, Ingo Molnar <mingo@elte.hu>:
>
> * Antonino Ingargiola <tritemio@gmail.com> wrote:
>
> > 2007/6/23, Ingo Molnar <mingo@elte.hu>:
> > >
> > >i'm pleased to announce release -v18 of the CFS scheduler patchset.
> >
> > I'm running -v18 on 2.6.22-rc5, no problems so far. How can I change a
> > task to SCHED_IDLE or SCHED_BATCH priority under CFS?
>
> pick up schedtool, and these are the choices it gives:
>
>     -N                    for SCHED_NORMAL
>     -F -p PRIO            for SCHED_FIFO       only as root
>     -R -p PRIO            for SCHED_RR         only as root
>     -B                    for SCHED_BATCH
>     -I -p PRIO            for SCHED_ISO
>     -D                    for SCHED_IDLEPRIO
>
> then for example to start up something as SCHED_IDLE:
>
>   schedtool -D -e ./somecommand.sh
>

Thank you very much! I was thinking that schedtool was suitable only
for -ck. I've installed schedtool and it works fine.

Anyway, I've discovered with great pleasure that CFS has also the
SCHED_ISO priority. I may have missed something, but I don't remember
to have read this in any of the CFS release notes :). For me this is a
really useful feature. Thanks.


Regards,

    ~ Antonio

>         Ingo
>

^ permalink raw reply	[flat|nested] 43+ messages in thread

* Re: [patch] CFS scheduler, -v18
  2007-06-24 10:02     ` Antonino Ingargiola
@ 2007-06-24 11:07       ` Ingo Molnar
  2007-06-25  7:27         ` Antonino Ingargiola
  0 siblings, 1 reply; 43+ messages in thread
From: Ingo Molnar @ 2007-06-24 11:07 UTC (permalink / raw)
  To: Antonino Ingargiola; +Cc: linux-kernel


* Antonino Ingargiola <tritemio@gmail.com> wrote:

> Anyway, I've discovered with great pleasure that CFS has also the 
> SCHED_ISO priority. I may have missed something, but I don't remember 
> to have read this in any of the CFS release notes :). For me this is a 
> really useful feature. Thanks.

well, it's only a hack and emulated: SCHED_ISO in CFS is recognized as a 
policy but it falls back to SCHED_NORMAL. Could you check how well this 
(i.e. SCHED_NORMAL) works for your workload, compared to SD's SCHED_ISO? 
If you'd like to increase the priority of a task, i'd suggest to use 
negative nice levels. (use the 'nice' option in 
/etc/security/limits.conf with a newer version of PAM to allow 
unprivileged users to use negative nice levels.)

	Ingo

^ permalink raw reply	[flat|nested] 43+ messages in thread

* Re: [patch] CFS scheduler, -v18
  2007-06-23 13:24 ` Willy Tarreau
@ 2007-06-24 15:52   ` Ingo Molnar
  2007-06-24 17:08     ` Willy Tarreau
  0 siblings, 1 reply; 43+ messages in thread
From: Ingo Molnar @ 2007-06-24 15:52 UTC (permalink / raw)
  To: Willy Tarreau
  Cc: linux-kernel, Linus Torvalds, Andrew Morton, Mike Galbraith,
	Arjan van de Ven, Thomas Gleixner, Dmitry Adamushko,
	Srivatsa Vaddagiri


* Willy Tarreau <w@1wt.eu> wrote:

> Today I had a little time to try CFS again (last time it was -v9!). I 
> ran it on top of 2.6.20.14, and simply tried ocbench again. You 
> remember ? With -v9, I ran 64 processes which all progressed very 
> smoothly. With -v18, it's not the case anymore. When I run 64 
> processes, only 7 of them show smooth rounds, while all the other ones 
> are only updated once a second. Sometimes they only progress by one 
> iteration, sometimes by a full round. Some are even updated once ever 
> 2 seconds, because if I drag an xterm above them and quickly remove 
> it, the xterm leaves a trace there for up to 2 seconds.
> 
> Also, only one of my 2 CPUs is used. I see the rq vary between 1 and 
> 5, with a permanent 50% idle... :
> 
>    procs                      memory      swap          io     system      cpu
>  r  b  w   swpd   free   buff  cache   si   so    bi    bo   in    cs us sy id
>  1  0  0      0 874400   7864  90436    0    0     0     0  279  2204 50  0 50
>  3  0  0      0 874408   7864  90436    0    0     0     0  273  2122 50  1 50
>  1  0  0      0 874408   7864  90436    0    0     0     0  253  1660 49  1 50
>  3  0  0      0 874408   7864  90436    0    0     0     0  252  1977 50  0 50
>  2  0  0      0 874408   7864  90436    0    0     0     0  253  2274 49  1 50
>  3  0  0      0 874408   7864  90436    0    0     0     0  252  1846 49  1 50
>  1  0  0      0 874408   7864  90436    0    0     0     0  339  1782 49  1 50
> 
> I have no idea about what version brought that unexpected behaviour, 
> but it's clearly something which needs to be tracked down.

hm, the two problems might be related. Could you try v17 perhaps? In v18 
i have 'unified' all the sched.c's between the various kernel releases, 
maybe that brought in something unexpected on 2.6.20.14. (perhaps try 
v2.6.21.5 based cfs too?)

> My scheduler is at 250 Hz, and here are the values I found in 
> /proc/sys/kernel:

could you send me the file the cfs-debug-info.sh script produced. You 
can pick the script up from:

  http://people.redhat.com/mingo/cfs-scheduler/tools/cfs-debug-info.sh

(i'd suggest to send it in private mail, output is large and detailed. 
If your kernel has no /proc/config.gz then please send me your .config 
file too.)

	Ingo

^ permalink raw reply	[flat|nested] 43+ messages in thread

* Re: [patch] CFS scheduler, -v18
  2007-06-24 15:52   ` Ingo Molnar
@ 2007-06-24 17:08     ` Willy Tarreau
  2007-06-24 20:31       ` Ingo Molnar
  0 siblings, 1 reply; 43+ messages in thread
From: Willy Tarreau @ 2007-06-24 17:08 UTC (permalink / raw)
  To: Ingo Molnar
  Cc: linux-kernel, Linus Torvalds, Andrew Morton, Mike Galbraith,
	Arjan van de Ven, Thomas Gleixner, Dmitry Adamushko,
	Srivatsa Vaddagiri

Hi Ingo,

On Sun, Jun 24, 2007 at 05:52:14PM +0200, Ingo Molnar wrote:
> 
> * Willy Tarreau <w@1wt.eu> wrote:
> 
> > Today I had a little time to try CFS again (last time it was -v9!). I 
> > ran it on top of 2.6.20.14, and simply tried ocbench again. You 
> > remember ? With -v9, I ran 64 processes which all progressed very 
> > smoothly. With -v18, it's not the case anymore. When I run 64 
> > processes, only 7 of them show smooth rounds, while all the other ones 
> > are only updated once a second. Sometimes they only progress by one 
> > iteration, sometimes by a full round. Some are even updated once ever 
> > 2 seconds, because if I drag an xterm above them and quickly remove 
> > it, the xterm leaves a trace there for up to 2 seconds.
> > 
> > Also, only one of my 2 CPUs is used. I see the rq vary between 1 and 
> > 5, with a permanent 50% idle... :
> > 
> >    procs                      memory      swap          io     system      cpu
> >  r  b  w   swpd   free   buff  cache   si   so    bi    bo   in    cs us sy id
> >  1  0  0      0 874400   7864  90436    0    0     0     0  279  2204 50  0 50
> >  3  0  0      0 874408   7864  90436    0    0     0     0  273  2122 50  1 50
> >  1  0  0      0 874408   7864  90436    0    0     0     0  253  1660 49  1 50
> >  3  0  0      0 874408   7864  90436    0    0     0     0  252  1977 50  0 50
> >  2  0  0      0 874408   7864  90436    0    0     0     0  253  2274 49  1 50
> >  3  0  0      0 874408   7864  90436    0    0     0     0  252  1846 49  1 50
> >  1  0  0      0 874408   7864  90436    0    0     0     0  339  1782 49  1 50
> > 
> > I have no idea about what version brought that unexpected behaviour, 
> > but it's clearly something which needs to be tracked down.
> 
> hm, the two problems might be related. Could you try v17 perhaps? In v18 
> i have 'unified' all the sched.c's between the various kernel releases, 
> maybe that brought in something unexpected on 2.6.20.14. (perhaps try 
> v2.6.21.5 based cfs too?)

Well, forget this, I'm nuts. I'm sorry, but I did not set any of the -R
and -S parameter on ocbench, which means that all the processes ran at
full speed and did not sleep. The load distribution was not fair, but
since they put a lot of stress on the X server, I think it might be one
of the reasons for the unfairness. I got the same behaviour with -v17,
-v9 and even 2.4 ! It told me something was wrong on my side ;-)

I've retried with 50%/50% run/sleep, and it now works like a charm. It's
perfectly smooth with both small and long run/sleep times (between 1 and 100
ms). I think that with X saturated, it might explain why I only had one CPU
running at 100% !

Next time, I'll try to take a bit more time for such a test.

> could you send me the file the cfs-debug-info.sh script produced. You 
> can pick the script up from:
> 
>   http://people.redhat.com/mingo/cfs-scheduler/tools/cfs-debug-info.sh

OK I got it, but I've not run it since the problem was between the keyboard
and the chair. If you want an output anyway, I can give it a run.

Sorry again for the wrong alert.

regards,
willy


^ permalink raw reply	[flat|nested] 43+ messages in thread

* Re: [patch] CFS scheduler, -v18
  2007-06-24 17:08     ` Willy Tarreau
@ 2007-06-24 20:31       ` Ingo Molnar
  0 siblings, 0 replies; 43+ messages in thread
From: Ingo Molnar @ 2007-06-24 20:31 UTC (permalink / raw)
  To: Willy Tarreau
  Cc: linux-kernel, Linus Torvalds, Andrew Morton, Mike Galbraith,
	Arjan van de Ven, Thomas Gleixner, Dmitry Adamushko,
	Srivatsa Vaddagiri


* Willy Tarreau <w@1wt.eu> wrote:

> > > I have no idea about what version brought that unexpected 
> > > behaviour, but it's clearly something which needs to be tracked 
> > > down.
> > 
> > hm, the two problems might be related. Could you try v17 perhaps? In 
> > v18 i have 'unified' all the sched.c's between the various kernel 
> > releases, maybe that brought in something unexpected on 2.6.20.14. 
> > (perhaps try v2.6.21.5 based cfs too?)
> 
> Well, forget this, I'm nuts. I'm sorry, but I did not set any of the 
> -R and -S parameter on ocbench, which means that all the processes ran 
> at full speed and did not sleep. The load distribution was not fair, 
> but since they put a lot of stress on the X server, I think it might 
> be one of the reasons for the unfairness. I got the same behaviour 
> with -v17, -v9 and even 2.4 ! It told me something was wrong on my 
> side ;-)
> 
> I've retried with 50%/50% run/sleep, and it now works like a charm. 
> It's perfectly smooth with both small and long run/sleep times 
> (between 1 and 100 ms). I think that with X saturated, it might 
> explain why I only had one CPU running at 100% !

ah, great! :-) My testbox needs a 90% / 10% ratio between sleep/run for 
an 8x8 matrix of ocbench tasks to not overload the X server. Once the 
overload happens X starts penalizing certain clients it finds abusive (i 
think), and that mechanism seems to be wall-clock based and it thus 
brings in alot of non-determinism and skews the clients.

	Ingo

^ permalink raw reply	[flat|nested] 43+ messages in thread

* Re: [patch] CFS scheduler, -v18
  2007-06-24 11:07       ` Ingo Molnar
@ 2007-06-25  7:27         ` Antonino Ingargiola
  0 siblings, 0 replies; 43+ messages in thread
From: Antonino Ingargiola @ 2007-06-25  7:27 UTC (permalink / raw)
  To: Ingo Molnar; +Cc: linux-kernel

2007/6/24, Ingo Molnar <mingo@elte.hu>:
>
> * Antonino Ingargiola <tritemio@gmail.com> wrote:
>
> > Anyway, I've discovered with great pleasure that CFS has also the
> > SCHED_ISO priority. I may have missed something, but I don't remember
> > to have read this in any of the CFS release notes :). For me this is a
> > really useful feature. Thanks.
>
> well, it's only a hack and emulated: SCHED_ISO in CFS is recognized as a
> policy but it falls back to SCHED_NORMAL. Could you check how well this
> (i.e. SCHED_NORMAL) works for your workload, compared to SD's SCHED_ISO?

To be fair, my workload is not really "critical". I'm used to
skip-free audio listening (no matter what) since long time running my
audio player with SCHED_ISO. Even in mainline the skips aren't so
frequent, but still annoying. I'm using SCHED_ISO for the confidence
it gives in providing skip-free audio.

For my modest needs also CFS SCHED_NORMAL has been just fine (in these
latest days). I'll report if I can find a more critical workload that
can possibly stress CFS SCHED_NORMAL.


Regards,

    ~ Antonio

^ permalink raw reply	[flat|nested] 43+ messages in thread

* Re: [patch] CFS scheduler, -v18
  2007-06-22 22:20     ` Ingo Molnar
@ 2007-06-26  3:02       ` Andrew Morton
  2007-06-26  8:38         ` Ingo Molnar
  0 siblings, 1 reply; 43+ messages in thread
From: Andrew Morton @ 2007-06-26  3:02 UTC (permalink / raw)
  To: Ingo Molnar
  Cc: S.Çağlar Onur, linux-kernel, Linus Torvalds,
	Mike Galbraith, Arjan van de Ven, Thomas Gleixner,
	Dmitry Adamushko, Srivatsa Vaddagiri

On Sat, 23 Jun 2007 00:20:36 +0200 Ingo Molnar <mingo@elte.hu> wrote:

> 
> * S.Çağlar Onur <caglar@pardus.org.tr> wrote:
> 
> > > kernel/sched.c:745:28: sched_idletask.c: No such file or directory
> > 
> > Ahh and this happens with [1], grabbing sched_idletask.c from .18 one solves 
> > the problem...
> 
> oops, indeed - i've fixed up the -git patch:
> 
>   http://people.redhat.com/mingo/cfs-scheduler/sched-cfs-v2.6.22-git-v18.patch
> 

So I locally generated the diff to take -mm up to the above version of CFS.


- sys_sched_yield_to() went away?  I guess I missed that.

- Curious.  the simplification of task_tick_rt() seems to go only
  halfway.  Could do

	if (p->policy != SCHED_RR)
		return;

	if (--p->time_slice)
		return;

	/* stuff goes here */

- dud macro:
					
#define is_rt_policy(p)		((p) == SCHED_FIFO || (p) == SCHED_RR)

  It evaluates its arg twice and could and should be coded in C.

  There are a bunch of other don't-need-to-be-implemented-as-a-macro
  macros around there too.  Generally, I suggest you review all the
  patchset for macros-which-don't-need-to-be-macros.

- Extraneous newline:

enum cpu_idle_type
{

- Style thing:

struct sched_entity {
	struct load_weight load;	/* for nice- load-balancing purposes */
	int on_rq;
	struct rb_node run_node;
	unsigned long delta_exec;
	s64 delta_fair;

	u64 wait_start_fair;
	u64 wait_start;
	u64 exec_start;
	u64 sleep_start, sleep_start_fair;
	u64 block_start;
	u64 sleep_max;
	u64 block_max;
	u64 exec_max;
	u64 wait_max;
	u64 last_ran;

	s64 wait_runtime;
	u64 sum_exec_runtime;
	s64 fair_key;
	s64 sum_wait_runtime, sum_sleep_runtime;
	unsigned long wait_runtime_overruns, wait_runtime_underruns;
};

  I think the one-definition-per-line style is better than the `unsigned
  long foo,bar,zot,zit;' thing.  Easier to read, easier to read subsequent
  patches and it leaves more room for a comment describing what the field
  does.

- None of these fields have comments describing what they do ;)

- __exit_signal() does apparently-unlocked 64-bit arith.  Is there some
  implicit locking here or do we not care about the occasional race-induced
  inaccuracy?

  (ditto, lots of places, I expect)

  (Gee, there's shitloads of 64-bit stuff in there.  Does it all _really_
  need to be 64-bit on 32-bit?)

- weight_s64() (what does this do?) looks too big to inline on 32-bit.

- update_stats_enqueue() looks too big to inline even on 64-bit.

- Overall, this change is tremendously huge for something which is
  supposedly ready-to-merge.  Looks like a lot of that is the sched_entity
  conversion, but afaict there's quite a lot besides.

- Should "4" in

	(sysctl_sched_features & 4)

  be enumerated?

- Maybe even __check_preempt_curr_fair() is too porky to inline.

- There really is an astonishing amount of 64-bit arith in here...

- Some opportunities for useful comments have been missed ;)

#define NICE_0_LOAD	SCHED_LOAD_SCALE
#define NICE_0_SHIFT	SCHED_LOAD_SHIFT

  <wonders what these mean>

- Should any of those new 64-bit arith functions in sched.c be pulled out
  and made generic?

- update_curr_load() is huge, inlined and has several callsites?

- lots more macros-which-dont-need-to-be-macros in sched.c:
  load_weight(), PRIO_TO_load_weight(), RTPRIO_TO_load_weight(), maybe
  others.  People are more inclined to comment functions than they are
  macros, for some reason.

- inc_load(), dec_load(), inc_nr_running(), dec_nr_running(): these will
  generate plenty of code on 32-bit and they're all inlined with multiple
  callsites.

- overall, CFS takes sched.o from 41157 of .text up to 48781 on x86_64,
  which at 18% is rather a large bloat.  Hopefully a lot of this is the new
  debug stuff.

- On i386 sched.o went from 33755 up to 43660 which is 29% growth. 
  Possibly acceptable, but why did it increase a lot more than the x86_64
  version?  All that 64-bit arith, I assume?

- style (or the lack thereof):

	p->se.sum_wait_runtime = p->se.sum_sleep_runtime = 0;
	p->se.sleep_start = p->se.sleep_start_fair = p->se.block_start = 0;
	p->se.sleep_max = p->se.block_max = p->se.exec_max = p->se.wait_max = 0;
	p->se.wait_runtime_overruns = p->se.wait_runtime_underruns = 0;

  bit of an eyesore?

- in sched_init() this looks funny:

		rq->ls.load_update_last = sched_clock();
		rq->ls.load_update_start = sched_clock();

  was it intended that these both get the same value?

	

^ permalink raw reply	[flat|nested] 43+ messages in thread

* Re: [patch] CFS scheduler, -v18
  2007-06-26  3:02       ` Andrew Morton
@ 2007-06-26  8:38         ` Ingo Molnar
  2007-06-26  9:00           ` Andrew Morton
  0 siblings, 1 reply; 43+ messages in thread
From: Ingo Molnar @ 2007-06-26  8:38 UTC (permalink / raw)
  To: Andrew Morton
  Cc: S.Çağlar Onur, linux-kernel, Linus Torvalds,
	Mike Galbraith, Arjan van de Ven, Thomas Gleixner,
	Dmitry Adamushko, Srivatsa Vaddagiri


* Andrew Morton <akpm@linux-foundation.org> wrote:

> So I locally generated the diff to take -mm up to the above version of 
> CFS.

thx. I released a diff against mm2:

 http://people.redhat.com/mingo/cfs-scheduler/sched-cfs-v2.6.22-rc4-mm2-v18.patch

but indeed the -git diff serves you better if you updated -mm to Linus' 
latest.

firstly, thanks a ton for your review feedback!

> - sys_sched_yield_to() went away?  I guess I missed that.

yep. Nobody tried it and sent any feedback on it, it was causing 
patch-logistical complications both in -mm and for packagers that bundle 
CFS (the experimental-schedulers site has a CFS repo and Fedora rawhide 
started carrying CFS recently as well), and i dont really agree with 
adding yet another yield interface anyway. So we can and should do this 
independently of CFS.

> - Curious.  the simplification of task_tick_rt() seems to go only
>   halfway.  Could do
> 
> 	if (p->policy != SCHED_RR)
> 		return;
> 
> 	if (--p->time_slice)
> 		return;
> 
> 	/* stuff goes here */

yeah. I have fixed it in my v19 tree for it to look like you suggest.

> - dud macro:
> 					
> #define is_rt_policy(p)		((p) == SCHED_FIFO || (p) == SCHED_RR)
> 
>   It evaluates its arg twice and could and should be coded in C.
> 
>   There are a bunch of other don't-need-to-be-implemented-as-a-macro
>   macros around there too.  Generally, I suggest you review all the
>   patchset for macros-which-don't-need-to-be-macros.

yep, fixed. (is a historic macro)

> - Extraneous newline:
> 
> enum cpu_idle_type
> {

fixed. (is a pre-existing enum)

> - Style thing:
> 
> struct sched_entity {

> 	u64 sleep_start, sleep_start_fair;

fixed.

> - None of these fields have comments describing what they do ;)

one of them has ;-) Will fill this in.

> - __exit_signal() does apparently-unlocked 64-bit arith.  Is there 
>   some implicit locking here or do we not care about the occasional 
>   race-induced inaccuracy?

do you mean the tsk->se.sum_exec_runtime addition, etc? That runs with 
interrupts disabled so sum_sched_runtime is protected.

>   (ditto, lots of places, I expect)

which places do you mean?

>   (Gee, there's shitloads of 64-bit stuff in there.  Does it all 
>   _really_ need to be 64-bit on 32-bit?)

yes - CFS is fundamentally designed for 64-bit, with still pretty OK 
arithmetics performance for 32-bit.

> - weight_s64() (what does this do?) looks too big to inline on 32-bit.

ok, i've uninlined it.

> - update_stats_enqueue() looks too big to inline even on 64-bit.

done.

> - Overall, this change is tremendously huge for something which is
>   supposedly ready-to-merge. [...]

hey, that's not fair, your review comments just made it 10K larger ;-)

> [...] Looks like a lot of that is the sched_entity conversion, but 
> afaict there's quite a lot besides.
> 
> - Should "4" in
> 
> 	(sysctl_sched_features & 4)
> 
>   be enumerated?

yep, done.

> - Maybe even __check_preempt_curr_fair() is too porky to inline.

yep - undone.

> - There really is an astonishing amount of 64-bit arith in here...
> 
> - Some opportunities for useful comments have been missed ;)
> 
> #define NICE_0_LOAD	SCHED_LOAD_SCALE
> #define NICE_0_SHIFT	SCHED_LOAD_SHIFT
> 
>   <wonders what these mean>

SCHED_LOAD_SCALE is the smpnice stuff. CFS reuses that and also makes it 
clear via this define that a nice-0 task has a 'load' contribution to 
the CPU as of NICE_0_LOAD. Sometimes, when doing smpnice load-balancing 
calculations we want to use 'SCHED_LOAD_SCALE', sometimes we want to 
stress it's NICE_0_LOAD.

> - Should any of those new 64-bit arith functions in sched.c be pulled 
>   out and made generic?

yep, the plan is to put this all into reciprocal_div.h and to convert 
existing users of reciprocal_div to the cleaner stuff from Thomas. The 
patch wont get any smaller due to that though ;-)

> - update_curr_load() is huge, inlined and has several callsites?

this is a reasonable tradeoff i think - update_curr_load()'s slowpath is 
in __update_curr_load(). Anyway, it probably wont get inlined when the 
kernel is built with -Os and without forced-inlining.

> - lots more macros-which-dont-need-to-be-macros in sched.c:
>   load_weight(), PRIO_TO_load_weight(), RTPRIO_TO_load_weight(), maybe
>   others.  People are more inclined to comment functions than they are
>   macros, for some reason.

these are mostly ancient macros. I fixed up some of them in my current 
tree.

> - inc_load(), dec_load(), inc_nr_running(), dec_nr_running(): these will
>   generate plenty of code on 32-bit and they're all inlined with 
>   multiple callsites.

yep - i'll revisit the inlining picture. This is not really a primary 
worry i think because it's easy to tweak and people can already express 
their inlining preference via CONFIG_CC_OPTIMIZE_FOR_SIZE and 
CONFIG_FORCED_INLINING.

> - overall, CFS takes sched.o from 41157 of .text up to 48781 on x86_64,
>   which at 18% is rather a large bloat.  Hopefully a lot of this is 
>   the new debug stuff.

> - On i386 sched.o went from 33755 up to 43660 which is 29% growth. 
>   Possibly acceptable, but why did it increase a lot more than the x86_64
>   version?  All that 64-bit arith, I assume?

the main reason is the sched debugging stuff:

   text    data     bss     dec     hex filename
  37570    2538      20   40128    9cc0 kernel/sched.o
  30692    2426      20   33138    8172 kernel/sched-no_sched_debug.o

i can make it depend on CONFIG_SCHEDSTATS, although i'd prefer it to be 
always on.

> - style (or the lack thereof):
> 
> 	p->se.sum_wait_runtime = p->se.sum_sleep_runtime = 0;
> 	p->se.sleep_start = p->se.sleep_start_fair = p->se.block_start = 0;
> 	p->se.sleep_max = p->se.block_max = p->se.exec_max = p->se.wait_max = 0;
> 	p->se.wait_runtime_overruns = p->se.wait_runtime_underruns = 0;
> 
>   bit of an eyesore?

fixed. (this heap grew gradually and now is/was an eyesore indeed.)

> - in sched_init() this looks funny:
> 
> 		rq->ls.load_update_last = sched_clock();
> 		rq->ls.load_update_start = sched_clock();
> 
>   was it intended that these both get the same value?

it doesnt really matter, i fixed them to be initialized to the same 
'now' value.

i've attached my current fixes. (Please dont apply it yet.)

	Ingo

Not-Signed-off-by: Ingo Molnar <mingo@elte.hu>

---
 Makefile              |    2 
 include/linux/sched.h |  120 +++++++++++++++++++++++++++++---------------------
 kernel/exit.c         |    2 
 kernel/sched.c        |   58 ++++++++++++++++--------
 kernel/sched_debug.c  |    2 
 kernel/sched_fair.c   |   61 ++++++++++++-------------
 kernel/sched_rt.c     |   15 +++---
 7 files changed, 149 insertions(+), 111 deletions(-)

Index: linux/Makefile
===================================================================
--- linux.orig/Makefile
+++ linux/Makefile
@@ -1,7 +1,7 @@
 VERSION = 2
 PATCHLEVEL = 6
 SUBLEVEL = 22
-EXTRAVERSION = -rc6-cfs-v18
+EXTRAVERSION = -rc6-cfs-v19
 NAME = Holy Dancing Manatees, Batman!
 
 # *DOCUMENTATION*
Index: linux/include/linux/sched.h
===================================================================
--- linux.orig/include/linux/sched.h
+++ linux/include/linux/sched.h
@@ -528,31 +528,6 @@ struct signal_struct {
 #define SIGNAL_STOP_CONTINUED	0x00000004 /* SIGCONT since WCONTINUED reap */
 #define SIGNAL_GROUP_EXIT	0x00000008 /* group exit in progress */
 
-
-/*
- * Priority of a process goes from 0..MAX_PRIO-1, valid RT
- * priority is 0..MAX_RT_PRIO-1, and SCHED_NORMAL/SCHED_BATCH
- * tasks are in the range MAX_RT_PRIO..MAX_PRIO-1. Priority
- * values are inverted: lower p->prio value means higher priority.
- *
- * The MAX_USER_RT_PRIO value allows the actual maximum
- * RT priority to be separate from the value exported to
- * user-space.  This allows kernel threads to set their
- * priority to a value higher than any user task. Note:
- * MAX_RT_PRIO must not be smaller than MAX_USER_RT_PRIO.
- */
-
-#define MAX_USER_RT_PRIO	100
-#define MAX_RT_PRIO		MAX_USER_RT_PRIO
-
-#define MAX_PRIO		(MAX_RT_PRIO + 40)
-#define DEFAULT_PRIO		(MAX_RT_PRIO + 20)
-
-#define rt_prio(prio)		unlikely((prio) < MAX_RT_PRIO)
-#define rt_task(p)		rt_prio((p)->prio)
-#define is_rt_policy(p)		((p) == SCHED_FIFO || (p) == SCHED_RR)
-#define has_rt_policy(p)	unlikely(is_rt_policy((p)->policy))
-
 /*
  * Some day this will be a full-fledged user tracking system..
  */
@@ -646,8 +621,7 @@ static inline int sched_info_on(void)
 #endif
 }
 
-enum cpu_idle_type
-{
+enum cpu_idle_type {
 	CPU_IDLE,
 	CPU_NOT_IDLE,
 	CPU_NEWLY_IDLE,
@@ -843,30 +817,45 @@ struct load_weight {
 	unsigned long weight, inv_weight;
 };
 
-/* CFS stats for a schedulable entity (task, task-group etc) */
+/*
+ * CFS stats for a schedulable entity (task, task-group etc)
+ *
+ * Current field usage histogram:
+ *
+ *     4 se->block_start
+ *     4 se->run_node
+ *     4 se->sleep_start
+ *     4 se->sleep_start_fair
+ *     6 se->load.weight
+ *     7 se->delta_fair
+ *    15 se->wait_runtime
+ */
 struct sched_entity {
-	struct load_weight load;	/* for nice- load-balancing purposes */
-	int on_rq;
-	struct rb_node run_node;
-	unsigned long delta_exec;
-	s64 delta_fair;
-
-	u64 wait_start_fair;
-	u64 wait_start;
-	u64 exec_start;
-	u64 sleep_start, sleep_start_fair;
-	u64 block_start;
-	u64 sleep_max;
-	u64 block_max;
-	u64 exec_max;
-	u64 wait_max;
-	u64 last_ran;
-
-	s64 wait_runtime;
-	u64 sum_exec_runtime;
-	s64 fair_key;
-	s64 sum_wait_runtime, sum_sleep_runtime;
-	unsigned long wait_runtime_overruns, wait_runtime_underruns;
+	s64			wait_runtime;
+	s64			delta_fair;
+	struct load_weight	load;		/* for load-balancing */
+	struct rb_node		run_node;
+	int			on_rq;
+	unsigned long		delta_exec;
+
+	u64			wait_start_fair;
+	u64			wait_start;
+	u64			exec_start;
+	u64			sleep_start;
+	u64			sleep_start_fair;
+	u64			block_start;
+	u64			sleep_max;
+	u64			block_max;
+	u64			exec_max;
+	u64			wait_max;
+	u64			last_ran;
+
+	u64			sum_exec_runtime;
+	s64			fair_key;
+	s64			sum_wait_runtime;
+	s64			sum_sleep_runtime;
+	unsigned long		wait_runtime_overruns;
+	unsigned long		wait_runtime_underruns;
 };
 
 struct task_struct {
@@ -1126,6 +1115,37 @@ struct task_struct {
 #endif
 };
 
+/*
+ * Priority of a process goes from 0..MAX_PRIO-1, valid RT
+ * priority is 0..MAX_RT_PRIO-1, and SCHED_NORMAL/SCHED_BATCH
+ * tasks are in the range MAX_RT_PRIO..MAX_PRIO-1. Priority
+ * values are inverted: lower p->prio value means higher priority.
+ *
+ * The MAX_USER_RT_PRIO value allows the actual maximum
+ * RT priority to be separate from the value exported to
+ * user-space.  This allows kernel threads to set their
+ * priority to a value higher than any user task. Note:
+ * MAX_RT_PRIO must not be smaller than MAX_USER_RT_PRIO.
+ */
+
+#define MAX_USER_RT_PRIO	100
+#define MAX_RT_PRIO		MAX_USER_RT_PRIO
+
+#define MAX_PRIO		(MAX_RT_PRIO + 40)
+#define DEFAULT_PRIO		(MAX_RT_PRIO + 20)
+
+static inline int rt_prio(int prio)
+{
+	if (unlikely(prio < MAX_RT_PRIO))
+		return 1;
+	return 0;
+}
+
+static inline int rt_task(struct task_struct *p)
+{
+	return rt_prio(p->prio);
+}
+
 static inline pid_t process_group(struct task_struct *tsk)
 {
 	return tsk->signal->pgrp;
Index: linux/kernel/exit.c
===================================================================
--- linux.orig/kernel/exit.c
+++ linux/kernel/exit.c
@@ -290,7 +290,7 @@ static void reparent_to_kthreadd(void)
 	/* Set the exit signal to SIGCHLD so we signal init on exit */
 	current->exit_signal = SIGCHLD;
 
-	if (!has_rt_policy(current) && (task_nice(current) < 0))
+	if (task_nice(current) < 0)
 		set_user_nice(current, 0);
 	/* cpus_allowed? */
 	/* rt_priority? */
Index: linux/kernel/sched.c
===================================================================
--- linux.orig/kernel/sched.c
+++ linux/kernel/sched.c
@@ -106,6 +106,18 @@ unsigned long long __attribute__((weak))
 #define MIN_TIMESLICE		max(5 * HZ / 1000, 1)
 #define DEF_TIMESLICE		(100 * HZ / 1000)
 
+static inline int rt_policy(int policy)
+{
+	if (unlikely(policy == SCHED_FIFO) || unlikely(policy == SCHED_RR))
+		return 1;
+	return 0;
+}
+
+static inline int task_has_rt_policy(struct task_struct *p)
+{
+	return rt_policy(p->policy);
+}
+
 /*
  * This is the priority-queue data structure of the RT scheduling class:
  */
@@ -752,7 +764,7 @@ static void set_load_weight(struct task_
 	task_rq(p)->cfs.wait_runtime -= p->se.wait_runtime;
 	p->se.wait_runtime = 0;
 
-	if (has_rt_policy(p)) {
+	if (task_has_rt_policy(p)) {
 		p->se.load.weight = prio_to_weight[0] * 2;
 		p->se.load.inv_weight = prio_to_wmult[0] >> 1;
 		return;
@@ -805,7 +817,7 @@ static inline int normal_prio(struct tas
 {
 	int prio;
 
-	if (has_rt_policy(p))
+	if (task_has_rt_policy(p))
 		prio = MAX_RT_PRIO-1 - p->rt_priority;
 	else
 		prio = __normal_prio(p);
@@ -1476,17 +1488,24 @@ int fastcall wake_up_state(struct task_s
  */
 static void __sched_fork(struct task_struct *p)
 {
-	p->se.wait_start_fair = p->se.wait_start = p->se.exec_start = 0;
-	p->se.sum_exec_runtime = 0;
-	p->se.delta_exec = 0;
-	p->se.delta_fair = 0;
-
-	p->se.wait_runtime = 0;
-
-	p->se.sum_wait_runtime = p->se.sum_sleep_runtime = 0;
-	p->se.sleep_start = p->se.sleep_start_fair = p->se.block_start = 0;
-	p->se.sleep_max = p->se.block_max = p->se.exec_max = p->se.wait_max = 0;
-	p->se.wait_runtime_overruns = p->se.wait_runtime_underruns = 0;
+	p->se.wait_start_fair		= 0;
+	p->se.wait_start		= 0;
+	p->se.exec_start		= 0;
+	p->se.sum_exec_runtime		= 0;
+	p->se.delta_exec		= 0;
+	p->se.delta_fair		= 0;
+	p->se.wait_runtime		= 0;
+	p->se.sum_wait_runtime		= 0;
+	p->se.sum_sleep_runtime		= 0;
+	p->se.sleep_start		= 0;
+	p->se.sleep_start_fair		= 0;
+	p->se.block_start		= 0;
+	p->se.sleep_max			= 0;
+	p->se.block_max			= 0;
+	p->se.exec_max			= 0;
+	p->se.wait_max			= 0;
+	p->se.wait_runtime_overruns	= 0;
+	p->se.wait_runtime_underruns	= 0;
 
 	INIT_LIST_HEAD(&p->run_list);
 	p->se.on_rq = 0;
@@ -1799,7 +1818,7 @@ static void update_cpu_load(struct rq *t
 	int i, scale;
 
 	this_rq->nr_load_updates++;
-	if (sysctl_sched_features & 64)
+	if (unlikely(!(sysctl_sched_features & SCHED_FEAT_PRECISE_CPU_LOAD)))
 		goto do_avg;
 
 	/* Update delta_fair/delta_exec fields first */
@@ -3801,7 +3820,7 @@ void set_user_nice(struct task_struct *p
 	 * it wont have any effect on scheduling until the task is
 	 * SCHED_FIFO/SCHED_RR:
 	 */
-	if (has_rt_policy(p)) {
+	if (task_has_rt_policy(p)) {
 		p->static_prio = NICE_TO_PRIO(nice);
 		goto out_unlock;
 	}
@@ -3999,14 +4018,14 @@ recheck:
 	    (p->mm && param->sched_priority > MAX_USER_RT_PRIO-1) ||
 	    (!p->mm && param->sched_priority > MAX_RT_PRIO-1))
 		return -EINVAL;
-	if (is_rt_policy(policy) != (param->sched_priority != 0))
+	if (rt_policy(policy) != (param->sched_priority != 0))
 		return -EINVAL;
 
 	/*
 	 * Allow unprivileged RT tasks to decrease priority:
 	 */
 	if (!capable(CAP_SYS_NICE)) {
-		if (is_rt_policy(policy)) {
+		if (rt_policy(policy)) {
 			unsigned long rlim_rtprio;
 			unsigned long flags;
 
@@ -6186,6 +6205,7 @@ int in_sched_functions(unsigned long add
 
 void __init sched_init(void)
 {
+	u64 now = sched_clock();
 	int highest_cpu = 0;
 	int i, j;
 
@@ -6206,8 +6226,8 @@ void __init sched_init(void)
 		rq->nr_running = 0;
 		rq->cfs.tasks_timeline = RB_ROOT;
 		rq->clock = rq->cfs.fair_clock = 1;
-		rq->ls.load_update_last = sched_clock();
-		rq->ls.load_update_start = sched_clock();
+		rq->ls.load_update_last = now;
+		rq->ls.load_update_start = now;
 
 		for (j = 0; j < CPU_LOAD_IDX_MAX; j++)
 			rq->cpu_load[j] = 0;
Index: linux/kernel/sched_debug.c
===================================================================
--- linux.orig/kernel/sched_debug.c
+++ linux/kernel/sched_debug.c
@@ -157,7 +157,7 @@ static int sched_debug_show(struct seq_f
 	u64 now = ktime_to_ns(ktime_get());
 	int cpu;
 
-	SEQ_printf(m, "Sched Debug Version: v0.03, cfs-v18, %s %.*s\n",
+	SEQ_printf(m, "Sched Debug Version: v0.03, cfs-v19, %s %.*s\n",
 		init_utsname()->release,
 		(int)strcspn(init_utsname()->version, " "),
 		init_utsname()->version);
Index: linux/kernel/sched_fair.c
===================================================================
--- linux.orig/kernel/sched_fair.c
+++ linux/kernel/sched_fair.c
@@ -61,8 +61,24 @@ unsigned int sysctl_sched_stat_granulari
  */
 unsigned int sysctl_sched_runtime_limit __read_mostly;
 
+/*
+ * Debugging: various feature bits
+ */
+enum {
+	SCHED_FEAT_IGNORE_PREEMPTED	= 1,
+	SCHED_FEAT_DISTRIBUTE		= 2,
+	SCHED_FEAT_FAIR_SLEEPERS	= 4,
+	SCHED_FEAT_SLEEPER_AVG		= 32,
+	SCHED_FEAT_PRECISE_CPU_LOAD	= 64,
+	SCHED_FEAT_START_DEBIT		= 128,
+	SCHED_FEAT_SKIP_INITIAL		= 256,
+};
+
 unsigned int sysctl_sched_features __read_mostly =
-			0 | 2 | 4 | 8 | 0 | 0 | 0 | 0;
+		SCHED_FEAT_DISTRIBUTE		|
+		SCHED_FEAT_FAIR_SLEEPERS	|
+		SCHED_FEAT_SLEEPER_AVG		|
+		SCHED_FEAT_PRECISE_CPU_LOAD;
 
 extern struct sched_class fair_sched_class;
 
@@ -145,7 +161,7 @@ static inline void
 __dequeue_entity(struct cfs_rq *cfs_rq, struct sched_entity *se)
 {
 	if (cfs_rq->rb_leftmost == &se->run_node)
-		cfs_rq->rb_leftmost = NULL;
+		cfs_rq->rb_leftmost = rb_next(&se->run_node);
 	rb_erase(&se->run_node, &cfs_rq->tasks_timeline);
 	update_load_sub(&cfs_rq->load, se->load.weight);
 	cfs_rq->nr_running--;
@@ -258,7 +274,7 @@ __update_curr(struct cfs_rq *cfs_rq, str
 	 * Task already marked for preemption, do not burden
 	 * it with the cost of not having left the CPU yet:
 	 */
-	if (unlikely(sysctl_sched_features & 1))
+	if (unlikely(sysctl_sched_features & SCHED_FEAT_IGNORE_PREEMPTED))
 		if (unlikely(test_tsk_thread_flag(curtask, TIF_NEED_RESCHED)))
 			return;
 
@@ -305,7 +321,7 @@ update_stats_wait_start(struct cfs_rq *c
 	se->wait_start = now;
 }
 
-static inline s64 weight_s64(s64 calc, unsigned long weight, int shift)
+static s64 weight_s64(s64 calc, unsigned long weight, int shift)
 {
 	if (calc < 0) {
 		calc = - calc * weight;
@@ -317,7 +333,7 @@ static inline s64 weight_s64(s64 calc, u
 /*
  * Task is being enqueued - update stats:
  */
-static inline void
+static void
 update_stats_enqueue(struct cfs_rq *cfs_rq, struct sched_entity *se, u64 now)
 {
 	s64 key;
@@ -438,7 +454,7 @@ static void distribute_fair_add(struct c
 	struct sched_entity *curr = cfs_rq_curr(cfs_rq);
 	s64 delta_fair = 0;
 
-	if (!(sysctl_sched_features & 2))
+	if (!(sysctl_sched_features & SCHED_FEAT_DISTRIBUTE))
 		return;
 
 	if (cfs_rq->nr_running) {
@@ -469,7 +485,7 @@ __enqueue_sleeper(struct cfs_rq *cfs_rq,
 	 * Fix up delta_fair with the effect of us running
 	 * during the whole sleep period:
 	 */
-	if (!(sysctl_sched_features & 32))
+	if (sysctl_sched_features & SCHED_FEAT_SLEEPER_AVG)
 		delta_fair = div64_s(delta_fair * load, load + se->load.weight);
 
 	delta_fair = weight_s64(delta_fair, se->load.weight, NICE_0_SHIFT);
@@ -495,7 +511,7 @@ enqueue_sleeper(struct cfs_rq *cfs_rq, s
 	s64 delta_fair;
 
 	if ((entity_is_task(se) && tsk->policy == SCHED_BATCH) ||
-						 !(sysctl_sched_features & 4))
+			 !(sysctl_sched_features & SCHED_FEAT_FAIR_SLEEPERS))
 		return;
 
 	delta_fair = cfs_rq->fair_clock - se->sleep_start_fair;
@@ -574,7 +590,7 @@ static void dequeue_entity(struct cfs_rq
 /*
  * Preempt the current task with a newly woken task if needed:
  */
-static inline void
+static void
 __check_preempt_curr_fair(struct cfs_rq *cfs_rq, struct sched_entity *se,
 			  struct sched_entity *curr, unsigned long granularity)
 {
@@ -612,23 +628,6 @@ put_prev_entity(struct cfs_rq *cfs_rq, s
 	int updated = 0;
 
 	/*
-	 * If the task is still waiting for the CPU (it just got
-	 * preempted), update its position within the tree and
-	 * start the wait period:
-	 */
-	if ((sysctl_sched_features & 16) && entity_is_task(prev))  {
-		struct task_struct *prevtask = task_of(prev);
-
-		if (prev->on_rq &&
-			test_tsk_thread_flag(prevtask, TIF_NEED_RESCHED)) {
-
-			dequeue_entity(cfs_rq, prev, 0, now);
-			enqueue_entity(cfs_rq, prev, 0, now);
-			updated = 1;
-		}
-	}
-
-	/*
 	 * If still on the runqueue then deactivate_task()
 	 * was not called and update_curr() has to be done:
 	 */
@@ -741,10 +740,8 @@ static void check_preempt_curr_fair(stru
 	unsigned long gran;
 
 	if (unlikely(rt_prio(p->prio))) {
-		if (sysctl_sched_features & 8) {
-			if (rt_prio(p->prio))
-				update_curr(cfs_rq, rq_clock(rq));
-		}
+		if (rt_prio(p->prio))
+			update_curr(cfs_rq, rq_clock(rq));
 		resched_task(curr);
 		return;
 	}
@@ -850,14 +847,14 @@ static void task_new_fair(struct rq *rq,
 	 * The first wait is dominated by the child-runs-first logic,
 	 * so do not credit it with that waiting time yet:
 	 */
-	if (sysctl_sched_features & 256)
+	if (sysctl_sched_features & SCHED_FEAT_SKIP_INITIAL)
 		p->se.wait_start_fair = 0;
 
 	/*
 	 * The statistical average of wait_runtime is about
 	 * -granularity/2, so initialize the task with that:
 	 */
-	if (sysctl_sched_features & 128)
+	if (sysctl_sched_features & SCHED_FEAT_START_DEBIT)
 		p->se.wait_runtime = -(s64)(sysctl_sched_granularity / 2);
 
 	__enqueue_entity(cfs_rq, se);
Index: linux/kernel/sched_rt.c
===================================================================
--- linux.orig/kernel/sched_rt.c
+++ linux/kernel/sched_rt.c
@@ -12,7 +12,7 @@ static inline void update_curr_rt(struct
 	struct task_struct *curr = rq->curr;
 	u64 delta_exec;
 
-	if (!has_rt_policy(curr))
+	if (!task_has_rt_policy(curr))
 		return;
 
 	delta_exec = now - curr->se.exec_start;
@@ -179,13 +179,14 @@ static void task_tick_rt(struct rq *rq, 
 	if (p->policy != SCHED_RR)
 		return;
 
-	if (!(--p->time_slice)) {
-		p->time_slice = static_prio_timeslice(p->static_prio);
-		set_tsk_need_resched(p);
+	if (--p->time_slice)
+		return;
 
-		/* put it at the end of the queue: */
-		requeue_task_rt(rq, p);
-	}
+	p->time_slice = static_prio_timeslice(p->static_prio);
+	set_tsk_need_resched(p);
+
+	/* put it at the end of the queue: */
+	requeue_task_rt(rq, p);
 }
 
 /*

^ permalink raw reply	[flat|nested] 43+ messages in thread

* Re: [patch] CFS scheduler, -v18
  2007-06-26  8:38         ` Ingo Molnar
@ 2007-06-26  9:00           ` Andrew Morton
  2007-06-26  9:38             ` Ingo Molnar
  0 siblings, 1 reply; 43+ messages in thread
From: Andrew Morton @ 2007-06-26  9:00 UTC (permalink / raw)
  To: Ingo Molnar
  Cc: S.Çağlar Onur, linux-kernel, Linus Torvalds,
	Mike Galbraith, Arjan van de Ven, Thomas Gleixner,
	Dmitry Adamushko, Srivatsa Vaddagiri

On Tue, 26 Jun 2007 10:38:13 +0200 Ingo Molnar <mingo@elte.hu> wrote:

> 
> > - __exit_signal() does apparently-unlocked 64-bit arith.  Is there 
> >   some implicit locking here or do we not care about the occasional 
> >   race-induced inaccuracy?
> 
> do you mean the tsk->se.sum_exec_runtime addition, etc? That runs with 
> interrupts disabled so sum_sched_runtime is protected.
> 
> >   (ditto, lots of places, I expect)
> 
> which places do you mean?

I forget ;) There seemed to be rather a lot of 64-bit addition with no
obvious locking in sight, that's all.

> > ...
> >   (Gee, there's shitloads of 64-bit stuff in there.  Does it all 
> >   _really_ need to be 64-bit on 32-bit?)
>
> yes - CFS is fundamentally designed for 64-bit, with still pretty OK 
> arithmetics performance for 32-bit.

It may have been designed for 64-bit, but was that the correct design?  The
cost on 32-bit appears to be pretty high.  Perhaps a round of uninlining
will help.

> > - overall, CFS takes sched.o from 41157 of .text up to 48781 on x86_64,
> >   which at 18% is rather a large bloat.  Hopefully a lot of this is 
> >   the new debug stuff.
> 
> > - On i386 sched.o went from 33755 up to 43660 which is 29% growth. 
> >   Possibly acceptable, but why did it increase a lot more than the x86_64
> >   version?  All that 64-bit arith, I assume?
> 
> the main reason is the sched debugging stuff:

That would serve to explain the 18% growth on x86_64.  But why did i386
grow by much more: 29%?  I'd be suspecting all the new 64-bit arithmetic.



^ permalink raw reply	[flat|nested] 43+ messages in thread

* Re: [patch] CFS scheduler, -v18
  2007-06-26  9:00           ` Andrew Morton
@ 2007-06-26  9:38             ` Ingo Molnar
  0 siblings, 0 replies; 43+ messages in thread
From: Ingo Molnar @ 2007-06-26  9:38 UTC (permalink / raw)
  To: Andrew Morton
  Cc: S.??a??lar Onur, linux-kernel, Linus Torvalds, Mike Galbraith,
	Arjan van de Ven, Thomas Gleixner, Dmitry Adamushko,
	Srivatsa Vaddagiri


* Andrew Morton <akpm@linux-foundation.org> wrote:

> > the main reason is the sched debugging stuff:
> 
> That would serve to explain the 18% growth on x86_64.  But why did 
> i386 grow by much more: 29%?  I'd be suspecting all the new 64-bit 
> arithmetic.

this is what i see on 32-bit:

    text    data     bss     dec     hex filename
   28732    3905      24   32661    7f95 kernel/sched.o-vanilla
   37986    2538      20   40544    9e60 kernel/sched.o-v18
   31092    2426      20   33538    8302 kernel/sched.o-v18-no_sched_debug

text is larger but data got smaller. While they are not equivalent in 
function, the two almost even out each other (and that's without any of 
the uninlining that is in v19). In fact, there's a 1.5K per CPU percpu 
data size win with CFS, which is not visible in this stat. So on 
dual-core the net cost should already be zero.

> The cost on 32-bit appears to be pretty high.  Perhaps a round of 
> uninlining will help.

agreed, i've done one more round of uninlining.

	Ingo

^ permalink raw reply	[flat|nested] 43+ messages in thread

* RE: [patch] CFS scheduler, -v18
  2007-06-22 22:02 [patch] CFS scheduler, -v18 Ingo Molnar
                   ` (3 preceding siblings ...)
  2007-06-23 13:24 ` Willy Tarreau
@ 2007-06-26 20:17 ` Fortier,Vincent [Montreal]
  2007-06-27 10:51   ` Ingo Molnar
  2007-06-30 21:06 ` Willy Tarreau
  2007-07-02 11:44 ` Vegard Nossum
  6 siblings, 1 reply; 43+ messages in thread
From: Fortier,Vincent [Montreal] @ 2007-06-26 20:17 UTC (permalink / raw)
  To: Ingo Molnar, linux-kernel

> -----Message d'origine-----
> De : linux-kernel-owner@vger.kernel.org 
> [mailto:linux-kernel-owner@vger.kernel.org] De la part de Ingo Molnar
> Envoyé : 22 juin 2007 18:02
> 
> i'm pleased to announce release -v18 of the CFS scheduler patchset.
> 
> The rolled-up CFS patch against today's -git kernel, 
> v2.6.22-rc5, v2.6.22-rc4-mm2, v2.6.21.5 or v2.6.20.14 can be 
> downloaded from the usual place:
> 
>    http://people.redhat.com/mingo/cfs-scheduler/
> 
> ...

Pre-built kernels for Fedora 7 & Fedora Core 6 (yum reporsitory available) and Debian Sarge/Etch using latest CFS v18 available at http://linux-dev.qc.ec.gc.ca

> 
> As usual, any sort of feedback, bugreport, fix and suggestion 
> is more than welcome!
>    
> 	Ingo

- vin

^ permalink raw reply	[flat|nested] 43+ messages in thread

* Re: [patch] CFS scheduler, -v18
  2007-06-26 20:17 ` Fortier,Vincent [Montreal]
@ 2007-06-27 10:51   ` Ingo Molnar
  0 siblings, 0 replies; 43+ messages in thread
From: Ingo Molnar @ 2007-06-27 10:51 UTC (permalink / raw)
  To: Fortier,Vincent [Montreal]; +Cc: linux-kernel


* Fortier,Vincent [Montreal] <Vincent.Fortier1@EC.GC.CA> wrote:

> > -----Message d'origine-----
> > De : linux-kernel-owner@vger.kernel.org 
> > [mailto:linux-kernel-owner@vger.kernel.org] De la part de Ingo Molnar
> > Envoyé : 22 juin 2007 18:02
> > 
> > i'm pleased to announce release -v18 of the CFS scheduler patchset.
> > 
> > The rolled-up CFS patch against today's -git kernel, 
> > v2.6.22-rc5, v2.6.22-rc4-mm2, v2.6.21.5 or v2.6.20.14 can be 
> > downloaded from the usual place:
> > 
> >    http://people.redhat.com/mingo/cfs-scheduler/
> > 
> > ...
> 
> Pre-built kernels for Fedora 7 & Fedora Core 6 (yum reporsitory 
> available) and Debian Sarge/Etch using latest CFS v18 available at 
> http://linux-dev.qc.ec.gc.ca

thanks Vincent!

	Ingo

^ permalink raw reply	[flat|nested] 43+ messages in thread

* Re: [patch] CFS scheduler, -v18
  2007-06-22 22:02 [patch] CFS scheduler, -v18 Ingo Molnar
                   ` (4 preceding siblings ...)
  2007-06-26 20:17 ` Fortier,Vincent [Montreal]
@ 2007-06-30 21:06 ` Willy Tarreau
  2007-07-01  8:31   ` Ingo Molnar
  2007-07-02 11:44 ` Vegard Nossum
  6 siblings, 1 reply; 43+ messages in thread
From: Willy Tarreau @ 2007-06-30 21:06 UTC (permalink / raw)
  To: Ingo Molnar
  Cc: linux-kernel, Linus Torvalds, Andrew Morton, Mike Galbraith,
	Arjan van de Ven, Thomas Gleixner, Dmitry Adamushko,
	Srivatsa Vaddagiri

Ingo,

I've accidentally discovered a problem with -v18.

Some time ago, I wrote a small program to prevent my laptop from entering
low-power mode, and noticed that after upgrading my laptop's kernel from
2.4.20.9+cfs-v6 to 2.4.20.14+cfs-v18, it completely freezes if I run this
program.

The program is trivial, it just sets its prio to nice +20 and forks a busy
loop. I've added the ability to stop the loop after a user-definable number
of iterations, and I can confirm that it unfreezes when the loop ends. I'm
not even root when I run it.

Everything freezes, including the frame-buffer. It's not 100% reproducible, I
would say 90% only. Sometimes it requires a few seconds before freezing. It
*seems* to me that running another task in parallel (such as vmstat) increases
its chances to freeze. It seems like nicing to +19 does not cause any trouble.

I've tried it on my dual athlon, and with 1 process I see occasional pauses of
1 or 2 seconds, and with 2 processes, I see fairly larger pauses.

Here's the trivial program. Probably you'll find an obvious bug.

Regards,
Willy

---

/*
 * cfs-freeze.c
 * Fork a busy loop running with idle prio. This often results
 * in complete freezes with CFS-v18.
 *
 * $ gcc -O2 -s -o cfs-freeze cfs-freeze.c
 * $ ./cfs-freeze
 */

#include <stdio.h>
#include <stdlib.h>
#include <unistd.h>
#include <sched.h>
#include <sys/time.h>
#include <sys/resource.h>


int main(int argc, char **argv) {
	struct sched_param sch;
	long long i;

	if (argc > 1)
		i = atoll(argv[1]);

	if (i <= 0)
		i = 4 * 1000 * 1000 * 1000ULL;

	sch.sched_priority = 0;
	sched_setscheduler(0, SCHED_OTHER, &sch);
	setpriority(PRIO_PROCESS, 0, 20);
	if (fork() == 0)
		while (i--);
	return 0;
}

---


^ permalink raw reply	[flat|nested] 43+ messages in thread

* Re: [patch] CFS scheduler, -v18
  2007-06-30 21:06 ` Willy Tarreau
@ 2007-07-01  8:31   ` Ingo Molnar
  2007-07-01  8:45     ` Ingo Molnar
  0 siblings, 1 reply; 43+ messages in thread
From: Ingo Molnar @ 2007-07-01  8:31 UTC (permalink / raw)
  To: Willy Tarreau
  Cc: linux-kernel, Linus Torvalds, Andrew Morton, Mike Galbraith,
	Arjan van de Ven, Thomas Gleixner, Dmitry Adamushko,
	Srivatsa Vaddagiri


* Willy Tarreau <w@1wt.eu> wrote:

> Ingo,
> 
> I've accidentally discovered a problem with -v18.
> 
> Some time ago, I wrote a small program to prevent my laptop from 
> entering low-power mode, and noticed that after upgrading my laptop's 
> kernel from 2.4.20.9+cfs-v6 to 2.4.20.14+cfs-v18, it completely 
> freezes if I run this program.
> 
> The program is trivial, it just sets its prio to nice +20 and forks a 
> busy loop. I've added the ability to stop the loop after a 
> user-definable number of iterations, and I can confirm that it 
> unfreezes when the loop ends. I'm not even root when I run it.

hm, i tried your test-app and it causes no problems here. (which is not 
a surprise - your app starts a nice +19 busy loop, which is one of the 
common tests i do here too.)

To further debug this, could you try to create a 'high priority shell' 
on a text console (i.e. not under X) that is SCHED_FIFO prio 98? 
Something like:

  chrt -f -p 98 $$

should do the trick. And then run this script:

  http://people.redhat.com/mingo/cfs-scheduler/tools/cfs-debug-info.sh

before the test, during the test and after the test, from the high-prio 
shell session. (the shell runs at SCHED_FIFO, so the expectation would 
be for that it will be able to run during the test too) Then please send 
me the resulting 3 debug files. Thanks,

	Ingo

^ permalink raw reply	[flat|nested] 43+ messages in thread

* Re: [patch] CFS scheduler, -v18
  2007-07-01  8:31   ` Ingo Molnar
@ 2007-07-01  8:45     ` Ingo Molnar
  2007-07-01  9:00       ` Willy Tarreau
  0 siblings, 1 reply; 43+ messages in thread
From: Ingo Molnar @ 2007-07-01  8:45 UTC (permalink / raw)
  To: Willy Tarreau
  Cc: linux-kernel, Linus Torvalds, Andrew Morton, Mike Galbraith,
	Arjan van de Ven, Thomas Gleixner, Dmitry Adamushko,
	Srivatsa Vaddagiri


Willy,

could you check whether your current v18 CFS tree has the fix below 
included? I discovered it right after having released v18 so i updated 
the v18 files in place - but maybe you downloaded an early version? I 
thought it's relatively harmless, that it would only affect SCHED_IDLE 
tasks, but maybe it affects nice +19 tasks too on your box!

	Ingo

Index: linux/kernel/sched_fair.c
===================================================================
--- linux.orig/kernel/sched_fair.c
+++ linux/kernel/sched_fair.c
@@ -342,8 +342,9 @@ update_stats_enqueue(struct cfs_rq *cfs_
 		s64 tmp;
 
 		if (se->wait_runtime < 0) {
-			tmp = (0 - se->wait_runtime) << NICE_0_SHIFT;
-			key += (tmp * se->load.inv_weight) >> WMULT_SHIFT;
+			tmp = -se->wait_runtime;
+			key += (tmp * se->load.inv_weight) >>
+					(WMULT_SHIFT - NICE_0_SHIFT);
 		} else {
 			tmp = se->wait_runtime * se->load.weight;
 			key -= tmp >> NICE_0_SHIFT;

^ permalink raw reply	[flat|nested] 43+ messages in thread

* Re: [patch] CFS scheduler, -v18
  2007-07-01  8:45     ` Ingo Molnar
@ 2007-07-01  9:00       ` Willy Tarreau
  0 siblings, 0 replies; 43+ messages in thread
From: Willy Tarreau @ 2007-07-01  9:00 UTC (permalink / raw)
  To: Ingo Molnar
  Cc: linux-kernel, Linus Torvalds, Andrew Morton, Mike Galbraith,
	Arjan van de Ven, Thomas Gleixner, Dmitry Adamushko,
	Srivatsa Vaddagiri

Hi Ingo,

On Sun, Jul 01, 2007 at 10:45:01AM +0200, Ingo Molnar wrote:
> could you check whether your current v18 CFS tree has the fix below 
> included? I discovered it right after having released v18 so i updated 
> the v18 files in place - but maybe you downloaded an early version? I 
> thought it's relatively harmless, that it would only affect SCHED_IDLE 
> tasks, but maybe it affects nice +19 tasks too on your box!

Good catch, it was the cause of the problem. I've just applied
your fix below and rebuilt and the system behaves perfectly now.

Thanks very much !
Willy


^ permalink raw reply	[flat|nested] 43+ messages in thread

* Re: [patch] CFS scheduler, -v18
  2007-06-22 22:02 [patch] CFS scheduler, -v18 Ingo Molnar
                   ` (5 preceding siblings ...)
  2007-06-30 21:06 ` Willy Tarreau
@ 2007-07-02 11:44 ` Vegard Nossum
  2007-07-02 13:01   ` Dmitry Adamushko
                     ` (2 more replies)
  6 siblings, 3 replies; 43+ messages in thread
From: Vegard Nossum @ 2007-07-02 11:44 UTC (permalink / raw)
  To: Ingo Molnar; +Cc: linux-kernel

Hello,

On 6/23/07, Ingo Molnar <mingo@elte.hu> wrote:
> i'm pleased to announce release -v18 of the CFS scheduler patchset.

> As usual, any sort of feedback, bugreport, fix and suggestion is more
> than welcome!

I have been running cfs-v18 for a couple of days now, and today I
stumbled upon a rather strange problem. Consider the following short
program:

while(1)
        printf("%ld\r", 1000 * clock() / CLOCKS_PER_SEC);

Running this in an xterm makes the xterm totally unresponsive. Ctrl-C
takes about two seconds to terminate the program, during which the
program will keep running. In fact, it seems that the longer it runs,
the longer it takes to terminate (towards 5 seconds after running for
a couple of minutes). This is rather surprising, as the rest of the
system is quite responsive (even remarkably so). I think this is also
in contrast with the expected behaviour, that Ctrl-C/program
termination should be prioritized somehow.

Some other observations: X.Org seems to be running at about 75% CPU on
CPU 1, the xterm at about 45% on CPU 0, and a.out at about 20% on CPU
0. (HT processor)

Killing with -2 or -9 from another terminal works immediately. Ctrl-Z
takes the same time as Ctrl-C.

Another thing to note is that simply looping with no output retains
the expected responsiveness of the xterm. Printing i++ is somewhere
halfway in between.

Is this behaviour expected or even intended? My main point is that
Ctrl-C is a safety fallback which suddenly doesn't work as usual. I
might even go so far as to call it a regression.

I'd also like to point out that Folding@Home seems to draw more CPU
than it should. Or, at least, in top, it shows up as using 50% CPU
even though other processes are demanding as much as they can get. The
FAH program should be running with idle priority. I expect it to fall
to near 0% when other programs are running at full speed, but it keeps
trotting along. And I am pretty sure that this is not due to SMP/HT (I
made sure to utilize both CPUs).

Lastly, I'd like to mention that I got BUGs (soft lockups) with -v8,
though it has not been reproducible with -v18, so I suppose it must
have been fixed already.

Otherwise, I am satisfied with the performance of CFS. Especially the
desktop is noticably smoother. Thanks!

Kind regards,
Vegard Nossum

^ permalink raw reply	[flat|nested] 43+ messages in thread

* Re: [patch] CFS scheduler, -v18
  2007-07-02 11:44 ` Vegard Nossum
@ 2007-07-02 13:01   ` Dmitry Adamushko
  2007-07-02 13:43     ` Vegard Nossum
  2007-07-02 14:13   ` Bill Davidsen
  2007-07-03  7:15   ` Ingo Molnar
  2 siblings, 1 reply; 43+ messages in thread
From: Dmitry Adamushko @ 2007-07-02 13:01 UTC (permalink / raw)
  To: Vegard Nossum; +Cc: Ingo Molnar, Linux Kernel

On 02/07/07, Vegard Nossum <vegard.nossum@gmail.com> wrote:
> Hello,
>
> On 6/23/07, Ingo Molnar <mingo@elte.hu> wrote:
> > i'm pleased to announce release -v18 of the CFS scheduler patchset.
>
> > As usual, any sort of feedback, bugreport, fix and suggestion is more
> > than welcome!
>
> I have been running cfs-v18 for a couple of days now, and today I
> stumbled upon a rather strange problem. Consider the following short
> program:
>
> while(1)
>         printf("%ld\r", 1000 * clock() / CLOCKS_PER_SEC);

Is it running with a default (0) nice value?

could you please run the following script when your application is running?
As you have pointed out :

"... In fact, it seems that the longer it runs, the longer it takes to
terminate (towards 5 seconds after running for a couple of minutes
..." ,

please run the script a few times : say, before starting up your
application, 10 sec. after it's got started, 1 minute, a few minute...

http://people.redhat.com/mingo/cfs-scheduler/tools/cfs-debug-info.sh

then send us the resulting files. TIA,


-- 
Best regards,
Dmitry Adamushko

^ permalink raw reply	[flat|nested] 43+ messages in thread

* Re: [patch] CFS scheduler, -v18
  2007-07-02 13:01   ` Dmitry Adamushko
@ 2007-07-02 13:43     ` Vegard Nossum
  2007-07-02 15:50       ` Ingo Molnar
  0 siblings, 1 reply; 43+ messages in thread
From: Vegard Nossum @ 2007-07-02 13:43 UTC (permalink / raw)
  To: Dmitry Adamushko; +Cc: Ingo Molnar, Linux Kernel

On 7/2/07, Dmitry Adamushko <dmitry.adamushko@gmail.com> wrote:
> On 02/07/07, Vegard Nossum <vegard.nossum@gmail.com> wrote:
> > I have been running cfs-v18 for a couple of days now, and today I
> > stumbled upon a rather strange problem. Consider the following short
> > program:
> >
> > while(1)
> >         printf("%ld\r", 1000 * clock() / CLOCKS_PER_SEC);
>
> Is it running with a default (0) nice value?

Yes.

> could you please run the following script when your application is running?
> As you have pointed out :
>
> "... In fact, it seems that the longer it runs, the longer it takes to
> terminate (towards 5 seconds after running for a couple of minutes
> ..." ,

(I don't think it'll ever take more than 5 seconds, though.)

> please run the script a few times : say, before starting up your
> application, 10 sec. after it's got started, 1 minute, a few minute...
>
> http://people.redhat.com/mingo/cfs-scheduler/tools/cfs-debug-info.sh

Resulting files at
http://vegard.afraid.org:1104/pub/cfs/

cfs-debug-info-2007.07.02-15:18:13    Before running program
cfs-debug-info-2007.07.02-15:19:51    ~10 secs after start
cfs-debug-info-2007.07.02-15:20:54    ~1 minute after start
cfs-debug-info-2007.07.02-15:25:52    ~5 minutes after start
cfs-debug-info-2007.07.02-15:30:54    ~10 minutes after start

a.out is my program, FahCore_78 is the f@h client.

Hope this helps.

Vegard

^ permalink raw reply	[flat|nested] 43+ messages in thread

* Re: [patch] CFS scheduler, -v18
  2007-07-02 11:44 ` Vegard Nossum
  2007-07-02 13:01   ` Dmitry Adamushko
@ 2007-07-02 14:13   ` Bill Davidsen
  2007-07-03  7:15   ` Ingo Molnar
  2 siblings, 0 replies; 43+ messages in thread
From: Bill Davidsen @ 2007-07-02 14:13 UTC (permalink / raw)
  To: Vegard Nossum; +Cc: Ingo Molnar, linux-kernel

Vegard Nossum wrote:
> Hello,
> 
> On 6/23/07, Ingo Molnar <mingo@elte.hu> wrote:
>> i'm pleased to announce release -v18 of the CFS scheduler patchset.
> 
>> As usual, any sort of feedback, bugreport, fix and suggestion is more
>> than welcome!
> 
> I have been running cfs-v18 for a couple of days now, and today I
> stumbled upon a rather strange problem. Consider the following short
> program:
> 
> while(1)
>        printf("%ld\r", 1000 * clock() / CLOCKS_PER_SEC);
> 
> Running this in an xterm makes the xterm totally unresponsive. Ctrl-C
> takes about two seconds to terminate the program, during which the
> program will keep running. In fact, it seems that the longer it runs,
> the longer it takes to terminate (towards 5 seconds after running for
> a couple of minutes). This is rather surprising, as the rest of the
> system is quite responsive (even remarkably so). I think this is also
> in contrast with the expected behaviour, that Ctrl-C/program
> termination should be prioritized somehow.
> 
This sounds as though it might be related to the issues I see with my 
"glitch1" script, posted here a while ago. With cfs-v18 the effect of 
having multiple xterms scrolling is obvious, occasionally they behave as 
if they were "owed" more CPU and get paid back all at once. I've seen 
this effect to one degree or another since cfs-v13, which did NOT show 
the effect.

> Some other observations: X.Org seems to be running at about 75% CPU on
> CPU 1, the xterm at about 45% on CPU 0, and a.out at about 20% on CPU
> 0. (HT processor)
> 
> Killing with -2 or -9 from another terminal works immediately. Ctrl-Z
> takes the same time as Ctrl-C.
> 
I think this is because the shell to read the keypress is getting high 
latency, rather than the process taking a long time to react. I have 
been wrong before...

I read Ingo's reply to this, I'll gather the same information when the 
test machine is available later this morning and send it off to Ingo.

> Another thing to note is that simply looping with no output retains
> the expected responsiveness of the xterm. Printing i++ is somewhere
> halfway in between.
> 
See http://www.tmr.com/~public/source  (note the tilde) for glitch1.

> Is this behaviour expected or even intended? My main point is that
> Ctrl-C is a safety fallback which suddenly doesn't work as usual. I
> might even go so far as to call it a regression.
> 
> I'd also like to point out that Folding@Home seems to draw more CPU
> than it should. Or, at least, in top, it shows up as using 50% CPU
> even though other processes are demanding as much as they can get. The
> FAH program should be running with idle priority. I expect it to fall
> to near 0% when other programs are running at full speed, but it keeps
> trotting along. And I am pretty sure that this is not due to SMP/HT (I
> made sure to utilize both CPUs).
> 
> Lastly, I'd like to mention that I got BUGs (soft lockups) with -v8,
> though it has not been reproducible with -v18, so I suppose it must
> have been fixed already.
> 
> Otherwise, I am satisfied with the performance of CFS. Especially the
> desktop is noticably smoother. Thanks!
> 
> Kind regards,
> Vegard Nossum


-- 
Bill Davidsen <davidsen@tmr.com>
   "We have more to fear from the bungling of the incompetent than from
the machinations of the wicked."  - from Slashdot

^ permalink raw reply	[flat|nested] 43+ messages in thread

* Re: [patch] CFS scheduler, -v18
  2007-07-02 13:43     ` Vegard Nossum
@ 2007-07-02 15:50       ` Ingo Molnar
  2007-07-02 16:40         ` Vegard Nossum
  0 siblings, 1 reply; 43+ messages in thread
From: Ingo Molnar @ 2007-07-02 15:50 UTC (permalink / raw)
  To: Vegard Nossum; +Cc: Dmitry Adamushko, Linux Kernel


* Vegard Nossum <vegard.nossum@gmail.com> wrote:

> Resulting files at
> http://vegard.afraid.org:1104/pub/cfs/
> 
> cfs-debug-info-2007.07.02-15:18:13    Before running program
> cfs-debug-info-2007.07.02-15:19:51    ~10 secs after start
> cfs-debug-info-2007.07.02-15:20:54    ~1 minute after start
> cfs-debug-info-2007.07.02-15:25:52    ~5 minutes after start
> cfs-debug-info-2007.07.02-15:30:54    ~10 minutes after start
> 
> a.out is my program, FahCore_78 is the f@h client.

thx. As an initial matter, could you double-check whether your v18 
kernel source has the patch below applied already?

	Ingo

Index: linux/kernel/sched_fair.c
===================================================================
--- linux.orig/kernel/sched_fair.c
+++ linux/kernel/sched_fair.c
@@ -342,8 +342,9 @@ update_stats_enqueue(struct cfs_rq *cfs_
 		s64 tmp;
 
 		if (se->wait_runtime < 0) {
-			tmp = (0 - se->wait_runtime) << NICE_0_SHIFT;
-			key += (tmp * se->load.inv_weight) >> WMULT_SHIFT;
+			tmp = -se->wait_runtime;
+			key += (tmp * se->load.inv_weight) >>
+					(WMULT_SHIFT - NICE_0_SHIFT);
 		} else {
 			tmp = se->wait_runtime * se->load.weight;
 			key -= tmp >> NICE_0_SHIFT;


^ permalink raw reply	[flat|nested] 43+ messages in thread

* Re: [patch] CFS scheduler, -v18
  2007-07-02 15:50       ` Ingo Molnar
@ 2007-07-02 16:40         ` Vegard Nossum
  2007-07-02 18:13           ` Ingo Molnar
  2007-07-03  7:12           ` Mike Galbraith
  0 siblings, 2 replies; 43+ messages in thread
From: Vegard Nossum @ 2007-07-02 16:40 UTC (permalink / raw)
  To: Ingo Molnar; +Cc: Dmitry Adamushko, Linux Kernel

On 7/2/07, Ingo Molnar <mingo@elte.hu> wrote:
> thx. As an initial matter, could you double-check whether your v18
> kernel source has the patch below applied already?
>
>         Ingo
>
> Index: linux/kernel/sched_fair.c
> ===================================================================
> --- linux.orig/kernel/sched_fair.c
> +++ linux/kernel/sched_fair.c
> @@ -342,8 +342,9 @@ update_stats_enqueue(struct cfs_rq *cfs_
>                 s64 tmp;
>
>                 if (se->wait_runtime < 0) {
> -                       tmp = (0 - se->wait_runtime) << NICE_0_SHIFT;
> -                       key += (tmp * se->load.inv_weight) >> WMULT_SHIFT;
> +                       tmp = -se->wait_runtime;
> +                       key += (tmp * se->load.inv_weight) >>
> +                                       (WMULT_SHIFT - NICE_0_SHIFT);
>                 } else {
>                         tmp = se->wait_runtime * se->load.weight;
>                         key -= tmp >> NICE_0_SHIFT;
>
>

It does.

Vegard

^ permalink raw reply	[flat|nested] 43+ messages in thread

* Re: [patch] CFS scheduler, -v18
  2007-07-02 16:40         ` Vegard Nossum
@ 2007-07-02 18:13           ` Ingo Molnar
  2007-07-03  7:01             ` Vegard Nossum
  2007-07-03  7:12           ` Mike Galbraith
  1 sibling, 1 reply; 43+ messages in thread
From: Ingo Molnar @ 2007-07-02 18:13 UTC (permalink / raw)
  To: Vegard Nossum; +Cc: Dmitry Adamushko, Linux Kernel


* Vegard Nossum <vegard.nossum@gmail.com> wrote:

> > thx. As an initial matter, could you double-check whether your v18
> > kernel source has the patch below applied already?
> 
> It does.

ok. Does the xterm slowdown get any better if you do:

  echo 46 > /proc/sys/kernel/sched_features

? The default on v18 is:

  echo 14 > /proc/sys/kernel/sched_features

	Ingo

^ permalink raw reply	[flat|nested] 43+ messages in thread

* Re: [patch] CFS scheduler, -v18
  2007-07-02 18:13           ` Ingo Molnar
@ 2007-07-03  7:01             ` Vegard Nossum
  0 siblings, 0 replies; 43+ messages in thread
From: Vegard Nossum @ 2007-07-03  7:01 UTC (permalink / raw)
  To: Ingo Molnar; +Cc: Dmitry Adamushko, Linux Kernel

On 7/2/07, Ingo Molnar <mingo@elte.hu> wrote:
> ok. Does the xterm slowdown get any better if you do:
>
>   echo 46 > /proc/sys/kernel/sched_features
>
> ? The default on v18 is:
>
>   echo 14 > /proc/sys/kernel/sched_features

No. The Ctrl-C still hangs between 1 and 3 seconds, again seemingly
depending on how long the program runs before I terminate it.

Vegard

^ permalink raw reply	[flat|nested] 43+ messages in thread

* Re: [patch] CFS scheduler, -v18
  2007-07-02 16:40         ` Vegard Nossum
  2007-07-02 18:13           ` Ingo Molnar
@ 2007-07-03  7:12           ` Mike Galbraith
  2007-07-03  7:22             ` Ingo Molnar
  1 sibling, 1 reply; 43+ messages in thread
From: Mike Galbraith @ 2007-07-03  7:12 UTC (permalink / raw)
  To: Vegard Nossum; +Cc: Ingo Molnar, Dmitry Adamushko, Linux Kernel

On Mon, 2007-07-02 at 18:40 +0200, Vegard Nossum wrote:
> On 7/2/07, Ingo Molnar <mingo@elte.hu> wrote:
> > thx. As an initial matter, could you double-check whether your v18
> > kernel source has the patch below applied already?
> >
> >         Ingo
> >
> > Index: linux/kernel/sched_fair.c
> > ===================================================================
> > --- linux.orig/kernel/sched_fair.c
> > +++ linux/kernel/sched_fair.c
> > @@ -342,8 +342,9 @@ update_stats_enqueue(struct cfs_rq *cfs_
> >                 s64 tmp;
> >
> >                 if (se->wait_runtime < 0) {
> > -                       tmp = (0 - se->wait_runtime) << NICE_0_SHIFT;
> > -                       key += (tmp * se->load.inv_weight) >> WMULT_SHIFT;
> > +                       tmp = -se->wait_runtime;
> > +                       key += (tmp * se->load.inv_weight) >>
> > +                                       (WMULT_SHIFT - NICE_0_SHIFT);
> >                 } else {
> >                         tmp = se->wait_runtime * se->load.weight;
> >                         key -= tmp >> NICE_0_SHIFT;
> >
> >
> 
> It does.

Hi,

This doesn't appear to be a CFS problem.  I can reproduce the problem
easily in virgin 2.6.22-rc7 by starting xterm-spam at nice -1 or better.
As soon as xterm-spam can get enough CPU to keep the xterm fully busy,
it's game over, the xterm freezes.  The more accurate fairness of CFS to
sleepers just tips the balance quicker.  In mainline, the xterm has an
unfair advantage and maintains it indefinitely... until you tip the
scales just a wee bit, at which time it inverts.

	-Mike


^ permalink raw reply	[flat|nested] 43+ messages in thread

* Re: [patch] CFS scheduler, -v18
  2007-07-02 11:44 ` Vegard Nossum
  2007-07-02 13:01   ` Dmitry Adamushko
  2007-07-02 14:13   ` Bill Davidsen
@ 2007-07-03  7:15   ` Ingo Molnar
  2007-07-03  9:11     ` Vegard Nossum
  2 siblings, 1 reply; 43+ messages in thread
From: Ingo Molnar @ 2007-07-03  7:15 UTC (permalink / raw)
  To: Vegard Nossum; +Cc: linux-kernel


* Vegard Nossum <vegard.nossum@gmail.com> wrote:

> I'd also like to point out that Folding@Home seems to draw more CPU 
> than it should. Or, at least, in top, it shows up as using 50% CPU 
> even though other processes are demanding as much as they can get. The 
> FAH program should be running with idle priority. I expect it to fall 
> to near 0% when other programs are running at full speed, but it keeps 
> trotting along. And I am pretty sure that this is not due to SMP/HT (I 
> made sure to utilize both CPUs).

your FAH processes are running with nice +19 - that should be enough to 
throttle them. With v18 you can also mark it SCHED_IDLE:

    schedtool -D $$  # mark the shell idle

SCHED_IDLE gets inherited by child tasks so if you mark the shell that 
starts up FEH as SCHED_IDLE, all FEH threads should be SCHED_IDLE too. 
(or you can start it up via schedtool -D -e ... )

does it still get more CPU time than you'd expect it to get? A reniced 
or SCHED_IDLE task will 'fill in' any idle time that it senses, so in 
itself it's not an anomaly if a task gets 50% and FEH fills in the 
remaining 50%. Does it still get CPU time if you start two CPU hogs:

   for (( N=0; N < 2; N++ )); do ( while :; do :; done ) & done

? If yes, that would be an anomaly.

> Otherwise, I am satisfied with the performance of CFS. Especially the 
> desktop is noticably smoother. Thanks!

great! :-)

	Ingo

^ permalink raw reply	[flat|nested] 43+ messages in thread

* Re: [patch] CFS scheduler, -v18
  2007-07-03  7:12           ` Mike Galbraith
@ 2007-07-03  7:22             ` Ingo Molnar
  2007-07-03  8:08               ` Keith Packard
  2007-07-04 12:11               ` Andi Kleen
  0 siblings, 2 replies; 43+ messages in thread
From: Ingo Molnar @ 2007-07-03  7:22 UTC (permalink / raw)
  To: Mike Galbraith
  Cc: Vegard Nossum, Dmitry Adamushko, Linux Kernel, Keith Packard

[-- Attachment #1: Type: text/plain, Size: 1214 bytes --]


* Mike Galbraith <efault@gmx.de> wrote:

> This doesn't appear to be a CFS problem.  I can reproduce the problem 
> easily in virgin 2.6.22-rc7 by starting xterm-spam at nice -1 or 
> better. As soon as xterm-spam can get enough CPU to keep the xterm 
> fully busy, it's game over, the xterm freezes.  The more accurate 
> fairness of CFS to sleepers just tips the balance quicker.  In 
> mainline, the xterm has an unfair advantage and maintains it 
> indefinitely... until you tip the scales just a wee bit, at which time 
> it inverts.

ah. That indeed makes sense. It seems like the xterm doesnt process the 
Ctrl-C/Z keypresses _at all_ when it is 'spammed' with output. Normally, 
output 'spam' is throttled by the scroll buffer's overhead. But in 
Vegard's case, the printout involves a \r carriage return:

       printf("%ld\r", 1000 * clock() / CLOCKS_PER_SEC);

which allows xterm-spam (attached) to easily flood the xterm (without 
any scrolling that would act as a throttle) and the xterm to flood Xorg. 

I suspect we need the help of an xterm/Xorg expert? (maybe Keith can 
give us further pointers? I can reproduce the problem on a T60 with i940 
and Core2Duo running Fedora 7 + Xorg 7.1.)

	Ingo

[-- Attachment #2: xterm-spam.c --]
[-- Type: text/plain, Size: 101 bytes --]


#include <time.h>

main ()
{
   while(1)
       printf("%ld\r", 1000 * clock() / CLOCKS_PER_SEC);
}

^ permalink raw reply	[flat|nested] 43+ messages in thread

* Re: [patch] CFS scheduler, -v18
  2007-07-03  7:22             ` Ingo Molnar
@ 2007-07-03  8:08               ` Keith Packard
  2007-07-03  8:31                 ` Ingo Molnar
  2007-07-04 12:11               ` Andi Kleen
  1 sibling, 1 reply; 43+ messages in thread
From: Keith Packard @ 2007-07-03  8:08 UTC (permalink / raw)
  To: Ingo Molnar
  Cc: keith.packard, Mike Galbraith, Vegard Nossum, Dmitry Adamushko,
	Linux Kernel

[-- Attachment #1: Type: text/plain, Size: 1439 bytes --]

On Tue, 2007-07-03 at 09:22 +0200, Ingo Molnar wrote:

> which allows xterm-spam (attached) to easily flood the xterm (without 
> any scrolling that would act as a throttle) and the xterm to flood Xorg. 

It's just an Xterm bug.

Xterm will look for X input if it ever manages to fill the input buffer
past 32768 bytes. If it manages to get more than 4096 bytes in one read,
it will invoke sched_yield. and then check for input. Gotta love that
sched_yield call.

As it always processes all of the incoming data before trying to read
again, there doesn't appear to be any way it can ever have more than
32768 characters in the buffer. 

And, as the kernel will not buffer more than 4095 bytes from a pty,
there isn't any way it will ever read 4096 bytes.

So, it sits there carefully reading every byte from the pty and painting
them on the screen.

You can 'fix' xterm with:

$ xterm -xrm '*minBufSize: 4095'

I hesitate to even suggest a patch to xterm that would solve this
problem correctly. Note that xterm has kludges in several of the output
processing steps which explicitly look for input (most vertical cursor
motion, it seems), which is why any application which scrolls doesn't
cause this problem.

Do you need more reasons to switch to another terminal emulator?
gnome-terminal has finally gotten reasonable; I expect rxvt or konsole
would work just as well.

-- 
keith.packard@intel.com

[-- Attachment #2: This is a digitally signed message part --]
[-- Type: application/pgp-signature, Size: 189 bytes --]

^ permalink raw reply	[flat|nested] 43+ messages in thread

* Re: [patch] CFS scheduler, -v18
  2007-07-03  8:08               ` Keith Packard
@ 2007-07-03  8:31                 ` Ingo Molnar
  0 siblings, 0 replies; 43+ messages in thread
From: Ingo Molnar @ 2007-07-03  8:31 UTC (permalink / raw)
  To: Keith Packard
  Cc: Mike Galbraith, Vegard Nossum, Dmitry Adamushko, Linux Kernel


* Keith Packard <keith.packard@intel.com> wrote:

> On Tue, 2007-07-03 at 09:22 +0200, Ingo Molnar wrote:
> 
> > which allows xterm-spam (attached) to easily flood the xterm 
> > (without any scrolling that would act as a throttle) and the xterm 
> > to flood Xorg.
> 
> It's just an Xterm bug.
> 
> Xterm will look for X input if it ever manages to fill the input 
> buffer past 32768 bytes. If it manages to get more than 4096 bytes in 
> one read, it will invoke sched_yield. and then check for input. Gotta 
> love that sched_yield call.
> 
> As it always processes all of the incoming data before trying to read 
> again, there doesn't appear to be any way it can ever have more than 
> 32768 characters in the buffer.
> 
> And, as the kernel will not buffer more than 4095 bytes from a pty, 
> there isn't any way it will ever read 4096 bytes.
> 
> So, it sits there carefully reading every byte from the pty and 
> painting them on the screen.

ah. Thanks for the explanation!

> You can 'fix' xterm with:
> 
> $ xterm -xrm '*minBufSize: 4095'

indeed, that solves the xterm-spam Ctrl-C/Ctrl-Z problem here!

> I hesitate to even suggest a patch to xterm that would solve this 
> problem correctly. Note that xterm has kludges in several of the 
> output processing steps which explicitly look for input (most vertical 
> cursor motion, it seems), which is why any application which scrolls 
> doesn't cause this problem.
> 
> Do you need more reasons to switch to another terminal emulator? 
> gnome-terminal has finally gotten reasonable; I expect rxvt or konsole 
> would work just as well.

yeah, i use gnome-terminal exclusively. But testers looking for CFS 
regressions do run every shell on the planet :-)

gnome-terminal is also faster all around (at least on my box):

 $ (echo '#!/bin/bash' ; echo 'for ((i=0; i<100000; i++)); do echo $i; done') > 1.sh; chmod +x 1.sh; time xterm $HOME/1.sh; time gnome-terminal  -x ./1.sh

 real    0m3.193s
 user    0m2.840s
 sys     0m0.460s

 real    0m2.495s
 user    0m2.430s
 sys     0m1.520s

	Ingo

^ permalink raw reply	[flat|nested] 43+ messages in thread

* Re: [patch] CFS scheduler, -v18
  2007-07-03  7:15   ` Ingo Molnar
@ 2007-07-03  9:11     ` Vegard Nossum
  0 siblings, 0 replies; 43+ messages in thread
From: Vegard Nossum @ 2007-07-03  9:11 UTC (permalink / raw)
  To: Ingo Molnar; +Cc: linux-kernel

On 7/3/07, Ingo Molnar <mingo@elte.hu> wrote:
> does it still get more CPU time than you'd expect it to get? A reniced
> or SCHED_IDLE task will 'fill in' any idle time that it senses, so in
> itself it's not an anomaly if a task gets 50% and FEH fills in the
> remaining 50%. Does it still get CPU time if you start two CPU hogs:
>
>    for (( N=0; N < 2; N++ )); do ( while :; do :; done ) & done
>
> ? If yes, that would be an anomaly.

No, with this test, it gets the expected amount of CPU. Thanks for
clearing it up!

Vegard

^ permalink raw reply	[flat|nested] 43+ messages in thread

* Re: [patch] CFS scheduler, -v18
  2007-07-03  7:22             ` Ingo Molnar
  2007-07-03  8:08               ` Keith Packard
@ 2007-07-04 12:11               ` Andi Kleen
  1 sibling, 0 replies; 43+ messages in thread
From: Andi Kleen @ 2007-07-04 12:11 UTC (permalink / raw)
  To: Ingo Molnar
  Cc: Mike Galbraith, Vegard Nossum, Dmitry Adamushko, Linux Kernel,
	Keith Packard

Ingo Molnar <mingo@elte.hu> writes:
> 
> ah. That indeed makes sense. It seems like the xterm doesnt process the 
> Ctrl-C/Z keypresses _at all_ when it is 'spammed' with output. Normally, 
> output 'spam' is throttled by the scroll buffer's overhead. But in 
> Vegard's case, the printout involves a \r carriage return:
> 
>        printf("%ld\r", 1000 * clock() / CLOCKS_PER_SEC);
> 
> which allows xterm-spam (attached) to easily flood the xterm (without 
> any scrolling that would act as a throttle) and the xterm to flood Xorg. 
> 
> I suspect we need the help of an xterm/Xorg expert? (maybe Keith can 
> give us further pointers? I can reproduce the problem on a T60 with i940 
> and Core2Duo running Fedora 7 + Xorg 7.1.)

Xorg seems to have a couple of starvation issues. e.g. I found
the Gantt view in icemon during a busy compile session can starve all
other X clients for tenths of seconds.

-Andi

^ permalink raw reply	[flat|nested] 43+ messages in thread

* Re: [patch] CFS scheduler, -v18
       [not found]                 ` <8CM5x-19K-3@gated-at.bofh.it>
@ 2007-07-05 13:39                   ` Thomas Dickey
  0 siblings, 0 replies; 43+ messages in thread
From: Thomas Dickey @ 2007-07-05 13:39 UTC (permalink / raw)
  To: Linux Kernel; +Cc: Ingo Molnar

[-- Attachment #1: Type: text/plain, Size: 782 bytes --]

On Tue, Jul 03, 2007 at 10:40:07AM +0200, Ingo Molnar wrote:
> yeah, i use gnome-terminal exclusively. But testers looking for CFS 
> regressions do run every shell on the planet :-)

...and people running older kernels get different results (no surprise)

fwiw, I ran 'top' on 5 terminals with xterm-spam running concurrently
on 2.6.15 (rxvt, pterm, xterm, konsole and gnome-terminal).

For that case, gnome-terminal was definitely the slowest,
and used the most CPU time (more than a factor of three slower
than xterm).  konsole was about 2.5, pterm was about the same
as xterm, and rxvt about half the CPU (ymmv).
> 
> gnome-terminal is also faster all around (at least on my box):

-- 
Thomas E. Dickey
http://invisible-island.net
ftp://invisible-island.net

[-- Attachment #2: Type: application/pgp-signature, Size: 228 bytes --]

^ permalink raw reply	[flat|nested] 43+ messages in thread

* Re: [patch] CFS scheduler, -v18
       [not found]               ` <8CLCt-vX-3@gated-at.bofh.it>
@ 2007-07-05 13:29                 ` Thomas Dickey
       [not found]                 ` <8CM5x-19K-3@gated-at.bofh.it>
  1 sibling, 0 replies; 43+ messages in thread
From: Thomas Dickey @ 2007-07-05 13:29 UTC (permalink / raw)
  To: Linux Kernel; +Cc: Keith Packard

[-- Attachment #1: Type: text/plain, Size: 829 bytes --]

On Tue, Jul 03, 2007 at 10:10:05AM +0200, Keith Packard wrote:
> On Tue, 2007-07-03 at 09:22 +0200, Ingo Molnar wrote:
> 
> > which allows xterm-spam (attached) to easily flood the xterm (without 
> > any scrolling that would act as a throttle) and the xterm to flood Xorg. 

...deleting the offhand remarks

...here's a little content:

> And, as the kernel will not buffer more than 4095 bytes from a pty,
> there isn't any way it will ever read 4096 bytes.

The sched_yield() call was added to address problems in an earlier
2.6.x kernel (and the report then stated that it could read more than
4096 because it would have to perform several reads to keep up with
a burst of output).
 
...deleting the other offhand remarks


-- 
Thomas E. Dickey
http://invisible-island.net
ftp://invisible-island.net

[-- Attachment #2: Type: application/pgp-signature, Size: 228 bytes --]

^ permalink raw reply	[flat|nested] 43+ messages in thread

end of thread, other threads:[~2007-07-05 13:58 UTC | newest]

Thread overview: 43+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2007-06-22 22:02 [patch] CFS scheduler, -v18 Ingo Molnar
2007-06-22 22:09 ` S.Çağlar Onur
2007-06-22 22:16   ` S.Çağlar Onur
2007-06-22 22:20     ` Ingo Molnar
2007-06-26  3:02       ` Andrew Morton
2007-06-26  8:38         ` Ingo Molnar
2007-06-26  9:00           ` Andrew Morton
2007-06-26  9:38             ` Ingo Molnar
2007-06-22 23:08 ` Gene Heskett
2007-06-23  7:11   ` Ingo Molnar
2007-06-23  9:55     ` Gene Heskett
2007-06-23 10:22 ` Antonino Ingargiola
2007-06-23 17:25   ` Ingo Molnar
2007-06-24 10:02     ` Antonino Ingargiola
2007-06-24 11:07       ` Ingo Molnar
2007-06-25  7:27         ` Antonino Ingargiola
2007-06-23 13:24 ` Willy Tarreau
2007-06-24 15:52   ` Ingo Molnar
2007-06-24 17:08     ` Willy Tarreau
2007-06-24 20:31       ` Ingo Molnar
2007-06-26 20:17 ` Fortier,Vincent [Montreal]
2007-06-27 10:51   ` Ingo Molnar
2007-06-30 21:06 ` Willy Tarreau
2007-07-01  8:31   ` Ingo Molnar
2007-07-01  8:45     ` Ingo Molnar
2007-07-01  9:00       ` Willy Tarreau
2007-07-02 11:44 ` Vegard Nossum
2007-07-02 13:01   ` Dmitry Adamushko
2007-07-02 13:43     ` Vegard Nossum
2007-07-02 15:50       ` Ingo Molnar
2007-07-02 16:40         ` Vegard Nossum
2007-07-02 18:13           ` Ingo Molnar
2007-07-03  7:01             ` Vegard Nossum
2007-07-03  7:12           ` Mike Galbraith
2007-07-03  7:22             ` Ingo Molnar
2007-07-03  8:08               ` Keith Packard
2007-07-03  8:31                 ` Ingo Molnar
2007-07-04 12:11               ` Andi Kleen
2007-07-02 14:13   ` Bill Davidsen
2007-07-03  7:15   ` Ingo Molnar
2007-07-03  9:11     ` Vegard Nossum
     [not found] <8yZun-1bO-5@gated-at.bofh.it>
     [not found] ` <8CszT-4hd-11@gated-at.bofh.it>
     [not found]   ` <8CtPi-6qj-9@gated-at.bofh.it>
     [not found]     ` <8Cus1-7eL-11@gated-at.bofh.it>
     [not found]       ` <8CwtU-1Z5-3@gated-at.bofh.it>
     [not found]         ` <8Cxgd-3fN-9@gated-at.bofh.it>
     [not found]           ` <8CKQ8-7HN-13@gated-at.bofh.it>
     [not found]             ` <8CKZV-7Ur-11@gated-at.bofh.it>
     [not found]               ` <8CLCt-vX-3@gated-at.bofh.it>
2007-07-05 13:29                 ` Thomas Dickey
     [not found]                 ` <8CM5x-19K-3@gated-at.bofh.it>
2007-07-05 13:39                   ` Thomas Dickey

This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).