DEV Community

Satoru Takeuchi
Satoru Takeuchi

Posted on

Visualize the Linux kernel's behavior: process scheduler

Preface

This article is to visualize the behavior of the Linux kernel. Specifically, it visualize the behavior of Linux's process scheduler.

Many people know the concept of round-robin scheduling, but, how many people actually confirmed that processes are really scheduled as this way? You can confirm it with fancy kernel features like ftrace. However, this article does it with a simple userland program, without the deep knowledge about OS kernels.

The Measurement Program

This article uses a measurement program, called sched. This program launches the one ore more worker processes and waits for the termination of worker processes. Each worker processes consumes some CPU time and exit after that.

This specification of this program is as follows.

  • sched nproc total resolution
  • The meaning of the command line arguments is as follows.
    • nproc(1st argument): the number of processes running in parallel
    • total(2nd argument): the CPU time which each process consumes in milliseconds
    • resolution(3rd argument): the measurement interval in milliseconds
  • The meaning of the output of this program is as follows.
    • One entry per one measurement point
    • The meaning of each measurement point is as follows.
    • 1st field: the ID of each worker thread
    • 2nd field: elapse time from the start time
    • 3rd field: the progress of worker thread in percentage, in other words, already consumed CPU time/<total>

Here is the source code.

#include <sys/types.h>
#include <sys/wait.h>
#include <time.h>
#include <unistd.h>
#include <stdio.h>
#include <stdlib.h>
#include <string.h>
#include <err.h>

#define NLOOP_FOR_ESTIMATION 1000000000UL
#define NSECS_PER_MSEC 1000000UL
#define NSECS_PER_SEC 1000000000UL

static inline long diff_nsec(struct timespec before, struct timespec after)
{
        return ((after.tv_sec * NSECS_PER_SEC + after.tv_nsec)
                - (before.tv_sec * NSECS_PER_SEC + before.tv_nsec));

}

static unsigned long loops_per_msec()
{
        struct timespec before, after;
        clock_gettime(CLOCK_MONOTONIC, &before);

        unsigned long i;
        for (i = 0; i < NLOOP_FOR_ESTIMATION; i++)
        ;

        clock_gettime(CLOCK_MONOTONIC, &after);

    int ret;
        return  NLOOP_FOR_ESTIMATION * NSECS_PER_MSEC / diff_nsec(before, after);
}

static inline void load(unsigned long nloop)
{
        unsigned long i;
        for (i = 0; i < nloop; i++)
                ;
}

static void child_fn(int id, struct timespec *buf, int nrecord, unsigned long nloop_per_resol, struct timespec start)
{
        int i;
        for (i = 0; i < nrecord; i++) {
                struct timespec ts;

                load(nloop_per_resol);
                clock_gettime(CLOCK_MONOTONIC, &ts);
                buf[i] = ts;
        }
        for (i = 0; i < nrecord; i++) {
                printf("%d\t%ld\t%d\n", id, diff_nsec(start, buf[i]) / NSECS_PER_MSEC, (i + 1) * 100 / nrecord);
        }
        exit(EXIT_SUCCESS);
}

static void parent_fn(int nproc)
{
        int i;
        for (i = 0; i < nproc; i++)
                wait(NULL);
}

static pid_t *pids;

int main(int argc, char *argv[])
{
        int ret = EXIT_FAILURE;

        if (argc < 4) {
                fprintf(stderr, "usage: %s <nproc> <total[ms]> <resolution[ms]>\n", argv[0]);
                exit(EXIT_FAILURE);
        }

        int nproc = atoi(argv[1]);
        int total = atoi(argv[2]);
        int resol = atoi(argv[3]);

        if (nproc < 1) {
                fprintf(stderr, "<nproc>(%d) should be >= 1\n", nproc);
                exit(EXIT_FAILURE);
        }

        if (total < 1) {
                fprintf(stderr, "<total>(%d) should be >= 1\n", total);
                exit(EXIT_FAILURE);
        }

        if (resol < 1) {
                fprintf(stderr, "<resol>(%d) should be >= 1\n", resol);
                exit(EXIT_FAILURE);
        }

        if (total % resol) {
                fprintf(stderr, "<total>(%d) should be multiple of <resolution>(%d)\n", total, resol);
                exit(EXIT_FAILURE);
        }
        int nrecord = total / resol;

        struct timespec *logbuf = malloc(nrecord * sizeof(struct timespec));
    if (!logbuf)
        err(EXIT_FAILURE, "malloc(logbuf) failed");

    puts("estimating workload which takes just one milisecond");
        unsigned long nloop_per_resol = loops_per_msec() * resol;
    puts("end estimation");
    fflush(stdout);

        pids = malloc(nproc * sizeof(pid_t));
        if (pids == NULL) {
                warn("malloc(pids) failed");
                goto free_logbuf;
        }

        struct timespec start;
        clock_gettime(CLOCK_MONOTONIC, &start);

        int i, ncreated;
        for (i = 0, ncreated = 0; i < nproc; i++, ncreated++) {
                pids[i] = fork();
                if (pids[i] < 0) {
                        goto wait_children;
                } else if (pids[i] == 0) {
                        // children

                        child_fn(i, logbuf, nrecord, nloop_per_resol, start);
                        /* shouldn't reach here */
                }
        }
        ret = EXIT_SUCCESS;

        // parent

wait_children:
        if (ret == EXIT_FAILURE)
                for (i = 0; i < ncreated; i++)
                        if (kill(pids[i], SIGINT) < 0)
                                warn("kill(%d) failed", pids[i]);

        for (i = 0; i < ncreated; i++)
                if (wait(NULL) < 0)
                        warn("wait() failed.");

free_pids:
        free(pids);

free_logbuf:
        free(logbuf);

        exit(ret);
}

How to measure

I used the following bash script named capture to get the data.

#!/bin/bash 

NCPU=$(grep -c processor /proc/cpuinfo) 
LASTCPU=$(($NCPU - 1)) 
PATTERN="1 2 3 4" 
TOTAL_US=1000
RESOL_US=10

for i in ${PATTERN} ; do 
    taskset --cpu-list ${LASTCPU} ./sched $i ${TOTAL_US} ${RESOL_US} >$i.txt 
done

With this script, I got the data of the following conditions.

  • the number of CPUs which sched runs on: 1
  • the arguments of sched
    • nproc: 1, 2, 3, 4
    • total: 1000
    • resolution: 10

I run configure as follows.

$ make sched
$ ./capture

Hardware/Software Environment

  • CPU: Intel(R) Core(TM) i5-2400 CPU @ 3.10GHz x 1 (4 core 1 thread)
  • OS: Debian GNU/Linux stretch
  • kernel: 4.8.0-1-amd64

The Measurement Result

I make the two graphs for each nproc as follows.

  • graph1: What worker thread running on CPU

    • x-axis: the elapsed time from the start time in microseconds
    • y-axis: the ID of each worker threads
  • graph2: the progress of worker threads

    • x-axis: the same as the graph1
    • y-axis: the progress of worker threads. 0 means 0 percent and 1 means 100 percent.

nproc == 1

  • graph1

Alt Text

  • graph2

Alt Text

nproc == 2

  • graph1

Alt Text

  • graph2

Alt Text

nproc == 3

  • graph1

Alt Text

  • graph2

Alt Text

nproc == 4

  • graph1

Alt Text

  • graph2

Alt Text

Consideration

You found the following things from the measurement result.

  • Only one worker process can run at the same time
  • each worker thread runs as round-robin way

Conclusion

By this article, you found Linux schedules processes in a round-robin way. If you're interested in the result, please confirm the following things.

  • How the timeslice of each worker processes are changed if nproc is changed
  • How about the result if you run sched without pinning each worker threads to only one CPU.
  • How the timeslice of each worker processes are changed if the nice value of them are changed

Top comments (0)