.-- .. - .... .-.. --- ...- .

home archive about

Linux kernel overview

09 Dec 2021

An overview and a study road map of operating systems and the Linux kernel.

The operating system

An operating system is a piece of software that controls the computer hardware, be it a desktop, a mobile phone, or an embedded device. It manages the hardware, and provides interfaces for the applications to use the hardware.


Hardware is at the base of a computer system.

We have CPU that runs the computation tasks, and memory that stores the data. These are the main computation resources of a computer system. Apart from CPU and memory, we also have peripheral devices, such as disk controllers, network controllers, keyboards, and so on.

Operating system

Operating system architecture:

Operating System Architecture


Service daemons, compiler, editor, browser, and other user applications.


A command-line user interface (CLI) or a graphical user interface (GUI) for access to the operating system's services.

Sh, Bash, Zsh, Gnome, etc.

Library routines

GNU C Library (glibc), Basic Mathematical Functions (libm), etc.

Applications, shell, and library routines are user space components. From the perspective of the kernel, they are all upper level user processes that the kernel manages.


The kernel is the core component of the operating system. It resides in the memory after the computer is booted, and tells the CPU what is the next task.

Kernel runs in privileged mode and has unrestricted access to the hardware. Kernel has its own memory space called the kernel space. User processes runs in user mode, and have restricted access to the hardware, such as their own memory and a subset of safe CPU operations.

The kernel is the primary interface between the hardware and the user processes. Processes normally use system calls to communicate with the kernel.

System calls

System calls are the programming interface between the user process and the kernel. It transfers the control from unprivileged user process to privileged kernel process via a software interrupt.

System calls can be divided into 5 categories:

What does the kernel do

The kernel is organized into a number of modules. The modules organization reflects the kernel's responsibilities.

Process management

The kernel is responsible for determining the next task to be executed (CPU scheduling). It controls process creation and termination, signal handling, etc.

Process management includes the starting, pausing, resuming, scheduling, and terminating of processes. In modern computer, processes run simultaneously by context switch. Each process use the CPU for a small fraction of time (a time slice), then pauses (interrupted by the CPU) and gives up the CPU, so that the kernel can choose another process to run for another time slice. Because the time slices are so small that humans cannot perceive them, the system appears to be multitasking.

During the context switch, the kernel saves the state of the current process, performs kernel tasks, such as I/O operations, that came up during the preceding time slice, then prepares memory and CPU for the next process, and switches the CPU to the next process.

Memory management

The kernel keeps track of all memory allocations, including which process owns which memory, and which memory is free. It is responsible for page allocation, page fault handling, virtual memory management, etc.

The kernel has its own kernel space memory that only the kernel threads can access. Each user process has its own user space memory, and one user process cannot access the unshared memory of another user process. Modern operating system also uses virtual memory (swap) to provide more memory than the physical memory.

I/O management

It is usually the kernel's job to operate the hardware, and it acts as an interface between the user processes and the I/O devices. The kernel provides an abstraction layer for hardware devices, such as the disk, the network interface, and the user I/O devices, in the form of file system, sockets, network protocols, and I/O controllers. I/O events generate interrupts to the CPU, which then calls the kernel to perform the I/O operation.

Linux kernel modules

Modern Linux kernel are modular kernel. You can add new device drivers, filesystem drivers, network drivers, system calls, or other kernel modules without the need of rebooting.

Kernel booting process

  1. BIOS/UEFI reads MBR/GPT to find and run a boot loader (such as GRUB).
  2. The boot loader core initializes to access disks and file systems.
  3. The boot loader finds the kernel image (vmlinuz) on the disk, loads it to the main memory, and runs the kernel.
  4. The kernel inspects and initializes the devices and drivers.
  5. The kernel mounts the root file system (initrd.img).
  6. The kernel starts the init process (process ID 1). From this point the user space starts.
  7. init starts other processes (fork() and exec()). User space processes are usually managed by a service management service, such as systemd.
  8. init starts a login process to allow the user to login.

Kernel development environment

References: 1, 2, 3.

Install tools and libraries

# Build essentials and libraries
sudo apt install bison build-essential ccache ctags fakeroot flex \
    kernel-package libelf-dev libncurses5-dev libssl-dev

# Development tools
sudo apt install git gdb

# QEMU emulator
sudo apt install qemu qemu-system

Clone the kernel source code

cd repository/
git clone git://git.kernel.org/pub/scm/linux/kernel/git/next/linux-next.git

Configure the kernel build options

Use the default configuration for x86_64 architecture.

cd linux-next/
make ARCH=x86_64 x86_64_defconfig

Check that the kernel configuration is saved in .config.

Then use the text based UI to configure the kernel manually, adding GDB debugger options.

make ARCH=x86_64 menuconfig
Menu Config
Menu Config
Menu Config

Save the configuration to .config and exit the TUI.

Compile the kernel

make -j8

The compressed kernel image is saved in arch/x86/boot/bzImage, vmlinux in the source root folder is an uncompressed version.

Build a root file system

We need to build a root file system to boot the kernel. Buildroot is a 'simple, efficient and easy-to-use tool to generate embedded Linux systems through cross-compilation'.

git clone git://git.buildroot.net/buildroot

cd buildroot
make menuconfig

Choose the x86_64 architecture.

Menu Config

And ext2/3/4 file system.

Menu Config

Build the file system.

make -j8

The file system image is saved in output/images/rootfs.ext2.

Start the kernel with the root file system in QEMU

qemu-system-x86_64 -kernel arch/x86/boot/bzImage -boot c -m 2049M \
    -hda ../buildroot/output/images/rootfs.ext2 \
    -append "root=/dev/sda rw console=ttyS0,115200 acpi=off nokaslr" \
    -serial stdio -display none

The system will boot with the kernel in the bzImage file and the root filesystem rootfs.ext2.

Use the -s option to allow gdb to connect to the QEMU instance (through TCP port 1234). Use the -S option to stop the execution until you continue from the gdb. This allows you to debug the kernel initialization process from start_kernel().

Optionally, you can enable KVM with QEMU to improve the performance (--enable-kvm).

Example: kernel debugging

Start the kernel with QEMU:

qemu-system-x86_64 -s -kernel arch/x86/boot/bzImage -boot c -m 2049M \
    -hda ../buildroot/output/images/rootfs.ext2 \
    -append "root=/dev/sda rw console=ttyS0,115200 acpi=off nokaslr" \
    -serial stdio -display none

Configure your GDB to allow the startup script to run. For simplicity, I just add the below command to my ~/.gdbinit file.

set auto-load safe-path /

Them from another shell instance, run the gdb command:

gdb ./vmlinux

(gdb) target remote :1234

Then you can use GDB (Tutorial 1, Tutorial 2, Tutorial 3, TUI) to set break points and debug the kernel.

Example: debug kernel with VS Code

Install the C/C++ for Visual Studio Code extension on VS Code.

If you are debugging from a remote VS Code, install the Remote - SSH extension on VS Code. For example I can use Remote - SSH on my Mac to open the source code directory on my Ubuntu host, and debug a qemu instance on the Ubuntu host.

Open VSCode and open the linux-next source code folder. If you are debugging remotely, use the VS Code command Remote-SSH: Connect to Host... to connect to the remote host first, then open the linux-next folder on the remote host.

Use the menu item Run -> Add Configuration ..., or just manually create an empty launch.json file in the .vscode folder under your source code root. Use the below content in your launch.json:

    "version": "0.2.0",
    "configurations": [
            "name": "(gdb) Debug Kernel",
            "type": "cppdbg",
            "request": "launch",
            "miDebuggerServerAddress": "",
            "program": "${workspaceFolder}/vmlinux",
            "args": [],
            "stopAtEntry": false,
            "cwd": "${workspaceFolder}",
            "environment": [],
            "externalConsole": false,
            "MIMode": "gdb",
            "logging": {
                "engineLogging": false
            "setupCommands": [
                    "description": "Enable pretty-printing for gdb",
                    "text": "-enable-pretty-printing",
                    "ignoreFailures": true
                    "description": "Set Disassembly Flavor to Intel",
                    "text": "-gdb-set disassembly-flavor intel",
                    "ignoreFailures": true

Then start the kernel you just built in QEMU. Remember to use the -s option to open the debugger port 1234.

From VS Code, press F5 to start debugging the kernel. Try add a breakpoint in the __schedule() function in file kernel/sched/core.c, the kernel will stop at the breakpoint upon the next scheduling.

Debugging the Linux kernel in VS Code

Tip: find a symbol in the kernel source

make tags
vim -t sys_fork
Creative Commons License
comments powered by Disqus