Compare commits
17 Commits
Author | SHA1 | Date | |
---|---|---|---|
|
dff1fb46cc | ||
|
d638829538 | ||
|
f95793c534 | ||
|
2b5eda7968 | ||
|
1f3a417ff7 | ||
|
b13b5bbcf3 | ||
|
764dc320d5 | ||
|
feb9373578 | ||
|
1d2c0e5c66 | ||
|
ee6cc74e03 | ||
|
e9bc6b34e4 | ||
|
e1372ce04b | ||
|
ccbf7e5a82 | ||
|
0213d7503f | ||
|
3fd1eaddc4 | ||
|
cf3ee4415a | ||
|
ccc1e5cb69 |
@@ -3,7 +3,7 @@
|
|||||||
# Linux distribution you are using, options are "Arch", "Ubuntu", "Debian", "Fedora", "Suse", "Gentoo", "Generic".
|
# Linux distribution you are using, options are "Arch", "Ubuntu", "Debian", "Fedora", "Suse", "Gentoo", "Generic".
|
||||||
# It is automatically set to "Arch" when using PKGBUILD.
|
# It is automatically set to "Arch" when using PKGBUILD.
|
||||||
# If left empty, the script will prompt
|
# If left empty, the script will prompt
|
||||||
_distro=""
|
_distro="Arch"
|
||||||
|
|
||||||
# Kernel Version - Options are "5.4", and from "5.7" to "5.19"
|
# Kernel Version - Options are "5.4", and from "5.7" to "5.19"
|
||||||
# you can also set a specific kernel version, e.g. "6.0-rc4" or "5.10.51",
|
# you can also set a specific kernel version, e.g. "6.0-rc4" or "5.10.51",
|
||||||
@@ -46,7 +46,7 @@ CUSTOM_GCC_PATH=""
|
|||||||
CUSTOM_LLVM_PATH=""
|
CUSTOM_LLVM_PATH=""
|
||||||
|
|
||||||
# Set to true to bypass makepkg.conf and use all available threads for compilation. False will respect your makepkg.conf options.
|
# Set to true to bypass makepkg.conf and use all available threads for compilation. False will respect your makepkg.conf options.
|
||||||
_force_all_threads="true"
|
_force_all_threads="false"
|
||||||
|
|
||||||
# Set to true to prevent ccache from being used and set CONFIG_GCC_PLUGINS=y (which needs to be disabled for ccache to work properly)
|
# Set to true to prevent ccache from being used and set CONFIG_GCC_PLUGINS=y (which needs to be disabled for ccache to work properly)
|
||||||
_noccache="false"
|
_noccache="false"
|
||||||
@@ -60,10 +60,10 @@ _modprobeddb="false"
|
|||||||
_modprobeddb_db_path=~/.config/modprobed.db
|
_modprobeddb_db_path=~/.config/modprobed.db
|
||||||
|
|
||||||
# Set to "1" to call make menuconfig, "2" to call make nconfig, "3" to call make xconfig, before building the kernel. Set to false to disable and skip the prompt.
|
# Set to "1" to call make menuconfig, "2" to call make nconfig, "3" to call make xconfig, before building the kernel. Set to false to disable and skip the prompt.
|
||||||
_menunconfig=""
|
_menunconfig="false"
|
||||||
|
|
||||||
# Set to true to generate a kernel config fragment from your changes in menuconfig/nconfig. Set to false to disable and skip the prompt.
|
# Set to true to generate a kernel config fragment from your changes in menuconfig/nconfig. Set to false to disable and skip the prompt.
|
||||||
_diffconfig=""
|
_diffconfig="false"
|
||||||
|
|
||||||
# Set to the file name where the generated config fragment should be written to. Only used if _diffconfig is active.
|
# Set to the file name where the generated config fragment should be written to. Only used if _diffconfig is active.
|
||||||
_diffconfig_name=""
|
_diffconfig_name=""
|
||||||
@@ -97,11 +97,11 @@ _STRIP="true"
|
|||||||
# LEAVE AN EMPTY VALUE TO BE PROMPTED ABOUT FOLLOWING OPTIONS AT BUILD TIME
|
# LEAVE AN EMPTY VALUE TO BE PROMPTED ABOUT FOLLOWING OPTIONS AT BUILD TIME
|
||||||
|
|
||||||
# CPU scheduler - Options are "upds" (TkG's Undead PDS), "pds", "bmq", "muqss", "cacule", "tt", "bore" or "cfs" (kernel's default)
|
# CPU scheduler - Options are "upds" (TkG's Undead PDS), "pds", "bmq", "muqss", "cacule", "tt", "bore" or "cfs" (kernel's default)
|
||||||
_cpusched=""
|
_cpusched="pds"
|
||||||
|
|
||||||
# Compiler to use - Options are "gcc" or "llvm".
|
# Compiler to use - Options are "gcc" or "llvm".
|
||||||
# For advanced users.
|
# For advanced users.
|
||||||
_compiler=""
|
_compiler="gcc"
|
||||||
|
|
||||||
# Force the use of the LLVM Integrated Assembler whether using LLVM, LTO or not.
|
# Force the use of the LLVM Integrated Assembler whether using LLVM, LTO or not.
|
||||||
# Set to "1" to enable.
|
# Set to "1" to enable.
|
||||||
@@ -131,7 +131,7 @@ _preempt_rt_force=""
|
|||||||
# For BMQ: 0: No yield.
|
# For BMQ: 0: No yield.
|
||||||
# 1: Deboost and requeue task. (Default)
|
# 1: Deboost and requeue task. (Default)
|
||||||
# 2: Set rq skip task.
|
# 2: Set rq skip task.
|
||||||
_sched_yield_type=""
|
_sched_yield_type="0"
|
||||||
|
|
||||||
# Round Robin interval is the longest duration two tasks with the same nice level will be delayed for. When CPU time is requested by a task, it receives a time slice equal
|
# Round Robin interval is the longest duration two tasks with the same nice level will be delayed for. When CPU time is requested by a task, it receives a time slice equal
|
||||||
# to the rr_interval in addition to a virtual deadline. When using yield_type 2, a low value can help offset the disadvantages of rescheduling a process that has yielded.
|
# to the rr_interval in addition to a virtual deadline. When using yield_type 2, a low value can help offset the disadvantages of rescheduling a process that has yielded.
|
||||||
@@ -139,7 +139,7 @@ _sched_yield_type=""
|
|||||||
# PDS default: 4ms"
|
# PDS default: 4ms"
|
||||||
# BMQ default: 2ms"
|
# BMQ default: 2ms"
|
||||||
# Set to "1" for 2ms, "2" for 4ms, "3" for 6ms, "4" for 8ms, or "default" to keep the chosen scheduler defaults.
|
# Set to "1" for 2ms, "2" for 4ms, "3" for 6ms, "4" for 8ms, or "default" to keep the chosen scheduler defaults.
|
||||||
_rr_interval=""
|
_rr_interval="2"
|
||||||
|
|
||||||
# Set to "true" to disable FUNCTION_TRACER/GRAPH_TRACER, lowering overhead but limiting debugging and analyzing of kernel functions - Kernel default is "false"
|
# Set to "true" to disable FUNCTION_TRACER/GRAPH_TRACER, lowering overhead but limiting debugging and analyzing of kernel functions - Kernel default is "false"
|
||||||
_ftracedisable="false"
|
_ftracedisable="false"
|
||||||
@@ -154,45 +154,27 @@ _misc_adds="true"
|
|||||||
# Full tickless can give higher performances in case you use isolation of CPUs for tasks
|
# Full tickless can give higher performances in case you use isolation of CPUs for tasks
|
||||||
# and it works only when using the nohz_full kernel parameter, otherwise behaves like idle.
|
# and it works only when using the nohz_full kernel parameter, otherwise behaves like idle.
|
||||||
# Just tickless idle perform better for most platforms.
|
# Just tickless idle perform better for most platforms.
|
||||||
_tickless=""
|
_tickless="2"
|
||||||
|
|
||||||
# Set to "true" to use ACS override patch - https://wiki.archlinux.org/index.php/PCI_passthrough_via_OVMF#Bypassing_the_IOMMU_groups_.28ACS_override_patch.29 - Kernel default is "false"
|
# Set to "true" to use ACS override patch - https://wiki.archlinux.org/index.php/PCI_passthrough_via_OVMF#Bypassing_the_IOMMU_groups_.28ACS_override_patch.29 - Kernel default is "false"
|
||||||
_acs_override=""
|
_acs_override="false"
|
||||||
|
|
||||||
# Set to "true" to add Bcache filesystem support. You'll have to install bcachefs-tools-git from AUR for utilities - https://bcachefs.org/ - If in doubt, set to "false"
|
# Set to "true" to add Bcache filesystem support. You'll have to install bcachefs-tools-git from AUR for utilities - https://bcachefs.org/ - If in doubt, set to "false"
|
||||||
# This can be buggy and isn't recommended on a production machine, also enabling this option will not allow you to enable MGLRU.
|
# This can be buggy and isn't recommended on a production machine, also enabling this option will not allow you to enable MGLRU.
|
||||||
_bcachefs="false"
|
_bcachefs="false"
|
||||||
|
|
||||||
# Set to "true" to add multi-generational LRU framework support on kernel 5.18+ - Improves memory pressure handling - https://lore.kernel.org/lkml/20220706220022.968789-1-yuzhao@google.com/
|
|
||||||
# Older kernel versions might have a patch available in the community-patches repo
|
|
||||||
# ! This option will be disabled when bcachefs is enabled for now !
|
|
||||||
_mglru="true"
|
|
||||||
|
|
||||||
# Set to "true" to enable support for fsync, an experimental replacement for esync found in Valve Proton 4.11+ - https://steamcommunity.com/games/221410/announcements/detail/2957094910196249305
|
|
||||||
# Can be enabled alongside _futex_waitv on 5.13+ to use it as a fallback for older Proton builds
|
|
||||||
_fsync="true"
|
|
||||||
|
|
||||||
# Set to "true" to enable support for futex2, an experimental interface that can be used by proton-tkg and proton 5.13 experimental through Fsync - Can be enabled alongside fsync to use it as a fallback
|
|
||||||
# https://gitlab.collabora.com/tonyk/linux/-/tree/futex2-dev
|
|
||||||
_futex2="true"
|
|
||||||
|
|
||||||
# Set to "true" to enable backported patches to add support for the futex_waitv() syscall, a new interface for fsync. It will appear in mainline at Linux 5.16 release and requires a wine/proton with builtin support for it. It's expected to be available in Valve Proton 6.3 stable soon - https://github.com/ValveSoftware/wine/pull/128
|
|
||||||
# !! Disables futex2 interfaces support !!
|
|
||||||
# https://github.com/andrealmeid/futex_waitv_patches
|
|
||||||
_futex_waitv="false"
|
|
||||||
|
|
||||||
# Set to "true" to enable support for winesync, an experimental replacement for esync - requires patched wine - https://repo.or.cz/linux/zf.git/shortlog/refs/heads/winesync4
|
# Set to "true" to enable support for winesync, an experimental replacement for esync - requires patched wine - https://repo.or.cz/linux/zf.git/shortlog/refs/heads/winesync4
|
||||||
# ! Can't be used on multiple kernels installed side-by-side, which will require https://aur.archlinux.org/packages/winesync-dkms/ instead of this option !
|
# ! Can't be used on multiple kernels installed side-by-side, which will require https://aur.archlinux.org/packages/winesync-dkms/ instead of this option !
|
||||||
_winesync="false"
|
_winesync="false"
|
||||||
|
|
||||||
# Set to "true" to enable Binder and Ashmem, the kernel modules required to use the android emulator Anbox. ! This doesn't apply to 5.4.y !
|
# Set to "true" to enable Binder and Ashmem, the kernel modules required to use the android emulator Anbox. ! This doesn't apply to 5.4.y !
|
||||||
_anbox=""
|
_anbox="false"
|
||||||
|
|
||||||
# A selection of patches from Zen/Liquorix kernel and additional tweaks for a better gaming experience (ZENIFY) - Default is "true"
|
# A selection of patches from Zen/Liquorix kernel and additional tweaks for a better gaming experience (ZENIFY) - Default is "true"
|
||||||
_zenify="true"
|
_zenify="true"
|
||||||
|
|
||||||
# compiler optimization level - 1. Optimize for performance (-O2); 2. Optimize harder (-O3); 3. Optimize for size (-Os) - Kernel default is "1"
|
# compiler optimization level - 1. Optimize for performance (-O2); 2. Optimize harder (-O3); 3. Optimize for size (-Os) - Kernel default is "1"
|
||||||
_compileroptlevel="1"
|
_compileroptlevel="2"
|
||||||
|
|
||||||
# CPU compiler optimizations - Defaults to prompt at kernel config if left empty
|
# CPU compiler optimizations - Defaults to prompt at kernel config if left empty
|
||||||
# AMD CPUs : "k8" "k8sse3" "k10" "barcelona" "bobcat" "jaguar" "bulldozer" "piledriver" "steamroller" "excavator" "zen" "zen2" "zen3" "zen4" (zen3 opt support depends on GCC11) (zen4 opt support depends on GCC13)
|
# AMD CPUs : "k8" "k8sse3" "k10" "barcelona" "bobcat" "jaguar" "bulldozer" "piledriver" "steamroller" "excavator" "zen" "zen2" "zen3" "zen4" (zen3 opt support depends on GCC11) (zen4 opt support depends on GCC13)
|
||||||
@@ -206,10 +188,7 @@ _compileroptlevel="1"
|
|||||||
# - "generic_v2" (depends on GCC11 - to share the package between machines with different CPU µarch supporting at least x86-64-v2
|
# - "generic_v2" (depends on GCC11 - to share the package between machines with different CPU µarch supporting at least x86-64-v2
|
||||||
# - "generic_v3" (depends on GCC11 - to share the package between machines with different CPU µarch supporting at least x86-64-v3
|
# - "generic_v3" (depends on GCC11 - to share the package between machines with different CPU µarch supporting at least x86-64-v3
|
||||||
# - "generic_v4" (depends on GCC11 - to share the package between machines with different CPU µarch supporting at least x86-64-v4
|
# - "generic_v4" (depends on GCC11 - to share the package between machines with different CPU µarch supporting at least x86-64-v4
|
||||||
_processor_opt=""
|
_processor_opt="skylake"
|
||||||
|
|
||||||
# MuQSS only - Make IRQ threading compulsory (FORCE_IRQ_THREADING) - Default is "false"
|
|
||||||
_irq_threading="false"
|
|
||||||
|
|
||||||
# CacULE only - Enable Response Driven Balancer, an experimental load balancer for CacULE
|
# CacULE only - Enable Response Driven Balancer, an experimental load balancer for CacULE
|
||||||
_cacule_rdb="false"
|
_cacule_rdb="false"
|
||||||
@@ -222,17 +201,13 @@ _cacule_rdb_interval="19"
|
|||||||
_tt_high_hz="false"
|
_tt_high_hz="false"
|
||||||
|
|
||||||
# MuQSS and PDS only - SMT (Hyperthreading) aware nice priority and policy support (SMT_NICE) - Kernel default is "true" - You can disable this on non-SMT/HT CPUs for lower overhead
|
# MuQSS and PDS only - SMT (Hyperthreading) aware nice priority and policy support (SMT_NICE) - Kernel default is "true" - You can disable this on non-SMT/HT CPUs for lower overhead
|
||||||
_smt_nice=""
|
_smt_nice="true"
|
||||||
|
|
||||||
# Trust the CPU manufacturer to initialize Linux's CRNG (RANDOM_TRUST_CPU) - Kernel default is "false"
|
# Trust the CPU manufacturer to initialize Linux's CRNG (RANDOM_TRUST_CPU) - Kernel default is "false"
|
||||||
_random_trust_cpu="true"
|
_random_trust_cpu="true"
|
||||||
|
|
||||||
# MuQSS only - CPU scheduler runqueue sharing - No sharing (RQ_NONE), SMT (hyperthread) siblings (RQ_SMT), Multicore siblings (RQ_MC), Symmetric Multi-Processing (RQ_SMP), NUMA (RQ_ALL)
|
|
||||||
# Valid values are "none", "smt", "mc", "mc-llc"(for zen), "smp", "all" - Kernel default is "smt"
|
|
||||||
_runqueue_sharing=""
|
|
||||||
|
|
||||||
# Timer frequency - "100" "250" "300" "500" "750" "1000" ("2000" is available for cacule cpusched only) - More options available in kernel config prompt when left empty depending on selected cpusched with the default option pointed with a ">" (2000 for cacule, 100 for muqss and 1000 for other cpu schedulers)
|
# Timer frequency - "100" "250" "300" "500" "750" "1000" ("2000" is available for cacule cpusched only) - More options available in kernel config prompt when left empty depending on selected cpusched with the default option pointed with a ">" (2000 for cacule, 100 for muqss and 1000 for other cpu schedulers)
|
||||||
_timer_freq=""
|
_timer_freq="500"
|
||||||
|
|
||||||
# Default CPU governor - "performance", "ondemand", "schedutil" or leave empty for default (schedutil)
|
# Default CPU governor - "performance", "ondemand", "schedutil" or leave empty for default (schedutil)
|
||||||
_default_cpu_gov="ondemand"
|
_default_cpu_gov="ondemand"
|
||||||
@@ -264,14 +239,41 @@ _custom_pkgbase=""
|
|||||||
# If left empty, it will use "-tkg-${_cpusched}${_compiler}" where "${_cpusched}" will be replaced by the user chosen scheduler, ${_compiler} will be replaced by "-llvm" if clang is used (nothing for GCC).
|
# If left empty, it will use "-tkg-${_cpusched}${_compiler}" where "${_cpusched}" will be replaced by the user chosen scheduler, ${_compiler} will be replaced by "-llvm" if clang is used (nothing for GCC).
|
||||||
_kernel_localversion=""
|
_kernel_localversion=""
|
||||||
|
|
||||||
# Set to "true" to add back missing symbol for AES-NI/AVX support on ZFS - This is a legacy option that can be ignored on 5.10+ kernels - https://github.com/NixOS/nixpkgs/blob/master/pkgs/os-specific/linux/kernel/export_kernel_fpu_functions.patch
|
|
||||||
_zfsfix="true"
|
|
||||||
|
|
||||||
# Set to your maximum number of CPUs (physical + logical cores) - Lower means less overhead - You can set it to "$(nproc)" to use the current host's CPU(s) core count, or leave empty to use default
|
# Set to your maximum number of CPUs (physical + logical cores) - Lower means less overhead - You can set it to "$(nproc)" to use the current host's CPU(s) core count, or leave empty to use default
|
||||||
# If you set this to a lower value than you have cores, some cores will be disabled
|
# If you set this to a lower value than you have cores, some cores will be disabled
|
||||||
# Default Arch kernel value is 320
|
# Default Arch kernel value is 320
|
||||||
_NR_CPUS_value=""
|
_NR_CPUS_value=""
|
||||||
|
|
||||||
|
#### LEGACY OPTIONS ####
|
||||||
|
|
||||||
|
# Upstreamed version of Fsync from Linux 5.16 for previous kernel versions - https://github.com/andrealmeid/futex_waitv_patches
|
||||||
|
# ! Only affect 5.13, 5.14 and 5.15 kernel branches. Safely ignored for 5.16 or newer !
|
||||||
|
_fsync_backport="true"
|
||||||
|
|
||||||
|
# Fsync legacy, known as FUTEX_WAIT_MULTIPLE (opcode 31) - previous version of fsync required for Valve Proton 4.11, 5.0 and 5.13 - https://steamcommunity.com/games/221410/announcements/detail/2957094910196249305
|
||||||
|
_fsync_legacy="true"
|
||||||
|
|
||||||
|
# Set to "true" to enable support for futex2, a DEPRECATED interface that can be used by proton-tkg and proton 5.13 experimental through Fsync - Can be enabled alongside fsync legacy to use it as a fallback
|
||||||
|
# https://gitlab.collabora.com/tonyk/linux/-/tree/futex2-dev
|
||||||
|
# ! Only affect 5.10-5.14 kernel branches. Safely ignored for 5.15 or newer !
|
||||||
|
# ! required _fsync_backport="false" !
|
||||||
|
_fsync_futex2="false"
|
||||||
|
|
||||||
|
# Set to "true" to add back missing symbol for AES-NI/AVX support on ZFS - This is a legacy option that can be ignored on 5.10+ kernels - https://github.com/NixOS/nixpkgs/blob/master/pkgs/os-specific/linux/kernel/export_kernel_fpu_functions.patch
|
||||||
|
_zfsfix="true"
|
||||||
|
|
||||||
|
# MuQSS only - CPU scheduler runqueue sharing - No sharing (RQ_NONE), SMT (hyperthread) siblings (RQ_SMT), Multicore siblings (RQ_MC), Symmetric Multi-Processing (RQ_SMP), NUMA (RQ_ALL)
|
||||||
|
# Valid values are "none", "smt", "mc", "mc-llc"(for zen), "smp", "all" - Kernel default is "smt"
|
||||||
|
_runqueue_sharing=""
|
||||||
|
|
||||||
|
# MuQSS only - Make IRQ threading compulsory (FORCE_IRQ_THREADING) - Default is "false"
|
||||||
|
_irq_threading="false"
|
||||||
|
|
||||||
|
# Set to "true" to add multi-generational LRU framework support on kernel 5.18+ - Improves memory pressure handling - https://lore.kernel.org/lkml/20220706220022.968789-1-yuzhao@google.com/
|
||||||
|
# Older kernel versions might have a patch available in the community-patches repo
|
||||||
|
# Upstream as of 6.1
|
||||||
|
# ! This option will be disabled when bcachefs is enabled !
|
||||||
|
_mglru="true"
|
||||||
|
|
||||||
#### USER PATCHES ####
|
#### USER PATCHES ####
|
||||||
|
|
||||||
|
@@ -6974,6 +6974,7 @@ CONFIG_SND_HDA_CODEC_SI3054=m
|
|||||||
CONFIG_SND_HDA_GENERIC=m
|
CONFIG_SND_HDA_GENERIC=m
|
||||||
CONFIG_SND_HDA_POWER_SAVE_DEFAULT=1
|
CONFIG_SND_HDA_POWER_SAVE_DEFAULT=1
|
||||||
CONFIG_SND_HDA_INTEL_HDMI_SILENT_STREAM=y
|
CONFIG_SND_HDA_INTEL_HDMI_SILENT_STREAM=y
|
||||||
|
# CONFIG_SND_HDA_CTL_DEV_ID is not set
|
||||||
# end of HD-Audio
|
# end of HD-Audio
|
||||||
|
|
||||||
CONFIG_SND_HDA_CORE=m
|
CONFIG_SND_HDA_CORE=m
|
||||||
|
@@ -1,15 +1,15 @@
|
|||||||
#
|
#
|
||||||
# Automatically generated file; DO NOT EDIT.
|
# Automatically generated file; DO NOT EDIT.
|
||||||
# Linux/x86 6.2.0-rc4 Kernel Configuration
|
# Linux/x86 6.2.5-arch1 Kernel Configuration
|
||||||
#
|
#
|
||||||
CONFIG_CC_VERSION_TEXT="gcc (GCC) 12.2.0"
|
CONFIG_CC_VERSION_TEXT="gcc (GCC) 12.2.1 20230201"
|
||||||
CONFIG_CC_IS_GCC=y
|
CONFIG_CC_IS_GCC=y
|
||||||
CONFIG_GCC_VERSION=120200
|
CONFIG_GCC_VERSION=120201
|
||||||
CONFIG_CLANG_VERSION=0
|
CONFIG_CLANG_VERSION=0
|
||||||
CONFIG_AS_IS_GNU=y
|
CONFIG_AS_IS_GNU=y
|
||||||
CONFIG_AS_VERSION=23900
|
CONFIG_AS_VERSION=24000
|
||||||
CONFIG_LD_IS_BFD=y
|
CONFIG_LD_IS_BFD=y
|
||||||
CONFIG_LD_VERSION=23900
|
CONFIG_LD_VERSION=24000
|
||||||
CONFIG_LLD_VERSION=0
|
CONFIG_LLD_VERSION=0
|
||||||
CONFIG_CC_CAN_LINK=y
|
CONFIG_CC_CAN_LINK=y
|
||||||
CONFIG_CC_CAN_LINK_STATIC=y
|
CONFIG_CC_CAN_LINK_STATIC=y
|
||||||
@@ -182,10 +182,9 @@ CONFIG_RCU_NOCB_CPU=y
|
|||||||
# CONFIG_RCU_NOCB_CPU_DEFAULT_ALL is not set
|
# CONFIG_RCU_NOCB_CPU_DEFAULT_ALL is not set
|
||||||
# CONFIG_RCU_NOCB_CPU_CB_BOOST is not set
|
# CONFIG_RCU_NOCB_CPU_CB_BOOST is not set
|
||||||
# CONFIG_TASKS_TRACE_RCU_READ_MB is not set
|
# CONFIG_TASKS_TRACE_RCU_READ_MB is not set
|
||||||
# CONFIG_RCU_LAZY is not set
|
CONFIG_RCU_LAZY=y
|
||||||
# end of RCU Subsystem
|
# end of RCU Subsystem
|
||||||
|
|
||||||
CONFIG_BUILD_BIN2C=y
|
|
||||||
CONFIG_IKCONFIG=y
|
CONFIG_IKCONFIG=y
|
||||||
CONFIG_IKCONFIG_PROC=y
|
CONFIG_IKCONFIG_PROC=y
|
||||||
CONFIG_IKHEADERS=m
|
CONFIG_IKHEADERS=m
|
||||||
@@ -638,7 +637,7 @@ CONFIG_ACPI_ADXL=y
|
|||||||
CONFIG_ACPI_CONFIGFS=m
|
CONFIG_ACPI_CONFIGFS=m
|
||||||
CONFIG_ACPI_PFRUT=m
|
CONFIG_ACPI_PFRUT=m
|
||||||
CONFIG_ACPI_PCC=y
|
CONFIG_ACPI_PCC=y
|
||||||
# CONFIG_ACPI_FFH is not set
|
CONFIG_ACPI_FFH=y
|
||||||
CONFIG_PMIC_OPREGION=y
|
CONFIG_PMIC_OPREGION=y
|
||||||
CONFIG_BYTCRC_PMIC_OPREGION=y
|
CONFIG_BYTCRC_PMIC_OPREGION=y
|
||||||
CONFIG_CHTCRC_PMIC_OPREGION=y
|
CONFIG_CHTCRC_PMIC_OPREGION=y
|
||||||
@@ -936,7 +935,7 @@ CONFIG_MODULE_SIG_HASH="sha512"
|
|||||||
# CONFIG_MODULE_COMPRESS_GZIP is not set
|
# CONFIG_MODULE_COMPRESS_GZIP is not set
|
||||||
# CONFIG_MODULE_COMPRESS_XZ is not set
|
# CONFIG_MODULE_COMPRESS_XZ is not set
|
||||||
CONFIG_MODULE_COMPRESS_ZSTD=y
|
CONFIG_MODULE_COMPRESS_ZSTD=y
|
||||||
# CONFIG_MODULE_DECOMPRESS is not set
|
CONFIG_MODULE_DECOMPRESS=y
|
||||||
CONFIG_MODULE_ALLOW_MISSING_NAMESPACE_IMPORTS=y
|
CONFIG_MODULE_ALLOW_MISSING_NAMESPACE_IMPORTS=y
|
||||||
CONFIG_MODPROBE_PATH="/sbin/modprobe"
|
CONFIG_MODPROBE_PATH="/sbin/modprobe"
|
||||||
CONFIG_MODULES_TREE_LOOKUP=y
|
CONFIG_MODULES_TREE_LOOKUP=y
|
||||||
@@ -1805,7 +1804,6 @@ CONFIG_DEFAULT_NET_SCH="fq_codel"
|
|||||||
#
|
#
|
||||||
CONFIG_NET_CLS=y
|
CONFIG_NET_CLS=y
|
||||||
CONFIG_NET_CLS_BASIC=m
|
CONFIG_NET_CLS_BASIC=m
|
||||||
CONFIG_NET_CLS_TCINDEX=m
|
|
||||||
CONFIG_NET_CLS_ROUTE4=m
|
CONFIG_NET_CLS_ROUTE4=m
|
||||||
CONFIG_NET_CLS_FW=m
|
CONFIG_NET_CLS_FW=m
|
||||||
CONFIG_NET_CLS_U32=m
|
CONFIG_NET_CLS_U32=m
|
||||||
@@ -1989,7 +1987,7 @@ CONFIG_BT_HCIUART_QCA=y
|
|||||||
CONFIG_BT_HCIUART_AG6XX=y
|
CONFIG_BT_HCIUART_AG6XX=y
|
||||||
CONFIG_BT_HCIUART_MRVL=y
|
CONFIG_BT_HCIUART_MRVL=y
|
||||||
CONFIG_BT_HCIBCM203X=m
|
CONFIG_BT_HCIBCM203X=m
|
||||||
# CONFIG_BT_HCIBCM4377 is not set
|
CONFIG_BT_HCIBCM4377=m
|
||||||
CONFIG_BT_HCIBPA10X=m
|
CONFIG_BT_HCIBPA10X=m
|
||||||
CONFIG_BT_HCIBFUSB=m
|
CONFIG_BT_HCIBFUSB=m
|
||||||
CONFIG_BT_HCIDTL1=m
|
CONFIG_BT_HCIDTL1=m
|
||||||
@@ -2327,7 +2325,7 @@ CONFIG_SYSFB=y
|
|||||||
CONFIG_FW_CS_DSP=m
|
CONFIG_FW_CS_DSP=m
|
||||||
CONFIG_GOOGLE_FIRMWARE=y
|
CONFIG_GOOGLE_FIRMWARE=y
|
||||||
# CONFIG_GOOGLE_SMI is not set
|
# CONFIG_GOOGLE_SMI is not set
|
||||||
# CONFIG_GOOGLE_CBMEM is not set
|
CONFIG_GOOGLE_CBMEM=m
|
||||||
CONFIG_GOOGLE_COREBOOT_TABLE=m
|
CONFIG_GOOGLE_COREBOOT_TABLE=m
|
||||||
CONFIG_GOOGLE_MEMCONSOLE=m
|
CONFIG_GOOGLE_MEMCONSOLE=m
|
||||||
# CONFIG_GOOGLE_MEMCONSOLE_X86_LEGACY is not set
|
# CONFIG_GOOGLE_MEMCONSOLE_X86_LEGACY is not set
|
||||||
@@ -2539,7 +2537,7 @@ CONFIG_ZRAM_DEF_COMP_LZORLE=y
|
|||||||
CONFIG_ZRAM_DEF_COMP="lzo-rle"
|
CONFIG_ZRAM_DEF_COMP="lzo-rle"
|
||||||
CONFIG_ZRAM_WRITEBACK=y
|
CONFIG_ZRAM_WRITEBACK=y
|
||||||
# CONFIG_ZRAM_MEMORY_TRACKING is not set
|
# CONFIG_ZRAM_MEMORY_TRACKING is not set
|
||||||
# CONFIG_ZRAM_MULTI_COMP is not set
|
CONFIG_ZRAM_MULTI_COMP=y
|
||||||
CONFIG_BLK_DEV_LOOP=m
|
CONFIG_BLK_DEV_LOOP=m
|
||||||
CONFIG_BLK_DEV_LOOP_MIN_COUNT=0
|
CONFIG_BLK_DEV_LOOP_MIN_COUNT=0
|
||||||
CONFIG_BLK_DEV_DRBD=m
|
CONFIG_BLK_DEV_DRBD=m
|
||||||
@@ -2548,7 +2546,9 @@ CONFIG_BLK_DEV_NBD=m
|
|||||||
CONFIG_BLK_DEV_RAM=m
|
CONFIG_BLK_DEV_RAM=m
|
||||||
CONFIG_BLK_DEV_RAM_COUNT=16
|
CONFIG_BLK_DEV_RAM_COUNT=16
|
||||||
CONFIG_BLK_DEV_RAM_SIZE=16384
|
CONFIG_BLK_DEV_RAM_SIZE=16384
|
||||||
# CONFIG_CDROM_PKTCDVD is not set
|
CONFIG_CDROM_PKTCDVD=m
|
||||||
|
CONFIG_CDROM_PKTCDVD_BUFFERS=8
|
||||||
|
# CONFIG_CDROM_PKTCDVD_WCACHE is not set
|
||||||
CONFIG_ATA_OVER_ETH=m
|
CONFIG_ATA_OVER_ETH=m
|
||||||
CONFIG_XEN_BLKDEV_FRONTEND=m
|
CONFIG_XEN_BLKDEV_FRONTEND=m
|
||||||
CONFIG_XEN_BLKDEV_BACKEND=m
|
CONFIG_XEN_BLKDEV_BACKEND=m
|
||||||
@@ -2597,6 +2597,8 @@ CONFIG_TIFM_CORE=m
|
|||||||
CONFIG_TIFM_7XX1=m
|
CONFIG_TIFM_7XX1=m
|
||||||
CONFIG_ICS932S401=m
|
CONFIG_ICS932S401=m
|
||||||
CONFIG_ENCLOSURE_SERVICES=m
|
CONFIG_ENCLOSURE_SERVICES=m
|
||||||
|
CONFIG_SMPRO_ERRMON=m
|
||||||
|
CONFIG_SMPRO_MISC=m
|
||||||
CONFIG_HP_ILO=m
|
CONFIG_HP_ILO=m
|
||||||
CONFIG_APDS9802ALS=m
|
CONFIG_APDS9802ALS=m
|
||||||
CONFIG_ISL29003=m
|
CONFIG_ISL29003=m
|
||||||
@@ -3321,7 +3323,7 @@ CONFIG_ENC28J60=m
|
|||||||
# CONFIG_ENC28J60_WRITEVERIFY is not set
|
# CONFIG_ENC28J60_WRITEVERIFY is not set
|
||||||
CONFIG_ENCX24J600=m
|
CONFIG_ENCX24J600=m
|
||||||
CONFIG_LAN743X=m
|
CONFIG_LAN743X=m
|
||||||
# CONFIG_VCAP is not set
|
CONFIG_VCAP=y
|
||||||
CONFIG_NET_VENDOR_MICROSEMI=y
|
CONFIG_NET_VENDOR_MICROSEMI=y
|
||||||
CONFIG_MSCC_OCELOT_SWITCH_LIB=m
|
CONFIG_MSCC_OCELOT_SWITCH_LIB=m
|
||||||
CONFIG_NET_VENDOR_MICROSOFT=y
|
CONFIG_NET_VENDOR_MICROSOFT=y
|
||||||
@@ -3329,6 +3331,7 @@ CONFIG_MICROSOFT_MANA=m
|
|||||||
CONFIG_NET_VENDOR_MYRI=y
|
CONFIG_NET_VENDOR_MYRI=y
|
||||||
CONFIG_MYRI10GE=m
|
CONFIG_MYRI10GE=m
|
||||||
CONFIG_MYRI10GE_DCA=y
|
CONFIG_MYRI10GE_DCA=y
|
||||||
|
CONFIG_FEALNX=m
|
||||||
CONFIG_NET_VENDOR_NI=y
|
CONFIG_NET_VENDOR_NI=y
|
||||||
CONFIG_NI_XGE_MANAGEMENT_ENET=m
|
CONFIG_NI_XGE_MANAGEMENT_ENET=m
|
||||||
CONFIG_NET_VENDOR_NATSEMI=y
|
CONFIG_NET_VENDOR_NATSEMI=y
|
||||||
@@ -3895,7 +3898,7 @@ CONFIG_MT7921_COMMON=m
|
|||||||
CONFIG_MT7921E=m
|
CONFIG_MT7921E=m
|
||||||
CONFIG_MT7921S=m
|
CONFIG_MT7921S=m
|
||||||
CONFIG_MT7921U=m
|
CONFIG_MT7921U=m
|
||||||
# CONFIG_MT7996E is not set
|
CONFIG_MT7996E=m
|
||||||
CONFIG_WLAN_VENDOR_MICROCHIP=y
|
CONFIG_WLAN_VENDOR_MICROCHIP=y
|
||||||
CONFIG_WILC1000=m
|
CONFIG_WILC1000=m
|
||||||
CONFIG_WILC1000_SDIO=m
|
CONFIG_WILC1000_SDIO=m
|
||||||
@@ -3959,27 +3962,29 @@ CONFIG_RTL8XXXU_UNTESTED=y
|
|||||||
CONFIG_RTW88=m
|
CONFIG_RTW88=m
|
||||||
CONFIG_RTW88_CORE=m
|
CONFIG_RTW88_CORE=m
|
||||||
CONFIG_RTW88_PCI=m
|
CONFIG_RTW88_PCI=m
|
||||||
|
CONFIG_RTW88_USB=m
|
||||||
CONFIG_RTW88_8822B=m
|
CONFIG_RTW88_8822B=m
|
||||||
CONFIG_RTW88_8822C=m
|
CONFIG_RTW88_8822C=m
|
||||||
CONFIG_RTW88_8723D=m
|
CONFIG_RTW88_8723D=m
|
||||||
CONFIG_RTW88_8821C=m
|
CONFIG_RTW88_8821C=m
|
||||||
CONFIG_RTW88_8822BE=m
|
CONFIG_RTW88_8822BE=m
|
||||||
# CONFIG_RTW88_8822BU is not set
|
CONFIG_RTW88_8822BU=m
|
||||||
CONFIG_RTW88_8822CE=m
|
CONFIG_RTW88_8822CE=m
|
||||||
# CONFIG_RTW88_8822CU is not set
|
CONFIG_RTW88_8822CU=m
|
||||||
CONFIG_RTW88_8723DE=m
|
CONFIG_RTW88_8723DE=m
|
||||||
# CONFIG_RTW88_8723DU is not set
|
CONFIG_RTW88_8723DU=m
|
||||||
CONFIG_RTW88_8821CE=m
|
CONFIG_RTW88_8821CE=m
|
||||||
# CONFIG_RTW88_8821CU is not set
|
CONFIG_RTW88_8821CU=m
|
||||||
CONFIG_RTW88_DEBUG=y
|
CONFIG_RTW88_DEBUG=y
|
||||||
CONFIG_RTW88_DEBUGFS=y
|
CONFIG_RTW88_DEBUGFS=y
|
||||||
CONFIG_RTW89=m
|
CONFIG_RTW89=m
|
||||||
CONFIG_RTW89_CORE=m
|
CONFIG_RTW89_CORE=m
|
||||||
CONFIG_RTW89_PCI=m
|
CONFIG_RTW89_PCI=m
|
||||||
CONFIG_RTW89_8852A=m
|
CONFIG_RTW89_8852A=m
|
||||||
|
CONFIG_RTW89_8852B=m
|
||||||
CONFIG_RTW89_8852C=m
|
CONFIG_RTW89_8852C=m
|
||||||
CONFIG_RTW89_8852AE=m
|
CONFIG_RTW89_8852AE=m
|
||||||
# CONFIG_RTW89_8852BE is not set
|
CONFIG_RTW89_8852BE=m
|
||||||
CONFIG_RTW89_8852CE=m
|
CONFIG_RTW89_8852CE=m
|
||||||
CONFIG_RTW89_DEBUG=y
|
CONFIG_RTW89_DEBUG=y
|
||||||
CONFIG_RTW89_DEBUGMSG=y
|
CONFIG_RTW89_DEBUGMSG=y
|
||||||
@@ -4227,7 +4232,7 @@ CONFIG_TOUCHSCREEN_CYTTSP_SPI=m
|
|||||||
CONFIG_TOUCHSCREEN_CYTTSP4_CORE=m
|
CONFIG_TOUCHSCREEN_CYTTSP4_CORE=m
|
||||||
CONFIG_TOUCHSCREEN_CYTTSP4_I2C=m
|
CONFIG_TOUCHSCREEN_CYTTSP4_I2C=m
|
||||||
CONFIG_TOUCHSCREEN_CYTTSP4_SPI=m
|
CONFIG_TOUCHSCREEN_CYTTSP4_SPI=m
|
||||||
# CONFIG_TOUCHSCREEN_CYTTSP5 is not set
|
CONFIG_TOUCHSCREEN_CYTTSP5=m
|
||||||
CONFIG_TOUCHSCREEN_DA9034=m
|
CONFIG_TOUCHSCREEN_DA9034=m
|
||||||
CONFIG_TOUCHSCREEN_DA9052=m
|
CONFIG_TOUCHSCREEN_DA9052=m
|
||||||
CONFIG_TOUCHSCREEN_DYNAPRO=m
|
CONFIG_TOUCHSCREEN_DYNAPRO=m
|
||||||
@@ -4239,7 +4244,7 @@ CONFIG_TOUCHSCREEN_FUJITSU=m
|
|||||||
CONFIG_TOUCHSCREEN_GOODIX=m
|
CONFIG_TOUCHSCREEN_GOODIX=m
|
||||||
CONFIG_TOUCHSCREEN_HIDEEP=m
|
CONFIG_TOUCHSCREEN_HIDEEP=m
|
||||||
CONFIG_TOUCHSCREEN_HYCON_HY46XX=m
|
CONFIG_TOUCHSCREEN_HYCON_HY46XX=m
|
||||||
# CONFIG_TOUCHSCREEN_HYNITRON_CSTXXX is not set
|
CONFIG_TOUCHSCREEN_HYNITRON_CSTXXX=m
|
||||||
CONFIG_TOUCHSCREEN_ILI210X=m
|
CONFIG_TOUCHSCREEN_ILI210X=m
|
||||||
CONFIG_TOUCHSCREEN_ILITEK=m
|
CONFIG_TOUCHSCREEN_ILITEK=m
|
||||||
CONFIG_TOUCHSCREEN_S6SY761=m
|
CONFIG_TOUCHSCREEN_S6SY761=m
|
||||||
@@ -4314,7 +4319,7 @@ CONFIG_TOUCHSCREEN_COLIBRI_VF50=m
|
|||||||
CONFIG_TOUCHSCREEN_ROHM_BU21023=m
|
CONFIG_TOUCHSCREEN_ROHM_BU21023=m
|
||||||
CONFIG_TOUCHSCREEN_IQS5XX=m
|
CONFIG_TOUCHSCREEN_IQS5XX=m
|
||||||
CONFIG_TOUCHSCREEN_ZINITIX=m
|
CONFIG_TOUCHSCREEN_ZINITIX=m
|
||||||
# CONFIG_TOUCHSCREEN_HIMAX_HX83112B is not set
|
CONFIG_TOUCHSCREEN_HIMAX_HX83112B=m
|
||||||
CONFIG_INPUT_MISC=y
|
CONFIG_INPUT_MISC=y
|
||||||
CONFIG_INPUT_88PM860X_ONKEY=m
|
CONFIG_INPUT_88PM860X_ONKEY=m
|
||||||
CONFIG_INPUT_88PM80X_ONKEY=m
|
CONFIG_INPUT_88PM80X_ONKEY=m
|
||||||
@@ -4432,7 +4437,7 @@ CONFIG_HW_CONSOLE=y
|
|||||||
CONFIG_VT_HW_CONSOLE_BINDING=y
|
CONFIG_VT_HW_CONSOLE_BINDING=y
|
||||||
CONFIG_UNIX98_PTYS=y
|
CONFIG_UNIX98_PTYS=y
|
||||||
# CONFIG_LEGACY_PTYS is not set
|
# CONFIG_LEGACY_PTYS is not set
|
||||||
CONFIG_LEGACY_TIOCSTI=y
|
# CONFIG_LEGACY_TIOCSTI is not set
|
||||||
CONFIG_LDISC_AUTOLOAD=y
|
CONFIG_LDISC_AUTOLOAD=y
|
||||||
|
|
||||||
#
|
#
|
||||||
@@ -4525,7 +4530,7 @@ CONFIG_IPMI_SSIF=m
|
|||||||
CONFIG_IPMI_IPMB=m
|
CONFIG_IPMI_IPMB=m
|
||||||
CONFIG_IPMI_WATCHDOG=m
|
CONFIG_IPMI_WATCHDOG=m
|
||||||
CONFIG_IPMI_POWEROFF=m
|
CONFIG_IPMI_POWEROFF=m
|
||||||
# CONFIG_SSIF_IPMI_BMC is not set
|
CONFIG_SSIF_IPMI_BMC=m
|
||||||
CONFIG_IPMB_DEVICE_INTERFACE=m
|
CONFIG_IPMB_DEVICE_INTERFACE=m
|
||||||
CONFIG_HW_RANDOM=y
|
CONFIG_HW_RANDOM=y
|
||||||
CONFIG_HW_RANDOM_TIMERIOMEM=m
|
CONFIG_HW_RANDOM_TIMERIOMEM=m
|
||||||
@@ -4721,7 +4726,7 @@ CONFIG_SPI_MICROCHIP_CORE=m
|
|||||||
CONFIG_SPI_MICROCHIP_CORE_QSPI=m
|
CONFIG_SPI_MICROCHIP_CORE_QSPI=m
|
||||||
# CONFIG_SPI_LANTIQ_SSC is not set
|
# CONFIG_SPI_LANTIQ_SSC is not set
|
||||||
CONFIG_SPI_OC_TINY=m
|
CONFIG_SPI_OC_TINY=m
|
||||||
# CONFIG_SPI_PCI1XXXX is not set
|
CONFIG_SPI_PCI1XXXX=m
|
||||||
CONFIG_SPI_PXA2XX=m
|
CONFIG_SPI_PXA2XX=m
|
||||||
CONFIG_SPI_PXA2XX_PCI=m
|
CONFIG_SPI_PXA2XX_PCI=m
|
||||||
# CONFIG_SPI_ROCKCHIP is not set
|
# CONFIG_SPI_ROCKCHIP is not set
|
||||||
@@ -4938,7 +4943,7 @@ CONFIG_GPIO_VIPERBOARD=m
|
|||||||
# Virtual GPIO drivers
|
# Virtual GPIO drivers
|
||||||
#
|
#
|
||||||
CONFIG_GPIO_AGGREGATOR=m
|
CONFIG_GPIO_AGGREGATOR=m
|
||||||
# CONFIG_GPIO_LATCH is not set
|
CONFIG_GPIO_LATCH=m
|
||||||
CONFIG_GPIO_MOCKUP=m
|
CONFIG_GPIO_MOCKUP=m
|
||||||
CONFIG_GPIO_VIRTIO=m
|
CONFIG_GPIO_VIRTIO=m
|
||||||
CONFIG_GPIO_SIM=m
|
CONFIG_GPIO_SIM=m
|
||||||
@@ -5076,6 +5081,7 @@ CONFIG_HWMON_VID=m
|
|||||||
#
|
#
|
||||||
CONFIG_SENSORS_ABITUGURU=m
|
CONFIG_SENSORS_ABITUGURU=m
|
||||||
CONFIG_SENSORS_ABITUGURU3=m
|
CONFIG_SENSORS_ABITUGURU3=m
|
||||||
|
CONFIG_SENSORS_SMPRO=m
|
||||||
CONFIG_SENSORS_AD7314=m
|
CONFIG_SENSORS_AD7314=m
|
||||||
CONFIG_SENSORS_AD7414=m
|
CONFIG_SENSORS_AD7414=m
|
||||||
CONFIG_SENSORS_AD7418=m
|
CONFIG_SENSORS_AD7418=m
|
||||||
@@ -5195,8 +5201,9 @@ CONFIG_SENSORS_NCT7904=m
|
|||||||
CONFIG_SENSORS_NPCM7XX=m
|
CONFIG_SENSORS_NPCM7XX=m
|
||||||
CONFIG_SENSORS_NZXT_KRAKEN2=m
|
CONFIG_SENSORS_NZXT_KRAKEN2=m
|
||||||
CONFIG_SENSORS_NZXT_SMART2=m
|
CONFIG_SENSORS_NZXT_SMART2=m
|
||||||
# CONFIG_SENSORS_OCC_P8_I2C is not set
|
CONFIG_SENSORS_OCC_P8_I2C=m
|
||||||
# CONFIG_SENSORS_OXP is not set
|
CONFIG_SENSORS_OCC=m
|
||||||
|
CONFIG_SENSORS_OXP=m
|
||||||
CONFIG_SENSORS_PCF8591=m
|
CONFIG_SENSORS_PCF8591=m
|
||||||
CONFIG_PMBUS=m
|
CONFIG_PMBUS=m
|
||||||
CONFIG_SENSORS_PMBUS=m
|
CONFIG_SENSORS_PMBUS=m
|
||||||
@@ -5401,7 +5408,7 @@ CONFIG_MAX63XX_WATCHDOG=m
|
|||||||
CONFIG_RETU_WATCHDOG=m
|
CONFIG_RETU_WATCHDOG=m
|
||||||
CONFIG_ACQUIRE_WDT=m
|
CONFIG_ACQUIRE_WDT=m
|
||||||
CONFIG_ADVANTECH_WDT=m
|
CONFIG_ADVANTECH_WDT=m
|
||||||
# CONFIG_ADVANTECH_EC_WDT is not set
|
CONFIG_ADVANTECH_EC_WDT=m
|
||||||
CONFIG_ALIM1535_WDT=m
|
CONFIG_ALIM1535_WDT=m
|
||||||
CONFIG_ALIM7101_WDT=m
|
CONFIG_ALIM7101_WDT=m
|
||||||
CONFIG_EBC_C384_WDT=m
|
CONFIG_EBC_C384_WDT=m
|
||||||
@@ -5483,7 +5490,7 @@ CONFIG_BCMA_DRIVER_GPIO=y
|
|||||||
#
|
#
|
||||||
CONFIG_MFD_CORE=y
|
CONFIG_MFD_CORE=y
|
||||||
CONFIG_MFD_AS3711=y
|
CONFIG_MFD_AS3711=y
|
||||||
# CONFIG_MFD_SMPRO is not set
|
CONFIG_MFD_SMPRO=m
|
||||||
CONFIG_PMIC_ADP5520=y
|
CONFIG_PMIC_ADP5520=y
|
||||||
CONFIG_MFD_AAT2870_CORE=y
|
CONFIG_MFD_AAT2870_CORE=y
|
||||||
CONFIG_MFD_BCM590XX=m
|
CONFIG_MFD_BCM590XX=m
|
||||||
@@ -5664,7 +5671,7 @@ CONFIG_REGULATOR_MT6311=m
|
|||||||
CONFIG_REGULATOR_MT6323=m
|
CONFIG_REGULATOR_MT6323=m
|
||||||
CONFIG_REGULATOR_MT6331=m
|
CONFIG_REGULATOR_MT6331=m
|
||||||
CONFIG_REGULATOR_MT6332=m
|
CONFIG_REGULATOR_MT6332=m
|
||||||
# CONFIG_REGULATOR_MT6357 is not set
|
CONFIG_REGULATOR_MT6357=m
|
||||||
CONFIG_REGULATOR_MT6358=m
|
CONFIG_REGULATOR_MT6358=m
|
||||||
CONFIG_REGULATOR_MT6359=m
|
CONFIG_REGULATOR_MT6359=m
|
||||||
CONFIG_REGULATOR_MT6360=m
|
CONFIG_REGULATOR_MT6360=m
|
||||||
@@ -5686,7 +5693,7 @@ CONFIG_REGULATOR_RT5120=m
|
|||||||
CONFIG_REGULATOR_RT5190A=m
|
CONFIG_REGULATOR_RT5190A=m
|
||||||
CONFIG_REGULATOR_RT5759=m
|
CONFIG_REGULATOR_RT5759=m
|
||||||
CONFIG_REGULATOR_RT6160=m
|
CONFIG_REGULATOR_RT6160=m
|
||||||
# CONFIG_REGULATOR_RT6190 is not set
|
CONFIG_REGULATOR_RT6190=m
|
||||||
CONFIG_REGULATOR_RT6245=m
|
CONFIG_REGULATOR_RT6245=m
|
||||||
CONFIG_REGULATOR_RTQ2134=m
|
CONFIG_REGULATOR_RTQ2134=m
|
||||||
CONFIG_REGULATOR_RTMV20=m
|
CONFIG_REGULATOR_RTMV20=m
|
||||||
@@ -6180,7 +6187,8 @@ CONFIG_VIDEO_VIMC=m
|
|||||||
CONFIG_VIDEO_VIVID=m
|
CONFIG_VIDEO_VIVID=m
|
||||||
CONFIG_VIDEO_VIVID_CEC=y
|
CONFIG_VIDEO_VIVID_CEC=y
|
||||||
CONFIG_VIDEO_VIVID_MAX_DEVS=64
|
CONFIG_VIDEO_VIVID_MAX_DEVS=64
|
||||||
# CONFIG_VIDEO_VISL is not set
|
CONFIG_VIDEO_VISL=m
|
||||||
|
# CONFIG_VISL_DEBUGFS is not set
|
||||||
CONFIG_DVB_TEST_DRIVERS=y
|
CONFIG_DVB_TEST_DRIVERS=y
|
||||||
CONFIG_DVB_VIDTV=m
|
CONFIG_DVB_VIDTV=m
|
||||||
|
|
||||||
@@ -6255,7 +6263,7 @@ CONFIG_VIDEO_NOON010PC30=m
|
|||||||
CONFIG_VIDEO_OG01A1B=m
|
CONFIG_VIDEO_OG01A1B=m
|
||||||
CONFIG_VIDEO_OV02A10=m
|
CONFIG_VIDEO_OV02A10=m
|
||||||
CONFIG_VIDEO_OV08D10=m
|
CONFIG_VIDEO_OV08D10=m
|
||||||
# CONFIG_VIDEO_OV08X40 is not set
|
CONFIG_VIDEO_OV08X40=m
|
||||||
CONFIG_VIDEO_OV13858=m
|
CONFIG_VIDEO_OV13858=m
|
||||||
CONFIG_VIDEO_OV13B10=m
|
CONFIG_VIDEO_OV13B10=m
|
||||||
CONFIG_VIDEO_OV2640=m
|
CONFIG_VIDEO_OV2640=m
|
||||||
@@ -6263,7 +6271,7 @@ CONFIG_VIDEO_OV2659=m
|
|||||||
CONFIG_VIDEO_OV2680=m
|
CONFIG_VIDEO_OV2680=m
|
||||||
CONFIG_VIDEO_OV2685=m
|
CONFIG_VIDEO_OV2685=m
|
||||||
CONFIG_VIDEO_OV2740=m
|
CONFIG_VIDEO_OV2740=m
|
||||||
# CONFIG_VIDEO_OV4689 is not set
|
CONFIG_VIDEO_OV4689=m
|
||||||
CONFIG_VIDEO_OV5647=m
|
CONFIG_VIDEO_OV5647=m
|
||||||
CONFIG_VIDEO_OV5648=m
|
CONFIG_VIDEO_OV5648=m
|
||||||
CONFIG_VIDEO_OV5670=m
|
CONFIG_VIDEO_OV5670=m
|
||||||
@@ -6573,7 +6581,6 @@ CONFIG_DRM=y
|
|||||||
CONFIG_DRM_MIPI_DBI=m
|
CONFIG_DRM_MIPI_DBI=m
|
||||||
CONFIG_DRM_MIPI_DSI=y
|
CONFIG_DRM_MIPI_DSI=y
|
||||||
# CONFIG_DRM_DEBUG_MM is not set
|
# CONFIG_DRM_DEBUG_MM is not set
|
||||||
CONFIG_DRM_USE_DYNAMIC_DEBUG=y
|
|
||||||
CONFIG_DRM_KMS_HELPER=y
|
CONFIG_DRM_KMS_HELPER=y
|
||||||
CONFIG_DRM_FBDEV_EMULATION=y
|
CONFIG_DRM_FBDEV_EMULATION=y
|
||||||
CONFIG_DRM_FBDEV_OVERALLOC=100
|
CONFIG_DRM_FBDEV_OVERALLOC=100
|
||||||
@@ -6644,7 +6651,6 @@ CONFIG_DRM_I915_FORCE_PROBE="*"
|
|||||||
CONFIG_DRM_I915_CAPTURE_ERROR=y
|
CONFIG_DRM_I915_CAPTURE_ERROR=y
|
||||||
CONFIG_DRM_I915_COMPRESS_ERROR=y
|
CONFIG_DRM_I915_COMPRESS_ERROR=y
|
||||||
CONFIG_DRM_I915_USERPTR=y
|
CONFIG_DRM_I915_USERPTR=y
|
||||||
CONFIG_DRM_I915_GVT=y
|
|
||||||
CONFIG_DRM_I915_GVT_KVMGT=m
|
CONFIG_DRM_I915_GVT_KVMGT=m
|
||||||
CONFIG_DRM_I915_PXP=y
|
CONFIG_DRM_I915_PXP=y
|
||||||
CONFIG_DRM_I915_REQUEST_TIMEOUT=20000
|
CONFIG_DRM_I915_REQUEST_TIMEOUT=20000
|
||||||
@@ -6656,6 +6662,7 @@ CONFIG_DRM_I915_PREEMPT_TIMEOUT_COMPUTE=7500
|
|||||||
CONFIG_DRM_I915_MAX_REQUEST_BUSYWAIT=8000
|
CONFIG_DRM_I915_MAX_REQUEST_BUSYWAIT=8000
|
||||||
CONFIG_DRM_I915_STOP_TIMEOUT=100
|
CONFIG_DRM_I915_STOP_TIMEOUT=100
|
||||||
CONFIG_DRM_I915_TIMESLICE_DURATION=1
|
CONFIG_DRM_I915_TIMESLICE_DURATION=1
|
||||||
|
CONFIG_DRM_I915_GVT=y
|
||||||
CONFIG_DRM_VGEM=m
|
CONFIG_DRM_VGEM=m
|
||||||
CONFIG_DRM_VKMS=m
|
CONFIG_DRM_VKMS=m
|
||||||
CONFIG_DRM_VMWGFX=m
|
CONFIG_DRM_VMWGFX=m
|
||||||
@@ -6851,7 +6858,7 @@ CONFIG_FRAMEBUFFER_CONSOLE_DEFERRED_TAKEOVER=y
|
|||||||
# CONFIG_LOGO is not set
|
# CONFIG_LOGO is not set
|
||||||
# end of Graphics support
|
# end of Graphics support
|
||||||
|
|
||||||
# CONFIG_DRM_ACCEL is not set
|
CONFIG_DRM_ACCEL=y
|
||||||
CONFIG_SOUND=m
|
CONFIG_SOUND=m
|
||||||
CONFIG_SOUND_OSS_CORE=y
|
CONFIG_SOUND_OSS_CORE=y
|
||||||
# CONFIG_SOUND_OSS_CORE_PRECLAIM is not set
|
# CONFIG_SOUND_OSS_CORE_PRECLAIM is not set
|
||||||
@@ -7020,6 +7027,7 @@ CONFIG_SND_HDA_CODEC_SI3054=m
|
|||||||
CONFIG_SND_HDA_GENERIC=m
|
CONFIG_SND_HDA_GENERIC=m
|
||||||
CONFIG_SND_HDA_POWER_SAVE_DEFAULT=1
|
CONFIG_SND_HDA_POWER_SAVE_DEFAULT=1
|
||||||
CONFIG_SND_HDA_INTEL_HDMI_SILENT_STREAM=y
|
CONFIG_SND_HDA_INTEL_HDMI_SILENT_STREAM=y
|
||||||
|
# CONFIG_SND_HDA_CTL_DEV_ID is not set
|
||||||
# end of HD-Audio
|
# end of HD-Audio
|
||||||
|
|
||||||
CONFIG_SND_HDA_CORE=m
|
CONFIG_SND_HDA_CORE=m
|
||||||
@@ -7163,11 +7171,11 @@ CONFIG_SND_SOC_INTEL_AVS_MACH_DA7219=m
|
|||||||
CONFIG_SND_SOC_INTEL_AVS_MACH_DMIC=m
|
CONFIG_SND_SOC_INTEL_AVS_MACH_DMIC=m
|
||||||
CONFIG_SND_SOC_INTEL_AVS_MACH_HDAUDIO=m
|
CONFIG_SND_SOC_INTEL_AVS_MACH_HDAUDIO=m
|
||||||
CONFIG_SND_SOC_INTEL_AVS_MACH_I2S_TEST=m
|
CONFIG_SND_SOC_INTEL_AVS_MACH_I2S_TEST=m
|
||||||
# CONFIG_SND_SOC_INTEL_AVS_MACH_MAX98927 is not set
|
CONFIG_SND_SOC_INTEL_AVS_MACH_MAX98927=m
|
||||||
CONFIG_SND_SOC_INTEL_AVS_MACH_MAX98357A=m
|
CONFIG_SND_SOC_INTEL_AVS_MACH_MAX98357A=m
|
||||||
CONFIG_SND_SOC_INTEL_AVS_MACH_MAX98373=m
|
CONFIG_SND_SOC_INTEL_AVS_MACH_MAX98373=m
|
||||||
CONFIG_SND_SOC_INTEL_AVS_MACH_NAU8825=m
|
CONFIG_SND_SOC_INTEL_AVS_MACH_NAU8825=m
|
||||||
# CONFIG_SND_SOC_INTEL_AVS_MACH_PROBE is not set
|
CONFIG_SND_SOC_INTEL_AVS_MACH_PROBE=m
|
||||||
CONFIG_SND_SOC_INTEL_AVS_MACH_RT274=m
|
CONFIG_SND_SOC_INTEL_AVS_MACH_RT274=m
|
||||||
CONFIG_SND_SOC_INTEL_AVS_MACH_RT286=m
|
CONFIG_SND_SOC_INTEL_AVS_MACH_RT286=m
|
||||||
CONFIG_SND_SOC_INTEL_AVS_MACH_RT298=m
|
CONFIG_SND_SOC_INTEL_AVS_MACH_RT298=m
|
||||||
@@ -7520,7 +7528,7 @@ CONFIG_SND_SOC_WM8903=m
|
|||||||
CONFIG_SND_SOC_WM8904=m
|
CONFIG_SND_SOC_WM8904=m
|
||||||
CONFIG_SND_SOC_WM8940=m
|
CONFIG_SND_SOC_WM8940=m
|
||||||
CONFIG_SND_SOC_WM8960=m
|
CONFIG_SND_SOC_WM8960=m
|
||||||
# CONFIG_SND_SOC_WM8961 is not set
|
CONFIG_SND_SOC_WM8961=m
|
||||||
CONFIG_SND_SOC_WM8962=m
|
CONFIG_SND_SOC_WM8962=m
|
||||||
CONFIG_SND_SOC_WM8974=m
|
CONFIG_SND_SOC_WM8974=m
|
||||||
CONFIG_SND_SOC_WM8978=m
|
CONFIG_SND_SOC_WM8978=m
|
||||||
@@ -8350,7 +8358,7 @@ CONFIG_INFINIBAND_HFI1=m
|
|||||||
# CONFIG_HFI1_DEBUG_SDMA_ORDER is not set
|
# CONFIG_HFI1_DEBUG_SDMA_ORDER is not set
|
||||||
# CONFIG_SDMA_VERBOSITY is not set
|
# CONFIG_SDMA_VERBOSITY is not set
|
||||||
CONFIG_INFINIBAND_IRDMA=m
|
CONFIG_INFINIBAND_IRDMA=m
|
||||||
# CONFIG_MANA_INFINIBAND is not set
|
CONFIG_MANA_INFINIBAND=m
|
||||||
CONFIG_MLX4_INFINIBAND=m
|
CONFIG_MLX4_INFINIBAND=m
|
||||||
CONFIG_MLX5_INFINIBAND=m
|
CONFIG_MLX5_INFINIBAND=m
|
||||||
CONFIG_INFINIBAND_MTHCA=m
|
CONFIG_INFINIBAND_MTHCA=m
|
||||||
@@ -8626,7 +8634,7 @@ CONFIG_NITRO_ENCLAVES=m
|
|||||||
CONFIG_ACRN_HSM=m
|
CONFIG_ACRN_HSM=m
|
||||||
CONFIG_EFI_SECRET=m
|
CONFIG_EFI_SECRET=m
|
||||||
CONFIG_SEV_GUEST=m
|
CONFIG_SEV_GUEST=m
|
||||||
# CONFIG_TDX_GUEST_DRIVER is not set
|
CONFIG_TDX_GUEST_DRIVER=m
|
||||||
CONFIG_VIRTIO_ANCHOR=y
|
CONFIG_VIRTIO_ANCHOR=y
|
||||||
CONFIG_VIRTIO=y
|
CONFIG_VIRTIO=y
|
||||||
CONFIG_VIRTIO_PCI_LIB=m
|
CONFIG_VIRTIO_PCI_LIB=m
|
||||||
@@ -8824,7 +8832,7 @@ CONFIG_CROS_EC_DEBUGFS=m
|
|||||||
CONFIG_CROS_EC_SENSORHUB=m
|
CONFIG_CROS_EC_SENSORHUB=m
|
||||||
CONFIG_CROS_EC_SYSFS=m
|
CONFIG_CROS_EC_SYSFS=m
|
||||||
CONFIG_CROS_EC_TYPEC=m
|
CONFIG_CROS_EC_TYPEC=m
|
||||||
# CONFIG_CROS_HPS_I2C is not set
|
CONFIG_CROS_HPS_I2C=m
|
||||||
CONFIG_CROS_USBPD_LOGGER=m
|
CONFIG_CROS_USBPD_LOGGER=m
|
||||||
CONFIG_CROS_USBPD_NOTIFY=m
|
CONFIG_CROS_USBPD_NOTIFY=m
|
||||||
CONFIG_CHROMEOS_PRIVACY_SCREEN=m
|
CONFIG_CHROMEOS_PRIVACY_SCREEN=m
|
||||||
@@ -8901,7 +8909,9 @@ CONFIG_AMILO_RFKILL=m
|
|||||||
CONFIG_FUJITSU_LAPTOP=m
|
CONFIG_FUJITSU_LAPTOP=m
|
||||||
CONFIG_FUJITSU_TABLET=m
|
CONFIG_FUJITSU_TABLET=m
|
||||||
CONFIG_GPD_POCKET_FAN=m
|
CONFIG_GPD_POCKET_FAN=m
|
||||||
# CONFIG_X86_PLATFORM_DRIVERS_HP is not set
|
CONFIG_X86_PLATFORM_DRIVERS_HP=y
|
||||||
|
CONFIG_HP_ACCEL=m
|
||||||
|
CONFIG_HP_WMI=m
|
||||||
CONFIG_WIRELESS_HOTKEY=m
|
CONFIG_WIRELESS_HOTKEY=m
|
||||||
CONFIG_IBM_RTL=m
|
CONFIG_IBM_RTL=m
|
||||||
CONFIG_IDEAPAD_LAPTOP=m
|
CONFIG_IDEAPAD_LAPTOP=m
|
||||||
@@ -8916,7 +8926,7 @@ CONFIG_THINKPAD_ACPI_HOTKEY_POLL=y
|
|||||||
CONFIG_THINKPAD_LMI=m
|
CONFIG_THINKPAD_LMI=m
|
||||||
CONFIG_INTEL_ATOMISP2_PDX86=y
|
CONFIG_INTEL_ATOMISP2_PDX86=y
|
||||||
CONFIG_INTEL_ATOMISP2_LED=m
|
CONFIG_INTEL_ATOMISP2_LED=m
|
||||||
# CONFIG_INTEL_IFS is not set
|
CONFIG_INTEL_IFS=m
|
||||||
CONFIG_INTEL_SAR_INT1092=m
|
CONFIG_INTEL_SAR_INT1092=m
|
||||||
CONFIG_INTEL_SKL_INT3472=m
|
CONFIG_INTEL_SKL_INT3472=m
|
||||||
CONFIG_INTEL_PMC_CORE=y
|
CONFIG_INTEL_PMC_CORE=y
|
||||||
@@ -9042,7 +9052,7 @@ CONFIG_INTEL_IOMMU_SVM=y
|
|||||||
# CONFIG_INTEL_IOMMU_DEFAULT_ON is not set
|
# CONFIG_INTEL_IOMMU_DEFAULT_ON is not set
|
||||||
CONFIG_INTEL_IOMMU_FLOPPY_WA=y
|
CONFIG_INTEL_IOMMU_FLOPPY_WA=y
|
||||||
CONFIG_INTEL_IOMMU_SCALABLE_MODE_DEFAULT_ON=y
|
CONFIG_INTEL_IOMMU_SCALABLE_MODE_DEFAULT_ON=y
|
||||||
# CONFIG_IOMMUFD is not set
|
CONFIG_IOMMUFD=m
|
||||||
CONFIG_IRQ_REMAP=y
|
CONFIG_IRQ_REMAP=y
|
||||||
CONFIG_HYPERV_IOMMU=y
|
CONFIG_HYPERV_IOMMU=y
|
||||||
CONFIG_VIRTIO_IOMMU=m
|
CONFIG_VIRTIO_IOMMU=m
|
||||||
@@ -9219,8 +9229,9 @@ CONFIG_IIO_CROS_EC_ACCEL_LEGACY=m
|
|||||||
CONFIG_IIO_ST_ACCEL_3AXIS=m
|
CONFIG_IIO_ST_ACCEL_3AXIS=m
|
||||||
CONFIG_IIO_ST_ACCEL_I2C_3AXIS=m
|
CONFIG_IIO_ST_ACCEL_I2C_3AXIS=m
|
||||||
CONFIG_IIO_ST_ACCEL_SPI_3AXIS=m
|
CONFIG_IIO_ST_ACCEL_SPI_3AXIS=m
|
||||||
# CONFIG_IIO_KX022A_SPI is not set
|
CONFIG_IIO_KX022A=m
|
||||||
# CONFIG_IIO_KX022A_I2C is not set
|
CONFIG_IIO_KX022A_SPI=m
|
||||||
|
CONFIG_IIO_KX022A_I2C=m
|
||||||
CONFIG_KXSD9=m
|
CONFIG_KXSD9=m
|
||||||
CONFIG_KXSD9_SPI=m
|
CONFIG_KXSD9_SPI=m
|
||||||
CONFIG_KXSD9_I2C=m
|
CONFIG_KXSD9_I2C=m
|
||||||
@@ -9247,7 +9258,7 @@ CONFIG_STK8BA50=m
|
|||||||
# Analog to digital converters
|
# Analog to digital converters
|
||||||
#
|
#
|
||||||
CONFIG_AD_SIGMA_DELTA=m
|
CONFIG_AD_SIGMA_DELTA=m
|
||||||
# CONFIG_AD4130 is not set
|
CONFIG_AD4130=m
|
||||||
CONFIG_AD7091R5=m
|
CONFIG_AD7091R5=m
|
||||||
CONFIG_AD7124=m
|
CONFIG_AD7124=m
|
||||||
CONFIG_AD7192=m
|
CONFIG_AD7192=m
|
||||||
@@ -9288,7 +9299,7 @@ CONFIG_MAX1027=m
|
|||||||
CONFIG_MAX11100=m
|
CONFIG_MAX11100=m
|
||||||
CONFIG_MAX1118=m
|
CONFIG_MAX1118=m
|
||||||
CONFIG_MAX11205=m
|
CONFIG_MAX11205=m
|
||||||
# CONFIG_MAX11410 is not set
|
CONFIG_MAX11410=m
|
||||||
CONFIG_MAX1241=m
|
CONFIG_MAX1241=m
|
||||||
CONFIG_MAX1363=m
|
CONFIG_MAX1363=m
|
||||||
CONFIG_MAX9611=m
|
CONFIG_MAX9611=m
|
||||||
@@ -9296,7 +9307,7 @@ CONFIG_MCP320X=m
|
|||||||
CONFIG_MCP3422=m
|
CONFIG_MCP3422=m
|
||||||
CONFIG_MCP3911=m
|
CONFIG_MCP3911=m
|
||||||
CONFIG_MEDIATEK_MT6360_ADC=m
|
CONFIG_MEDIATEK_MT6360_ADC=m
|
||||||
# CONFIG_MEDIATEK_MT6370_ADC is not set
|
CONFIG_MEDIATEK_MT6370_ADC=m
|
||||||
CONFIG_MEN_Z188_ADC=m
|
CONFIG_MEN_Z188_ADC=m
|
||||||
CONFIG_MP2629_ADC=m
|
CONFIG_MP2629_ADC=m
|
||||||
CONFIG_NAU7802=m
|
CONFIG_NAU7802=m
|
||||||
@@ -9329,7 +9340,7 @@ CONFIG_XILINX_XADC=m
|
|||||||
#
|
#
|
||||||
# Analog to digital and digital to analog converters
|
# Analog to digital and digital to analog converters
|
||||||
#
|
#
|
||||||
# CONFIG_AD74115 is not set
|
CONFIG_AD74115=m
|
||||||
CONFIG_AD74413R=m
|
CONFIG_AD74413R=m
|
||||||
# end of Analog to digital and digital to analog converters
|
# end of Analog to digital and digital to analog converters
|
||||||
|
|
||||||
@@ -9479,7 +9490,7 @@ CONFIG_AD9523=m
|
|||||||
#
|
#
|
||||||
CONFIG_ADF4350=m
|
CONFIG_ADF4350=m
|
||||||
CONFIG_ADF4371=m
|
CONFIG_ADF4371=m
|
||||||
# CONFIG_ADF4377 is not set
|
CONFIG_ADF4377=m
|
||||||
CONFIG_ADMV1013=m
|
CONFIG_ADMV1013=m
|
||||||
CONFIG_ADMV1014=m
|
CONFIG_ADMV1014=m
|
||||||
CONFIG_ADMV4420=m
|
CONFIG_ADMV4420=m
|
||||||
@@ -9788,7 +9799,7 @@ CONFIG_TMP007=m
|
|||||||
CONFIG_TMP117=m
|
CONFIG_TMP117=m
|
||||||
CONFIG_TSYS01=m
|
CONFIG_TSYS01=m
|
||||||
CONFIG_TSYS02D=m
|
CONFIG_TSYS02D=m
|
||||||
# CONFIG_MAX30208 is not set
|
CONFIG_MAX30208=m
|
||||||
CONFIG_MAX31856=m
|
CONFIG_MAX31856=m
|
||||||
CONFIG_MAX31865=m
|
CONFIG_MAX31865=m
|
||||||
# end of Temperature sensors
|
# end of Temperature sensors
|
||||||
@@ -9942,7 +9953,8 @@ CONFIG_FPGA_DFL_NIOS_INTEL_PAC_N3000=m
|
|||||||
CONFIG_FPGA_DFL_PCI=m
|
CONFIG_FPGA_DFL_PCI=m
|
||||||
CONFIG_FPGA_M10_BMC_SEC_UPDATE=m
|
CONFIG_FPGA_M10_BMC_SEC_UPDATE=m
|
||||||
CONFIG_FPGA_MGR_MICROCHIP_SPI=m
|
CONFIG_FPGA_MGR_MICROCHIP_SPI=m
|
||||||
# CONFIG_FPGA_MGR_LATTICE_SYSCONFIG_SPI is not set
|
CONFIG_FPGA_MGR_LATTICE_SYSCONFIG=m
|
||||||
|
CONFIG_FPGA_MGR_LATTICE_SYSCONFIG_SPI=m
|
||||||
CONFIG_TEE=m
|
CONFIG_TEE=m
|
||||||
CONFIG_AMDTEE=m
|
CONFIG_AMDTEE=m
|
||||||
CONFIG_MULTIPLEXER=m
|
CONFIG_MULTIPLEXER=m
|
||||||
@@ -10187,10 +10199,10 @@ CONFIG_SQUASHFS=m
|
|||||||
# CONFIG_SQUASHFS_FILE_CACHE is not set
|
# CONFIG_SQUASHFS_FILE_CACHE is not set
|
||||||
CONFIG_SQUASHFS_FILE_DIRECT=y
|
CONFIG_SQUASHFS_FILE_DIRECT=y
|
||||||
CONFIG_SQUASHFS_DECOMP_SINGLE=y
|
CONFIG_SQUASHFS_DECOMP_SINGLE=y
|
||||||
# CONFIG_SQUASHFS_CHOICE_DECOMP_BY_MOUNT is not set
|
CONFIG_SQUASHFS_DECOMP_MULTI=y
|
||||||
CONFIG_SQUASHFS_COMPILE_DECOMP_SINGLE=y
|
CONFIG_SQUASHFS_DECOMP_MULTI_PERCPU=y
|
||||||
# CONFIG_SQUASHFS_COMPILE_DECOMP_MULTI is not set
|
CONFIG_SQUASHFS_CHOICE_DECOMP_BY_MOUNT=y
|
||||||
# CONFIG_SQUASHFS_COMPILE_DECOMP_MULTI_PERCPU is not set
|
CONFIG_SQUASHFS_MOUNT_DECOMP_THREADS=y
|
||||||
CONFIG_SQUASHFS_XATTR=y
|
CONFIG_SQUASHFS_XATTR=y
|
||||||
CONFIG_SQUASHFS_ZLIB=y
|
CONFIG_SQUASHFS_ZLIB=y
|
||||||
CONFIG_SQUASHFS_LZ4=y
|
CONFIG_SQUASHFS_LZ4=y
|
||||||
@@ -10869,6 +10881,7 @@ CONFIG_TEXTSEARCH_BM=m
|
|||||||
CONFIG_TEXTSEARCH_FSM=m
|
CONFIG_TEXTSEARCH_FSM=m
|
||||||
CONFIG_BTREE=y
|
CONFIG_BTREE=y
|
||||||
CONFIG_INTERVAL_TREE=y
|
CONFIG_INTERVAL_TREE=y
|
||||||
|
CONFIG_INTERVAL_TREE_SPAN_ITER=y
|
||||||
CONFIG_XARRAY_MULTI=y
|
CONFIG_XARRAY_MULTI=y
|
||||||
CONFIG_ASSOCIATIVE_ARRAY=y
|
CONFIG_ASSOCIATIVE_ARRAY=y
|
||||||
CONFIG_HAS_IOMEM=y
|
CONFIG_HAS_IOMEM=y
|
||||||
|
14
linux-tkg-config/6.3/90-cleanup.hook
Normal file
14
linux-tkg-config/6.3/90-cleanup.hook
Normal file
@@ -0,0 +1,14 @@
|
|||||||
|
[Trigger]
|
||||||
|
Type = File
|
||||||
|
Operation = Install
|
||||||
|
Operation = Upgrade
|
||||||
|
Operation = Remove
|
||||||
|
Target = usr/lib/modules/*/
|
||||||
|
Target = !usr/lib/modules/*/?*
|
||||||
|
|
||||||
|
[Action]
|
||||||
|
Description = Cleaning up...
|
||||||
|
When = PostTransaction
|
||||||
|
Exec = /usr/share/libalpm/scripts/cleanup
|
||||||
|
NeedsTargets
|
||||||
|
|
10
linux-tkg-config/6.3/cleanup
Executable file
10
linux-tkg-config/6.3/cleanup
Executable file
@@ -0,0 +1,10 @@
|
|||||||
|
#!/bin/bash
|
||||||
|
|
||||||
|
for _f in /usr/lib/modules/*tkg*; do
|
||||||
|
if [[ ! -e ${_f}/vmlinuz ]]; then
|
||||||
|
rm -rf "$_f"
|
||||||
|
fi
|
||||||
|
done
|
||||||
|
|
||||||
|
# vim:set ft=sh sw=2 et:
|
||||||
|
|
11396
linux-tkg-config/6.3/config.x86_64
Normal file
11396
linux-tkg-config/6.3/config.x86_64
Normal file
File diff suppressed because it is too large
Load Diff
@@ -1,7 +1,7 @@
|
|||||||
#!/bin/bash
|
#!/bin/bash
|
||||||
|
|
||||||
# List of kernels that are maintained upstream
|
# List of kernels that are maintained upstream
|
||||||
_current_kernels=("6.2" "6.1" "6.0" "5.15" "5.10" "5.4")
|
_current_kernels=("6.3" "6.2" "6.1" "6.0" "5.15" "5.10" "5.4")
|
||||||
|
|
||||||
# List of kernels that are no longer maintained upstream
|
# List of kernels that are no longer maintained upstream
|
||||||
_eol_kernels=("5.19" "5.18" "5.17" "5.16" "5.14" "5.13" "5.12" "5.11" "5.9" "5.8" "5.7")
|
_eol_kernels=("5.19" "5.18" "5.17" "5.16" "5.14" "5.13" "5.12" "5.11" "5.9" "5.8" "5.7")
|
||||||
@@ -39,32 +39,32 @@ done
|
|||||||
# PREEMPT_RT's supported kernel subversion
|
# PREEMPT_RT's supported kernel subversion
|
||||||
typeset -Ag _rt_subver_map
|
typeset -Ag _rt_subver_map
|
||||||
_rt_subver_map=(
|
_rt_subver_map=(
|
||||||
["5.4"]="221"
|
["5.4"]="230"
|
||||||
["5.9"]="1"
|
["5.9"]="1"
|
||||||
["5.10"]="153"
|
["5.10"]="168"
|
||||||
["5.11"]="4"
|
["5.11"]="4"
|
||||||
["5.14"]="2"
|
["5.14"]="2"
|
||||||
["5.15"]="79"
|
["5.15"]="96"
|
||||||
["5.16"]="2"
|
["5.16"]="2"
|
||||||
["5.17"]="1"
|
["5.17"]="1"
|
||||||
["6.0"]="5"
|
["6.0"]="5"
|
||||||
["6.1"]="rc7"
|
["6.1"]="12"
|
||||||
)
|
)
|
||||||
|
|
||||||
# PREEMPT_RT's patch revision for the kernel
|
# PREEMPT_RT's patch revision for the kernel
|
||||||
# We separated this to allow for forcing the application of the patch when _preempt_rt_force=1 on version mismatch
|
# We separated this to allow for forcing the application of the patch when _preempt_rt_force=1 on version mismatch
|
||||||
typeset -Ag _rt_rev_map
|
typeset -Ag _rt_rev_map
|
||||||
_rt_rev_map=(
|
_rt_rev_map=(
|
||||||
["5.4"]="79"
|
["5.4"]="80"
|
||||||
["5.9"]="20"
|
["5.9"]="20"
|
||||||
["5.10"]="76"
|
["5.10"]="83"
|
||||||
["5.11"]="11"
|
["5.11"]="11"
|
||||||
["5.14"]="21"
|
["5.14"]="21"
|
||||||
["5.15"]="54"
|
["5.15"]="61"
|
||||||
["5.16"]="19"
|
["5.16"]="19"
|
||||||
["5.17"]="17"
|
["5.17"]="17"
|
||||||
["6.0"]="14"
|
["6.0"]="14"
|
||||||
["6.1"]="5"
|
["6.1"]="7"
|
||||||
)
|
)
|
||||||
|
|
||||||
_undefine() {
|
_undefine() {
|
||||||
@@ -593,8 +593,10 @@ _tkg_srcprep() {
|
|||||||
|
|
||||||
cd "$_kernel_work_folder_abs"
|
cd "$_kernel_work_folder_abs"
|
||||||
|
|
||||||
msg2 "Setting version..."
|
if (( "$_kver" <= 602 )); then
|
||||||
scripts/setlocalversion --save-scmversion
|
msg2 "Setting version..."
|
||||||
|
scripts/setlocalversion --save-scmversion
|
||||||
|
fi
|
||||||
|
|
||||||
if [ "${_distro}" = "Arch" ]; then
|
if [ "${_distro}" = "Arch" ]; then
|
||||||
echo "-$pkgrel-tkg-${_cpusched}${_compiler_name}" > localversion.10-pkgrel
|
echo "-$pkgrel-tkg-${_cpusched}${_compiler_name}" > localversion.10-pkgrel
|
||||||
@@ -723,7 +725,7 @@ _tkg_srcprep() {
|
|||||||
elif [ "$_kver" = "518" ]; then
|
elif [ "$_kver" = "518" ]; then
|
||||||
rev=2
|
rev=2
|
||||||
elif [ "$_kver" = "601" ]; then
|
elif [ "$_kver" = "601" ]; then
|
||||||
rev=3
|
rev=1
|
||||||
else
|
else
|
||||||
rev=0
|
rev=0
|
||||||
fi
|
fi
|
||||||
@@ -866,6 +868,11 @@ _tkg_srcprep() {
|
|||||||
_disable "CPU_FREQ_DEFAULT_GOV_ONDEMAND" "CPU_FREQ_DEFAULT_GOV_CONSERVATIVE" "CPU_FREQ_DEFAULT_GOV_PERFORMANCE" "CPU_FREQ_DEFAULT_GOV_PERFORMANCE_NODEF"
|
_disable "CPU_FREQ_DEFAULT_GOV_ONDEMAND" "CPU_FREQ_DEFAULT_GOV_CONSERVATIVE" "CPU_FREQ_DEFAULT_GOV_PERFORMANCE" "CPU_FREQ_DEFAULT_GOV_PERFORMANCE_NODEF"
|
||||||
_module "BLK_DEV_LOOP"
|
_module "BLK_DEV_LOOP"
|
||||||
|
|
||||||
|
# buggy project C/PSI interaction workaround
|
||||||
|
if [ "${_cpusched}" = "pds" ] || [ "${_cpusched}" = "bmq" ]; then
|
||||||
|
_enable "PSI_DEFAULT_DISABLED"
|
||||||
|
fi
|
||||||
|
|
||||||
if [ -n "$_custom_commandline" ]; then
|
if [ -n "$_custom_commandline" ]; then
|
||||||
_enable "CMDLINE_BOOL"
|
_enable "CMDLINE_BOOL"
|
||||||
_disable "CMDLINE_OVERRIDE"
|
_disable "CMDLINE_OVERRIDE"
|
||||||
@@ -1384,10 +1391,10 @@ _tkg_srcprep() {
|
|||||||
fi
|
fi
|
||||||
fi
|
fi
|
||||||
|
|
||||||
# futex_waitv support
|
# fsync (futex_waitv) support
|
||||||
tkgpatch="$srcdir/0007-v${_basekernel}-futex_waitv.patch"
|
tkgpatch="$srcdir/0007-v${_basekernel}-futex_waitv.patch"
|
||||||
if [ -e "$tkgpatch" ]; then
|
if [ -e "$tkgpatch" ]; then
|
||||||
if [ -z "$_futex_waitv" ]; then
|
if [ -z "$_fsync_backport" ]; then
|
||||||
plain ""
|
plain ""
|
||||||
plain "Enable support for futex_waitv, backported patches for fsync from 5.16 Kernel"
|
plain "Enable support for futex_waitv, backported patches for fsync from 5.16 Kernel"
|
||||||
plain "! Will disable futex2 patchset !"
|
plain "! Will disable futex2 patchset !"
|
||||||
@@ -1395,33 +1402,33 @@ _tkg_srcprep() {
|
|||||||
plain "https://github.com/ValveSoftware/wine/pull/128"
|
plain "https://github.com/ValveSoftware/wine/pull/128"
|
||||||
read -rp "`echo $' > N/y : '`" CONDITION9;
|
read -rp "`echo $' > N/y : '`" CONDITION9;
|
||||||
fi
|
fi
|
||||||
if [[ "$CONDITION9" =~ [yY] ]] || [ "$_futex_waitv" = "true" ]; then
|
if [[ "$CONDITION9" =~ [yY] ]] || [ "$_fsync_backport" = "true" ]; then
|
||||||
_msg="Patching futex_waitv support"
|
_msg="Patching fsync support"
|
||||||
_tkg_patcher
|
_tkg_patcher
|
||||||
_futex2="false"
|
_fsync_futex2="false"
|
||||||
fi
|
fi
|
||||||
else
|
else
|
||||||
_futex_waitv="false"
|
_fsync_backport="false"
|
||||||
fi
|
fi
|
||||||
|
|
||||||
# fsync support
|
# fsync legacy support
|
||||||
if [[ $_kver > 515 ]] || [[ "$CONDITION9" =~ [yY] ]] || [ "$_futex_waitv" = "true" ]; then
|
if [[ $_kver > 515 ]] || [[ "$CONDITION9" =~ [yY] ]] || [ "$_fsync_backport" = "true" ]; then
|
||||||
tkgpatch="$srcdir/0007-v${_basekernel}-fsync1_via_futex_waitv.patch"
|
tkgpatch="$srcdir/0007-v${_basekernel}-fsync_legacy_via_futex_waitv.patch"
|
||||||
else
|
else
|
||||||
tkgpatch="$srcdir/0007-v${_basekernel}-fsync.patch"
|
tkgpatch="$srcdir/0007-v${_basekernel}-fsync_legacy.patch"
|
||||||
fi
|
fi
|
||||||
if [ -e "$tkgpatch" ]; then
|
if [ -e "$tkgpatch" ]; then
|
||||||
if [ -z "$_fsync" ]; then
|
if [ -z "$_fsync_legacy" ]; then
|
||||||
plain ""
|
plain ""
|
||||||
plain "Enable support for fsync, an experimental replacement for esync in Valve Proton 4.11+"
|
plain "Enable support for FUTEX_WAIT_MULTIPLE (opcode 31) - fsync legacy used in Valve Proton 4.11, 5.0 and 5.13"
|
||||||
plain "https://steamcommunity.com/games/221410/announcements/detail/2957094910196249305"
|
plain "https://steamcommunity.com/games/221410/announcements/detail/2957094910196249305"
|
||||||
if [[ "$CONDITION9" =~ [yY] ]] || [ "$_futex_waitv" = "true" ]; then
|
if [[ "$CONDITION9" =~ [yY] ]] || [ "$_fsync_backport" = "true" ]; then
|
||||||
plain "Will be used as a fallback to futex_waitv on older Proton builds if enabled"
|
plain "Will be used as a fallback to futex_waitv on older Proton builds if enabled"
|
||||||
fi
|
fi
|
||||||
read -rp "`echo $' > N/y : '`" CONDITION10;
|
read -rp "`echo $' > N/y : '`" CONDITION10;
|
||||||
fi
|
fi
|
||||||
if [[ "$CONDITION10" =~ [yY] ]] || [ "$_fsync" = "true" ]; then
|
if [[ "$CONDITION10" =~ [yY] ]] || [ "$_fsync_legacy" = "true" ]; then
|
||||||
_msg="Patching Fsync support"
|
_msg="Patching fsync legacy support"
|
||||||
_tkg_patcher
|
_tkg_patcher
|
||||||
fi
|
fi
|
||||||
fi
|
fi
|
||||||
@@ -1429,15 +1436,15 @@ _tkg_srcprep() {
|
|||||||
# futex2 support
|
# futex2 support
|
||||||
tkgpatch="$srcdir/0007-v${_basekernel}-futex2_interface.patch"
|
tkgpatch="$srcdir/0007-v${_basekernel}-futex2_interface.patch"
|
||||||
if [ -e "$tkgpatch" ]; then
|
if [ -e "$tkgpatch" ]; then
|
||||||
if [ -z "$_futex2" ]; then
|
if [ -z "$_fsync_futex2" ]; then
|
||||||
plain ""
|
plain ""
|
||||||
plain "Enable support for futex2, an experimental replacement for esync and fsync in Valve Proton 5.13 experimental"
|
plain "Enable support for futex2, a DEPRECATED replacement for esync and fsync in Valve Proton 5.13 experimental"
|
||||||
plain "Can be enabled alongside regular fsync patchset to have a fallback option"
|
plain "Can be enabled alongside fsync legacy patchset to have a fallback option"
|
||||||
plain "https://gitlab.collabora.com/tonyk/linux/-/tree/futex2-dev"
|
plain "https://gitlab.collabora.com/tonyk/linux/-/tree/futex2-dev"
|
||||||
plain "https://github.com/ValveSoftware/Proton/issues/4568"
|
plain "https://github.com/ValveSoftware/Proton/issues/4568"
|
||||||
read -rp "`echo $' > N/y : '`" CONDITION11;
|
read -rp "`echo $' > N/y : '`" CONDITION11;
|
||||||
fi
|
fi
|
||||||
if [[ "$CONDITION11" =~ [yY] ]] || [ "$_futex2" = "true" ]; then
|
if [[ "$CONDITION11" =~ [yY] ]] || [ "$_fsync_futex2" = "true" ]; then
|
||||||
_msg="Patching futex2 support"
|
_msg="Patching futex2 support"
|
||||||
_tkg_patcher
|
_tkg_patcher
|
||||||
_enable "FUTEX2"
|
_enable "FUTEX2"
|
||||||
|
@@ -1,165 +0,0 @@
|
|||||||
From b70e738f08403950aa3053c36b98c6b0eeb0eb90 Mon Sep 17 00:00:00 2001
|
|
||||||
From: =?UTF-8?q?Andr=C3=A9=20Almeida?= <andrealmeid@collabora.com>
|
|
||||||
Date: Mon, 25 Oct 2021 09:49:42 -0300
|
|
||||||
Subject: [PATCH] futex: Add entry point for FUTEX_WAIT_MULTIPLE (opcode 31)
|
|
||||||
MIME-Version: 1.0
|
|
||||||
Content-Type: text/plain; charset=UTF-8
|
|
||||||
Content-Transfer-Encoding: 8bit
|
|
||||||
|
|
||||||
Add an option to wait on multiple futexes using the old interface, that
|
|
||||||
uses opcode 31 through futex() syscall. Do that by just translation the
|
|
||||||
old interface to use the new code. This allows old and stable versions
|
|
||||||
of Proton to still use fsync in new kernel releases.
|
|
||||||
|
|
||||||
Signed-off-by: André Almeida <andrealmeid@collabora.com>
|
|
||||||
---
|
|
||||||
include/uapi/linux/futex.h | 12 ++++++
|
|
||||||
kernel/futex/core.c | 75 +++++++++++++++++++++++++++++++++++++-
|
|
||||||
2 files changed, 86 insertions(+), 1 deletion(-)
|
|
||||||
|
|
||||||
diff --git a/include/uapi/linux/futex.h b/include/uapi/linux/futex.h
|
|
||||||
index 2a06b99f9803..417c5d89b745 100644
|
|
||||||
--- a/include/uapi/linux/futex.h
|
|
||||||
+++ b/include/uapi/linux/futex.h
|
|
||||||
@@ -22,6 +22,7 @@
|
|
||||||
#define FUTEX_WAIT_REQUEUE_PI 11
|
|
||||||
#define FUTEX_CMP_REQUEUE_PI 12
|
|
||||||
#define FUTEX_LOCK_PI2 13
|
|
||||||
+#define FUTEX_WAIT_MULTIPLE 31
|
|
||||||
|
|
||||||
#define FUTEX_PRIVATE_FLAG 128
|
|
||||||
#define FUTEX_CLOCK_REALTIME 256
|
|
||||||
@@ -68,6 +69,17 @@ struct futex_waitv {
|
|
||||||
__u32 __reserved;
|
|
||||||
};
|
|
||||||
|
|
||||||
+/**
|
|
||||||
+ * struct futex_wait_block - Block of futexes to be waited for
|
|
||||||
+ * @uaddr: User address of the futex
|
|
||||||
+ * @val: Futex value expected by userspace
|
|
||||||
+ * @bitset: Bitset for the optional bitmasked wakeup
|
|
||||||
+ */
|
|
||||||
+struct futex_wait_block {
|
|
||||||
+ __u32 __user *uaddr;
|
|
||||||
+ __u32 val;
|
|
||||||
+ __u32 bitset;
|
|
||||||
+};
|
|
||||||
|
|
||||||
/*
|
|
||||||
* Support for robust futexes: the kernel cleans up held futexes at
|
|
||||||
diff --git a/kernel/futex/core.c b/kernel/futex/core.c
|
|
||||||
index 4a9e7ce3714a..c3f2e65afab8 100644
|
|
||||||
--- a/kernel/futex/core.c
|
|
||||||
+++ b/kernel/futex/core.c
|
|
||||||
@@ -4012,6 +4012,7 @@ static __always_inline bool futex_cmd_has_timeout(u32 cmd)
|
|
||||||
case FUTEX_LOCK_PI2:
|
|
||||||
case FUTEX_WAIT_BITSET:
|
|
||||||
case FUTEX_WAIT_REQUEUE_PI:
|
|
||||||
+ case FUTEX_WAIT_MULTIPLE:
|
|
||||||
return true;
|
|
||||||
}
|
|
||||||
return false;
|
|
||||||
@@ -4024,13 +4025,79 @@ futex_init_timeout(u32 cmd, u32 op, struct timespec64 *ts, ktime_t *t)
|
|
||||||
return -EINVAL;
|
|
||||||
|
|
||||||
*t = timespec64_to_ktime(*ts);
|
|
||||||
- if (cmd == FUTEX_WAIT)
|
|
||||||
+ if (cmd == FUTEX_WAIT || cmd == FUTEX_WAIT_MULTIPLE)
|
|
||||||
*t = ktime_add_safe(ktime_get(), *t);
|
|
||||||
else if (cmd != FUTEX_LOCK_PI && !(op & FUTEX_CLOCK_REALTIME))
|
|
||||||
*t = timens_ktime_to_host(CLOCK_MONOTONIC, *t);
|
|
||||||
return 0;
|
|
||||||
}
|
|
||||||
|
|
||||||
+/**
|
|
||||||
+ * futex_read_wait_block - Read an array of futex_wait_block from userspace
|
|
||||||
+ * @uaddr: Userspace address of the block
|
|
||||||
+ * @count: Number of blocks to be read
|
|
||||||
+ *
|
|
||||||
+ * This function creates and allocate an array of futex_q (we zero it to
|
|
||||||
+ * initialize the fields) and then, for each futex_wait_block element from
|
|
||||||
+ * userspace, fill a futex_q element with proper values.
|
|
||||||
+ */
|
|
||||||
+inline struct futex_vector *futex_read_wait_block(u32 __user *uaddr, u32 count)
|
|
||||||
+{
|
|
||||||
+ unsigned int i;
|
|
||||||
+ struct futex_vector *futexv;
|
|
||||||
+ struct futex_wait_block fwb;
|
|
||||||
+ struct futex_wait_block __user *entry =
|
|
||||||
+ (struct futex_wait_block __user *)uaddr;
|
|
||||||
+
|
|
||||||
+ if (!count || count > FUTEX_WAITV_MAX)
|
|
||||||
+ return ERR_PTR(-EINVAL);
|
|
||||||
+
|
|
||||||
+ futexv = kcalloc(count, sizeof(*futexv), GFP_KERNEL);
|
|
||||||
+ if (!futexv)
|
|
||||||
+ return ERR_PTR(-ENOMEM);
|
|
||||||
+
|
|
||||||
+ for (i = 0; i < count; i++) {
|
|
||||||
+ if (copy_from_user(&fwb, &entry[i], sizeof(fwb))) {
|
|
||||||
+ kfree(futexv);
|
|
||||||
+ return ERR_PTR(-EFAULT);
|
|
||||||
+ }
|
|
||||||
+
|
|
||||||
+ futexv[i].w.flags = FUTEX_32;
|
|
||||||
+ futexv[i].w.val = fwb.val;
|
|
||||||
+ futexv[i].w.uaddr = (uintptr_t) (fwb.uaddr);
|
|
||||||
+ futexv[i].q = futex_q_init;
|
|
||||||
+ }
|
|
||||||
+
|
|
||||||
+ return futexv;
|
|
||||||
+}
|
|
||||||
+
|
|
||||||
+int futex_wait_multiple(struct futex_vector *vs, unsigned int count,
|
|
||||||
+ struct hrtimer_sleeper *to);
|
|
||||||
+
|
|
||||||
+int futex_opcode_31(ktime_t *abs_time, u32 __user *uaddr, int count)
|
|
||||||
+{
|
|
||||||
+ int ret;
|
|
||||||
+ struct futex_vector *vs;
|
|
||||||
+ struct hrtimer_sleeper *to = NULL, timeout;
|
|
||||||
+
|
|
||||||
+ to = futex_setup_timer(abs_time, &timeout, 0, 0);
|
|
||||||
+
|
|
||||||
+ vs = futex_read_wait_block(uaddr, count);
|
|
||||||
+
|
|
||||||
+ if (IS_ERR(vs))
|
|
||||||
+ return PTR_ERR(vs);
|
|
||||||
+
|
|
||||||
+ ret = futex_wait_multiple(vs, count, abs_time ? to : NULL);
|
|
||||||
+ kfree(vs);
|
|
||||||
+
|
|
||||||
+ if (to) {
|
|
||||||
+ hrtimer_cancel(&to->timer);
|
|
||||||
+ destroy_hrtimer_on_stack(&to->timer);
|
|
||||||
+ }
|
|
||||||
+
|
|
||||||
+ return ret;
|
|
||||||
+}
|
|
||||||
+
|
|
||||||
SYSCALL_DEFINE6(futex, u32 __user *, uaddr, int, op, u32, val,
|
|
||||||
const struct __kernel_timespec __user *, utime,
|
|
||||||
u32 __user *, uaddr2, u32, val3)
|
|
||||||
@@ -4050,6 +4117,9 @@ SYSCALL_DEFINE6(futex, u32 __user *, uaddr, int, op, u32, val,
|
|
||||||
tp = &t;
|
|
||||||
}
|
|
||||||
|
|
||||||
+ if (cmd == FUTEX_WAIT_MULTIPLE)
|
|
||||||
+ return futex_opcode_31(tp, uaddr, val);
|
|
||||||
+
|
|
||||||
return do_futex(uaddr, op, val, tp, uaddr2, (unsigned long)utime, val3);
|
|
||||||
}
|
|
||||||
|
|
||||||
@@ -4551,6 +4621,9 @@ SYSCALL_DEFINE6(futex_time32, u32 __user *, uaddr, int, op, u32, val,
|
|
||||||
tp = &t;
|
|
||||||
}
|
|
||||||
|
|
||||||
+ if (cmd == FUTEX_WAIT_MULTIPLE)
|
|
||||||
+ return futex_opcode_31(tp, uaddr, val);
|
|
||||||
+
|
|
||||||
return do_futex(uaddr, op, val, tp, uaddr2, (unsigned long)utime, val3);
|
|
||||||
}
|
|
||||||
#endif /* CONFIG_COMPAT_32BIT_TIME */
|
|
||||||
--
|
|
||||||
2.33.1
|
|
||||||
|
|
@@ -5,7 +5,7 @@ index 42af9ca0127e..31747ec54f9d 100644
|
|||||||
@@ -5406,6 +5406,12 @@
|
@@ -5406,6 +5406,12 @@
|
||||||
sa1100ir [NET]
|
sa1100ir [NET]
|
||||||
See drivers/net/irda/sa1100_ir.c.
|
See drivers/net/irda/sa1100_ir.c.
|
||||||
|
|
||||||
+ sched_timeslice=
|
+ sched_timeslice=
|
||||||
+ [KNL] Time slice in ms for Project C BMQ/PDS scheduler.
|
+ [KNL] Time slice in ms for Project C BMQ/PDS scheduler.
|
||||||
+ Format: integer 2, 4
|
+ Format: integer 2, 4
|
||||||
@@ -13,14 +13,14 @@ index 42af9ca0127e..31747ec54f9d 100644
|
|||||||
+ See Documentation/scheduler/sched-BMQ.txt
|
+ See Documentation/scheduler/sched-BMQ.txt
|
||||||
+
|
+
|
||||||
sched_verbose [KNL] Enables verbose scheduler debug messages.
|
sched_verbose [KNL] Enables verbose scheduler debug messages.
|
||||||
|
|
||||||
schedstats= [KNL,X86] Enable or disable scheduled statistics.
|
schedstats= [KNL,X86] Enable or disable scheduled statistics.
|
||||||
diff --git a/Documentation/admin-guide/sysctl/kernel.rst b/Documentation/admin-guide/sysctl/kernel.rst
|
diff --git a/Documentation/admin-guide/sysctl/kernel.rst b/Documentation/admin-guide/sysctl/kernel.rst
|
||||||
index 98d1b198b2b4..d7c78a107f93 100644
|
index 98d1b198b2b4..d7c78a107f93 100644
|
||||||
--- a/Documentation/admin-guide/sysctl/kernel.rst
|
--- a/Documentation/admin-guide/sysctl/kernel.rst
|
||||||
+++ b/Documentation/admin-guide/sysctl/kernel.rst
|
+++ b/Documentation/admin-guide/sysctl/kernel.rst
|
||||||
@@ -1552,3 +1552,13 @@ is 10 seconds.
|
@@ -1552,3 +1552,13 @@ is 10 seconds.
|
||||||
|
|
||||||
The softlockup threshold is (``2 * watchdog_thresh``). Setting this
|
The softlockup threshold is (``2 * watchdog_thresh``). Setting this
|
||||||
tunable to zero will disable lockup detection altogether.
|
tunable to zero will disable lockup detection altogether.
|
||||||
+
|
+
|
||||||
@@ -161,7 +161,7 @@ index 9e479d7d202b..2a8530021b23 100644
|
|||||||
+ (unsigned long long)tsk_seruntime(task),
|
+ (unsigned long long)tsk_seruntime(task),
|
||||||
(unsigned long long)task->sched_info.run_delay,
|
(unsigned long long)task->sched_info.run_delay,
|
||||||
task->sched_info.pcount);
|
task->sched_info.pcount);
|
||||||
|
|
||||||
diff --git a/include/asm-generic/resource.h b/include/asm-generic/resource.h
|
diff --git a/include/asm-generic/resource.h b/include/asm-generic/resource.h
|
||||||
index 8874f681b056..59eb72bf7d5f 100644
|
index 8874f681b056..59eb72bf7d5f 100644
|
||||||
--- a/include/asm-generic/resource.h
|
--- a/include/asm-generic/resource.h
|
||||||
@@ -181,7 +181,7 @@ index ffb6eb55cd13..2e730a59caa2 100644
|
|||||||
+++ b/include/linux/sched.h
|
+++ b/include/linux/sched.h
|
||||||
@@ -762,8 +762,14 @@ struct task_struct {
|
@@ -762,8 +762,14 @@ struct task_struct {
|
||||||
unsigned int ptrace;
|
unsigned int ptrace;
|
||||||
|
|
||||||
#ifdef CONFIG_SMP
|
#ifdef CONFIG_SMP
|
||||||
- int on_cpu;
|
- int on_cpu;
|
||||||
struct __call_single_node wake_entry;
|
struct __call_single_node wake_entry;
|
||||||
@@ -202,11 +202,11 @@ index ffb6eb55cd13..2e730a59caa2 100644
|
|||||||
+#endif /* !CONFIG_SCHED_ALT */
|
+#endif /* !CONFIG_SCHED_ALT */
|
||||||
#endif
|
#endif
|
||||||
int on_rq;
|
int on_rq;
|
||||||
|
|
||||||
@@ -785,6 +792,20 @@ struct task_struct {
|
@@ -785,6 +792,20 @@ struct task_struct {
|
||||||
int normal_prio;
|
int normal_prio;
|
||||||
unsigned int rt_priority;
|
unsigned int rt_priority;
|
||||||
|
|
||||||
+#ifdef CONFIG_SCHED_ALT
|
+#ifdef CONFIG_SCHED_ALT
|
||||||
+ u64 last_ran;
|
+ u64 last_ran;
|
||||||
+ s64 time_slice;
|
+ s64 time_slice;
|
||||||
@@ -229,13 +229,13 @@ index ffb6eb55cd13..2e730a59caa2 100644
|
|||||||
unsigned int core_occupation;
|
unsigned int core_occupation;
|
||||||
#endif
|
#endif
|
||||||
+#endif /* !CONFIG_SCHED_ALT */
|
+#endif /* !CONFIG_SCHED_ALT */
|
||||||
|
|
||||||
#ifdef CONFIG_CGROUP_SCHED
|
#ifdef CONFIG_CGROUP_SCHED
|
||||||
struct task_group *sched_task_group;
|
struct task_group *sched_task_group;
|
||||||
@@ -1545,6 +1567,15 @@ struct task_struct {
|
@@ -1545,6 +1567,15 @@ struct task_struct {
|
||||||
*/
|
*/
|
||||||
};
|
};
|
||||||
|
|
||||||
+#ifdef CONFIG_SCHED_ALT
|
+#ifdef CONFIG_SCHED_ALT
|
||||||
+#define tsk_seruntime(t) ((t)->sched_time)
|
+#define tsk_seruntime(t) ((t)->sched_time)
|
||||||
+/* replace the uncertian rt_timeout with 0UL */
|
+/* replace the uncertian rt_timeout with 0UL */
|
||||||
@@ -254,7 +254,7 @@ index 7c83d4d5a971..fa30f98cb2be 100644
|
|||||||
+++ b/include/linux/sched/deadline.h
|
+++ b/include/linux/sched/deadline.h
|
||||||
@@ -1,5 +1,24 @@
|
@@ -1,5 +1,24 @@
|
||||||
/* SPDX-License-Identifier: GPL-2.0 */
|
/* SPDX-License-Identifier: GPL-2.0 */
|
||||||
|
|
||||||
+#ifdef CONFIG_SCHED_ALT
|
+#ifdef CONFIG_SCHED_ALT
|
||||||
+
|
+
|
||||||
+static inline int dl_task(struct task_struct *p)
|
+static inline int dl_task(struct task_struct *p)
|
||||||
@@ -282,7 +282,7 @@ index 7c83d4d5a971..fa30f98cb2be 100644
|
|||||||
return dl_prio(p->prio);
|
return dl_prio(p->prio);
|
||||||
}
|
}
|
||||||
+#endif /* CONFIG_SCHED_ALT */
|
+#endif /* CONFIG_SCHED_ALT */
|
||||||
|
|
||||||
static inline bool dl_time_before(u64 a, u64 b)
|
static inline bool dl_time_before(u64 a, u64 b)
|
||||||
{
|
{
|
||||||
diff --git a/include/linux/sched/prio.h b/include/linux/sched/prio.h
|
diff --git a/include/linux/sched/prio.h b/include/linux/sched/prio.h
|
||||||
@@ -292,7 +292,7 @@ index ab83d85e1183..6af9ae681116 100644
|
|||||||
@@ -18,6 +18,32 @@
|
@@ -18,6 +18,32 @@
|
||||||
#define MAX_PRIO (MAX_RT_PRIO + NICE_WIDTH)
|
#define MAX_PRIO (MAX_RT_PRIO + NICE_WIDTH)
|
||||||
#define DEFAULT_PRIO (MAX_RT_PRIO + NICE_WIDTH / 2)
|
#define DEFAULT_PRIO (MAX_RT_PRIO + NICE_WIDTH / 2)
|
||||||
|
|
||||||
+#ifdef CONFIG_SCHED_ALT
|
+#ifdef CONFIG_SCHED_ALT
|
||||||
+
|
+
|
||||||
+/* Undefine MAX_PRIO and DEFAULT_PRIO */
|
+/* Undefine MAX_PRIO and DEFAULT_PRIO */
|
||||||
@@ -327,7 +327,7 @@ index 994c25640e15..8c050a59ece1 100644
|
|||||||
--- a/include/linux/sched/rt.h
|
--- a/include/linux/sched/rt.h
|
||||||
+++ b/include/linux/sched/rt.h
|
+++ b/include/linux/sched/rt.h
|
||||||
@@ -24,8 +24,10 @@ static inline bool task_is_realtime(struct task_struct *tsk)
|
@@ -24,8 +24,10 @@ static inline bool task_is_realtime(struct task_struct *tsk)
|
||||||
|
|
||||||
if (policy == SCHED_FIFO || policy == SCHED_RR)
|
if (policy == SCHED_FIFO || policy == SCHED_RR)
|
||||||
return true;
|
return true;
|
||||||
+#ifndef CONFIG_SCHED_ALT
|
+#ifndef CONFIG_SCHED_ALT
|
||||||
@@ -336,15 +336,15 @@ index 994c25640e15..8c050a59ece1 100644
|
|||||||
+#endif
|
+#endif
|
||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
|
|
||||||
diff --git a/include/linux/sched/topology.h b/include/linux/sched/topology.h
|
diff --git a/include/linux/sched/topology.h b/include/linux/sched/topology.h
|
||||||
index 816df6cc444e..c8da08e18c91 100644
|
index 816df6cc444e..c8da08e18c91 100644
|
||||||
--- a/include/linux/sched/topology.h
|
--- a/include/linux/sched/topology.h
|
||||||
+++ b/include/linux/sched/topology.h
|
+++ b/include/linux/sched/topology.h
|
||||||
@@ -234,7 +234,8 @@ static inline bool cpus_share_cache(int this_cpu, int that_cpu)
|
@@ -234,7 +234,8 @@ static inline bool cpus_share_cache(int this_cpu, int that_cpu)
|
||||||
|
|
||||||
#endif /* !CONFIG_SMP */
|
#endif /* !CONFIG_SMP */
|
||||||
|
|
||||||
-#if defined(CONFIG_ENERGY_MODEL) && defined(CONFIG_CPU_FREQ_GOV_SCHEDUTIL)
|
-#if defined(CONFIG_ENERGY_MODEL) && defined(CONFIG_CPU_FREQ_GOV_SCHEDUTIL)
|
||||||
+#if defined(CONFIG_ENERGY_MODEL) && defined(CONFIG_CPU_FREQ_GOV_SCHEDUTIL) && \
|
+#if defined(CONFIG_ENERGY_MODEL) && defined(CONFIG_CPU_FREQ_GOV_SCHEDUTIL) && \
|
||||||
+ !defined(CONFIG_SCHED_ALT)
|
+ !defined(CONFIG_SCHED_ALT)
|
||||||
@@ -364,9 +364,9 @@ index 94125d3b6893..c87ba766d354 100644
|
|||||||
This feature enables the scheduler to track the clamped utilization
|
This feature enables the scheduler to track the clamped utilization
|
||||||
of each CPU based on RUNNABLE tasks scheduled on that CPU.
|
of each CPU based on RUNNABLE tasks scheduled on that CPU.
|
||||||
@@ -865,6 +866,35 @@ config UCLAMP_BUCKETS_COUNT
|
@@ -865,6 +866,35 @@ config UCLAMP_BUCKETS_COUNT
|
||||||
|
|
||||||
If in doubt, use the default value.
|
If in doubt, use the default value.
|
||||||
|
|
||||||
+menuconfig SCHED_ALT
|
+menuconfig SCHED_ALT
|
||||||
+ bool "Alternative CPU Schedulers"
|
+ bool "Alternative CPU Schedulers"
|
||||||
+ default y
|
+ default y
|
||||||
@@ -397,7 +397,7 @@ index 94125d3b6893..c87ba766d354 100644
|
|||||||
+endif
|
+endif
|
||||||
+
|
+
|
||||||
endmenu
|
endmenu
|
||||||
|
|
||||||
#
|
#
|
||||||
@@ -918,6 +948,7 @@ config NUMA_BALANCING
|
@@ -918,6 +948,7 @@ config NUMA_BALANCING
|
||||||
depends on ARCH_SUPPORTS_NUMA_BALANCING
|
depends on ARCH_SUPPORTS_NUMA_BALANCING
|
||||||
@@ -410,7 +410,7 @@ index 94125d3b6893..c87ba766d354 100644
|
|||||||
@@ -1015,6 +1046,7 @@ config FAIR_GROUP_SCHED
|
@@ -1015,6 +1046,7 @@ config FAIR_GROUP_SCHED
|
||||||
depends on CGROUP_SCHED
|
depends on CGROUP_SCHED
|
||||||
default CGROUP_SCHED
|
default CGROUP_SCHED
|
||||||
|
|
||||||
+if !SCHED_ALT
|
+if !SCHED_ALT
|
||||||
config CFS_BANDWIDTH
|
config CFS_BANDWIDTH
|
||||||
bool "CPU bandwidth provisioning for FAIR_GROUP_SCHED"
|
bool "CPU bandwidth provisioning for FAIR_GROUP_SCHED"
|
||||||
@@ -418,13 +418,13 @@ index 94125d3b6893..c87ba766d354 100644
|
|||||||
@@ -1037,6 +1069,7 @@ config RT_GROUP_SCHED
|
@@ -1037,6 +1069,7 @@ config RT_GROUP_SCHED
|
||||||
realtime bandwidth for them.
|
realtime bandwidth for them.
|
||||||
See Documentation/scheduler/sched-rt-group.rst for more information.
|
See Documentation/scheduler/sched-rt-group.rst for more information.
|
||||||
|
|
||||||
+endif #!SCHED_ALT
|
+endif #!SCHED_ALT
|
||||||
endif #CGROUP_SCHED
|
endif #CGROUP_SCHED
|
||||||
|
|
||||||
config UCLAMP_TASK_GROUP
|
config UCLAMP_TASK_GROUP
|
||||||
@@ -1281,6 +1314,7 @@ config CHECKPOINT_RESTORE
|
@@ -1281,6 +1314,7 @@ config CHECKPOINT_RESTORE
|
||||||
|
|
||||||
config SCHED_AUTOGROUP
|
config SCHED_AUTOGROUP
|
||||||
bool "Automatic process group scheduling"
|
bool "Automatic process group scheduling"
|
||||||
+ depends on !SCHED_ALT
|
+ depends on !SCHED_ALT
|
||||||
@@ -482,7 +482,7 @@ index c2f1fd95a821..41654679b1b2 100644
|
|||||||
--- a/kernel/Kconfig.preempt
|
--- a/kernel/Kconfig.preempt
|
||||||
+++ b/kernel/Kconfig.preempt
|
+++ b/kernel/Kconfig.preempt
|
||||||
@@ -117,7 +117,7 @@ config PREEMPT_DYNAMIC
|
@@ -117,7 +117,7 @@ config PREEMPT_DYNAMIC
|
||||||
|
|
||||||
config SCHED_CORE
|
config SCHED_CORE
|
||||||
bool "Core Scheduling for SMT"
|
bool "Core Scheduling for SMT"
|
||||||
- depends on SCHED_SMT
|
- depends on SCHED_SMT
|
||||||
@@ -497,7 +497,7 @@ index b474289c15b8..a23224b45b03 100644
|
|||||||
@@ -787,7 +787,7 @@ static int validate_change(struct cpuset *cur, struct cpuset *trial)
|
@@ -787,7 +787,7 @@ static int validate_change(struct cpuset *cur, struct cpuset *trial)
|
||||||
return ret;
|
return ret;
|
||||||
}
|
}
|
||||||
|
|
||||||
-#ifdef CONFIG_SMP
|
-#ifdef CONFIG_SMP
|
||||||
+#if defined(CONFIG_SMP) && !defined(CONFIG_SCHED_ALT)
|
+#if defined(CONFIG_SMP) && !defined(CONFIG_SCHED_ALT)
|
||||||
/*
|
/*
|
||||||
@@ -522,9 +522,9 @@ index e39cb696cfbd..463423572e09 100644
|
|||||||
t2 = tsk->sched_info.run_delay;
|
t2 = tsk->sched_info.run_delay;
|
||||||
- t3 = tsk->se.sum_exec_runtime;
|
- t3 = tsk->se.sum_exec_runtime;
|
||||||
+ t3 = tsk_seruntime(tsk);
|
+ t3 = tsk_seruntime(tsk);
|
||||||
|
|
||||||
d->cpu_count += t1;
|
d->cpu_count += t1;
|
||||||
|
|
||||||
diff --git a/kernel/exit.c b/kernel/exit.c
|
diff --git a/kernel/exit.c b/kernel/exit.c
|
||||||
index 35e0a31a0315..64e368441cf4 100644
|
index 35e0a31a0315..64e368441cf4 100644
|
||||||
--- a/kernel/exit.c
|
--- a/kernel/exit.c
|
||||||
@@ -532,11 +532,11 @@ index 35e0a31a0315..64e368441cf4 100644
|
|||||||
@@ -125,7 +125,7 @@ static void __exit_signal(struct task_struct *tsk)
|
@@ -125,7 +125,7 @@ static void __exit_signal(struct task_struct *tsk)
|
||||||
sig->curr_target = next_thread(tsk);
|
sig->curr_target = next_thread(tsk);
|
||||||
}
|
}
|
||||||
|
|
||||||
- add_device_randomness((const void*) &tsk->se.sum_exec_runtime,
|
- add_device_randomness((const void*) &tsk->se.sum_exec_runtime,
|
||||||
+ add_device_randomness((const void*) &tsk_seruntime(tsk),
|
+ add_device_randomness((const void*) &tsk_seruntime(tsk),
|
||||||
sizeof(unsigned long long));
|
sizeof(unsigned long long));
|
||||||
|
|
||||||
/*
|
/*
|
||||||
@@ -146,7 +146,7 @@ static void __exit_signal(struct task_struct *tsk)
|
@@ -146,7 +146,7 @@ static void __exit_signal(struct task_struct *tsk)
|
||||||
sig->inblock += task_io_get_inblock(tsk);
|
sig->inblock += task_io_get_inblock(tsk);
|
||||||
@@ -558,14 +558,14 @@ index 7779ee8abc2a..5b9893cdfb1b 100644
|
|||||||
- waiter->deadline = task->dl.deadline;
|
- waiter->deadline = task->dl.deadline;
|
||||||
+ waiter->deadline = __tsk_deadline(task);
|
+ waiter->deadline = __tsk_deadline(task);
|
||||||
}
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Only use with rt_mutex_waiter_{less,equal}()
|
* Only use with rt_mutex_waiter_{less,equal}()
|
||||||
*/
|
*/
|
||||||
#define task_to_waiter(p) \
|
#define task_to_waiter(p) \
|
||||||
- &(struct rt_mutex_waiter){ .prio = __waiter_prio(p), .deadline = (p)->dl.deadline }
|
- &(struct rt_mutex_waiter){ .prio = __waiter_prio(p), .deadline = (p)->dl.deadline }
|
||||||
+ &(struct rt_mutex_waiter){ .prio = __waiter_prio(p), .deadline = __tsk_deadline(p) }
|
+ &(struct rt_mutex_waiter){ .prio = __waiter_prio(p), .deadline = __tsk_deadline(p) }
|
||||||
|
|
||||||
static __always_inline int rt_mutex_waiter_less(struct rt_mutex_waiter *left,
|
static __always_inline int rt_mutex_waiter_less(struct rt_mutex_waiter *left,
|
||||||
struct rt_mutex_waiter *right)
|
struct rt_mutex_waiter *right)
|
||||||
{
|
{
|
||||||
@@ -574,7 +574,7 @@ index 7779ee8abc2a..5b9893cdfb1b 100644
|
|||||||
+#else
|
+#else
|
||||||
if (left->prio < right->prio)
|
if (left->prio < right->prio)
|
||||||
return 1;
|
return 1;
|
||||||
|
|
||||||
+#ifndef CONFIG_SCHED_BMQ
|
+#ifndef CONFIG_SCHED_BMQ
|
||||||
/*
|
/*
|
||||||
* If both waiters have dl_prio(), we check the deadlines of the
|
* If both waiters have dl_prio(), we check the deadlines of the
|
||||||
@@ -584,11 +584,11 @@ index 7779ee8abc2a..5b9893cdfb1b 100644
|
|||||||
if (dl_prio(left->prio))
|
if (dl_prio(left->prio))
|
||||||
return dl_time_before(left->deadline, right->deadline);
|
return dl_time_before(left->deadline, right->deadline);
|
||||||
+#endif
|
+#endif
|
||||||
|
|
||||||
return 0;
|
return 0;
|
||||||
+#endif
|
+#endif
|
||||||
}
|
}
|
||||||
|
|
||||||
static __always_inline int rt_mutex_waiter_equal(struct rt_mutex_waiter *left,
|
static __always_inline int rt_mutex_waiter_equal(struct rt_mutex_waiter *left,
|
||||||
struct rt_mutex_waiter *right)
|
struct rt_mutex_waiter *right)
|
||||||
{
|
{
|
||||||
@@ -597,7 +597,7 @@ index 7779ee8abc2a..5b9893cdfb1b 100644
|
|||||||
+#else
|
+#else
|
||||||
if (left->prio != right->prio)
|
if (left->prio != right->prio)
|
||||||
return 0;
|
return 0;
|
||||||
|
|
||||||
+#ifndef CONFIG_SCHED_BMQ
|
+#ifndef CONFIG_SCHED_BMQ
|
||||||
/*
|
/*
|
||||||
* If both waiters have dl_prio(), we check the deadlines of the
|
* If both waiters have dl_prio(), we check the deadlines of the
|
||||||
@@ -607,11 +607,11 @@ index 7779ee8abc2a..5b9893cdfb1b 100644
|
|||||||
if (dl_prio(left->prio))
|
if (dl_prio(left->prio))
|
||||||
return left->deadline == right->deadline;
|
return left->deadline == right->deadline;
|
||||||
+#endif
|
+#endif
|
||||||
|
|
||||||
return 1;
|
return 1;
|
||||||
+#endif
|
+#endif
|
||||||
}
|
}
|
||||||
|
|
||||||
static inline bool rt_mutex_steal(struct rt_mutex_waiter *waiter,
|
static inline bool rt_mutex_steal(struct rt_mutex_waiter *waiter,
|
||||||
diff --git a/kernel/sched/Makefile b/kernel/sched/Makefile
|
diff --git a/kernel/sched/Makefile b/kernel/sched/Makefile
|
||||||
index 976092b7bd45..31d587c16ec1 100644
|
index 976092b7bd45..31d587c16ec1 100644
|
||||||
@@ -632,10 +632,10 @@ index 976092b7bd45..31d587c16ec1 100644
|
|||||||
obj-y += build_utility.o
|
obj-y += build_utility.o
|
||||||
diff --git a/kernel/sched/alt_core.c b/kernel/sched/alt_core.c
|
diff --git a/kernel/sched/alt_core.c b/kernel/sched/alt_core.c
|
||||||
new file mode 100644
|
new file mode 100644
|
||||||
index 000000000000..acb8657e811d
|
index 000000000000..572eab74418f
|
||||||
--- /dev/null
|
--- /dev/null
|
||||||
+++ b/kernel/sched/alt_core.c
|
+++ b/kernel/sched/alt_core.c
|
||||||
@@ -0,0 +1,7978 @@
|
@@ -0,0 +1,7961 @@
|
||||||
+/*
|
+/*
|
||||||
+ * kernel/sched/alt_core.c
|
+ * kernel/sched/alt_core.c
|
||||||
+ *
|
+ *
|
||||||
@@ -665,6 +665,7 @@ index 000000000000..acb8657e811d
|
|||||||
+#include <linux/init_task.h>
|
+#include <linux/init_task.h>
|
||||||
+#include <linux/kcov.h>
|
+#include <linux/kcov.h>
|
||||||
+#include <linux/kprobes.h>
|
+#include <linux/kprobes.h>
|
||||||
|
+#include <linux/profile.h>
|
||||||
+#include <linux/nmi.h>
|
+#include <linux/nmi.h>
|
||||||
+#include <linux/scs.h>
|
+#include <linux/scs.h>
|
||||||
+
|
+
|
||||||
@@ -705,7 +706,7 @@ index 000000000000..acb8657e811d
|
|||||||
+#define sched_feat(x) (0)
|
+#define sched_feat(x) (0)
|
||||||
+#endif /* CONFIG_SCHED_DEBUG */
|
+#endif /* CONFIG_SCHED_DEBUG */
|
||||||
+
|
+
|
||||||
+#define ALT_SCHED_VERSION "v6.1-r3"
|
+#define ALT_SCHED_VERSION "v6.1-r1"
|
||||||
+
|
+
|
||||||
+/* rt_prio(prio) defined in include/linux/sched/rt.h */
|
+/* rt_prio(prio) defined in include/linux/sched/rt.h */
|
||||||
+#define rt_task(p) rt_prio((p)->prio)
|
+#define rt_task(p) rt_prio((p)->prio)
|
||||||
@@ -814,14 +815,14 @@ index 000000000000..acb8657e811d
|
|||||||
+clear_recorded_preempt_mask(int pr, int low, int high, int cpu)
|
+clear_recorded_preempt_mask(int pr, int low, int high, int cpu)
|
||||||
+{
|
+{
|
||||||
+ if (low < pr && pr <= high)
|
+ if (low < pr && pr <= high)
|
||||||
+ cpumask_clear_cpu(cpu, sched_preempt_mask + SCHED_QUEUE_BITS - pr);
|
+ cpumask_clear_cpu(cpu, sched_preempt_mask + SCHED_QUEUE_BITS - 1 - pr);
|
||||||
+}
|
+}
|
||||||
+
|
+
|
||||||
+static inline void
|
+static inline void
|
||||||
+set_recorded_preempt_mask(int pr, int low, int high, int cpu)
|
+set_recorded_preempt_mask(int pr, int low, int high, int cpu)
|
||||||
+{
|
+{
|
||||||
+ if (low < pr && pr <= high)
|
+ if (low < pr && pr <= high)
|
||||||
+ cpumask_set_cpu(cpu, sched_preempt_mask + SCHED_QUEUE_BITS - pr);
|
+ cpumask_set_cpu(cpu, sched_preempt_mask + SCHED_QUEUE_BITS - 1 - pr);
|
||||||
+}
|
+}
|
||||||
+
|
+
|
||||||
+static atomic_t sched_prio_record = ATOMIC_INIT(0);
|
+static atomic_t sched_prio_record = ATOMIC_INIT(0);
|
||||||
@@ -1391,8 +1392,8 @@ index 000000000000..acb8657e811d
|
|||||||
+ * Context: rq->lock
|
+ * Context: rq->lock
|
||||||
+ */
|
+ */
|
||||||
+#define __SCHED_DEQUEUE_TASK(p, rq, flags) \
|
+#define __SCHED_DEQUEUE_TASK(p, rq, flags) \
|
||||||
+ sched_info_dequeue(rq, p); \
|
|
||||||
+ psi_dequeue(p, flags & DEQUEUE_SLEEP); \
|
+ psi_dequeue(p, flags & DEQUEUE_SLEEP); \
|
||||||
|
+ sched_info_dequeue(rq, p); \
|
||||||
+ \
|
+ \
|
||||||
+ list_del(&p->sq_node); \
|
+ list_del(&p->sq_node); \
|
||||||
+ if (list_empty(&rq->queue.heads[p->sq_idx])) \
|
+ if (list_empty(&rq->queue.heads[p->sq_idx])) \
|
||||||
@@ -2029,13 +2030,11 @@ index 000000000000..acb8657e811d
|
|||||||
+
|
+
|
||||||
+ WARN_ON_ONCE(is_migration_disabled(p));
|
+ WARN_ON_ONCE(is_migration_disabled(p));
|
||||||
+#endif
|
+#endif
|
||||||
|
+ if (task_cpu(p) == new_cpu)
|
||||||
|
+ return;
|
||||||
+ trace_sched_migrate_task(p, new_cpu);
|
+ trace_sched_migrate_task(p, new_cpu);
|
||||||
+
|
+ rseq_migrate(p);
|
||||||
+ if (task_cpu(p) != new_cpu)
|
+ perf_event_task_migrate(p);
|
||||||
+ {
|
|
||||||
+ rseq_migrate(p);
|
|
||||||
+ perf_event_task_migrate(p);
|
|
||||||
+ }
|
|
||||||
+
|
+
|
||||||
+ __set_task_cpu(p, new_cpu);
|
+ __set_task_cpu(p, new_cpu);
|
||||||
+}
|
+}
|
||||||
@@ -5111,15 +5110,15 @@ index 000000000000..acb8657e811d
|
|||||||
+ if (src_rq->nr_running < 2)
|
+ if (src_rq->nr_running < 2)
|
||||||
+ cpumask_clear_cpu(i, &sched_rq_pending_mask);
|
+ cpumask_clear_cpu(i, &sched_rq_pending_mask);
|
||||||
+
|
+
|
||||||
+ spin_release(&src_rq->lock.dep_map, _RET_IP_);
|
|
||||||
+ do_raw_spin_unlock(&src_rq->lock);
|
|
||||||
+
|
|
||||||
+ rq->nr_running += nr_migrated;
|
+ rq->nr_running += nr_migrated;
|
||||||
+ if (rq->nr_running > 1)
|
+ if (rq->nr_running > 1)
|
||||||
+ cpumask_set_cpu(cpu, &sched_rq_pending_mask);
|
+ cpumask_set_cpu(cpu, &sched_rq_pending_mask);
|
||||||
+
|
+
|
||||||
+ cpufreq_update_util(rq, 0);
|
+ cpufreq_update_util(rq, 0);
|
||||||
+
|
+
|
||||||
|
+ spin_release(&src_rq->lock.dep_map, _RET_IP_);
|
||||||
|
+ do_raw_spin_unlock(&src_rq->lock);
|
||||||
|
+
|
||||||
+ return 1;
|
+ return 1;
|
||||||
+ }
|
+ }
|
||||||
+
|
+
|
||||||
@@ -5148,7 +5147,7 @@ index 000000000000..acb8657e811d
|
|||||||
+}
|
+}
|
||||||
+
|
+
|
||||||
+static inline struct task_struct *
|
+static inline struct task_struct *
|
||||||
+choose_next_task(struct rq *rq, int cpu)
|
+choose_next_task(struct rq *rq, int cpu, struct task_struct *prev)
|
||||||
+{
|
+{
|
||||||
+ struct task_struct *next;
|
+ struct task_struct *next;
|
||||||
+
|
+
|
||||||
@@ -5335,7 +5334,7 @@ index 000000000000..acb8657e811d
|
|||||||
+
|
+
|
||||||
+ check_curr(prev, rq);
|
+ check_curr(prev, rq);
|
||||||
+
|
+
|
||||||
+ next = choose_next_task(rq, cpu);
|
+ next = choose_next_task(rq, cpu, prev);
|
||||||
+ clear_tsk_need_resched(prev);
|
+ clear_tsk_need_resched(prev);
|
||||||
+ clear_preempt_need_resched();
|
+ clear_preempt_need_resched();
|
||||||
+#ifdef CONFIG_SCHED_DEBUG
|
+#ifdef CONFIG_SCHED_DEBUG
|
||||||
@@ -5765,7 +5764,6 @@ index 000000000000..acb8657e811d
|
|||||||
+ return;
|
+ return;
|
||||||
+
|
+
|
||||||
+ rq = __task_access_lock(p, &lock);
|
+ rq = __task_access_lock(p, &lock);
|
||||||
+ update_rq_clock(rq);
|
|
||||||
+ /*
|
+ /*
|
||||||
+ * Set under pi_lock && rq->lock, such that the value can be used under
|
+ * Set under pi_lock && rq->lock, such that the value can be used under
|
||||||
+ * either lock.
|
+ * either lock.
|
||||||
@@ -6645,13 +6643,6 @@ index 000000000000..acb8657e811d
|
|||||||
+ return retval;
|
+ return retval;
|
||||||
+}
|
+}
|
||||||
+
|
+
|
||||||
+#ifdef CONFIG_SMP
|
|
||||||
+int dl_task_check_affinity(struct task_struct *p, const struct cpumask *mask)
|
|
||||||
+{
|
|
||||||
+ return 0;
|
|
||||||
+}
|
|
||||||
+#endif
|
|
||||||
+
|
|
||||||
+static int
|
+static int
|
||||||
+__sched_setaffinity(struct task_struct *p, const struct cpumask *mask)
|
+__sched_setaffinity(struct task_struct *p, const struct cpumask *mask)
|
||||||
+{
|
+{
|
||||||
@@ -8036,14 +8027,6 @@ index 000000000000..acb8657e811d
|
|||||||
+
|
+
|
||||||
+ sched_smp_initialized = true;
|
+ sched_smp_initialized = true;
|
||||||
+}
|
+}
|
||||||
+
|
|
||||||
+static int __init migration_init(void)
|
|
||||||
+{
|
|
||||||
+ sched_cpu_starting(smp_processor_id());
|
|
||||||
+ return 0;
|
|
||||||
+}
|
|
||||||
+early_initcall(migration_init);
|
|
||||||
+
|
|
||||||
+#else
|
+#else
|
||||||
+void __init sched_init_smp(void)
|
+void __init sched_init_smp(void)
|
||||||
+{
|
+{
|
||||||
@@ -8653,15 +8636,14 @@ index 000000000000..1212a031700e
|
|||||||
+{}
|
+{}
|
||||||
diff --git a/kernel/sched/alt_sched.h b/kernel/sched/alt_sched.h
|
diff --git a/kernel/sched/alt_sched.h b/kernel/sched/alt_sched.h
|
||||||
new file mode 100644
|
new file mode 100644
|
||||||
index 000000000000..c32403ed82b6
|
index 000000000000..e3b6320a397a
|
||||||
--- /dev/null
|
--- /dev/null
|
||||||
+++ b/kernel/sched/alt_sched.h
|
+++ b/kernel/sched/alt_sched.h
|
||||||
@@ -0,0 +1,668 @@
|
@@ -0,0 +1,667 @@
|
||||||
+#ifndef ALT_SCHED_H
|
+#ifndef ALT_SCHED_H
|
||||||
+#define ALT_SCHED_H
|
+#define ALT_SCHED_H
|
||||||
+
|
+
|
||||||
+#include <linux/context_tracking.h>
|
+#include <linux/context_tracking.h>
|
||||||
+#include <linux/profile.h>
|
|
||||||
+#include <linux/psi.h>
|
+#include <linux/psi.h>
|
||||||
+#include <linux/stop_machine.h>
|
+#include <linux/stop_machine.h>
|
||||||
+#include <linux/syscalls.h>
|
+#include <linux/syscalls.h>
|
||||||
@@ -9446,23 +9428,23 @@ index d9dc9ab3773f..71a25540d65e 100644
|
|||||||
--- a/kernel/sched/build_policy.c
|
--- a/kernel/sched/build_policy.c
|
||||||
+++ b/kernel/sched/build_policy.c
|
+++ b/kernel/sched/build_policy.c
|
||||||
@@ -42,13 +42,19 @@
|
@@ -42,13 +42,19 @@
|
||||||
|
|
||||||
#include "idle.c"
|
#include "idle.c"
|
||||||
|
|
||||||
+#ifndef CONFIG_SCHED_ALT
|
+#ifndef CONFIG_SCHED_ALT
|
||||||
#include "rt.c"
|
#include "rt.c"
|
||||||
+#endif
|
+#endif
|
||||||
|
|
||||||
#ifdef CONFIG_SMP
|
#ifdef CONFIG_SMP
|
||||||
+#ifndef CONFIG_SCHED_ALT
|
+#ifndef CONFIG_SCHED_ALT
|
||||||
# include "cpudeadline.c"
|
# include "cpudeadline.c"
|
||||||
+#endif
|
+#endif
|
||||||
# include "pelt.c"
|
# include "pelt.c"
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
#include "cputime.c"
|
#include "cputime.c"
|
||||||
-#include "deadline.c"
|
-#include "deadline.c"
|
||||||
|
|
||||||
+#ifndef CONFIG_SCHED_ALT
|
+#ifndef CONFIG_SCHED_ALT
|
||||||
+#include "deadline.c"
|
+#include "deadline.c"
|
||||||
+#endif
|
+#endif
|
||||||
@@ -9471,7 +9453,7 @@ index 99bdd96f454f..23f80a86d2d7 100644
|
|||||||
--- a/kernel/sched/build_utility.c
|
--- a/kernel/sched/build_utility.c
|
||||||
+++ b/kernel/sched/build_utility.c
|
+++ b/kernel/sched/build_utility.c
|
||||||
@@ -85,7 +85,9 @@
|
@@ -85,7 +85,9 @@
|
||||||
|
|
||||||
#ifdef CONFIG_SMP
|
#ifdef CONFIG_SMP
|
||||||
# include "cpupri.c"
|
# include "cpupri.c"
|
||||||
+#ifndef CONFIG_SCHED_ALT
|
+#ifndef CONFIG_SCHED_ALT
|
||||||
@@ -9479,14 +9461,14 @@ index 99bdd96f454f..23f80a86d2d7 100644
|
|||||||
+#endif
|
+#endif
|
||||||
# include "topology.c"
|
# include "topology.c"
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
diff --git a/kernel/sched/cpufreq_schedutil.c b/kernel/sched/cpufreq_schedutil.c
|
diff --git a/kernel/sched/cpufreq_schedutil.c b/kernel/sched/cpufreq_schedutil.c
|
||||||
index 1207c78f85c1..68812e0756cb 100644
|
index 1207c78f85c1..68812e0756cb 100644
|
||||||
--- a/kernel/sched/cpufreq_schedutil.c
|
--- a/kernel/sched/cpufreq_schedutil.c
|
||||||
+++ b/kernel/sched/cpufreq_schedutil.c
|
+++ b/kernel/sched/cpufreq_schedutil.c
|
||||||
@@ -159,9 +159,14 @@ static void sugov_get_util(struct sugov_cpu *sg_cpu)
|
@@ -159,9 +159,14 @@ static void sugov_get_util(struct sugov_cpu *sg_cpu)
|
||||||
struct rq *rq = cpu_rq(sg_cpu->cpu);
|
struct rq *rq = cpu_rq(sg_cpu->cpu);
|
||||||
|
|
||||||
sg_cpu->max = arch_scale_cpu_capacity(sg_cpu->cpu);
|
sg_cpu->max = arch_scale_cpu_capacity(sg_cpu->cpu);
|
||||||
+#ifndef CONFIG_SCHED_ALT
|
+#ifndef CONFIG_SCHED_ALT
|
||||||
sg_cpu->bw_dl = cpu_bw_dl(rq);
|
sg_cpu->bw_dl = cpu_bw_dl(rq);
|
||||||
@@ -9497,7 +9479,7 @@ index 1207c78f85c1..68812e0756cb 100644
|
|||||||
+ sg_cpu->util = rq_load_util(rq, sg_cpu->max);
|
+ sg_cpu->util = rq_load_util(rq, sg_cpu->max);
|
||||||
+#endif /* CONFIG_SCHED_ALT */
|
+#endif /* CONFIG_SCHED_ALT */
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
@@ -305,8 +310,10 @@ static inline bool sugov_cpu_is_busy(struct sugov_cpu *sg_cpu) { return false; }
|
@@ -305,8 +310,10 @@ static inline bool sugov_cpu_is_busy(struct sugov_cpu *sg_cpu) { return false; }
|
||||||
*/
|
*/
|
||||||
@@ -9508,11 +9490,11 @@ index 1207c78f85c1..68812e0756cb 100644
|
|||||||
sg_cpu->sg_policy->limits_changed = true;
|
sg_cpu->sg_policy->limits_changed = true;
|
||||||
+#endif
|
+#endif
|
||||||
}
|
}
|
||||||
|
|
||||||
static inline bool sugov_update_single_common(struct sugov_cpu *sg_cpu,
|
static inline bool sugov_update_single_common(struct sugov_cpu *sg_cpu,
|
||||||
@@ -606,6 +613,7 @@ static int sugov_kthread_create(struct sugov_policy *sg_policy)
|
@@ -606,6 +613,7 @@ static int sugov_kthread_create(struct sugov_policy *sg_policy)
|
||||||
}
|
}
|
||||||
|
|
||||||
ret = sched_setattr_nocheck(thread, &attr);
|
ret = sched_setattr_nocheck(thread, &attr);
|
||||||
+
|
+
|
||||||
if (ret) {
|
if (ret) {
|
||||||
@@ -9527,7 +9509,7 @@ index 1207c78f85c1..68812e0756cb 100644
|
|||||||
+#endif /* CONFIG_SCHED_ALT */
|
+#endif /* CONFIG_SCHED_ALT */
|
||||||
}
|
}
|
||||||
static DECLARE_WORK(rebuild_sd_work, rebuild_sd_workfn);
|
static DECLARE_WORK(rebuild_sd_work, rebuild_sd_workfn);
|
||||||
|
|
||||||
diff --git a/kernel/sched/cputime.c b/kernel/sched/cputime.c
|
diff --git a/kernel/sched/cputime.c b/kernel/sched/cputime.c
|
||||||
index 95fc77853743..b48b3f9ed47f 100644
|
index 95fc77853743..b48b3f9ed47f 100644
|
||||||
--- a/kernel/sched/cputime.c
|
--- a/kernel/sched/cputime.c
|
||||||
@@ -9535,15 +9517,15 @@ index 95fc77853743..b48b3f9ed47f 100644
|
|||||||
@@ -122,7 +122,7 @@ void account_user_time(struct task_struct *p, u64 cputime)
|
@@ -122,7 +122,7 @@ void account_user_time(struct task_struct *p, u64 cputime)
|
||||||
p->utime += cputime;
|
p->utime += cputime;
|
||||||
account_group_user_time(p, cputime);
|
account_group_user_time(p, cputime);
|
||||||
|
|
||||||
- index = (task_nice(p) > 0) ? CPUTIME_NICE : CPUTIME_USER;
|
- index = (task_nice(p) > 0) ? CPUTIME_NICE : CPUTIME_USER;
|
||||||
+ index = task_running_nice(p) ? CPUTIME_NICE : CPUTIME_USER;
|
+ index = task_running_nice(p) ? CPUTIME_NICE : CPUTIME_USER;
|
||||||
|
|
||||||
/* Add user time to cpustat. */
|
/* Add user time to cpustat. */
|
||||||
task_group_account_field(p, index, cputime);
|
task_group_account_field(p, index, cputime);
|
||||||
@@ -146,7 +146,7 @@ void account_guest_time(struct task_struct *p, u64 cputime)
|
@@ -146,7 +146,7 @@ void account_guest_time(struct task_struct *p, u64 cputime)
|
||||||
p->gtime += cputime;
|
p->gtime += cputime;
|
||||||
|
|
||||||
/* Add guest time to cpustat. */
|
/* Add guest time to cpustat. */
|
||||||
- if (task_nice(p) > 0) {
|
- if (task_nice(p) > 0) {
|
||||||
+ if (task_running_nice(p)) {
|
+ if (task_running_nice(p)) {
|
||||||
@@ -9561,12 +9543,12 @@ index 95fc77853743..b48b3f9ed47f 100644
|
|||||||
static u64 read_sum_exec_runtime(struct task_struct *t)
|
static u64 read_sum_exec_runtime(struct task_struct *t)
|
||||||
@@ -294,7 +294,7 @@ static u64 read_sum_exec_runtime(struct task_struct *t)
|
@@ -294,7 +294,7 @@ static u64 read_sum_exec_runtime(struct task_struct *t)
|
||||||
struct rq *rq;
|
struct rq *rq;
|
||||||
|
|
||||||
rq = task_rq_lock(t, &rf);
|
rq = task_rq_lock(t, &rf);
|
||||||
- ns = t->se.sum_exec_runtime;
|
- ns = t->se.sum_exec_runtime;
|
||||||
+ ns = tsk_seruntime(t);
|
+ ns = tsk_seruntime(t);
|
||||||
task_rq_unlock(rq, t, &rf);
|
task_rq_unlock(rq, t, &rf);
|
||||||
|
|
||||||
return ns;
|
return ns;
|
||||||
@@ -626,7 +626,7 @@ void cputime_adjust(struct task_cputime *curr, struct prev_cputime *prev,
|
@@ -626,7 +626,7 @@ void cputime_adjust(struct task_cputime *curr, struct prev_cputime *prev,
|
||||||
void task_cputime_adjusted(struct task_struct *p, u64 *ut, u64 *st)
|
void task_cputime_adjusted(struct task_struct *p, u64 *ut, u64 *st)
|
||||||
@@ -9575,7 +9557,7 @@ index 95fc77853743..b48b3f9ed47f 100644
|
|||||||
- .sum_exec_runtime = p->se.sum_exec_runtime,
|
- .sum_exec_runtime = p->se.sum_exec_runtime,
|
||||||
+ .sum_exec_runtime = tsk_seruntime(p),
|
+ .sum_exec_runtime = tsk_seruntime(p),
|
||||||
};
|
};
|
||||||
|
|
||||||
if (task_cputime(p, &cputime.utime, &cputime.stime))
|
if (task_cputime(p, &cputime.utime, &cputime.stime))
|
||||||
diff --git a/kernel/sched/debug.c b/kernel/sched/debug.c
|
diff --git a/kernel/sched/debug.c b/kernel/sched/debug.c
|
||||||
index 1637b65ba07a..033c6deeb515 100644
|
index 1637b65ba07a..033c6deeb515 100644
|
||||||
@@ -9584,39 +9566,39 @@ index 1637b65ba07a..033c6deeb515 100644
|
|||||||
@@ -7,6 +7,7 @@
|
@@ -7,6 +7,7 @@
|
||||||
* Copyright(C) 2007, Red Hat, Inc., Ingo Molnar
|
* Copyright(C) 2007, Red Hat, Inc., Ingo Molnar
|
||||||
*/
|
*/
|
||||||
|
|
||||||
+#ifndef CONFIG_SCHED_ALT
|
+#ifndef CONFIG_SCHED_ALT
|
||||||
/*
|
/*
|
||||||
* This allows printing both to /proc/sched_debug and
|
* This allows printing both to /proc/sched_debug and
|
||||||
* to the console
|
* to the console
|
||||||
@@ -215,6 +216,7 @@ static const struct file_operations sched_scaling_fops = {
|
@@ -215,6 +216,7 @@ static const struct file_operations sched_scaling_fops = {
|
||||||
};
|
};
|
||||||
|
|
||||||
#endif /* SMP */
|
#endif /* SMP */
|
||||||
+#endif /* !CONFIG_SCHED_ALT */
|
+#endif /* !CONFIG_SCHED_ALT */
|
||||||
|
|
||||||
#ifdef CONFIG_PREEMPT_DYNAMIC
|
#ifdef CONFIG_PREEMPT_DYNAMIC
|
||||||
|
|
||||||
@@ -278,6 +280,7 @@ static const struct file_operations sched_dynamic_fops = {
|
@@ -278,6 +280,7 @@ static const struct file_operations sched_dynamic_fops = {
|
||||||
|
|
||||||
#endif /* CONFIG_PREEMPT_DYNAMIC */
|
#endif /* CONFIG_PREEMPT_DYNAMIC */
|
||||||
|
|
||||||
+#ifndef CONFIG_SCHED_ALT
|
+#ifndef CONFIG_SCHED_ALT
|
||||||
__read_mostly bool sched_debug_verbose;
|
__read_mostly bool sched_debug_verbose;
|
||||||
|
|
||||||
static const struct seq_operations sched_debug_sops;
|
static const struct seq_operations sched_debug_sops;
|
||||||
@@ -293,6 +296,7 @@ static const struct file_operations sched_debug_fops = {
|
@@ -293,6 +296,7 @@ static const struct file_operations sched_debug_fops = {
|
||||||
.llseek = seq_lseek,
|
.llseek = seq_lseek,
|
||||||
.release = seq_release,
|
.release = seq_release,
|
||||||
};
|
};
|
||||||
+#endif /* !CONFIG_SCHED_ALT */
|
+#endif /* !CONFIG_SCHED_ALT */
|
||||||
|
|
||||||
static struct dentry *debugfs_sched;
|
static struct dentry *debugfs_sched;
|
||||||
|
|
||||||
@@ -302,12 +306,15 @@ static __init int sched_init_debug(void)
|
@@ -302,12 +306,15 @@ static __init int sched_init_debug(void)
|
||||||
|
|
||||||
debugfs_sched = debugfs_create_dir("sched", NULL);
|
debugfs_sched = debugfs_create_dir("sched", NULL);
|
||||||
|
|
||||||
+#ifndef CONFIG_SCHED_ALT
|
+#ifndef CONFIG_SCHED_ALT
|
||||||
debugfs_create_file("features", 0644, debugfs_sched, NULL, &sched_feat_fops);
|
debugfs_create_file("features", 0644, debugfs_sched, NULL, &sched_feat_fops);
|
||||||
debugfs_create_bool("verbose", 0644, debugfs_sched, &sched_debug_verbose);
|
debugfs_create_bool("verbose", 0644, debugfs_sched, &sched_debug_verbose);
|
||||||
@@ -9624,31 +9606,31 @@ index 1637b65ba07a..033c6deeb515 100644
|
|||||||
#ifdef CONFIG_PREEMPT_DYNAMIC
|
#ifdef CONFIG_PREEMPT_DYNAMIC
|
||||||
debugfs_create_file("preempt", 0644, debugfs_sched, NULL, &sched_dynamic_fops);
|
debugfs_create_file("preempt", 0644, debugfs_sched, NULL, &sched_dynamic_fops);
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
+#ifndef CONFIG_SCHED_ALT
|
+#ifndef CONFIG_SCHED_ALT
|
||||||
debugfs_create_u32("latency_ns", 0644, debugfs_sched, &sysctl_sched_latency);
|
debugfs_create_u32("latency_ns", 0644, debugfs_sched, &sysctl_sched_latency);
|
||||||
debugfs_create_u32("min_granularity_ns", 0644, debugfs_sched, &sysctl_sched_min_granularity);
|
debugfs_create_u32("min_granularity_ns", 0644, debugfs_sched, &sysctl_sched_min_granularity);
|
||||||
debugfs_create_u32("idle_min_granularity_ns", 0644, debugfs_sched, &sysctl_sched_idle_min_granularity);
|
debugfs_create_u32("idle_min_granularity_ns", 0644, debugfs_sched, &sysctl_sched_idle_min_granularity);
|
||||||
@@ -337,11 +344,13 @@ static __init int sched_init_debug(void)
|
@@ -337,11 +344,13 @@ static __init int sched_init_debug(void)
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
debugfs_create_file("debug", 0444, debugfs_sched, NULL, &sched_debug_fops);
|
debugfs_create_file("debug", 0444, debugfs_sched, NULL, &sched_debug_fops);
|
||||||
+#endif /* !CONFIG_SCHED_ALT */
|
+#endif /* !CONFIG_SCHED_ALT */
|
||||||
|
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
late_initcall(sched_init_debug);
|
late_initcall(sched_init_debug);
|
||||||
|
|
||||||
+#ifndef CONFIG_SCHED_ALT
|
+#ifndef CONFIG_SCHED_ALT
|
||||||
#ifdef CONFIG_SMP
|
#ifdef CONFIG_SMP
|
||||||
|
|
||||||
static cpumask_var_t sd_sysctl_cpus;
|
static cpumask_var_t sd_sysctl_cpus;
|
||||||
@@ -1068,6 +1077,7 @@ void proc_sched_set_task(struct task_struct *p)
|
@@ -1068,6 +1077,7 @@ void proc_sched_set_task(struct task_struct *p)
|
||||||
memset(&p->stats, 0, sizeof(p->stats));
|
memset(&p->stats, 0, sizeof(p->stats));
|
||||||
#endif
|
#endif
|
||||||
}
|
}
|
||||||
+#endif /* !CONFIG_SCHED_ALT */
|
+#endif /* !CONFIG_SCHED_ALT */
|
||||||
|
|
||||||
void resched_latency_warn(int cpu, u64 latency)
|
void resched_latency_warn(int cpu, u64 latency)
|
||||||
{
|
{
|
||||||
diff --git a/kernel/sched/idle.c b/kernel/sched/idle.c
|
diff --git a/kernel/sched/idle.c b/kernel/sched/idle.c
|
||||||
@@ -9658,7 +9640,7 @@ index f26ab2675f7d..480d4ad16d45 100644
|
|||||||
@@ -400,6 +400,7 @@ void cpu_startup_entry(enum cpuhp_state state)
|
@@ -400,6 +400,7 @@ void cpu_startup_entry(enum cpuhp_state state)
|
||||||
do_idle();
|
do_idle();
|
||||||
}
|
}
|
||||||
|
|
||||||
+#ifndef CONFIG_SCHED_ALT
|
+#ifndef CONFIG_SCHED_ALT
|
||||||
/*
|
/*
|
||||||
* idle-task scheduling class.
|
* idle-task scheduling class.
|
||||||
@@ -9808,17 +9790,17 @@ index 0f310768260c..bd38bf738fe9 100644
|
|||||||
@@ -266,6 +266,7 @@ ___update_load_avg(struct sched_avg *sa, unsigned long load)
|
@@ -266,6 +266,7 @@ ___update_load_avg(struct sched_avg *sa, unsigned long load)
|
||||||
WRITE_ONCE(sa->util_avg, sa->util_sum / divider);
|
WRITE_ONCE(sa->util_avg, sa->util_sum / divider);
|
||||||
}
|
}
|
||||||
|
|
||||||
+#ifndef CONFIG_SCHED_ALT
|
+#ifndef CONFIG_SCHED_ALT
|
||||||
/*
|
/*
|
||||||
* sched_entity:
|
* sched_entity:
|
||||||
*
|
*
|
||||||
@@ -383,8 +384,9 @@ int update_dl_rq_load_avg(u64 now, struct rq *rq, int running)
|
@@ -383,8 +384,9 @@ int update_dl_rq_load_avg(u64 now, struct rq *rq, int running)
|
||||||
|
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
+#endif
|
+#endif
|
||||||
|
|
||||||
-#ifdef CONFIG_SCHED_THERMAL_PRESSURE
|
-#ifdef CONFIG_SCHED_THERMAL_PRESSURE
|
||||||
+#if defined(CONFIG_SCHED_THERMAL_PRESSURE) && !defined(CONFIG_SCHED_ALT)
|
+#if defined(CONFIG_SCHED_THERMAL_PRESSURE) && !defined(CONFIG_SCHED_ALT)
|
||||||
/*
|
/*
|
||||||
@@ -9831,7 +9813,7 @@ index 3a0e0dc28721..e8a7d84aa5a5 100644
|
|||||||
@@ -1,13 +1,15 @@
|
@@ -1,13 +1,15 @@
|
||||||
#ifdef CONFIG_SMP
|
#ifdef CONFIG_SMP
|
||||||
#include "sched-pelt.h"
|
#include "sched-pelt.h"
|
||||||
|
|
||||||
+#ifndef CONFIG_SCHED_ALT
|
+#ifndef CONFIG_SCHED_ALT
|
||||||
int __update_load_avg_blocked_se(u64 now, struct sched_entity *se);
|
int __update_load_avg_blocked_se(u64 now, struct sched_entity *se);
|
||||||
int __update_load_avg_se(u64 now, struct cfs_rq *cfs_rq, struct sched_entity *se);
|
int __update_load_avg_se(u64 now, struct cfs_rq *cfs_rq, struct sched_entity *se);
|
||||||
@@ -9839,16 +9821,16 @@ index 3a0e0dc28721..e8a7d84aa5a5 100644
|
|||||||
int update_rt_rq_load_avg(u64 now, struct rq *rq, int running);
|
int update_rt_rq_load_avg(u64 now, struct rq *rq, int running);
|
||||||
int update_dl_rq_load_avg(u64 now, struct rq *rq, int running);
|
int update_dl_rq_load_avg(u64 now, struct rq *rq, int running);
|
||||||
+#endif
|
+#endif
|
||||||
|
|
||||||
-#ifdef CONFIG_SCHED_THERMAL_PRESSURE
|
-#ifdef CONFIG_SCHED_THERMAL_PRESSURE
|
||||||
+#if defined(CONFIG_SCHED_THERMAL_PRESSURE) && !defined(CONFIG_SCHED_ALT)
|
+#if defined(CONFIG_SCHED_THERMAL_PRESSURE) && !defined(CONFIG_SCHED_ALT)
|
||||||
int update_thermal_load_avg(u64 now, struct rq *rq, u64 capacity);
|
int update_thermal_load_avg(u64 now, struct rq *rq, u64 capacity);
|
||||||
|
|
||||||
static inline u64 thermal_load_avg(struct rq *rq)
|
static inline u64 thermal_load_avg(struct rq *rq)
|
||||||
@@ -44,6 +46,7 @@ static inline u32 get_pelt_divider(struct sched_avg *avg)
|
@@ -44,6 +46,7 @@ static inline u32 get_pelt_divider(struct sched_avg *avg)
|
||||||
return PELT_MIN_DIVIDER + avg->period_contrib;
|
return PELT_MIN_DIVIDER + avg->period_contrib;
|
||||||
}
|
}
|
||||||
|
|
||||||
+#ifndef CONFIG_SCHED_ALT
|
+#ifndef CONFIG_SCHED_ALT
|
||||||
static inline void cfs_se_util_change(struct sched_avg *avg)
|
static inline void cfs_se_util_change(struct sched_avg *avg)
|
||||||
{
|
{
|
||||||
@@ -9858,9 +9840,9 @@ index 3a0e0dc28721..e8a7d84aa5a5 100644
|
|||||||
}
|
}
|
||||||
#endif
|
#endif
|
||||||
+#endif /* CONFIG_SCHED_ALT */
|
+#endif /* CONFIG_SCHED_ALT */
|
||||||
|
|
||||||
#else
|
#else
|
||||||
|
|
||||||
+#ifndef CONFIG_SCHED_ALT
|
+#ifndef CONFIG_SCHED_ALT
|
||||||
static inline int
|
static inline int
|
||||||
update_cfs_rq_load_avg(u64 now, struct cfs_rq *cfs_rq)
|
update_cfs_rq_load_avg(u64 now, struct cfs_rq *cfs_rq)
|
||||||
@@ -9870,7 +9852,7 @@ index 3a0e0dc28721..e8a7d84aa5a5 100644
|
|||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
+#endif
|
+#endif
|
||||||
|
|
||||||
static inline int
|
static inline int
|
||||||
update_thermal_load_avg(u64 now, struct rq *rq, u64 capacity)
|
update_thermal_load_avg(u64 now, struct rq *rq, u64 capacity)
|
||||||
diff --git a/kernel/sched/sched.h b/kernel/sched/sched.h
|
diff --git a/kernel/sched/sched.h b/kernel/sched/sched.h
|
||||||
@@ -9880,7 +9862,7 @@ index a4a20046e586..c363693cd869 100644
|
|||||||
@@ -5,6 +5,10 @@
|
@@ -5,6 +5,10 @@
|
||||||
#ifndef _KERNEL_SCHED_SCHED_H
|
#ifndef _KERNEL_SCHED_SCHED_H
|
||||||
#define _KERNEL_SCHED_SCHED_H
|
#define _KERNEL_SCHED_SCHED_H
|
||||||
|
|
||||||
+#ifdef CONFIG_SCHED_ALT
|
+#ifdef CONFIG_SCHED_ALT
|
||||||
+#include "alt_sched.h"
|
+#include "alt_sched.h"
|
||||||
+#else
|
+#else
|
||||||
@@ -9891,7 +9873,7 @@ index a4a20046e586..c363693cd869 100644
|
|||||||
@@ -3183,4 +3187,9 @@ static inline void update_current_exec_runtime(struct task_struct *curr,
|
@@ -3183,4 +3187,9 @@ static inline void update_current_exec_runtime(struct task_struct *curr,
|
||||||
cgroup_account_cputime(curr, delta_exec);
|
cgroup_account_cputime(curr, delta_exec);
|
||||||
}
|
}
|
||||||
|
|
||||||
+static inline int task_running_nice(struct task_struct *p)
|
+static inline int task_running_nice(struct task_struct *p)
|
||||||
+{
|
+{
|
||||||
+ return (task_nice(p) > 0);
|
+ return (task_nice(p) > 0);
|
||||||
@@ -9915,7 +9897,7 @@ index 857f837f52cb..5486c63e4790 100644
|
|||||||
rq = cpu_rq(cpu);
|
rq = cpu_rq(cpu);
|
||||||
@@ -143,6 +145,7 @@ static int show_schedstat(struct seq_file *seq, void *v)
|
@@ -143,6 +145,7 @@ static int show_schedstat(struct seq_file *seq, void *v)
|
||||||
seq_printf(seq, "\n");
|
seq_printf(seq, "\n");
|
||||||
|
|
||||||
#ifdef CONFIG_SMP
|
#ifdef CONFIG_SMP
|
||||||
+#ifndef CONFIG_SCHED_ALT
|
+#ifndef CONFIG_SCHED_ALT
|
||||||
/* domain-specific stats */
|
/* domain-specific stats */
|
||||||
@@ -9934,9 +9916,9 @@ index 84a188913cc9..53934e7ef5db 100644
|
|||||||
--- a/kernel/sched/stats.h
|
--- a/kernel/sched/stats.h
|
||||||
+++ b/kernel/sched/stats.h
|
+++ b/kernel/sched/stats.h
|
||||||
@@ -89,6 +89,7 @@ static inline void rq_sched_info_depart (struct rq *rq, unsigned long long delt
|
@@ -89,6 +89,7 @@ static inline void rq_sched_info_depart (struct rq *rq, unsigned long long delt
|
||||||
|
|
||||||
#endif /* CONFIG_SCHEDSTATS */
|
#endif /* CONFIG_SCHEDSTATS */
|
||||||
|
|
||||||
+#ifndef CONFIG_SCHED_ALT
|
+#ifndef CONFIG_SCHED_ALT
|
||||||
#ifdef CONFIG_FAIR_GROUP_SCHED
|
#ifdef CONFIG_FAIR_GROUP_SCHED
|
||||||
struct sched_entity_stats {
|
struct sched_entity_stats {
|
||||||
@@ -9946,7 +9928,7 @@ index 84a188913cc9..53934e7ef5db 100644
|
|||||||
return &task_of(se)->stats;
|
return &task_of(se)->stats;
|
||||||
}
|
}
|
||||||
+#endif /* CONFIG_SCHED_ALT */
|
+#endif /* CONFIG_SCHED_ALT */
|
||||||
|
|
||||||
#ifdef CONFIG_PSI
|
#ifdef CONFIG_PSI
|
||||||
void psi_task_change(struct task_struct *task, int clear, int set);
|
void psi_task_change(struct task_struct *task, int clear, int set);
|
||||||
diff --git a/kernel/sched/topology.c b/kernel/sched/topology.c
|
diff --git a/kernel/sched/topology.c b/kernel/sched/topology.c
|
||||||
@@ -9956,37 +9938,37 @@ index 8739c2a5a54e..d8dd6c15eb47 100644
|
|||||||
@@ -3,6 +3,7 @@
|
@@ -3,6 +3,7 @@
|
||||||
* Scheduler topology setup/handling methods
|
* Scheduler topology setup/handling methods
|
||||||
*/
|
*/
|
||||||
|
|
||||||
+#ifndef CONFIG_SCHED_ALT
|
+#ifndef CONFIG_SCHED_ALT
|
||||||
DEFINE_MUTEX(sched_domains_mutex);
|
DEFINE_MUTEX(sched_domains_mutex);
|
||||||
|
|
||||||
/* Protected by sched_domains_mutex: */
|
/* Protected by sched_domains_mutex: */
|
||||||
@@ -1413,8 +1414,10 @@ static void asym_cpu_capacity_scan(void)
|
@@ -1413,8 +1414,10 @@ static void asym_cpu_capacity_scan(void)
|
||||||
*/
|
*/
|
||||||
|
|
||||||
static int default_relax_domain_level = -1;
|
static int default_relax_domain_level = -1;
|
||||||
+#endif /* CONFIG_SCHED_ALT */
|
+#endif /* CONFIG_SCHED_ALT */
|
||||||
int sched_domain_level_max;
|
int sched_domain_level_max;
|
||||||
|
|
||||||
+#ifndef CONFIG_SCHED_ALT
|
+#ifndef CONFIG_SCHED_ALT
|
||||||
static int __init setup_relax_domain_level(char *str)
|
static int __init setup_relax_domain_level(char *str)
|
||||||
{
|
{
|
||||||
if (kstrtoint(str, 0, &default_relax_domain_level))
|
if (kstrtoint(str, 0, &default_relax_domain_level))
|
||||||
@@ -1647,6 +1650,7 @@ sd_init(struct sched_domain_topology_level *tl,
|
@@ -1647,6 +1650,7 @@ sd_init(struct sched_domain_topology_level *tl,
|
||||||
|
|
||||||
return sd;
|
return sd;
|
||||||
}
|
}
|
||||||
+#endif /* CONFIG_SCHED_ALT */
|
+#endif /* CONFIG_SCHED_ALT */
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Topology list, bottom-up.
|
* Topology list, bottom-up.
|
||||||
@@ -1683,6 +1687,7 @@ void set_sched_topology(struct sched_domain_topology_level *tl)
|
@@ -1683,6 +1687,7 @@ void set_sched_topology(struct sched_domain_topology_level *tl)
|
||||||
sched_domain_topology_saved = NULL;
|
sched_domain_topology_saved = NULL;
|
||||||
}
|
}
|
||||||
|
|
||||||
+#ifndef CONFIG_SCHED_ALT
|
+#ifndef CONFIG_SCHED_ALT
|
||||||
#ifdef CONFIG_NUMA
|
#ifdef CONFIG_NUMA
|
||||||
|
|
||||||
static const struct cpumask *sd_numa_mask(int cpu)
|
static const struct cpumask *sd_numa_mask(int cpu)
|
||||||
@@ -2645,3 +2650,15 @@ void partition_sched_domains(int ndoms_new, cpumask_var_t doms_new[],
|
@@ -2645,3 +2650,15 @@ void partition_sched_domains(int ndoms_new, cpumask_var_t doms_new[],
|
||||||
partition_sched_domains_locked(ndoms_new, doms_new, dattr_new);
|
partition_sched_domains_locked(ndoms_new, doms_new, dattr_new);
|
||||||
@@ -10009,9 +9991,9 @@ index c6d9dec11b74..2bc42ce8b48e 100644
|
|||||||
--- a/kernel/sysctl.c
|
--- a/kernel/sysctl.c
|
||||||
+++ b/kernel/sysctl.c
|
+++ b/kernel/sysctl.c
|
||||||
@@ -93,6 +93,10 @@ EXPORT_SYMBOL_GPL(sysctl_long_vals);
|
@@ -93,6 +93,10 @@ EXPORT_SYMBOL_GPL(sysctl_long_vals);
|
||||||
|
|
||||||
/* Constants used for minimum and maximum */
|
/* Constants used for minimum and maximum */
|
||||||
|
|
||||||
+#ifdef CONFIG_SCHED_ALT
|
+#ifdef CONFIG_SCHED_ALT
|
||||||
+extern int sched_yield_type;
|
+extern int sched_yield_type;
|
||||||
+#endif
|
+#endif
|
||||||
@@ -10021,7 +10003,7 @@ index c6d9dec11b74..2bc42ce8b48e 100644
|
|||||||
#endif
|
#endif
|
||||||
@@ -1633,6 +1637,7 @@ int proc_do_static_key(struct ctl_table *table, int write,
|
@@ -1633,6 +1637,7 @@ int proc_do_static_key(struct ctl_table *table, int write,
|
||||||
}
|
}
|
||||||
|
|
||||||
static struct ctl_table kern_table[] = {
|
static struct ctl_table kern_table[] = {
|
||||||
+#ifndef CONFIG_SCHED_ALT
|
+#ifndef CONFIG_SCHED_ALT
|
||||||
#ifdef CONFIG_NUMA_BALANCING
|
#ifdef CONFIG_NUMA_BALANCING
|
||||||
@@ -10060,13 +10042,13 @@ index 3ae661ab6260..35f0176dcdb0 100644
|
|||||||
@@ -2088,8 +2088,10 @@ long hrtimer_nanosleep(ktime_t rqtp, const enum hrtimer_mode mode,
|
@@ -2088,8 +2088,10 @@ long hrtimer_nanosleep(ktime_t rqtp, const enum hrtimer_mode mode,
|
||||||
int ret = 0;
|
int ret = 0;
|
||||||
u64 slack;
|
u64 slack;
|
||||||
|
|
||||||
+#ifndef CONFIG_SCHED_ALT
|
+#ifndef CONFIG_SCHED_ALT
|
||||||
slack = current->timer_slack_ns;
|
slack = current->timer_slack_ns;
|
||||||
if (dl_task(current) || rt_task(current))
|
if (dl_task(current) || rt_task(current))
|
||||||
+#endif
|
+#endif
|
||||||
slack = 0;
|
slack = 0;
|
||||||
|
|
||||||
hrtimer_init_sleeper_on_stack(&t, clockid, mode);
|
hrtimer_init_sleeper_on_stack(&t, clockid, mode);
|
||||||
diff --git a/kernel/time/posix-cpu-timers.c b/kernel/time/posix-cpu-timers.c
|
diff --git a/kernel/time/posix-cpu-timers.c b/kernel/time/posix-cpu-timers.c
|
||||||
index cb925e8ef9a8..67d823510f5c 100644
|
index cb925e8ef9a8..67d823510f5c 100644
|
||||||
@@ -10074,17 +10056,17 @@ index cb925e8ef9a8..67d823510f5c 100644
|
|||||||
+++ b/kernel/time/posix-cpu-timers.c
|
+++ b/kernel/time/posix-cpu-timers.c
|
||||||
@@ -223,7 +223,7 @@ static void task_sample_cputime(struct task_struct *p, u64 *samples)
|
@@ -223,7 +223,7 @@ static void task_sample_cputime(struct task_struct *p, u64 *samples)
|
||||||
u64 stime, utime;
|
u64 stime, utime;
|
||||||
|
|
||||||
task_cputime(p, &utime, &stime);
|
task_cputime(p, &utime, &stime);
|
||||||
- store_samples(samples, stime, utime, p->se.sum_exec_runtime);
|
- store_samples(samples, stime, utime, p->se.sum_exec_runtime);
|
||||||
+ store_samples(samples, stime, utime, tsk_seruntime(p));
|
+ store_samples(samples, stime, utime, tsk_seruntime(p));
|
||||||
}
|
}
|
||||||
|
|
||||||
static void proc_sample_cputime_atomic(struct task_cputime_atomic *at,
|
static void proc_sample_cputime_atomic(struct task_cputime_atomic *at,
|
||||||
@@ -866,6 +866,7 @@ static void collect_posix_cputimers(struct posix_cputimers *pct, u64 *samples,
|
@@ -866,6 +866,7 @@ static void collect_posix_cputimers(struct posix_cputimers *pct, u64 *samples,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
+#ifndef CONFIG_SCHED_ALT
|
+#ifndef CONFIG_SCHED_ALT
|
||||||
static inline void check_dl_overrun(struct task_struct *tsk)
|
static inline void check_dl_overrun(struct task_struct *tsk)
|
||||||
{
|
{
|
||||||
@@ -10094,18 +10076,18 @@ index cb925e8ef9a8..67d823510f5c 100644
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
+#endif
|
+#endif
|
||||||
|
|
||||||
static bool check_rlimit(u64 time, u64 limit, int signo, bool rt, bool hard)
|
static bool check_rlimit(u64 time, u64 limit, int signo, bool rt, bool hard)
|
||||||
{
|
{
|
||||||
@@ -900,8 +902,10 @@ static void check_thread_timers(struct task_struct *tsk,
|
@@ -900,8 +902,10 @@ static void check_thread_timers(struct task_struct *tsk,
|
||||||
u64 samples[CPUCLOCK_MAX];
|
u64 samples[CPUCLOCK_MAX];
|
||||||
unsigned long soft;
|
unsigned long soft;
|
||||||
|
|
||||||
+#ifndef CONFIG_SCHED_ALT
|
+#ifndef CONFIG_SCHED_ALT
|
||||||
if (dl_task(tsk))
|
if (dl_task(tsk))
|
||||||
check_dl_overrun(tsk);
|
check_dl_overrun(tsk);
|
||||||
+#endif
|
+#endif
|
||||||
|
|
||||||
if (expiry_cache_is_inactive(pct))
|
if (expiry_cache_is_inactive(pct))
|
||||||
return;
|
return;
|
||||||
@@ -915,7 +919,7 @@ static void check_thread_timers(struct task_struct *tsk,
|
@@ -915,7 +919,7 @@ static void check_thread_timers(struct task_struct *tsk,
|
||||||
@@ -10115,17 +10097,17 @@ index cb925e8ef9a8..67d823510f5c 100644
|
|||||||
- unsigned long rttime = tsk->rt.timeout * (USEC_PER_SEC / HZ);
|
- unsigned long rttime = tsk->rt.timeout * (USEC_PER_SEC / HZ);
|
||||||
+ unsigned long rttime = tsk_rttimeout(tsk) * (USEC_PER_SEC / HZ);
|
+ unsigned long rttime = tsk_rttimeout(tsk) * (USEC_PER_SEC / HZ);
|
||||||
unsigned long hard = task_rlimit_max(tsk, RLIMIT_RTTIME);
|
unsigned long hard = task_rlimit_max(tsk, RLIMIT_RTTIME);
|
||||||
|
|
||||||
/* At the hard limit, send SIGKILL. No further action. */
|
/* At the hard limit, send SIGKILL. No further action. */
|
||||||
@@ -1151,8 +1155,10 @@ static inline bool fastpath_timer_check(struct task_struct *tsk)
|
@@ -1151,8 +1155,10 @@ static inline bool fastpath_timer_check(struct task_struct *tsk)
|
||||||
return true;
|
return true;
|
||||||
}
|
}
|
||||||
|
|
||||||
+#ifndef CONFIG_SCHED_ALT
|
+#ifndef CONFIG_SCHED_ALT
|
||||||
if (dl_task(tsk) && tsk->dl.dl_overrun)
|
if (dl_task(tsk) && tsk->dl.dl_overrun)
|
||||||
return true;
|
return true;
|
||||||
+#endif
|
+#endif
|
||||||
|
|
||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
diff --git a/kernel/trace/trace_selftest.c b/kernel/trace/trace_selftest.c
|
diff --git a/kernel/trace/trace_selftest.c b/kernel/trace/trace_selftest.c
|
||||||
@@ -10147,4 +10129,4 @@ index a2d301f58ced..2ccdede8585c 100644
|
|||||||
+#endif
|
+#endif
|
||||||
};
|
};
|
||||||
struct wakeup_test_data *x = data;
|
struct wakeup_test_data *x = data;
|
||||||
|
|
File diff suppressed because it is too large
Load Diff
@@ -358,3 +358,263 @@ index c0cd1b9..af1e2fb 100644
|
|||||||
--
|
--
|
||||||
https://clearlinux.org
|
https://clearlinux.org
|
||||||
|
|
||||||
|
From 676c2dc63592f52b716515573a3a825582a371e9 Mon Sep 17 00:00:00 2001
|
||||||
|
From: Arjan van de Ven <arjan@linux.intel.com>
|
||||||
|
Date: Sat, 8 Dec 2018 18:21:32 +0000
|
||||||
|
Subject: [PATCH 1/9] x86/vdso: Use lfence instead of rep and nop
|
||||||
|
|
||||||
|
Signed-off-by: Alexandre Frade <kernel@xanmod.org>
|
||||||
|
---
|
||||||
|
arch/x86/include/asm/vdso/processor.h | 2 +-
|
||||||
|
1 file changed, 1 insertion(+), 1 deletion(-)
|
||||||
|
|
||||||
|
diff --git a/arch/x86/include/asm/vdso/processor.h b/arch/x86/include/asm/vdso/processor.h
|
||||||
|
index 57b1a7034c64..e2c45674f989 100644
|
||||||
|
--- a/arch/x86/include/asm/vdso/processor.h
|
||||||
|
+++ b/arch/x86/include/asm/vdso/processor.h
|
||||||
|
@@ -10,7 +10,7 @@
|
||||||
|
/* REP NOP (PAUSE) is a good thing to insert into busy-wait loops. */
|
||||||
|
static __always_inline void rep_nop(void)
|
||||||
|
{
|
||||||
|
- asm volatile("rep; nop" ::: "memory");
|
||||||
|
+ asm volatile("lfence" ::: "memory");
|
||||||
|
}
|
||||||
|
|
||||||
|
static __always_inline void cpu_relax(void)
|
||||||
|
--
|
||||||
|
2.39.1
|
||||||
|
|
||||||
|
From 48dc9669f8db68adc480ffc2698ed8204440e45b Mon Sep 17 00:00:00 2001
|
||||||
|
From: Arjan van de Ven <arjan@linux.intel.com>
|
||||||
|
Date: Thu, 13 Dec 2018 01:00:49 +0000
|
||||||
|
Subject: [PATCH 2/9] sched/wait: Do accept() in LIFO order for cache
|
||||||
|
efficiency
|
||||||
|
|
||||||
|
Signed-off-by: Alexandre Frade <kernel@xanmod.org>
|
||||||
|
---
|
||||||
|
include/linux/wait.h | 2 ++
|
||||||
|
kernel/sched/wait.c | 24 ++++++++++++++++++++++++
|
||||||
|
net/ipv4/inet_connection_sock.c | 2 +-
|
||||||
|
3 files changed, 27 insertions(+), 1 deletion(-)
|
||||||
|
|
||||||
|
diff --git a/include/linux/wait.h b/include/linux/wait.h
|
||||||
|
index a0307b516b09..edc21128f387 100644
|
||||||
|
--- a/include/linux/wait.h
|
||||||
|
+++ b/include/linux/wait.h
|
||||||
|
@@ -165,6 +165,7 @@ static inline bool wq_has_sleeper(struct wait_queue_head *wq_head)
|
||||||
|
|
||||||
|
extern void add_wait_queue(struct wait_queue_head *wq_head, struct wait_queue_entry *wq_entry);
|
||||||
|
extern void add_wait_queue_exclusive(struct wait_queue_head *wq_head, struct wait_queue_entry *wq_entry);
|
||||||
|
+extern void add_wait_queue_exclusive_lifo(struct wait_queue_head *wq_head, struct wait_queue_entry *wq_entry);
|
||||||
|
extern void add_wait_queue_priority(struct wait_queue_head *wq_head, struct wait_queue_entry *wq_entry);
|
||||||
|
extern void remove_wait_queue(struct wait_queue_head *wq_head, struct wait_queue_entry *wq_entry);
|
||||||
|
|
||||||
|
@@ -1192,6 +1193,7 @@ do { \
|
||||||
|
*/
|
||||||
|
void prepare_to_wait(struct wait_queue_head *wq_head, struct wait_queue_entry *wq_entry, int state);
|
||||||
|
bool prepare_to_wait_exclusive(struct wait_queue_head *wq_head, struct wait_queue_entry *wq_entry, int state);
|
||||||
|
+void prepare_to_wait_exclusive_lifo(struct wait_queue_head *wq_head, struct wait_queue_entry *wq_entry, int state);
|
||||||
|
long prepare_to_wait_event(struct wait_queue_head *wq_head, struct wait_queue_entry *wq_entry, int state);
|
||||||
|
void finish_wait(struct wait_queue_head *wq_head, struct wait_queue_entry *wq_entry);
|
||||||
|
long wait_woken(struct wait_queue_entry *wq_entry, unsigned mode, long timeout);
|
||||||
|
diff --git a/kernel/sched/wait.c b/kernel/sched/wait.c
|
||||||
|
index 133b74730738..1647fb8662eb 100644
|
||||||
|
--- a/kernel/sched/wait.c
|
||||||
|
+++ b/kernel/sched/wait.c
|
||||||
|
@@ -47,6 +47,17 @@ void add_wait_queue_priority(struct wait_queue_head *wq_head, struct wait_queue_
|
||||||
|
}
|
||||||
|
EXPORT_SYMBOL_GPL(add_wait_queue_priority);
|
||||||
|
|
||||||
|
+void add_wait_queue_exclusive_lifo(struct wait_queue_head *wq_head, struct wait_queue_entry *wq_entry)
|
||||||
|
+{
|
||||||
|
+ unsigned long flags;
|
||||||
|
+
|
||||||
|
+ wq_entry->flags |= WQ_FLAG_EXCLUSIVE;
|
||||||
|
+ spin_lock_irqsave(&wq_head->lock, flags);
|
||||||
|
+ __add_wait_queue(wq_head, wq_entry);
|
||||||
|
+ spin_unlock_irqrestore(&wq_head->lock, flags);
|
||||||
|
+}
|
||||||
|
+EXPORT_SYMBOL(add_wait_queue_exclusive_lifo);
|
||||||
|
+
|
||||||
|
void remove_wait_queue(struct wait_queue_head *wq_head, struct wait_queue_entry *wq_entry)
|
||||||
|
{
|
||||||
|
unsigned long flags;
|
||||||
|
@@ -293,6 +304,19 @@ prepare_to_wait_exclusive(struct wait_queue_head *wq_head, struct wait_queue_ent
|
||||||
|
}
|
||||||
|
EXPORT_SYMBOL(prepare_to_wait_exclusive);
|
||||||
|
|
||||||
|
+void prepare_to_wait_exclusive_lifo(struct wait_queue_head *wq_head, struct wait_queue_entry *wq_entry, int state)
|
||||||
|
+{
|
||||||
|
+ unsigned long flags;
|
||||||
|
+
|
||||||
|
+ wq_entry->flags |= WQ_FLAG_EXCLUSIVE;
|
||||||
|
+ spin_lock_irqsave(&wq_head->lock, flags);
|
||||||
|
+ if (list_empty(&wq_entry->entry))
|
||||||
|
+ __add_wait_queue(wq_head, wq_entry);
|
||||||
|
+ set_current_state(state);
|
||||||
|
+ spin_unlock_irqrestore(&wq_head->lock, flags);
|
||||||
|
+}
|
||||||
|
+EXPORT_SYMBOL(prepare_to_wait_exclusive_lifo);
|
||||||
|
+
|
||||||
|
void init_wait_entry(struct wait_queue_entry *wq_entry, int flags)
|
||||||
|
{
|
||||||
|
wq_entry->flags = flags;
|
||||||
|
diff --git a/net/ipv4/inet_connection_sock.c b/net/ipv4/inet_connection_sock.c
|
||||||
|
index f2c43f67187d..9885bfb429a2 100644
|
||||||
|
--- a/net/ipv4/inet_connection_sock.c
|
||||||
|
+++ b/net/ipv4/inet_connection_sock.c
|
||||||
|
@@ -606,7 +606,7 @@ static int inet_csk_wait_for_connect(struct sock *sk, long timeo)
|
||||||
|
* having to remove and re-insert us on the wait queue.
|
||||||
|
*/
|
||||||
|
for (;;) {
|
||||||
|
- prepare_to_wait_exclusive(sk_sleep(sk), &wait,
|
||||||
|
+ prepare_to_wait_exclusive_lifo(sk_sleep(sk), &wait,
|
||||||
|
TASK_INTERRUPTIBLE);
|
||||||
|
release_sock(sk);
|
||||||
|
if (reqsk_queue_empty(&icsk->icsk_accept_queue))
|
||||||
|
--
|
||||||
|
2.39.1
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
From afa213811c5490906caf394b20bb4b616fc6f12a Mon Sep 17 00:00:00 2001
|
||||||
|
From: Ricardo Neri <ricardo.neri-calderon@linux.intel.com>
|
||||||
|
Date: Thu, 25 Aug 2022 15:55:26 -0700
|
||||||
|
Subject: [PATCH 3/9] sched/fair: Simplify asym_packing logic for SMT sched
|
||||||
|
groups
|
||||||
|
|
||||||
|
When the destination CPU is an SMT sibling and idle, it can only help the
|
||||||
|
busiest group if all of its other SMT siblings are also idle. Otherwise,
|
||||||
|
there is not increase in throughput.
|
||||||
|
|
||||||
|
It does not matter whether the busiest group has SMT siblings. Simply
|
||||||
|
check if there are any tasks running on the local group before proceeding.
|
||||||
|
|
||||||
|
Cc: Ben Segall <bsegall@google.com>
|
||||||
|
Cc: Daniel Bristot de Oliveira <bristot@redhat.com>
|
||||||
|
Cc: Dietmar Eggemann <dietmar.eggemann@arm.com>
|
||||||
|
Cc: Len Brown <len.brown@intel.com>
|
||||||
|
Cc: Mel Gorman <mgorman@suse.de>
|
||||||
|
Cc: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
|
||||||
|
Cc: Srinivas Pandruvada <srinivas.pandruvada@linux.intel.com>
|
||||||
|
Cc: Steven Rostedt <rostedt@goodmis.org>
|
||||||
|
Cc: Tim C. Chen <tim.c.chen@intel.com>
|
||||||
|
Cc: Valentin Schneider <vschneid@redhat.com>
|
||||||
|
Cc: x86@kernel.org
|
||||||
|
Cc: linux-kernel@vger.kernel.org
|
||||||
|
Reviewed-by: Len Brown <len.brown@intel.com>
|
||||||
|
Signed-off-by: Ricardo Neri <ricardo.neri-calderon@linux.intel.com>
|
||||||
|
Signed-off-by: Alexandre Frade <kernel@xanmod.org>
|
||||||
|
---
|
||||||
|
kernel/sched/fair.c | 29 +++++++++--------------------
|
||||||
|
1 file changed, 9 insertions(+), 20 deletions(-)
|
||||||
|
|
||||||
|
diff --git a/kernel/sched/fair.c b/kernel/sched/fair.c
|
||||||
|
index 345cc5e9fa6e..60f9690a5626 100644
|
||||||
|
--- a/kernel/sched/fair.c
|
||||||
|
+++ b/kernel/sched/fair.c
|
||||||
|
@@ -8921,12 +8921,10 @@ static bool asym_smt_can_pull_tasks(int dst_cpu, struct sd_lb_stats *sds,
|
||||||
|
struct sched_group *sg)
|
||||||
|
{
|
||||||
|
#ifdef CONFIG_SCHED_SMT
|
||||||
|
- bool local_is_smt, sg_is_smt;
|
||||||
|
+ bool local_is_smt;
|
||||||
|
int sg_busy_cpus;
|
||||||
|
|
||||||
|
local_is_smt = sds->local->flags & SD_SHARE_CPUCAPACITY;
|
||||||
|
- sg_is_smt = sg->flags & SD_SHARE_CPUCAPACITY;
|
||||||
|
-
|
||||||
|
sg_busy_cpus = sgs->group_weight - sgs->idle_cpus;
|
||||||
|
|
||||||
|
if (!local_is_smt) {
|
||||||
|
@@ -8947,25 +8945,16 @@ static bool asym_smt_can_pull_tasks(int dst_cpu, struct sd_lb_stats *sds,
|
||||||
|
return sched_asym_prefer(dst_cpu, sg->asym_prefer_cpu);
|
||||||
|
}
|
||||||
|
|
||||||
|
- /* @dst_cpu has SMT siblings. */
|
||||||
|
-
|
||||||
|
- if (sg_is_smt) {
|
||||||
|
- int local_busy_cpus = sds->local->group_weight -
|
||||||
|
- sds->local_stat.idle_cpus;
|
||||||
|
- int busy_cpus_delta = sg_busy_cpus - local_busy_cpus;
|
||||||
|
-
|
||||||
|
- if (busy_cpus_delta == 1)
|
||||||
|
- return sched_asym_prefer(dst_cpu, sg->asym_prefer_cpu);
|
||||||
|
-
|
||||||
|
- return false;
|
||||||
|
- }
|
||||||
|
-
|
||||||
|
/*
|
||||||
|
- * @sg does not have SMT siblings. Ensure that @sds::local does not end
|
||||||
|
- * up with more than one busy SMT sibling and only pull tasks if there
|
||||||
|
- * are not busy CPUs (i.e., no CPU has running tasks).
|
||||||
|
+ * @dst_cpu has SMT siblings. When both @dst_cpu and the busiest core
|
||||||
|
+ * have one or more busy siblings, moving tasks between them results
|
||||||
|
+ * in the same throughput. Only if all the siblings of @dst_cpu are
|
||||||
|
+ * idle throughput can increase.
|
||||||
|
+ *
|
||||||
|
+ * If the difference in the number of busy CPUs is two or more, let
|
||||||
|
+ * find_busiest_group() take care of it.
|
||||||
|
*/
|
||||||
|
- if (!sds->local_stat.sum_nr_running)
|
||||||
|
+ if (sg_busy_cpus == 1 && !sds->local_stat.sum_nr_running)
|
||||||
|
return sched_asym_prefer(dst_cpu, sg->asym_prefer_cpu);
|
||||||
|
|
||||||
|
return false;
|
||||||
|
--
|
||||||
|
2.39.1
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
From a1f627fd10ced4f5eeae678bc4ba96ea7fa01d7e Mon Sep 17 00:00:00 2001
|
||||||
|
From: Ricardo Neri <ricardo.neri-calderon@linux.intel.com>
|
||||||
|
Date: Thu, 25 Aug 2022 15:55:28 -0700
|
||||||
|
Subject: [PATCH 4/9] sched/fair: Let lower-priority CPUs do active balancing
|
||||||
|
|
||||||
|
When more than one SMT siblings of a physical core are busy, an idle CPU
|
||||||
|
of lower priority can help.
|
||||||
|
|
||||||
|
Indicate that the low priority CPU can do active balancing from the high-
|
||||||
|
priority CPU only if they belong to separate cores.
|
||||||
|
|
||||||
|
Cc: Ben Segall <bsegall@google.com>
|
||||||
|
Cc: Daniel Bristot de Oliveira <bristot@redhat.com>
|
||||||
|
Cc: Dietmar Eggemann <dietmar.eggemann@arm.com>
|
||||||
|
Cc: Len Brown <len.brown@intel.com>
|
||||||
|
Cc: Mel Gorman <mgorman@suse.de>
|
||||||
|
Cc: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
|
||||||
|
Cc: Srinivas Pandruvada <srinivas.pandruvada@linux.intel.com>
|
||||||
|
Cc: Steven Rostedt <rostedt@goodmis.org>
|
||||||
|
Cc: Tim C. Chen <tim.c.chen@intel.com>
|
||||||
|
Cc: Valentin Schneider <vschneid@redhat.com>
|
||||||
|
Cc: x86@kernel.org
|
||||||
|
Cc: linux-kernel@vger.kernel.org
|
||||||
|
Reviewed-by: Len Brown <len.brown@intel.com>
|
||||||
|
Signed-off-by: Ricardo Neri <ricardo.neri-calderon@linux.intel.com>
|
||||||
|
Signed-off-by: Alexandre Frade <kernel@xanmod.org>
|
||||||
|
---
|
||||||
|
kernel/sched/fair.c | 7 ++++++-
|
||||||
|
1 file changed, 6 insertions(+), 1 deletion(-)
|
||||||
|
|
||||||
|
diff --git a/kernel/sched/fair.c b/kernel/sched/fair.c
|
||||||
|
index 60f9690a5626..67b0eacad0e9 100644
|
||||||
|
--- a/kernel/sched/fair.c
|
||||||
|
+++ b/kernel/sched/fair.c
|
||||||
|
@@ -10176,9 +10176,14 @@ asym_active_balance(struct lb_env *env)
|
||||||
|
* ASYM_PACKING needs to force migrate tasks from busy but
|
||||||
|
* lower priority CPUs in order to pack all tasks in the
|
||||||
|
* highest priority CPUs.
|
||||||
|
+ *
|
||||||
|
+ * If the busy CPU has higher priority but is an SMT sibling
|
||||||
|
+ * in which other SMT siblings are also busy, a lower-priority
|
||||||
|
+ * CPU in a separate core can help.
|
||||||
|
*/
|
||||||
|
return env->idle != CPU_NOT_IDLE && (env->sd->flags & SD_ASYM_PACKING) &&
|
||||||
|
- sched_asym_prefer(env->dst_cpu, env->src_cpu);
|
||||||
|
+ (sched_asym_prefer(env->dst_cpu, env->src_cpu) ||
|
||||||
|
+ !(env->sd->flags & SD_SHARE_CPUCAPACITY));
|
||||||
|
}
|
||||||
|
|
||||||
|
static inline bool
|
||||||
|
--
|
||||||
|
2.39.1
|
||||||
|
@@ -820,3 +820,31 @@ index a0b0397e29ee4c..87a983a356530c 100644
|
|||||||
spin_unlock(&zone->lock);
|
spin_unlock(&zone->lock);
|
||||||
return allocated;
|
return allocated;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
From 6329525a0fa10cd13f39b76948b1296150f75c95 Mon Sep 17 00:00:00 2001
|
||||||
|
From: Alexandre Frade <kernel@xanmod.org>
|
||||||
|
Date: Mon, 29 Aug 2022 16:47:26 +0000
|
||||||
|
Subject: [PATCH 14/16] XANMOD: Makefile: Disable GCC vectorization on trees
|
||||||
|
|
||||||
|
Signed-off-by: Alexandre Frade <kernel@xanmod.org>
|
||||||
|
---
|
||||||
|
Makefile | 3 +++
|
||||||
|
1 file changed, 3 insertions(+)
|
||||||
|
|
||||||
|
diff --git a/Makefile b/Makefile
|
||||||
|
index 3f6628780eb2..35a5ae1ede42 100644
|
||||||
|
--- a/Makefile
|
||||||
|
+++ b/Makefile
|
||||||
|
@@ -1069,6 +1069,9 @@ endif
|
||||||
|
KBUILD_CFLAGS-$(call gcc-min-version, 90100) += -Wno-alloc-size-larger-than
|
||||||
|
KBUILD_CFLAGS += $(KBUILD_CFLAGS-y) $(CONFIG_CC_IMPLICIT_FALLTHROUGH)
|
||||||
|
|
||||||
|
+# disable GCC vectorization on trees
|
||||||
|
+KBUILD_CFLAGS += $(call cc-option, -fno-tree-vectorize)
|
||||||
|
+
|
||||||
|
# disable invalid "can't wrap" optimizations for signed / pointers
|
||||||
|
KBUILD_CFLAGS += -fno-strict-overflow
|
||||||
|
|
||||||
|
--
|
||||||
|
2.39.1
|
||||||
|
|
||||||
|
@@ -20,3 +20,265 @@ index 6b3b59cc51d6..2a0072192c3d 100644
|
|||||||
|
|
||||||
int sched_thermal_decay_shift;
|
int sched_thermal_decay_shift;
|
||||||
static int __init setup_sched_thermal_decay_shift(char *str)
|
static int __init setup_sched_thermal_decay_shift(char *str)
|
||||||
|
|
||||||
|
From 5d5b708e3731e135ea7ae168571ad78d883e63e8 Mon Sep 17 00:00:00 2001
|
||||||
|
From: Alexandre Frade <kernel@xanmod.org>
|
||||||
|
Date: Wed, 1 Feb 2023 10:17:47 +0000
|
||||||
|
Subject: [PATCH 02/16] XANMOD: fair: Remove all energy efficiency functions
|
||||||
|
|
||||||
|
Signed-off-by: Alexandre Frade <kernel@xanmod.org>
|
||||||
|
---
|
||||||
|
kernel/sched/fair.c | 224 +-------------------------------------------
|
||||||
|
1 file changed, 3 insertions(+), 221 deletions(-)
|
||||||
|
|
||||||
|
diff --git a/kernel/sched/fair.c b/kernel/sched/fair.c
|
||||||
|
index 0f8736991427..345cc5e9fa6e 100644
|
||||||
|
--- a/kernel/sched/fair.c
|
||||||
|
+++ b/kernel/sched/fair.c
|
||||||
|
@@ -19,6 +19,9 @@
|
||||||
|
*
|
||||||
|
* Adaptive scheduling granularity, math enhancements by Peter Zijlstra
|
||||||
|
* Copyright (C) 2007 Red Hat, Inc., Peter Zijlstra
|
||||||
|
+ *
|
||||||
|
+ * Remove energy efficiency functions by Alexandre Frade
|
||||||
|
+ * (C) 2021 Alexandre Frade <kernel@xanmod.org>
|
||||||
|
*/
|
||||||
|
#include <linux/energy_model.h>
|
||||||
|
#include <linux/mmap_lock.h>
|
||||||
|
@@ -7136,219 +7139,6 @@ eenv_pd_max_util(struct energy_env *eenv, struct cpumask *pd_cpus,
|
||||||
|
return min(max_util, eenv->cpu_cap);
|
||||||
|
}
|
||||||
|
|
||||||
|
-/*
|
||||||
|
- * compute_energy(): Use the Energy Model to estimate the energy that @pd would
|
||||||
|
- * consume for a given utilization landscape @eenv. When @dst_cpu < 0, the task
|
||||||
|
- * contribution is ignored.
|
||||||
|
- */
|
||||||
|
-static inline unsigned long
|
||||||
|
-compute_energy(struct energy_env *eenv, struct perf_domain *pd,
|
||||||
|
- struct cpumask *pd_cpus, struct task_struct *p, int dst_cpu)
|
||||||
|
-{
|
||||||
|
- unsigned long max_util = eenv_pd_max_util(eenv, pd_cpus, p, dst_cpu);
|
||||||
|
- unsigned long busy_time = eenv->pd_busy_time;
|
||||||
|
-
|
||||||
|
- if (dst_cpu >= 0)
|
||||||
|
- busy_time = min(eenv->pd_cap, busy_time + eenv->task_busy_time);
|
||||||
|
-
|
||||||
|
- return em_cpu_energy(pd->em_pd, max_util, busy_time, eenv->cpu_cap);
|
||||||
|
-}
|
||||||
|
-
|
||||||
|
-/*
|
||||||
|
- * find_energy_efficient_cpu(): Find most energy-efficient target CPU for the
|
||||||
|
- * waking task. find_energy_efficient_cpu() looks for the CPU with maximum
|
||||||
|
- * spare capacity in each performance domain and uses it as a potential
|
||||||
|
- * candidate to execute the task. Then, it uses the Energy Model to figure
|
||||||
|
- * out which of the CPU candidates is the most energy-efficient.
|
||||||
|
- *
|
||||||
|
- * The rationale for this heuristic is as follows. In a performance domain,
|
||||||
|
- * all the most energy efficient CPU candidates (according to the Energy
|
||||||
|
- * Model) are those for which we'll request a low frequency. When there are
|
||||||
|
- * several CPUs for which the frequency request will be the same, we don't
|
||||||
|
- * have enough data to break the tie between them, because the Energy Model
|
||||||
|
- * only includes active power costs. With this model, if we assume that
|
||||||
|
- * frequency requests follow utilization (e.g. using schedutil), the CPU with
|
||||||
|
- * the maximum spare capacity in a performance domain is guaranteed to be among
|
||||||
|
- * the best candidates of the performance domain.
|
||||||
|
- *
|
||||||
|
- * In practice, it could be preferable from an energy standpoint to pack
|
||||||
|
- * small tasks on a CPU in order to let other CPUs go in deeper idle states,
|
||||||
|
- * but that could also hurt our chances to go cluster idle, and we have no
|
||||||
|
- * ways to tell with the current Energy Model if this is actually a good
|
||||||
|
- * idea or not. So, find_energy_efficient_cpu() basically favors
|
||||||
|
- * cluster-packing, and spreading inside a cluster. That should at least be
|
||||||
|
- * a good thing for latency, and this is consistent with the idea that most
|
||||||
|
- * of the energy savings of EAS come from the asymmetry of the system, and
|
||||||
|
- * not so much from breaking the tie between identical CPUs. That's also the
|
||||||
|
- * reason why EAS is enabled in the topology code only for systems where
|
||||||
|
- * SD_ASYM_CPUCAPACITY is set.
|
||||||
|
- *
|
||||||
|
- * NOTE: Forkees are not accepted in the energy-aware wake-up path because
|
||||||
|
- * they don't have any useful utilization data yet and it's not possible to
|
||||||
|
- * forecast their impact on energy consumption. Consequently, they will be
|
||||||
|
- * placed by find_idlest_cpu() on the least loaded CPU, which might turn out
|
||||||
|
- * to be energy-inefficient in some use-cases. The alternative would be to
|
||||||
|
- * bias new tasks towards specific types of CPUs first, or to try to infer
|
||||||
|
- * their util_avg from the parent task, but those heuristics could hurt
|
||||||
|
- * other use-cases too. So, until someone finds a better way to solve this,
|
||||||
|
- * let's keep things simple by re-using the existing slow path.
|
||||||
|
- */
|
||||||
|
-static int find_energy_efficient_cpu(struct task_struct *p, int prev_cpu)
|
||||||
|
-{
|
||||||
|
- struct cpumask *cpus = this_cpu_cpumask_var_ptr(select_rq_mask);
|
||||||
|
- unsigned long prev_delta = ULONG_MAX, best_delta = ULONG_MAX;
|
||||||
|
- unsigned long p_util_min = uclamp_is_used() ? uclamp_eff_value(p, UCLAMP_MIN) : 0;
|
||||||
|
- unsigned long p_util_max = uclamp_is_used() ? uclamp_eff_value(p, UCLAMP_MAX) : 1024;
|
||||||
|
- struct root_domain *rd = this_rq()->rd;
|
||||||
|
- int cpu, best_energy_cpu, target = -1;
|
||||||
|
- struct sched_domain *sd;
|
||||||
|
- struct perf_domain *pd;
|
||||||
|
- struct energy_env eenv;
|
||||||
|
-
|
||||||
|
- rcu_read_lock();
|
||||||
|
- pd = rcu_dereference(rd->pd);
|
||||||
|
- if (!pd || READ_ONCE(rd->overutilized))
|
||||||
|
- goto unlock;
|
||||||
|
-
|
||||||
|
- /*
|
||||||
|
- * Energy-aware wake-up happens on the lowest sched_domain starting
|
||||||
|
- * from sd_asym_cpucapacity spanning over this_cpu and prev_cpu.
|
||||||
|
- */
|
||||||
|
- sd = rcu_dereference(*this_cpu_ptr(&sd_asym_cpucapacity));
|
||||||
|
- while (sd && !cpumask_test_cpu(prev_cpu, sched_domain_span(sd)))
|
||||||
|
- sd = sd->parent;
|
||||||
|
- if (!sd)
|
||||||
|
- goto unlock;
|
||||||
|
-
|
||||||
|
- target = prev_cpu;
|
||||||
|
-
|
||||||
|
- sync_entity_load_avg(&p->se);
|
||||||
|
- if (!uclamp_task_util(p, p_util_min, p_util_max))
|
||||||
|
- goto unlock;
|
||||||
|
-
|
||||||
|
- eenv_task_busy_time(&eenv, p, prev_cpu);
|
||||||
|
-
|
||||||
|
- for (; pd; pd = pd->next) {
|
||||||
|
- unsigned long util_min = p_util_min, util_max = p_util_max;
|
||||||
|
- unsigned long cpu_cap, cpu_thermal_cap, util;
|
||||||
|
- unsigned long cur_delta, max_spare_cap = 0;
|
||||||
|
- unsigned long rq_util_min, rq_util_max;
|
||||||
|
- unsigned long prev_spare_cap = 0;
|
||||||
|
- int max_spare_cap_cpu = -1;
|
||||||
|
- unsigned long base_energy;
|
||||||
|
-
|
||||||
|
- cpumask_and(cpus, perf_domain_span(pd), cpu_online_mask);
|
||||||
|
-
|
||||||
|
- if (cpumask_empty(cpus))
|
||||||
|
- continue;
|
||||||
|
-
|
||||||
|
- /* Account thermal pressure for the energy estimation */
|
||||||
|
- cpu = cpumask_first(cpus);
|
||||||
|
- cpu_thermal_cap = arch_scale_cpu_capacity(cpu);
|
||||||
|
- cpu_thermal_cap -= arch_scale_thermal_pressure(cpu);
|
||||||
|
-
|
||||||
|
- eenv.cpu_cap = cpu_thermal_cap;
|
||||||
|
- eenv.pd_cap = 0;
|
||||||
|
-
|
||||||
|
- for_each_cpu(cpu, cpus) {
|
||||||
|
- struct rq *rq = cpu_rq(cpu);
|
||||||
|
-
|
||||||
|
- eenv.pd_cap += cpu_thermal_cap;
|
||||||
|
-
|
||||||
|
- if (!cpumask_test_cpu(cpu, sched_domain_span(sd)))
|
||||||
|
- continue;
|
||||||
|
-
|
||||||
|
- if (!cpumask_test_cpu(cpu, p->cpus_ptr))
|
||||||
|
- continue;
|
||||||
|
-
|
||||||
|
- util = cpu_util_next(cpu, p, cpu);
|
||||||
|
- cpu_cap = capacity_of(cpu);
|
||||||
|
-
|
||||||
|
- /*
|
||||||
|
- * Skip CPUs that cannot satisfy the capacity request.
|
||||||
|
- * IOW, placing the task there would make the CPU
|
||||||
|
- * overutilized. Take uclamp into account to see how
|
||||||
|
- * much capacity we can get out of the CPU; this is
|
||||||
|
- * aligned with sched_cpu_util().
|
||||||
|
- */
|
||||||
|
- if (uclamp_is_used() && !uclamp_rq_is_idle(rq)) {
|
||||||
|
- /*
|
||||||
|
- * Open code uclamp_rq_util_with() except for
|
||||||
|
- * the clamp() part. Ie: apply max aggregation
|
||||||
|
- * only. util_fits_cpu() logic requires to
|
||||||
|
- * operate on non clamped util but must use the
|
||||||
|
- * max-aggregated uclamp_{min, max}.
|
||||||
|
- */
|
||||||
|
- rq_util_min = uclamp_rq_get(rq, UCLAMP_MIN);
|
||||||
|
- rq_util_max = uclamp_rq_get(rq, UCLAMP_MAX);
|
||||||
|
-
|
||||||
|
- util_min = max(rq_util_min, p_util_min);
|
||||||
|
- util_max = max(rq_util_max, p_util_max);
|
||||||
|
- }
|
||||||
|
- if (!util_fits_cpu(util, util_min, util_max, cpu))
|
||||||
|
- continue;
|
||||||
|
-
|
||||||
|
- lsub_positive(&cpu_cap, util);
|
||||||
|
-
|
||||||
|
- if (cpu == prev_cpu) {
|
||||||
|
- /* Always use prev_cpu as a candidate. */
|
||||||
|
- prev_spare_cap = cpu_cap;
|
||||||
|
- } else if (cpu_cap > max_spare_cap) {
|
||||||
|
- /*
|
||||||
|
- * Find the CPU with the maximum spare capacity
|
||||||
|
- * among the remaining CPUs in the performance
|
||||||
|
- * domain.
|
||||||
|
- */
|
||||||
|
- max_spare_cap = cpu_cap;
|
||||||
|
- max_spare_cap_cpu = cpu;
|
||||||
|
- }
|
||||||
|
- }
|
||||||
|
-
|
||||||
|
- if (max_spare_cap_cpu < 0 && prev_spare_cap == 0)
|
||||||
|
- continue;
|
||||||
|
-
|
||||||
|
- eenv_pd_busy_time(&eenv, cpus, p);
|
||||||
|
- /* Compute the 'base' energy of the pd, without @p */
|
||||||
|
- base_energy = compute_energy(&eenv, pd, cpus, p, -1);
|
||||||
|
-
|
||||||
|
- /* Evaluate the energy impact of using prev_cpu. */
|
||||||
|
- if (prev_spare_cap > 0) {
|
||||||
|
- prev_delta = compute_energy(&eenv, pd, cpus, p,
|
||||||
|
- prev_cpu);
|
||||||
|
- /* CPU utilization has changed */
|
||||||
|
- if (prev_delta < base_energy)
|
||||||
|
- goto unlock;
|
||||||
|
- prev_delta -= base_energy;
|
||||||
|
- best_delta = min(best_delta, prev_delta);
|
||||||
|
- }
|
||||||
|
-
|
||||||
|
- /* Evaluate the energy impact of using max_spare_cap_cpu. */
|
||||||
|
- if (max_spare_cap_cpu >= 0 && max_spare_cap > prev_spare_cap) {
|
||||||
|
- cur_delta = compute_energy(&eenv, pd, cpus, p,
|
||||||
|
- max_spare_cap_cpu);
|
||||||
|
- /* CPU utilization has changed */
|
||||||
|
- if (cur_delta < base_energy)
|
||||||
|
- goto unlock;
|
||||||
|
- cur_delta -= base_energy;
|
||||||
|
- if (cur_delta < best_delta) {
|
||||||
|
- best_delta = cur_delta;
|
||||||
|
- best_energy_cpu = max_spare_cap_cpu;
|
||||||
|
- }
|
||||||
|
- }
|
||||||
|
- }
|
||||||
|
- rcu_read_unlock();
|
||||||
|
-
|
||||||
|
- if (best_delta < prev_delta)
|
||||||
|
- target = best_energy_cpu;
|
||||||
|
-
|
||||||
|
- return target;
|
||||||
|
-
|
||||||
|
-unlock:
|
||||||
|
- rcu_read_unlock();
|
||||||
|
-
|
||||||
|
- return target;
|
||||||
|
-}
|
||||||
|
-
|
||||||
|
/*
|
||||||
|
* select_task_rq_fair: Select target runqueue for the waking task in domains
|
||||||
|
* that have the relevant SD flag set. In practice, this is SD_BALANCE_WAKE,
|
||||||
|
@@ -7376,14 +7166,6 @@ select_task_rq_fair(struct task_struct *p, int prev_cpu, int wake_flags)
|
||||||
|
lockdep_assert_held(&p->pi_lock);
|
||||||
|
if (wake_flags & WF_TTWU) {
|
||||||
|
record_wakee(p);
|
||||||
|
-
|
||||||
|
- if (sched_energy_enabled()) {
|
||||||
|
- new_cpu = find_energy_efficient_cpu(p, prev_cpu);
|
||||||
|
- if (new_cpu >= 0)
|
||||||
|
- return new_cpu;
|
||||||
|
- new_cpu = prev_cpu;
|
||||||
|
- }
|
||||||
|
-
|
||||||
|
want_affine = !wake_wide(p) && cpumask_test_cpu(cpu, p->cpus_ptr);
|
||||||
|
}
|
||||||
|
|
||||||
|
--
|
||||||
|
2.39.1
|
||||||
|
@@ -88,3 +88,31 @@ index 6b423eebfd5d..61e3271675d6 100644
|
|||||||
#define MICRO_FREQUENCY_MIN_SAMPLE_RATE (10000)
|
#define MICRO_FREQUENCY_MIN_SAMPLE_RATE (10000)
|
||||||
#define MIN_FREQUENCY_UP_THRESHOLD (1)
|
#define MIN_FREQUENCY_UP_THRESHOLD (1)
|
||||||
#define MAX_FREQUENCY_UP_THRESHOLD (100)
|
#define MAX_FREQUENCY_UP_THRESHOLD (100)
|
||||||
|
|
||||||
|
From cba31b19f8c38696b13ba48e0e8b6dbe747d6bae Mon Sep 17 00:00:00 2001
|
||||||
|
From: Alexandre Frade <admfrade@gmail.com>
|
||||||
|
Date: Mon, 29 Jan 2018 17:31:25 +0000
|
||||||
|
Subject: [PATCH 10/16] XANMOD: mm/vmscan: vm_swappiness = 30 decreases the
|
||||||
|
amount of swapping
|
||||||
|
|
||||||
|
Signed-off-by: Alexandre Frade <admfrade@gmail.com>
|
||||||
|
Signed-off-by: Alexandre Frade <kernel@xanmod.org>
|
||||||
|
---
|
||||||
|
mm/vmscan.c | 2 +-
|
||||||
|
1 file changed, 1 insertion(+), 1 deletion(-)
|
||||||
|
|
||||||
|
diff --git a/mm/vmscan.c b/mm/vmscan.c
|
||||||
|
index 5b7b8d4f5297..549684b29418 100644
|
||||||
|
--- a/mm/vmscan.c
|
||||||
|
+++ b/mm/vmscan.c
|
||||||
|
@@ -190,7 +190,7 @@ struct scan_control {
|
||||||
|
/*
|
||||||
|
* From 0 .. 200. Higher means more swappy.
|
||||||
|
*/
|
||||||
|
-int vm_swappiness = 60;
|
||||||
|
+int vm_swappiness = 30;
|
||||||
|
|
||||||
|
static void set_task_reclaim_state(struct task_struct *task,
|
||||||
|
struct reclaim_state *rs)
|
||||||
|
--
|
||||||
|
2.39.1
|
||||||
|
@@ -78,7 +78,7 @@ index 4700d24e5d55..8f7a3d7fd9c1 100644
|
|||||||
@@ -3372,6 +3372,106 @@ static void quirk_no_bus_reset(struct pci_dev *dev)
|
@@ -3372,6 +3372,106 @@ static void quirk_no_bus_reset(struct pci_dev *dev)
|
||||||
dev->dev_flags |= PCI_DEV_FLAGS_NO_BUS_RESET;
|
dev->dev_flags |= PCI_DEV_FLAGS_NO_BUS_RESET;
|
||||||
}
|
}
|
||||||
|
|
||||||
+static bool acs_on_downstream;
|
+static bool acs_on_downstream;
|
||||||
+static bool acs_on_multifunction;
|
+static bool acs_on_multifunction;
|
||||||
+
|
+
|
||||||
@@ -180,14 +180,14 @@ index 4700d24e5d55..8f7a3d7fd9c1 100644
|
|||||||
+ return -ENOTTY;
|
+ return -ENOTTY;
|
||||||
+}
|
+}
|
||||||
/*
|
/*
|
||||||
* Some Atheros AR9xxx and QCA988x chips do not behave after a bus reset.
|
* Some NVIDIA GPU devices do not work with bus reset, SBR needs to be
|
||||||
* The device will throw a Link Down error on AER-capable systems and
|
* prevented for those affected devices.
|
||||||
@@ -4513,6 +4613,7 @@ static const struct pci_dev_acs_enabled {
|
@@ -5002,6 +5102,8 @@ static const struct pci_dev_acs_enabled {
|
||||||
{ PCI_VENDOR_ID_ZHAOXIN, 0x9083, pci_quirk_mf_endpoint_acs },
|
|
||||||
/* Zhaoxin Root/Downstream Ports */
|
|
||||||
{ PCI_VENDOR_ID_ZHAOXIN, PCI_ANY_ID, pci_quirk_zhaoxin_pcie_ports_acs },
|
{ PCI_VENDOR_ID_ZHAOXIN, PCI_ANY_ID, pci_quirk_zhaoxin_pcie_ports_acs },
|
||||||
+ { PCI_ANY_ID, PCI_ANY_ID, pcie_acs_overrides },
|
/* Wangxun nics */
|
||||||
|
{ PCI_VENDOR_ID_WANGXUN, PCI_ANY_ID, pci_quirk_wangxun_nic_acs },
|
||||||
|
+ /* ACS override */
|
||||||
|
+ { PCI_ANY_ID, PCI_ANY_ID, pcie_acs_overrides },
|
||||||
{ 0 }
|
{ 0 }
|
||||||
};
|
};
|
||||||
|
|
||||||
|
|
||||||
|
@@ -1,8 +1,62 @@
|
|||||||
|
From e44ef62b127f6a161a131c84db92a7527d8fc72d Mon Sep 17 00:00:00 2001
|
||||||
|
From: Peter Jung <admin@ptr1337.dev>
|
||||||
|
Date: Wed, 22 Feb 2023 19:24:36 +0100
|
||||||
|
Subject: [PATCH] prjc
|
||||||
|
|
||||||
|
Signed-off-by: Peter Jung <admin@ptr1337.dev>
|
||||||
|
---
|
||||||
|
.../admin-guide/kernel-parameters.txt | 6 +
|
||||||
|
Documentation/admin-guide/sysctl/kernel.rst | 10 +
|
||||||
|
Documentation/scheduler/sched-BMQ.txt | 110 +
|
||||||
|
fs/proc/base.c | 2 +-
|
||||||
|
include/asm-generic/resource.h | 2 +-
|
||||||
|
include/linux/sched.h | 33 +-
|
||||||
|
include/linux/sched/deadline.h | 20 +
|
||||||
|
include/linux/sched/prio.h | 26 +
|
||||||
|
include/linux/sched/rt.h | 2 +
|
||||||
|
include/linux/sched/topology.h | 3 +-
|
||||||
|
init/Kconfig | 34 +
|
||||||
|
init/init_task.c | 18 +
|
||||||
|
kernel/Kconfig.preempt | 2 +-
|
||||||
|
kernel/cgroup/cpuset.c | 4 +-
|
||||||
|
kernel/delayacct.c | 2 +-
|
||||||
|
kernel/exit.c | 4 +-
|
||||||
|
kernel/locking/rtmutex.c | 16 +-
|
||||||
|
kernel/sched/Makefile | 5 +
|
||||||
|
kernel/sched/alt_core.c | 8111 +++++++++++++++++
|
||||||
|
kernel/sched/alt_debug.c | 31 +
|
||||||
|
kernel/sched/alt_sched.h | 671 ++
|
||||||
|
kernel/sched/bmq.h | 110 +
|
||||||
|
kernel/sched/build_policy.c | 8 +-
|
||||||
|
kernel/sched/build_utility.c | 2 +
|
||||||
|
kernel/sched/cpufreq_schedutil.c | 10 +
|
||||||
|
kernel/sched/cputime.c | 10 +-
|
||||||
|
kernel/sched/debug.c | 10 +
|
||||||
|
kernel/sched/idle.c | 2 +
|
||||||
|
kernel/sched/pds.h | 127 +
|
||||||
|
kernel/sched/pelt.c | 4 +-
|
||||||
|
kernel/sched/pelt.h | 8 +-
|
||||||
|
kernel/sched/sched.h | 9 +
|
||||||
|
kernel/sched/stats.c | 4 +
|
||||||
|
kernel/sched/stats.h | 2 +
|
||||||
|
kernel/sched/topology.c | 17 +
|
||||||
|
kernel/sysctl.c | 15 +
|
||||||
|
kernel/time/hrtimer.c | 2 +
|
||||||
|
kernel/time/posix-cpu-timers.c | 10 +-
|
||||||
|
kernel/trace/trace_selftest.c | 5 +
|
||||||
|
39 files changed, 9445 insertions(+), 22 deletions(-)
|
||||||
|
create mode 100644 Documentation/scheduler/sched-BMQ.txt
|
||||||
|
create mode 100644 kernel/sched/alt_core.c
|
||||||
|
create mode 100644 kernel/sched/alt_debug.c
|
||||||
|
create mode 100644 kernel/sched/alt_sched.h
|
||||||
|
create mode 100644 kernel/sched/bmq.h
|
||||||
|
create mode 100644 kernel/sched/pds.h
|
||||||
|
|
||||||
diff --git a/Documentation/admin-guide/kernel-parameters.txt b/Documentation/admin-guide/kernel-parameters.txt
|
diff --git a/Documentation/admin-guide/kernel-parameters.txt b/Documentation/admin-guide/kernel-parameters.txt
|
||||||
index 42af9ca0127e..31747ec54f9d 100644
|
index 6cfa6e3996cf..1b6a407213da 100644
|
||||||
--- a/Documentation/admin-guide/kernel-parameters.txt
|
--- a/Documentation/admin-guide/kernel-parameters.txt
|
||||||
+++ b/Documentation/admin-guide/kernel-parameters.txt
|
+++ b/Documentation/admin-guide/kernel-parameters.txt
|
||||||
@@ -5406,6 +5406,12 @@
|
@@ -5437,6 +5437,12 @@
|
||||||
sa1100ir [NET]
|
sa1100ir [NET]
|
||||||
See drivers/net/irda/sa1100_ir.c.
|
See drivers/net/irda/sa1100_ir.c.
|
||||||
|
|
||||||
@@ -16,10 +70,10 @@ index 42af9ca0127e..31747ec54f9d 100644
|
|||||||
|
|
||||||
schedstats= [KNL,X86] Enable or disable scheduled statistics.
|
schedstats= [KNL,X86] Enable or disable scheduled statistics.
|
||||||
diff --git a/Documentation/admin-guide/sysctl/kernel.rst b/Documentation/admin-guide/sysctl/kernel.rst
|
diff --git a/Documentation/admin-guide/sysctl/kernel.rst b/Documentation/admin-guide/sysctl/kernel.rst
|
||||||
index 98d1b198b2b4..d7c78a107f93 100644
|
index 46e3d62c0eea..fb4568c919d0 100644
|
||||||
--- a/Documentation/admin-guide/sysctl/kernel.rst
|
--- a/Documentation/admin-guide/sysctl/kernel.rst
|
||||||
+++ b/Documentation/admin-guide/sysctl/kernel.rst
|
+++ b/Documentation/admin-guide/sysctl/kernel.rst
|
||||||
@@ -1552,3 +1552,13 @@ is 10 seconds.
|
@@ -1597,3 +1597,13 @@ is 10 seconds.
|
||||||
|
|
||||||
The softlockup threshold is (``2 * watchdog_thresh``). Setting this
|
The softlockup threshold is (``2 * watchdog_thresh``). Setting this
|
||||||
tunable to zero will disable lockup detection altogether.
|
tunable to zero will disable lockup detection altogether.
|
||||||
@@ -176,7 +230,7 @@ index 8874f681b056..59eb72bf7d5f 100644
|
|||||||
[RLIMIT_RTTIME] = { RLIM_INFINITY, RLIM_INFINITY }, \
|
[RLIMIT_RTTIME] = { RLIM_INFINITY, RLIM_INFINITY }, \
|
||||||
}
|
}
|
||||||
diff --git a/include/linux/sched.h b/include/linux/sched.h
|
diff --git a/include/linux/sched.h b/include/linux/sched.h
|
||||||
index ffb6eb55cd13..2e730a59caa2 100644
|
index 853d08f7562b..ad7e050d7455 100644
|
||||||
--- a/include/linux/sched.h
|
--- a/include/linux/sched.h
|
||||||
+++ b/include/linux/sched.h
|
+++ b/include/linux/sched.h
|
||||||
@@ -762,8 +762,14 @@ struct task_struct {
|
@@ -762,8 +762,14 @@ struct task_struct {
|
||||||
@@ -232,7 +286,7 @@ index ffb6eb55cd13..2e730a59caa2 100644
|
|||||||
|
|
||||||
#ifdef CONFIG_CGROUP_SCHED
|
#ifdef CONFIG_CGROUP_SCHED
|
||||||
struct task_group *sched_task_group;
|
struct task_group *sched_task_group;
|
||||||
@@ -1545,6 +1567,15 @@ struct task_struct {
|
@@ -1539,6 +1561,15 @@ struct task_struct {
|
||||||
*/
|
*/
|
||||||
};
|
};
|
||||||
|
|
||||||
@@ -352,10 +406,10 @@ index 816df6cc444e..c8da08e18c91 100644
|
|||||||
#else
|
#else
|
||||||
static inline void rebuild_sched_domains_energy(void)
|
static inline void rebuild_sched_domains_energy(void)
|
||||||
diff --git a/init/Kconfig b/init/Kconfig
|
diff --git a/init/Kconfig b/init/Kconfig
|
||||||
index 94125d3b6893..c87ba766d354 100644
|
index 44e90b28a30f..af24591984ab 100644
|
||||||
--- a/init/Kconfig
|
--- a/init/Kconfig
|
||||||
+++ b/init/Kconfig
|
+++ b/init/Kconfig
|
||||||
@@ -819,6 +819,7 @@ menu "Scheduler features"
|
@@ -821,6 +821,7 @@ menu "Scheduler features"
|
||||||
config UCLAMP_TASK
|
config UCLAMP_TASK
|
||||||
bool "Enable utilization clamping for RT/FAIR tasks"
|
bool "Enable utilization clamping for RT/FAIR tasks"
|
||||||
depends on CPU_FREQ_GOV_SCHEDUTIL
|
depends on CPU_FREQ_GOV_SCHEDUTIL
|
||||||
@@ -363,7 +417,7 @@ index 94125d3b6893..c87ba766d354 100644
|
|||||||
help
|
help
|
||||||
This feature enables the scheduler to track the clamped utilization
|
This feature enables the scheduler to track the clamped utilization
|
||||||
of each CPU based on RUNNABLE tasks scheduled on that CPU.
|
of each CPU based on RUNNABLE tasks scheduled on that CPU.
|
||||||
@@ -865,6 +866,35 @@ config UCLAMP_BUCKETS_COUNT
|
@@ -867,6 +868,35 @@ config UCLAMP_BUCKETS_COUNT
|
||||||
|
|
||||||
If in doubt, use the default value.
|
If in doubt, use the default value.
|
||||||
|
|
||||||
@@ -399,7 +453,7 @@ index 94125d3b6893..c87ba766d354 100644
|
|||||||
endmenu
|
endmenu
|
||||||
|
|
||||||
#
|
#
|
||||||
@@ -918,6 +948,7 @@ config NUMA_BALANCING
|
@@ -924,6 +954,7 @@ config NUMA_BALANCING
|
||||||
depends on ARCH_SUPPORTS_NUMA_BALANCING
|
depends on ARCH_SUPPORTS_NUMA_BALANCING
|
||||||
depends on !ARCH_WANT_NUMA_VARIABLE_LOCALITY
|
depends on !ARCH_WANT_NUMA_VARIABLE_LOCALITY
|
||||||
depends on SMP && NUMA && MIGRATION && !PREEMPT_RT
|
depends on SMP && NUMA && MIGRATION && !PREEMPT_RT
|
||||||
@@ -407,7 +461,7 @@ index 94125d3b6893..c87ba766d354 100644
|
|||||||
help
|
help
|
||||||
This option adds support for automatic NUMA aware memory/task placement.
|
This option adds support for automatic NUMA aware memory/task placement.
|
||||||
The mechanism is quite primitive and is based on migrating memory when
|
The mechanism is quite primitive and is based on migrating memory when
|
||||||
@@ -1015,6 +1046,7 @@ config FAIR_GROUP_SCHED
|
@@ -1021,6 +1052,7 @@ config FAIR_GROUP_SCHED
|
||||||
depends on CGROUP_SCHED
|
depends on CGROUP_SCHED
|
||||||
default CGROUP_SCHED
|
default CGROUP_SCHED
|
||||||
|
|
||||||
@@ -415,7 +469,7 @@ index 94125d3b6893..c87ba766d354 100644
|
|||||||
config CFS_BANDWIDTH
|
config CFS_BANDWIDTH
|
||||||
bool "CPU bandwidth provisioning for FAIR_GROUP_SCHED"
|
bool "CPU bandwidth provisioning for FAIR_GROUP_SCHED"
|
||||||
depends on FAIR_GROUP_SCHED
|
depends on FAIR_GROUP_SCHED
|
||||||
@@ -1037,6 +1069,7 @@ config RT_GROUP_SCHED
|
@@ -1043,6 +1075,7 @@ config RT_GROUP_SCHED
|
||||||
realtime bandwidth for them.
|
realtime bandwidth for them.
|
||||||
See Documentation/scheduler/sched-rt-group.rst for more information.
|
See Documentation/scheduler/sched-rt-group.rst for more information.
|
||||||
|
|
||||||
@@ -423,7 +477,7 @@ index 94125d3b6893..c87ba766d354 100644
|
|||||||
endif #CGROUP_SCHED
|
endif #CGROUP_SCHED
|
||||||
|
|
||||||
config UCLAMP_TASK_GROUP
|
config UCLAMP_TASK_GROUP
|
||||||
@@ -1281,6 +1314,7 @@ config CHECKPOINT_RESTORE
|
@@ -1287,6 +1320,7 @@ config CHECKPOINT_RESTORE
|
||||||
|
|
||||||
config SCHED_AUTOGROUP
|
config SCHED_AUTOGROUP
|
||||||
bool "Automatic process group scheduling"
|
bool "Automatic process group scheduling"
|
||||||
@@ -491,10 +545,10 @@ index c2f1fd95a821..41654679b1b2 100644
|
|||||||
This option permits Core Scheduling, a means of coordinated task
|
This option permits Core Scheduling, a means of coordinated task
|
||||||
selection across SMT siblings. When enabled -- see
|
selection across SMT siblings. When enabled -- see
|
||||||
diff --git a/kernel/cgroup/cpuset.c b/kernel/cgroup/cpuset.c
|
diff --git a/kernel/cgroup/cpuset.c b/kernel/cgroup/cpuset.c
|
||||||
index b474289c15b8..a23224b45b03 100644
|
index ca826bd1eba3..60e194f1d6d8 100644
|
||||||
--- a/kernel/cgroup/cpuset.c
|
--- a/kernel/cgroup/cpuset.c
|
||||||
+++ b/kernel/cgroup/cpuset.c
|
+++ b/kernel/cgroup/cpuset.c
|
||||||
@@ -787,7 +787,7 @@ static int validate_change(struct cpuset *cur, struct cpuset *trial)
|
@@ -791,7 +791,7 @@ static int validate_change(struct cpuset *cur, struct cpuset *trial)
|
||||||
return ret;
|
return ret;
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -503,7 +557,7 @@ index b474289c15b8..a23224b45b03 100644
|
|||||||
/*
|
/*
|
||||||
* Helper routine for generate_sched_domains().
|
* Helper routine for generate_sched_domains().
|
||||||
* Do cpusets a, b have overlapping effective cpus_allowed masks?
|
* Do cpusets a, b have overlapping effective cpus_allowed masks?
|
||||||
@@ -1183,7 +1183,7 @@ static void rebuild_sched_domains_locked(void)
|
@@ -1187,7 +1187,7 @@ static void rebuild_sched_domains_locked(void)
|
||||||
/* Have scheduler rebuild the domains */
|
/* Have scheduler rebuild the domains */
|
||||||
partition_and_rebuild_sched_domains(ndoms, doms, attr);
|
partition_and_rebuild_sched_domains(ndoms, doms, attr);
|
||||||
}
|
}
|
||||||
@@ -526,10 +580,10 @@ index e39cb696cfbd..463423572e09 100644
|
|||||||
d->cpu_count += t1;
|
d->cpu_count += t1;
|
||||||
|
|
||||||
diff --git a/kernel/exit.c b/kernel/exit.c
|
diff --git a/kernel/exit.c b/kernel/exit.c
|
||||||
index 35e0a31a0315..64e368441cf4 100644
|
index 15dc2ec80c46..1e583e0f89a7 100644
|
||||||
--- a/kernel/exit.c
|
--- a/kernel/exit.c
|
||||||
+++ b/kernel/exit.c
|
+++ b/kernel/exit.c
|
||||||
@@ -125,7 +125,7 @@ static void __exit_signal(struct task_struct *tsk)
|
@@ -172,7 +172,7 @@ static void __exit_signal(struct task_struct *tsk)
|
||||||
sig->curr_target = next_thread(tsk);
|
sig->curr_target = next_thread(tsk);
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -538,7 +592,7 @@ index 35e0a31a0315..64e368441cf4 100644
|
|||||||
sizeof(unsigned long long));
|
sizeof(unsigned long long));
|
||||||
|
|
||||||
/*
|
/*
|
||||||
@@ -146,7 +146,7 @@ static void __exit_signal(struct task_struct *tsk)
|
@@ -193,7 +193,7 @@ static void __exit_signal(struct task_struct *tsk)
|
||||||
sig->inblock += task_io_get_inblock(tsk);
|
sig->inblock += task_io_get_inblock(tsk);
|
||||||
sig->oublock += task_io_get_oublock(tsk);
|
sig->oublock += task_io_get_oublock(tsk);
|
||||||
task_io_accounting_add(&sig->ioac, &tsk->ioac);
|
task_io_accounting_add(&sig->ioac, &tsk->ioac);
|
||||||
@@ -548,10 +602,10 @@ index 35e0a31a0315..64e368441cf4 100644
|
|||||||
__unhash_process(tsk, group_dead);
|
__unhash_process(tsk, group_dead);
|
||||||
write_sequnlock(&sig->stats_lock);
|
write_sequnlock(&sig->stats_lock);
|
||||||
diff --git a/kernel/locking/rtmutex.c b/kernel/locking/rtmutex.c
|
diff --git a/kernel/locking/rtmutex.c b/kernel/locking/rtmutex.c
|
||||||
index 7779ee8abc2a..5b9893cdfb1b 100644
|
index 728f434de2bb..0e1082a4e878 100644
|
||||||
--- a/kernel/locking/rtmutex.c
|
--- a/kernel/locking/rtmutex.c
|
||||||
+++ b/kernel/locking/rtmutex.c
|
+++ b/kernel/locking/rtmutex.c
|
||||||
@@ -300,21 +300,25 @@ static __always_inline void
|
@@ -337,21 +337,25 @@ static __always_inline void
|
||||||
waiter_update_prio(struct rt_mutex_waiter *waiter, struct task_struct *task)
|
waiter_update_prio(struct rt_mutex_waiter *waiter, struct task_struct *task)
|
||||||
{
|
{
|
||||||
waiter->prio = __waiter_prio(task);
|
waiter->prio = __waiter_prio(task);
|
||||||
@@ -579,7 +633,7 @@ index 7779ee8abc2a..5b9893cdfb1b 100644
|
|||||||
/*
|
/*
|
||||||
* If both waiters have dl_prio(), we check the deadlines of the
|
* If both waiters have dl_prio(), we check the deadlines of the
|
||||||
* associated tasks.
|
* associated tasks.
|
||||||
@@ -323,16 +327,22 @@ static __always_inline int rt_mutex_waiter_less(struct rt_mutex_waiter *left,
|
@@ -360,16 +364,22 @@ static __always_inline int rt_mutex_waiter_less(struct rt_mutex_waiter *left,
|
||||||
*/
|
*/
|
||||||
if (dl_prio(left->prio))
|
if (dl_prio(left->prio))
|
||||||
return dl_time_before(left->deadline, right->deadline);
|
return dl_time_before(left->deadline, right->deadline);
|
||||||
@@ -602,7 +656,7 @@ index 7779ee8abc2a..5b9893cdfb1b 100644
|
|||||||
/*
|
/*
|
||||||
* If both waiters have dl_prio(), we check the deadlines of the
|
* If both waiters have dl_prio(), we check the deadlines of the
|
||||||
* associated tasks.
|
* associated tasks.
|
||||||
@@ -341,8 +351,10 @@ static __always_inline int rt_mutex_waiter_equal(struct rt_mutex_waiter *left,
|
@@ -378,8 +388,10 @@ static __always_inline int rt_mutex_waiter_equal(struct rt_mutex_waiter *left,
|
||||||
*/
|
*/
|
||||||
if (dl_prio(left->prio))
|
if (dl_prio(left->prio))
|
||||||
return left->deadline == right->deadline;
|
return left->deadline == right->deadline;
|
||||||
@@ -632,10 +686,10 @@ index 976092b7bd45..31d587c16ec1 100644
|
|||||||
obj-y += build_utility.o
|
obj-y += build_utility.o
|
||||||
diff --git a/kernel/sched/alt_core.c b/kernel/sched/alt_core.c
|
diff --git a/kernel/sched/alt_core.c b/kernel/sched/alt_core.c
|
||||||
new file mode 100644
|
new file mode 100644
|
||||||
index 000000000000..acb8657e811d
|
index 000000000000..f5e9c01f9382
|
||||||
--- /dev/null
|
--- /dev/null
|
||||||
+++ b/kernel/sched/alt_core.c
|
+++ b/kernel/sched/alt_core.c
|
||||||
@@ -0,0 +1,7978 @@
|
@@ -0,0 +1,8111 @@
|
||||||
+/*
|
+/*
|
||||||
+ * kernel/sched/alt_core.c
|
+ * kernel/sched/alt_core.c
|
||||||
+ *
|
+ *
|
||||||
@@ -705,7 +759,7 @@ index 000000000000..acb8657e811d
|
|||||||
+#define sched_feat(x) (0)
|
+#define sched_feat(x) (0)
|
||||||
+#endif /* CONFIG_SCHED_DEBUG */
|
+#endif /* CONFIG_SCHED_DEBUG */
|
||||||
+
|
+
|
||||||
+#define ALT_SCHED_VERSION "v6.1-r3"
|
+#define ALT_SCHED_VERSION "v6.2-r0"
|
||||||
+
|
+
|
||||||
+/* rt_prio(prio) defined in include/linux/sched/rt.h */
|
+/* rt_prio(prio) defined in include/linux/sched/rt.h */
|
||||||
+#define rt_task(p) rt_prio((p)->prio)
|
+#define rt_task(p) rt_prio((p)->prio)
|
||||||
@@ -726,6 +780,12 @@ index 000000000000..acb8657e811d
|
|||||||
+#include "pds.h"
|
+#include "pds.h"
|
||||||
+#endif
|
+#endif
|
||||||
+
|
+
|
||||||
|
+struct affinity_context {
|
||||||
|
+ const struct cpumask *new_mask;
|
||||||
|
+ struct cpumask *user_mask;
|
||||||
|
+ unsigned int flags;
|
||||||
|
+};
|
||||||
|
+
|
||||||
+static int __init sched_timeslice(char *str)
|
+static int __init sched_timeslice(char *str)
|
||||||
+{
|
+{
|
||||||
+ int timeslice_ms;
|
+ int timeslice_ms;
|
||||||
@@ -788,6 +848,14 @@ index 000000000000..acb8657e811d
|
|||||||
+static cpumask_t sched_preempt_mask[SCHED_QUEUE_BITS] ____cacheline_aligned_in_smp;
|
+static cpumask_t sched_preempt_mask[SCHED_QUEUE_BITS] ____cacheline_aligned_in_smp;
|
||||||
+static cpumask_t *const sched_idle_mask = &sched_preempt_mask[0];
|
+static cpumask_t *const sched_idle_mask = &sched_preempt_mask[0];
|
||||||
+
|
+
|
||||||
|
+/* task function */
|
||||||
|
+static inline const struct cpumask *task_user_cpus(struct task_struct *p)
|
||||||
|
+{
|
||||||
|
+ if (!p->user_cpus_ptr)
|
||||||
|
+ return cpu_possible_mask; /* &init_task.cpus_mask */
|
||||||
|
+ return p->user_cpus_ptr;
|
||||||
|
+}
|
||||||
|
+
|
||||||
+/* sched_queue related functions */
|
+/* sched_queue related functions */
|
||||||
+static inline void sched_queue_init(struct sched_queue *q)
|
+static inline void sched_queue_init(struct sched_queue *q)
|
||||||
+{
|
+{
|
||||||
@@ -1400,7 +1468,7 @@ index 000000000000..acb8657e811d
|
|||||||
+
|
+
|
||||||
+#define __SCHED_ENQUEUE_TASK(p, rq, flags) \
|
+#define __SCHED_ENQUEUE_TASK(p, rq, flags) \
|
||||||
+ sched_info_enqueue(rq, p); \
|
+ sched_info_enqueue(rq, p); \
|
||||||
+ psi_enqueue(p, flags); \
|
+ psi_enqueue(p, flags & ENQUEUE_WAKEUP); \
|
||||||
+ \
|
+ \
|
||||||
+ p->sq_idx = task_sched_prio_idx(p, rq); \
|
+ p->sq_idx = task_sched_prio_idx(p, rq); \
|
||||||
+ list_add_tail(&p->sq_node, &rq->queue.heads[p->sq_idx]); \
|
+ list_add_tail(&p->sq_node, &rq->queue.heads[p->sq_idx]); \
|
||||||
@@ -2268,35 +2336,101 @@ index 000000000000..acb8657e811d
|
|||||||
+}
|
+}
|
||||||
+
|
+
|
||||||
+static inline void
|
+static inline void
|
||||||
+set_cpus_allowed_common(struct task_struct *p, const struct cpumask *new_mask)
|
+set_cpus_allowed_common(struct task_struct *p, struct affinity_context *ctx)
|
||||||
+{
|
+{
|
||||||
+ cpumask_copy(&p->cpus_mask, new_mask);
|
+ cpumask_copy(&p->cpus_mask, ctx->new_mask);
|
||||||
+ p->nr_cpus_allowed = cpumask_weight(new_mask);
|
+ p->nr_cpus_allowed = cpumask_weight(ctx->new_mask);
|
||||||
|
+
|
||||||
|
+ /*
|
||||||
|
+ * Swap in a new user_cpus_ptr if SCA_USER flag set
|
||||||
|
+ */
|
||||||
|
+ if (ctx->flags & SCA_USER)
|
||||||
|
+ swap(p->user_cpus_ptr, ctx->user_mask);
|
||||||
+}
|
+}
|
||||||
+
|
+
|
||||||
+static void
|
+static void
|
||||||
+__do_set_cpus_allowed(struct task_struct *p, const struct cpumask *new_mask)
|
+__do_set_cpus_allowed(struct task_struct *p, struct affinity_context *ctx)
|
||||||
+{
|
+{
|
||||||
+ lockdep_assert_held(&p->pi_lock);
|
+ lockdep_assert_held(&p->pi_lock);
|
||||||
+ set_cpus_allowed_common(p, new_mask);
|
+ set_cpus_allowed_common(p, ctx);
|
||||||
+}
|
+}
|
||||||
+
|
+
|
||||||
|
+/*
|
||||||
|
+ * Used for kthread_bind() and select_fallback_rq(), in both cases the user
|
||||||
|
+ * affinity (if any) should be destroyed too.
|
||||||
|
+ */
|
||||||
+void do_set_cpus_allowed(struct task_struct *p, const struct cpumask *new_mask)
|
+void do_set_cpus_allowed(struct task_struct *p, const struct cpumask *new_mask)
|
||||||
+{
|
+{
|
||||||
+ __do_set_cpus_allowed(p, new_mask);
|
+ struct affinity_context ac = {
|
||||||
|
+ .new_mask = new_mask,
|
||||||
|
+ .user_mask = NULL,
|
||||||
|
+ .flags = SCA_USER, /* clear the user requested mask */
|
||||||
|
+ };
|
||||||
|
+ union cpumask_rcuhead {
|
||||||
|
+ cpumask_t cpumask;
|
||||||
|
+ struct rcu_head rcu;
|
||||||
|
+ };
|
||||||
|
+
|
||||||
|
+ __do_set_cpus_allowed(p, &ac);
|
||||||
|
+
|
||||||
|
+ /*
|
||||||
|
+ * Because this is called with p->pi_lock held, it is not possible
|
||||||
|
+ * to use kfree() here (when PREEMPT_RT=y), therefore punt to using
|
||||||
|
+ * kfree_rcu().
|
||||||
|
+ */
|
||||||
|
+ kfree_rcu((union cpumask_rcuhead *)ac.user_mask, rcu);
|
||||||
|
+}
|
||||||
|
+
|
||||||
|
+static cpumask_t *alloc_user_cpus_ptr(int node)
|
||||||
|
+{
|
||||||
|
+ /*
|
||||||
|
+ * See do_set_cpus_allowed() above for the rcu_head usage.
|
||||||
|
+ */
|
||||||
|
+ int size = max_t(int, cpumask_size(), sizeof(struct rcu_head));
|
||||||
|
+
|
||||||
|
+ return kmalloc_node(size, GFP_KERNEL, node);
|
||||||
+}
|
+}
|
||||||
+
|
+
|
||||||
+int dup_user_cpus_ptr(struct task_struct *dst, struct task_struct *src,
|
+int dup_user_cpus_ptr(struct task_struct *dst, struct task_struct *src,
|
||||||
+ int node)
|
+ int node)
|
||||||
+{
|
+{
|
||||||
+ if (!src->user_cpus_ptr)
|
+ cpumask_t *user_mask;
|
||||||
|
+ unsigned long flags;
|
||||||
|
+
|
||||||
|
+ /*
|
||||||
|
+ * Always clear dst->user_cpus_ptr first as their user_cpus_ptr's
|
||||||
|
+ * may differ by now due to racing.
|
||||||
|
+ */
|
||||||
|
+ dst->user_cpus_ptr = NULL;
|
||||||
|
+
|
||||||
|
+ /*
|
||||||
|
+ * This check is racy and losing the race is a valid situation.
|
||||||
|
+ * It is not worth the extra overhead of taking the pi_lock on
|
||||||
|
+ * every fork/clone.
|
||||||
|
+ */
|
||||||
|
+ if (data_race(!src->user_cpus_ptr))
|
||||||
+ return 0;
|
+ return 0;
|
||||||
+
|
+
|
||||||
+ dst->user_cpus_ptr = kmalloc_node(cpumask_size(), GFP_KERNEL, node);
|
+ user_mask = alloc_user_cpus_ptr(node);
|
||||||
+ if (!dst->user_cpus_ptr)
|
+ if (!user_mask)
|
||||||
+ return -ENOMEM;
|
+ return -ENOMEM;
|
||||||
+
|
+
|
||||||
+ cpumask_copy(dst->user_cpus_ptr, src->user_cpus_ptr);
|
+ /*
|
||||||
|
+ * Use pi_lock to protect content of user_cpus_ptr
|
||||||
|
+ *
|
||||||
|
+ * Though unlikely, user_cpus_ptr can be reset to NULL by a concurrent
|
||||||
|
+ * do_set_cpus_allowed().
|
||||||
|
+ */
|
||||||
|
+ raw_spin_lock_irqsave(&src->pi_lock, flags);
|
||||||
|
+ if (src->user_cpus_ptr) {
|
||||||
|
+ swap(dst->user_cpus_ptr, user_mask);
|
||||||
|
+ cpumask_copy(dst->user_cpus_ptr, src->user_cpus_ptr);
|
||||||
|
+ }
|
||||||
|
+ raw_spin_unlock_irqrestore(&src->pi_lock, flags);
|
||||||
|
+
|
||||||
|
+ if (unlikely(user_mask))
|
||||||
|
+ kfree(user_mask);
|
||||||
|
+
|
||||||
+ return 0;
|
+ return 0;
|
||||||
+}
|
+}
|
||||||
+
|
+
|
||||||
@@ -2641,6 +2775,8 @@ index 000000000000..acb8657e811d
|
|||||||
+
|
+
|
||||||
+static int affine_move_task(struct rq *rq, struct task_struct *p, int dest_cpu,
|
+static int affine_move_task(struct rq *rq, struct task_struct *p, int dest_cpu,
|
||||||
+ raw_spinlock_t *lock, unsigned long irq_flags)
|
+ raw_spinlock_t *lock, unsigned long irq_flags)
|
||||||
|
+ __releases(rq->lock)
|
||||||
|
+ __releases(p->pi_lock)
|
||||||
+{
|
+{
|
||||||
+ /* Can the task run on the task's current CPU? If so, we're done */
|
+ /* Can the task run on the task's current CPU? If so, we're done */
|
||||||
+ if (!cpumask_test_cpu(task_cpu(p), &p->cpus_mask)) {
|
+ if (!cpumask_test_cpu(task_cpu(p), &p->cpus_mask)) {
|
||||||
@@ -2678,8 +2814,7 @@ index 000000000000..acb8657e811d
|
|||||||
+}
|
+}
|
||||||
+
|
+
|
||||||
+static int __set_cpus_allowed_ptr_locked(struct task_struct *p,
|
+static int __set_cpus_allowed_ptr_locked(struct task_struct *p,
|
||||||
+ const struct cpumask *new_mask,
|
+ struct affinity_context *ctx,
|
||||||
+ u32 flags,
|
|
||||||
+ struct rq *rq,
|
+ struct rq *rq,
|
||||||
+ raw_spinlock_t *lock,
|
+ raw_spinlock_t *lock,
|
||||||
+ unsigned long irq_flags)
|
+ unsigned long irq_flags)
|
||||||
@@ -2687,7 +2822,6 @@ index 000000000000..acb8657e811d
|
|||||||
+ const struct cpumask *cpu_allowed_mask = task_cpu_possible_mask(p);
|
+ const struct cpumask *cpu_allowed_mask = task_cpu_possible_mask(p);
|
||||||
+ const struct cpumask *cpu_valid_mask = cpu_active_mask;
|
+ const struct cpumask *cpu_valid_mask = cpu_active_mask;
|
||||||
+ bool kthread = p->flags & PF_KTHREAD;
|
+ bool kthread = p->flags & PF_KTHREAD;
|
||||||
+ struct cpumask *user_mask = NULL;
|
|
||||||
+ int dest_cpu;
|
+ int dest_cpu;
|
||||||
+ int ret = 0;
|
+ int ret = 0;
|
||||||
+
|
+
|
||||||
@@ -2705,7 +2839,7 @@ index 000000000000..acb8657e811d
|
|||||||
+ cpu_valid_mask = cpu_online_mask;
|
+ cpu_valid_mask = cpu_online_mask;
|
||||||
+ }
|
+ }
|
||||||
+
|
+
|
||||||
+ if (!kthread && !cpumask_subset(new_mask, cpu_allowed_mask)) {
|
+ if (!kthread && !cpumask_subset(ctx->new_mask, cpu_allowed_mask)) {
|
||||||
+ ret = -EINVAL;
|
+ ret = -EINVAL;
|
||||||
+ goto out;
|
+ goto out;
|
||||||
+ }
|
+ }
|
||||||
@@ -2714,30 +2848,23 @@ index 000000000000..acb8657e811d
|
|||||||
+ * Must re-check here, to close a race against __kthread_bind(),
|
+ * Must re-check here, to close a race against __kthread_bind(),
|
||||||
+ * sched_setaffinity() is not guaranteed to observe the flag.
|
+ * sched_setaffinity() is not guaranteed to observe the flag.
|
||||||
+ */
|
+ */
|
||||||
+ if ((flags & SCA_CHECK) && (p->flags & PF_NO_SETAFFINITY)) {
|
+ if ((ctx->flags & SCA_CHECK) && (p->flags & PF_NO_SETAFFINITY)) {
|
||||||
+ ret = -EINVAL;
|
+ ret = -EINVAL;
|
||||||
+ goto out;
|
+ goto out;
|
||||||
+ }
|
+ }
|
||||||
+
|
+
|
||||||
+ if (cpumask_equal(&p->cpus_mask, new_mask))
|
+ if (cpumask_equal(&p->cpus_mask, ctx->new_mask))
|
||||||
+ goto out;
|
+ goto out;
|
||||||
+
|
+
|
||||||
+ dest_cpu = cpumask_any_and(cpu_valid_mask, new_mask);
|
+ dest_cpu = cpumask_any_and(cpu_valid_mask, ctx->new_mask);
|
||||||
+ if (dest_cpu >= nr_cpu_ids) {
|
+ if (dest_cpu >= nr_cpu_ids) {
|
||||||
+ ret = -EINVAL;
|
+ ret = -EINVAL;
|
||||||
+ goto out;
|
+ goto out;
|
||||||
+ }
|
+ }
|
||||||
+
|
+
|
||||||
+ __do_set_cpus_allowed(p, new_mask);
|
+ __do_set_cpus_allowed(p, ctx);
|
||||||
+
|
+
|
||||||
+ if (flags & SCA_USER)
|
+ return affine_move_task(rq, p, dest_cpu, lock, irq_flags);
|
||||||
+ user_mask = clear_user_cpus_ptr(p);
|
|
||||||
+
|
|
||||||
+ ret = affine_move_task(rq, p, dest_cpu, lock, irq_flags);
|
|
||||||
+
|
|
||||||
+ kfree(user_mask);
|
|
||||||
+
|
|
||||||
+ return ret;
|
|
||||||
+
|
+
|
||||||
+out:
|
+out:
|
||||||
+ __task_access_unlock(p, lock);
|
+ __task_access_unlock(p, lock);
|
||||||
@@ -2748,7 +2875,6 @@ index 000000000000..acb8657e811d
|
|||||||
+
|
+
|
||||||
+/*
|
+/*
|
||||||
+ * Change a given task's CPU affinity. Migrate the thread to a
|
+ * Change a given task's CPU affinity. Migrate the thread to a
|
||||||
+ * proper CPU and schedule it away if the CPU it's executing on
|
|
||||||
+ * is removed from the allowed bitmask.
|
+ * is removed from the allowed bitmask.
|
||||||
+ *
|
+ *
|
||||||
+ * NOTE: the caller must have a valid reference to the task, the
|
+ * NOTE: the caller must have a valid reference to the task, the
|
||||||
@@ -2756,7 +2882,7 @@ index 000000000000..acb8657e811d
|
|||||||
+ * call is not atomic; no spinlocks may be held.
|
+ * call is not atomic; no spinlocks may be held.
|
||||||
+ */
|
+ */
|
||||||
+static int __set_cpus_allowed_ptr(struct task_struct *p,
|
+static int __set_cpus_allowed_ptr(struct task_struct *p,
|
||||||
+ const struct cpumask *new_mask, u32 flags)
|
+ struct affinity_context *ctx)
|
||||||
+{
|
+{
|
||||||
+ unsigned long irq_flags;
|
+ unsigned long irq_flags;
|
||||||
+ struct rq *rq;
|
+ struct rq *rq;
|
||||||
@@ -2764,20 +2890,36 @@ index 000000000000..acb8657e811d
|
|||||||
+
|
+
|
||||||
+ raw_spin_lock_irqsave(&p->pi_lock, irq_flags);
|
+ raw_spin_lock_irqsave(&p->pi_lock, irq_flags);
|
||||||
+ rq = __task_access_lock(p, &lock);
|
+ rq = __task_access_lock(p, &lock);
|
||||||
|
+ /*
|
||||||
|
+ * Masking should be skipped if SCA_USER or any of the SCA_MIGRATE_*
|
||||||
|
+ * flags are set.
|
||||||
|
+ */
|
||||||
|
+ if (p->user_cpus_ptr &&
|
||||||
|
+ !(ctx->flags & SCA_USER) &&
|
||||||
|
+ cpumask_and(rq->scratch_mask, ctx->new_mask, p->user_cpus_ptr))
|
||||||
|
+ ctx->new_mask = rq->scratch_mask;
|
||||||
+
|
+
|
||||||
+ return __set_cpus_allowed_ptr_locked(p, new_mask, flags, rq, lock, irq_flags);
|
+
|
||||||
|
+ return __set_cpus_allowed_ptr_locked(p, ctx, rq, lock, irq_flags);
|
||||||
+}
|
+}
|
||||||
+
|
+
|
||||||
+int set_cpus_allowed_ptr(struct task_struct *p, const struct cpumask *new_mask)
|
+int set_cpus_allowed_ptr(struct task_struct *p, const struct cpumask *new_mask)
|
||||||
+{
|
+{
|
||||||
+ return __set_cpus_allowed_ptr(p, new_mask, 0);
|
+ struct affinity_context ac = {
|
||||||
|
+ .new_mask = new_mask,
|
||||||
|
+ .flags = 0,
|
||||||
|
+ };
|
||||||
|
+
|
||||||
|
+ return __set_cpus_allowed_ptr(p, &ac);
|
||||||
+}
|
+}
|
||||||
+EXPORT_SYMBOL_GPL(set_cpus_allowed_ptr);
|
+EXPORT_SYMBOL_GPL(set_cpus_allowed_ptr);
|
||||||
+
|
+
|
||||||
+/*
|
+/*
|
||||||
+ * Change a given task's CPU affinity to the intersection of its current
|
+ * Change a given task's CPU affinity to the intersection of its current
|
||||||
+ * affinity mask and @subset_mask, writing the resulting mask to @new_mask
|
+ * affinity mask and @subset_mask, writing the resulting mask to @new_mask.
|
||||||
+ * and pointing @p->user_cpus_ptr to a copy of the old mask.
|
+ * If user_cpus_ptr is defined, use it as the basis for restricting CPU
|
||||||
|
+ * affinity or use cpu_online_mask instead.
|
||||||
|
+ *
|
||||||
+ * If the resulting mask is empty, leave the affinity unchanged and return
|
+ * If the resulting mask is empty, leave the affinity unchanged and return
|
||||||
+ * -EINVAL.
|
+ * -EINVAL.
|
||||||
+ */
|
+ */
|
||||||
@@ -2785,48 +2927,34 @@ index 000000000000..acb8657e811d
|
|||||||
+ struct cpumask *new_mask,
|
+ struct cpumask *new_mask,
|
||||||
+ const struct cpumask *subset_mask)
|
+ const struct cpumask *subset_mask)
|
||||||
+{
|
+{
|
||||||
+ struct cpumask *user_mask = NULL;
|
+ struct affinity_context ac = {
|
||||||
|
+ .new_mask = new_mask,
|
||||||
|
+ .flags = 0,
|
||||||
|
+ };
|
||||||
+ unsigned long irq_flags;
|
+ unsigned long irq_flags;
|
||||||
+ raw_spinlock_t *lock;
|
+ raw_spinlock_t *lock;
|
||||||
+ struct rq *rq;
|
+ struct rq *rq;
|
||||||
+ int err;
|
+ int err;
|
||||||
+
|
+
|
||||||
+ if (!p->user_cpus_ptr) {
|
|
||||||
+ user_mask = kmalloc(cpumask_size(), GFP_KERNEL);
|
|
||||||
+ if (!user_mask)
|
|
||||||
+ return -ENOMEM;
|
|
||||||
+ }
|
|
||||||
+
|
|
||||||
+ raw_spin_lock_irqsave(&p->pi_lock, irq_flags);
|
+ raw_spin_lock_irqsave(&p->pi_lock, irq_flags);
|
||||||
+ rq = __task_access_lock(p, &lock);
|
+ rq = __task_access_lock(p, &lock);
|
||||||
+
|
+
|
||||||
+ if (!cpumask_and(new_mask, &p->cpus_mask, subset_mask)) {
|
+ if (!cpumask_and(new_mask, task_user_cpus(p), subset_mask)) {
|
||||||
+ err = -EINVAL;
|
+ err = -EINVAL;
|
||||||
+ goto err_unlock;
|
+ goto err_unlock;
|
||||||
+ }
|
+ }
|
||||||
+
|
+
|
||||||
+ /*
|
+ return __set_cpus_allowed_ptr_locked(p, &ac, rq, lock, irq_flags);
|
||||||
+ * We're about to butcher the task affinity, so keep track of what
|
|
||||||
+ * the user asked for in case we're able to restore it later on.
|
|
||||||
+ */
|
|
||||||
+ if (user_mask) {
|
|
||||||
+ cpumask_copy(user_mask, p->cpus_ptr);
|
|
||||||
+ p->user_cpus_ptr = user_mask;
|
|
||||||
+ }
|
|
||||||
+
|
|
||||||
+ /*return __set_cpus_allowed_ptr_locked(p, new_mask, 0, rq, &rf);*/
|
|
||||||
+ return __set_cpus_allowed_ptr_locked(p, new_mask, 0, rq, lock, irq_flags);
|
|
||||||
+
|
+
|
||||||
+err_unlock:
|
+err_unlock:
|
||||||
+ __task_access_unlock(p, lock);
|
+ __task_access_unlock(p, lock);
|
||||||
+ raw_spin_unlock_irqrestore(&p->pi_lock, irq_flags);
|
+ raw_spin_unlock_irqrestore(&p->pi_lock, irq_flags);
|
||||||
+ kfree(user_mask);
|
|
||||||
+ return err;
|
+ return err;
|
||||||
+}
|
+}
|
||||||
+
|
+
|
||||||
+/*
|
+/*
|
||||||
+ * Restrict the CPU affinity of task @p so that it is a subset of
|
+ * Restrict the CPU affinity of task @p so that it is a subset of
|
||||||
+ * task_cpu_possible_mask() and point @p->user_cpu_ptr to a copy of the
|
+ * task_cpu_possible_mask() and point @p->user_cpus_ptr to a copy of the
|
||||||
+ * old affinity mask. If the resulting mask is empty, we warn and walk
|
+ * old affinity mask. If the resulting mask is empty, we warn and walk
|
||||||
+ * up the cpuset hierarchy until we find a suitable mask.
|
+ * up the cpuset hierarchy until we find a suitable mask.
|
||||||
+ */
|
+ */
|
||||||
@@ -2870,34 +2998,29 @@ index 000000000000..acb8657e811d
|
|||||||
+}
|
+}
|
||||||
+
|
+
|
||||||
+static int
|
+static int
|
||||||
+__sched_setaffinity(struct task_struct *p, const struct cpumask *mask);
|
+__sched_setaffinity(struct task_struct *p, struct affinity_context *ctx);
|
||||||
+
|
+
|
||||||
+/*
|
+/*
|
||||||
+ * Restore the affinity of a task @p which was previously restricted by a
|
+ * Restore the affinity of a task @p which was previously restricted by a
|
||||||
+ * call to force_compatible_cpus_allowed_ptr(). This will clear (and free)
|
+ * call to force_compatible_cpus_allowed_ptr().
|
||||||
+ * @p->user_cpus_ptr.
|
|
||||||
+ *
|
+ *
|
||||||
+ * It is the caller's responsibility to serialise this with any calls to
|
+ * It is the caller's responsibility to serialise this with any calls to
|
||||||
+ * force_compatible_cpus_allowed_ptr(@p).
|
+ * force_compatible_cpus_allowed_ptr(@p).
|
||||||
+ */
|
+ */
|
||||||
+void relax_compatible_cpus_allowed_ptr(struct task_struct *p)
|
+void relax_compatible_cpus_allowed_ptr(struct task_struct *p)
|
||||||
+{
|
+{
|
||||||
+ struct cpumask *user_mask = p->user_cpus_ptr;
|
+ struct affinity_context ac = {
|
||||||
+ unsigned long flags;
|
+ .new_mask = task_user_cpus(p),
|
||||||
|
+ .flags = 0,
|
||||||
|
+ };
|
||||||
|
+ int ret;
|
||||||
+
|
+
|
||||||
+ /*
|
+ /*
|
||||||
+ * Try to restore the old affinity mask. If this fails, then
|
+ * Try to restore the old affinity mask with __sched_setaffinity().
|
||||||
+ * we free the mask explicitly to avoid it being inherited across
|
+ * Cpuset masking will be done there too.
|
||||||
+ * a subsequent fork().
|
|
||||||
+ */
|
+ */
|
||||||
+ if (!user_mask || !__sched_setaffinity(p, user_mask))
|
+ ret = __sched_setaffinity(p, &ac);
|
||||||
+ return;
|
+ WARN_ON_ONCE(ret);
|
||||||
+
|
|
||||||
+ raw_spin_lock_irqsave(&p->pi_lock, flags);
|
|
||||||
+ user_mask = clear_user_cpus_ptr(p);
|
|
||||||
+ raw_spin_unlock_irqrestore(&p->pi_lock, flags);
|
|
||||||
+
|
|
||||||
+ kfree(user_mask);
|
|
||||||
+}
|
+}
|
||||||
+
|
+
|
||||||
+#else /* CONFIG_SMP */
|
+#else /* CONFIG_SMP */
|
||||||
@@ -2909,9 +3032,9 @@ index 000000000000..acb8657e811d
|
|||||||
+
|
+
|
||||||
+static inline int
|
+static inline int
|
||||||
+__set_cpus_allowed_ptr(struct task_struct *p,
|
+__set_cpus_allowed_ptr(struct task_struct *p,
|
||||||
+ const struct cpumask *new_mask, u32 flags)
|
+ struct affinity_context *ctx)
|
||||||
+{
|
+{
|
||||||
+ return set_cpus_allowed_ptr(p, new_mask);
|
+ return set_cpus_allowed_ptr(p, ctx->new_mask);
|
||||||
+}
|
+}
|
||||||
+
|
+
|
||||||
+static inline bool rq_has_pinned_tasks(struct rq *rq)
|
+static inline bool rq_has_pinned_tasks(struct rq *rq)
|
||||||
@@ -2919,6 +3042,11 @@ index 000000000000..acb8657e811d
|
|||||||
+ return false;
|
+ return false;
|
||||||
+}
|
+}
|
||||||
+
|
+
|
||||||
|
+static inline cpumask_t *alloc_user_cpus_ptr(int node)
|
||||||
|
+{
|
||||||
|
+ return NULL;
|
||||||
|
+}
|
||||||
|
+
|
||||||
+#endif /* !CONFIG_SMP */
|
+#endif /* !CONFIG_SMP */
|
||||||
+
|
+
|
||||||
+static void
|
+static void
|
||||||
@@ -3030,13 +3158,6 @@ index 000000000000..acb8657e811d
|
|||||||
+ if (!llist)
|
+ if (!llist)
|
||||||
+ return;
|
+ return;
|
||||||
+
|
+
|
||||||
+ /*
|
|
||||||
+ * rq::ttwu_pending racy indication of out-standing wakeups.
|
|
||||||
+ * Races such that false-negatives are possible, since they
|
|
||||||
+ * are shorter lived that false-positives would be.
|
|
||||||
+ */
|
|
||||||
+ WRITE_ONCE(rq->ttwu_pending, 0);
|
|
||||||
+
|
|
||||||
+ rq_lock_irqsave(rq, &rf);
|
+ rq_lock_irqsave(rq, &rf);
|
||||||
+ update_rq_clock(rq);
|
+ update_rq_clock(rq);
|
||||||
+
|
+
|
||||||
@@ -3050,6 +3171,17 @@ index 000000000000..acb8657e811d
|
|||||||
+ ttwu_do_activate(rq, p, p->sched_remote_wakeup ? WF_MIGRATED : 0);
|
+ ttwu_do_activate(rq, p, p->sched_remote_wakeup ? WF_MIGRATED : 0);
|
||||||
+ }
|
+ }
|
||||||
+
|
+
|
||||||
|
+ /*
|
||||||
|
+ * Must be after enqueueing at least once task such that
|
||||||
|
+ * idle_cpu() does not observe a false-negative -- if it does,
|
||||||
|
+ * it is possible for select_idle_siblings() to stack a number
|
||||||
|
+ * of tasks on this CPU during that window.
|
||||||
|
+ *
|
||||||
|
+ * It is ok to clear ttwu_pending when another task pending.
|
||||||
|
+ * We will receive IPI after local irq enabled and then enqueue it.
|
||||||
|
+ * Since now nr_running > 0, idle_cpu() will always get correct result.
|
||||||
|
+ */
|
||||||
|
+ WRITE_ONCE(rq->ttwu_pending, 0);
|
||||||
+ rq_unlock_irqrestore(rq, &rf);
|
+ rq_unlock_irqrestore(rq, &rf);
|
||||||
+}
|
+}
|
||||||
+
|
+
|
||||||
@@ -4635,7 +4767,9 @@ index 000000000000..acb8657e811d
|
|||||||
+ struct rq *rq = cpu_rq(cpu);
|
+ struct rq *rq = cpu_rq(cpu);
|
||||||
+ u64 resched_latency;
|
+ u64 resched_latency;
|
||||||
+
|
+
|
||||||
+ arch_scale_freq_tick();
|
+ if (housekeeping_cpu(cpu, HK_TYPE_TICK))
|
||||||
|
+ arch_scale_freq_tick();
|
||||||
|
+
|
||||||
+ sched_clock_tick();
|
+ sched_clock_tick();
|
||||||
+
|
+
|
||||||
+ raw_spin_lock(&rq->lock);
|
+ raw_spin_lock(&rq->lock);
|
||||||
@@ -4734,7 +4868,7 @@ index 000000000000..acb8657e811d
|
|||||||
+ int i;
|
+ int i;
|
||||||
+
|
+
|
||||||
+ for_each_cpu_wrap(i, &chk, cpu) {
|
+ for_each_cpu_wrap(i, &chk, cpu) {
|
||||||
+ if (cpumask_subset(cpu_smt_mask(i), &chk) &&
|
+ if (!cpumask_intersects(cpu_smt_mask(i), sched_idle_mask) &&\
|
||||||
+ sg_balance_trigger(i))
|
+ sg_balance_trigger(i))
|
||||||
+ return;
|
+ return;
|
||||||
+ }
|
+ }
|
||||||
@@ -4857,6 +4991,7 @@ index 000000000000..acb8657e811d
|
|||||||
+static void sched_tick_stop(int cpu)
|
+static void sched_tick_stop(int cpu)
|
||||||
+{
|
+{
|
||||||
+ struct tick_work *twork;
|
+ struct tick_work *twork;
|
||||||
|
+ int os;
|
||||||
+
|
+
|
||||||
+ if (housekeeping_cpu(cpu, HK_TYPE_TICK))
|
+ if (housekeeping_cpu(cpu, HK_TYPE_TICK))
|
||||||
+ return;
|
+ return;
|
||||||
@@ -4864,7 +4999,10 @@ index 000000000000..acb8657e811d
|
|||||||
+ WARN_ON_ONCE(!tick_work_cpu);
|
+ WARN_ON_ONCE(!tick_work_cpu);
|
||||||
+
|
+
|
||||||
+ twork = per_cpu_ptr(tick_work_cpu, cpu);
|
+ twork = per_cpu_ptr(tick_work_cpu, cpu);
|
||||||
+ cancel_delayed_work_sync(&twork->work);
|
+ /* There cannot be competing actions, but don't rely on stop-machine. */
|
||||||
|
+ os = atomic_xchg(&twork->state, TICK_SCHED_REMOTE_OFFLINING);
|
||||||
|
+ WARN_ON_ONCE(os != TICK_SCHED_REMOTE_RUNNING);
|
||||||
|
+ /* Don't cancel, as this would mess up the state machine. */
|
||||||
+}
|
+}
|
||||||
+#endif /* CONFIG_HOTPLUG_CPU */
|
+#endif /* CONFIG_HOTPLUG_CPU */
|
||||||
+
|
+
|
||||||
@@ -4988,8 +5126,7 @@ index 000000000000..acb8657e811d
|
|||||||
+ pr_err("Preemption disabled at:");
|
+ pr_err("Preemption disabled at:");
|
||||||
+ print_ip_sym(KERN_ERR, preempt_disable_ip);
|
+ print_ip_sym(KERN_ERR, preempt_disable_ip);
|
||||||
+ }
|
+ }
|
||||||
+ if (panic_on_warn)
|
+ check_panic_on_warn("scheduling while atomic");
|
||||||
+ panic("scheduling while atomic\n");
|
|
||||||
+
|
+
|
||||||
+ dump_stack();
|
+ dump_stack();
|
||||||
+ add_taint(TAINT_WARN, LOCKDEP_STILL_OK);
|
+ add_taint(TAINT_WARN, LOCKDEP_STILL_OK);
|
||||||
@@ -5305,7 +5442,7 @@ index 000000000000..acb8657e811d
|
|||||||
+ prev->sched_contributes_to_load =
|
+ prev->sched_contributes_to_load =
|
||||||
+ (prev_state & TASK_UNINTERRUPTIBLE) &&
|
+ (prev_state & TASK_UNINTERRUPTIBLE) &&
|
||||||
+ !(prev_state & TASK_NOLOAD) &&
|
+ !(prev_state & TASK_NOLOAD) &&
|
||||||
+ !(prev->flags & TASK_FROZEN);
|
+ !(prev_state & TASK_FROZEN);
|
||||||
+
|
+
|
||||||
+ if (prev->sched_contributes_to_load)
|
+ if (prev->sched_contributes_to_load)
|
||||||
+ rq->nr_uninterruptible++;
|
+ rq->nr_uninterruptible++;
|
||||||
@@ -6653,7 +6790,7 @@ index 000000000000..acb8657e811d
|
|||||||
+#endif
|
+#endif
|
||||||
+
|
+
|
||||||
+static int
|
+static int
|
||||||
+__sched_setaffinity(struct task_struct *p, const struct cpumask *mask)
|
+__sched_setaffinity(struct task_struct *p, struct affinity_context *ctx)
|
||||||
+{
|
+{
|
||||||
+ int retval;
|
+ int retval;
|
||||||
+ cpumask_var_t cpus_allowed, new_mask;
|
+ cpumask_var_t cpus_allowed, new_mask;
|
||||||
@@ -6667,9 +6804,12 @@ index 000000000000..acb8657e811d
|
|||||||
+ }
|
+ }
|
||||||
+
|
+
|
||||||
+ cpuset_cpus_allowed(p, cpus_allowed);
|
+ cpuset_cpus_allowed(p, cpus_allowed);
|
||||||
+ cpumask_and(new_mask, mask, cpus_allowed);
|
+ cpumask_and(new_mask, ctx->new_mask, cpus_allowed);
|
||||||
+again:
|
+
|
||||||
+ retval = __set_cpus_allowed_ptr(p, new_mask, SCA_CHECK | SCA_USER);
|
+ ctx->new_mask = new_mask;
|
||||||
|
+ ctx->flags |= SCA_CHECK;
|
||||||
|
+
|
||||||
|
+ retval = __set_cpus_allowed_ptr(p, ctx);
|
||||||
+ if (retval)
|
+ if (retval)
|
||||||
+ goto out_free_new_mask;
|
+ goto out_free_new_mask;
|
||||||
+
|
+
|
||||||
@@ -6681,7 +6821,24 @@ index 000000000000..acb8657e811d
|
|||||||
+ * cpuset's cpus_allowed
|
+ * cpuset's cpus_allowed
|
||||||
+ */
|
+ */
|
||||||
+ cpumask_copy(new_mask, cpus_allowed);
|
+ cpumask_copy(new_mask, cpus_allowed);
|
||||||
+ goto again;
|
+
|
||||||
|
+ /*
|
||||||
|
+ * If SCA_USER is set, a 2nd call to __set_cpus_allowed_ptr()
|
||||||
|
+ * will restore the previous user_cpus_ptr value.
|
||||||
|
+ *
|
||||||
|
+ * In the unlikely event a previous user_cpus_ptr exists,
|
||||||
|
+ * we need to further restrict the mask to what is allowed
|
||||||
|
+ * by that old user_cpus_ptr.
|
||||||
|
+ */
|
||||||
|
+ if (unlikely((ctx->flags & SCA_USER) && ctx->user_mask)) {
|
||||||
|
+ bool empty = !cpumask_and(new_mask, new_mask,
|
||||||
|
+ ctx->user_mask);
|
||||||
|
+
|
||||||
|
+ if (WARN_ON_ONCE(empty))
|
||||||
|
+ cpumask_copy(new_mask, cpus_allowed);
|
||||||
|
+ }
|
||||||
|
+ __set_cpus_allowed_ptr(p, ctx);
|
||||||
|
+ retval = -EINVAL;
|
||||||
+ }
|
+ }
|
||||||
+
|
+
|
||||||
+out_free_new_mask:
|
+out_free_new_mask:
|
||||||
@@ -6693,6 +6850,8 @@ index 000000000000..acb8657e811d
|
|||||||
+
|
+
|
||||||
+long sched_setaffinity(pid_t pid, const struct cpumask *in_mask)
|
+long sched_setaffinity(pid_t pid, const struct cpumask *in_mask)
|
||||||
+{
|
+{
|
||||||
|
+ struct affinity_context ac;
|
||||||
|
+ struct cpumask *user_mask;
|
||||||
+ struct task_struct *p;
|
+ struct task_struct *p;
|
||||||
+ int retval;
|
+ int retval;
|
||||||
+
|
+
|
||||||
@@ -6727,7 +6886,27 @@ index 000000000000..acb8657e811d
|
|||||||
+ if (retval)
|
+ if (retval)
|
||||||
+ goto out_put_task;
|
+ goto out_put_task;
|
||||||
+
|
+
|
||||||
+ retval = __sched_setaffinity(p, in_mask);
|
+ /*
|
||||||
|
+ * With non-SMP configs, user_cpus_ptr/user_mask isn't used and
|
||||||
|
+ * alloc_user_cpus_ptr() returns NULL.
|
||||||
|
+ */
|
||||||
|
+ user_mask = alloc_user_cpus_ptr(NUMA_NO_NODE);
|
||||||
|
+ if (user_mask) {
|
||||||
|
+ cpumask_copy(user_mask, in_mask);
|
||||||
|
+ } else if (IS_ENABLED(CONFIG_SMP)) {
|
||||||
|
+ retval = -ENOMEM;
|
||||||
|
+ goto out_put_task;
|
||||||
|
+ }
|
||||||
|
+
|
||||||
|
+ ac = (struct affinity_context){
|
||||||
|
+ .new_mask = in_mask,
|
||||||
|
+ .user_mask = user_mask,
|
||||||
|
+ .flags = SCA_USER,
|
||||||
|
+ };
|
||||||
|
+
|
||||||
|
+ retval = __sched_setaffinity(p, &ac);
|
||||||
|
+ kfree(ac.user_mask);
|
||||||
|
+
|
||||||
+out_put_task:
|
+out_put_task:
|
||||||
+ put_task_struct(p);
|
+ put_task_struct(p);
|
||||||
+ return retval;
|
+ return retval;
|
||||||
@@ -7483,6 +7662,12 @@ index 000000000000..acb8657e811d
|
|||||||
+ */
|
+ */
|
||||||
+void __init init_idle(struct task_struct *idle, int cpu)
|
+void __init init_idle(struct task_struct *idle, int cpu)
|
||||||
+{
|
+{
|
||||||
|
+#ifdef CONFIG_SMP
|
||||||
|
+ struct affinity_context ac = (struct affinity_context) {
|
||||||
|
+ .new_mask = cpumask_of(cpu),
|
||||||
|
+ .flags = 0,
|
||||||
|
+ };
|
||||||
|
+#endif
|
||||||
+ struct rq *rq = cpu_rq(cpu);
|
+ struct rq *rq = cpu_rq(cpu);
|
||||||
+ unsigned long flags;
|
+ unsigned long flags;
|
||||||
+
|
+
|
||||||
@@ -7509,7 +7694,7 @@ index 000000000000..acb8657e811d
|
|||||||
+ *
|
+ *
|
||||||
+ * And since this is boot we can forgo the serialisation.
|
+ * And since this is boot we can forgo the serialisation.
|
||||||
+ */
|
+ */
|
||||||
+ set_cpus_allowed_common(idle, cpumask_of(cpu));
|
+ set_cpus_allowed_common(idle, &ac);
|
||||||
+#endif
|
+#endif
|
||||||
+
|
+
|
||||||
+ /* Silence PROVE_RCU */
|
+ /* Silence PROVE_RCU */
|
||||||
@@ -8137,6 +8322,8 @@ index 000000000000..acb8657e811d
|
|||||||
+
|
+
|
||||||
+ hrtick_rq_init(rq);
|
+ hrtick_rq_init(rq);
|
||||||
+ atomic_set(&rq->nr_iowait, 0);
|
+ atomic_set(&rq->nr_iowait, 0);
|
||||||
|
+
|
||||||
|
+ zalloc_cpumask_var_node(&rq->scratch_mask, GFP_KERNEL, cpu_to_node(i));
|
||||||
+ }
|
+ }
|
||||||
+#ifdef CONFIG_SMP
|
+#ifdef CONFIG_SMP
|
||||||
+ /* Set rq->online for cpu 0 */
|
+ /* Set rq->online for cpu 0 */
|
||||||
@@ -8653,10 +8840,10 @@ index 000000000000..1212a031700e
|
|||||||
+{}
|
+{}
|
||||||
diff --git a/kernel/sched/alt_sched.h b/kernel/sched/alt_sched.h
|
diff --git a/kernel/sched/alt_sched.h b/kernel/sched/alt_sched.h
|
||||||
new file mode 100644
|
new file mode 100644
|
||||||
index 000000000000..c32403ed82b6
|
index 000000000000..0b563999d4c1
|
||||||
--- /dev/null
|
--- /dev/null
|
||||||
+++ b/kernel/sched/alt_sched.h
|
+++ b/kernel/sched/alt_sched.h
|
||||||
@@ -0,0 +1,668 @@
|
@@ -0,0 +1,671 @@
|
||||||
+#ifndef ALT_SCHED_H
|
+#ifndef ALT_SCHED_H
|
||||||
+#define ALT_SCHED_H
|
+#define ALT_SCHED_H
|
||||||
+
|
+
|
||||||
@@ -8903,6 +9090,9 @@ index 000000000000..c32403ed82b6
|
|||||||
+#endif
|
+#endif
|
||||||
+ atomic_t nohz_flags;
|
+ atomic_t nohz_flags;
|
||||||
+#endif /* CONFIG_NO_HZ_COMMON */
|
+#endif /* CONFIG_NO_HZ_COMMON */
|
||||||
|
+
|
||||||
|
+ /* Scratch cpumask to be temporarily used under rq_lock */
|
||||||
|
+ cpumask_var_t scratch_mask;
|
||||||
+};
|
+};
|
||||||
+
|
+
|
||||||
+extern unsigned long rq_load_util(struct rq *rq, unsigned long max);
|
+extern unsigned long rq_load_util(struct rq *rq, unsigned long max);
|
||||||
@@ -9874,7 +10064,7 @@ index 3a0e0dc28721..e8a7d84aa5a5 100644
|
|||||||
static inline int
|
static inline int
|
||||||
update_thermal_load_avg(u64 now, struct rq *rq, u64 capacity)
|
update_thermal_load_avg(u64 now, struct rq *rq, u64 capacity)
|
||||||
diff --git a/kernel/sched/sched.h b/kernel/sched/sched.h
|
diff --git a/kernel/sched/sched.h b/kernel/sched/sched.h
|
||||||
index a4a20046e586..c363693cd869 100644
|
index 771f8ddb7053..787a5069d69a 100644
|
||||||
--- a/kernel/sched/sched.h
|
--- a/kernel/sched/sched.h
|
||||||
+++ b/kernel/sched/sched.h
|
+++ b/kernel/sched/sched.h
|
||||||
@@ -5,6 +5,10 @@
|
@@ -5,6 +5,10 @@
|
||||||
@@ -9888,7 +10078,7 @@ index a4a20046e586..c363693cd869 100644
|
|||||||
#include <linux/sched/affinity.h>
|
#include <linux/sched/affinity.h>
|
||||||
#include <linux/sched/autogroup.h>
|
#include <linux/sched/autogroup.h>
|
||||||
#include <linux/sched/cpufreq.h>
|
#include <linux/sched/cpufreq.h>
|
||||||
@@ -3183,4 +3187,9 @@ static inline void update_current_exec_runtime(struct task_struct *curr,
|
@@ -3261,4 +3265,9 @@ static inline void update_current_exec_runtime(struct task_struct *curr,
|
||||||
cgroup_account_cputime(curr, delta_exec);
|
cgroup_account_cputime(curr, delta_exec);
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -9930,7 +10120,7 @@ index 857f837f52cb..5486c63e4790 100644
|
|||||||
}
|
}
|
||||||
return 0;
|
return 0;
|
||||||
diff --git a/kernel/sched/stats.h b/kernel/sched/stats.h
|
diff --git a/kernel/sched/stats.h b/kernel/sched/stats.h
|
||||||
index 84a188913cc9..53934e7ef5db 100644
|
index 38f3698f5e5b..b9d597394316 100644
|
||||||
--- a/kernel/sched/stats.h
|
--- a/kernel/sched/stats.h
|
||||||
+++ b/kernel/sched/stats.h
|
+++ b/kernel/sched/stats.h
|
||||||
@@ -89,6 +89,7 @@ static inline void rq_sched_info_depart (struct rq *rq, unsigned long long delt
|
@@ -89,6 +89,7 @@ static inline void rq_sched_info_depart (struct rq *rq, unsigned long long delt
|
||||||
@@ -10005,7 +10195,7 @@ index 8739c2a5a54e..d8dd6c15eb47 100644
|
|||||||
+#endif /* CONFIG_NUMA */
|
+#endif /* CONFIG_NUMA */
|
||||||
+#endif
|
+#endif
|
||||||
diff --git a/kernel/sysctl.c b/kernel/sysctl.c
|
diff --git a/kernel/sysctl.c b/kernel/sysctl.c
|
||||||
index c6d9dec11b74..2bc42ce8b48e 100644
|
index 137d4abe3eda..6bada3a6d571 100644
|
||||||
--- a/kernel/sysctl.c
|
--- a/kernel/sysctl.c
|
||||||
+++ b/kernel/sysctl.c
|
+++ b/kernel/sysctl.c
|
||||||
@@ -93,6 +93,10 @@ EXPORT_SYMBOL_GPL(sysctl_long_vals);
|
@@ -93,6 +93,10 @@ EXPORT_SYMBOL_GPL(sysctl_long_vals);
|
||||||
@@ -10019,7 +10209,7 @@ index c6d9dec11b74..2bc42ce8b48e 100644
|
|||||||
#ifdef CONFIG_PERF_EVENTS
|
#ifdef CONFIG_PERF_EVENTS
|
||||||
static const int six_hundred_forty_kb = 640 * 1024;
|
static const int six_hundred_forty_kb = 640 * 1024;
|
||||||
#endif
|
#endif
|
||||||
@@ -1953,6 +1959,17 @@ static struct ctl_table kern_table[] = {
|
@@ -1934,6 +1938,17 @@ static struct ctl_table kern_table[] = {
|
||||||
.proc_handler = proc_dointvec,
|
.proc_handler = proc_dointvec,
|
||||||
},
|
},
|
||||||
#endif
|
#endif
|
||||||
@@ -10113,10 +10303,10 @@ index cb925e8ef9a8..67d823510f5c 100644
|
|||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
diff --git a/kernel/trace/trace_selftest.c b/kernel/trace/trace_selftest.c
|
diff --git a/kernel/trace/trace_selftest.c b/kernel/trace/trace_selftest.c
|
||||||
index a2d301f58ced..2ccdede8585c 100644
|
index ff0536cea968..ce266990006d 100644
|
||||||
--- a/kernel/trace/trace_selftest.c
|
--- a/kernel/trace/trace_selftest.c
|
||||||
+++ b/kernel/trace/trace_selftest.c
|
+++ b/kernel/trace/trace_selftest.c
|
||||||
@@ -1143,10 +1143,15 @@ static int trace_wakeup_test_thread(void *data)
|
@@ -1150,10 +1150,15 @@ static int trace_wakeup_test_thread(void *data)
|
||||||
{
|
{
|
||||||
/* Make this a -deadline thread */
|
/* Make this a -deadline thread */
|
||||||
static const struct sched_attr attr = {
|
static const struct sched_attr attr = {
|
||||||
@@ -10132,3 +10322,6 @@ index a2d301f58ced..2ccdede8585c 100644
|
|||||||
};
|
};
|
||||||
struct wakeup_test_data *x = data;
|
struct wakeup_test_data *x = data;
|
||||||
|
|
||||||
|
--
|
||||||
|
2.39.2
|
||||||
|
|
||||||
|
@@ -0,0 +1,151 @@
|
|||||||
|
From d50977b164e708bf523a35ef53315355528c3ca6 Mon Sep 17 00:00:00 2001
|
||||||
|
From: "Jan Alexander Steffens (heftig)" <jan.steffens@gmail.com>
|
||||||
|
Date: Mon, 16 Sep 2019 04:53:20 +0200
|
||||||
|
Subject: [PATCH] ZEN: Add sysctl and CONFIG to disallow unprivileged
|
||||||
|
CLONE_NEWUSER
|
||||||
|
|
||||||
|
Our default behavior continues to match the vanilla kernel.
|
||||||
|
---
|
||||||
|
include/linux/user_namespace.h | 4 ++++
|
||||||
|
init/Kconfig | 16 ++++++++++++++++
|
||||||
|
kernel/fork.c | 14 ++++++++++++++
|
||||||
|
kernel/sysctl.c | 12 ++++++++++++
|
||||||
|
kernel/user_namespace.c | 7 +++++++
|
||||||
|
5 files changed, 53 insertions(+)
|
||||||
|
|
||||||
|
diff --git a/include/linux/user_namespace.h b/include/linux/user_namespace.h
|
||||||
|
index 45f09bec02c485..87b20e2ee27445 100644
|
||||||
|
--- a/include/linux/user_namespace.h
|
||||||
|
+++ b/include/linux/user_namespace.h
|
||||||
|
@@ -148,6 +148,8 @@ static inline void set_userns_rlimit_max(struct user_namespace *ns,
|
||||||
|
|
||||||
|
#ifdef CONFIG_USER_NS
|
||||||
|
|
||||||
|
+extern int unprivileged_userns_clone;
|
||||||
|
+
|
||||||
|
static inline struct user_namespace *get_user_ns(struct user_namespace *ns)
|
||||||
|
{
|
||||||
|
if (ns)
|
||||||
|
@@ -181,6 +183,8 @@ extern bool current_in_userns(const struct user_namespace *target_ns);
|
||||||
|
struct ns_common *ns_get_owner(struct ns_common *ns);
|
||||||
|
#else
|
||||||
|
|
||||||
|
+#define unprivileged_userns_clone 0
|
||||||
|
+
|
||||||
|
static inline struct user_namespace *get_user_ns(struct user_namespace *ns)
|
||||||
|
{
|
||||||
|
return &init_user_ns;
|
||||||
|
diff --git a/init/Kconfig b/init/Kconfig
|
||||||
|
index 94125d3b6893c7..9f7139b536f638 100644
|
||||||
|
--- a/init/Kconfig
|
||||||
|
+++ b/init/Kconfig
|
||||||
|
@@ -1247,6 +1247,22 @@ config USER_NS
|
||||||
|
|
||||||
|
If unsure, say N.
|
||||||
|
|
||||||
|
+config USER_NS_UNPRIVILEGED
|
||||||
|
+ bool "Allow unprivileged users to create namespaces"
|
||||||
|
+ default y
|
||||||
|
+ depends on USER_NS
|
||||||
|
+ help
|
||||||
|
+ When disabled, unprivileged users will not be able to create
|
||||||
|
+ new namespaces. Allowing users to create their own namespaces
|
||||||
|
+ has been part of several recent local privilege escalation
|
||||||
|
+ exploits, so if you need user namespaces but are
|
||||||
|
+ paranoid^Wsecurity-conscious you want to disable this.
|
||||||
|
+
|
||||||
|
+ This setting can be overridden at runtime via the
|
||||||
|
+ kernel.unprivileged_userns_clone sysctl.
|
||||||
|
+
|
||||||
|
+ If unsure, say Y.
|
||||||
|
+
|
||||||
|
config PID_NS
|
||||||
|
bool "PID Namespaces"
|
||||||
|
default y
|
||||||
|
diff --git a/kernel/fork.c b/kernel/fork.c
|
||||||
|
index 08969f5aa38d59..ff601cb7a1fae0 100644
|
||||||
|
--- a/kernel/fork.c
|
||||||
|
+++ b/kernel/fork.c
|
||||||
|
@@ -98,6 +98,10 @@
|
||||||
|
#include <linux/io_uring.h>
|
||||||
|
#include <linux/bpf.h>
|
||||||
|
|
||||||
|
+#ifdef CONFIG_USER_NS
|
||||||
|
+#include <linux/user_namespace.h>
|
||||||
|
+#endif
|
||||||
|
+
|
||||||
|
#include <asm/pgalloc.h>
|
||||||
|
#include <linux/uaccess.h>
|
||||||
|
#include <asm/mmu_context.h>
|
||||||
|
@@ -2008,6 +2012,10 @@ static __latent_entropy struct task_struct *copy_process(
|
||||||
|
if ((clone_flags & (CLONE_NEWUSER|CLONE_FS)) == (CLONE_NEWUSER|CLONE_FS))
|
||||||
|
return ERR_PTR(-EINVAL);
|
||||||
|
|
||||||
|
+ if ((clone_flags & CLONE_NEWUSER) && !unprivileged_userns_clone)
|
||||||
|
+ if (!capable(CAP_SYS_ADMIN))
|
||||||
|
+ return ERR_PTR(-EPERM);
|
||||||
|
+
|
||||||
|
/*
|
||||||
|
* Thread groups must share signals as well, and detached threads
|
||||||
|
* can only be started up within the thread group.
|
||||||
|
@@ -3166,6 +3174,12 @@ int ksys_unshare(unsigned long unshare_flags)
|
||||||
|
if (unshare_flags & CLONE_NEWNS)
|
||||||
|
unshare_flags |= CLONE_FS;
|
||||||
|
|
||||||
|
+ if ((unshare_flags & CLONE_NEWUSER) && !unprivileged_userns_clone) {
|
||||||
|
+ err = -EPERM;
|
||||||
|
+ if (!capable(CAP_SYS_ADMIN))
|
||||||
|
+ goto bad_unshare_out;
|
||||||
|
+ }
|
||||||
|
+
|
||||||
|
err = check_unshare_flags(unshare_flags);
|
||||||
|
if (err)
|
||||||
|
goto bad_unshare_out;
|
||||||
|
diff --git a/kernel/sysctl.c b/kernel/sysctl.c
|
||||||
|
index c6d9dec11b749d..9a4514ad481b21 100644
|
||||||
|
--- a/kernel/sysctl.c
|
||||||
|
+++ b/kernel/sysctl.c
|
||||||
|
@@ -81,6 +81,9 @@
|
||||||
|
#ifdef CONFIG_RT_MUTEXES
|
||||||
|
#include <linux/rtmutex.h>
|
||||||
|
#endif
|
||||||
|
+#ifdef CONFIG_USER_NS
|
||||||
|
+#include <linux/user_namespace.h>
|
||||||
|
+#endif
|
||||||
|
|
||||||
|
/* shared constants to be used in various sysctls */
|
||||||
|
const int sysctl_vals[] = { 0, 1, 2, 3, 4, 100, 200, 1000, 3000, INT_MAX, 65535, -1 };
|
||||||
|
@@ -1659,6 +1662,15 @@ static struct ctl_table kern_table[] = {
|
||||||
|
.mode = 0644,
|
||||||
|
.proc_handler = proc_dointvec,
|
||||||
|
},
|
||||||
|
+#ifdef CONFIG_USER_NS
|
||||||
|
+ {
|
||||||
|
+ .procname = "unprivileged_userns_clone",
|
||||||
|
+ .data = &unprivileged_userns_clone,
|
||||||
|
+ .maxlen = sizeof(int),
|
||||||
|
+ .mode = 0644,
|
||||||
|
+ .proc_handler = proc_dointvec,
|
||||||
|
+ },
|
||||||
|
+#endif
|
||||||
|
#ifdef CONFIG_PROC_SYSCTL
|
||||||
|
{
|
||||||
|
.procname = "tainted",
|
||||||
|
diff --git a/kernel/user_namespace.c b/kernel/user_namespace.c
|
||||||
|
index 54211dbd516c57..16ca0c1516298d 100644
|
||||||
|
--- a/kernel/user_namespace.c
|
||||||
|
+++ b/kernel/user_namespace.c
|
||||||
|
@@ -22,6 +22,13 @@
|
||||||
|
#include <linux/bsearch.h>
|
||||||
|
#include <linux/sort.h>
|
||||||
|
|
||||||
|
+/* sysctl */
|
||||||
|
+#ifdef CONFIG_USER_NS_UNPRIVILEGED
|
||||||
|
+int unprivileged_userns_clone = 1;
|
||||||
|
+#else
|
||||||
|
+int unprivileged_userns_clone;
|
||||||
|
+#endif
|
||||||
|
+
|
||||||
|
static struct kmem_cache *user_ns_cachep __read_mostly;
|
||||||
|
static DEFINE_MUTEX(userns_state_mutex);
|
||||||
|
|
@@ -0,0 +1,244 @@
|
|||||||
|
From 5ae86c8436b83762bc6cf46bea1da6ace2d3f50e Mon Sep 17 00:00:00 2001
|
||||||
|
From: Paul Gofman <pgofman@codeweavers.com>
|
||||||
|
Date: Wed, 6 May 2020 14:37:44 +0300
|
||||||
|
Subject: [PATCH 1/2] mm: Support soft dirty flag reset for VA range.
|
||||||
|
|
||||||
|
---
|
||||||
|
fs/proc/task_mmu.c | 129 ++++++++++++++++++++++++++++++++++++---------
|
||||||
|
1 file changed, 103 insertions(+), 26 deletions(-)
|
||||||
|
|
||||||
|
diff --git a/fs/proc/task_mmu.c b/fs/proc/task_mmu.c
|
||||||
|
index 3cec6fbef725..7c7865028f10 100644
|
||||||
|
--- a/fs/proc/task_mmu.c
|
||||||
|
+++ b/fs/proc/task_mmu.c
|
||||||
|
@@ -1032,6 +1032,8 @@ enum clear_refs_types {
|
||||||
|
|
||||||
|
struct clear_refs_private {
|
||||||
|
enum clear_refs_types type;
|
||||||
|
+ unsigned long start, end;
|
||||||
|
+ bool clear_range;
|
||||||
|
};
|
||||||
|
|
||||||
|
#ifdef CONFIG_MEM_SOFT_DIRTY
|
||||||
|
@@ -1125,6 +1127,8 @@ static int clear_refs_pte_range(pmd_t *pmd, unsigned long addr,
|
||||||
|
spinlock_t *ptl;
|
||||||
|
struct page *page;
|
||||||
|
|
||||||
|
+ BUG_ON(addr < cp->start || end > cp->end);
|
||||||
|
+
|
||||||
|
ptl = pmd_trans_huge_lock(pmd, vma);
|
||||||
|
if (ptl) {
|
||||||
|
if (cp->type == CLEAR_REFS_SOFT_DIRTY) {
|
||||||
|
@@ -1181,9 +1185,11 @@ static int clear_refs_test_walk(unsigned long start, unsigned long end,
|
||||||
|
struct clear_refs_private *cp = walk->private;
|
||||||
|
struct vm_area_struct *vma = walk->vma;
|
||||||
|
|
||||||
|
- if (vma->vm_flags & VM_PFNMAP)
|
||||||
|
+ if (!cp->clear_range && (vma->vm_flags & VM_PFNMAP))
|
||||||
|
return 1;
|
||||||
|
|
||||||
|
+ BUG_ON(start < cp->start || end > cp->end);
|
||||||
|
+
|
||||||
|
/*
|
||||||
|
* Writing 1 to /proc/pid/clear_refs affects all pages.
|
||||||
|
* Writing 2 to /proc/pid/clear_refs only affects anonymous pages.
|
||||||
|
@@ -1206,10 +1212,12 @@ static ssize_t clear_refs_write(struct file *file, const char __user *buf,
|
||||||
|
size_t count, loff_t *ppos)
|
||||||
|
{
|
||||||
|
struct task_struct *task;
|
||||||
|
- char buffer[PROC_NUMBUF];
|
||||||
|
+ char buffer[18];
|
||||||
|
struct mm_struct *mm;
|
||||||
|
struct vm_area_struct *vma;
|
||||||
|
enum clear_refs_types type;
|
||||||
|
+ unsigned long start, end;
|
||||||
|
+ bool clear_range;
|
||||||
|
int itype;
|
||||||
|
int rv;
|
||||||
|
|
||||||
|
@@ -1218,12 +1226,34 @@ static ssize_t clear_refs_write(struct file *file, const char __user *buf,
|
||||||
|
count = sizeof(buffer) - 1;
|
||||||
|
if (copy_from_user(buffer, buf, count))
|
||||||
|
return -EFAULT;
|
||||||
|
- rv = kstrtoint(strstrip(buffer), 10, &itype);
|
||||||
|
- if (rv < 0)
|
||||||
|
- return rv;
|
||||||
|
- type = (enum clear_refs_types)itype;
|
||||||
|
- if (type < CLEAR_REFS_ALL || type >= CLEAR_REFS_LAST)
|
||||||
|
- return -EINVAL;
|
||||||
|
+
|
||||||
|
+ if (buffer[0] == '6')
|
||||||
|
+ {
|
||||||
|
+ static int once;
|
||||||
|
+
|
||||||
|
+ if (!once++)
|
||||||
|
+ printk(KERN_DEBUG "task_mmu: Using POC clear refs range implementation.\n");
|
||||||
|
+
|
||||||
|
+ if (count != 17)
|
||||||
|
+ return -EINVAL;
|
||||||
|
+
|
||||||
|
+ type = CLEAR_REFS_SOFT_DIRTY;
|
||||||
|
+ start = *(unsigned long *)(buffer + 1);
|
||||||
|
+ end = *(unsigned long *)(buffer + 1 + 8);
|
||||||
|
+ }
|
||||||
|
+ else
|
||||||
|
+ {
|
||||||
|
+ rv = kstrtoint(strstrip(buffer), 10, &itype);
|
||||||
|
+ if (rv < 0)
|
||||||
|
+ return rv;
|
||||||
|
+ type = (enum clear_refs_types)itype;
|
||||||
|
+
|
||||||
|
+ if (type < CLEAR_REFS_ALL || type >= CLEAR_REFS_LAST)
|
||||||
|
+ return -EINVAL;
|
||||||
|
+
|
||||||
|
+ start = 0;
|
||||||
|
+ end = -1UL;
|
||||||
|
+ }
|
||||||
|
|
||||||
|
task = get_proc_task(file_inode(file));
|
||||||
|
if (!task)
|
||||||
|
@@ -1235,40 +1265,86 @@ static ssize_t clear_refs_write(struct file *file, const char __user *buf,
|
||||||
|
.type = type,
|
||||||
|
};
|
||||||
|
|
||||||
|
- if (mmap_write_lock_killable(mm)) {
|
||||||
|
- count = -EINTR;
|
||||||
|
- goto out_mm;
|
||||||
|
+ if (start || end != -1UL)
|
||||||
|
+ {
|
||||||
|
+ start = min(start, -1) & PAGE_MASK;
|
||||||
|
+ end = min(end, -1) & PAGE_MASK;
|
||||||
|
+
|
||||||
|
+ if (start >= end)
|
||||||
|
+ {
|
||||||
|
+ count = -EINVAL;
|
||||||
|
+ goto out_mm;
|
||||||
|
+ }
|
||||||
|
+ clear_range = true;
|
||||||
|
}
|
||||||
|
+ else
|
||||||
|
+ {
|
||||||
|
+ clear_range = false;
|
||||||
|
+ }
|
||||||
|
+
|
||||||
|
+ cp.start = start;
|
||||||
|
+ cp.end = end;
|
||||||
|
+ cp.clear_range = clear_range;
|
||||||
|
+
|
||||||
|
if (type == CLEAR_REFS_MM_HIWATER_RSS) {
|
||||||
|
+ if (mmap_write_lock_killable(mm)) {
|
||||||
|
+ count = -EINTR;
|
||||||
|
+ goto out_mm;
|
||||||
|
+ }
|
||||||
|
+
|
||||||
|
/*
|
||||||
|
* Writing 5 to /proc/pid/clear_refs resets the peak
|
||||||
|
* resident set size to this mm's current rss value.
|
||||||
|
*/
|
||||||
|
reset_mm_hiwater_rss(mm);
|
||||||
|
- goto out_unlock;
|
||||||
|
+ mmap_write_unlock(mm);
|
||||||
|
+ goto out_mm;
|
||||||
|
}
|
||||||
|
|
||||||
|
if (type == CLEAR_REFS_SOFT_DIRTY) {
|
||||||
|
- for_each_vma(vmi, vma) {
|
||||||
|
- if (!(vma->vm_flags & VM_SOFTDIRTY))
|
||||||
|
- continue;
|
||||||
|
- vm_flags_clear(vma, VM_SOFTDIRTY);
|
||||||
|
- vma_set_page_prot(vma);
|
||||||
|
+ if (mmap_read_lock_killable(mm)) {
|
||||||
|
+ count = -EINTR;
|
||||||
|
+ goto out_mm;
|
||||||
|
}
|
||||||
|
-
|
||||||
|
+ if (!clear_range)
|
||||||
|
+ for_each_vma(vmi, vma) {
|
||||||
|
+ if (!(vma->vm_flags & VM_SOFTDIRTY))
|
||||||
|
+ continue;
|
||||||
|
+ mmap_read_unlock(mm);
|
||||||
|
+ if (mmap_write_lock_killable(mm)) {
|
||||||
|
+ count = -EINTR;
|
||||||
|
+ goto out_mm;
|
||||||
|
+ }
|
||||||
|
+ for_each_vma(vmi, vma) {
|
||||||
|
+ vm_flags_clear(vma, VM_SOFTDIRTY);
|
||||||
|
+ vma_set_page_prot(vma);
|
||||||
|
+ }
|
||||||
|
+ mmap_write_downgrade(mm);
|
||||||
|
+ break;
|
||||||
|
+ }
|
||||||
|
inc_tlb_flush_pending(mm);
|
||||||
|
mmu_notifier_range_init(&range, MMU_NOTIFY_SOFT_DIRTY,
|
||||||
|
- 0, mm, 0, -1UL);
|
||||||
|
+ 0, mm, start, end);
|
||||||
|
mmu_notifier_invalidate_range_start(&range);
|
||||||
|
}
|
||||||
|
- walk_page_range(mm, 0, -1, &clear_refs_walk_ops, &cp);
|
||||||
|
+ else
|
||||||
|
+ {
|
||||||
|
+ if (mmap_write_lock_killable(mm)) {
|
||||||
|
+ count = -EINTR;
|
||||||
|
+ goto out_mm;
|
||||||
|
+ }
|
||||||
|
+ }
|
||||||
|
+ walk_page_range(mm, start, end == -1UL ? -1 : end, &clear_refs_walk_ops, &cp);
|
||||||
|
if (type == CLEAR_REFS_SOFT_DIRTY) {
|
||||||
|
mmu_notifier_invalidate_range_end(&range);
|
||||||
|
flush_tlb_mm(mm);
|
||||||
|
dec_tlb_flush_pending(mm);
|
||||||
|
+ mmap_read_unlock(mm);
|
||||||
|
+ }
|
||||||
|
+ else
|
||||||
|
+ {
|
||||||
|
+ mmap_write_unlock(mm);
|
||||||
|
}
|
||||||
|
-out_unlock:
|
||||||
|
- mmap_write_unlock(mm);
|
||||||
|
out_mm:
|
||||||
|
mmput(mm);
|
||||||
|
}
|
||||||
|
@@ -1301,6 +1377,7 @@ struct pagemapread {
|
||||||
|
#define PM_PFRAME_MASK GENMASK_ULL(PM_PFRAME_BITS - 1, 0)
|
||||||
|
#define PM_SOFT_DIRTY BIT_ULL(55)
|
||||||
|
#define PM_MMAP_EXCLUSIVE BIT_ULL(56)
|
||||||
|
+#define PM_SOFT_DIRTY_PAGE BIT_ULL(57)
|
||||||
|
#define PM_UFFD_WP BIT_ULL(57)
|
||||||
|
#define PM_FILE BIT_ULL(61)
|
||||||
|
#define PM_SWAP BIT_ULL(62)
|
||||||
|
@@ -1373,13 +1450,13 @@ static pagemap_entry_t pte_to_pagemap_entry(struct pagemapread *pm,
|
||||||
|
flags |= PM_PRESENT;
|
||||||
|
page = vm_normal_page(vma, addr, pte);
|
||||||
|
if (pte_soft_dirty(pte))
|
||||||
|
- flags |= PM_SOFT_DIRTY;
|
||||||
|
+ flags |= PM_SOFT_DIRTY | PM_SOFT_DIRTY_PAGE;
|
||||||
|
if (pte_uffd_wp(pte))
|
||||||
|
flags |= PM_UFFD_WP;
|
||||||
|
} else if (is_swap_pte(pte)) {
|
||||||
|
swp_entry_t entry;
|
||||||
|
if (pte_swp_soft_dirty(pte))
|
||||||
|
- flags |= PM_SOFT_DIRTY;
|
||||||
|
+ flags |= PM_SOFT_DIRTY | PM_SOFT_DIRTY_PAGE;
|
||||||
|
if (pte_swp_uffd_wp(pte))
|
||||||
|
flags |= PM_UFFD_WP;
|
||||||
|
entry = pte_to_swp_entry(pte);
|
||||||
|
@@ -1500,7 +1500,7 @@
|
||||||
|
|
||||||
|
flags |= PM_PRESENT;
|
||||||
|
if (pmd_soft_dirty(pmd))
|
||||||
|
- flags |= PM_SOFT_DIRTY;
|
||||||
|
+ flags |= PM_SOFT_DIRTY | PM_SOFT_DIRTY_PAGE;
|
||||||
|
if (pmd_uffd_wp(pmd))
|
||||||
|
flags |= PM_UFFD_WP;
|
||||||
|
if (pm->show_pfn)
|
||||||
|
@@ -1442,7 +1519,7 @@ static int pagemap_pmd_range(pmd_t *pmdp, unsigned long addr, unsigned long end,
|
||||||
|
}
|
||||||
|
flags |= PM_SWAP;
|
||||||
|
if (pmd_swp_soft_dirty(pmd))
|
||||||
|
- flags |= PM_SOFT_DIRTY;
|
||||||
|
+ flags |= PM_SOFT_DIRTY | PM_SOFT_DIRTY_PAGE;
|
||||||
|
if (pmd_swp_uffd_wp(pmd))
|
||||||
|
flags |= PM_UFFD_WP;
|
||||||
|
VM_BUG_ON(!is_pmd_migration_entry(pmd));
|
||||||
|
--
|
||||||
|
2.30.2
|
620
linux-tkg-patches/6.3/0002-clear-patches.patch
Normal file
620
linux-tkg-patches/6.3/0002-clear-patches.patch
Normal file
@@ -0,0 +1,620 @@
|
|||||||
|
From 0000000000000000000000000000000000000000 Mon Sep 17 00:00:00 2001
|
||||||
|
From: Arjan van de Ven <arjan@linux.intel.com>
|
||||||
|
Date: Mon, 14 Mar 2016 11:10:58 -0600
|
||||||
|
Subject: [PATCH] pci pme wakeups
|
||||||
|
|
||||||
|
Reduce wakeups for PME checks, which are a workaround for miswired
|
||||||
|
boards (sadly, too many of them) in laptops.
|
||||||
|
---
|
||||||
|
drivers/pci/pci.c | 2 +-
|
||||||
|
1 file changed, 1 insertion(+), 1 deletion(-)
|
||||||
|
|
||||||
|
diff --git a/drivers/pci/pci.c b/drivers/pci/pci.c
|
||||||
|
index c9338f9..6974fbf 100644
|
||||||
|
--- a/drivers/pci/pci.c
|
||||||
|
+++ b/drivers/pci/pci.c
|
||||||
|
@@ -62,7 +62,7 @@ struct pci_pme_device {
|
||||||
|
struct pci_dev *dev;
|
||||||
|
};
|
||||||
|
|
||||||
|
-#define PME_TIMEOUT 1000 /* How long between PME checks */
|
||||||
|
+#define PME_TIMEOUT 4000 /* How long between PME checks */
|
||||||
|
|
||||||
|
static void pci_dev_d3_sleep(struct pci_dev *dev)
|
||||||
|
{
|
||||||
|
--
|
||||||
|
https://clearlinux.org
|
||||||
|
|
||||||
|
From 0000000000000000000000000000000000000000 Mon Sep 17 00:00:00 2001
|
||||||
|
From: Arjan van de Ven <arjan@linux.intel.com>
|
||||||
|
Date: Sat, 19 Mar 2016 21:32:19 -0400
|
||||||
|
Subject: [PATCH] intel_idle: tweak cpuidle cstates
|
||||||
|
|
||||||
|
Increase target_residency in cpuidle cstate
|
||||||
|
|
||||||
|
Tune intel_idle to be a bit less agressive;
|
||||||
|
Clear linux is cleaner in hygiene (wakupes) than the average linux,
|
||||||
|
so we can afford changing these in a way that increases
|
||||||
|
performance while keeping power efficiency
|
||||||
|
---
|
||||||
|
drivers/idle/intel_idle.c | 44 +++++++++++++++++++--------------------
|
||||||
|
1 file changed, 22 insertions(+), 22 deletions(-)
|
||||||
|
|
||||||
|
diff --git a/drivers/idle/intel_idle.c b/drivers/idle/intel_idle.c
|
||||||
|
index f449584..c994d24 100644
|
||||||
|
--- a/drivers/idle/intel_idle.c
|
||||||
|
+++ b/drivers/idle/intel_idle.c
|
||||||
|
@@ -531,7 +531,7 @@ static struct cpuidle_state hsw_cstates[] __initdata = {
|
||||||
|
.desc = "MWAIT 0x01",
|
||||||
|
.flags = MWAIT2flg(0x01) | CPUIDLE_FLAG_ALWAYS_ENABLE,
|
||||||
|
.exit_latency = 10,
|
||||||
|
- .target_residency = 20,
|
||||||
|
+ .target_residency = 120,
|
||||||
|
.enter = &intel_idle,
|
||||||
|
.enter_s2idle = intel_idle_s2idle, },
|
||||||
|
{
|
||||||
|
@@ -539,7 +539,7 @@ static struct cpuidle_state hsw_cstates[] __initdata = {
|
||||||
|
.desc = "MWAIT 0x10",
|
||||||
|
.flags = MWAIT2flg(0x10) | CPUIDLE_FLAG_TLB_FLUSHED,
|
||||||
|
.exit_latency = 33,
|
||||||
|
- .target_residency = 100,
|
||||||
|
+ .target_residency = 900,
|
||||||
|
.enter = &intel_idle,
|
||||||
|
.enter_s2idle = intel_idle_s2idle, },
|
||||||
|
{
|
||||||
|
@@ -547,7 +547,7 @@ static struct cpuidle_state hsw_cstates[] __initdata = {
|
||||||
|
.desc = "MWAIT 0x20",
|
||||||
|
.flags = MWAIT2flg(0x20) | CPUIDLE_FLAG_TLB_FLUSHED,
|
||||||
|
.exit_latency = 133,
|
||||||
|
- .target_residency = 400,
|
||||||
|
+ .target_residency = 1000,
|
||||||
|
.enter = &intel_idle,
|
||||||
|
.enter_s2idle = intel_idle_s2idle, },
|
||||||
|
{
|
||||||
|
@@ -555,7 +555,7 @@ static struct cpuidle_state hsw_cstates[] __initdata = {
|
||||||
|
.desc = "MWAIT 0x32",
|
||||||
|
.flags = MWAIT2flg(0x32) | CPUIDLE_FLAG_TLB_FLUSHED,
|
||||||
|
.exit_latency = 166,
|
||||||
|
- .target_residency = 500,
|
||||||
|
+ .target_residency = 1500,
|
||||||
|
.enter = &intel_idle,
|
||||||
|
.enter_s2idle = intel_idle_s2idle, },
|
||||||
|
{
|
||||||
|
@@ -563,7 +563,7 @@ static struct cpuidle_state hsw_cstates[] __initdata = {
|
||||||
|
.desc = "MWAIT 0x40",
|
||||||
|
.flags = MWAIT2flg(0x40) | CPUIDLE_FLAG_TLB_FLUSHED,
|
||||||
|
.exit_latency = 300,
|
||||||
|
- .target_residency = 900,
|
||||||
|
+ .target_residency = 2000,
|
||||||
|
.enter = &intel_idle,
|
||||||
|
.enter_s2idle = intel_idle_s2idle, },
|
||||||
|
{
|
||||||
|
@@ -571,7 +571,7 @@ static struct cpuidle_state hsw_cstates[] __initdata = {
|
||||||
|
.desc = "MWAIT 0x50",
|
||||||
|
.flags = MWAIT2flg(0x50) | CPUIDLE_FLAG_TLB_FLUSHED,
|
||||||
|
.exit_latency = 600,
|
||||||
|
- .target_residency = 1800,
|
||||||
|
+ .target_residency = 5000,
|
||||||
|
.enter = &intel_idle,
|
||||||
|
.enter_s2idle = intel_idle_s2idle, },
|
||||||
|
{
|
||||||
|
@@ -579,7 +579,7 @@ static struct cpuidle_state hsw_cstates[] __initdata = {
|
||||||
|
.desc = "MWAIT 0x60",
|
||||||
|
.flags = MWAIT2flg(0x60) | CPUIDLE_FLAG_TLB_FLUSHED,
|
||||||
|
.exit_latency = 2600,
|
||||||
|
- .target_residency = 7700,
|
||||||
|
+ .target_residency = 9000,
|
||||||
|
.enter = &intel_idle,
|
||||||
|
.enter_s2idle = intel_idle_s2idle, },
|
||||||
|
{
|
||||||
|
@@ -599,7 +599,7 @@ static struct cpuidle_state bdw_cstates[] __initdata = {
|
||||||
|
.desc = "MWAIT 0x01",
|
||||||
|
.flags = MWAIT2flg(0x01) | CPUIDLE_FLAG_ALWAYS_ENABLE,
|
||||||
|
.exit_latency = 10,
|
||||||
|
- .target_residency = 20,
|
||||||
|
+ .target_residency = 120,
|
||||||
|
.enter = &intel_idle,
|
||||||
|
.enter_s2idle = intel_idle_s2idle, },
|
||||||
|
{
|
||||||
|
@@ -607,7 +607,7 @@ static struct cpuidle_state bdw_cstates[] __initdata = {
|
||||||
|
.desc = "MWAIT 0x10",
|
||||||
|
.flags = MWAIT2flg(0x10) | CPUIDLE_FLAG_TLB_FLUSHED,
|
||||||
|
.exit_latency = 40,
|
||||||
|
- .target_residency = 100,
|
||||||
|
+ .target_residency = 1000,
|
||||||
|
.enter = &intel_idle,
|
||||||
|
.enter_s2idle = intel_idle_s2idle, },
|
||||||
|
{
|
||||||
|
@@ -615,7 +615,7 @@ static struct cpuidle_state bdw_cstates[] __initdata = {
|
||||||
|
.desc = "MWAIT 0x20",
|
||||||
|
.flags = MWAIT2flg(0x20) | CPUIDLE_FLAG_TLB_FLUSHED,
|
||||||
|
.exit_latency = 133,
|
||||||
|
- .target_residency = 400,
|
||||||
|
+ .target_residency = 1000,
|
||||||
|
.enter = &intel_idle,
|
||||||
|
.enter_s2idle = intel_idle_s2idle, },
|
||||||
|
{
|
||||||
|
@@ -623,7 +623,7 @@ static struct cpuidle_state bdw_cstates[] __initdata = {
|
||||||
|
.desc = "MWAIT 0x32",
|
||||||
|
.flags = MWAIT2flg(0x32) | CPUIDLE_FLAG_TLB_FLUSHED,
|
||||||
|
.exit_latency = 166,
|
||||||
|
- .target_residency = 500,
|
||||||
|
+ .target_residency = 2000,
|
||||||
|
.enter = &intel_idle,
|
||||||
|
.enter_s2idle = intel_idle_s2idle, },
|
||||||
|
{
|
||||||
|
@@ -631,7 +631,7 @@ static struct cpuidle_state bdw_cstates[] __initdata = {
|
||||||
|
.desc = "MWAIT 0x40",
|
||||||
|
.flags = MWAIT2flg(0x40) | CPUIDLE_FLAG_TLB_FLUSHED,
|
||||||
|
.exit_latency = 300,
|
||||||
|
- .target_residency = 900,
|
||||||
|
+ .target_residency = 4000,
|
||||||
|
.enter = &intel_idle,
|
||||||
|
.enter_s2idle = intel_idle_s2idle, },
|
||||||
|
{
|
||||||
|
@@ -639,7 +639,7 @@ static struct cpuidle_state bdw_cstates[] __initdata = {
|
||||||
|
.desc = "MWAIT 0x50",
|
||||||
|
.flags = MWAIT2flg(0x50) | CPUIDLE_FLAG_TLB_FLUSHED,
|
||||||
|
.exit_latency = 600,
|
||||||
|
- .target_residency = 1800,
|
||||||
|
+ .target_residency = 7000,
|
||||||
|
.enter = &intel_idle,
|
||||||
|
.enter_s2idle = intel_idle_s2idle, },
|
||||||
|
{
|
||||||
|
@@ -647,7 +647,7 @@ static struct cpuidle_state bdw_cstates[] __initdata = {
|
||||||
|
.desc = "MWAIT 0x60",
|
||||||
|
.flags = MWAIT2flg(0x60) | CPUIDLE_FLAG_TLB_FLUSHED,
|
||||||
|
.exit_latency = 2600,
|
||||||
|
- .target_residency = 7700,
|
||||||
|
+ .target_residency = 9000,
|
||||||
|
.enter = &intel_idle,
|
||||||
|
.enter_s2idle = intel_idle_s2idle, },
|
||||||
|
{
|
||||||
|
@@ -668,7 +668,7 @@ static struct cpuidle_state skl_cstates[] __initdata = {
|
||||||
|
.desc = "MWAIT 0x01",
|
||||||
|
.flags = MWAIT2flg(0x01) | CPUIDLE_FLAG_ALWAYS_ENABLE,
|
||||||
|
.exit_latency = 10,
|
||||||
|
- .target_residency = 20,
|
||||||
|
+ .target_residency = 120,
|
||||||
|
.enter = &intel_idle,
|
||||||
|
.enter_s2idle = intel_idle_s2idle, },
|
||||||
|
{
|
||||||
|
@@ -676,7 +676,7 @@ static struct cpuidle_state skl_cstates[] __initdata = {
|
||||||
|
.desc = "MWAIT 0x10",
|
||||||
|
.flags = MWAIT2flg(0x10) | CPUIDLE_FLAG_TLB_FLUSHED,
|
||||||
|
.exit_latency = 70,
|
||||||
|
- .target_residency = 100,
|
||||||
|
+ .target_residency = 1000,
|
||||||
|
.enter = &intel_idle,
|
||||||
|
.enter_s2idle = intel_idle_s2idle, },
|
||||||
|
{
|
||||||
|
@@ -684,7 +684,7 @@ static struct cpuidle_state skl_cstates[] __initdata = {
|
||||||
|
.desc = "MWAIT 0x20",
|
||||||
|
.flags = MWAIT2flg(0x20) | CPUIDLE_FLAG_TLB_FLUSHED,
|
||||||
|
.exit_latency = 85,
|
||||||
|
- .target_residency = 200,
|
||||||
|
+ .target_residency = 600,
|
||||||
|
.enter = &intel_idle,
|
||||||
|
.enter_s2idle = intel_idle_s2idle, },
|
||||||
|
{
|
||||||
|
@@ -692,7 +692,7 @@ static struct cpuidle_state skl_cstates[] __initdata = {
|
||||||
|
.desc = "MWAIT 0x33",
|
||||||
|
.flags = MWAIT2flg(0x33) | CPUIDLE_FLAG_TLB_FLUSHED,
|
||||||
|
.exit_latency = 124,
|
||||||
|
- .target_residency = 800,
|
||||||
|
+ .target_residency = 3000,
|
||||||
|
.enter = &intel_idle,
|
||||||
|
.enter_s2idle = intel_idle_s2idle, },
|
||||||
|
{
|
||||||
|
@@ -700,7 +700,7 @@ static struct cpuidle_state skl_cstates[] __initdata = {
|
||||||
|
.desc = "MWAIT 0x40",
|
||||||
|
.flags = MWAIT2flg(0x40) | CPUIDLE_FLAG_TLB_FLUSHED,
|
||||||
|
.exit_latency = 200,
|
||||||
|
- .target_residency = 800,
|
||||||
|
+ .target_residency = 3200,
|
||||||
|
.enter = &intel_idle,
|
||||||
|
.enter_s2idle = intel_idle_s2idle, },
|
||||||
|
{
|
||||||
|
@@ -708,7 +708,7 @@ static struct cpuidle_state skl_cstates[] __initdata = {
|
||||||
|
.desc = "MWAIT 0x50",
|
||||||
|
.flags = MWAIT2flg(0x50) | CPUIDLE_FLAG_TLB_FLUSHED,
|
||||||
|
.exit_latency = 480,
|
||||||
|
- .target_residency = 5000,
|
||||||
|
+ .target_residency = 9000,
|
||||||
|
.enter = &intel_idle,
|
||||||
|
.enter_s2idle = intel_idle_s2idle, },
|
||||||
|
{
|
||||||
|
@@ -716,7 +716,7 @@ static struct cpuidle_state skl_cstates[] __initdata = {
|
||||||
|
.desc = "MWAIT 0x60",
|
||||||
|
.flags = MWAIT2flg(0x60) | CPUIDLE_FLAG_TLB_FLUSHED,
|
||||||
|
.exit_latency = 890,
|
||||||
|
- .target_residency = 5000,
|
||||||
|
+ .target_residency = 9000,
|
||||||
|
.enter = &intel_idle,
|
||||||
|
.enter_s2idle = intel_idle_s2idle, },
|
||||||
|
{
|
||||||
|
@@ -737,7 +737,7 @@ static struct cpuidle_state skx_cstates[] __initdata = {
|
||||||
|
.desc = "MWAIT 0x01",
|
||||||
|
.flags = MWAIT2flg(0x01) | CPUIDLE_FLAG_ALWAYS_ENABLE,
|
||||||
|
.exit_latency = 10,
|
||||||
|
- .target_residency = 20,
|
||||||
|
+ .target_residency = 300,
|
||||||
|
.enter = &intel_idle,
|
||||||
|
.enter_s2idle = intel_idle_s2idle, },
|
||||||
|
{
|
||||||
|
--
|
||||||
|
https://clearlinux.org
|
||||||
|
|
||||||
|
From 0000000000000000000000000000000000000000 Mon Sep 17 00:00:00 2001
|
||||||
|
From: Arjan van de Ven <arjan@linux.intel.com>
|
||||||
|
Date: Fri, 6 Jan 2017 15:34:09 +0000
|
||||||
|
Subject: [PATCH] ipv4/tcp: allow the memory tuning for tcp to go a little
|
||||||
|
bigger than default
|
||||||
|
|
||||||
|
---
|
||||||
|
net/ipv4/tcp.c | 4 ++--
|
||||||
|
1 file changed, 2 insertions(+), 2 deletions(-)
|
||||||
|
|
||||||
|
diff --git a/net/ipv4/tcp.c b/net/ipv4/tcp.c
|
||||||
|
index 30c1142..4345075 100644
|
||||||
|
--- a/net/ipv4/tcp.c
|
||||||
|
+++ b/net/ipv4/tcp.c
|
||||||
|
@@ -4201,8 +4201,8 @@ void __init tcp_init(void)
|
||||||
|
tcp_init_mem();
|
||||||
|
/* Set per-socket limits to no more than 1/128 the pressure threshold */
|
||||||
|
limit = nr_free_buffer_pages() << (PAGE_SHIFT - 7);
|
||||||
|
- max_wshare = min(4UL*1024*1024, limit);
|
||||||
|
- max_rshare = min(6UL*1024*1024, limit);
|
||||||
|
+ max_wshare = min(16UL*1024*1024, limit);
|
||||||
|
+ max_rshare = min(16UL*1024*1024, limit);
|
||||||
|
|
||||||
|
init_net.ipv4.sysctl_tcp_wmem[0] = SK_MEM_QUANTUM;
|
||||||
|
init_net.ipv4.sysctl_tcp_wmem[1] = 16*1024;
|
||||||
|
--
|
||||||
|
https://clearlinux.org
|
||||||
|
|
||||||
|
From 0000000000000000000000000000000000000000 Mon Sep 17 00:00:00 2001
|
||||||
|
From: Arjan van de Ven <arjan@linux.intel.com>
|
||||||
|
Date: Sun, 18 Feb 2018 23:35:41 +0000
|
||||||
|
Subject: [PATCH] locking: rwsem: spin faster
|
||||||
|
|
||||||
|
tweak rwsem owner spinning a bit
|
||||||
|
---
|
||||||
|
kernel/locking/rwsem.c | 4 +++-
|
||||||
|
1 file changed, 3 insertions(+), 1 deletion(-)
|
||||||
|
|
||||||
|
diff --git a/kernel/locking/rwsem.c b/kernel/locking/rwsem.c
|
||||||
|
index f11b9bd..1bbfcc1 100644
|
||||||
|
--- a/kernel/locking/rwsem.c
|
||||||
|
+++ b/kernel/locking/rwsem.c
|
||||||
|
@@ -717,6 +717,7 @@ rwsem_spin_on_owner(struct rw_semaphore *sem, unsigned long nonspinnable)
|
||||||
|
struct task_struct *new, *owner;
|
||||||
|
unsigned long flags, new_flags;
|
||||||
|
enum owner_state state;
|
||||||
|
+ int i = 0;
|
||||||
|
|
||||||
|
owner = rwsem_owner_flags(sem, &flags);
|
||||||
|
state = rwsem_owner_state(owner, flags, nonspinnable);
|
||||||
|
@@ -750,7 +751,8 @@ rwsem_spin_on_owner(struct rw_semaphore *sem, unsigned long nonspinnable)
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
|
||||||
|
- cpu_relax();
|
||||||
|
+ if (i++ > 1000)
|
||||||
|
+ cpu_relax();
|
||||||
|
}
|
||||||
|
rcu_read_unlock();
|
||||||
|
|
||||||
|
--
|
||||||
|
https://clearlinux.org
|
||||||
|
|
||||||
|
From 0000000000000000000000000000000000000000 Mon Sep 17 00:00:00 2001
|
||||||
|
From: Arjan van de Ven <arjan@linux.intel.com>
|
||||||
|
Date: Thu, 2 Jun 2016 23:36:32 -0500
|
||||||
|
Subject: [PATCH] initialize ata before graphics
|
||||||
|
|
||||||
|
ATA init is the long pole in the boot process, and its asynchronous.
|
||||||
|
move the graphics init after it so that ata and graphics initialize
|
||||||
|
in parallel
|
||||||
|
---
|
||||||
|
drivers/Makefile | 15 ++++++++-------
|
||||||
|
1 file changed, 8 insertions(+), 7 deletions(-)
|
||||||
|
|
||||||
|
diff --git a/drivers/Makefile b/drivers/Makefile
|
||||||
|
index c0cd1b9..af1e2fb 100644
|
||||||
|
--- a/drivers/Makefile
|
||||||
|
+++ b/drivers/Makefile
|
||||||
|
@@ -59,15 +59,8 @@ obj-y += char/
|
||||||
|
# iommu/ comes before gpu as gpu are using iommu controllers
|
||||||
|
obj-y += iommu/
|
||||||
|
|
||||||
|
-# gpu/ comes after char for AGP vs DRM startup and after iommu
|
||||||
|
-obj-y += gpu/
|
||||||
|
-
|
||||||
|
obj-$(CONFIG_CONNECTOR) += connector/
|
||||||
|
|
||||||
|
-# i810fb and intelfb depend on char/agp/
|
||||||
|
-obj-$(CONFIG_FB_I810) += video/fbdev/i810/
|
||||||
|
-obj-$(CONFIG_FB_INTEL) += video/fbdev/intelfb/
|
||||||
|
-
|
||||||
|
obj-$(CONFIG_PARPORT) += parport/
|
||||||
|
obj-$(CONFIG_NVM) += lightnvm/
|
||||||
|
obj-y += base/ block/ misc/ mfd/ nfc/
|
||||||
|
@@ -80,6 +73,14 @@ obj-$(CONFIG_IDE) += ide/
|
||||||
|
obj-y += scsi/
|
||||||
|
obj-y += nvme/
|
||||||
|
obj-$(CONFIG_ATA) += ata/
|
||||||
|
+
|
||||||
|
+# gpu/ comes after char for AGP vs DRM startup and after iommu
|
||||||
|
+obj-y += gpu/
|
||||||
|
+
|
||||||
|
+# i810fb and intelfb depend on char/agp/
|
||||||
|
+obj-$(CONFIG_FB_I810) += video/fbdev/i810/
|
||||||
|
+obj-$(CONFIG_FB_INTEL) += video/fbdev/intelfb/
|
||||||
|
+
|
||||||
|
obj-$(CONFIG_TARGET_CORE) += target/
|
||||||
|
obj-$(CONFIG_MTD) += mtd/
|
||||||
|
obj-$(CONFIG_SPI) += spi/
|
||||||
|
--
|
||||||
|
https://clearlinux.org
|
||||||
|
|
||||||
|
From 676c2dc63592f52b716515573a3a825582a371e9 Mon Sep 17 00:00:00 2001
|
||||||
|
From: Arjan van de Ven <arjan@linux.intel.com>
|
||||||
|
Date: Sat, 8 Dec 2018 18:21:32 +0000
|
||||||
|
Subject: [PATCH 1/9] x86/vdso: Use lfence instead of rep and nop
|
||||||
|
|
||||||
|
Signed-off-by: Alexandre Frade <kernel@xanmod.org>
|
||||||
|
---
|
||||||
|
arch/x86/include/asm/vdso/processor.h | 2 +-
|
||||||
|
1 file changed, 1 insertion(+), 1 deletion(-)
|
||||||
|
|
||||||
|
diff --git a/arch/x86/include/asm/vdso/processor.h b/arch/x86/include/asm/vdso/processor.h
|
||||||
|
index 57b1a7034c64..e2c45674f989 100644
|
||||||
|
--- a/arch/x86/include/asm/vdso/processor.h
|
||||||
|
+++ b/arch/x86/include/asm/vdso/processor.h
|
||||||
|
@@ -10,7 +10,7 @@
|
||||||
|
/* REP NOP (PAUSE) is a good thing to insert into busy-wait loops. */
|
||||||
|
static __always_inline void rep_nop(void)
|
||||||
|
{
|
||||||
|
- asm volatile("rep; nop" ::: "memory");
|
||||||
|
+ asm volatile("lfence" ::: "memory");
|
||||||
|
}
|
||||||
|
|
||||||
|
static __always_inline void cpu_relax(void)
|
||||||
|
--
|
||||||
|
2.39.1
|
||||||
|
|
||||||
|
From 48dc9669f8db68adc480ffc2698ed8204440e45b Mon Sep 17 00:00:00 2001
|
||||||
|
From: Arjan van de Ven <arjan@linux.intel.com>
|
||||||
|
Date: Thu, 13 Dec 2018 01:00:49 +0000
|
||||||
|
Subject: [PATCH 2/9] sched/wait: Do accept() in LIFO order for cache
|
||||||
|
efficiency
|
||||||
|
|
||||||
|
Signed-off-by: Alexandre Frade <kernel@xanmod.org>
|
||||||
|
---
|
||||||
|
include/linux/wait.h | 2 ++
|
||||||
|
kernel/sched/wait.c | 24 ++++++++++++++++++++++++
|
||||||
|
net/ipv4/inet_connection_sock.c | 2 +-
|
||||||
|
3 files changed, 27 insertions(+), 1 deletion(-)
|
||||||
|
|
||||||
|
diff --git a/include/linux/wait.h b/include/linux/wait.h
|
||||||
|
index a0307b516b09..edc21128f387 100644
|
||||||
|
--- a/include/linux/wait.h
|
||||||
|
+++ b/include/linux/wait.h
|
||||||
|
@@ -165,6 +165,7 @@ static inline bool wq_has_sleeper(struct wait_queue_head *wq_head)
|
||||||
|
|
||||||
|
extern void add_wait_queue(struct wait_queue_head *wq_head, struct wait_queue_entry *wq_entry);
|
||||||
|
extern void add_wait_queue_exclusive(struct wait_queue_head *wq_head, struct wait_queue_entry *wq_entry);
|
||||||
|
+extern void add_wait_queue_exclusive_lifo(struct wait_queue_head *wq_head, struct wait_queue_entry *wq_entry);
|
||||||
|
extern void add_wait_queue_priority(struct wait_queue_head *wq_head, struct wait_queue_entry *wq_entry);
|
||||||
|
extern void remove_wait_queue(struct wait_queue_head *wq_head, struct wait_queue_entry *wq_entry);
|
||||||
|
|
||||||
|
@@ -1192,6 +1193,7 @@ do { \
|
||||||
|
*/
|
||||||
|
void prepare_to_wait(struct wait_queue_head *wq_head, struct wait_queue_entry *wq_entry, int state);
|
||||||
|
bool prepare_to_wait_exclusive(struct wait_queue_head *wq_head, struct wait_queue_entry *wq_entry, int state);
|
||||||
|
+void prepare_to_wait_exclusive_lifo(struct wait_queue_head *wq_head, struct wait_queue_entry *wq_entry, int state);
|
||||||
|
long prepare_to_wait_event(struct wait_queue_head *wq_head, struct wait_queue_entry *wq_entry, int state);
|
||||||
|
void finish_wait(struct wait_queue_head *wq_head, struct wait_queue_entry *wq_entry);
|
||||||
|
long wait_woken(struct wait_queue_entry *wq_entry, unsigned mode, long timeout);
|
||||||
|
diff --git a/kernel/sched/wait.c b/kernel/sched/wait.c
|
||||||
|
index 133b74730738..1647fb8662eb 100644
|
||||||
|
--- a/kernel/sched/wait.c
|
||||||
|
+++ b/kernel/sched/wait.c
|
||||||
|
@@ -47,6 +47,17 @@ void add_wait_queue_priority(struct wait_queue_head *wq_head, struct wait_queue_
|
||||||
|
}
|
||||||
|
EXPORT_SYMBOL_GPL(add_wait_queue_priority);
|
||||||
|
|
||||||
|
+void add_wait_queue_exclusive_lifo(struct wait_queue_head *wq_head, struct wait_queue_entry *wq_entry)
|
||||||
|
+{
|
||||||
|
+ unsigned long flags;
|
||||||
|
+
|
||||||
|
+ wq_entry->flags |= WQ_FLAG_EXCLUSIVE;
|
||||||
|
+ spin_lock_irqsave(&wq_head->lock, flags);
|
||||||
|
+ __add_wait_queue(wq_head, wq_entry);
|
||||||
|
+ spin_unlock_irqrestore(&wq_head->lock, flags);
|
||||||
|
+}
|
||||||
|
+EXPORT_SYMBOL(add_wait_queue_exclusive_lifo);
|
||||||
|
+
|
||||||
|
void remove_wait_queue(struct wait_queue_head *wq_head, struct wait_queue_entry *wq_entry)
|
||||||
|
{
|
||||||
|
unsigned long flags;
|
||||||
|
@@ -293,6 +304,19 @@ prepare_to_wait_exclusive(struct wait_queue_head *wq_head, struct wait_queue_ent
|
||||||
|
}
|
||||||
|
EXPORT_SYMBOL(prepare_to_wait_exclusive);
|
||||||
|
|
||||||
|
+void prepare_to_wait_exclusive_lifo(struct wait_queue_head *wq_head, struct wait_queue_entry *wq_entry, int state)
|
||||||
|
+{
|
||||||
|
+ unsigned long flags;
|
||||||
|
+
|
||||||
|
+ wq_entry->flags |= WQ_FLAG_EXCLUSIVE;
|
||||||
|
+ spin_lock_irqsave(&wq_head->lock, flags);
|
||||||
|
+ if (list_empty(&wq_entry->entry))
|
||||||
|
+ __add_wait_queue(wq_head, wq_entry);
|
||||||
|
+ set_current_state(state);
|
||||||
|
+ spin_unlock_irqrestore(&wq_head->lock, flags);
|
||||||
|
+}
|
||||||
|
+EXPORT_SYMBOL(prepare_to_wait_exclusive_lifo);
|
||||||
|
+
|
||||||
|
void init_wait_entry(struct wait_queue_entry *wq_entry, int flags)
|
||||||
|
{
|
||||||
|
wq_entry->flags = flags;
|
||||||
|
diff --git a/net/ipv4/inet_connection_sock.c b/net/ipv4/inet_connection_sock.c
|
||||||
|
index f2c43f67187d..9885bfb429a2 100644
|
||||||
|
--- a/net/ipv4/inet_connection_sock.c
|
||||||
|
+++ b/net/ipv4/inet_connection_sock.c
|
||||||
|
@@ -606,7 +606,7 @@ static int inet_csk_wait_for_connect(struct sock *sk, long timeo)
|
||||||
|
* having to remove and re-insert us on the wait queue.
|
||||||
|
*/
|
||||||
|
for (;;) {
|
||||||
|
- prepare_to_wait_exclusive(sk_sleep(sk), &wait,
|
||||||
|
+ prepare_to_wait_exclusive_lifo(sk_sleep(sk), &wait,
|
||||||
|
TASK_INTERRUPTIBLE);
|
||||||
|
release_sock(sk);
|
||||||
|
if (reqsk_queue_empty(&icsk->icsk_accept_queue))
|
||||||
|
--
|
||||||
|
2.39.1
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
From afa213811c5490906caf394b20bb4b616fc6f12a Mon Sep 17 00:00:00 2001
|
||||||
|
From: Ricardo Neri <ricardo.neri-calderon@linux.intel.com>
|
||||||
|
Date: Thu, 25 Aug 2022 15:55:26 -0700
|
||||||
|
Subject: [PATCH 3/9] sched/fair: Simplify asym_packing logic for SMT sched
|
||||||
|
groups
|
||||||
|
|
||||||
|
When the destination CPU is an SMT sibling and idle, it can only help the
|
||||||
|
busiest group if all of its other SMT siblings are also idle. Otherwise,
|
||||||
|
there is not increase in throughput.
|
||||||
|
|
||||||
|
It does not matter whether the busiest group has SMT siblings. Simply
|
||||||
|
check if there are any tasks running on the local group before proceeding.
|
||||||
|
|
||||||
|
Cc: Ben Segall <bsegall@google.com>
|
||||||
|
Cc: Daniel Bristot de Oliveira <bristot@redhat.com>
|
||||||
|
Cc: Dietmar Eggemann <dietmar.eggemann@arm.com>
|
||||||
|
Cc: Len Brown <len.brown@intel.com>
|
||||||
|
Cc: Mel Gorman <mgorman@suse.de>
|
||||||
|
Cc: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
|
||||||
|
Cc: Srinivas Pandruvada <srinivas.pandruvada@linux.intel.com>
|
||||||
|
Cc: Steven Rostedt <rostedt@goodmis.org>
|
||||||
|
Cc: Tim C. Chen <tim.c.chen@intel.com>
|
||||||
|
Cc: Valentin Schneider <vschneid@redhat.com>
|
||||||
|
Cc: x86@kernel.org
|
||||||
|
Cc: linux-kernel@vger.kernel.org
|
||||||
|
Reviewed-by: Len Brown <len.brown@intel.com>
|
||||||
|
Signed-off-by: Ricardo Neri <ricardo.neri-calderon@linux.intel.com>
|
||||||
|
Signed-off-by: Alexandre Frade <kernel@xanmod.org>
|
||||||
|
---
|
||||||
|
kernel/sched/fair.c | 29 +++++++++--------------------
|
||||||
|
1 file changed, 9 insertions(+), 20 deletions(-)
|
||||||
|
|
||||||
|
diff --git a/kernel/sched/fair.c b/kernel/sched/fair.c
|
||||||
|
index 345cc5e9fa6e..60f9690a5626 100644
|
||||||
|
--- a/kernel/sched/fair.c
|
||||||
|
+++ b/kernel/sched/fair.c
|
||||||
|
@@ -8921,12 +8921,10 @@ static bool asym_smt_can_pull_tasks(int dst_cpu, struct sd_lb_stats *sds,
|
||||||
|
struct sched_group *sg)
|
||||||
|
{
|
||||||
|
#ifdef CONFIG_SCHED_SMT
|
||||||
|
- bool local_is_smt, sg_is_smt;
|
||||||
|
+ bool local_is_smt;
|
||||||
|
int sg_busy_cpus;
|
||||||
|
|
||||||
|
local_is_smt = sds->local->flags & SD_SHARE_CPUCAPACITY;
|
||||||
|
- sg_is_smt = sg->flags & SD_SHARE_CPUCAPACITY;
|
||||||
|
-
|
||||||
|
sg_busy_cpus = sgs->group_weight - sgs->idle_cpus;
|
||||||
|
|
||||||
|
if (!local_is_smt) {
|
||||||
|
@@ -8947,25 +8945,16 @@ static bool asym_smt_can_pull_tasks(int dst_cpu, struct sd_lb_stats *sds,
|
||||||
|
return sched_asym_prefer(dst_cpu, sg->asym_prefer_cpu);
|
||||||
|
}
|
||||||
|
|
||||||
|
- /* @dst_cpu has SMT siblings. */
|
||||||
|
-
|
||||||
|
- if (sg_is_smt) {
|
||||||
|
- int local_busy_cpus = sds->local->group_weight -
|
||||||
|
- sds->local_stat.idle_cpus;
|
||||||
|
- int busy_cpus_delta = sg_busy_cpus - local_busy_cpus;
|
||||||
|
-
|
||||||
|
- if (busy_cpus_delta == 1)
|
||||||
|
- return sched_asym_prefer(dst_cpu, sg->asym_prefer_cpu);
|
||||||
|
-
|
||||||
|
- return false;
|
||||||
|
- }
|
||||||
|
-
|
||||||
|
/*
|
||||||
|
- * @sg does not have SMT siblings. Ensure that @sds::local does not end
|
||||||
|
- * up with more than one busy SMT sibling and only pull tasks if there
|
||||||
|
- * are not busy CPUs (i.e., no CPU has running tasks).
|
||||||
|
+ * @dst_cpu has SMT siblings. When both @dst_cpu and the busiest core
|
||||||
|
+ * have one or more busy siblings, moving tasks between them results
|
||||||
|
+ * in the same throughput. Only if all the siblings of @dst_cpu are
|
||||||
|
+ * idle throughput can increase.
|
||||||
|
+ *
|
||||||
|
+ * If the difference in the number of busy CPUs is two or more, let
|
||||||
|
+ * find_busiest_group() take care of it.
|
||||||
|
*/
|
||||||
|
- if (!sds->local_stat.sum_nr_running)
|
||||||
|
+ if (sg_busy_cpus == 1 && !sds->local_stat.sum_nr_running)
|
||||||
|
return sched_asym_prefer(dst_cpu, sg->asym_prefer_cpu);
|
||||||
|
|
||||||
|
return false;
|
||||||
|
--
|
||||||
|
2.39.1
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
From a1f627fd10ced4f5eeae678bc4ba96ea7fa01d7e Mon Sep 17 00:00:00 2001
|
||||||
|
From: Ricardo Neri <ricardo.neri-calderon@linux.intel.com>
|
||||||
|
Date: Thu, 25 Aug 2022 15:55:28 -0700
|
||||||
|
Subject: [PATCH 4/9] sched/fair: Let lower-priority CPUs do active balancing
|
||||||
|
|
||||||
|
When more than one SMT siblings of a physical core are busy, an idle CPU
|
||||||
|
of lower priority can help.
|
||||||
|
|
||||||
|
Indicate that the low priority CPU can do active balancing from the high-
|
||||||
|
priority CPU only if they belong to separate cores.
|
||||||
|
|
||||||
|
Cc: Ben Segall <bsegall@google.com>
|
||||||
|
Cc: Daniel Bristot de Oliveira <bristot@redhat.com>
|
||||||
|
Cc: Dietmar Eggemann <dietmar.eggemann@arm.com>
|
||||||
|
Cc: Len Brown <len.brown@intel.com>
|
||||||
|
Cc: Mel Gorman <mgorman@suse.de>
|
||||||
|
Cc: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
|
||||||
|
Cc: Srinivas Pandruvada <srinivas.pandruvada@linux.intel.com>
|
||||||
|
Cc: Steven Rostedt <rostedt@goodmis.org>
|
||||||
|
Cc: Tim C. Chen <tim.c.chen@intel.com>
|
||||||
|
Cc: Valentin Schneider <vschneid@redhat.com>
|
||||||
|
Cc: x86@kernel.org
|
||||||
|
Cc: linux-kernel@vger.kernel.org
|
||||||
|
Reviewed-by: Len Brown <len.brown@intel.com>
|
||||||
|
Signed-off-by: Ricardo Neri <ricardo.neri-calderon@linux.intel.com>
|
||||||
|
Signed-off-by: Alexandre Frade <kernel@xanmod.org>
|
||||||
|
---
|
||||||
|
kernel/sched/fair.c | 7 ++++++-
|
||||||
|
1 file changed, 6 insertions(+), 1 deletion(-)
|
||||||
|
|
||||||
|
diff --git a/kernel/sched/fair.c b/kernel/sched/fair.c
|
||||||
|
index 60f9690a5626..67b0eacad0e9 100644
|
||||||
|
--- a/kernel/sched/fair.c
|
||||||
|
+++ b/kernel/sched/fair.c
|
||||||
|
@@ -10176,9 +10176,14 @@ asym_active_balance(struct lb_env *env)
|
||||||
|
* ASYM_PACKING needs to force migrate tasks from busy but
|
||||||
|
* lower priority CPUs in order to pack all tasks in the
|
||||||
|
* highest priority CPUs.
|
||||||
|
+ *
|
||||||
|
+ * If the busy CPU has higher priority but is an SMT sibling
|
||||||
|
+ * in which other SMT siblings are also busy, a lower-priority
|
||||||
|
+ * CPU in a separate core can help.
|
||||||
|
*/
|
||||||
|
return env->idle != CPU_NOT_IDLE && (env->sd->flags & SD_ASYM_PACKING) &&
|
||||||
|
- sched_asym_prefer(env->dst_cpu, env->src_cpu);
|
||||||
|
+ (sched_asym_prefer(env->dst_cpu, env->src_cpu) ||
|
||||||
|
+ !(env->sd->flags & SD_SHARE_CPUCAPACITY));
|
||||||
|
}
|
||||||
|
|
||||||
|
static inline bool
|
||||||
|
--
|
||||||
|
2.39.1
|
@@ -0,0 +1,363 @@
|
|||||||
|
From 9c85113cf4019e7b277a44e72bda8b78347aa72f Mon Sep 17 00:00:00 2001
|
||||||
|
From: Paul Gofman <pgofman@codeweavers.com>
|
||||||
|
Date: Thu, 7 May 2020 14:05:31 +0300
|
||||||
|
Subject: [PATCH 2/2] mm: Support soft dirty flag read with reset.
|
||||||
|
|
||||||
|
---
|
||||||
|
fs/proc/base.c | 3 +
|
||||||
|
fs/proc/internal.h | 1 +
|
||||||
|
fs/proc/task_mmu.c | 144 +++++++++++++++++++++++++++++++++++++++------
|
||||||
|
3 files changed, 130 insertions(+), 18 deletions(-)
|
||||||
|
|
||||||
|
diff --git a/fs/proc/base.c b/fs/proc/base.c
|
||||||
|
index b3422cda2a91..8199ae2411ca 100644
|
||||||
|
--- a/fs/proc/base.c
|
||||||
|
+++ b/fs/proc/base.c
|
||||||
|
@@ -3202,6 +3202,9 @@ static const struct pid_entry tgid_base_stuff[] = {
|
||||||
|
REG("smaps", S_IRUGO, proc_pid_smaps_operations),
|
||||||
|
REG("smaps_rollup", S_IRUGO, proc_pid_smaps_rollup_operations),
|
||||||
|
REG("pagemap", S_IRUSR, proc_pagemap_operations),
|
||||||
|
+#ifdef CONFIG_MEM_SOFT_DIRTY
|
||||||
|
+ REG("pagemap_reset", S_IRUSR, proc_pagemap_reset_operations),
|
||||||
|
+#endif
|
||||||
|
#endif
|
||||||
|
#ifdef CONFIG_SECURITY
|
||||||
|
DIR("attr", S_IRUGO|S_IXUGO, proc_attr_dir_inode_operations, proc_attr_dir_operations),
|
||||||
|
diff --git a/fs/proc/internal.h b/fs/proc/internal.h
|
||||||
|
index f60b379dcdc7..36a901cf0e7f 100644
|
||||||
|
--- a/fs/proc/internal.h
|
||||||
|
+++ b/fs/proc/internal.h
|
||||||
|
@@ -303,6 +303,7 @@ extern const struct file_operations proc_pid_smaps_operations;
|
||||||
|
extern const struct file_operations proc_pid_smaps_rollup_operations;
|
||||||
|
extern const struct file_operations proc_clear_refs_operations;
|
||||||
|
extern const struct file_operations proc_pagemap_operations;
|
||||||
|
+extern const struct file_operations proc_pagemap_reset_operations;
|
||||||
|
|
||||||
|
extern unsigned long task_vsize(struct mm_struct *);
|
||||||
|
extern unsigned long task_statm(struct mm_struct *,
|
||||||
|
diff --git a/fs/proc/task_mmu.c b/fs/proc/task_mmu.c
|
||||||
|
index 7c7865028f10..a21694967915 100644
|
||||||
|
--- a/fs/proc/task_mmu.c
|
||||||
|
+++ b/fs/proc/task_mmu.c
|
||||||
|
@@ -1056,8 +1056,8 @@ static inline bool pte_is_pinned(struct vm_area_struct *vma, unsigned long addr,
|
||||||
|
return page_maybe_dma_pinned(page);
|
||||||
|
}
|
||||||
|
|
||||||
|
-static inline void clear_soft_dirty(struct vm_area_struct *vma,
|
||||||
|
- unsigned long addr, pte_t *pte)
|
||||||
|
+static inline bool clear_soft_dirty(struct vm_area_struct *vma,
|
||||||
|
+ unsigned long addr, pte_t *pte)
|
||||||
|
{
|
||||||
|
/*
|
||||||
|
* The soft-dirty tracker uses #PF-s to catch writes
|
||||||
|
@@ -1066,37 +1066,46 @@ static inline void clear_soft_dirty(struct vm_area_struct *vma,
|
||||||
|
* of how soft-dirty works.
|
||||||
|
*/
|
||||||
|
pte_t ptent = *pte;
|
||||||
|
+ bool ret = false;
|
||||||
|
|
||||||
|
if (pte_present(ptent)) {
|
||||||
|
pte_t old_pte;
|
||||||
|
|
||||||
|
if (pte_is_pinned(vma, addr, ptent))
|
||||||
|
- return;
|
||||||
|
+ return ret;
|
||||||
|
old_pte = ptep_modify_prot_start(vma, addr, pte);
|
||||||
|
+ ret = pte_soft_dirty(old_pte);
|
||||||
|
ptent = pte_wrprotect(old_pte);
|
||||||
|
ptent = pte_clear_soft_dirty(ptent);
|
||||||
|
ptep_modify_prot_commit(vma, addr, pte, old_pte, ptent);
|
||||||
|
} else if (is_swap_pte(ptent)) {
|
||||||
|
+ ret = pte_swp_soft_dirty(ptent);
|
||||||
|
ptent = pte_swp_clear_soft_dirty(ptent);
|
||||||
|
set_pte_at(vma->vm_mm, addr, pte, ptent);
|
||||||
|
}
|
||||||
|
+ return ret;
|
||||||
|
}
|
||||||
|
#else
|
||||||
|
-static inline void clear_soft_dirty(struct vm_area_struct *vma,
|
||||||
|
+static inline bool clear_soft_dirty(struct vm_area_struct *vma,
|
||||||
|
unsigned long addr, pte_t *pte)
|
||||||
|
{
|
||||||
|
+ return false;
|
||||||
|
}
|
||||||
|
#endif
|
||||||
|
|
||||||
|
#if defined(CONFIG_MEM_SOFT_DIRTY) && defined(CONFIG_TRANSPARENT_HUGEPAGE)
|
||||||
|
-static inline void clear_soft_dirty_pmd(struct vm_area_struct *vma,
|
||||||
|
+static inline bool clear_soft_dirty_pmd(struct vm_area_struct *vma,
|
||||||
|
unsigned long addr, pmd_t *pmdp)
|
||||||
|
{
|
||||||
|
pmd_t old, pmd = *pmdp;
|
||||||
|
+ bool ret = false;
|
||||||
|
|
||||||
|
if (pmd_present(pmd)) {
|
||||||
|
/* See comment in change_huge_pmd() */
|
||||||
|
old = pmdp_invalidate(vma, addr, pmdp);
|
||||||
|
+
|
||||||
|
+ ret = pmd_soft_dirty(old);
|
||||||
|
+
|
||||||
|
if (pmd_dirty(old))
|
||||||
|
pmd = pmd_mkdirty(pmd);
|
||||||
|
if (pmd_young(old))
|
||||||
|
@@ -1107,14 +1116,17 @@ static inline void clear_soft_dirty_pmd(struct vm_area_struct *vma,
|
||||||
|
|
||||||
|
set_pmd_at(vma->vm_mm, addr, pmdp, pmd);
|
||||||
|
} else if (is_migration_entry(pmd_to_swp_entry(pmd))) {
|
||||||
|
+ ret = pmd_swp_soft_dirty(pmd);
|
||||||
|
pmd = pmd_swp_clear_soft_dirty(pmd);
|
||||||
|
set_pmd_at(vma->vm_mm, addr, pmdp, pmd);
|
||||||
|
}
|
||||||
|
+ return ret;
|
||||||
|
}
|
||||||
|
#else
|
||||||
|
-static inline void clear_soft_dirty_pmd(struct vm_area_struct *vma,
|
||||||
|
+static inline bool clear_soft_dirty_pmd(struct vm_area_struct *vma,
|
||||||
|
unsigned long addr, pmd_t *pmdp)
|
||||||
|
{
|
||||||
|
+ return false;
|
||||||
|
}
|
||||||
|
#endif
|
||||||
|
|
||||||
|
@@ -1367,6 +1379,7 @@ struct pagemapread {
|
||||||
|
int pos, len; /* units: PM_ENTRY_BYTES, not bytes */
|
||||||
|
pagemap_entry_t *buffer;
|
||||||
|
bool show_pfn;
|
||||||
|
+ bool reset;
|
||||||
|
};
|
||||||
|
|
||||||
|
#define PAGEMAP_WALK_SIZE (PMD_SIZE)
|
||||||
|
@@ -1398,6 +1411,14 @@ static int add_to_pagemap(unsigned long addr, pagemap_entry_t *pme,
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
+static int add_addr_to_pagemap(unsigned long addr, struct pagemapread *pm)
|
||||||
|
+{
|
||||||
|
+ ((unsigned long *)pm->buffer)[pm->pos++] = addr;
|
||||||
|
+ if (pm->pos >= pm->len)
|
||||||
|
+ return PM_END_OF_BUFFER;
|
||||||
|
+ return 0;
|
||||||
|
+}
|
||||||
|
+
|
||||||
|
static int pagemap_pte_hole(unsigned long start, unsigned long end,
|
||||||
|
__always_unused int depth, struct mm_walk *walk)
|
||||||
|
{
|
||||||
|
@@ -1405,6 +1426,9 @@ static int pagemap_pte_hole(unsigned long start, unsigned long end,
|
||||||
|
unsigned long addr = start;
|
||||||
|
int err = 0;
|
||||||
|
|
||||||
|
+ if (pm->reset)
|
||||||
|
+ goto out;
|
||||||
|
+
|
||||||
|
while (addr < end) {
|
||||||
|
struct vm_area_struct *vma = find_vma(walk->mm, addr);
|
||||||
|
pagemap_entry_t pme = make_pme(0, 0);
|
||||||
|
@@ -1439,8 +1463,9 @@ static int pagemap_pte_hole(unsigned long start, unsigned long end,
|
||||||
|
}
|
||||||
|
|
||||||
|
static pagemap_entry_t pte_to_pagemap_entry(struct pagemapread *pm,
|
||||||
|
- struct vm_area_struct *vma, unsigned long addr, pte_t pte)
|
||||||
|
+ struct vm_area_struct *vma, unsigned long addr, pte_t *pte_addr)
|
||||||
|
{
|
||||||
|
+ pte_t pte = *pte_addr;
|
||||||
|
u64 frame = 0, flags = 0;
|
||||||
|
struct page *page = NULL;
|
||||||
|
|
||||||
|
@@ -1493,6 +1518,20 @@ static int pagemap_pmd_range(pmd_t *pmdp, unsigned long addr, unsigned long end,
|
||||||
|
pmd_t pmd = *pmdp;
|
||||||
|
struct page *page = NULL;
|
||||||
|
|
||||||
|
+ if (pm->reset)
|
||||||
|
+ {
|
||||||
|
+ if (clear_soft_dirty_pmd(vma, addr, pmdp))
|
||||||
|
+ {
|
||||||
|
+ for (; addr != end; addr += PAGE_SIZE)
|
||||||
|
+ {
|
||||||
|
+ err = add_addr_to_pagemap(addr, pm);
|
||||||
|
+ if (err)
|
||||||
|
+ break;
|
||||||
|
+ }
|
||||||
|
+ }
|
||||||
|
+ goto trans_huge_done;
|
||||||
|
+ }
|
||||||
|
+
|
||||||
|
if (vma->vm_flags & VM_SOFTDIRTY)
|
||||||
|
flags |= PM_SOFT_DIRTY;
|
||||||
|
|
||||||
|
@@ -1541,6 +1580,7 @@ static int pagemap_pmd_range(pmd_t *pmdp, unsigned long addr, unsigned long end,
|
||||||
|
frame += (1 << MAX_SWAPFILES_SHIFT);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
+trans_huge_done:
|
||||||
|
spin_unlock(ptl);
|
||||||
|
return err;
|
||||||
|
}
|
||||||
|
@@ -1555,10 +1595,18 @@ static int pagemap_pmd_range(pmd_t *pmdp, unsigned long addr, unsigned long end,
|
||||||
|
*/
|
||||||
|
orig_pte = pte = pte_offset_map_lock(walk->mm, pmdp, addr, &ptl);
|
||||||
|
for (; addr < end; pte++, addr += PAGE_SIZE) {
|
||||||
|
- pagemap_entry_t pme;
|
||||||
|
+ if (pm->reset)
|
||||||
|
+ {
|
||||||
|
+ if (clear_soft_dirty(vma, addr, pte))
|
||||||
|
+ err = add_addr_to_pagemap(addr, pm);
|
||||||
|
+ }
|
||||||
|
+ else
|
||||||
|
+ {
|
||||||
|
+ pagemap_entry_t pme;
|
||||||
|
|
||||||
|
- pme = pte_to_pagemap_entry(pm, vma, addr, *pte);
|
||||||
|
- err = add_to_pagemap(addr, &pme, pm);
|
||||||
|
+ pme = pte_to_pagemap_entry(pm, vma, addr, pte);
|
||||||
|
+ err = add_to_pagemap(addr, &pme, pm);
|
||||||
|
+ }
|
||||||
|
if (err)
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
@@ -1650,8 +1698,8 @@ static const struct mm_walk_ops pagemap_ops = {
|
||||||
|
* determine which areas of memory are actually mapped and llseek to
|
||||||
|
* skip over unmapped regions.
|
||||||
|
*/
|
||||||
|
-static ssize_t pagemap_read(struct file *file, char __user *buf,
|
||||||
|
- size_t count, loff_t *ppos)
|
||||||
|
+static ssize_t do_pagemap_read(struct file *file, char __user *buf,
|
||||||
|
+ size_t count, loff_t *ppos, bool reset)
|
||||||
|
{
|
||||||
|
struct mm_struct *mm = file->private_data;
|
||||||
|
struct pagemapread pm;
|
||||||
|
@@ -1660,6 +1708,8 @@ static ssize_t pagemap_read(struct file *file, char __user *buf,
|
||||||
|
unsigned long start_vaddr;
|
||||||
|
unsigned long end_vaddr;
|
||||||
|
int ret = 0, copied = 0;
|
||||||
|
+ struct mmu_notifier_range range;
|
||||||
|
+ size_t buffer_len;
|
||||||
|
|
||||||
|
if (!mm || !mmget_not_zero(mm))
|
||||||
|
goto out;
|
||||||
|
@@ -1675,19 +1725,38 @@ static ssize_t pagemap_read(struct file *file, char __user *buf,
|
||||||
|
|
||||||
|
/* do not disclose physical addresses: attack vector */
|
||||||
|
pm.show_pfn = file_ns_capable(file, &init_user_ns, CAP_SYS_ADMIN);
|
||||||
|
+ pm.reset = reset;
|
||||||
|
|
||||||
|
- pm.len = (PAGEMAP_WALK_SIZE >> PAGE_SHIFT);
|
||||||
|
- pm.buffer = kmalloc_array(pm.len, PM_ENTRY_BYTES, GFP_KERNEL);
|
||||||
|
+ buffer_len = min(PAGEMAP_WALK_SIZE >> PAGE_SHIFT, count / PM_ENTRY_BYTES);
|
||||||
|
+
|
||||||
|
+ pm.buffer = kmalloc_array(buffer_len, PM_ENTRY_BYTES, GFP_KERNEL);
|
||||||
|
ret = -ENOMEM;
|
||||||
|
if (!pm.buffer)
|
||||||
|
goto out_mm;
|
||||||
|
|
||||||
|
src = *ppos;
|
||||||
|
svpfn = src / PM_ENTRY_BYTES;
|
||||||
|
- end_vaddr = mm->task_size;
|
||||||
|
+
|
||||||
|
+ start_vaddr = svpfn << PAGE_SHIFT;
|
||||||
|
+
|
||||||
|
+ if (reset)
|
||||||
|
+ {
|
||||||
|
+ if (count < sizeof(end_vaddr))
|
||||||
|
+ {
|
||||||
|
+ ret = -EINVAL;
|
||||||
|
+ goto out_mm;
|
||||||
|
+ }
|
||||||
|
+ if (copy_from_user(&end_vaddr, buf, sizeof(end_vaddr)))
|
||||||
|
+ return -EFAULT;
|
||||||
|
+ end_vaddr = min(end_vaddr, mm->task_size);
|
||||||
|
+ }
|
||||||
|
+ else
|
||||||
|
+ {
|
||||||
|
+ end_vaddr = mm->task_size;
|
||||||
|
+ start_vaddr = end_vaddr;
|
||||||
|
+ }
|
||||||
|
|
||||||
|
/* watch out for wraparound */
|
||||||
|
- start_vaddr = end_vaddr;
|
||||||
|
if (svpfn <= (ULONG_MAX >> PAGE_SHIFT))
|
||||||
|
start_vaddr = untagged_addr(svpfn << PAGE_SHIFT);
|
||||||
|
|
||||||
|
@@ -1707,18 +1776,35 @@ static ssize_t pagemap_read(struct file *file, char __user *buf,
|
||||||
|
unsigned long end;
|
||||||
|
|
||||||
|
pm.pos = 0;
|
||||||
|
- end = (start_vaddr + PAGEMAP_WALK_SIZE) & PAGEMAP_WALK_MASK;
|
||||||
|
+ pm.len = min(buffer_len, count / PM_ENTRY_BYTES);
|
||||||
|
+
|
||||||
|
+ end = reset ? end_vaddr : (start_vaddr + (pm.len << PAGE_SHIFT));
|
||||||
|
/* overflow ? */
|
||||||
|
if (end < start_vaddr || end > end_vaddr)
|
||||||
|
end = end_vaddr;
|
||||||
|
+
|
||||||
|
ret = mmap_read_lock_killable(mm);
|
||||||
|
if (ret)
|
||||||
|
goto out_free;
|
||||||
|
+
|
||||||
|
+ if (reset)
|
||||||
|
+ {
|
||||||
|
+ inc_tlb_flush_pending(mm);
|
||||||
|
+ mmu_notifier_range_init(&range, MMU_NOTIFY_SOFT_DIRTY,
|
||||||
|
+ 0, mm, start_vaddr, end);
|
||||||
|
+ mmu_notifier_invalidate_range_start(&range);
|
||||||
|
+ }
|
||||||
|
ret = walk_page_range(mm, start_vaddr, end, &pagemap_ops, &pm);
|
||||||
|
+ if (reset)
|
||||||
|
+ {
|
||||||
|
+ mmu_notifier_invalidate_range_end(&range);
|
||||||
|
+ flush_tlb_mm(mm);
|
||||||
|
+ dec_tlb_flush_pending(mm);
|
||||||
|
+ }
|
||||||
|
mmap_read_unlock(mm);
|
||||||
|
- start_vaddr = end;
|
||||||
|
|
||||||
|
len = min(count, PM_ENTRY_BYTES * pm.pos);
|
||||||
|
+ BUG_ON(ret && ret != PM_END_OF_BUFFER);
|
||||||
|
if (copy_to_user(buf, pm.buffer, len)) {
|
||||||
|
ret = -EFAULT;
|
||||||
|
goto out_free;
|
||||||
|
@@ -1726,6 +1812,8 @@ static ssize_t pagemap_read(struct file *file, char __user *buf,
|
||||||
|
copied += len;
|
||||||
|
buf += len;
|
||||||
|
count -= len;
|
||||||
|
+
|
||||||
|
+ start_vaddr = reset && pm.pos == pm.len ? ((unsigned long *)pm.buffer)[pm.pos - 1] + PAGE_SIZE : end;
|
||||||
|
}
|
||||||
|
*ppos += copied;
|
||||||
|
if (!ret || ret == PM_END_OF_BUFFER)
|
||||||
|
@@ -1739,6 +1827,18 @@ static ssize_t pagemap_read(struct file *file, char __user *buf,
|
||||||
|
return ret;
|
||||||
|
}
|
||||||
|
|
||||||
|
+static ssize_t pagemap_read(struct file *file, char __user *buf,
|
||||||
|
+ size_t count, loff_t *ppos)
|
||||||
|
+{
|
||||||
|
+ return do_pagemap_read(file, buf, count, ppos, false);
|
||||||
|
+}
|
||||||
|
+
|
||||||
|
+static ssize_t pagemap_reset_read(struct file *file, char __user *buf,
|
||||||
|
+ size_t count, loff_t *ppos)
|
||||||
|
+{
|
||||||
|
+ return do_pagemap_read(file, buf, count, ppos, true);
|
||||||
|
+}
|
||||||
|
+
|
||||||
|
static int pagemap_open(struct inode *inode, struct file *file)
|
||||||
|
{
|
||||||
|
struct mm_struct *mm;
|
||||||
|
@@ -1765,6 +1865,14 @@ const struct file_operations proc_pagemap_operations = {
|
||||||
|
.open = pagemap_open,
|
||||||
|
.release = pagemap_release,
|
||||||
|
};
|
||||||
|
+
|
||||||
|
+const struct file_operations proc_pagemap_reset_operations = {
|
||||||
|
+ .llseek = mem_lseek, /* borrow this */
|
||||||
|
+ .read = pagemap_reset_read,
|
||||||
|
+ .open = pagemap_open,
|
||||||
|
+ .release = pagemap_release,
|
||||||
|
+};
|
||||||
|
+
|
||||||
|
#endif /* CONFIG_PROC_PAGE_MONITOR */
|
||||||
|
|
||||||
|
#ifdef CONFIG_NUMA
|
||||||
|
--
|
||||||
|
2.30.2
|
||||||
|
|
879
linux-tkg-patches/6.3/0003-glitched-base.patch
Normal file
879
linux-tkg-patches/6.3/0003-glitched-base.patch
Normal file
@@ -0,0 +1,879 @@
|
|||||||
|
From f7f49141a5dbe9c99d78196b58c44307fb2e6be3 Mon Sep 17 00:00:00 2001
|
||||||
|
From: Tk-Glitch <ti3nou@gmail.com>
|
||||||
|
Date: Wed, 4 Jul 2018 04:30:08 +0200
|
||||||
|
Subject: [PATCH 01/17] glitched
|
||||||
|
|
||||||
|
---
|
||||||
|
init/Makefile | 2 +-
|
||||||
|
1 file changed, 1 insertions(+), 1 deletions(-)
|
||||||
|
|
||||||
|
diff --git a/init/Makefile b/init/Makefile
|
||||||
|
index baf3ab8d9d49..854e32e6aec7 100755
|
||||||
|
--- a/init/Makefile
|
||||||
|
+++ b/init/Makefile
|
||||||
|
@@ -19,7 +19,7 @@ else
|
||||||
|
|
||||||
|
# Maximum length of UTS_VERSION is 64 chars
|
||||||
|
filechk_uts_version = \
|
||||||
|
- utsver=$$(echo '$(pound)'"$(build-version)" $(smp-flag-y) $(preempt-flag-y) "$(build-timestamp)" | cut -b -64); \
|
||||||
|
+ utsver=$$(echo '$(pound)'"$(build-version)" $(smp-flag-y) $(preempt-flag-y) "TKG" "$(build-timestamp)" | cut -b -64); \
|
||||||
|
echo '$(pound)'define UTS_VERSION \""$${utsver}"\"
|
||||||
|
|
||||||
|
#
|
||||||
|
--
|
||||||
|
2.28.0
|
||||||
|
|
||||||
|
|
||||||
|
From c304f43d14e98d4bf1215fc10bc5012f554bdd8a Mon Sep 17 00:00:00 2001
|
||||||
|
From: Alexandre Frade <admfrade@gmail.com>
|
||||||
|
Date: Mon, 29 Jan 2018 16:59:22 +0000
|
||||||
|
Subject: [PATCH 02/17] dcache: cache_pressure = 50 decreases the rate at which
|
||||||
|
VFS caches are reclaimed
|
||||||
|
|
||||||
|
Signed-off-by: Alexandre Frade <admfrade@gmail.com>
|
||||||
|
---
|
||||||
|
fs/dcache.c | 2 +-
|
||||||
|
1 file changed, 1 insertion(+), 1 deletion(-)
|
||||||
|
|
||||||
|
diff --git a/fs/dcache.c b/fs/dcache.c
|
||||||
|
index 361ea7ab30ea..0c5cf69b241a 100644
|
||||||
|
--- a/fs/dcache.c
|
||||||
|
+++ b/fs/dcache.c
|
||||||
|
@@ -71,7 +71,7 @@
|
||||||
|
* If no ancestor relationship:
|
||||||
|
* arbitrary, since it's serialized on rename_lock
|
||||||
|
*/
|
||||||
|
-int sysctl_vfs_cache_pressure __read_mostly = 100;
|
||||||
|
+int sysctl_vfs_cache_pressure __read_mostly = 50;
|
||||||
|
EXPORT_SYMBOL_GPL(sysctl_vfs_cache_pressure);
|
||||||
|
|
||||||
|
__cacheline_aligned_in_smp DEFINE_SEQLOCK(rename_lock);
|
||||||
|
--
|
||||||
|
2.28.0
|
||||||
|
|
||||||
|
|
||||||
|
diff --git a/kernel/sched/rt.c b/kernel/sched/rt.c
|
||||||
|
index f788cd61df21..2bfbb4213707 100644
|
||||||
|
--- a/kernel/sched/rt.c
|
||||||
|
+++ b/kernel/sched/rt.c
|
||||||
|
@@ -15,9 +15,9 @@ __read_mostly int scheduler_running;
|
||||||
|
|
||||||
|
/*
|
||||||
|
* part of the period that we allow rt tasks to run in us.
|
||||||
|
- * default: 0.95s
|
||||||
|
+ * XanMod default: 0.98s
|
||||||
|
*/
|
||||||
|
-int sysctl_sched_rt_runtime = 950000;
|
||||||
|
+int sysctl_sched_rt_runtime = 980000;
|
||||||
|
|
||||||
|
#ifdef CONFIG_SYSCTL
|
||||||
|
static int sysctl_sched_rr_timeslice = (MSEC_PER_SEC / HZ) * RR_TIMESLICE;
|
||||||
|
--
|
||||||
|
2.28.0
|
||||||
|
|
||||||
|
|
||||||
|
From acc49f33a10f61dc66c423888cbb883ba46710e4 Mon Sep 17 00:00:00 2001
|
||||||
|
From: Alexandre Frade <admfrade@gmail.com>
|
||||||
|
Date: Mon, 29 Jan 2018 17:41:29 +0000
|
||||||
|
Subject: [PATCH 04/17] scripts: disable the localversion "+" tag of a git repo
|
||||||
|
|
||||||
|
Signed-off-by: Alexandre Frade <admfrade@gmail.com>
|
||||||
|
---
|
||||||
|
scripts/setlocalversion | 2 +-
|
||||||
|
1 file changed, 1 insertion(+), 1 deletion(-)
|
||||||
|
|
||||||
|
diff --git a/scripts/setlocalversion b/scripts/setlocalversion
|
||||||
|
index 20f2efd57b11..0552d8b9f582 100755
|
||||||
|
--- a/scripts/setlocalversion
|
||||||
|
+++ b/scripts/setlocalversion
|
||||||
|
@@ -54,7 +54,7 @@ scm_version()
|
||||||
|
# If only the short version is requested, don't bother
|
||||||
|
# running further git commands
|
||||||
|
if $short; then
|
||||||
|
- echo "+"
|
||||||
|
+ #echo "+"
|
||||||
|
return
|
||||||
|
fi
|
||||||
|
# If we are past the tagged commit, we pretty print it.
|
||||||
|
--
|
||||||
|
2.28.0
|
||||||
|
|
||||||
|
|
||||||
|
From 360c6833e07cc9fdef5746f6bc45bdbc7212288d Mon Sep 17 00:00:00 2001
|
||||||
|
From: "Jan Alexander Steffens (heftig)" <jan.steffens@gmail.com>
|
||||||
|
Date: Fri, 26 Oct 2018 11:22:33 +0100
|
||||||
|
Subject: [PATCH 06/17] infiniband: Fix __read_overflow2 error with -O3
|
||||||
|
inlining
|
||||||
|
|
||||||
|
---
|
||||||
|
drivers/infiniband/core/addr.c | 1 +
|
||||||
|
1 file changed, 1 insertion(+)
|
||||||
|
|
||||||
|
diff --git a/drivers/infiniband/core/addr.c b/drivers/infiniband/core/addr.c
|
||||||
|
index 3a98439bba83..6efc4f907f58 100644
|
||||||
|
--- a/drivers/infiniband/core/addr.c
|
||||||
|
+++ b/drivers/infiniband/core/addr.c
|
||||||
|
@@ -820,6 +820,7 @@ int rdma_addr_find_l2_eth_by_grh(const union ib_gid *sgid,
|
||||||
|
union {
|
||||||
|
struct sockaddr_in _sockaddr_in;
|
||||||
|
struct sockaddr_in6 _sockaddr_in6;
|
||||||
|
+ struct sockaddr_ib _sockaddr_ib;
|
||||||
|
} sgid_addr, dgid_addr;
|
||||||
|
int ret;
|
||||||
|
|
||||||
|
--
|
||||||
|
2.28.0
|
||||||
|
|
||||||
|
|
||||||
|
From f85ed068b4d0e6c31edce8574a95757a60e58b87 Mon Sep 17 00:00:00 2001
|
||||||
|
From: Etienne Juvigny <Ti3noU@gmail.com>
|
||||||
|
Date: Mon, 3 Sep 2018 17:36:25 +0200
|
||||||
|
Subject: [PATCH 07/17] Zenify & stuff
|
||||||
|
|
||||||
|
---
|
||||||
|
init/Kconfig | 32 ++++++++++++++++++++++++++++++++
|
||||||
|
kernel/sched/fair.c | 25 +++++++++++++++++++++++++
|
||||||
|
mm/page-writeback.c | 8 ++++++++
|
||||||
|
3 files changed, 65 insertions(+)
|
||||||
|
|
||||||
|
diff --git a/init/Kconfig b/init/Kconfig
|
||||||
|
index 3ae8678e1145..da708eed0f1e 100644
|
||||||
|
--- a/init/Kconfig
|
||||||
|
+++ b/init/Kconfig
|
||||||
|
@@ -92,6 +92,38 @@ config THREAD_INFO_IN_TASK
|
||||||
|
|
||||||
|
menu "General setup"
|
||||||
|
|
||||||
|
+config ZENIFY
|
||||||
|
+ bool "A selection of patches from Zen/Liquorix kernel and additional tweaks for a better gaming experience"
|
||||||
|
+ default y
|
||||||
|
+ help
|
||||||
|
+ Tunes the kernel for responsiveness at the cost of throughput and power usage.
|
||||||
|
+
|
||||||
|
+ --- Virtual Memory Subsystem ---------------------------
|
||||||
|
+
|
||||||
|
+ Mem dirty before bg writeback..: 10 % -> 20 %
|
||||||
|
+ Mem dirty before sync writeback: 20 % -> 50 %
|
||||||
|
+
|
||||||
|
+ --- Block Layer ----------------------------------------
|
||||||
|
+
|
||||||
|
+ Queue depth...............: 128 -> 512
|
||||||
|
+ Default MQ scheduler......: mq-deadline -> bfq
|
||||||
|
+
|
||||||
|
+ --- CFS CPU Scheduler ----------------------------------
|
||||||
|
+
|
||||||
|
+ Scheduling latency.............: 6 -> 3 ms
|
||||||
|
+ Minimal granularity............: 0.75 -> 0.3 ms
|
||||||
|
+ Wakeup granularity.............: 1 -> 0.5 ms
|
||||||
|
+ CPU migration cost.............: 0.5 -> 0.25 ms
|
||||||
|
+ Bandwidth slice size...........: 5 -> 3 ms
|
||||||
|
+ Ondemand fine upscaling limit..: 95 % -> 85 %
|
||||||
|
+
|
||||||
|
+ --- MuQSS CPU Scheduler --------------------------------
|
||||||
|
+
|
||||||
|
+ Scheduling interval............: 6 -> 3 ms
|
||||||
|
+ ISO task max realtime use......: 70 % -> 25 %
|
||||||
|
+ Ondemand coarse upscaling limit: 80 % -> 45 %
|
||||||
|
+ Ondemand fine upscaling limit..: 95 % -> 45 %
|
||||||
|
+
|
||||||
|
config BROKEN
|
||||||
|
bool
|
||||||
|
|
||||||
|
diff --git a/kernel/sched/fair.c b/kernel/sched/fair.c
|
||||||
|
index 6b3b59cc51d6..2a0072192c3d 100644
|
||||||
|
--- a/kernel/sched/fair.c
|
||||||
|
+++ b/kernel/sched/fair.c
|
||||||
|
@@ -37,8 +37,13 @@
|
||||||
|
*
|
||||||
|
* (default: 6ms * (1 + ilog(ncpus)), units: nanoseconds)
|
||||||
|
*/
|
||||||
|
+#ifdef CONFIG_ZENIFY
|
||||||
|
+unsigned int sysctl_sched_latency = 3000000ULL;
|
||||||
|
+static unsigned int normalized_sysctl_sched_latency = 3000000ULL;
|
||||||
|
+#else
|
||||||
|
unsigned int sysctl_sched_latency = 6000000ULL;
|
||||||
|
static unsigned int normalized_sysctl_sched_latency = 6000000ULL;
|
||||||
|
+#endif
|
||||||
|
|
||||||
|
/*
|
||||||
|
* The initial- and re-scaling of tunables is configurable
|
||||||
|
@@ -58,21 +63,34 @@ enum sched_tunable_scaling sysctl_sched_tunable_scaling = SCHED_TUNABLESCALING_L
|
||||||
|
*
|
||||||
|
* (default: 0.75 msec * (1 + ilog(ncpus)), units: nanoseconds)
|
||||||
|
*/
|
||||||
|
+#ifdef CONFIG_ZENIFY
|
||||||
|
+unsigned int sysctl_sched_min_granularity = 300000ULL;
|
||||||
|
+static unsigned int normalized_sysctl_sched_min_granularity = 300000ULL;
|
||||||
|
+#else
|
||||||
|
unsigned int sysctl_sched_min_granularity = 750000ULL;
|
||||||
|
static unsigned int normalized_sysctl_sched_min_granularity = 750000ULL;
|
||||||
|
+#endif
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Minimal preemption granularity for CPU-bound SCHED_IDLE tasks.
|
||||||
|
* Applies only when SCHED_IDLE tasks compete with normal tasks.
|
||||||
|
*
|
||||||
|
* (default: 0.75 msec)
|
||||||
|
*/
|
||||||
|
+#ifdef CONFIG_ZENIFY
|
||||||
|
+unsigned int sysctl_sched_idle_min_granularity = 300000ULL;
|
||||||
|
+#else
|
||||||
|
unsigned int sysctl_sched_idle_min_granularity = 750000ULL;
|
||||||
|
+#endif
|
||||||
|
|
||||||
|
/*
|
||||||
|
* This value is kept at sysctl_sched_latency/sysctl_sched_min_granularity
|
||||||
|
*/
|
||||||
|
+#ifdef CONFIG_ZENIFY
|
||||||
|
+static unsigned int sched_nr_latency = 10;
|
||||||
|
+#else
|
||||||
|
static unsigned int sched_nr_latency = 8;
|
||||||
|
+#endif
|
||||||
|
|
||||||
|
/*
|
||||||
|
* After fork, child runs first. If set to 0 (default) then
|
||||||
|
@@ -128,8 +149,12 @@ int __weak arch_asym_cpu_priority(int cpu)
|
||||||
|
*
|
||||||
|
* (default: 5 msec, units: microseconds)
|
||||||
|
*/
|
||||||
|
+#ifdef CONFIG_ZENIFY
|
||||||
|
+static unsigned int sysctl_sched_cfs_bandwidth_slice = 3000UL;
|
||||||
|
+#else
|
||||||
|
static unsigned int sysctl_sched_cfs_bandwidth_slice = 5000UL;
|
||||||
|
#endif
|
||||||
|
+#endif
|
||||||
|
|
||||||
|
#ifdef CONFIG_SYSCTL
|
||||||
|
static struct ctl_table sched_fair_sysctls[] = {
|
||||||
|
diff --git a/mm/page-writeback.c b/mm/page-writeback.c
|
||||||
|
index 28b3e7a67565..01a1aef2b9b1 100644
|
||||||
|
--- a/mm/page-writeback.c
|
||||||
|
+++ b/mm/page-writeback.c
|
||||||
|
@@ -71,7 +71,11 @@ static long ratelimit_pages = 32;
|
||||||
|
/*
|
||||||
|
* Start background writeback (via writeback threads) at this percentage
|
||||||
|
*/
|
||||||
|
+#ifdef CONFIG_ZENIFY
|
||||||
|
+static int dirty_background_ratio = 20;
|
||||||
|
+#else
|
||||||
|
static int dirty_background_ratio = 10;
|
||||||
|
+#endif
|
||||||
|
|
||||||
|
/*
|
||||||
|
* dirty_background_bytes starts at 0 (disabled) so that it is a function of
|
||||||
|
@@ -88,7 +92,11 @@ int vm_highmem_is_dirtyable;
|
||||||
|
/*
|
||||||
|
* The generator of dirty data starts writeback at this percentage
|
||||||
|
*/
|
||||||
|
+#ifdef CONFIG_ZENIFY
|
||||||
|
+static int vm_dirty_ratio = 50;
|
||||||
|
+#else
|
||||||
|
static int vm_dirty_ratio = 20;
|
||||||
|
+#endif
|
||||||
|
|
||||||
|
/*
|
||||||
|
* vm_dirty_bytes starts at 0 (disabled) so that it is a function of
|
||||||
|
--
|
||||||
|
2.28.0
|
||||||
|
|
||||||
|
|
||||||
|
From e92e67143385cf285851e12aa8b7f083dd38dd24 Mon Sep 17 00:00:00 2001
|
||||||
|
From: Steven Barrett <damentz@liquorix.net>
|
||||||
|
Date: Sun, 16 Jan 2011 18:57:32 -0600
|
||||||
|
Subject: [PATCH 08/17] ZEN: Allow TCP YeAH as default congestion control
|
||||||
|
|
||||||
|
4.4: In my tests YeAH dramatically slowed down transfers over a WLAN,
|
||||||
|
reducing throughput from ~65Mbps (CUBIC) to ~7MBps (YeAH) over 10
|
||||||
|
seconds (netperf TCP_STREAM) including long stalls.
|
||||||
|
|
||||||
|
Be careful when choosing this. ~heftig
|
||||||
|
---
|
||||||
|
net/ipv4/Kconfig | 4 ++++
|
||||||
|
1 file changed, 4 insertions(+)
|
||||||
|
|
||||||
|
diff --git a/net/ipv4/Kconfig b/net/ipv4/Kconfig
|
||||||
|
index e64e59b536d3..bfb55ef7ebbe 100644
|
||||||
|
--- a/net/ipv4/Kconfig
|
||||||
|
+++ b/net/ipv4/Kconfig
|
||||||
|
@@ -691,6 +691,9 @@ choice
|
||||||
|
config DEFAULT_VEGAS
|
||||||
|
bool "Vegas" if TCP_CONG_VEGAS=y
|
||||||
|
|
||||||
|
+ config DEFAULT_YEAH
|
||||||
|
+ bool "YeAH" if TCP_CONG_YEAH=y
|
||||||
|
+
|
||||||
|
config DEFAULT_VENO
|
||||||
|
bool "Veno" if TCP_CONG_VENO=y
|
||||||
|
|
||||||
|
@@ -724,6 +727,7 @@ config DEFAULT_TCP_CONG
|
||||||
|
default "htcp" if DEFAULT_HTCP
|
||||||
|
default "hybla" if DEFAULT_HYBLA
|
||||||
|
default "vegas" if DEFAULT_VEGAS
|
||||||
|
+ default "yeah" if DEFAULT_YEAH
|
||||||
|
default "westwood" if DEFAULT_WESTWOOD
|
||||||
|
default "veno" if DEFAULT_VENO
|
||||||
|
default "reno" if DEFAULT_RENO
|
||||||
|
--
|
||||||
|
2.28.0
|
||||||
|
|
||||||
|
|
||||||
|
From 76dbe7477bfde1b5e8bf29a71b5af7ab2be9b98e Mon Sep 17 00:00:00 2001
|
||||||
|
From: Steven Barrett <steven@liquorix.net>
|
||||||
|
Date: Wed, 28 Nov 2018 19:01:27 -0600
|
||||||
|
Subject: [PATCH 09/17] zen: Use [defer+madvise] as default khugepaged defrag
|
||||||
|
strategy
|
||||||
|
|
||||||
|
For some reason, the default strategy to respond to THP fault fallbacks
|
||||||
|
is still just madvise, meaning stall if the program wants transparent
|
||||||
|
hugepages, but don't trigger a background reclaim / compaction if THP
|
||||||
|
begins to fail allocations. This creates a snowball affect where we
|
||||||
|
still use the THP code paths, but we almost always fail once a system
|
||||||
|
has been active and busy for a while.
|
||||||
|
|
||||||
|
The option "defer" was created for interactive systems where THP can
|
||||||
|
still improve performance. If we have to fallback to a regular page due
|
||||||
|
to an allocation failure or anything else, we will trigger a background
|
||||||
|
reclaim and compaction so future THP attempts succeed and previous
|
||||||
|
attempts eventually have their smaller pages combined without stalling
|
||||||
|
running applications.
|
||||||
|
|
||||||
|
We still want madvise to stall applications that explicitely want THP,
|
||||||
|
so defer+madvise _does_ make a ton of sense. Make it the default for
|
||||||
|
interactive systems, especially if the kernel maintainer left
|
||||||
|
transparent hugepages on "always".
|
||||||
|
|
||||||
|
Reasoning and details in the original patch: https://lwn.net/Articles/711248/
|
||||||
|
---
|
||||||
|
mm/huge_memory.c | 4 ++++
|
||||||
|
1 file changed, 4 insertions(+)
|
||||||
|
|
||||||
|
diff --git a/mm/huge_memory.c b/mm/huge_memory.c
|
||||||
|
index 74300e337c3c..9277f22c10a7 100644
|
||||||
|
--- a/mm/huge_memory.c
|
||||||
|
+++ b/mm/huge_memory.c
|
||||||
|
@@ -53,7 +53,11 @@ unsigned long transparent_hugepage_flags __read_mostly =
|
||||||
|
#ifdef CONFIG_TRANSPARENT_HUGEPAGE_MADVISE
|
||||||
|
(1<<TRANSPARENT_HUGEPAGE_REQ_MADV_FLAG)|
|
||||||
|
#endif
|
||||||
|
+#ifdef CONFIG_ZENIFY
|
||||||
|
+ (1<<TRANSPARENT_HUGEPAGE_DEFRAG_KSWAPD_OR_MADV_FLAG)|
|
||||||
|
+#else
|
||||||
|
(1<<TRANSPARENT_HUGEPAGE_DEFRAG_REQ_MADV_FLAG)|
|
||||||
|
+#endif
|
||||||
|
(1<<TRANSPARENT_HUGEPAGE_DEFRAG_KHUGEPAGED_FLAG)|
|
||||||
|
(1<<TRANSPARENT_HUGEPAGE_USE_ZERO_PAGE_FLAG);
|
||||||
|
|
||||||
|
--
|
||||||
|
2.28.0
|
||||||
|
|
||||||
|
|
||||||
|
From 2b65a1329cb220b43c19c4d0de5833fae9e2b22d Mon Sep 17 00:00:00 2001
|
||||||
|
From: Alexandre Frade <admfrade@gmail.com>
|
||||||
|
Date: Wed, 24 Oct 2018 16:58:52 -0300
|
||||||
|
Subject: [PATCH 10/17] net/sched: allow configuring cake qdisc as default
|
||||||
|
|
||||||
|
Signed-off-by: Alexandre Frade <admfrade@gmail.com>
|
||||||
|
---
|
||||||
|
net/sched/Kconfig | 4 ++++
|
||||||
|
1 file changed, 4 insertions(+)
|
||||||
|
|
||||||
|
diff --git a/net/sched/Kconfig b/net/sched/Kconfig
|
||||||
|
index 84badf00647e..6a922bca9f39 100644
|
||||||
|
--- a/net/sched/Kconfig
|
||||||
|
+++ b/net/sched/Kconfig
|
||||||
|
@@ -471,6 +471,9 @@ choice
|
||||||
|
config DEFAULT_SFQ
|
||||||
|
bool "Stochastic Fair Queue" if NET_SCH_SFQ
|
||||||
|
|
||||||
|
+ config DEFAULT_CAKE
|
||||||
|
+ bool "Common Applications Kept Enhanced" if NET_SCH_CAKE
|
||||||
|
+
|
||||||
|
config DEFAULT_PFIFO_FAST
|
||||||
|
bool "Priority FIFO Fast"
|
||||||
|
endchoice
|
||||||
|
@@ -481,6 +484,7 @@ config DEFAULT_NET_SCH
|
||||||
|
default "fq" if DEFAULT_FQ
|
||||||
|
default "fq_codel" if DEFAULT_FQ_CODEL
|
||||||
|
default "sfq" if DEFAULT_SFQ
|
||||||
|
+ default "cake" if DEFAULT_CAKE
|
||||||
|
default "pfifo_fast"
|
||||||
|
endif
|
||||||
|
|
||||||
|
--
|
||||||
|
2.28.0
|
||||||
|
|
||||||
|
|
||||||
|
From 90240bcd90a568878738e66c0d45bed3e38e347b Mon Sep 17 00:00:00 2001
|
||||||
|
From: Tk-Glitch <ti3nou@gmail.com>
|
||||||
|
Date: Fri, 19 Apr 2019 12:33:38 +0200
|
||||||
|
Subject: [PATCH 12/17] Set vm.max_map_count to 262144 by default
|
||||||
|
|
||||||
|
The value is still pretty low, and AMD64-ABI and ELF extended numbering
|
||||||
|
supports that, so we should be fine on modern x86 systems.
|
||||||
|
|
||||||
|
This fixes crashes in some applications using more than 65535 vmas (also
|
||||||
|
affects some windows games running in wine, such as Star Citizen).
|
||||||
|
---
|
||||||
|
include/linux/mm.h | 3 +--
|
||||||
|
1 file changed, 1 insertion(+), 2 deletions(-)
|
||||||
|
|
||||||
|
diff --git a/include/linux/mm.h b/include/linux/mm.h
|
||||||
|
index bc05c3588aa3..b0cefe94920d 100644
|
||||||
|
--- a/include/linux/mm.h
|
||||||
|
+++ b/include/linux/mm.h
|
||||||
|
@@ -190,8 +190,7 @@ static inline void __mm_zero_struct_page(struct page *page)
|
||||||
|
* not a hard limit any more. Although some userspace tools can be surprised by
|
||||||
|
* that.
|
||||||
|
*/
|
||||||
|
-#define MAPCOUNT_ELF_CORE_MARGIN (5)
|
||||||
|
-#define DEFAULT_MAX_MAP_COUNT (USHRT_MAX - MAPCOUNT_ELF_CORE_MARGIN)
|
||||||
|
+#define DEFAULT_MAX_MAP_COUNT (262144)
|
||||||
|
|
||||||
|
extern int sysctl_max_map_count;
|
||||||
|
|
||||||
|
--
|
||||||
|
2.28.0
|
||||||
|
|
||||||
|
|
||||||
|
From 3a34034dba5efe91bcec491efe8c66e8087f509b Mon Sep 17 00:00:00 2001
|
||||||
|
From: Tk-Glitch <ti3nou@gmail.com>
|
||||||
|
Date: Mon, 27 Jul 2020 00:19:18 +0200
|
||||||
|
Subject: [PATCH 13/17] mm: bump DEFAULT_MAX_MAP_COUNT
|
||||||
|
|
||||||
|
Some games such as Detroit: Become Human tend to be very crash prone with
|
||||||
|
lower values.
|
||||||
|
---
|
||||||
|
include/linux/mm.h | 2 +-
|
||||||
|
1 file changed, 1 insertion(+), 1 deletion(-)
|
||||||
|
|
||||||
|
diff --git a/include/linux/mm.h b/include/linux/mm.h
|
||||||
|
index b0cefe94920d..890165099b07 100644
|
||||||
|
--- a/include/linux/mm.h
|
||||||
|
+++ b/include/linux/mm.h
|
||||||
|
@@ -190,7 +190,7 @@ static inline void __mm_zero_struct_page(struct page *page)
|
||||||
|
* not a hard limit any more. Although some userspace tools can be surprised by
|
||||||
|
* that.
|
||||||
|
*/
|
||||||
|
-#define DEFAULT_MAX_MAP_COUNT (262144)
|
||||||
|
+#define DEFAULT_MAX_MAP_COUNT (16777216)
|
||||||
|
|
||||||
|
extern int sysctl_max_map_count;
|
||||||
|
|
||||||
|
--
|
||||||
|
2.28.0
|
||||||
|
|
||||||
|
From 977812938da7c7226415778c340832141d9278b7 Mon Sep 17 00:00:00 2001
|
||||||
|
From: Alexandre Frade <admfrade@gmail.com>
|
||||||
|
Date: Mon, 25 Nov 2019 15:13:06 -0300
|
||||||
|
Subject: [PATCH 14/17] elevator: set default scheduler to bfq for blk-mq
|
||||||
|
|
||||||
|
Signed-off-by: Alexandre Frade <admfrade@gmail.com>
|
||||||
|
---
|
||||||
|
block/elevator.c | 6 +++---
|
||||||
|
1 file changed, 3 insertions(+), 3 deletions(-)
|
||||||
|
|
||||||
|
diff --git a/block/elevator.c b/block/elevator.c
|
||||||
|
index 4eab3d70e880..79669aa39d79 100644
|
||||||
|
--- a/block/elevator.c
|
||||||
|
+++ b/block/elevator.c
|
||||||
|
@@ -623,19 +623,19 @@ static inline bool elv_support_iosched(struct request_queue *q)
|
||||||
|
}
|
||||||
|
|
||||||
|
/*
|
||||||
|
- * For single queue devices, default to using mq-deadline. If we have multiple
|
||||||
|
- * queues or mq-deadline is not available, default to "none".
|
||||||
|
+ * For single queue devices, default to using bfq. If we have multiple
|
||||||
|
+ * queues or bfq is not available, default to "none".
|
||||||
|
*/
|
||||||
|
static struct elevator_type *elevator_get_default(struct request_queue *q)
|
||||||
|
{
|
||||||
|
if (q->tag_set && q->tag_set->flags & BLK_MQ_F_NO_SCHED_BY_DEFAULT)
|
||||||
|
return NULL;
|
||||||
|
|
||||||
|
if (q->nr_hw_queues != 1 &&
|
||||||
|
!blk_mq_is_shared_tags(q->tag_set->flags))
|
||||||
|
return NULL;
|
||||||
|
|
||||||
|
- return elevator_find_get(q, "mq-deadline");
|
||||||
|
+ return elevator_find_get(q, "bfq");
|
||||||
|
}
|
||||||
|
|
||||||
|
/*
|
||||||
|
--
|
||||||
|
2.28.0
|
||||||
|
|
||||||
|
From 3c229f434aca65c4ca61772bc03c3e0370817b92 Mon Sep 17 00:00:00 2001
|
||||||
|
From: Alexandre Frade <kernel@xanmod.org>
|
||||||
|
Date: Mon, 3 Aug 2020 17:05:04 +0000
|
||||||
|
Subject: [PATCH 16/17] mm: set 2 megabytes for address_space-level file
|
||||||
|
read-ahead pages size
|
||||||
|
|
||||||
|
Signed-off-by: Alexandre Frade <kernel@xanmod.org>
|
||||||
|
---
|
||||||
|
include/linux/pagemap.h | 2 +-
|
||||||
|
1 file changed, 1 insertion(+), 1 deletion(-)
|
||||||
|
|
||||||
|
diff --git a/include/linux/pagemap.h b/include/linux/pagemap.h
|
||||||
|
index cf2468da68e9..007dea784451 100644
|
||||||
|
--- a/include/linux/pagemap.h
|
||||||
|
+++ b/include/linux/pagemap.h
|
||||||
|
@@ -655,7 +655,7 @@ int replace_page_cache_page(struct page *old, struct page *new, gfp_t gfp_mask);
|
||||||
|
void delete_from_page_cache_batch(struct address_space *mapping,
|
||||||
|
struct pagevec *pvec);
|
||||||
|
|
||||||
|
-#define VM_READAHEAD_PAGES (SZ_128K / PAGE_SIZE)
|
||||||
|
+#define VM_READAHEAD_PAGES (SZ_2M / PAGE_SIZE)
|
||||||
|
|
||||||
|
void page_cache_sync_readahead(struct address_space *, struct file_ra_state *,
|
||||||
|
struct file *, pgoff_t index, unsigned long req_count);
|
||||||
|
--
|
||||||
|
2.28.0
|
||||||
|
|
||||||
|
|
||||||
|
From 716f41cf6631f3a85834dcb67b4ce99185b6387f Mon Sep 17 00:00:00 2001
|
||||||
|
From: Steven Barrett <steven@liquorix.net>
|
||||||
|
Date: Wed, 15 Jan 2020 20:43:56 -0600
|
||||||
|
Subject: [PATCH 17/17] ZEN: intel-pstate: Implement "enable" parameter
|
||||||
|
|
||||||
|
If intel-pstate is compiled into the kernel, it will preempt the loading
|
||||||
|
of acpi-cpufreq so you can take advantage of hardware p-states without
|
||||||
|
any friction.
|
||||||
|
|
||||||
|
However, intel-pstate is not completely superior to cpufreq's ondemand
|
||||||
|
for one reason. There's no concept of an up_threshold property.
|
||||||
|
|
||||||
|
In ondemand, up_threshold essentially reduces the maximum utilization to
|
||||||
|
compare against, allowing you to hit max frequencies and turbo boost
|
||||||
|
from a much lower core utilization.
|
||||||
|
|
||||||
|
With intel-pstate, you have the concept of minimum and maximum
|
||||||
|
performance, but no tunable that lets you define, maximum frequency
|
||||||
|
means 50% core utilization. For just this oversight, there's reasons
|
||||||
|
you may want ondemand.
|
||||||
|
|
||||||
|
Lets support setting "enable" in kernel boot parameters. This lets
|
||||||
|
kernel maintainers include "intel_pstate=disable" statically in the
|
||||||
|
static boot parameters, but let users of the kernel override this
|
||||||
|
selection.
|
||||||
|
---
|
||||||
|
Documentation/admin-guide/kernel-parameters.txt | 3 +++
|
||||||
|
drivers/cpufreq/intel_pstate.c | 2 ++
|
||||||
|
2 files changed, 5 insertions(+)
|
||||||
|
|
||||||
|
diff --git a/Documentation/admin-guide/kernel-parameters.txt b/Documentation/admin-guide/kernel-parameters.txt
|
||||||
|
index fb95fad81c79..3e92fee81e33 100644
|
||||||
|
--- a/Documentation/admin-guide/kernel-parameters.txt
|
||||||
|
+++ b/Documentation/admin-guide/kernel-parameters.txt
|
||||||
|
@@ -1857,6 +1857,9 @@
|
||||||
|
disable
|
||||||
|
Do not enable intel_pstate as the default
|
||||||
|
scaling driver for the supported processors
|
||||||
|
+ enable
|
||||||
|
+ Enable intel_pstate in-case "disable" was passed
|
||||||
|
+ previously in the kernel boot parameters
|
||||||
|
passive
|
||||||
|
Use intel_pstate as a scaling driver, but configure it
|
||||||
|
to work with generic cpufreq governors (instead of
|
||||||
|
diff --git a/drivers/cpufreq/intel_pstate.c b/drivers/cpufreq/intel_pstate.c
|
||||||
|
index 36a469150ff9..aee891c9b78a 100644
|
||||||
|
--- a/drivers/cpufreq/intel_pstate.c
|
||||||
|
+++ b/drivers/cpufreq/intel_pstate.c
|
||||||
|
@@ -2845,6 +2845,8 @@ static int __init intel_pstate_setup(char *str)
|
||||||
|
if (!strcmp(str, "no_hwp"))
|
||||||
|
no_hwp = 1;
|
||||||
|
|
||||||
|
+ if (!strcmp(str, "enable"))
|
||||||
|
+ no_load = 0;
|
||||||
|
if (!strcmp(str, "force"))
|
||||||
|
force_load = 1;
|
||||||
|
if (!strcmp(str, "hwp_only"))
|
||||||
|
--
|
||||||
|
2.28.0
|
||||||
|
|
||||||
|
From 379cbab18b5c75c622b93e2c5abdfac141fe9654 Mon Sep 17 00:00:00 2001
|
||||||
|
From: Kenny Levinsen <kl@kl.wtf>
|
||||||
|
Date: Sun, 27 Dec 2020 14:43:13 +0000
|
||||||
|
Subject: [PATCH] ZEN: Input: evdev - use call_rcu when detaching client
|
||||||
|
|
||||||
|
Significant time was spent on synchronize_rcu in evdev_detach_client
|
||||||
|
when applications closed evdev devices. Switching VT away from a
|
||||||
|
graphical environment commonly leads to mass input device closures,
|
||||||
|
which could lead to noticable delays on systems with many input devices.
|
||||||
|
|
||||||
|
Replace synchronize_rcu with call_rcu, deferring reclaim of the evdev
|
||||||
|
client struct till after the RCU grace period instead of blocking the
|
||||||
|
calling application.
|
||||||
|
|
||||||
|
While this does not solve all slow evdev fd closures, it takes care of a
|
||||||
|
good portion of them, including this simple test:
|
||||||
|
|
||||||
|
#include <fcntl.h>
|
||||||
|
#include <unistd.h>
|
||||||
|
|
||||||
|
int main(int argc, char *argv[])
|
||||||
|
{
|
||||||
|
int idx, fd;
|
||||||
|
const char *path = "/dev/input/event0";
|
||||||
|
for (idx = 0; idx < 1000; idx++) {
|
||||||
|
if ((fd = open(path, O_RDWR)) == -1) {
|
||||||
|
return -1;
|
||||||
|
}
|
||||||
|
close(fd);
|
||||||
|
}
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
Time to completion of above test when run locally:
|
||||||
|
|
||||||
|
Before: 0m27.111s
|
||||||
|
After: 0m0.018s
|
||||||
|
|
||||||
|
Signed-off-by: Kenny Levinsen <kl@kl.wtf>
|
||||||
|
---
|
||||||
|
drivers/input/evdev.c | 19 +++++++++++--------
|
||||||
|
1 file changed, 11 insertions(+), 8 deletions(-)
|
||||||
|
|
||||||
|
diff --git a/drivers/input/evdev.c b/drivers/input/evdev.c
|
||||||
|
index 95f90699d2b17b..2b10fe29d2c8d9 100644
|
||||||
|
--- a/drivers/input/evdev.c
|
||||||
|
+++ b/drivers/input/evdev.c
|
||||||
|
@@ -46,6 +46,7 @@ struct evdev_client {
|
||||||
|
struct fasync_struct *fasync;
|
||||||
|
struct evdev *evdev;
|
||||||
|
struct list_head node;
|
||||||
|
+ struct rcu_head rcu;
|
||||||
|
enum input_clock_type clk_type;
|
||||||
|
bool revoked;
|
||||||
|
unsigned long *evmasks[EV_CNT];
|
||||||
|
@@ -377,13 +378,22 @@ static void evdev_attach_client(struct evdev *evdev,
|
||||||
|
spin_unlock(&evdev->client_lock);
|
||||||
|
}
|
||||||
|
|
||||||
|
+static void evdev_reclaim_client(struct rcu_head *rp)
|
||||||
|
+{
|
||||||
|
+ struct evdev_client *client = container_of(rp, struct evdev_client, rcu);
|
||||||
|
+ unsigned int i;
|
||||||
|
+ for (i = 0; i < EV_CNT; ++i)
|
||||||
|
+ bitmap_free(client->evmasks[i]);
|
||||||
|
+ kvfree(client);
|
||||||
|
+}
|
||||||
|
+
|
||||||
|
static void evdev_detach_client(struct evdev *evdev,
|
||||||
|
struct evdev_client *client)
|
||||||
|
{
|
||||||
|
spin_lock(&evdev->client_lock);
|
||||||
|
list_del_rcu(&client->node);
|
||||||
|
spin_unlock(&evdev->client_lock);
|
||||||
|
- synchronize_rcu();
|
||||||
|
+ call_rcu(&client->rcu, evdev_reclaim_client);
|
||||||
|
}
|
||||||
|
|
||||||
|
static int evdev_open_device(struct evdev *evdev)
|
||||||
|
@@ -436,7 +446,6 @@ static int evdev_release(struct inode *inode, struct file *file)
|
||||||
|
{
|
||||||
|
struct evdev_client *client = file->private_data;
|
||||||
|
struct evdev *evdev = client->evdev;
|
||||||
|
- unsigned int i;
|
||||||
|
|
||||||
|
mutex_lock(&evdev->mutex);
|
||||||
|
|
||||||
|
@@ -448,11 +457,6 @@ static int evdev_release(struct inode *inode, struct file *file)
|
||||||
|
|
||||||
|
evdev_detach_client(evdev, client);
|
||||||
|
|
||||||
|
- for (i = 0; i < EV_CNT; ++i)
|
||||||
|
- bitmap_free(client->evmasks[i]);
|
||||||
|
-
|
||||||
|
- kvfree(client);
|
||||||
|
-
|
||||||
|
evdev_close_device(evdev);
|
||||||
|
|
||||||
|
return 0;
|
||||||
|
@@ -495,7 +499,6 @@ static int evdev_open(struct inode *inode, struct file *file)
|
||||||
|
|
||||||
|
err_free_client:
|
||||||
|
evdev_detach_client(evdev, client);
|
||||||
|
- kvfree(client);
|
||||||
|
return error;
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
From 2aafb56f20e4b63d8c4af172fe9d017c64bc4129 Mon Sep 17 00:00:00 2001
|
||||||
|
From: Sultan Alsawaf <sultan@kerneltoast.com>
|
||||||
|
Date: Wed, 20 Oct 2021 20:50:11 -0700
|
||||||
|
Subject: [PATCH] ZEN: mm: Lower the non-hugetlbpage pageblock size to reduce
|
||||||
|
scheduling delays
|
||||||
|
|
||||||
|
The page allocator processes free pages in groups of pageblocks, where
|
||||||
|
the size of a pageblock is typically quite large (1024 pages without
|
||||||
|
hugetlbpage support). Pageblocks are processed atomically with the zone
|
||||||
|
lock held, which can cause severe scheduling delays on both the CPU
|
||||||
|
going through the pageblock and any other CPUs waiting to acquire the
|
||||||
|
zone lock. A frequent offender is move_freepages_block(), which is used
|
||||||
|
by rmqueue() for page allocation.
|
||||||
|
|
||||||
|
As it turns out, there's no requirement for pageblocks to be so large,
|
||||||
|
so the pageblock order can simply be reduced to ease the scheduling
|
||||||
|
delays and zone lock contention. PAGE_ALLOC_COSTLY_ORDER is used as a
|
||||||
|
reasonable setting to ensure non-costly page allocation requests can
|
||||||
|
still be serviced without always needing to free up more than one
|
||||||
|
pageblock's worth of pages at a time.
|
||||||
|
|
||||||
|
This has a noticeable effect on overall system latency when memory
|
||||||
|
pressure is elevated. The various mm functions which operate on
|
||||||
|
pageblocks no longer appear in the preemptoff tracer, where previously
|
||||||
|
they would spend up to 100 ms on a mobile arm64 CPU processing a
|
||||||
|
pageblock with preemption disabled and the zone lock held.
|
||||||
|
|
||||||
|
Signed-off-by: Sultan Alsawaf <sultan@kerneltoast.com>
|
||||||
|
---
|
||||||
|
include/linux/pageblock-flags.h | 2 +-
|
||||||
|
1 file changed, 1 insertion(+), 1 deletion(-)
|
||||||
|
|
||||||
|
diff --git a/include/linux/pageblock-flags.h b/include/linux/pageblock-flags.h
|
||||||
|
index 5f1ae07d724b88..97cda629c9e909 100644
|
||||||
|
--- a/include/linux/pageblock-flags.h
|
||||||
|
+++ b/include/linux/pageblock-flags.h
|
||||||
|
@@ -48,7 +48,7 @@ extern unsigned int pageblock_order;
|
||||||
|
#else /* CONFIG_HUGETLB_PAGE */
|
||||||
|
|
||||||
|
/* If huge pages are not used, group by MAX_ORDER_NR_PAGES */
|
||||||
|
-#define pageblock_order (MAX_ORDER-1)
|
||||||
|
+#define pageblock_order PAGE_ALLOC_COSTLY_ORDER
|
||||||
|
|
||||||
|
#endif /* CONFIG_HUGETLB_PAGE */
|
||||||
|
|
||||||
|
|
||||||
|
From f22bc56be85e69c71c8e36041193856bb8b01525 Mon Sep 17 00:00:00 2001
|
||||||
|
From: Sultan Alsawaf <sultan@kerneltoast.com>
|
||||||
|
Date: Wed, 20 Oct 2021 20:50:32 -0700
|
||||||
|
Subject: [PATCH] ZEN: mm: Don't hog the CPU and zone lock in rmqueue_bulk()
|
||||||
|
|
||||||
|
There is noticeable scheduling latency and heavy zone lock contention
|
||||||
|
stemming from rmqueue_bulk's single hold of the zone lock while doing
|
||||||
|
its work, as seen with the preemptoff tracer. There's no actual need for
|
||||||
|
rmqueue_bulk() to hold the zone lock the entire time; it only does so
|
||||||
|
for supposed efficiency. As such, we can relax the zone lock and even
|
||||||
|
reschedule when IRQs are enabled in order to keep the scheduling delays
|
||||||
|
and zone lock contention at bay. Forward progress is still guaranteed,
|
||||||
|
as the zone lock can only be relaxed after page removal.
|
||||||
|
|
||||||
|
With this change, rmqueue_bulk() no longer appears as a serious offender
|
||||||
|
in the preemptoff tracer, and system latency is noticeably improved.
|
||||||
|
|
||||||
|
Signed-off-by: Sultan Alsawaf <sultan@kerneltoast.com>
|
||||||
|
---
|
||||||
|
mm/page_alloc.c | 23 ++++++++++++++++++-----
|
||||||
|
1 file changed, 18 insertions(+), 5 deletions(-)
|
||||||
|
|
||||||
|
diff --git a/mm/page_alloc.c b/mm/page_alloc.c
|
||||||
|
index a0b0397e29ee4c..87a983a356530c 100644
|
||||||
|
--- a/mm/page_alloc.c
|
||||||
|
+++ b/mm/page_alloc.c
|
||||||
|
@@ -3118,15 +3119,16 @@ __rmqueue(struct zone *zone, unsigned int order, int migratetype,
|
||||||
|
}
|
||||||
|
|
||||||
|
/*
|
||||||
|
- * Obtain a specified number of elements from the buddy allocator, all under
|
||||||
|
- * a single hold of the lock, for efficiency. Add them to the supplied list.
|
||||||
|
- * Returns the number of new pages which were placed at *list.
|
||||||
|
+ * Obtain a specified number of elements from the buddy allocator, and relax the
|
||||||
|
+ * zone lock when needed. Add them to the supplied list. Returns the number of
|
||||||
|
+ * new pages which were placed at *list.
|
||||||
|
*/
|
||||||
|
static int rmqueue_bulk(struct zone *zone, unsigned int order,
|
||||||
|
unsigned long count, struct list_head *list,
|
||||||
|
int migratetype, unsigned int alloc_flags)
|
||||||
|
{
|
||||||
|
unsigned long flags;
|
||||||
|
- int i, allocated = 0;
|
||||||
|
+ const bool can_resched = !preempt_count() && !irqs_disabled();
|
||||||
|
+ int i, allocated = 0, last_mod = 0;
|
||||||
|
|
||||||
|
/* Caller must hold IRQ-safe pcp->lock so IRQs are disabled. */
|
||||||
|
spin_lock(&zone->lock);
|
||||||
|
@@ -3137,6 +3138,18 @@ static int rmqueue_bulk(struct zone *zone, unsigned int order,
|
||||||
|
if (unlikely(page == NULL))
|
||||||
|
break;
|
||||||
|
|
||||||
|
+ /* Reschedule and ease the contention on the lock if needed */
|
||||||
|
+ if (i + 1 < count && ((can_resched && need_resched()) ||
|
||||||
|
+ spin_needbreak(&zone->lock))) {
|
||||||
|
+ __mod_zone_page_state(zone, NR_FREE_PAGES,
|
||||||
|
+ -((i + 1 - last_mod) << order));
|
||||||
|
+ last_mod = i + 1;
|
||||||
|
+ spin_unlock(&zone->lock);
|
||||||
|
+ if (can_resched)
|
||||||
|
+ cond_resched();
|
||||||
|
+ spin_lock(&zone->lock);
|
||||||
|
+ }
|
||||||
|
+
|
||||||
|
if (unlikely(check_pcp_refill(page, order)))
|
||||||
|
continue;
|
||||||
|
|
||||||
|
@@ -3163,7 +3176,7 @@ static int rmqueue_bulk(struct zone *zone, unsigned int order,
|
||||||
|
* on i. Do not confuse with 'allocated' which is the number of
|
||||||
|
* pages added to the pcp list.
|
||||||
|
*/
|
||||||
|
- __mod_zone_page_state(zone, NR_FREE_PAGES, -(i << order));
|
||||||
|
+ __mod_zone_page_state(zone, NR_FREE_PAGES, -((i - last_mod) << order));
|
||||||
|
spin_unlock(&zone->lock);
|
||||||
|
return allocated;
|
||||||
|
}
|
||||||
|
|
||||||
|
From 6329525a0fa10cd13f39b76948b1296150f75c95 Mon Sep 17 00:00:00 2001
|
||||||
|
From: Alexandre Frade <kernel@xanmod.org>
|
||||||
|
Date: Mon, 29 Aug 2022 16:47:26 +0000
|
||||||
|
Subject: [PATCH 14/16] XANMOD: Makefile: Disable GCC vectorization on trees
|
||||||
|
|
||||||
|
Signed-off-by: Alexandre Frade <kernel@xanmod.org>
|
||||||
|
---
|
||||||
|
Makefile | 3 +++
|
||||||
|
1 file changed, 3 insertions(+)
|
||||||
|
|
||||||
|
diff --git a/Makefile b/Makefile
|
||||||
|
index 3f6628780eb2..35a5ae1ede42 100644
|
||||||
|
--- a/Makefile
|
||||||
|
+++ b/Makefile
|
||||||
|
@@ -1069,6 +1069,9 @@ endif
|
||||||
|
KBUILD_CFLAGS-$(call gcc-min-version, 90100) += -Wno-alloc-size-larger-than
|
||||||
|
KBUILD_CFLAGS += $(KBUILD_CFLAGS-y) $(CONFIG_CC_IMPLICIT_FALLTHROUGH)
|
||||||
|
|
||||||
|
+# disable GCC vectorization on trees
|
||||||
|
+KBUILD_CFLAGS += $(call cc-option, -fno-tree-vectorize)
|
||||||
|
+
|
||||||
|
# disable invalid "can't wrap" optimizations for signed / pointers
|
||||||
|
KBUILD_CFLAGS += -fno-strict-overflow
|
||||||
|
|
||||||
|
--
|
||||||
|
2.39.1
|
||||||
|
|
||||||
|
From f997578464b2c4c63e7bd1afbfef56212ee44f2d Mon Sep 17 00:00:00 2001
|
||||||
|
From: Etienne JUVIGNY <ti3nou@gmail.com>
|
||||||
|
Date: Mon, 6 Mar 2023 13:54:09 +0100
|
||||||
|
Subject: Don't add -dirty versioning on unclean trees
|
||||||
|
|
||||||
|
|
||||||
|
diff --git a/scripts/setlocalversion b/scripts/setlocalversion
|
||||||
|
index ca5795e16..ad0d94477 100755
|
||||||
|
--- a/scripts/setlocalversion
|
||||||
|
+++ b/scripts/setlocalversion
|
||||||
|
@@ -85,12 +85,12 @@ scm_version()
|
||||||
|
# git-diff-index does not refresh the index, so it may give misleading
|
||||||
|
# results.
|
||||||
|
# See git-update-index(1), git-diff-index(1), and git-status(1).
|
||||||
|
- if {
|
||||||
|
- git --no-optional-locks status -uno --porcelain 2>/dev/null ||
|
||||||
|
- git diff-index --name-only HEAD
|
||||||
|
- } | read dummy; then
|
||||||
|
- printf '%s' -dirty
|
||||||
|
- fi
|
||||||
|
+ #if {
|
||||||
|
+ # git --no-optional-locks status -uno --porcelain 2>/dev/null ||
|
||||||
|
+ # git diff-index --name-only HEAD
|
||||||
|
+ #} | read dummy; then
|
||||||
|
+ # printf '%s' -dirty
|
||||||
|
+ #fi
|
||||||
|
}
|
||||||
|
|
||||||
|
collect_files()
|
317
linux-tkg-patches/6.3/0003-glitched-cfs-additions.patch
Normal file
317
linux-tkg-patches/6.3/0003-glitched-cfs-additions.patch
Normal file
@@ -0,0 +1,317 @@
|
|||||||
|
diff --git a/kernel/sched/fair.c b/kernel/sched/fair.c
|
||||||
|
index 6b3b59cc51d6..2a0072192c3d 100644
|
||||||
|
--- a/kernel/sched/fair.c
|
||||||
|
+++ b/kernel/sched/fair.c
|
||||||
|
@@ -81,10 +95,17 @@ unsigned int sysctl_sched_child_runs_first __read_mostly;
|
||||||
|
*
|
||||||
|
* (default: 1 msec * (1 + ilog(ncpus)), units: nanoseconds)
|
||||||
|
*/
|
||||||
|
+#ifdef CONFIG_ZENIFY
|
||||||
|
+unsigned int sysctl_sched_wakeup_granularity = 500000UL;
|
||||||
|
+static unsigned int normalized_sysctl_sched_wakeup_granularity = 500000UL;
|
||||||
|
+
|
||||||
|
+const_debug unsigned int sysctl_sched_migration_cost = 50000UL;
|
||||||
|
+#else
|
||||||
|
unsigned int sysctl_sched_wakeup_granularity = 1000000UL;
|
||||||
|
static unsigned int normalized_sysctl_sched_wakeup_granularity = 1000000UL;
|
||||||
|
|
||||||
|
const_debug unsigned int sysctl_sched_migration_cost = 500000UL;
|
||||||
|
+#endif
|
||||||
|
|
||||||
|
int sched_thermal_decay_shift;
|
||||||
|
static int __init setup_sched_thermal_decay_shift(char *str)
|
||||||
|
|
||||||
|
From 5d5b708e3731e135ea7ae168571ad78d883e63e8 Mon Sep 17 00:00:00 2001
|
||||||
|
From: Alexandre Frade <kernel@xanmod.org>
|
||||||
|
Date: Wed, 1 Feb 2023 10:17:47 +0000
|
||||||
|
Subject: [PATCH 02/16] XANMOD: fair: Remove all energy efficiency functions
|
||||||
|
|
||||||
|
Signed-off-by: Alexandre Frade <kernel@xanmod.org>
|
||||||
|
---
|
||||||
|
kernel/sched/fair.c | 224 +-------------------------------------------
|
||||||
|
1 file changed, 3 insertions(+), 221 deletions(-)
|
||||||
|
|
||||||
|
diff --git a/kernel/sched/fair.c b/kernel/sched/fair.c
|
||||||
|
index 0f8736991427..345cc5e9fa6e 100644
|
||||||
|
--- a/kernel/sched/fair.c
|
||||||
|
+++ b/kernel/sched/fair.c
|
||||||
|
@@ -19,6 +19,9 @@
|
||||||
|
*
|
||||||
|
* Adaptive scheduling granularity, math enhancements by Peter Zijlstra
|
||||||
|
* Copyright (C) 2007 Red Hat, Inc., Peter Zijlstra
|
||||||
|
+ *
|
||||||
|
+ * Remove energy efficiency functions by Alexandre Frade
|
||||||
|
+ * (C) 2021 Alexandre Frade <kernel@xanmod.org>
|
||||||
|
*/
|
||||||
|
#include <linux/energy_model.h>
|
||||||
|
#include <linux/mmap_lock.h>
|
||||||
|
@@ -7327,252 +7327,6 @@ eenv_pd_max_util(struct energy_env *eenv, struct cpumask *pd_cpus,
|
||||||
|
return min(max_util, eenv->cpu_cap);
|
||||||
|
}
|
||||||
|
|
||||||
|
-/*
|
||||||
|
- * compute_energy(): Use the Energy Model to estimate the energy that @pd would
|
||||||
|
- * consume for a given utilization landscape @eenv. When @dst_cpu < 0, the task
|
||||||
|
- * contribution is ignored.
|
||||||
|
- */
|
||||||
|
-static inline unsigned long
|
||||||
|
-compute_energy(struct energy_env *eenv, struct perf_domain *pd,
|
||||||
|
- struct cpumask *pd_cpus, struct task_struct *p, int dst_cpu)
|
||||||
|
-{
|
||||||
|
- unsigned long max_util = eenv_pd_max_util(eenv, pd_cpus, p, dst_cpu);
|
||||||
|
- unsigned long busy_time = eenv->pd_busy_time;
|
||||||
|
-
|
||||||
|
- if (dst_cpu >= 0)
|
||||||
|
- busy_time = min(eenv->pd_cap, busy_time + eenv->task_busy_time);
|
||||||
|
-
|
||||||
|
- return em_cpu_energy(pd->em_pd, max_util, busy_time, eenv->cpu_cap);
|
||||||
|
-}
|
||||||
|
-
|
||||||
|
-/*
|
||||||
|
- * find_energy_efficient_cpu(): Find most energy-efficient target CPU for the
|
||||||
|
- * waking task. find_energy_efficient_cpu() looks for the CPU with maximum
|
||||||
|
- * spare capacity in each performance domain and uses it as a potential
|
||||||
|
- * candidate to execute the task. Then, it uses the Energy Model to figure
|
||||||
|
- * out which of the CPU candidates is the most energy-efficient.
|
||||||
|
- *
|
||||||
|
- * The rationale for this heuristic is as follows. In a performance domain,
|
||||||
|
- * all the most energy efficient CPU candidates (according to the Energy
|
||||||
|
- * Model) are those for which we'll request a low frequency. When there are
|
||||||
|
- * several CPUs for which the frequency request will be the same, we don't
|
||||||
|
- * have enough data to break the tie between them, because the Energy Model
|
||||||
|
- * only includes active power costs. With this model, if we assume that
|
||||||
|
- * frequency requests follow utilization (e.g. using schedutil), the CPU with
|
||||||
|
- * the maximum spare capacity in a performance domain is guaranteed to be among
|
||||||
|
- * the best candidates of the performance domain.
|
||||||
|
- *
|
||||||
|
- * In practice, it could be preferable from an energy standpoint to pack
|
||||||
|
- * small tasks on a CPU in order to let other CPUs go in deeper idle states,
|
||||||
|
- * but that could also hurt our chances to go cluster idle, and we have no
|
||||||
|
- * ways to tell with the current Energy Model if this is actually a good
|
||||||
|
- * idea or not. So, find_energy_efficient_cpu() basically favors
|
||||||
|
- * cluster-packing, and spreading inside a cluster. That should at least be
|
||||||
|
- * a good thing for latency, and this is consistent with the idea that most
|
||||||
|
- * of the energy savings of EAS come from the asymmetry of the system, and
|
||||||
|
- * not so much from breaking the tie between identical CPUs. That's also the
|
||||||
|
- * reason why EAS is enabled in the topology code only for systems where
|
||||||
|
- * SD_ASYM_CPUCAPACITY is set.
|
||||||
|
- *
|
||||||
|
- * NOTE: Forkees are not accepted in the energy-aware wake-up path because
|
||||||
|
- * they don't have any useful utilization data yet and it's not possible to
|
||||||
|
- * forecast their impact on energy consumption. Consequently, they will be
|
||||||
|
- * placed by find_idlest_cpu() on the least loaded CPU, which might turn out
|
||||||
|
- * to be energy-inefficient in some use-cases. The alternative would be to
|
||||||
|
- * bias new tasks towards specific types of CPUs first, or to try to infer
|
||||||
|
- * their util_avg from the parent task, but those heuristics could hurt
|
||||||
|
- * other use-cases too. So, until someone finds a better way to solve this,
|
||||||
|
- * let's keep things simple by re-using the existing slow path.
|
||||||
|
- */
|
||||||
|
-static int find_energy_efficient_cpu(struct task_struct *p, int prev_cpu)
|
||||||
|
-{
|
||||||
|
- struct cpumask *cpus = this_cpu_cpumask_var_ptr(select_rq_mask);
|
||||||
|
- unsigned long prev_delta = ULONG_MAX, best_delta = ULONG_MAX;
|
||||||
|
- unsigned long p_util_min = uclamp_is_used() ? uclamp_eff_value(p, UCLAMP_MIN) : 0;
|
||||||
|
- unsigned long p_util_max = uclamp_is_used() ? uclamp_eff_value(p, UCLAMP_MAX) : 1024;
|
||||||
|
- struct root_domain *rd = this_rq()->rd;
|
||||||
|
- int cpu, best_energy_cpu, target = -1;
|
||||||
|
- int prev_fits = -1, best_fits = -1;
|
||||||
|
- unsigned long best_thermal_cap = 0;
|
||||||
|
- unsigned long prev_thermal_cap = 0;
|
||||||
|
- struct sched_domain *sd;
|
||||||
|
- struct perf_domain *pd;
|
||||||
|
- struct energy_env eenv;
|
||||||
|
-
|
||||||
|
- rcu_read_lock();
|
||||||
|
- pd = rcu_dereference(rd->pd);
|
||||||
|
- if (!pd || READ_ONCE(rd->overutilized))
|
||||||
|
- goto unlock;
|
||||||
|
-
|
||||||
|
- /*
|
||||||
|
- * Energy-aware wake-up happens on the lowest sched_domain starting
|
||||||
|
- * from sd_asym_cpucapacity spanning over this_cpu and prev_cpu.
|
||||||
|
- */
|
||||||
|
- sd = rcu_dereference(*this_cpu_ptr(&sd_asym_cpucapacity));
|
||||||
|
- while (sd && !cpumask_test_cpu(prev_cpu, sched_domain_span(sd)))
|
||||||
|
- sd = sd->parent;
|
||||||
|
- if (!sd)
|
||||||
|
- goto unlock;
|
||||||
|
-
|
||||||
|
- target = prev_cpu;
|
||||||
|
-
|
||||||
|
- sync_entity_load_avg(&p->se);
|
||||||
|
- if (!uclamp_task_util(p, p_util_min, p_util_max))
|
||||||
|
- goto unlock;
|
||||||
|
-
|
||||||
|
- eenv_task_busy_time(&eenv, p, prev_cpu);
|
||||||
|
-
|
||||||
|
- for (; pd; pd = pd->next) {
|
||||||
|
- unsigned long util_min = p_util_min, util_max = p_util_max;
|
||||||
|
- unsigned long cpu_cap, cpu_thermal_cap, util;
|
||||||
|
- unsigned long cur_delta, max_spare_cap = 0;
|
||||||
|
- unsigned long rq_util_min, rq_util_max;
|
||||||
|
- unsigned long prev_spare_cap = 0;
|
||||||
|
- int max_spare_cap_cpu = -1;
|
||||||
|
- unsigned long base_energy;
|
||||||
|
- int fits, max_fits = -1;
|
||||||
|
-
|
||||||
|
- cpumask_and(cpus, perf_domain_span(pd), cpu_online_mask);
|
||||||
|
-
|
||||||
|
- if (cpumask_empty(cpus))
|
||||||
|
- continue;
|
||||||
|
-
|
||||||
|
- /* Account thermal pressure for the energy estimation */
|
||||||
|
- cpu = cpumask_first(cpus);
|
||||||
|
- cpu_thermal_cap = arch_scale_cpu_capacity(cpu);
|
||||||
|
- cpu_thermal_cap -= arch_scale_thermal_pressure(cpu);
|
||||||
|
-
|
||||||
|
- eenv.cpu_cap = cpu_thermal_cap;
|
||||||
|
- eenv.pd_cap = 0;
|
||||||
|
-
|
||||||
|
- for_each_cpu(cpu, cpus) {
|
||||||
|
- struct rq *rq = cpu_rq(cpu);
|
||||||
|
-
|
||||||
|
- eenv.pd_cap += cpu_thermal_cap;
|
||||||
|
-
|
||||||
|
- if (!cpumask_test_cpu(cpu, sched_domain_span(sd)))
|
||||||
|
- continue;
|
||||||
|
-
|
||||||
|
- if (!cpumask_test_cpu(cpu, p->cpus_ptr))
|
||||||
|
- continue;
|
||||||
|
-
|
||||||
|
- util = cpu_util_next(cpu, p, cpu);
|
||||||
|
- cpu_cap = capacity_of(cpu);
|
||||||
|
-
|
||||||
|
- /*
|
||||||
|
- * Skip CPUs that cannot satisfy the capacity request.
|
||||||
|
- * IOW, placing the task there would make the CPU
|
||||||
|
- * overutilized. Take uclamp into account to see how
|
||||||
|
- * much capacity we can get out of the CPU; this is
|
||||||
|
- * aligned with sched_cpu_util().
|
||||||
|
- */
|
||||||
|
- if (uclamp_is_used() && !uclamp_rq_is_idle(rq)) {
|
||||||
|
- /*
|
||||||
|
- * Open code uclamp_rq_util_with() except for
|
||||||
|
- * the clamp() part. Ie: apply max aggregation
|
||||||
|
- * only. util_fits_cpu() logic requires to
|
||||||
|
- * operate on non clamped util but must use the
|
||||||
|
- * max-aggregated uclamp_{min, max}.
|
||||||
|
- */
|
||||||
|
- rq_util_min = uclamp_rq_get(rq, UCLAMP_MIN);
|
||||||
|
- rq_util_max = uclamp_rq_get(rq, UCLAMP_MAX);
|
||||||
|
-
|
||||||
|
- util_min = max(rq_util_min, p_util_min);
|
||||||
|
- util_max = max(rq_util_max, p_util_max);
|
||||||
|
- }
|
||||||
|
-
|
||||||
|
- fits = util_fits_cpu(util, util_min, util_max, cpu);
|
||||||
|
- if (!fits)
|
||||||
|
- continue;
|
||||||
|
-
|
||||||
|
- lsub_positive(&cpu_cap, util);
|
||||||
|
-
|
||||||
|
- if (cpu == prev_cpu) {
|
||||||
|
- /* Always use prev_cpu as a candidate. */
|
||||||
|
- prev_spare_cap = cpu_cap;
|
||||||
|
- prev_fits = fits;
|
||||||
|
- } else if ((fits > max_fits) ||
|
||||||
|
- ((fits == max_fits) && (cpu_cap > max_spare_cap))) {
|
||||||
|
- /*
|
||||||
|
- * Find the CPU with the maximum spare capacity
|
||||||
|
- * among the remaining CPUs in the performance
|
||||||
|
- * domain.
|
||||||
|
- */
|
||||||
|
- max_spare_cap = cpu_cap;
|
||||||
|
- max_spare_cap_cpu = cpu;
|
||||||
|
- max_fits = fits;
|
||||||
|
- }
|
||||||
|
- }
|
||||||
|
-
|
||||||
|
- if (max_spare_cap_cpu < 0 && prev_spare_cap == 0)
|
||||||
|
- continue;
|
||||||
|
-
|
||||||
|
- eenv_pd_busy_time(&eenv, cpus, p);
|
||||||
|
- /* Compute the 'base' energy of the pd, without @p */
|
||||||
|
- base_energy = compute_energy(&eenv, pd, cpus, p, -1);
|
||||||
|
-
|
||||||
|
- /* Evaluate the energy impact of using prev_cpu. */
|
||||||
|
- if (prev_spare_cap > 0) {
|
||||||
|
- prev_delta = compute_energy(&eenv, pd, cpus, p,
|
||||||
|
- prev_cpu);
|
||||||
|
- /* CPU utilization has changed */
|
||||||
|
- if (prev_delta < base_energy)
|
||||||
|
- goto unlock;
|
||||||
|
- prev_delta -= base_energy;
|
||||||
|
- prev_thermal_cap = cpu_thermal_cap;
|
||||||
|
- best_delta = min(best_delta, prev_delta);
|
||||||
|
- }
|
||||||
|
-
|
||||||
|
- /* Evaluate the energy impact of using max_spare_cap_cpu. */
|
||||||
|
- if (max_spare_cap_cpu >= 0 && max_spare_cap > prev_spare_cap) {
|
||||||
|
- /* Current best energy cpu fits better */
|
||||||
|
- if (max_fits < best_fits)
|
||||||
|
- continue;
|
||||||
|
-
|
||||||
|
- /*
|
||||||
|
- * Both don't fit performance hint (i.e. uclamp_min)
|
||||||
|
- * but best energy cpu has better capacity.
|
||||||
|
- */
|
||||||
|
- if ((max_fits < 0) &&
|
||||||
|
- (cpu_thermal_cap <= best_thermal_cap))
|
||||||
|
- continue;
|
||||||
|
-
|
||||||
|
- cur_delta = compute_energy(&eenv, pd, cpus, p,
|
||||||
|
- max_spare_cap_cpu);
|
||||||
|
- /* CPU utilization has changed */
|
||||||
|
- if (cur_delta < base_energy)
|
||||||
|
- goto unlock;
|
||||||
|
- cur_delta -= base_energy;
|
||||||
|
-
|
||||||
|
- /*
|
||||||
|
- * Both fit for the task but best energy cpu has lower
|
||||||
|
- * energy impact.
|
||||||
|
- */
|
||||||
|
- if ((max_fits > 0) && (best_fits > 0) &&
|
||||||
|
- (cur_delta >= best_delta))
|
||||||
|
- continue;
|
||||||
|
-
|
||||||
|
- best_delta = cur_delta;
|
||||||
|
- best_energy_cpu = max_spare_cap_cpu;
|
||||||
|
- best_fits = max_fits;
|
||||||
|
- best_thermal_cap = cpu_thermal_cap;
|
||||||
|
- }
|
||||||
|
- }
|
||||||
|
- rcu_read_unlock();
|
||||||
|
-
|
||||||
|
- if ((best_fits > prev_fits) ||
|
||||||
|
- ((best_fits > 0) && (best_delta < prev_delta)) ||
|
||||||
|
- ((best_fits < 0) && (best_thermal_cap > prev_thermal_cap)))
|
||||||
|
- target = best_energy_cpu;
|
||||||
|
-
|
||||||
|
- return target;
|
||||||
|
-
|
||||||
|
-unlock:
|
||||||
|
- rcu_read_unlock();
|
||||||
|
-
|
||||||
|
- return target;
|
||||||
|
-}
|
||||||
|
-
|
||||||
|
/*
|
||||||
|
* select_task_rq_fair: Select target runqueue for the waking task in domains
|
||||||
|
* that have the relevant SD flag set. In practice, this is SD_BALANCE_WAKE,
|
||||||
|
@@ -7376,14 +7166,6 @@ select_task_rq_fair(struct task_struct *p, int prev_cpu, int wake_flags)
|
||||||
|
lockdep_assert_held(&p->pi_lock);
|
||||||
|
if (wake_flags & WF_TTWU) {
|
||||||
|
record_wakee(p);
|
||||||
|
-
|
||||||
|
- if (sched_energy_enabled()) {
|
||||||
|
- new_cpu = find_energy_efficient_cpu(p, prev_cpu);
|
||||||
|
- if (new_cpu >= 0)
|
||||||
|
- return new_cpu;
|
||||||
|
- new_cpu = prev_cpu;
|
||||||
|
- }
|
||||||
|
-
|
||||||
|
want_affine = !wake_wide(p) && cpumask_test_cpu(cpu, p->cpus_ptr);
|
||||||
|
}
|
||||||
|
|
||||||
|
--
|
||||||
|
2.39.1
|
118
linux-tkg-patches/6.3/0003-glitched-cfs.patch
Normal file
118
linux-tkg-patches/6.3/0003-glitched-cfs.patch
Normal file
@@ -0,0 +1,118 @@
|
|||||||
|
diff --git a/kernel/Kconfig.hz b/kernel/Kconfig.hz
|
||||||
|
index 2a202a846757..1d9c7ed79b11 100644
|
||||||
|
--- a/kernel/Kconfig.hz
|
||||||
|
+++ b/kernel/Kconfig.hz
|
||||||
|
@@ -4,7 +4,7 @@
|
||||||
|
|
||||||
|
choice
|
||||||
|
prompt "Timer frequency"
|
||||||
|
- default HZ_250
|
||||||
|
+ default HZ_500
|
||||||
|
help
|
||||||
|
Allows the configuration of the timer frequency. It is customary
|
||||||
|
to have the timer interrupt run at 1000 Hz but 100 Hz may be more
|
||||||
|
@@ -39,6 +39,13 @@ choice
|
||||||
|
on SMP and NUMA systems and exactly dividing by both PAL and
|
||||||
|
NTSC frame rates for video and multimedia work.
|
||||||
|
|
||||||
|
+ config HZ_500
|
||||||
|
+ bool "500 HZ"
|
||||||
|
+ help
|
||||||
|
+ 500 Hz is a balanced timer frequency. Provides fast interactivity
|
||||||
|
+ on desktops with great smoothness without increasing CPU power
|
||||||
|
+ consumption and sacrificing the battery life on laptops.
|
||||||
|
+
|
||||||
|
config HZ_1000
|
||||||
|
bool "1000 HZ"
|
||||||
|
help
|
||||||
|
@@ -52,6 +59,7 @@ config HZ
|
||||||
|
default 100 if HZ_100
|
||||||
|
default 250 if HZ_250
|
||||||
|
default 300 if HZ_300
|
||||||
|
+ default 500 if HZ_500
|
||||||
|
default 1000 if HZ_1000
|
||||||
|
|
||||||
|
config SCHED_HRTICK
|
||||||
|
|
||||||
|
diff --git a/kernel/Kconfig.hz b/kernel/Kconfig.hz
|
||||||
|
index 2a202a846757..1d9c7ed79b11 100644
|
||||||
|
--- a/kernel/Kconfig.hz
|
||||||
|
+++ b/kernel/Kconfig.hz
|
||||||
|
@@ -4,7 +4,7 @@
|
||||||
|
|
||||||
|
choice
|
||||||
|
prompt "Timer frequency"
|
||||||
|
- default HZ_500
|
||||||
|
+ default HZ_750
|
||||||
|
help
|
||||||
|
Allows the configuration of the timer frequency. It is customary
|
||||||
|
to have the timer interrupt run at 1000 Hz but 100 Hz may be more
|
||||||
|
@@ -46,6 +46,13 @@ choice
|
||||||
|
on desktops with great smoothness without increasing CPU power
|
||||||
|
consumption and sacrificing the battery life on laptops.
|
||||||
|
|
||||||
|
+ config HZ_750
|
||||||
|
+ bool "750 HZ"
|
||||||
|
+ help
|
||||||
|
+ 750 Hz is a good timer frequency for desktops. Provides fast
|
||||||
|
+ interactivity with great smoothness without sacrificing too
|
||||||
|
+ much throughput.
|
||||||
|
+
|
||||||
|
config HZ_1000
|
||||||
|
bool "1000 HZ"
|
||||||
|
help
|
||||||
|
@@ -60,6 +67,7 @@ config HZ
|
||||||
|
default 250 if HZ_250
|
||||||
|
default 300 if HZ_300
|
||||||
|
default 500 if HZ_500
|
||||||
|
+ default 750 if HZ_750
|
||||||
|
default 1000 if HZ_1000
|
||||||
|
|
||||||
|
config SCHED_HRTICK
|
||||||
|
|
||||||
|
diff --git a/drivers/cpufreq/cpufreq_ondemand.c b/drivers/cpufreq/cpufreq_ondemand.c
|
||||||
|
index 6b423eebfd5d..61e3271675d6 100644
|
||||||
|
--- a/drivers/cpufreq/cpufreq_ondemand.c
|
||||||
|
+++ b/drivers/cpufreq/cpufreq_ondemand.c
|
||||||
|
@@ -21,10 +21,10 @@
|
||||||
|
#include "cpufreq_ondemand.h"
|
||||||
|
|
||||||
|
/* On-demand governor macros */
|
||||||
|
-#define DEF_FREQUENCY_UP_THRESHOLD (80)
|
||||||
|
-#define DEF_SAMPLING_DOWN_FACTOR (1)
|
||||||
|
+#define DEF_FREQUENCY_UP_THRESHOLD (55)
|
||||||
|
+#define DEF_SAMPLING_DOWN_FACTOR (5)
|
||||||
|
#define MAX_SAMPLING_DOWN_FACTOR (100000)
|
||||||
|
-#define MICRO_FREQUENCY_UP_THRESHOLD (95)
|
||||||
|
+#define MICRO_FREQUENCY_UP_THRESHOLD (63)
|
||||||
|
#define MICRO_FREQUENCY_MIN_SAMPLE_RATE (10000)
|
||||||
|
#define MIN_FREQUENCY_UP_THRESHOLD (1)
|
||||||
|
#define MAX_FREQUENCY_UP_THRESHOLD (100)
|
||||||
|
|
||||||
|
From cba31b19f8c38696b13ba48e0e8b6dbe747d6bae Mon Sep 17 00:00:00 2001
|
||||||
|
From: Alexandre Frade <admfrade@gmail.com>
|
||||||
|
Date: Mon, 29 Jan 2018 17:31:25 +0000
|
||||||
|
Subject: [PATCH 10/16] XANMOD: mm/vmscan: vm_swappiness = 30 decreases the
|
||||||
|
amount of swapping
|
||||||
|
|
||||||
|
Signed-off-by: Alexandre Frade <admfrade@gmail.com>
|
||||||
|
Signed-off-by: Alexandre Frade <kernel@xanmod.org>
|
||||||
|
---
|
||||||
|
mm/vmscan.c | 2 +-
|
||||||
|
1 file changed, 1 insertion(+), 1 deletion(-)
|
||||||
|
|
||||||
|
diff --git a/mm/vmscan.c b/mm/vmscan.c
|
||||||
|
index 5b7b8d4f5297..549684b29418 100644
|
||||||
|
--- a/mm/vmscan.c
|
||||||
|
+++ b/mm/vmscan.c
|
||||||
|
@@ -190,7 +190,7 @@ struct scan_control {
|
||||||
|
/*
|
||||||
|
* From 0 .. 200. Higher means more swappy.
|
||||||
|
*/
|
||||||
|
-int vm_swappiness = 60;
|
||||||
|
+int vm_swappiness = 30;
|
||||||
|
|
||||||
|
static void set_task_reclaim_state(struct task_struct *task,
|
||||||
|
struct reclaim_state *rs)
|
||||||
|
--
|
||||||
|
2.39.1
|
193
linux-tkg-patches/6.3/0006-add-acs-overrides_iommu.patch
Normal file
193
linux-tkg-patches/6.3/0006-add-acs-overrides_iommu.patch
Normal file
@@ -0,0 +1,193 @@
|
|||||||
|
From cdeab384f48dd9c88e2dff2e9ad8d57dca1a1b1c Mon Sep 17 00:00:00 2001
|
||||||
|
From: Mark Weiman <mark.weiman@markzz.com>
|
||||||
|
Date: Sun, 12 Aug 2018 11:36:21 -0400
|
||||||
|
Subject: [PATCH] pci: Enable overrides for missing ACS capabilities
|
||||||
|
|
||||||
|
This an updated version of Alex Williamson's patch from:
|
||||||
|
https://lkml.org/lkml/2013/5/30/513
|
||||||
|
|
||||||
|
Original commit message follows:
|
||||||
|
|
||||||
|
PCIe ACS (Access Control Services) is the PCIe 2.0+ feature that
|
||||||
|
allows us to control whether transactions are allowed to be redirected
|
||||||
|
in various subnodes of a PCIe topology. For instance, if two
|
||||||
|
endpoints are below a root port or downsteam switch port, the
|
||||||
|
downstream port may optionally redirect transactions between the
|
||||||
|
devices, bypassing upstream devices. The same can happen internally
|
||||||
|
on multifunction devices. The transaction may never be visible to the
|
||||||
|
upstream devices.
|
||||||
|
|
||||||
|
One upstream device that we particularly care about is the IOMMU. If
|
||||||
|
a redirection occurs in the topology below the IOMMU, then the IOMMU
|
||||||
|
cannot provide isolation between devices. This is why the PCIe spec
|
||||||
|
encourages topologies to include ACS support. Without it, we have to
|
||||||
|
assume peer-to-peer DMA within a hierarchy can bypass IOMMU isolation.
|
||||||
|
|
||||||
|
Unfortunately, far too many topologies do not support ACS to make this
|
||||||
|
a steadfast requirement. Even the latest chipsets from Intel are only
|
||||||
|
sporadically supporting ACS. We have trouble getting interconnect
|
||||||
|
vendors to include the PCIe spec required PCIe capability, let alone
|
||||||
|
suggested features.
|
||||||
|
|
||||||
|
Therefore, we need to add some flexibility. The pcie_acs_override=
|
||||||
|
boot option lets users opt-in specific devices or sets of devices to
|
||||||
|
assume ACS support. The "downstream" option assumes full ACS support
|
||||||
|
on root ports and downstream switch ports. The "multifunction"
|
||||||
|
option assumes the subset of ACS features available on multifunction
|
||||||
|
endpoints and upstream switch ports are supported. The "id:nnnn:nnnn"
|
||||||
|
option enables ACS support on devices matching the provided vendor
|
||||||
|
and device IDs, allowing more strategic ACS overrides. These options
|
||||||
|
may be combined in any order. A maximum of 16 id specific overrides
|
||||||
|
are available. It's suggested to use the most limited set of options
|
||||||
|
necessary to avoid completely disabling ACS across the topology.
|
||||||
|
Note to hardware vendors, we have facilities to permanently quirk
|
||||||
|
specific devices which enforce isolation but not provide an ACS
|
||||||
|
capability. Please contact me to have your devices added and save
|
||||||
|
your customers the hassle of this boot option.
|
||||||
|
|
||||||
|
Signed-off-by: Mark Weiman <mark.weiman@markzz.com>
|
||||||
|
---
|
||||||
|
.../admin-guide/kernel-parameters.txt | 9 ++
|
||||||
|
drivers/pci/quirks.c | 101 ++++++++++++++++++
|
||||||
|
2 files changed, 110 insertions(+)
|
||||||
|
|
||||||
|
diff --git a/Documentation/admin-guide/kernel-parameters.txt b/Documentation/admin-guide/kernel-parameters.txt
|
||||||
|
index aefd358a5ca3..173b3596fd9e 100644
|
||||||
|
--- a/Documentation/admin-guide/kernel-parameters.txt
|
||||||
|
+++ b/Documentation/admin-guide/kernel-parameters.txt
|
||||||
|
@@ -3190,6 +3190,15 @@
|
||||||
|
nomsi [MSI] If the PCI_MSI kernel config parameter is
|
||||||
|
enabled, this kernel boot option can be used to
|
||||||
|
disable the use of MSI interrupts system-wide.
|
||||||
|
+ pcie_acs_override =
|
||||||
|
+ [PCIE] Override missing PCIe ACS support for:
|
||||||
|
+ downstream
|
||||||
|
+ All downstream ports - full ACS capabilities
|
||||||
|
+ multifunction
|
||||||
|
+ All multifunction devices - multifunction ACS subset
|
||||||
|
+ id:nnnn:nnnn
|
||||||
|
+ Specific device - full ACS capabilities
|
||||||
|
+ Specified as vid:did (vendor/device ID) in hex
|
||||||
|
noioapicquirk [APIC] Disable all boot interrupt quirks.
|
||||||
|
Safety option to keep boot IRQs enabled. This
|
||||||
|
should never be necessary.
|
||||||
|
diff --git a/drivers/pci/quirks.c b/drivers/pci/quirks.c
|
||||||
|
index 4700d24e5d55..8f7a3d7fd9c1 100644
|
||||||
|
--- a/drivers/pci/quirks.c
|
||||||
|
+++ b/drivers/pci/quirks.c
|
||||||
|
@@ -3372,6 +3372,106 @@ static void quirk_no_bus_reset(struct pci_dev *dev)
|
||||||
|
dev->dev_flags |= PCI_DEV_FLAGS_NO_BUS_RESET;
|
||||||
|
}
|
||||||
|
|
||||||
|
+static bool acs_on_downstream;
|
||||||
|
+static bool acs_on_multifunction;
|
||||||
|
+
|
||||||
|
+#define NUM_ACS_IDS 16
|
||||||
|
+struct acs_on_id {
|
||||||
|
+ unsigned short vendor;
|
||||||
|
+ unsigned short device;
|
||||||
|
+};
|
||||||
|
+static struct acs_on_id acs_on_ids[NUM_ACS_IDS];
|
||||||
|
+static u8 max_acs_id;
|
||||||
|
+
|
||||||
|
+static __init int pcie_acs_override_setup(char *p)
|
||||||
|
+{
|
||||||
|
+ if (!p)
|
||||||
|
+ return -EINVAL;
|
||||||
|
+
|
||||||
|
+ while (*p) {
|
||||||
|
+ if (!strncmp(p, "downstream", 10))
|
||||||
|
+ acs_on_downstream = true;
|
||||||
|
+ if (!strncmp(p, "multifunction", 13))
|
||||||
|
+ acs_on_multifunction = true;
|
||||||
|
+ if (!strncmp(p, "id:", 3)) {
|
||||||
|
+ char opt[5];
|
||||||
|
+ int ret;
|
||||||
|
+ long val;
|
||||||
|
+
|
||||||
|
+ if (max_acs_id >= NUM_ACS_IDS - 1) {
|
||||||
|
+ pr_warn("Out of PCIe ACS override slots (%d)\n",
|
||||||
|
+ NUM_ACS_IDS);
|
||||||
|
+ goto next;
|
||||||
|
+ }
|
||||||
|
+
|
||||||
|
+ p += 3;
|
||||||
|
+ snprintf(opt, 5, "%s", p);
|
||||||
|
+ ret = kstrtol(opt, 16, &val);
|
||||||
|
+ if (ret) {
|
||||||
|
+ pr_warn("PCIe ACS ID parse error %d\n", ret);
|
||||||
|
+ goto next;
|
||||||
|
+ }
|
||||||
|
+ acs_on_ids[max_acs_id].vendor = val;
|
||||||
|
+
|
||||||
|
+ p += strcspn(p, ":");
|
||||||
|
+ if (*p != ':') {
|
||||||
|
+ pr_warn("PCIe ACS invalid ID\n");
|
||||||
|
+ goto next;
|
||||||
|
+ }
|
||||||
|
+
|
||||||
|
+ p++;
|
||||||
|
+ snprintf(opt, 5, "%s", p);
|
||||||
|
+ ret = kstrtol(opt, 16, &val);
|
||||||
|
+ if (ret) {
|
||||||
|
+ pr_warn("PCIe ACS ID parse error %d\n", ret);
|
||||||
|
+ goto next;
|
||||||
|
+ }
|
||||||
|
+ acs_on_ids[max_acs_id].device = val;
|
||||||
|
+ max_acs_id++;
|
||||||
|
+ }
|
||||||
|
+next:
|
||||||
|
+ p += strcspn(p, ",");
|
||||||
|
+ if (*p == ',')
|
||||||
|
+ p++;
|
||||||
|
+ }
|
||||||
|
+
|
||||||
|
+ if (acs_on_downstream || acs_on_multifunction || max_acs_id)
|
||||||
|
+ pr_warn("Warning: PCIe ACS overrides enabled; This may allow non-IOMMU protected peer-to-peer DMA\n");
|
||||||
|
+
|
||||||
|
+ return 0;
|
||||||
|
+}
|
||||||
|
+early_param("pcie_acs_override", pcie_acs_override_setup);
|
||||||
|
+
|
||||||
|
+static int pcie_acs_overrides(struct pci_dev *dev, u16 acs_flags)
|
||||||
|
+{
|
||||||
|
+ int i;
|
||||||
|
+
|
||||||
|
+ /* Never override ACS for legacy devices or devices with ACS caps */
|
||||||
|
+ if (!pci_is_pcie(dev) ||
|
||||||
|
+ pci_find_ext_capability(dev, PCI_EXT_CAP_ID_ACS))
|
||||||
|
+ return -ENOTTY;
|
||||||
|
+
|
||||||
|
+ for (i = 0; i < max_acs_id; i++)
|
||||||
|
+ if (acs_on_ids[i].vendor == dev->vendor &&
|
||||||
|
+ acs_on_ids[i].device == dev->device)
|
||||||
|
+ return 1;
|
||||||
|
+
|
||||||
|
+ switch (pci_pcie_type(dev)) {
|
||||||
|
+ case PCI_EXP_TYPE_DOWNSTREAM:
|
||||||
|
+ case PCI_EXP_TYPE_ROOT_PORT:
|
||||||
|
+ if (acs_on_downstream)
|
||||||
|
+ return 1;
|
||||||
|
+ break;
|
||||||
|
+ case PCI_EXP_TYPE_ENDPOINT:
|
||||||
|
+ case PCI_EXP_TYPE_UPSTREAM:
|
||||||
|
+ case PCI_EXP_TYPE_LEG_END:
|
||||||
|
+ case PCI_EXP_TYPE_RC_END:
|
||||||
|
+ if (acs_on_multifunction && dev->multifunction)
|
||||||
|
+ return 1;
|
||||||
|
+ }
|
||||||
|
+
|
||||||
|
+ return -ENOTTY;
|
||||||
|
+}
|
||||||
|
/*
|
||||||
|
* Some Atheros AR9xxx and QCA988x chips do not behave after a bus reset.
|
||||||
|
* The device will throw a Link Down error on AER-capable systems and
|
||||||
|
@@ -4513,6 +4613,7 @@ static const struct pci_dev_acs_enabled {
|
||||||
|
{ PCI_VENDOR_ID_ZHAOXIN, 0x9083, pci_quirk_mf_endpoint_acs },
|
||||||
|
/* Zhaoxin Root/Downstream Ports */
|
||||||
|
{ PCI_VENDOR_ID_ZHAOXIN, PCI_ANY_ID, pci_quirk_zhaoxin_pcie_ports_acs },
|
||||||
|
+ { PCI_ANY_ID, PCI_ANY_ID, pcie_acs_overrides },
|
||||||
|
{ 0 }
|
||||||
|
};
|
||||||
|
|
||||||
|
|
@@ -13,12 +13,12 @@ of Proton to still use fsync in new kernel releases.
|
|||||||
|
|
||||||
Signed-off-by: André Almeida <andrealmeid@collabora.com>
|
Signed-off-by: André Almeida <andrealmeid@collabora.com>
|
||||||
---
|
---
|
||||||
include/uapi/linux/futex.h | 12 ++++++
|
include/uapi/linux/futex.h | 13 +++++++
|
||||||
kernel/futex/core.c | 75 +++++++++++++++++++++++++++++++++++++-
|
kernel/futex/syscalls.c | 75 +++++++++++++++++++++++++++++++++++++-
|
||||||
2 files changed, 86 insertions(+), 1 deletion(-)
|
2 files changed, 87 insertions(+), 1 deletion(-)
|
||||||
|
|
||||||
diff --git a/include/uapi/linux/futex.h b/include/uapi/linux/futex.h
|
diff --git a/include/uapi/linux/futex.h b/include/uapi/linux/futex.h
|
||||||
index 2a06b99f9803..417c5d89b745 100644
|
index 71a5df8d2689..d375ab21cbf8 100644
|
||||||
--- a/include/uapi/linux/futex.h
|
--- a/include/uapi/linux/futex.h
|
||||||
+++ b/include/uapi/linux/futex.h
|
+++ b/include/uapi/linux/futex.h
|
||||||
@@ -22,6 +22,7 @@
|
@@ -22,6 +22,7 @@
|
||||||
@@ -29,7 +29,7 @@ index 2a06b99f9803..417c5d89b745 100644
|
|||||||
|
|
||||||
#define FUTEX_PRIVATE_FLAG 128
|
#define FUTEX_PRIVATE_FLAG 128
|
||||||
#define FUTEX_CLOCK_REALTIME 256
|
#define FUTEX_CLOCK_REALTIME 256
|
||||||
@@ -68,6 +69,17 @@ struct futex_waitv {
|
@@ -68,6 +69,18 @@ struct futex_waitv {
|
||||||
__u32 __reserved;
|
__u32 __reserved;
|
||||||
};
|
};
|
||||||
|
|
||||||
@@ -44,14 +44,15 @@ index 2a06b99f9803..417c5d89b745 100644
|
|||||||
+ __u32 val;
|
+ __u32 val;
|
||||||
+ __u32 bitset;
|
+ __u32 bitset;
|
||||||
+};
|
+};
|
||||||
|
+
|
||||||
/*
|
/*
|
||||||
* Support for robust futexes: the kernel cleans up held futexes at
|
* Support for robust futexes: the kernel cleans up held futexes at
|
||||||
diff --git a/kernel/futex/core.c b/kernel/futex/core.c
|
* thread exit time.
|
||||||
index 4a9e7ce3714a..c3f2e65afab8 100644
|
diff --git a/kernel/futex/syscalls.c b/kernel/futex/syscalls.c
|
||||||
--- a/kernel/futex/core.c
|
index 6f91a07a6a83..2f4d4c04ede2 100644
|
||||||
+++ b/kernel/futex/core.c
|
--- a/kernel/futex/syscalls.c
|
||||||
@@ -4012,6 +4012,7 @@ static __always_inline bool futex_cmd_has_timeout(u32 cmd)
|
+++ b/kernel/futex/syscalls.c
|
||||||
|
@@ -158,6 +158,7 @@ static __always_inline bool futex_cmd_has_timeout(u32 cmd)
|
||||||
case FUTEX_LOCK_PI2:
|
case FUTEX_LOCK_PI2:
|
||||||
case FUTEX_WAIT_BITSET:
|
case FUTEX_WAIT_BITSET:
|
||||||
case FUTEX_WAIT_REQUEUE_PI:
|
case FUTEX_WAIT_REQUEUE_PI:
|
||||||
@@ -59,7 +60,7 @@ index 4a9e7ce3714a..c3f2e65afab8 100644
|
|||||||
return true;
|
return true;
|
||||||
}
|
}
|
||||||
return false;
|
return false;
|
||||||
@@ -4024,13 +4025,79 @@ futex_init_timeout(u32 cmd, u32 op, struct timespec64 *ts, ktime_t *t)
|
@@ -170,13 +171,79 @@ futex_init_timeout(u32 cmd, u32 op, struct timespec64 *ts, ktime_t *t)
|
||||||
return -EINVAL;
|
return -EINVAL;
|
||||||
|
|
||||||
*t = timespec64_to_ktime(*ts);
|
*t = timespec64_to_ktime(*ts);
|
||||||
@@ -140,7 +141,7 @@ index 4a9e7ce3714a..c3f2e65afab8 100644
|
|||||||
SYSCALL_DEFINE6(futex, u32 __user *, uaddr, int, op, u32, val,
|
SYSCALL_DEFINE6(futex, u32 __user *, uaddr, int, op, u32, val,
|
||||||
const struct __kernel_timespec __user *, utime,
|
const struct __kernel_timespec __user *, utime,
|
||||||
u32 __user *, uaddr2, u32, val3)
|
u32 __user *, uaddr2, u32, val3)
|
||||||
@@ -4050,6 +4117,9 @@ SYSCALL_DEFINE6(futex, u32 __user *, uaddr, int, op, u32, val,
|
@@ -196,6 +263,9 @@ SYSCALL_DEFINE6(futex, u32 __user *, uaddr, int, op, u32, val,
|
||||||
tp = &t;
|
tp = &t;
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -150,7 +151,7 @@ index 4a9e7ce3714a..c3f2e65afab8 100644
|
|||||||
return do_futex(uaddr, op, val, tp, uaddr2, (unsigned long)utime, val3);
|
return do_futex(uaddr, op, val, tp, uaddr2, (unsigned long)utime, val3);
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -4551,6 +4621,9 @@ SYSCALL_DEFINE6(futex_time32, u32 __user *, uaddr, int, op, u32, val,
|
@@ -392,6 +462,9 @@ SYSCALL_DEFINE6(futex_time32, u32 __user *, uaddr, int, op, u32, val,
|
||||||
tp = &t;
|
tp = &t;
|
||||||
}
|
}
|
||||||
|
|
5105
linux-tkg-patches/6.3/0007-v6.3-winesync.patch
Normal file
5105
linux-tkg-patches/6.3/0007-v6.3-winesync.patch
Normal file
File diff suppressed because it is too large
Load Diff
66
linux-tkg-patches/6.3/0012-misc-additions.patch
Normal file
66
linux-tkg-patches/6.3/0012-misc-additions.patch
Normal file
@@ -0,0 +1,66 @@
|
|||||||
|
From e5e77ad2223f662e1615266d8ef39a8db7e65a70 Mon Sep 17 00:00:00 2001
|
||||||
|
From: =?UTF-8?q?Felix=20H=C3=A4dicke?= <felixhaedicke@web.de>
|
||||||
|
Date: Thu, 19 Nov 2020 09:22:32 +0100
|
||||||
|
Subject: HID: quirks: Add Apple Magic Trackpad 2 to hid_have_special_driver
|
||||||
|
list
|
||||||
|
MIME-Version: 1.0
|
||||||
|
Content-Type: text/plain; charset=UTF-8
|
||||||
|
Content-Transfer-Encoding: 8bit
|
||||||
|
|
||||||
|
The Apple Magic Trackpad 2 is handled by the magicmouse driver. And
|
||||||
|
there were severe stability issues when both drivers (hid-generic and
|
||||||
|
hid-magicmouse) were loaded for this device.
|
||||||
|
|
||||||
|
Fixes: https://bugzilla.kernel.org/show_bug.cgi?id=210241
|
||||||
|
|
||||||
|
Signed-off-by: Felix Hädicke <felixhaedicke@web.de>
|
||||||
|
---
|
||||||
|
drivers/hid/hid-quirks.c | 2 ++
|
||||||
|
1 file changed, 2 insertions(+)
|
||||||
|
|
||||||
|
diff --git a/drivers/hid/hid-quirks.c b/drivers/hid/hid-quirks.c
|
||||||
|
index bf7ecab5d9e5..142e9dae2837 100644
|
||||||
|
--- a/drivers/hid/hid-quirks.c
|
||||||
|
+++ b/drivers/hid/hid-quirks.c
|
||||||
|
@@ -478,6 +478,8 @@ static const struct hid_device_id hid_have_special_driver[] = {
|
||||||
|
#if IS_ENABLED(CONFIG_HID_MAGICMOUSE)
|
||||||
|
{ HID_BLUETOOTH_DEVICE(USB_VENDOR_ID_APPLE, USB_DEVICE_ID_APPLE_MAGICMOUSE) },
|
||||||
|
{ HID_BLUETOOTH_DEVICE(USB_VENDOR_ID_APPLE, USB_DEVICE_ID_APPLE_MAGICTRACKPAD) },
|
||||||
|
+ { HID_BLUETOOTH_DEVICE(BT_VENDOR_ID_APPLE, USB_DEVICE_ID_APPLE_MAGICTRACKPAD2) },
|
||||||
|
+ { HID_USB_DEVICE(USB_VENDOR_ID_APPLE, USB_DEVICE_ID_APPLE_MAGICTRACKPAD2) },
|
||||||
|
#endif
|
||||||
|
#if IS_ENABLED(CONFIG_HID_MAYFLASH)
|
||||||
|
{ HID_USB_DEVICE(USB_VENDOR_ID_DRAGONRISE, USB_DEVICE_ID_DRAGONRISE_PS3) },
|
||||||
|
--
|
||||||
|
cgit v1.2.3-1-gf6bb5
|
||||||
|
|
||||||
|
From f7f49141a5dbe9c99d78196b58c44307fb2e6be3 Mon Sep 17 00:00:00 2001
|
||||||
|
From: Tk-Glitch <ti3nou@gmail.com>
|
||||||
|
Date: Wed, 3 Feb 2021 11:20:12 +0200
|
||||||
|
Subject: Revert "cpufreq: Avoid configuring old governors as default with intel_pstate"
|
||||||
|
|
||||||
|
This is an undesirable behavior for us since our aggressive ondemand performs
|
||||||
|
better than schedutil for gaming when using intel_pstate in passive mode.
|
||||||
|
Also it interferes with the option to select the desired default governor we have.
|
||||||
|
|
||||||
|
diff --git a/drivers/cpufreq/Kconfig b/drivers/cpufreq/Kconfig
|
||||||
|
index 2c7171e0b0010..85de313ddec29 100644
|
||||||
|
--- a/drivers/cpufreq/Kconfig
|
||||||
|
+++ b/drivers/cpufreq/Kconfig
|
||||||
|
@@ -71,7 +71,6 @@ config CPU_FREQ_DEFAULT_GOV_USERSPACE
|
||||||
|
|
||||||
|
config CPU_FREQ_DEFAULT_GOV_ONDEMAND
|
||||||
|
bool "ondemand"
|
||||||
|
- depends on !(X86_INTEL_PSTATE && SMP)
|
||||||
|
select CPU_FREQ_GOV_ONDEMAND
|
||||||
|
select CPU_FREQ_GOV_PERFORMANCE
|
||||||
|
help
|
||||||
|
@@ -83,7 +84,6 @@ config CPU_FREQ_DEFAULT_GOV_ONDEMAND
|
||||||
|
|
||||||
|
config CPU_FREQ_DEFAULT_GOV_CONSERVATIVE
|
||||||
|
bool "conservative"
|
||||||
|
- depends on !(X86_INTEL_PSTATE && SMP)
|
||||||
|
select CPU_FREQ_GOV_CONSERVATIVE
|
||||||
|
select CPU_FREQ_GOV_PERFORMANCE
|
||||||
|
help
|
||||||
|
|
53
linux-tkg-patches/6.3/0013-fedora-rpm.patch
Normal file
53
linux-tkg-patches/6.3/0013-fedora-rpm.patch
Normal file
@@ -0,0 +1,53 @@
|
|||||||
|
diff --git a/scripts/package/mkspec b/scripts/package/mkspec
|
||||||
|
index 70392fd2f..34f98648f 100755
|
||||||
|
--- a/scripts/package/mkspec
|
||||||
|
+++ b/scripts/package/mkspec
|
||||||
|
@@ -25,7 +25,7 @@ fi
|
||||||
|
|
||||||
|
+PROVIDES_DRM=""
|
||||||
|
if grep -q CONFIG_DRM=y .config; then
|
||||||
|
- PROVIDES=kernel-drm
|
||||||
|
+ PROVIDES_DRM="Provides: kernel-drm = %{version}"
|
||||||
|
fi
|
||||||
|
|
||||||
|
-PROVIDES="$PROVIDES kernel-$KERNELRELEASE"
|
||||||
|
__KERNELRELEASE=$(echo $KERNELRELEASE | sed -e "s/-/_/g")
|
||||||
|
@@ -50,3 +50,6 @@ sed -e '/^DEL/d' -e 's/^\t*//' <<EOF
|
||||||
|
$S Source: kernel-$__KERNELRELEASE.tar.gz
|
||||||
|
- Provides: $PROVIDES
|
||||||
|
+ $PROVIDES_DRM
|
||||||
|
+ Provides: kernel = %{version}
|
||||||
|
+ Provides: kernel-uname-r = %{version}
|
||||||
|
+ Provides: installonlypkg(kernel) = %{version}
|
||||||
|
# $UTS_MACHINE as a fallback of _arch in case
|
||||||
|
@@ -63,4 +66,4 @@ $S Source: kernel-$__KERNELRELEASE.tar.gz
|
||||||
|
Group: Development/System
|
||||||
|
- Obsoletes: kernel-headers
|
||||||
|
Provides: kernel-headers = %{version}
|
||||||
|
+ Provides: installonlypkg(kernel) = %{version}
|
||||||
|
%description headers
|
||||||
|
@@ -75,2 +78,5 @@ $S$M Summary: Development package for building kernel modules to match the $__KE
|
||||||
|
$S$M Group: System Environment/Kernel
|
||||||
|
+$S$M Provides: kernel-devel = %{version}
|
||||||
|
+$S$M Provides: kernel-devel-uname-r = %{version}
|
||||||
|
+$S$M Provides: installonlypkg(kernel) = %{version}
|
||||||
|
$S$M AutoReqProv: no
|
||||||
|
@@ -80,2 +86,18 @@ $S$M against the $__KERNELRELEASE kernel package.
|
||||||
|
$S$M
|
||||||
|
+$S # Opt out of a lot of Fedora hardening flags etc...
|
||||||
|
+$S # See https://src.fedoraproject.org/rpms/redhat-rpm-config//blob/rawhide/f/buildflags.md
|
||||||
|
+$S %undefine _package_note_file
|
||||||
|
+$S %undefine _auto_set_build_flags
|
||||||
|
+$S %undefine _include_frame_pointers
|
||||||
|
+$S %define _build_id_flags -Wl,--build-id=none
|
||||||
|
+$S %undefine _annotated_build
|
||||||
|
+$S %undefine _fortify_level
|
||||||
|
+$S %undefine _hardened_build
|
||||||
|
+$S %global _lto_cflags %{nil}
|
||||||
|
+$S %global _configure_gnuconfig_hack 0
|
||||||
|
+$S %global _configure_libtool_hardening_hack 0
|
||||||
|
+$S # Nearly had to go to the deep web to find documentation on this one... Gosh
|
||||||
|
+$S # See https://github.com/rpm-software-management/rpm/blob/master/macros.in#L471
|
||||||
|
+$S %define _build_id_links none
|
||||||
|
+$S
|
||||||
|
$S %prep
|
46
linux-tkg-patches/6.3/0013-optimize_harder_O3.patch
Normal file
46
linux-tkg-patches/6.3/0013-optimize_harder_O3.patch
Normal file
@@ -0,0 +1,46 @@
|
|||||||
|
diff --git a/Makefile b/Makefile
|
||||||
|
--- a/Makefile
|
||||||
|
+++ b/Makefile
|
||||||
|
@@ -442,7 +442,7 @@ endif
|
||||||
|
HOSTPKG_CONFIG = pkg-config
|
||||||
|
|
||||||
|
KBUILD_USERHOSTCFLAGS := -Wall -Wmissing-prototypes -Wstrict-prototypes \
|
||||||
|
- -O2 -fomit-frame-pointer -std=gnu11 \
|
||||||
|
+ -O3 -fomit-frame-pointer -std=gnu11 \
|
||||||
|
-Wdeclaration-after-statement
|
||||||
|
KBUILD_USERCFLAGS := $(KBUILD_USERHOSTCFLAGS) $(USERCFLAGS)
|
||||||
|
KBUILD_USERLDFLAGS := $(USERLDFLAGS)
|
||||||
|
@@ -474,7 +474,7 @@ endif
|
||||||
|
-Wclippy::dbg_macro
|
||||||
|
|
||||||
|
KBUILD_HOSTCFLAGS := $(KBUILD_USERHOSTCFLAGS) $(HOST_LFS_CFLAGS) $(HOSTCFLAGS)
|
||||||
|
-KBUILD_HOSTCXXFLAGS := -Wall -O2 $(HOST_LFS_CFLAGS) $(HOSTCXXFLAGS)
|
||||||
|
+KBUILD_HOSTCXXFLAGS := -Wall -O3 $(HOST_LFS_CFLAGS) $(HOSTCXXFLAGS)
|
||||||
|
KBUILD_HOSTRUSTFLAGS := $(rust_common_flags) -O -Cstrip=debuginfo \
|
||||||
|
-Zallow-features= $(HOSTRUSTFLAGS)
|
||||||
|
KBUILD_HOSTLDFLAGS := $(HOST_LFS_LDFLAGS) $(HOSTLDFLAGS)
|
||||||
|
@@ -757,7 +757,7 @@ KBUILD_CFLAGS += $(call cc-disable-warning, format-overflow)
|
||||||
|
KBUILD_CFLAGS += $(call cc-disable-warning, address-of-packed-member)
|
||||||
|
|
||||||
|
ifdef CONFIG_CC_OPTIMIZE_FOR_PERFORMANCE
|
||||||
|
-KBUILD_CFLAGS += -O2
|
||||||
|
+KBUILD_CFLAGS += -O3
|
||||||
|
KBUILD_RUSTFLAGS += -Copt-level=2
|
||||||
|
else ifdef CONFIG_CC_OPTIMIZE_FOR_SIZE
|
||||||
|
KBUILD_CFLAGS += -Os
|
||||||
|
diff --git a/init/Kconfig b/init/Kconfig
|
||||||
|
--- a/init/Kconfig
|
||||||
|
+++ b/init/Kconfig
|
||||||
|
@@ -1401,10 +1401,10 @@ choice
|
||||||
|
default CC_OPTIMIZE_FOR_PERFORMANCE
|
||||||
|
|
||||||
|
config CC_OPTIMIZE_FOR_PERFORMANCE
|
||||||
|
- bool "Optimize for performance (-O2)"
|
||||||
|
+ bool "Optimize for performance (-O3)"
|
||||||
|
help
|
||||||
|
This is the default optimization level for the kernel, building
|
||||||
|
- with the "-O2" compiler flag for best performance and most
|
||||||
|
+ with the "-O3" compiler flag for best performance and most
|
||||||
|
helpful compile-time warnings.
|
||||||
|
|
||||||
|
config CC_OPTIMIZE_FOR_SIZE
|
Reference in New Issue
Block a user