/* Void Main's man pages */

{ phpMan } else { main(); }

Command: man perldoc info search(apropos)  


CPUSET(7)                                           Linux Programmer's Manual                                          CPUSET(7)



NAME
       cpuset - confine processes to processor and memory node subsets

DESCRIPTION
       The  cpuset  file  system  is a pseudo-file-system interface to the kernel cpuset mechanism, which is used to control the
       processor placement and memory placement of processes.  It is commonly mounted at /dev/cpuset.

       On systems with kernels compiled with built in support for cpusets, all processes are attached to a cpuset,  and  cpusets
       are  always  present.   If a system supports cpusets, then it will have the entry nodev cpuset in the file /proc/filesys-
       tems.  By mounting the cpuset file system (see the EXAMPLE section below), the administrator can configure the cpusets on
       a system to control the processor and memory placement of processes on that system.  By default, if the cpuset configura-
       tion on a system is not modified or if the cpuset file system is not even mounted,  then  the  cpuset  mechanism,  though
       present, has no affect on the system's behavior.

       A cpuset defines a list of CPUs and memory nodes.

       The CPUs of a system include all the logical processing units on which a process can execute, including, if present, mul-
       tiple processor cores within a package and Hyper-Threads within a processor core.   Memory  nodes  include  all  distinct
       banks  of main memory; small and SMP systems typically have just one memory node that contains all the system's main mem-
       ory, while NUMA (non-uniform memory access) systems have multiple memory nodes.

       Cpusets are represented as directories in a hierarchical pseudo-file system, where the top  directory  in  the  hierarchy
       (/dev/cpuset)  represents  the entire system (all online CPUs and memory nodes) and any cpuset that is the child (descen-
       dant) of another parent cpuset contains a subset of that parent's CPUs and memory nodes.  The directories and files  rep-
       resenting cpusets have normal file-system permissions.

       Every  process  in the system belongs to exactly one cpuset.  A process is confined to only run on the CPUs in the cpuset
       it belongs to, and to allocate memory only on the memory nodes in that  cpuset.   When  a  process  fork(2)s,  the  child
       process is placed in the same cpuset as its parent.  With sufficient privilege, a process may be moved from one cpuset to
       another and the allowed CPUs and memory nodes of an existing cpuset may be changed.

       When the system begins booting, a single cpuset is defined that includes all CPUs and memory nodes on the system, and all
       processes  are  in  that  cpuset.  During the boot process, or later during normal system operation, other cpusets may be
       created, as subdirectories of this top cpuset, under the control of the system administrator, and processes may be placed
       in these other cpusets.

       Cpusets  are integrated with the sched_setaffinity(2) scheduling affinity mechanism and the mbind(2) and set_mempolicy(2)
       memory-placement mechanisms in the kernel.  Neither of these mechanisms let a process make use of a CPU  or  memory  node
       that is not allowed by that process's cpuset.  If changes to a process's cpuset placement conflict with these other mech-
       anisms, then cpuset placement is enforced even if it means overriding these other mechanisms.   The  kernel  accomplishes
       this overriding by silently restricting the CPUs and memory nodes requested by these other mechanisms to those allowed by
       the invoking process's cpuset.  This can result in these other calls returning an error, if for example, such a call ends
       up requesting an empty set of CPUs or memory nodes, after that request is restricted to the invoking process's cpuset.

       Typically,  a  cpuset  is used to manage the CPU and memory-node confinement for a set of cooperating processes such as a
       batch scheduler job, and these other mechanisms are used to manage  the  placement  of  individual  processes  or  memory
       regions within that set or job.

FILES
       Each  directory  below  /dev/cpuset  represents a cpuset and contains a fixed set of pseudo-files describing the state of
       that cpuset.

       New cpusets are created using the mkdir(2) system call or the mkdir(1) command.  The properties of a cpuset, such as  its
       flags,  allowed  CPUs  and  memory  nodes,  and attached processes, are queried and modified by reading or writing to the
       appropriate file in that cpuset's directory, as listed below.

       The pseudo-files in each cpuset directory are automatically created when the cpuset  is  created,  as  a  result  of  the
       mkdir(2) invocation.  It is not possible to directly add or remove these pseudo-files.

       A  cpuset  directory  that  contains  no  child  cpuset  directories, and has no attached processes, can be removed using
       rmdir(2) or rmdir(1).  It is not necessary, or possible, to remove the pseudo-files inside the directory before  removing
       it.

       The pseudo-files in each cpuset directory are small text files that may be read and written using traditional shell util-
       ities such as cat(1), and echo(1), or from a program by using file  I/O  library  functions  or  system  calls,  such  as
       open(2), read(2), write(2), and close(2).

       The  pseudo-files  in  a  cpuset  directory represent internal kernel state and do not have any persistent image on disk.
       Each of these per-cpuset files is listed and described below.

       tasks  List of the process IDs (PIDs) of the processes in that cpuset.  The list is formatted as a series of ASCII  deci-
              mal  numbers,  each followed by a newline.  A process may be added to a cpuset (automatically removing it from the
              cpuset that previously contained it) by writing its PID to that cpuset's tasks file (with or  without  a  trailing
              newline.)

              Warning:  only one PID may be written to the tasks file at a time.  If a string is written that contains more than
              one PID, only the first one will be used.

       notify_on_release
              Flag (0 or 1).  If set (1), that cpuset will receive special handling after it is released,  that  is,  after  all
              processes  cease  using  it  (i.e., terminate or are moved to a different cpuset) and all child cpuset directories
              have been removed.  See the Notify On Release section, below.

       cpus   List of the physical numbers of the CPUs on which processes in that cpuset are allowed to execute.  See List  For-
              mat below for a description of the format of cpus.

              The CPUs allowed to a cpuset may be changed by writing a new list to its cpus file.

       cpu_exclusive
              Flag  (0  or  1).   If  set (1), the cpuset has exclusive use of its CPUs (no sibling or cousin cpuset may overlap
              CPUs).  By default this is off (0).  Newly created cpusets also initially default this to off (0).

              Two cpusets are sibling cpusets if they share the same parent cpuset in the /dev/cpuset  hierarchy.   Two  cpusets
              are  cousin  cpusets  if  neither  is  the ancestor of the other.  Regardless of the cpu_exclusive setting, if one
              cpuset is the ancestor of another, and if both of these cpusets have nonempty cpus, then their cpus must  overlap,
              because the cpus of any cpuset are always a subset of the cpus of its parent cpuset.

       mems   List  of memory nodes on which processes in this cpuset are allowed to allocate memory.  See List Format below for
              a description of the format of mems.

       mem_exclusive
              Flag (0 or 1).  If set (1), the cpuset has exclusive use of its memory nodes (no sibling or cousin  may  overlap).
              Also  if set (1), the cpuset is a Hardwall cpuset (see below.)  By default this is off (0).  Newly created cpusets
              also initially default this to off (0).

              Regardless of the mem_exclusive setting, if one cpuset is the ancestor of another, then their  memory  nodes  must
              overlap,  because  the  memory nodes of any cpuset are always a subset of the memory nodes of that cpuset's parent
              cpuset.

       mem_hardwall (since Linux 2.6.26)
              Flag (0 or 1).  If set (1), the cpuset is a Hardwall cpuset (see below.)  Unlike mem_exclusive, there is  no  con-
              straint  on  whether cpusets marked mem_hardwall may have overlapping memory nodes with sibling or cousin cpusets.
              By default this is off (0).  Newly created cpusets also initially default this to off (0).

       memory_migrate (since Linux 2.6.16)
              Flag (0 or 1).  If set (1), then memory migration is enabled.  By default this is off (0).  See the Memory  Migra-
              tion section, below.

       memory_pressure (since Linux 2.6.16)
              A  measure of how much memory pressure the processes in this cpuset are causing.  See the Memory Pressure section,
              below.  Unless memory_pressure_enabled is enabled, always has value zero (0).  This file is  read-only.   See  the
              WARNINGS section, below.

       memory_pressure_enabled (since Linux 2.6.16)
              Flag  (0 or 1).  This file is only present in the root cpuset, normally /dev/cpuset.  If set (1), the memory_pres-
              sure calculations are enabled for all cpusets in the system.  By default this is off (0).  See the Memory Pressure
              section, below.

       memory_spread_page (since Linux 2.6.17)
              Flag  (0  or 1).  If set (1), pages in the kernel page cache (file-system buffers) are uniformly spread across the
              cpuset.  By default this is off (0) in the top cpuset, and inherited from  the  parent  cpuset  in  newly  created
              cpusets.  See the Memory Spread section, below.

       memory_spread_slab (since Linux 2.6.17)
              Flag  (0  or  1).   If set (1), the kernel slab caches for file I/O (directory and inode structures) are uniformly
              spread across the cpuset.  By default this is off (0) in the top cpuset, and inherited from the parent  cpuset  in
              newly created cpusets.  See the Memory Spread section, below.

       sched_load_balance (since Linux 2.6.24)
              Flag  (0  or 1).  If set (1, the default) the kernel will automatically load balance processes in that cpuset over
              the allowed CPUs in that cpuset.  If cleared (0) the kernel will avoid load balancing processes  in  this  cpuset,
              unless some other cpuset with overlapping CPUs has its sched_load_balance flag set.  See Scheduler Load Balancing,
              below, for further details.

       sched_relax_domain_level (since Linux 2.6.26)
              Integer, between -1 and a small positive value.  The sched_relax_domain_level controls the width of the  range  of
              CPUs  over  which  the  kernel  scheduler  performs  immediate  rebalancing  of  runnable  tasks  across CPUs.  If
              sched_load_balance is disabled, then the setting of sched_relax_domain_level does not matter, as no such load bal-
              ancing  is done.  If sched_load_balance is enabled, then the higher the value of the sched_relax_domain_level, the
              wider the range of CPUs over which immediate load balancing is  attempted.   See  Scheduler  Relax  Domain  Level,
              below, for further details.

       In  addition  to  the  above  pseudo-files  in  each  directory  below  /dev/cpuset,  each  process  has  a  pseudo-file,
       /proc/<pid>/cpuset, that displays the path of the process's cpuset directory relative to the root of the cpuset file sys-
       tem.

       Also  the  /proc/<pid>/status file for each process has four added lines, displaying the process's Cpus_allowed (on which
       CPUs it may be scheduled) and Mems_allowed (on which memory nodes it may obtain memory), in the two formats  Mask  Format
       and List Format (see below) as shown in the following example:

              Cpus_allowed:   ffffffff,ffffffff,ffffffff,ffffffff
              Cpus_allowed_list:     0-127
              Mems_allowed:   ffffffff,ffffffff
              Mems_allowed_list:     0-63

       The "allowed" fields were added in Linux 2.6.24; the "allowed_list" fields were added in Linux 2.6.26.

EXTENDED CAPABILITIES
       In  addition to controlling which cpus and mems a process is allowed to use, cpusets provide the following extended capa-
       bilities.

   Exclusive Cpusets
       If a cpuset is marked cpu_exclusive or mem_exclusive, no other cpuset, other than a direct ancestor  or  descendant,  may
       share any of the same CPUs or memory nodes.

       A  cpuset  that is mem_exclusive restricts kernel allocations for buffer cache pages and other internal kernel data pages
       commonly shared by the kernel across multiple users.  All cpusets, whether mem_exclusive or not, restrict allocations  of
       memory  for user space.  This enables configuring a system so that several independent jobs can share common kernel data,
       while isolating each job's user allocation in its own cpuset.  To do this, construct a large mem_exclusive cpuset to hold
       all the jobs, and construct child, non-mem_exclusive cpusets for each individual job.  Only a small amount of kernel mem-
       ory, such as requests from interrupt handlers, is allowed to be placed on  memory  nodes  outside  even  a  mem_exclusive
       cpuset.

   Hardwall
       A cpuset that has mem_exclusive or mem_hardwall set is a hardwall cpuset.  A hardwall cpuset restricts kernel allocations
       for page, buffer, and other data commonly shared by the kernel across multiple users.  All cpusets, whether  hardwall  or
       not, restrict allocations of memory for user space.

       This  enables  configuring  a  system  so that several independent jobs can share common kernel data, such as file system
       pages, while isolating each job's user allocation in its own cpuset.  To do this, construct a large  hardwall  cpuset  to
       hold all the jobs, and construct child cpusets for each individual job which are not hardwall cpusets.

       Only  a  small  amount  of kernel memory, such as requests from interrupt handlers, is allowed to be taken outside even a
       hardwall cpuset.

   Notify On Release
       If the notify_on_release flag is enabled (1) in a cpuset, then whenever the last process in the cpuset leaves  (exits  or
       attaches  to  some  other  cpuset)  and  the last child cpuset of that cpuset is removed, the kernel will run the command
       /sbin/cpuset_release_agent, supplying the pathname (relative to the mount point of the cpuset file system) of  the  aban-
       doned cpuset.  This enables automatic removal of abandoned cpusets.

       The  default  value  of  notify_on_release in the root cpuset at system boot is disabled (0).  The default value of other
       cpusets at creation is the current value of their parent's notify_on_release setting.

       The command /sbin/cpuset_release_agent is invoked, with the name (/dev/cpuset relative path) of the to-be-released cpuset
       in argv[1].

       The usual contents of the command /sbin/cpuset_release_agent is simply the shell script:

           #!/bin/sh
           rmdir /dev/cpuset/$1

       As  with other flag values below, this flag can be changed by writing an ASCII number 0 or 1 (with optional trailing new-
       line) into the file, to clear or set the flag, respectively.

   Memory Pressure
       The memory_pressure of a cpuset provides a simple per-cpuset running average of the rate that the processes in  a  cpuset
       are attempting to free up in-use memory on the nodes of the cpuset to satisfy additional memory requests.

       This  enables  batch  managers  that are monitoring jobs running in dedicated cpusets to efficiently detect what level of
       memory pressure that job is causing.

       This is useful both on tightly managed systems running a wide mix of submitted jobs, which may  choose  to  terminate  or
       reprioritize  jobs  that are trying to use more memory than allowed on the nodes assigned them, and with tightly coupled,
       long-running, massively parallel scientific computing jobs that will dramatically fail to meet required performance goals
       if they start to use more memory than allowed to them.

       This  mechanism  provides  a  very economical way for the batch manager to monitor a cpuset for signs of memory pressure.
       It's up to the batch manager or other user code to decide what action to take if it detects signs of memory pressure.

       Unless memory pressure calculation is enabled by setting the pseudo-file /dev/cpuset/memory_pressure_enabled, it  is  not
       computed for any cpuset, and reads from any memory_pressure always return zero, as represented by the ASCII string "0\n".
       See the WARNINGS section, below.

       A per-cpuset, running average is employed for the following reasons:

       *  Because this meter is per-cpuset rather than per-process or per virtual memory region, the system load  imposed  by  a
          batch  scheduler  monitoring  this  metric  is sharply reduced on large systems, because a scan of the tasklist can be
          avoided on each set of queries.

       *  Because this meter is a running average rather than an accumulating counter, a batch scheduler can detect memory pres-
          sure with a single read, instead of having to read and accumulate results for a period of time.

       *  Because this meter is per-cpuset rather than per-process, the batch scheduler can obtain the key information -- memory
          pressure in a cpuset -- with a single read, rather than having to query and accumulate results over all  the  (dynami-
          cally changing) set of processes in the cpuset.

       The  memory_pressure  of  a cpuset is calculated using a per-cpuset simple digital filter that is kept within the kernel.
       For each cpuset, this filter tracks the recent rate at which processes attached to that cpuset enter  the  kernel  direct
       reclaim code.

       The  kernel  direct reclaim code is entered whenever a process has to satisfy a memory page request by first finding some
       other page to repurpose, due to lack of any readily available already free pages.  Dirty file system pages are repurposed
       by  first  writing  them  to disk.  Unmodified file system buffer pages are repurposed by simply dropping them, though if
       that page is needed again, it will have to be reread from disk.

       The memory_pressure file provides an integer number representing the recent (half-life of 10 seconds) rate of entries  to
       the direct reclaim code caused by any process in the cpuset, in units of reclaims attempted per second, times 1000.

   Memory Spread
       There are two Boolean flag files per cpuset that control where the kernel allocates pages for the file-system buffers and
       related in-kernel data structures.  They are called memory_spread_page and memory_spread_slab.

       If the per-cpuset Boolean flag file memory_spread_page is set, then the kernel will spread the file-system buffers  (page
       cache) evenly over all the nodes that the faulting process is allowed to use, instead of preferring to put those pages on
       the node where the process is running.

       If the per-cpuset Boolean flag file memory_spread_slab is set, then the kernel will spread some file-system-related  slab
       caches, such as those for inodes and directory entries, evenly over all the nodes that the faulting process is allowed to
       use, instead of preferring to put those pages on the node where the process is running.

       The setting of these flags does not affect the data segment (see brk(2)) or stack segment pages of a process.

       By default, both kinds of memory spreading are off and the kernel prefers to allocate memory pages on the node  local  to
       where  the requesting process is running.  If that node is not allowed by the process's NUMA memory policy or cpuset con-
       figuration or if there are insufficient free memory pages on that node, then the kernel looks for the nearest  node  that
       is allowed and has sufficient free memory.

       When new cpusets are created, they inherit the memory spread settings of their parent.

       Setting memory spreading causes allocations for the affected page or slab caches to ignore the process's NUMA memory pol-
       icy and be spread instead.  However, the effect of these changes in memory placement caused  by  cpuset-specified  memory
       spreading  is  hidden  from  the mbind(2) or set_mempolicy(2) calls.  These two NUMA memory policy calls always appear to
       behave as if no cpuset-specified memory spreading is in effect, even if it is.  If  cpuset  memory  spreading  is  subse-
       quently turned off, the NUMA memory policy most recently specified by these calls is automatically reapplied.

       Both  memory_spread_page  and  memory_spread_slab  are Boolean flag files.  By default they contain "0", meaning that the
       feature is off for that cpuset.  If a "1" is written to that file, that turns the named feature on.

       Cpuset-specified memory spreading behaves similarly to what is known (in other contexts)  as  round-robin  or  interleave
       memory placement.

       Cpuset-specified memory spreading can provide substantial performance improvements for jobs that:

       a) need  to  place thread-local data on memory nodes close to the CPUs which are running the threads that most frequently
          access that data; but also

       b) need to access large file-system data sets that must to be spread across the several nodes  in  the  job's  cpuset  in
          order to fit.

       Without  this  policy,  the memory allocation across the nodes in the job's cpuset can become very uneven, especially for
       jobs that might have just a single thread initializing or reading in the data set.

   Memory Migration
       Normally, under the default setting (disabled) of memory_migrate, once a page is allocated (given a physical page of main
       memory)  then  that  page  stays on whatever node it was allocated, so long as it remains allocated, even if the cpuset's
       memory-placement policy mems subsequently changes.

       When memory migration is enabled in a cpuset, if the mems setting of the cpuset is changed, then any memory page  in  use
       by any process in the cpuset that is on a memory node that is no longer allowed will be migrated to a memory node that is
       allowed.

       Furthermore, if a process is moved into a cpuset with memory_migrate enabled, any memory pages it uses that were on  mem-
       ory  nodes allowed in its previous cpuset, but which are not allowed in its new cpuset, will be migrated to a memory node
       allowed in the new cpuset.

       The relative placement of a migrated page within the cpuset is preserved during these migration operations  if  possible.
       For  example,  if  the  page was on the second valid node of the prior cpuset, then the page will be placed on the second
       valid node of the new cpuset, if possible.

   Scheduler Load Balancing
       The kernel scheduler automatically load balances processes.  If one CPU is underutilized, the kernel will look  for  pro-
       cesses  on  other  more overloaded CPUs and move those processes to the underutilized CPU, within the constraints of such
       placement mechanisms as cpusets and sched_setaffinity(2).

       The algorithmic cost of load balancing and its impact on key shared kernel data  structures  such  as  the  process  list
       increases  more  than linearly with the number of CPUs being balanced.  For example, it costs more to load balance across
       one large set of CPUs than it does to balance across two smaller sets of CPUs, each of half the size of the  larger  set.
       (The precise relationship between the number of CPUs being balanced and the cost of load balancing depends on implementa-
       tion details of the kernel process scheduler, which is subject to change over time, as improved  kernel  scheduler  algo-
       rithms are implemented.)

       The  per-cpuset flag sched_load_balance provides a mechanism to suppress this automatic scheduler load balancing in cases
       where it is not needed and suppressing it would have worthwhile performance benefits.

       By default, load balancing is done across all CPUs, except those marked isolated using the kernel boot  time  "isolcpus="
       argument.  (See Scheduler Relax Domain Level, below, to change this default.)

       This default load balancing across all CPUs is not well suited to the following two situations:

       *  On large systems, load balancing across many CPUs is expensive.  If the system is managed using cpusets to place inde-
          pendent jobs on separate sets of CPUs, full load balancing is unnecessary.

       *  Systems supporting real-time on some CPUs need to minimize system overhead on those CPUs, including  avoiding  process
          load balancing if that is not needed.

       When  the  per-cpuset flag sched_load_balance is enabled (the default setting), it requests load balancing across all the
       CPUs in that cpuset's allowed CPUs, ensuring that load balancing  can  move  a  process  (not  otherwise  pinned,  as  by
       sched_setaffinity(2)) from any CPU in that cpuset to any other.

       When  the per-cpuset flag sched_load_balance is disabled, then the scheduler will avoid load balancing across the CPUs in
       that cpuset, except in so far as is necessary because some overlapping cpuset has sched_load_balance enabled.

       So, for example, if the top cpuset has the flag sched_load_balance enabled, then the scheduler will load  balance  across
       all CPUs, and the setting of the sched_load_balance flag in other cpusets has no effect, as we're already fully load bal-
       ancing.

       Therefore in the above two situations, the flag sched_load_balance should be disabled in the top cpuset, and only some of
       the smaller, child cpusets would have this flag enabled.

       When  doing  this,  you  don't  usually  want to leave any unpinned processes in the top cpuset that might use nontrivial
       amounts of CPU, as such processes may be artificially constrained to some subset of CPUs, depending on the particulars of
       this  flag setting in descendant cpusets.  Even if such a process could use spare CPU cycles in some other CPUs, the ker-
       nel scheduler might not consider the possibility of load balancing that process to the underused CPU.

       Of course, processes pinned to a particular CPU can be left in a cpuset that disables sched_load_balance  as  those  pro-
       cesses aren't going anywhere else anyway.

   Scheduler Relax Domain Level
       The  kernel  scheduler  performs  immediate  load balancing whenever a CPU becomes free or another task becomes runnable.
       This load balancing works to ensure that as many CPUs as possible are usefully employed running tasks.  The  kernel  also
       performs  periodic  load  balancing off the software clock described in time(7).  The setting of sched_relax_domain_level
       only applies to immediate load balancing.  Regardless of the sched_relax_domain_level setting, periodic load balancing is
       attempted  over all CPUs (unless disabled by turning off sched_load_balance.)  In any case, of course, tasks will only be
       scheduled to run on CPUs allowed by their cpuset, as modified by sched_setaffinity(2) system calls.

       On small systems, such as those with just a few CPUs, immediate load balancing is useful to improve system  interactivity
       and  to minimize wasteful idle CPU cycles.  But on large systems, attempting immediate load balancing across a large num-
       ber of CPUs can be more costly than it is worth, depending on the particular performance characteristics of the  job  mix
       and the hardware.

       The  exact meaning of the small integer values of sched_relax_domain_level will depend on internal implementation details
       of the kernel scheduler code and on the non-uniform architecture of the hardware.  Both of these will  evolve  over  time
       and vary by system architecture and kernel version.

       As  of  this writing, when this capability was introduced in Linux 2.6.26, on certain popular architectures, the positive
       values of sched_relax_domain_level have the following meanings.

       (1) Perform immediate load balancing across Hyper-Thread siblings on the same core.
       (2) Perform immediate load balancing across other cores in the same package.
       (3) Perform immediate load balancing across other CPUs on the same node or blade.
       (4) Perform immediate load balancing across over several (implementation detail) nodes [On NUMA systems].
       (5) Perform immediate load balancing across over all CPUs in system [On NUMA systems].

       The sched_relax_domain_level value of zero (0) always means don't perform immediate load balancing, hence that load  bal-
       ancing is only done periodically, not immediately when a CPU becomes available or another task becomes runnable.

       The sched_relax_domain_level value of minus one (-1) always means use the system default value.  The system default value
       can vary  by  architecture  and  kernel  version.   This  system  default  value  can  be  changed  by  kernel  boot-time
       "relax_domain_level=" argument.

       In the case of multiple overlapping cpusets which have conflicting sched_relax_domain_level values, then the highest such
       value applies to all CPUs in any of the overlapping cpusets.  In such cases, the value  minus  one  (-1)  is  the  lowest
       value, overridden by any other value, and the value zero (0) is the next lowest value.

FORMATS
       The following formats are used to represent sets of CPUs and memory nodes.

   Mask Format
       The Mask Format is used to represent CPU and memory-node bitmasks in the /proc/<pid>/status file.

       This  format  displays each 32-bit word in hexadecimal (using ASCII characters "0" - "9" and "a" - "f"); words are filled
       with leading zeros, if required.  For masks longer than one word, a comma separator is used  between  words.   Words  are
       displayed  in  big-endian order, which has the most significant bit first.  The hex digits within a word are also in big-
       endian order.

       The number of 32-bit words displayed is the minimum number needed to display all bits of the bitmask, based on  the  size
       of the bitmask.

       Examples of the Mask Format:

              00000001                        # just bit 0 set
              40000000,00000000,00000000      # just bit 94 set
              00000001,00000000,00000000      # just bit 64 set
              000000ff,00000000               # bits 32-39 set
              00000000,000E3862               # 1,5,6,11-13,17-19 set

       A mask with bits 0, 1, 2, 4, 8, 16, 32, and 64 set displays as:

              00000001,00000001,00010117

       The  first "1" is for bit 64, the second for bit 32, the third for bit 16, the fourth for bit 8, the fifth for bit 4, and
       the "7" is for bits 2, 1, and 0.

   List Format
       The List Format for cpus and mems is a comma-separated list of CPU or memory-node numbers and ranges of numbers, in ASCII
       decimal.

       Examples of the List Format:

              0-4,9           # bits 0, 1, 2, 3, 4, and 9 set
              0-2,7,12-14     # bits 0, 1, 2, 7, 12, 13, and 14 set

RULES
       The following rules apply to each cpuset:

       *  Its CPUs and memory nodes must be a (possibly equal) subset of its parent's.

       *  It can only be marked cpu_exclusive if its parent is.

       *  It can only be marked mem_exclusive if its parent is.

       *  If it is cpu_exclusive, its CPUs may not overlap any sibling.

       *  If it is memory_exclusive, its memory nodes may not overlap any sibling.

PERMISSIONS
       The permissions of a cpuset are determined by the permissions of the directories and pseudo-files in the cpuset file sys-
       tem, normally mounted at /dev/cpuset.

       For instance, a process can put itself in some other cpuset (than its current one) if it can write  the  tasks  file  for
       that cpuset.  This requires execute permission on the encompassing directories and write permission on the tasks file.

       An  additional  constraint  is  applied  to requests to place some other process in a cpuset.  One process may not attach
       another to a cpuset unless it would have permission to send that process a signal (see kill(2)).

       A process may create a child cpuset if it can access and write the parent cpuset directory.  It can modify  the  CPUs  or
       memory nodes in a cpuset if it can access that cpuset's directory (execute permissions on the each of the parent directo-
       ries) and write the corresponding cpus or mems file.

       There is one minor difference between the manner in which these permissions are evaluated and the manner in which  normal
       file-system  operation  permissions are evaluated.  The kernel interprets relative pathnames starting at a process's cur-
       rent working directory.  Even if one is operating on a cpuset file, relative pathnames are interpreted  relative  to  the
       process's current working directory, not relative to the process's current cpuset.  The only ways that cpuset paths rela-
       tive to a process's current cpuset can be used are if either the process's current working directory is  its  cpuset  (it
       first did a cd or chdir(2) to its cpuset directory beneath /dev/cpuset, which is a bit unusual) or if some user code con-
       verts the relative cpuset path to a full file-system path.

       In theory, this means that user code should specify cpusets using absolute pathnames, which requires  knowing  the  mount
       point  of the cpuset file system (usually, but not necessarily, /dev/cpuset).  In practice, all user level code that this
       author is aware of simply assumes that if the cpuset file system is mounted, then it is mounted at /dev/cpuset.  Further-
       more,  it  is common practice for carefully written user code to verify the presence of the pseudo-file /dev/cpuset/tasks
       in order to verify that the cpuset pseudo-file system is currently mounted.

WARNINGS
   Enabling memory_pressure
       By default, the per-cpuset file memory_pressure always contains zero (0).  Unless this feature is enabled by writing  "1"
       to the pseudo-file /dev/cpuset/memory_pressure_enabled, the kernel does not compute per-cpuset memory_pressure.

   Using the echo command
       When  using the echo command at the shell prompt to change the values of cpuset files, beware that the built-in echo com-
       mand in some shells does not display an error message if the write(2) system call fails.  For example, if the command:

           echo 19 > mems

       failed because memory node 19 was not allowed (perhaps the current system does not have a memory node 19), then the  echo
       command  might not display any error.  It is better to use the /bin/echo external command to change cpuset file settings,
       as this command will display write(2) errors, as in the example:

           /bin/echo 19 > mems
           /bin/echo: write error: Invalid argument

EXCEPTIONS
   Memory placement
       Not all allocations of system memory are constrained by cpusets, for the following reasons.

       If hot-plug functionality is used to remove all the CPUs that are currently assigned to a cpuset, then  the  kernel  will
       automatically  update  the  cpus_allowed of all processes attached to CPUs in that cpuset to allow all CPUs.  When memory
       hot-plug functionality for removing memory nodes is available, a similar exception is expected to apply  there  as  well.
       In  general,  the kernel prefers to violate cpuset placement, rather than starving a process that has had all its allowed
       CPUs or memory nodes taken offline.  User code should reconfigure cpusets to only refer to online CPUs and  memory  nodes
       when using hot-plug to add or remove such resources.

       A few kernel-critical, internal memory-allocation requests, marked GFP_ATOMIC, must be satisfied immediately.  The kernel
       may drop some request or malfunction if one of these allocations fail.  If such a request cannot be satisfied within  the
       current  process's cpuset, then we relax the cpuset, and look for memory anywhere we can find it.  It's better to violate
       the cpuset than stress the kernel.

       Allocations of memory requested by kernel drivers while processing an interrupt lack any relevant  process  context,  and
       are not confined by cpusets.

   Renaming cpusets
       You  can  use the rename(2) system call to rename cpusets.  Only simple renaming is supported; that is, changing the name
       of a cpuset directory is permitted, but moving a directory into a different directory is not permitted.

ERRORS
       The Linux kernel implementation of cpusets sets errno to specify the reason for a failed system call affecting cpusets.

       The possible errno settings and their meaning when set on a failed cpuset call are as listed below.

       E2BIG  Attempted a write(2) on a special cpuset file with a length larger than some kernel-determined upper limit on  the
              length of such writes.

       EACCES Attempted  to  write(2) the process ID (PID) of a process to a cpuset tasks file when one lacks permission to move
              that process.

       EACCES Attempted to add, using write(2), a CPU or memory node to a cpuset, when that CPU or memory node was  not  already
              in its parent.

       EACCES Attempted to set, using write(2), cpu_exclusive or mem_exclusive on a cpuset whose parent lacks the same setting.

       EACCES Attempted to write(2) a memory_pressure file.

       EACCES Attempted to create a file in a cpuset directory.

       EBUSY  Attempted to remove, using rmdir(2), a cpuset with attached processes.

       EBUSY  Attempted to remove, using rmdir(2), a cpuset with child cpusets.

       EBUSY  Attempted to remove a CPU or memory node from a cpuset that is also in a child of that cpuset.

       EEXIST Attempted to create, using mkdir(2), a cpuset that already exists.

       EEXIST Attempted to rename(2) a cpuset to a name that already exists.

       EFAULT Attempted  to  read(2)  or  write(2) a cpuset file using a buffer that is outside the writing processes accessible
              address space.

       EINVAL Attempted to change a cpuset, using write(2), in a  way  that  would  violate  a  cpu_exclusive  or  mem_exclusive
              attribute of that cpuset or any of its siblings.

       EINVAL Attempted to write(2) an empty cpus or mems list to a cpuset which has attached processes or child cpusets.

       EINVAL Attempted  to  write(2)  a  cpus or mems list which included a range with the second number smaller than the first
              number.

       EINVAL Attempted to write(2) a cpus or mems list which included an invalid character in the string.

       EINVAL Attempted to write(2) a list to a cpus file that did not include any online CPUs.

       EINVAL Attempted to write(2) a list to a mems file that did not include any online memory nodes.

       EINVAL Attempted to write(2) a list to a mems file that included a node that held no memory.

       EIO    Attempted to write(2) a string to a cpuset tasks file that does not begin with an ASCII decimal integer.

       EIO    Attempted to rename(2) a cpuset into a different directory.

       ENAMETOOLONG
              Attempted to read(2) a /proc/<pid>/cpuset file for a cpuset path that is longer than the kernel page size.

       ENAMETOOLONG
              Attempted to create, using mkdir(2), a cpuset whose base directory name is longer than 255 characters.

       ENAMETOOLONG
              Attempted to create, using  mkdir(2),  a  cpuset  whose  full  pathname,  including  the  mount  point  (typically
              "/dev/cpuset/") prefix, is longer than 4095 characters.

       ENODEV The  cpuset was removed by another process at the same time as a write(2) was attempted on one of the pseudo-files
              in the cpuset directory.

       ENOENT Attempted to create, using mkdir(2), a cpuset in a parent cpuset that doesn't exist.

       ENOENT Attempted to access(2) or open(2) a nonexistent file in a cpuset directory.

       ENOMEM Insufficient memory is available within the kernel; can occur on a variety of system calls affecting cpusets,  but
              only if the system is extremely short of memory.

       ENOSPC Attempted  to  write(2) the process ID (PID) of a process to a cpuset tasks file when the cpuset had an empty cpus
              or empty mems setting.

       ENOSPC Attempted to write(2) an empty cpus or mems setting to a cpuset that has tasks attached.

       ENOTDIR
              Attempted to rename(2) a nonexistent cpuset.

       EPERM  Attempted to remove a file from a cpuset directory.

       ERANGE Specified a cpus or mems list to the kernel which included a number too large for the kernel to set  in  its  bit-
              masks.

       ESRCH  Attempted to write(2) the process ID (PID) of a nonexistent process to a cpuset tasks file.

VERSIONS
       Cpusets appeared in version 2.6.12 of the Linux kernel.

NOTES
       Despite  its  name,  the  pid parameter is actually a thread ID, and each thread in a threaded group can be attached to a
       different cpuset.  The value returned from a call to gettid(2) can be passed in the argument pid.

BUGS
       memory_pressure cpuset files can be opened for writing, creation, or truncation, but then the write(2) fails  with  errno
       set to EACCES, and the creation and truncation options on open(2) have no effect.

EXAMPLE
       The following examples demonstrate querying and setting cpuset options using shell commands.

   Creating and attaching to a cpuset.
       To create a new cpuset and attach the current command shell to it, the steps are:

       1)  mkdir /dev/cpuset (if not already done)
       2)  mount -t cpuset none /dev/cpuset (if not already done)
       3)  Create the new cpuset using mkdir(1).
       4)  Assign CPUs and memory nodes to the new cpuset.
       5)  Attach the shell to the new cpuset.

       For  example,  the following sequence of commands will set up a cpuset named "Charlie", containing just CPUs 2 and 3, and
       memory node 1, and then attach the current shell to that cpuset.

           $ mkdir /dev/cpuset
           $ mount -t cpuset cpuset /dev/cpuset
           $ cd /dev/cpuset
           $ mkdir Charlie
           $ cd Charlie
           $ /bin/echo 2-3 > cpus
           $ /bin/echo 1 > mems
           $ /bin/echo $$ > tasks
           # The current shell is now running in cpuset Charlie
           # The next line should display '/Charlie'
           $ cat /proc/self/cpuset

   Migrating a job to different memory nodes.
       To migrate a job (the set of processes attached to a cpuset) to different CPUs and memory nodes in the system,  including
       moving the memory pages currently allocated to that job, perform the following steps.

       1)  Let's  say  we  want to move the job in cpuset alpha (CPUs 4-7 and memory nodes 2-3) to a new cpuset beta (CPUs 16-19
           and memory nodes 8-9).
       2)  First create the new cpuset beta.
       3)  Then allow CPUs 16-19 and memory nodes 8-9 in beta.
       4)  Then enable memory_migration in beta.
       5)  Then move each process from alpha to beta.

       The following sequence of commands accomplishes this.

           $ cd /dev/cpuset
           $ mkdir beta
           $ cd beta
           $ /bin/echo 16-19 > cpus
           $ /bin/echo 8-9 > mems
           $ /bin/echo 1 > memory_migrate
           $ while read i; do /bin/echo $i; done < ../alpha/tasks > tasks

       The above should move any processes in alpha to beta, and any memory held by these processes on memory nodes 2-3 to  mem-
       ory nodes 8-9, respectively.

       Notice that the last step of the above sequence did not do:

           $ cp ../alpha/tasks tasks

       The  while loop, rather than the seemingly easier use of the cp(1) command, was necessary because only one process PID at
       a time may be written to the tasks file.

       The same effect (writing one PID at a time) as the while loop can be accomplished more efficiently, in  fewer  keystrokes
       and in syntax that works on any shell, but alas more obscurely, by using the -u (unbuffered) option of sed(1):

           $ sed -un p < ../alpha/tasks > tasks

SEE ALSO
       taskset(1),  get_mempolicy(2),  getcpu(2),  mbind(2),  sched_getaffinity(2), sched_setaffinity(2), sched_setscheduler(2),
       set_mempolicy(2), CPU_SET(3), proc(5), numa(7), migratepages(8), numactl(8)

       The kernel source file Documentation/cpusets.txt.

COLOPHON
       This page is part of release 3.25 of the Linux man-pages project.  A description of the project,  and  information  about
       reporting bugs, can be found at http://www.kernel.org/doc/man-pages/.



Linux                                                      2008-11-12                                                  CPUSET(7)

Valid XHTML 1.0!Valid CSS!