123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707708709710711712713714715716717718719720721722723724725726727728729730731732733734735736737738739740741742743744745746747748749750751752753754755756757758759760761762763764765766767768769770771772773774775776777778779780781782783784785786787788789790791792793794795796797798799800801802803804805806807808809810811812813814815816 |
- LIBNVDIMM: Non-Volatile Devices
- libnvdimm - kernel / libndctl - userspace helper library
- linux-nvdimm@lists.01.org
- v13
- Glossary
- Overview
- Supporting Documents
- Git Trees
- LIBNVDIMM PMEM and BLK
- Why BLK?
- PMEM vs BLK
- BLK-REGIONs, PMEM-REGIONs, Atomic Sectors, and DAX
- Example NVDIMM Platform
- LIBNVDIMM Kernel Device Model and LIBNDCTL Userspace API
- LIBNDCTL: Context
- libndctl: instantiate a new library context example
- LIBNVDIMM/LIBNDCTL: Bus
- libnvdimm: control class device in /sys/class
- libnvdimm: bus
- libndctl: bus enumeration example
- LIBNVDIMM/LIBNDCTL: DIMM (NMEM)
- libnvdimm: DIMM (NMEM)
- libndctl: DIMM enumeration example
- LIBNVDIMM/LIBNDCTL: Region
- libnvdimm: region
- libndctl: region enumeration example
- Why Not Encode the Region Type into the Region Name?
- How Do I Determine the Major Type of a Region?
- LIBNVDIMM/LIBNDCTL: Namespace
- libnvdimm: namespace
- libndctl: namespace enumeration example
- libndctl: namespace creation example
- Why the Term "namespace"?
- LIBNVDIMM/LIBNDCTL: Block Translation Table "btt"
- libnvdimm: btt layout
- libndctl: btt creation example
- Summary LIBNDCTL Diagram
- Glossary
- --------
- PMEM: A system-physical-address range where writes are persistent. A
- block device composed of PMEM is capable of DAX. A PMEM address range
- may span an interleave of several DIMMs.
- BLK: A set of one or more programmable memory mapped apertures provided
- by a DIMM to access its media. This indirection precludes the
- performance benefit of interleaving, but enables DIMM-bounded failure
- modes.
- DPA: DIMM Physical Address, is a DIMM-relative offset. With one DIMM in
- the system there would be a 1:1 system-physical-address:DPA association.
- Once more DIMMs are added a memory controller interleave must be
- decoded to determine the DPA associated with a given
- system-physical-address. BLK capacity always has a 1:1 relationship
- with a single-DIMM's DPA range.
- DAX: File system extensions to bypass the page cache and block layer to
- mmap persistent memory, from a PMEM block device, directly into a
- process address space.
- DSM: Device Specific Method: ACPI method to to control specific
- device - in this case the firmware.
- DCR: NVDIMM Control Region Structure defined in ACPI 6 Section 5.2.25.5.
- It defines a vendor-id, device-id, and interface format for a given DIMM.
- BTT: Block Translation Table: Persistent memory is byte addressable.
- Existing software may have an expectation that the power-fail-atomicity
- of writes is at least one sector, 512 bytes. The BTT is an indirection
- table with atomic update semantics to front a PMEM/BLK block device
- driver and present arbitrary atomic sector sizes.
- LABEL: Metadata stored on a DIMM device that partitions and identifies
- (persistently names) storage between PMEM and BLK. It also partitions
- BLK storage to host BTTs with different parameters per BLK-partition.
- Note that traditional partition tables, GPT/MBR, are layered on top of a
- BLK or PMEM device.
- Overview
- --------
- The LIBNVDIMM subsystem provides support for three types of NVDIMMs, namely,
- PMEM, BLK, and NVDIMM devices that can simultaneously support both PMEM
- and BLK mode access. These three modes of operation are described by
- the "NVDIMM Firmware Interface Table" (NFIT) in ACPI 6. While the LIBNVDIMM
- implementation is generic and supports pre-NFIT platforms, it was guided
- by the superset of capabilities need to support this ACPI 6 definition
- for NVDIMM resources. The bulk of the kernel implementation is in place
- to handle the case where DPA accessible via PMEM is aliased with DPA
- accessible via BLK. When that occurs a LABEL is needed to reserve DPA
- for exclusive access via one mode a time.
- Supporting Documents
- ACPI 6: http://www.uefi.org/sites/default/files/resources/ACPI_6.0.pdf
- NVDIMM Namespace: http://pmem.io/documents/NVDIMM_Namespace_Spec.pdf
- DSM Interface Example: http://pmem.io/documents/NVDIMM_DSM_Interface_Example.pdf
- Driver Writer's Guide: http://pmem.io/documents/NVDIMM_Driver_Writers_Guide.pdf
- Git Trees
- LIBNVDIMM: https://git.kernel.org/cgit/linux/kernel/git/djbw/nvdimm.git
- LIBNDCTL: https://github.com/pmem/ndctl.git
- PMEM: https://github.com/01org/prd
- LIBNVDIMM PMEM and BLK
- ------------------
- Prior to the arrival of the NFIT, non-volatile memory was described to a
- system in various ad-hoc ways. Usually only the bare minimum was
- provided, namely, a single system-physical-address range where writes
- are expected to be durable after a system power loss. Now, the NFIT
- specification standardizes not only the description of PMEM, but also
- BLK and platform message-passing entry points for control and
- configuration.
- For each NVDIMM access method (PMEM, BLK), LIBNVDIMM provides a block
- device driver:
- 1. PMEM (nd_pmem.ko): Drives a system-physical-address range. This
- range is contiguous in system memory and may be interleaved (hardware
- memory controller striped) across multiple DIMMs. When interleaved the
- platform may optionally provide details of which DIMMs are participating
- in the interleave.
- Note that while LIBNVDIMM describes system-physical-address ranges that may
- alias with BLK access as ND_NAMESPACE_PMEM ranges and those without
- alias as ND_NAMESPACE_IO ranges, to the nd_pmem driver there is no
- distinction. The different device-types are an implementation detail
- that userspace can exploit to implement policies like "only interface
- with address ranges from certain DIMMs". It is worth noting that when
- aliasing is present and a DIMM lacks a label, then no block device can
- be created by default as userspace needs to do at least one allocation
- of DPA to the PMEM range. In contrast ND_NAMESPACE_IO ranges, once
- registered, can be immediately attached to nd_pmem.
- 2. BLK (nd_blk.ko): This driver performs I/O using a set of platform
- defined apertures. A set of apertures will access just one DIMM.
- Multiple windows (apertures) allow multiple concurrent accesses, much like
- tagged-command-queuing, and would likely be used by different threads or
- different CPUs.
- The NFIT specification defines a standard format for a BLK-aperture, but
- the spec also allows for vendor specific layouts, and non-NFIT BLK
- implementations may have other designs for BLK I/O. For this reason
- "nd_blk" calls back into platform-specific code to perform the I/O.
- One such implementation is defined in the "Driver Writer's Guide" and "DSM
- Interface Example".
- Why BLK?
- --------
- While PMEM provides direct byte-addressable CPU-load/store access to
- NVDIMM storage, it does not provide the best system RAS (recovery,
- availability, and serviceability) model. An access to a corrupted
- system-physical-address address causes a CPU exception while an access
- to a corrupted address through an BLK-aperture causes that block window
- to raise an error status in a register. The latter is more aligned with
- the standard error model that host-bus-adapter attached disks present.
- Also, if an administrator ever wants to replace a memory it is easier to
- service a system at DIMM module boundaries. Compare this to PMEM where
- data could be interleaved in an opaque hardware specific manner across
- several DIMMs.
- PMEM vs BLK
- BLK-apertures solve these RAS problems, but their presence is also the
- major contributing factor to the complexity of the ND subsystem. They
- complicate the implementation because PMEM and BLK alias in DPA space.
- Any given DIMM's DPA-range may contribute to one or more
- system-physical-address sets of interleaved DIMMs, *and* may also be
- accessed in its entirety through its BLK-aperture. Accessing a DPA
- through a system-physical-address while simultaneously accessing the
- same DPA through a BLK-aperture has undefined results. For this reason,
- DIMMs with this dual interface configuration include a DSM function to
- store/retrieve a LABEL. The LABEL effectively partitions the DPA-space
- into exclusive system-physical-address and BLK-aperture accessible
- regions. For simplicity a DIMM is allowed a PMEM "region" per each
- interleave set in which it is a member. The remaining DPA space can be
- carved into an arbitrary number of BLK devices with discontiguous
- extents.
- BLK-REGIONs, PMEM-REGIONs, Atomic Sectors, and DAX
- --------------------------------------------------
- One of the few
- reasons to allow multiple BLK namespaces per REGION is so that each
- BLK-namespace can be configured with a BTT with unique atomic sector
- sizes. While a PMEM device can host a BTT the LABEL specification does
- not provide for a sector size to be specified for a PMEM namespace.
- This is due to the expectation that the primary usage model for PMEM is
- via DAX, and the BTT is incompatible with DAX. However, for the cases
- where an application or filesystem still needs atomic sector update
- guarantees it can register a BTT on a PMEM device or partition. See
- LIBNVDIMM/NDCTL: Block Translation Table "btt"
- Example NVDIMM Platform
- -----------------------
- For the remainder of this document the following diagram will be
- referenced for any example sysfs layouts.
- (a) (b) DIMM BLK-REGION
- +-------------------+--------+--------+--------+
- +------+ | pm0.0 | blk2.0 | pm1.0 | blk2.1 | 0 region2
- | imc0 +--+- - - region0- - - +--------+ +--------+
- +--+---+ | pm0.0 | blk3.0 | pm1.0 | blk3.1 | 1 region3
- | +-------------------+--------v v--------+
- +--+---+ | |
- | cpu0 | region1
- +--+---+ | |
- | +----------------------------^ ^--------+
- +--+---+ | blk4.0 | pm1.0 | blk4.0 | 2 region4
- | imc1 +--+----------------------------| +--------+
- +------+ | blk5.0 | pm1.0 | blk5.0 | 3 region5
- +----------------------------+--------+--------+
- In this platform we have four DIMMs and two memory controllers in one
- socket. Each unique interface (BLK or PMEM) to DPA space is identified
- by a region device with a dynamically assigned id (REGION0 - REGION5).
- 1. The first portion of DIMM0 and DIMM1 are interleaved as REGION0. A
- single PMEM namespace is created in the REGION0-SPA-range that spans most
- of DIMM0 and DIMM1 with a user-specified name of "pm0.0". Some of that
- interleaved system-physical-address range is reclaimed as BLK-aperture
- accessed space starting at DPA-offset (a) into each DIMM. In that
- reclaimed space we create two BLK-aperture "namespaces" from REGION2 and
- REGION3 where "blk2.0" and "blk3.0" are just human readable names that
- could be set to any user-desired name in the LABEL.
- 2. In the last portion of DIMM0 and DIMM1 we have an interleaved
- system-physical-address range, REGION1, that spans those two DIMMs as
- well as DIMM2 and DIMM3. Some of REGION1 is allocated to a PMEM namespace
- named "pm1.0", the rest is reclaimed in 4 BLK-aperture namespaces (for
- each DIMM in the interleave set), "blk2.1", "blk3.1", "blk4.0", and
- "blk5.0".
- 3. The portion of DIMM2 and DIMM3 that do not participate in the REGION1
- interleaved system-physical-address range (i.e. the DPA address past
- offset (b) are also included in the "blk4.0" and "blk5.0" namespaces.
- Note, that this example shows that BLK-aperture namespaces don't need to
- be contiguous in DPA-space.
- This bus is provided by the kernel under the device
- /sys/devices/platform/nfit_test.0 when CONFIG_NFIT_TEST is enabled and
- the nfit_test.ko module is loaded. This not only test LIBNVDIMM but the
- acpi_nfit.ko driver as well.
- LIBNVDIMM Kernel Device Model and LIBNDCTL Userspace API
- ----------------------------------------------------
- What follows is a description of the LIBNVDIMM sysfs layout and a
- corresponding object hierarchy diagram as viewed through the LIBNDCTL
- API. The example sysfs paths and diagrams are relative to the Example
- NVDIMM Platform which is also the LIBNVDIMM bus used in the LIBNDCTL unit
- test.
- LIBNDCTL: Context
- Every API call in the LIBNDCTL library requires a context that holds the
- logging parameters and other library instance state. The library is
- based on the libabc template:
- https://git.kernel.org/cgit/linux/kernel/git/kay/libabc.git
- LIBNDCTL: instantiate a new library context example
- struct ndctl_ctx *ctx;
- if (ndctl_new(&ctx) == 0)
- return ctx;
- else
- return NULL;
- LIBNVDIMM/LIBNDCTL: Bus
- -------------------
- A bus has a 1:1 relationship with an NFIT. The current expectation for
- ACPI based systems is that there is only ever one platform-global NFIT.
- That said, it is trivial to register multiple NFITs, the specification
- does not preclude it. The infrastructure supports multiple busses and
- we we use this capability to test multiple NFIT configurations in the
- unit test.
- LIBNVDIMM: control class device in /sys/class
- This character device accepts DSM messages to be passed to DIMM
- identified by its NFIT handle.
- /sys/class/nd/ndctl0
- |-- dev
- |-- device -> ../../../ndbus0
- |-- subsystem -> ../../../../../../../class/nd
- LIBNVDIMM: bus
- struct nvdimm_bus *nvdimm_bus_register(struct device *parent,
- struct nvdimm_bus_descriptor *nfit_desc);
- /sys/devices/platform/nfit_test.0/ndbus0
- |-- commands
- |-- nd
- |-- nfit
- |-- nmem0
- |-- nmem1
- |-- nmem2
- |-- nmem3
- |-- power
- |-- provider
- |-- region0
- |-- region1
- |-- region2
- |-- region3
- |-- region4
- |-- region5
- |-- uevent
- `-- wait_probe
- LIBNDCTL: bus enumeration example
- Find the bus handle that describes the bus from Example NVDIMM Platform
- static struct ndctl_bus *get_bus_by_provider(struct ndctl_ctx *ctx,
- const char *provider)
- {
- struct ndctl_bus *bus;
- ndctl_bus_foreach(ctx, bus)
- if (strcmp(provider, ndctl_bus_get_provider(bus)) == 0)
- return bus;
- return NULL;
- }
- bus = get_bus_by_provider(ctx, "nfit_test.0");
- LIBNVDIMM/LIBNDCTL: DIMM (NMEM)
- ---------------------------
- The DIMM device provides a character device for sending commands to
- hardware, and it is a container for LABELs. If the DIMM is defined by
- NFIT then an optional 'nfit' attribute sub-directory is available to add
- NFIT-specifics.
- Note that the kernel device name for "DIMMs" is "nmemX". The NFIT
- describes these devices via "Memory Device to System Physical Address
- Range Mapping Structure", and there is no requirement that they actually
- be physical DIMMs, so we use a more generic name.
- LIBNVDIMM: DIMM (NMEM)
- struct nvdimm *nvdimm_create(struct nvdimm_bus *nvdimm_bus, void *provider_data,
- const struct attribute_group **groups, unsigned long flags,
- unsigned long *dsm_mask);
- /sys/devices/platform/nfit_test.0/ndbus0
- |-- nmem0
- | |-- available_slots
- | |-- commands
- | |-- dev
- | |-- devtype
- | |-- driver -> ../../../../../bus/nd/drivers/nvdimm
- | |-- modalias
- | |-- nfit
- | | |-- device
- | | |-- format
- | | |-- handle
- | | |-- phys_id
- | | |-- rev_id
- | | |-- serial
- | | `-- vendor
- | |-- state
- | |-- subsystem -> ../../../../../bus/nd
- | `-- uevent
- |-- nmem1
- [..]
- LIBNDCTL: DIMM enumeration example
- Note, in this example we are assuming NFIT-defined DIMMs which are
- identified by an "nfit_handle" a 32-bit value where:
- Bit 3:0 DIMM number within the memory channel
- Bit 7:4 memory channel number
- Bit 11:8 memory controller ID
- Bit 15:12 socket ID (within scope of a Node controller if node controller is present)
- Bit 27:16 Node Controller ID
- Bit 31:28 Reserved
- static struct ndctl_dimm *get_dimm_by_handle(struct ndctl_bus *bus,
- unsigned int handle)
- {
- struct ndctl_dimm *dimm;
- ndctl_dimm_foreach(bus, dimm)
- if (ndctl_dimm_get_handle(dimm) == handle)
- return dimm;
- return NULL;
- }
- #define DIMM_HANDLE(n, s, i, c, d) \
- (((n & 0xfff) << 16) | ((s & 0xf) << 12) | ((i & 0xf) << 8) \
- | ((c & 0xf) << 4) | (d & 0xf))
- dimm = get_dimm_by_handle(bus, DIMM_HANDLE(0, 0, 0, 0, 0));
- LIBNVDIMM/LIBNDCTL: Region
- ----------------------
- A generic REGION device is registered for each PMEM range or BLK-aperture
- set. Per the example there are 6 regions: 2 PMEM and 4 BLK-aperture
- sets on the "nfit_test.0" bus. The primary role of regions are to be a
- container of "mappings". A mapping is a tuple of <DIMM,
- DPA-start-offset, length>.
- LIBNVDIMM provides a built-in driver for these REGION devices. This driver
- is responsible for reconciling the aliased DPA mappings across all
- regions, parsing the LABEL, if present, and then emitting NAMESPACE
- devices with the resolved/exclusive DPA-boundaries for the nd_pmem or
- nd_blk device driver to consume.
- In addition to the generic attributes of "mapping"s, "interleave_ways"
- and "size" the REGION device also exports some convenience attributes.
- "nstype" indicates the integer type of namespace-device this region
- emits, "devtype" duplicates the DEVTYPE variable stored by udev at the
- 'add' event, "modalias" duplicates the MODALIAS variable stored by udev
- at the 'add' event, and finally, the optional "spa_index" is provided in
- the case where the region is defined by a SPA.
- LIBNVDIMM: region
- struct nd_region *nvdimm_pmem_region_create(struct nvdimm_bus *nvdimm_bus,
- struct nd_region_desc *ndr_desc);
- struct nd_region *nvdimm_blk_region_create(struct nvdimm_bus *nvdimm_bus,
- struct nd_region_desc *ndr_desc);
- /sys/devices/platform/nfit_test.0/ndbus0
- |-- region0
- | |-- available_size
- | |-- btt0
- | |-- btt_seed
- | |-- devtype
- | |-- driver -> ../../../../../bus/nd/drivers/nd_region
- | |-- init_namespaces
- | |-- mapping0
- | |-- mapping1
- | |-- mappings
- | |-- modalias
- | |-- namespace0.0
- | |-- namespace_seed
- | |-- numa_node
- | |-- nfit
- | | `-- spa_index
- | |-- nstype
- | |-- set_cookie
- | |-- size
- | |-- subsystem -> ../../../../../bus/nd
- | `-- uevent
- |-- region1
- [..]
- LIBNDCTL: region enumeration example
- Sample region retrieval routines based on NFIT-unique data like
- "spa_index" (interleave set id) for PMEM and "nfit_handle" (dimm id) for
- BLK.
- static struct ndctl_region *get_pmem_region_by_spa_index(struct ndctl_bus *bus,
- unsigned int spa_index)
- {
- struct ndctl_region *region;
- ndctl_region_foreach(bus, region) {
- if (ndctl_region_get_type(region) != ND_DEVICE_REGION_PMEM)
- continue;
- if (ndctl_region_get_spa_index(region) == spa_index)
- return region;
- }
- return NULL;
- }
- static struct ndctl_region *get_blk_region_by_dimm_handle(struct ndctl_bus *bus,
- unsigned int handle)
- {
- struct ndctl_region *region;
- ndctl_region_foreach(bus, region) {
- struct ndctl_mapping *map;
- if (ndctl_region_get_type(region) != ND_DEVICE_REGION_BLOCK)
- continue;
- ndctl_mapping_foreach(region, map) {
- struct ndctl_dimm *dimm = ndctl_mapping_get_dimm(map);
- if (ndctl_dimm_get_handle(dimm) == handle)
- return region;
- }
- }
- return NULL;
- }
- Why Not Encode the Region Type into the Region Name?
- ----------------------------------------------------
- At first glance it seems since NFIT defines just PMEM and BLK interface
- types that we should simply name REGION devices with something derived
- from those type names. However, the ND subsystem explicitly keeps the
- REGION name generic and expects userspace to always consider the
- region-attributes for four reasons:
- 1. There are already more than two REGION and "namespace" types. For
- PMEM there are two subtypes. As mentioned previously we have PMEM where
- the constituent DIMM devices are known and anonymous PMEM. For BLK
- regions the NFIT specification already anticipates vendor specific
- implementations. The exact distinction of what a region contains is in
- the region-attributes not the region-name or the region-devtype.
- 2. A region with zero child-namespaces is a possible configuration. For
- example, the NFIT allows for a DCR to be published without a
- corresponding BLK-aperture. This equates to a DIMM that can only accept
- control/configuration messages, but no i/o through a descendant block
- device. Again, this "type" is advertised in the attributes ('mappings'
- == 0) and the name does not tell you much.
- 3. What if a third major interface type arises in the future? Outside
- of vendor specific implementations, it's not difficult to envision a
- third class of interface type beyond BLK and PMEM. With a generic name
- for the REGION level of the device-hierarchy old userspace
- implementations can still make sense of new kernel advertised
- region-types. Userspace can always rely on the generic region
- attributes like "mappings", "size", etc and the expected child devices
- named "namespace". This generic format of the device-model hierarchy
- allows the LIBNVDIMM and LIBNDCTL implementations to be more uniform and
- future-proof.
- 4. There are more robust mechanisms for determining the major type of a
- region than a device name. See the next section, How Do I Determine the
- Major Type of a Region?
- How Do I Determine the Major Type of a Region?
- ----------------------------------------------
- Outside of the blanket recommendation of "use libndctl", or simply
- looking at the kernel header (/usr/include/linux/ndctl.h) to decode the
- "nstype" integer attribute, here are some other options.
- 1. module alias lookup:
- The whole point of region/namespace device type differentiation is to
- decide which block-device driver will attach to a given LIBNVDIMM namespace.
- One can simply use the modalias to lookup the resulting module. It's
- important to note that this method is robust in the presence of a
- vendor-specific driver down the road. If a vendor-specific
- implementation wants to supplant the standard nd_blk driver it can with
- minimal impact to the rest of LIBNVDIMM.
- In fact, a vendor may also want to have a vendor-specific region-driver
- (outside of nd_region). For example, if a vendor defined its own LABEL
- format it would need its own region driver to parse that LABEL and emit
- the resulting namespaces. The output from module resolution is more
- accurate than a region-name or region-devtype.
- 2. udev:
- The kernel "devtype" is registered in the udev database
- # udevadm info --path=/devices/platform/nfit_test.0/ndbus0/region0
- P: /devices/platform/nfit_test.0/ndbus0/region0
- E: DEVPATH=/devices/platform/nfit_test.0/ndbus0/region0
- E: DEVTYPE=nd_pmem
- E: MODALIAS=nd:t2
- E: SUBSYSTEM=nd
- # udevadm info --path=/devices/platform/nfit_test.0/ndbus0/region4
- P: /devices/platform/nfit_test.0/ndbus0/region4
- E: DEVPATH=/devices/platform/nfit_test.0/ndbus0/region4
- E: DEVTYPE=nd_blk
- E: MODALIAS=nd:t3
- E: SUBSYSTEM=nd
- ...and is available as a region attribute, but keep in mind that the
- "devtype" does not indicate sub-type variations and scripts should
- really be understanding the other attributes.
- 3. type specific attributes:
- As it currently stands a BLK-aperture region will never have a
- "nfit/spa_index" attribute, but neither will a non-NFIT PMEM region. A
- BLK region with a "mappings" value of 0 is, as mentioned above, a DIMM
- that does not allow I/O. A PMEM region with a "mappings" value of zero
- is a simple system-physical-address range.
- LIBNVDIMM/LIBNDCTL: Namespace
- -------------------------
- A REGION, after resolving DPA aliasing and LABEL specified boundaries,
- surfaces one or more "namespace" devices. The arrival of a "namespace"
- device currently triggers either the nd_blk or nd_pmem driver to load
- and register a disk/block device.
- LIBNVDIMM: namespace
- Here is a sample layout from the three major types of NAMESPACE where
- namespace0.0 represents DIMM-info-backed PMEM (note that it has a 'uuid'
- attribute), namespace2.0 represents a BLK namespace (note it has a
- 'sector_size' attribute) that, and namespace6.0 represents an anonymous
- PMEM namespace (note that has no 'uuid' attribute due to not support a
- LABEL).
- /sys/devices/platform/nfit_test.0/ndbus0/region0/namespace0.0
- |-- alt_name
- |-- devtype
- |-- dpa_extents
- |-- force_raw
- |-- modalias
- |-- numa_node
- |-- resource
- |-- size
- |-- subsystem -> ../../../../../../bus/nd
- |-- type
- |-- uevent
- `-- uuid
- /sys/devices/platform/nfit_test.0/ndbus0/region2/namespace2.0
- |-- alt_name
- |-- devtype
- |-- dpa_extents
- |-- force_raw
- |-- modalias
- |-- numa_node
- |-- sector_size
- |-- size
- |-- subsystem -> ../../../../../../bus/nd
- |-- type
- |-- uevent
- `-- uuid
- /sys/devices/platform/nfit_test.1/ndbus1/region6/namespace6.0
- |-- block
- | `-- pmem0
- |-- devtype
- |-- driver -> ../../../../../../bus/nd/drivers/pmem
- |-- force_raw
- |-- modalias
- |-- numa_node
- |-- resource
- |-- size
- |-- subsystem -> ../../../../../../bus/nd
- |-- type
- `-- uevent
- LIBNDCTL: namespace enumeration example
- Namespaces are indexed relative to their parent region, example below.
- These indexes are mostly static from boot to boot, but subsystem makes
- no guarantees in this regard. For a static namespace identifier use its
- 'uuid' attribute.
- static struct ndctl_namespace *get_namespace_by_id(struct ndctl_region *region,
- unsigned int id)
- {
- struct ndctl_namespace *ndns;
- ndctl_namespace_foreach(region, ndns)
- if (ndctl_namespace_get_id(ndns) == id)
- return ndns;
- return NULL;
- }
- LIBNDCTL: namespace creation example
- Idle namespaces are automatically created by the kernel if a given
- region has enough available capacity to create a new namespace.
- Namespace instantiation involves finding an idle namespace and
- configuring it. For the most part the setting of namespace attributes
- can occur in any order, the only constraint is that 'uuid' must be set
- before 'size'. This enables the kernel to track DPA allocations
- internally with a static identifier.
- static int configure_namespace(struct ndctl_region *region,
- struct ndctl_namespace *ndns,
- struct namespace_parameters *parameters)
- {
- char devname[50];
- snprintf(devname, sizeof(devname), "namespace%d.%d",
- ndctl_region_get_id(region), paramaters->id);
- ndctl_namespace_set_alt_name(ndns, devname);
- /* 'uuid' must be set prior to setting size! */
- ndctl_namespace_set_uuid(ndns, paramaters->uuid);
- ndctl_namespace_set_size(ndns, paramaters->size);
- /* unlike pmem namespaces, blk namespaces have a sector size */
- if (parameters->lbasize)
- ndctl_namespace_set_sector_size(ndns, parameters->lbasize);
- ndctl_namespace_enable(ndns);
- }
- Why the Term "namespace"?
- 1. Why not "volume" for instance? "volume" ran the risk of confusing
- ND (libnvdimm subsystem) to a volume manager like device-mapper.
- 2. The term originated to describe the sub-devices that can be created
- within a NVME controller (see the nvme specification:
- http://www.nvmexpress.org/specifications/), and NFIT namespaces are
- meant to parallel the capabilities and configurability of
- NVME-namespaces.
- LIBNVDIMM/LIBNDCTL: Block Translation Table "btt"
- ---------------------------------------------
- A BTT (design document: http://pmem.io/2014/09/23/btt.html) is a stacked
- block device driver that fronts either the whole block device or a
- partition of a block device emitted by either a PMEM or BLK NAMESPACE.
- LIBNVDIMM: btt layout
- Every region will start out with at least one BTT device which is the
- seed device. To activate it set the "namespace", "uuid", and
- "sector_size" attributes and then bind the device to the nd_pmem or
- nd_blk driver depending on the region type.
- /sys/devices/platform/nfit_test.1/ndbus0/region0/btt0/
- |-- namespace
- |-- delete
- |-- devtype
- |-- modalias
- |-- numa_node
- |-- sector_size
- |-- subsystem -> ../../../../../bus/nd
- |-- uevent
- `-- uuid
- LIBNDCTL: btt creation example
- Similar to namespaces an idle BTT device is automatically created per
- region. Each time this "seed" btt device is configured and enabled a new
- seed is created. Creating a BTT configuration involves two steps of
- finding and idle BTT and assigning it to consume a PMEM or BLK namespace.
- static struct ndctl_btt *get_idle_btt(struct ndctl_region *region)
- {
- struct ndctl_btt *btt;
- ndctl_btt_foreach(region, btt)
- if (!ndctl_btt_is_enabled(btt)
- && !ndctl_btt_is_configured(btt))
- return btt;
- return NULL;
- }
- static int configure_btt(struct ndctl_region *region,
- struct btt_parameters *parameters)
- {
- btt = get_idle_btt(region);
- ndctl_btt_set_uuid(btt, parameters->uuid);
- ndctl_btt_set_sector_size(btt, parameters->sector_size);
- ndctl_btt_set_namespace(btt, parameters->ndns);
- /* turn off raw mode device */
- ndctl_namespace_disable(parameters->ndns);
- /* turn on btt access */
- ndctl_btt_enable(btt);
- }
- Once instantiated a new inactive btt seed device will appear underneath
- the region.
- Once a "namespace" is removed from a BTT that instance of the BTT device
- will be deleted or otherwise reset to default values. This deletion is
- only at the device model level. In order to destroy a BTT the "info
- block" needs to be destroyed. Note, that to destroy a BTT the media
- needs to be written in raw mode. By default, the kernel will autodetect
- the presence of a BTT and disable raw mode. This autodetect behavior
- can be suppressed by enabling raw mode for the namespace via the
- ndctl_namespace_set_raw_mode() API.
- Summary LIBNDCTL Diagram
- ------------------------
- For the given example above, here is the view of the objects as seen by the
- LIBNDCTL API:
- +---+
- |CTX| +---------+ +--------------+ +---------------+
- +-+-+ +-> REGION0 +---> NAMESPACE0.0 +--> PMEM8 "pm0.0" |
- | | +---------+ +--------------+ +---------------+
- +-------+ | | +---------+ +--------------+ +---------------+
- | DIMM0 <-+ | +-> REGION1 +---> NAMESPACE1.0 +--> PMEM6 "pm1.0" |
- +-------+ | | | +---------+ +--------------+ +---------------+
- | DIMM1 <-+ +-v--+ | +---------+ +--------------+ +---------------+
- +-------+ +-+BUS0+---> REGION2 +-+-> NAMESPACE2.0 +--> ND6 "blk2.0" |
- | DIMM2 <-+ +----+ | +---------+ | +--------------+ +----------------------+
- +-------+ | | +-> NAMESPACE2.1 +--> ND5 "blk2.1" | BTT2 |
- | DIMM3 <-+ | +--------------+ +----------------------+
- +-------+ | +---------+ +--------------+ +---------------+
- +-> REGION3 +-+-> NAMESPACE3.0 +--> ND4 "blk3.0" |
- | +---------+ | +--------------+ +----------------------+
- | +-> NAMESPACE3.1 +--> ND3 "blk3.1" | BTT1 |
- | +--------------+ +----------------------+
- | +---------+ +--------------+ +---------------+
- +-> REGION4 +---> NAMESPACE4.0 +--> ND2 "blk4.0" |
- | +---------+ +--------------+ +---------------+
- | +---------+ +--------------+ +----------------------+
- +-> REGION5 +---> NAMESPACE5.0 +--> ND1 "blk5.0" | BTT0 |
- +---------+ +--------------+ +---------------+------+
|