<?xml version='1.0' encoding='utf-8'?>
<!DOCTYPE rfc [
  <!ENTITY nbsp    "&#160;">
  <!ENTITY zwsp   "&#8203;">
  <!ENTITY nbhy   "&#8209;">
  <!ENTITY wj     "&#8288;">
]>
<?xml-stylesheet type="text/xsl" href="rfc2629.xslt" ?>
<!-- generated by https://github.com/cabo/kramdown-rfc version 1.7.29 (Ruby 3.4.2) -->
<?rfc docmapping="yes"?>
<?rfc comments="yes"?>
<rfc xmlns:xi="http://www.w3.org/2001/XInclude" ipr="trust200902" docName="draft-haynes-nfsv4-flexfiles-v2-01" category="std" consensus="true" tocInclude="true" sortRefs="true" symRefs="true" version="3">
  <!-- xml2rfc v2v3 conversion 3.28.1 -->
  <front>
    <title abbrev="Flex File Layout v2">Parallel NFS (pNFS) Flexible File Layout Version 2</title>
    <seriesInfo name="Internet-Draft" value="draft-haynes-nfsv4-flexfiles-v2-01"/>
    <author initials="T." surname="Haynes" fullname="Thomas Haynes">
      <organization>Hammerspace</organization>
      <address>
        <email>loghyr@gmail.com</email>
      </address>
    </author>
    <date/>
    <area>General</area>
    <workgroup>Network File System Version 4</workgroup>
    <keyword>Internet-Draft</keyword>
    <abstract>
      <?line 42?>

<t>Parallel NFS (pNFS) allows a separation between the metadata (onto a
metadata server) and data (onto a storage device) for a file.  The
flexible file layout type is defined in this document as an extension
to pNFS that allows the use of storage devices that require only a
limited degree of interaction with the metadata server and use
already-existing protocols.  Data protection is also added to provide
integrity.  Both Client-side mirroring and the Mojette algorithm are
used for data protection.</t>
    </abstract>
    <note>
      <name>Note to Readers</name>
      <?line 53?>

<t>Discussion of this draft takes place
on the NFSv4 working group mailing list (nfsv4@ietf.org),
which is archived at
<eref target="https://mailarchive.ietf.org/arch/search/?email_list=nfsv4"/>. Source
code and issues list for this draft can be found at
<eref target="https://github.com/ietf-wg-nfsv4/uncacheable"/>.</t>
      <t>Working Group information can be found at <eref target="https://github.com/ietf-wg-nfsv4"/>.</t>
      <t>This draft is currently a work in progress.  It needs to be
determined if we want to copy v1 text to v2 or if we want just a diff
of the new content.  For right now, we are copying the v1 text and
adding the new v1 text.  Also, expect sections to move as we push the
emphasis from flex files to protection types.</t>
      <t><em>As a WIP, the XDR extraction may not yet work.</em></t>
    </note>
  </front>
  <middle>
    <?line 72?>

<section anchor="introduction">
      <name>Introduction</name>
      <t>In Parallel NFS (pNFS), the metadata server returns layout type
structures that describe where file data is located.  There are
different layout types for different storage systems and methods of
arranging data on storage devices.  This document defines the
flexible file layout type used with file-based data servers that are
accessed using the NFS protocols: NFSv3 <xref target="RFC1813"/>, NFSv4.0 <xref target="RFC7530"/>,
NFSv4.1 <xref target="RFC8881"/>, and NFSv4.2 <xref target="RFC7862"/>.</t>
      <t>To provide a global state model equivalent to that of the files
layout type, a back-end control protocol might be implemented between
the metadata server and NFSv4.1+ storage devices.  An implementation
can either define its own proprietary mechanism or it could define a
control protocol in a Standards Track document.  The requirements for
a control protocol are specified in <xref target="RFC8881"/> and clarified in
<xref target="RFC8434"/>.</t>
      <t>The control protocol described in this document is based on NFS.  It
does not provide for knowledge of stateids to be passed between the
metadata server and the storage devices.  Instead, the storage
devices are configured such that the metadata server has full access
rights to the data file system and then the metadata server uses
synthetic ids to control client access to individual data files.</t>
      <t>In traditional mirroring of data, the server is responsible for
replicating, validating, and repairing copies of the data file.  With
client-side mirroring, the metadata server provides a layout that
presents the available mirrors to the client.  The client then picks
a mirror to read from and ensures that all writes go to all mirrors.
The client only considers the write transaction to have succeeded if
all mirrors are successfully updated.  In case of error, the client
can use the LAYOUTERROR operation to inform the metadata server,
which is then responsible for the repairing of the mirrored copies of
the file.</t>
      <section anchor="definitions">
        <name>Definitions</name>
        <dl>
          <dt>control communication requirements:</dt>
          <dd>
            <t>the specification for information on layouts, stateids, file metadata,
and file data that must be communicated between the metadata server and
the storage devices.  There is a separate set of requirements for each
layout type.</t>
          </dd>
          <dt>control protocol:</dt>
          <dd>
            <t>the particular mechanism that an implementation of a layout type would
use to meet the control communication requirement for that layout type.
This need not be a protocol as normally understood.  In some cases,
the same protocol may be used as a control protocol and storage protocol.</t>
          </dd>
          <dt>client-side mirroring:</dt>
          <dd>
            <t>a feature in which the client, not the server, is responsible for
updating all of the mirrored copies of a layout segment.</t>
          </dd>
          <dt>erasure encoding:</dt>
          <dd>
            <t>tbd</t>
          </dd>
          <dt>(file) data:</dt>
          <dd>
            <t>that part of the file system object that contains the data to be read
or written.  It is the contents of the object rather than the attributes
of the object.</t>
          </dd>
          <dt>data server (DS):</dt>
          <dd>
            <t>a pNFS server that provides the file's data when the file system
object is accessed over a file-based protocol.</t>
          </dd>
          <dt>fencing:</dt>
          <dd>
            <t>the process by which the metadata server prevents the storage devices
from processing I/O from a specific client to a specific file.</t>
          </dd>
          <dt>file layout type:</dt>
          <dd>
            <t>a layout type in which the storage devices are accessed via the NFS
protocol (see <xref target="sec-layouthint"/> of <xref target="RFC8881"/>).</t>
          </dd>
          <dt>gid:</dt>
          <dd>
            <t>the group id, a numeric value that identifies to which group a file
belongs.</t>
          </dd>
          <dt>layout:</dt>
          <dd>
            <t>the information a client uses to access file data on a storage device.
This information includes specification of the protocol (layout type)
and the identity of the storage devices to be used.</t>
          </dd>
          <dt>layout iomode:</dt>
          <dd>
            <t>a grant of either read-only or read/write I/O to the client.</t>
          </dd>
          <dt>layout segment:</dt>
          <dd>
            <t>a sub-division of a layout.  That sub-division might be by the layout
iomode (see Sections 3.3.20 and 12.2.9 of <xref target="RFC8881"/>), a striping pattern
(see Section 13.3 of <xref target="RFC8881"/>), or requested byte range.</t>
          </dd>
          <dt>layout stateid:</dt>
          <dd>
            <t>a 128-bit quantity returned by a server that uniquely defines the
layout state provided by the server for a specific layout that describes
a layout type and file (see Section 12.5.2 of <xref target="RFC8881"/>).  Further,
Section 12.5.3 of <xref target="RFC8881"/> describes differences in handling between
layout stateids and other stateid types.</t>
          </dd>
          <dt>layout type:</dt>
          <dd>
            <t>a specification of both the storage protocol used to access the data
and the aggregation scheme used to lay out the file data on the underlying
storage devices.</t>
          </dd>
          <dt>loose coupling:</dt>
          <dd>
            <t>when the control protocol is a storage protocol.</t>
          </dd>
          <dt>(file) metadata:</dt>
          <dd>
            <t>the part of the file system object that contains various descriptive
data relevant to the file object, as opposed to the file data itself.
This could include the time of last modification, access time, EOF
position, etc.</t>
          </dd>
          <dt>metadata server (MDS):</dt>
          <dd>
            <t>the pNFS server that provides metadata information for a file system
object.  It is also responsible for generating, recalling, and revoking
layouts for file system objects, for performing directory operations,
and for performing I/O operations to regular files when the clients
direct these to the metadata server itself.</t>
          </dd>
          <dt>mirror:</dt>
          <dd>
            <t>a copy of a layout segment.  Note that if one copy of the mirror is
updated, then all copies must be updated.</t>
          </dd>
          <dt>non-systematic encoding:</dt>
          <dd>
            <t>TBD</t>
          </dd>
          <dt>recalling a layout:</dt>
          <dd>
            <t>a graceful recall, via a callback, of a specific layout by the metadata
server to the client.  Graceful here means that the client would have
the opportunity to flush any WRITEs, etc., before returning the layout
to the metadata server.</t>
          </dd>
          <dt>revoking a layout:</dt>
          <dd>
            <t>an invalidation of a specific layout by the metadata server.
Once revocation occurs, the metadata server will not accept as valid any
reference to the revoked layout, and a storage device will not accept
any client access based on the layout.</t>
          </dd>
          <dt>resilvering:</dt>
          <dd>
            <t>the act of rebuilding a mirrored copy of a layout segment from a
known good copy of the layout segment.  Note that this can also be done
to create a new mirrored copy of the layout segment.</t>
          </dd>
          <dt>rsize:</dt>
          <dd>
            <t>the data transfer buffer size used for READs.</t>
          </dd>
          <dt>stateid:</dt>
          <dd>
            <t>a 128-bit quantity returned by a server that uniquely defines the set
of locking-related state provided by the server.  Stateids may designate
state related to open files, byte-range locks, delegations, or layouts.</t>
          </dd>
          <dt>storage device:</dt>
          <dd>
            <t>the target to which clients may direct I/O requests when they hold
an appropriate layout.  See Section 2.1 of <xref target="RFC8434"/> for further
discussion of the difference between a data server and a storage device.</t>
          </dd>
          <dt>storage protocol:</dt>
          <dd>
            <t>the protocol used by clients to do I/O operations to the storage
device.  Each layout type specifies the set of storage protocols.</t>
          </dd>
          <dt>systematic encoding:</dt>
          <dd>
            <t>TBD</t>
          </dd>
          <dt>tight coupling:</dt>
          <dd>
            <t>an arrangement in which the control protocol is one designed
specifically for control communication.  It may be either a proprietary
protocol adapted specifically to a particular metadata server or a
protocol based on a Standards Track document.</t>
          </dd>
          <dt>uid:</dt>
          <dd>
            <t>the user id, a numeric value that identifies which user owns a file.</t>
          </dd>
          <dt>write hole:</dt>
          <dd>
            <t>TBD</t>
          </dd>
          <dt>wsize:</dt>
          <dd>
            <t>the data transfer buffer size used for WRITEs.</t>
          </dd>
        </dl>
      </section>
      <section anchor="requirements-language">
        <name>Requirements Language</name>
        <t>The key words "<bcp14>MUST</bcp14>", "<bcp14>MUST NOT</bcp14>", "<bcp14>REQUIRED</bcp14>", "<bcp14>SHALL</bcp14>", "<bcp14>SHALL
NOT</bcp14>", "<bcp14>SHOULD</bcp14>", "<bcp14>SHOULD NOT</bcp14>", "<bcp14>RECOMMENDED</bcp14>", "<bcp14>NOT RECOMMENDED</bcp14>",
"<bcp14>MAY</bcp14>", and "<bcp14>OPTIONAL</bcp14>" in this document are to be interpreted as
described in BCP 14 <xref target="RFC2119"/> <xref target="RFC8174"/> when, and only when, they
appear in all capitals, as shown here.</t>
        <?line -18?>

</section>
    </section>
    <section anchor="coupling-of-storage-devices">
      <name>Coupling of Storage Devices</name>
      <t>A server implementation may choose either a loosely coupled model or
a tightly coupled model between the metadata server and the storage
devices.  <xref target="RFC8434"/> describes the general problems facing pNFS
implementations.  This document details how the new flexible file
layout type addresses these issues.  To implement the tightly coupled
model, a control protocol has to be defined.  As the flexible file
layout imposes no special requirements on the client, the control
protocol will need to provide:</t>
      <ol spacing="normal" type="1"><li>
          <t>management of both security and LAYOUTCOMMITs and</t>
        </li>
        <li>
          <t>a global stateid model and management of these stateids.</t>
        </li>
      </ol>
      <t>When implementing the loosely coupled model, the only control
protocol will be a version of NFS, with no ability to provide a
global stateid model or to prevent clients from using layouts
inappropriately.  To enable client use in that environment, this
document will specify how security, state, and locking are to be
managed.</t>
      <section anchor="layoutcommit">
        <name>LAYOUTCOMMIT</name>
        <t>Regardless of the coupling model, the metadata server has the
responsibility, upon receiving a LAYOUTCOMMIT (see Section 18.42 of
<xref target="RFC8881"/>) to ensure that the semantics of pNFS are respected (see
Section 3.1 of <xref target="RFC8434"/>).  These do include a requirement that data
written to a data storage device be stable before the occurrence of
the LAYOUTCOMMIT.</t>
        <t>It is the responsibility of the client to make sure the data file is
stable before the metadata server begins to query the storage devices
about the changes to the file.  If any WRITE to a storage device did
not result with stable_how equal to FILE_SYNC, a LAYOUTCOMMIT to the
metadata server <bcp14>MUST</bcp14> be preceded by a COMMIT to the storage devices
written to.  Note that if the client has not done a COMMIT to the
storage device, then the LAYOUTCOMMIT might not be synchronized to
the last WRITE operation to the storage device.</t>
      </section>
      <section anchor="fencing-clients-from-the-storage-device">
        <name>Fencing Clients from the Storage Device</name>
        <t>With loosely coupled storage devices, the metadata server uses
synthetic uids (user ids) and gids (group ids) for the data file,
where the uid owner of the data file is allowed read/write access and
the gid owner is allowed read-only access.  As part of the layout
(see ffds_user and ffds_group in <xref target="sec_ff_layout"/>), the client is
provided with the user and group to be used in the Remote Procedure Call
(RPC) <xref target="RFC5531"/> credentials needed to access the data file.  Fencing off
of clients is achieved by the metadata server changing the synthetic uid
and/or gid owners of the data file on the storage device to implicitly
revoke the outstanding RPC credentials.  A client presenting the wrong
credential for the desired access will get an NFS4ERR_ACCESS error.</t>
        <t>With this loosely coupled model, the metadata server is not able to fence
off a single client; it is forced to fence off all clients.  However,
as the other clients react to the fencing, returning their layouts and
trying to get new ones, the metadata server can hand out a new uid and
gid to allow access.</t>
        <t>It is <bcp14>RECOMMENDED</bcp14> to implement common access control methods at the
storage device file system to allow only the metadata server root
(super user) access to the storage device and to set the owner of all
directories holding data files to the root user.  This approach provides
a practical model to enforce access control and fence off cooperative
clients, but it cannot protect against malicious clients; hence, it
provides a level of security equivalent to AUTH_SYS.  It is <bcp14>RECOMMENDED</bcp14>
that the communication between the metadata server and storage device
be secure from eavesdroppers and man-in-the-middle protocol tampering.
The security measure could be physical security (e.g., the servers
are co-located in a physically secure area), encrypted communications,
or some other technique.</t>
        <t>With tightly coupled storage devices, the metadata server sets the
user and group owners, mode bits, and Access Control List (ACL) of
the data file to be the same as the metadata file.  And the client
must authenticate with the storage device and go through the same
authorization process it would go through via the metadata server.
In the case of tight coupling, fencing is the responsibility of the
control protocol and is not described in detail in this document.
However, implementations of the tightly coupled locking model (see
<xref target="sec-state-locking"/>) will need a way to prevent access by certain clients to
specific files by invalidating the corresponding stateids on the
storage device.  In such a scenario, the client will be given an
error of NFS4ERR_BAD_STATEID.</t>
        <t>The client need not know the model used between the metadata server
and the storage device.  It need only react consistently to any
errors in interacting with the storage device.  It should both return
the layout and error to the metadata server and ask for a new layout.
At that point, the metadata server can either hand out a new layout,
hand out no layout (forcing the I/O through it), or deny the client
further access to the file.</t>
        <section anchor="implementation-notes-for-synthetic-uidsgids">
          <name>Implementation Notes for Synthetic uids/gids</name>
          <t>The selection method for the synthetic uids and gids to be used for
fencing in loosely coupled storage devices is strictly an
implementation issue.  That is, an administrator might restrict a
range of such ids available to the Lightweight Directory Access
Protocol (LDAP) 'uid' field <xref target="RFC4519"/>.  The administrator might also
be able to choose an id that would never be used to grant access.
Then, when the metadata server had a request to access a file, a
SETATTR would be sent to the storage device to set the owner and
group of the data file.  The user and group might be selected in a
round-robin fashion from the range of available ids.</t>
          <t>Those ids would be sent back as ffds_user and ffds_group to the
client, who would present them as the RPC credentials to the storage
device.  When the client is done accessing the file and the metadata
server knows that no other client is accessing the file, it can reset
the owner and group to restrict access to the data file.</t>
          <t>When the metadata server wants to fence off a client, it changes the
synthetic uid and/or gid to the restricted ids.  Note that using a
restricted id ensures that there is a change of owner and at least
one id available that never gets allowed access.</t>
          <t>Under an AUTH_SYS security model, synthetic uids and gids of 0 <bcp14>SHOULD</bcp14>
be avoided.  These typically either grant super access to files on a
storage device or are mapped to an anonymous id.  In the first case,
even if the data file is fenced, the client might still be able to
access the file.  In the second case, multiple ids might be mapped to
the anonymous ids.</t>
        </section>
        <section anchor="example-of-using-synthetic-uidsgids">
          <name>Example of using Synthetic uids/gids</name>
          <t>The user loghyr creates a file "ompha.c" on the metadata server,
which then creates a corresponding data file on the storage device.</t>
          <t>The metadata server entry may look like:</t>
          <figure anchor="fig-meta-ompha">
            <name>Metadata's view of ompha.c</name>
            <sourcecode type="shell"><![CDATA[
-rw-r--r--    1 loghyr  staff    1697 Dec  4 11:31 ompha.c
]]></sourcecode>
          </figure>
          <t>On the storage device, the file may be assigned some unpredictable
synthetic uid/gid to deny access:</t>
          <figure anchor="fig-data-ompha">
            <name>Data's view of ompha.c</name>
            <sourcecode type="shell"><![CDATA[
-rw-r-----    1 19452   28418    1697 Dec  4 11:31 data_ompha.c
]]></sourcecode>
          </figure>
          <t>When the file is opened on a client and accessed, the user will try
to get a layout for the data file.  Since the layout knows nothing
about the user (and does not care), it does not matter whether the
user loghyr or garbo opens the file.  The client has to present an
uid of 19452 to get write permission.  If it presents any other value
for the uid, then it must give a gid of 28418 to get read access.</t>
          <t>Further, if the metadata server decides to fence the file, it should
change the uid and/or gid such that these values neither match
earlier values for that file nor match a predictable change based on
an earlier fencing.</t>
          <figure anchor="fig-fenced-ompha">
            <name>Fenced Data's view of ompha.c</name>
            <sourcecode type="shell"><![CDATA[
-rw-r-----    1 19453   28419    1697 Dec  4 11:31 data_ompha.c
]]></sourcecode>
          </figure>
          <t>The set of synthetic gids on the storage device should be selected
such that there is no mapping in any of the name services used by the
storage device, i.e., each group should have no members.</t>
          <t>If the layout segment has an iomode of LAYOUTIOMODE4_READ, then the
metadata server should return a synthetic uid that is not set on the
storage device.  Only the synthetic gid would be valid.</t>
          <t>The client is thus solely responsible for enforcing file permissions
in a loosely coupled model.  To allow loghyr write access, it will
send an RPC to the storage device with a credential of 1066:1067.  To
allow garbo read access, it will send an RPC to the storage device
with a credential of 1067:1067.  The value of the uid does not matter
as long as it is not the synthetic uid granted when getting the
layout.</t>
          <t>While pushing the enforcement of permission checking onto the client
may seem to weaken security, the client may already be responsible
for enforcing permissions before modifications are sent to a server.
With cached writes, the client is always responsible for tracking who
is modifying a file and making sure to not coalesce requests from
multiple users into one request.</t>
        </section>
      </section>
      <section anchor="sec-state-locking">
        <name>State and Locking Models</name>
        <t>An implementation can always be deployed as a loosely coupled model.
There is, however, no way for a storage device to indicate over an
NFS protocol that it can definitively participate in a tightly
coupled model:</t>
        <ul spacing="normal">
          <li>
            <t>Storage devices implementing the NFSv3 and NFSv4.0 protocols are
always treated as loosely coupled.</t>
          </li>
          <li>
            <t>NFSv4.1+ storage devices that do not return the
EXCHGID4_FLAG_USE_PNFS_DS flag set to EXCHANGE_ID are indicating
that they are to be treated as loosely coupled.  From the locking
viewpoint, they are treated in the same way as NFSv4.0 storage
devices.</t>
          </li>
          <li>
            <t>NFSv4.1+ storage devices that do identify themselves with the
EXCHGID4_FLAG_USE_PNFS_DS flag set to EXCHANGE_ID can potentially
be tightly coupled.  They would use a back-end control protocol to
implement the global stateid model as described in <xref target="RFC8881"/>.</t>
          </li>
        </ul>
        <t>A storage device would have to be either discovered or advertised
over the control protocol to enable a tightly coupled model.</t>
        <section anchor="loosely-coupled-locking-model">
          <name>Loosely Coupled Locking Model</name>
          <t>When locking-related operations are requested, they are primarily
dealt with by the metadata server, which generates the appropriate
stateids.  When an NFSv4 version is used as the data access protocol,
the metadata server may make stateid-related requests of the storage
devices.  However, it is not required to do so, and the resulting
stateids are known only to the metadata server and the storage
device.</t>
          <t>Given this basic structure, locking-related operations are handled as
follows:</t>
          <ul spacing="normal">
            <li>
              <t>OPENs are dealt with by the metadata server.  Stateids are
selected by the metadata server and associated with the client ID
describing the client's connection to the metadata server.  The
metadata server may need to interact with the storage device to
locate the file to be opened, but no locking-related functionality
need be used on the storage device.</t>
            </li>
            <li>
              <t>OPEN_DOWNGRADE and CLOSE only require local execution on the
metadata server.</t>
            </li>
            <li>
              <t>Advisory byte-range locks can be implemented locally on the
metadata server.  As in the case of OPENs, the stateids associated
with byte-range locks are assigned by the metadata server and only
used on the metadata server.</t>
            </li>
            <li>
              <t>Delegations are assigned by the metadata server that initiates
recalls when conflicting OPENs are processed.  No storage device
involvement is required.</t>
            </li>
            <li>
              <t>TEST_STATEID and FREE_STATEID are processed locally on the
metadata server, without storage device involvement.</t>
            </li>
          </ul>
          <t>All I/O operations to the storage device are done using the anonymous
stateid.  Thus, the storage device has no information about the
openowner and lockowner responsible for issuing a particular I/O
operation.  As a result:</t>
          <ul spacing="normal">
            <li>
              <t>Mandatory byte-range locking cannot be supported because the
storage device has no way of distinguishing I/O done on behalf of
the lock owner from those done by others.</t>
            </li>
            <li>
              <t>Enforcement of share reservations is the responsibility of the
client.  Even though I/O is done using the anonymous stateid, the
client must ensure that it has a valid stateid associated with the
openowner.</t>
            </li>
          </ul>
          <t>In the event that a stateid is revoked, the metadata server is
responsible for preventing client access, since it has no way of
being sure that the client is aware that the stateid in question has
been revoked.</t>
          <t>As the client never receives a stateid generated by a storage device,
there is no client lease on the storage device and no prospect of
lease expiration, even when access is via NFSv4 protocols.  Clients
will have leases on the metadata server.  In dealing with lease
expiration, the metadata server may need to use fencing to prevent
revoked stateids from being relied upon by a client unaware of the
fact that they have been revoked.</t>
        </section>
        <section anchor="tightly-coupled-locking-model">
          <name>Tightly Coupled Locking Model</name>
          <t>When locking-related operations are requested, they are primarily
dealt with by the metadata server, which generates the appropriate
stateids.  These stateids must be made known to the storage device
using control protocol facilities, the details of which are not
discussed in this document.</t>
          <t>Given this basic structure, locking-related operations are handled as
follows:</t>
          <ul spacing="normal">
            <li>
              <t>OPENs are dealt with primarily on the metadata server.  Stateids
are selected by the metadata server and associated with the client
ID describing the client's connection to the metadata server.  The
metadata server needs to interact with the storage device to
locate the file to be opened and to make the storage device aware
of the association between the metadata-server-chosen stateid and
the client and openowner that it represents.  OPEN_DOWNGRADE and
CLOSE are executed initially on the metadata server, but the state
change made must be propagated to the storage device.</t>
            </li>
            <li>
              <t>Advisory byte-range locks can be implemented locally on the
metadata server.  As in the case of OPENs, the stateids associated
with byte-range locks are assigned by the metadata server and are
available for use on the metadata server.  Because I/O operations
are allowed to present lock stateids, the metadata server needs
the ability to make the storage device aware of the association
between the metadata-server-chosen stateid and the corresponding
open stateid it is associated with.</t>
            </li>
            <li>
              <t>Mandatory byte-range locks can be supported when both the metadata
server and the storage devices have the appropriate support.  As
in the case of advisory byte-range locks, these are assigned by
the metadata server and are available for use on the metadata
server.  To enable mandatory lock enforcement on the storage
device, the metadata server needs the ability to make the storage
device aware of the association between the metadata-server-chosen
stateid and the client, openowner, and lock (i.e., lockowner,
byte-range, and lock-type) that it represents.  Because I/O
operations are allowed to present lock stateids, this information
needs to be propagated to all storage devices to which I/O might
be directed rather than only to storage device that contain the
locked region.</t>
            </li>
            <li>
              <t>Delegations are assigned by the metadata server that initiates
recalls when conflicting OPENs are processed.  Because I/O
operations are allowed to present delegation stateids, the
metadata server requires the ability:  </t>
              <ol spacing="normal" type="1"><li>
                  <t>to make the storage device aware of the association between
the metadata-server-chosen stateid and the filehandle and
delegation type it represents</t>
                </li>
                <li>
                  <t>to break such an association.</t>
                </li>
              </ol>
            </li>
            <li>
              <t>TEST_STATEID is processed locally on the metadata server, without
storage device involvement.</t>
            </li>
            <li>
              <t>FREE_STATEID is processed on the metadata server, but the metadata
server requires the ability to propagate the request to the
corresponding storage devices.</t>
            </li>
          </ul>
          <t>Because the client will possess and use stateids valid on the storage
device, there will be a client lease on the storage device, and the
possibility of lease expiration does exist.  The best approach for
the storage device is to retain these locks as a courtesy.  However,
if it does not do so, control protocol facilities need to provide the
means to synchronize lock state between the metadata server and
storage device.</t>
          <t>Clients will also have leases on the metadata server that are subject
to expiration.  In dealing with lease expiration, the metadata server
would be expected to use control protocol facilities enabling it to
invalidate revoked stateids on the storage device.  In the event the
client is not responsive, the metadata server may need to use fencing
to prevent revoked stateids from being acted upon by the storage
device.</t>
        </section>
      </section>
    </section>
    <section anchor="client-side-protection-modes">
      <name>Client-Side Protection Modes</name>
      <section anchor="client-side-mirroring">
        <name>Client-Side Mirroring</name>
        <t>Do I want this?</t>
      </section>
    </section>
    <section anchor="xdr-description-of-the-flexible-file-layout-type">
      <name>XDR Description of the Flexible File Layout Type</name>
      <t>This document contains the External Data Representation (XDR)
<xref target="RFC4506"/> description of the flexible file layout type.  The XDR
description is embedded in this document in a way that makes it simple
for the reader to extract into a ready-to-compile form.  The reader can
feed this document into the shell script in <xref target="fig-extract"/> to produce
the machine-readable XDR description of the flexible file layout type.</t>
      <figure anchor="fig-extract">
        <name>extract.sh</name>
        <sourcecode type="shell"><![CDATA[
#!/bin/sh
grep '^ *///' $* | sed 's?^ */// ??' | sed 's?^ *///$??'
]]></sourcecode>
      </figure>
      <t>That is, if the above script is stored in a file called "extract.sh"
and this document is in a file called "spec.txt", then the reader can
run the script as in <xref target="fig-extract-example"/>.</t>
      <figure anchor="fig-extract-example">
        <name>Example use of extract.sh</name>
        <sourcecode type="shell"><![CDATA[
sh extract.sh < spec.txt > flex_files2_prot.x
]]></sourcecode>
      </figure>
      <t>The effect of the script is to remove leading blank space from each
line, plus a sentinel sequence of "///".</t>
      <t>The embedded XDR file header follows.  Subsequent XDR descriptions
with the sentinel sequence are embedded throughout the document.</t>
      <t>Note that the XDR code contained in this document depends on types
from the NFSv4.1 nfs4_prot.x file <xref target="RFC5662"/>.  This includes both nfs
types that end with a 4, such as offset4, length4, etc., as well as
more generic types such as uint32_t and uint64_t.</t>
    </section>
    <section anchor="device-addressing-and-discovery">
      <name>Device Addressing and Discovery</name>
      <t>Data operations to a storage device require the client to know the
network address of the storage device.  The NFSv4.1+ GETDEVICEINFO
operation (Section 18.40 of <xref target="RFC8881"/>) is used by the client to
retrieve that information.</t>
      <section anchor="sec-ff_device_addr4">
        <name>ff_device_addr4</name>
        <t>The ff_device_addr4 data structure (see <xref target="fig-ff_device_addr4"/>) is returned by the
server as the layout-type-specific opaque field da_addr_body in the
device_addr4 structure by a successful GETDEVICEINFO operation.</t>
        <figure anchor="fig-ff_device_versions4">
          <name>ff_device_versions4</name>
          <sourcecode type="xdr"><![CDATA[
   /// struct ff_device_versions4 {
   ///         uint32_t        ffdv_version;
   ///         uint32_t        ffdv_minorversion;
   ///         uint32_t        ffdv_rsize;
   ///         uint32_t        ffdv_wsize;
   ///         bool            ffdv_tightly_coupled;
   /// };
   ///
]]></sourcecode>
        </figure>
        <figure anchor="fig-ff_device_addr4">
          <name>ff_device_addr4</name>
          <sourcecode type="xdr"><![CDATA[
   /// struct ff_device_addr4 {
   ///         multipath_list4     ffda_netaddrs;
   ///         ff_device_versions4 ffda_versions<>;
   /// };
   ///
]]></sourcecode>
        </figure>
        <t>The ffda_netaddrs field is used to locate the storage device.  It
<bcp14>MUST</bcp14> be set by the server to a list holding one or more of the device
network addresses.</t>
        <t>The ffda_versions array allows the metadata server to present choices
as to NFS version, minor version, and coupling strength to the
client.  The ffdv_version and ffdv_minorversion represent the NFS
protocol to be used to access the storage device.  This layout
specification defines the semantics for ffdv_versions 3 and 4.  If
ffdv_version equals 3, then the server <bcp14>MUST</bcp14> set ffdv_minorversion to
0 and ffdv_tightly_coupled to false.  The client <bcp14>MUST</bcp14> then access the
storage device using the NFSv3 protocol <xref target="RFC1813"/>.  If ffdv_version
equals 4, then the server <bcp14>MUST</bcp14> set ffdv_minorversion to one of the
NFSv4 minor version numbers, and the client <bcp14>MUST</bcp14> access the storage
device using NFSv4 with the specified minor version.</t>
        <t>Note that while the client might determine that it cannot use any of
the configured combinations of ffdv_version, ffdv_minorversion, and
ffdv_tightly_coupled, when it gets the device list from the metadata
server, there is no way to indicate to the metadata server as to
which device it is version incompatible.  However, if the client
waits until it retrieves the layout from the metadata server, it can
at that time clearly identify the storage device in question (see
<xref target="sec-version-errors"/>).</t>
        <t>The ffdv_rsize and ffdv_wsize are used to communicate the maximum
rsize and wsize supported by the storage device.  As the storage
device can have a different rsize or wsize than the metadata server,
the ffdv_rsize and ffdv_wsize allow the metadata server to
communicate that information on behalf of the storage device.</t>
        <t>ffdv_tightly_coupled informs the client as to whether or not the
metadata server is tightly coupled with the storage devices.  Note
that even if the data protocol is at least NFSv4.1, it may still be
the case that there is loose coupling in effect.  If
ffdv_tightly_coupled is not set, then the client <bcp14>MUST</bcp14> commit writes
to the storage devices for the file before sending a LAYOUTCOMMIT to
the metadata server.  That is, the writes <bcp14>MUST</bcp14> be committed by the
client to stable storage via issuing WRITEs with stable_how ==
FILE_SYNC or by issuing a COMMIT after WRITEs with stable_how !=
FILE_SYNC (see Section 3.3.7 of <xref target="RFC1813"/>).</t>
      </section>
      <section anchor="storage-device-multipathing">
        <name>Storage Device Multipathing</name>
        <t>The flexible file layout type supports multipathing to multiple
storage device addresses.  Storage-device-level multipathing is used
for bandwidth scaling via trunking and for higher availability of use
in the event of a storage device failure.  Multipathing allows the
client to switch to another storage device address that may be that
of another storage device that is exporting the same data stripe
unit, without having to contact the metadata server for a new layout.</t>
        <t>To support storage device multipathing, ffda_netaddrs contains an
array of one or more storage device network addresses.  This array
(data type multipath_list4) represents a list of storage devices
(each identified by a network address), with the possibility that
some storage device will appear in the list multiple times.</t>
        <t>The client is free to use any of the network addresses as a
destination to send storage device requests.  If some network
addresses are less desirable paths to the data than others, then the
metadata server <bcp14>SHOULD NOT</bcp14> include those network addresses in
ffda_netaddrs.  If less desirable network addresses exist to provide
failover, the <bcp14>RECOMMENDED</bcp14> method to offer the addresses is to provide
them in a replacement device-ID-to-device-address mapping or a
replacement device ID.  When a client finds no response from the
storage device using all addresses available in ffda_netaddrs, it
<bcp14>SHOULD</bcp14> send a GETDEVICEINFO to attempt to replace the existing
device-ID-to-device-address mappings.  If the metadata server detects
that all network paths represented by ffda_netaddrs are unavailable,
the metadata server <bcp14>SHOULD</bcp14> send a CB_NOTIFY_DEVICEID (if the client
has indicated it wants device ID notifications for changed device
IDs) to change the device-ID-to-device-address mappings to the
available addresses.  If the device ID itself will be replaced, the
metadata server <bcp14>SHOULD</bcp14> recall all layouts with the device ID and thus
force the client to get new layouts and device ID mappings via
LAYOUTGET and GETDEVICEINFO.</t>
        <t>Generally, if two network addresses appear in ffda_netaddrs, they
will designate the same storage device.  When the storage device is
accessed over NFSv4.1 or a higher minor version, the two storage
device addresses will support the implementation of client ID or
session trunking (the latter is <bcp14>RECOMMENDED</bcp14>) as defined in <xref target="RFC8881"/>.
The two storage device addresses will share the same server owner or
major ID of the server owner.  It is not always necessary for the two
storage device addresses to designate the same storage device with
trunking being used.  For example, the data could be read-only, and
the data consist of exact replicas.</t>
      </section>
    </section>
    <section anchor="flexible-file-layout-type">
      <name>Flexible File Layout Type</name>
      <t>The original layouttype4 introduced in <xref target="RFC5662"/> is modified to as in
<xref target="fig-orig-layout"/>.</t>
      <figure anchor="fig-orig-layout">
        <name>The original layout type</name>
        <sourcecode type="xdr"><![CDATA[
       enum layouttype4 {
           LAYOUT4_NFSV4_1_FILES   = 1,
           LAYOUT4_OSD2_OBJECTS    = 2,
           LAYOUT4_BLOCK_VOLUME    = 3,
           LAYOUT4_FLEX_FILES      = 4
       };

       struct layout_content4 {
           layouttype4             loc_type;
           opaque                  loc_body<>;
       };

       struct layout4 {
           offset4                 lo_offset;
           length4                 lo_length;
           layoutiomode4           lo_iomode;
           layout_content4         lo_content;
       };
]]></sourcecode>
      </figure>
      <t>This document defines structures associated with the layouttype4
value LAYOUT4_FLEX_FILES.  <xref target="RFC8881"/> specifies the loc_body structure
as an XDR type "opaque".  The opaque layout is uninterpreted by the
generic pNFS client layers but is interpreted by the flexible file
layout type implementation.  This section defines the structure of
this otherwise opaque value, ff_layout4.</t>
      <section anchor="sec_ff_layout">
        <name>ff_layout4</name>
        <figure anchor="fig-ff_layout4v2">
          <name>The flex files layout type v2</name>
          <sourcecode type="xdr"><![CDATA[
   /// const FF_FLAGS_NO_LAYOUTCOMMIT   = 0x00000001;
   /// const FF_FLAGS_NO_IO_THRU_MDS    = 0x00000002;
   /// const FF_FLAGS_NO_READ_IO        = 0x00000004;
   /// const FF_FLAGS_WRITE_ONE_MIRROR  = 0x00000008;
   ///
   /// typedef uint32_t            ff_flags4;
   ///
   /// /*
   ///  * NFsv4.0, NFSv4.1, and NFSv4.2 can all
   ///  * have unique stateids for the file.
   ///  */
   /// struct ff2_file_info4 {
   ///     stateid4                fffi_stateid;
   ///     nfs_fh4                 fffi_fh_vers;
   /// };
   ///
   /// /*
   ///  * For now, allow all protection types to
   ///  * be in the same flags space.
   ///  */
   /// const FF2_DS_FLAGS_ACTIVE        = 0x00000001;
   /// const FF2_DS_FLAGS_SPARE         = 0x00000002;
   /// const FF2_DS_FLAGS_REPAIR        = 0x00000004;
   /// typedef uint32_t            ff2_ds_flags4;
   ///
   /// struct ff2_data_server4 {
   ///     deviceid4               ffds_deviceid;
   ///     uint32_t                ffds_efficiency;
   ///     ff2_file_info4          ffds_file_info<>;
   ///     fattr4_owner            ffds_user;
   ///     fattr4_owner_group      ffds_group;
   ///     ff2_ds_flags4           ffds_flags;
   /// };
   ///
   /// struct ff2_mirror4 {
   ///     ff2_data_server4        ffm_data_servers<>;
   ///     ff2_protection_type     ffds_protection;
   /// };
   ///
   /// // X_Y: Need X+Y to write
   /// // Can lose Y files
   /// // So Y spares
   /// enum ff2_mojette_faulty_devices {
   ///         FF2_MOJETTE_FAULTY_DEVICES_2_1       = 0x1;
   ///         FF2_MOJETTE_FAULTY_DEVICES_4_1       = 0x2;
   ///         FF2_MOJETTE_FAULTY_DEVICES_4_2       = 0x3;
   ///         FF2_MOJETTE_FAULTY_DEVICES_8_1       = 0x4;
   ///         FF2_MOJETTE_FAULTY_DEVICES_8_2       = 0x5;
   ///         FF2_MOJETTE_FAULTY_DEVICES_8_3       = 0x6;
   ///         FF2_MOJETTE_FAULTY_DEVICES_8_4       = 0x7;
   /// };
   ///
   /// //
   /// // Need to define projection header for READ/WRITE
   /// //
   ///
   /// union ff2_protection_data switch (ff2_protection_type fpd_type) {
   ///     case FF2_PROTECTION_TYPE_MOJETTE:
   ///         uint32_t                        fpd_mojette_rsize;
   ///         uint32_t                        fpd_mojette_wsize;
   ///         ff2_mojette_faulty_devices      fpd_mojette_potection_configuration;
   ///     case FF2_PROTECTION_TYPE_MIRRORED:
   ///         void;
   /// };
   ///
   /// struct ff2_layout4 {
   ///     length4                 ffl_stripe_unit;
   ///     ff2_mirror4             ffl_mirrors<>;
   ///     ff_flags4               ffl_flags;
   ///     uint32_t                ffl_stats_collect_hint;
   ///     ff2_protection_data     ffl_protection_data;
   /// };
   ///
   ///
   ///
]]></sourcecode>
        </figure>
        <figure anchor="fig-ff_layout4">
          <name>The flex files layout type v1</name>
          <sourcecode type="xdr"><![CDATA[
   struct ff_data_server4 {
       deviceid4               ffds_deviceid;
       uint32_t                ffds_efficiency;
       stateid4                ffds_stateid;
       nfs_fh4                 ffds_fh_vers<>;
       fattr4_owner            ffds_user;
       fattr4_owner_group      ffds_group;
   };


   struct ff_mirror4 {
       ff_data_server4         ffm_data_servers<>;
   };


   struct ff_layout4 {
       length4                 ffl_stripe_unit;
       ff_mirror4              ffl_mirrors<>;
       ff_flags4               ffl_flags;
       uint32_t                ffl_stats_collect_hint;
   };
]]></sourcecode>
        </figure>
        <t>The ff_layout4 structure (see <xref target="fig-ff_layout4"/>) specifies a layout in that
portion of the data file described in the current layout segment.  It
is either a single instance or a set of mirrored copies of that
portion of the data file.  When mirroring is in effect, it protects
against loss of data in layout segments.</t>
        <t>While not explicitly shown in <xref target="fig-ff_layout4"/>, each layout4 element
returned in the logr_layout array of LAYOUTGET4res (see
Section 18.43.2 of <xref target="RFC8881"/>) describes a layout segment.  Hence,
each ff_layout4 also describes a layout segment.  It is possible that
the file is concatenated from more than one layout segment.  Each
layout segment <bcp14>MAY</bcp14> represent different striping parameters.</t>
        <t>The ffl_stripe_unit field is the stripe unit size in use for the
current layout segment.  The number of stripes is given inside each
mirror by the number of elements in ffm_data_servers.  If the number
of stripes is one, then the value for ffl_stripe_unit <bcp14>MUST</bcp14> default to
zero.  The only supported mapping scheme is sparse and is detailed in
<xref target="sec-striping"/>.  Note that there is an assumption here that both the
stripe unit size and the number of stripes are the same across all
mirrors.</t>
        <t>The ffl_mirrors field is the array of mirrored storage devices that
provide the storage for the current stripe; see <xref target="fig-parallel-fileystem"/>.</t>
        <t>The ffl_stats_collect_hint field provides a hint to the client on how
often the server wants it to report LAYOUTSTATS for a file.  The time
is in seconds.</t>
        <figure anchor="fig-parallel-fileystem">
          <name>The Relationship between MDS and DSes</name>
          <artwork><![CDATA[
            +-----------+
            |           |
            |           |
            |   File    |
            |           |
            |           |
            +-----+-----+
                  |
     +------------+------------+
     |                         |
+----+-----+             +-----+----+
| Mirror 1 |             | Mirror 2 |
+----+-----+             +-----+----+
     |                         |
+-----------+            +-----------+
|+-----------+           |+-----------+
||+-----------+          ||+-----------+
+||  Storage  |          +||  Storage  |
 +|  Devices  |           +|  Devices  |
  +-----------+            +-----------+
]]></artwork>
        </figure>
        <t>The ffs_mirrors field represents an array of state information for
each mirrored copy of the current layout segment.  Each element is
described by a ff_mirror4 type.</t>
        <t>ffds_deviceid provides the deviceid of the storage device holding the
data file.</t>
        <t>ffds_fh_vers is an array of filehandles of the data file matching the
available NFS versions on the given storage device.  There <bcp14>MUST</bcp14> be
exactly as many elements in ffds_fh_vers as there are in
ffda_versions.  Each element of the array corresponds to a particular
combination of ffdv_version, ffdv_minorversion, and
ffdv_tightly_coupled provided for the device.  The array allows for
server implementations that have different filehandles for different
combinations of version, minor version, and coupling strength.  See
<xref target="sec-version-errors"/> for how to handle versioning issues between the client
and storage devices.</t>
        <t>For tight coupling, ffds_stateid provides the stateid to be used by
the client to access the file.  For loose coupling and an NFSv4
storage device, the client will have to use an anonymous stateid to
perform I/O on the storage device.  With no control protocol, the
metadata server stateid cannot be used to provide a global stateid
model.  Thus, the server <bcp14>MUST</bcp14> set the ffds_stateid to be the
anonymous stateid.</t>
        <t>This specification of the ffds_stateid restricts both models for
NFSv4.x storage protocols:</t>
        <dl>
          <dt>loosely couple</dt>
          <dd>
            <t>the stateid has to be an anonymous stateid</t>
          </dd>
          <dt>tightly couple</dt>
          <dd>
            <t>the stateid has to be a global stateid</t>
          </dd>
        </dl>
        <t>A number of issues stem from a mismatch between the fact that
ffds_stateid is defined as a single item while ffds_fh_vers is
defined as an array.  It is possible for each open file on the
storage device to require its own open stateid.  Because there are
established loosely coupled implementations of the version of the
protocol described in this document, such potential issues have not
been addressed here.  It is possible for future layout types to be
defined that address these issues, should it become important to
provide multiple stateids for the same underlying file.</t>
        <t>For loosely coupled storage devices, ffds_user and ffds_group provide
the synthetic user and group to be used in the RPC credentials that
the client presents to the storage device to access the data files.
For tightly coupled storage devices, the user and group on the
storage device will be the same as on the metadata server; that is,
if ffdv_tightly_coupled (see <xref target="sec-ff_device_addr4"/>) is set, then the client
<bcp14>MUST</bcp14> ignore both ffds_user and ffds_group.</t>
        <t>The allowed values for both ffds_user and ffds_group are specified as
owner and owner_group, respectively, in Section 5.9 of <xref target="RFC8881"/>.
For NFSv3 compatibility, user and group strings that consist of
decimal numeric values with no leading zeros can be given a special
interpretation by clients and servers that choose to provide such
support.  The receiver may treat such a user or group string as
representing the same user as would be represented by an NFSv3 uid or
gid having the corresponding numeric value.  Note that if using
Kerberos for security, the expectation is that these values will be a
name@domain string.</t>
        <t>ffds_efficiency describes the metadata server's evaluation as to the
effectiveness of each mirror.  Note that this is per layout and not
per device as the metric may change due to perceived load,
availability to the metadata server, etc.  Higher values denote
higher perceived utility.  The way the client can select the best
mirror to access is discussed in <xref target="sec-select-mirror"/>.</t>
        <t>ffl_flags is a bitmap that allows the metadata server to inform the
client of particular conditions that may result from more or less
tight coupling of the storage devices.</t>
        <dl>
          <dt>FF_FLAGS_NO_LAYOUTCOMMIT</dt>
          <dd>
            <t>can be set to indicate that the client is not required to send
LAYOUTCOMMIT to the metadata server.</t>
          </dd>
          <dt>F_FLAGS_NO_IO_THRU_MDS</dt>
          <dd>
            <t>can be set to indicate that the client should not send I/O operations
to the metadata server.  That is, even if the client could determine
that there was a network disconnect to a storage device, the client
should not try to proxy the I/O through the metadata server.</t>
          </dd>
          <dt>FF_FLAGS_NO_READ_IO</dt>
          <dd>
            <t>can be set to indicate that the client should not send READ requests
with the layouts of iomode LAYOUTIOMODE4_RW.  Instead, it should request
a layout of iomode LAYOUTIOMODE4_READ from the metadata server.</t>
          </dd>
          <dt>FF_FLAGS_WRITE_ONE_MIRROR</dt>
          <dd>
            <t>can be set to indicate that the client only needs to update one of
the mirrors (see <xref target="sec-write-mirrors"/>).</t>
          </dd>
        </dl>
        <section anchor="error-codes-from-layoutget">
          <name>Error Codes from LAYOUTGET</name>
          <t><xref target="RFC8881"/> provides little guidance as to how the client is to proceed
with a LAYOUTGET that returns an error of either
NFS4ERR_LAYOUTTRYLATER, NFS4ERR_LAYOUTUNAVAILABLE, and NFS4ERR_DELAY.
Within the context of this document:</t>
          <dl>
            <dt>NFS4ERR_LAYOUTUNAVAILABLE</dt>
            <dd>
              <t>there is no layout available and the I/O is to go to the metadata
server.  Note that it is possible to have had a layout before a recall
and not after.</t>
            </dd>
            <dt>NFS4ERR_LAYOUTTRYLATER</dt>
            <dd>
              <t>there is some issue preventing the layout from being granted.  If the
client already has an appropriate layout, it should continue with I/O
to the storage devices.</t>
            </dd>
            <dt>NFS4ERR_DELAY</dt>
            <dd>
              <t>there is some issue preventing the layout from being granted.  If the
client already has an appropriate layout, it should not continue with
I/O to the storage devices.</t>
            </dd>
          </dl>
        </section>
        <section anchor="client-interactions-with-ffflagsnoiothrumds">
          <name>Client Interactions with FF_FLAGS_NO_IO_THRU_MDS</name>
          <t>Even if the metadata server provides the FF_FLAGS_NO_IO_THRU_MDS
flag, the client can still perform I/O to the metadata server.  The
flag functions as a hint.  The flag indicates to the client that the
metadata server prefers to separate the metadata I/O from the data I/
O, most likely for performance reasons.</t>
        </section>
      </section>
      <section anchor="layoutcommit-1">
        <name>LAYOUTCOMMIT</name>
        <t>The flexible file layout does not use lou_body inside the
loca_layoutupdate argument to LAYOUTCOMMIT.  If lou_type is
LAYOUT4_FLEX_FILES, the lou_body field <bcp14>MUST</bcp14> have a zero length (see
Section 18.42.1 of <xref target="RFC8881"/>).</t>
      </section>
      <section anchor="interactions-between-devices-and-layouts">
        <name>Interactions between Devices and Layouts</name>
        <t>The file layout type is defined such that the
relationship between multipathing and filehandles can result in
either 0, 1, or N filehandles (see Section 13.3 of <xref target="RFC8881"/>).  Some rationales
for this are clustered servers that share the same filehandle or
allow for multiple read-only copies of the file on the same storage
device.  In the flexible file layout type, while there is an array of
filehandles, they are independent of the multipathing being used.  If
the metadata server wants to provide multiple read-only copies of the
same file on the same storage device, then it should provide multiple
mirrored instances, each with a different ff_device_addr4.  The
client can then determine that, since the each of the ffds_fh_vers
are different, there are multiple copies of the file for the current
layout segment available.</t>
      </section>
      <section anchor="sec-version-errors">
        <name>Handling Version Errors</name>
        <t>When the metadata server provides the ffda_versions array in the
ff_device_addr4 (see <xref target="sec-ff_device_addr4"/>), the client is able to determine
whether or not it can access a storage device with any of the
supplied combinations of ffdv_version, ffdv_minorversion, and
ffdv_tightly_coupled.  However, due to the limitations of reporting
errors in GETDEVICEINFO (see Section 18.40 in <xref target="RFC8881"/>), the client
is not able to specify which specific device it cannot communicate
with over one of the provided ffdv_version and ffdv_minorversion
combinations.  Using ff_ioerr4 (<xref target="sec-ff_ioerr4"/>) inside either the
LAYOUTRETURN (see Section 18.44 of <xref target="RFC8881"/>) or the LAYOUTERROR (see
Section 15.6 of <xref target="RFC7862"/> and <xref target="sec-LAYOUTERROR"/> of this document), the
client can isolate the problematic storage device.</t>
        <t>The error code to return for LAYOUTRETURN and/or LAYOUTERROR is
NFS4ERR_MINOR_VERS_MISMATCH.  It does not matter whether the mismatch
is a major version (e.g., client can use NFSv3 but not NFSv4) or
minor version (e.g., client can use NFSv4.1 but not NFSv4.2), the
error indicates that for all the supplied combinations for
ffdv_version and ffdv_minorversion, the client cannot communicate
with the storage device.  The client can retry the GETDEVICEINFO to
see if the metadata server can provide a different combination, or it
can fall back to doing the I/O through the metadata server.</t>
      </section>
    </section>
    <section anchor="sec-striping">
      <name>Striping via Sparse Mapping</name>
      <t>While other layout types support both dense and sparse mapping of
logical offsets to physical offsets within a file (see, for example,
Section 13.4 of <xref target="RFC8881"/>), the flexible file layout type only
supports a sparse mapping.</t>
      <t>With sparse mappings, the logical offset within a file (L) is also
the physical offset on the storage device.  As detailed in
Section 13.4.4 of <xref target="RFC8881"/>, this results in holes across each
storage device that does not contain the current stripe index.</t>
      <figure anchor="fig-striping">
        <name>Stripe Mapping Math</name>
        <artwork><![CDATA[
L: logical offset within the file

W: stripe width
    W = number of elements in ffm_data_servers

S: number of bytes in a stripe
    S = W * ffl_stripe_unit

N: stripe number
    N = L / S
]]></artwork>
      </figure>
    </section>
    <section anchor="recovering-from-client-io-errors">
      <name>Recovering from Client I/O Errors</name>
      <t>The pNFS client may encounter errors when directly accessing the
storage devices.  However, it is the responsibility of the metadata
server to recover from the I/O errors.  When the LAYOUT4_FLEX_FILES
layout type is used, the client <bcp14>MUST</bcp14> report the I/O errors to the
server at LAYOUTRETURN time using the ff_ioerr4 structure (see
<xref target="sec-ff_ioerr4"/>).</t>
      <t>The metadata server analyzes the error and determines the required
recovery operations such as recovering media failures or
reconstructing missing data files.</t>
      <t>The metadata server <bcp14>MUST</bcp14> recall any outstanding layouts to allow it
exclusive write access to the stripes being recovered and to prevent
other clients from hitting the same error condition.  In these cases,
the server <bcp14>MUST</bcp14> complete recovery before handing out any new layouts
to the affected byte ranges.</t>
      <t>Although the client implementation has the option to propagate a
corresponding error to the application that initiated the I/O
operation and drop any unwritten data, the client should attempt to
retry the original I/O operation by either requesting a new layout or
sending the I/O via regular NFSv4.1+ READ or WRITE operations to the
metadata server.  The client <bcp14>SHOULD</bcp14> attempt to retrieve a new layout
and retry the I/O operation using the storage device first and only
retry the I/O operation via the metadata server if the error
persists.</t>
    </section>
    <section anchor="mirroring">
      <name>Mirroring</name>
      <t>The flexible file layout type has a simple model in place for the
mirroring of the file data constrained by a layout segment.  There is
no assumption that each copy of the mirror is stored identically on
the storage devices.  For example, one device might employ
compression or deduplication on the data.  However, the over-the-wire
transfer of the file contents <bcp14>MUST</bcp14> appear identical.  Note, this is a
constraint of the selected XDR representation in which each mirrored
copy of the layout segment has the same striping pattern (see
<xref target="fig-parallel-fileystem"/>).</t>
      <t>The metadata server is responsible for determining the number of
mirrored copies and the location of each mirror.  While the client
may provide a hint to how many copies it wants (see <xref target="sec-layouthint"/>), the
metadata server can ignore that hint; in any event, the client has no
means to dictate either the storage device (which also means the
coupling and/or protocol levels to access the layout segments) or the
location of said storage device.</t>
      <t>The updating of mirrored layout segments is done via client-side
mirroring.  With this approach, the client is responsible for making
sure modifications are made on all copies of the layout segments it
is informed of via the layout.  If a layout segment is being
resilvered to a storage device, that mirrored copy will not be in the
layout.  Thus, the metadata server <bcp14>MUST</bcp14> update that copy until the
client is presented it in a layout.  If the FF_FLAGS_WRITE_ONE_MIRROR
is set in ffl_flags, the client need only update one of the mirrors
(see <xref target="sec-write-mirrors"/>).  If the client is writing to the layout segments
via the metadata server, then the metadata server <bcp14>MUST</bcp14> update all
copies of the mirror.  As seen in <xref target="sec-mds-resilvering"/>, during the
resilvering, the layout is recalled, and the client has to make
modifications via the metadata server.</t>
      <section anchor="sec-select-mirror">
        <name>Selecting a Mirror</name>
        <t>When the metadata server grants a layout to a client, it <bcp14>MAY</bcp14> let the
client know how fast it expects each mirror to be once the request
arrives at the storage devices via the ffds_efficiency member.  While
the algorithms to calculate that value are left to the metadata
server implementations, factors that could contribute to that
calculation include speed of the storage device, physical memory
available to the device, operating system version, current load, etc.</t>
        <t>However, what should not be involved in that calculation is a
perceived network distance between the client and the storage device.
The client is better situated for making that determination based on
past interaction with the storage device over the different available
network interfaces between the two; that is, the metadata server
might not know about a transient outage between the client and
storage device because it has no presence on the given subnet.</t>
        <t>As such, it is the client that decides which mirror to access for
reading the file.  The requirements for writing to mirrored layout
segments are presented below.</t>
      </section>
      <section anchor="sec-write-mirrors">
        <name>Writing to Mirrors</name>
        <section anchor="single-storage-device-updates-mirrors">
          <name>Single Storage Device Updates Mirrors</name>
          <t>If the FF_FLAGS_WRITE_ONE_MIRROR flag in ffl_flags is set, the client
only needs to update one of the copies of the layout segment.  For
this case, the storage device <bcp14>MUST</bcp14> ensure that all copies of the
mirror are updated when any one of the mirrors is updated.  If the
storage device gets an error when updating one of the mirrors, then
it <bcp14>MUST</bcp14> inform the client that the original WRITE had an error.  The
client then <bcp14>MUST</bcp14> inform the metadata server (see <xref target="sec-write-errors"/>).  The
client's responsibility with respect to COMMIT is explained in
<xref target="sec-write-commits"/>.  The client may choose any one of the mirrors and may
use ffds_efficiency as described in <xref target="sec-select-mirror"/> when making this
choice.</t>
        </section>
        <section anchor="client-updates-all-mirrors">
          <name>Client Updates All Mirrors</name>
          <t>If the FF_FLAGS_WRITE_ONE_MIRROR flag in ffl_flags is not set, the
client is responsible for updating all mirrored copies of the layout
segments that it is given in the layout.  A single failed update is
sufficient to fail the entire operation.  If all but one copy is
updated successfully and the last one provides an error, then the
client needs to inform the metadata server about the error.  The
client can use either LAYOUTRETURN or LAYOUTERROR to inform the
metadata server that the update failed to that storage device.  If
the client is updating the mirrors serially, then it <bcp14>SHOULD</bcp14> stop at
the first error encountered and report that to the metadata server.
If the client is updating the mirrors in parallel, then it <bcp14>SHOULD</bcp14>
wait until all storage devices respond so that it can report all
errors encountered during the update.</t>
        </section>
        <section anchor="sec-write-errors">
          <name>Handling Write Errors</name>
          <t>When the client reports a write error to the metadata server, the
metadata server is responsible for determining if it wants to remove
the errant mirror from the layout, if the mirror has recovered from
some transient error, etc.  When the client tries to get a new
layout, the metadata server informs it of the decision by the
contents of the layout.  The client <bcp14>MUST NOT</bcp14> assume that the contents
of the previous layout will match those of the new one.  If it has
updates that were not committed to all mirrors, then it <bcp14>MUST</bcp14> resend
those updates to all mirrors.</t>
          <t>There is no provision in the protocol for the metadata server to
directly determine that the client has or has not recovered from an
error.  For example, if a storage device was network partitioned from
the client and the client reported the error to the metadata server,
then the network partition would be repaired, and all of the copies
would be successfully updated.  There is no mechanism for the client
to report that fact, and the metadata server is forced to repair the
file across the mirror.</t>
          <t>If the client supports NFSv4.2, it can use LAYOUTERROR and
LAYOUTRETURN to provide hints to the metadata server about the
recovery efforts.  A LAYOUTERROR on a file is for a non-fatal error.
A subsequent LAYOUTRETURN without a ff_ioerr4 indicates that the
client successfully replayed the I/O to all mirrors.  Any
LAYOUTRETURN with a ff_ioerr4 is an error that the metadata server
needs to repair.  The client <bcp14>MUST</bcp14> be prepared for the LAYOUTERROR to
trigger a CB_LAYOUTRECALL if the metadata server determines it needs
to start repairing the file.</t>
        </section>
        <section anchor="sec-write-commits">
          <name>Handling Write COMMITs</name>
          <t>When stable writes are done to the metadata server or to a single
replica (if allowed by the use of FF_FLAGS_WRITE_ONE_MIRROR), it is
the responsibility of the receiving node to propagate the written
data stably, before replying to the client.</t>
          <t>In the corresponding cases in which unstable writes are done, the
receiving node does not have any such obligation, although it may
choose to asynchronously propagate the updates.  However, once a
COMMIT is replied to, all replicas must reflect the writes that have
been done, and this data must have been committed to stable storage
on all replicas.</t>
          <t>In order to avoid situations in which stale data is read from
replicas to which writes have not been propagated:</t>
          <ul spacing="normal">
            <li>
              <t>A client that has outstanding unstable writes made to single node
(metadata server or storage device) <bcp14>MUST</bcp14> do all reads from that
same node.</t>
            </li>
            <li>
              <t>When writes are flushed to the server (for example, to implement
close-to-open semantics), a COMMIT must be done by the client to
ensure that up-to-date written data will be available irrespective
of the particular replica read.</t>
            </li>
          </ul>
        </section>
      </section>
      <section anchor="sec-mds-resilvering">
        <name>Metadata Server Resilvering of the File</name>
        <t>The metadata server may elect to create a new mirror of the layout
segments at any time.  This might be to resilver a copy on a storage
device that was down for servicing, to provide a copy of the layout
segments on storage with different storage performance
characteristics, etc.  As the client will not be aware of the new
mirror and the metadata server will not be aware of updates that the
client is making to the layout segments, the metadata server <bcp14>MUST</bcp14>
recall the writable layout segment(s) that it is resilvering.  If the
client issues a LAYOUTGET for a writable layout segment that is in
the process of being resilvered, then the metadata server can deny
that request with an NFS4ERR_LAYOUTUNAVAILABLE.  The client would
then have to perform the I/O through the metadata server.</t>
      </section>
    </section>
    <section anchor="sec-layouthint">
      <name>Flexible File Layout Type Return</name>
      <t>layoutreturn_file4 is used in the LAYOUTRETURN operation to convey
layout-type-specific information to the server.  It is defined in
Section 18.44.1 of <xref target="RFC8881"/> (also shown in <xref target="fig-LAYOUTRETURN"/>).</t>
      <figure anchor="fig-LAYOUTRETURN">
        <name>Layout Return XDR</name>
        <sourcecode type="xdr"><![CDATA[
      /* Constants used for LAYOUTRETURN and CB_LAYOUTRECALL */
      const LAYOUT4_RET_REC_FILE      = 1;
      const LAYOUT4_RET_REC_FSID      = 2;
      const LAYOUT4_RET_REC_ALL       = 3;

      enum layoutreturn_type4 {
              LAYOUTRETURN4_FILE = LAYOUT4_RET_REC_FILE,
              LAYOUTRETURN4_FSID = LAYOUT4_RET_REC_FSID,
              LAYOUTRETURN4_ALL  = LAYOUT4_RET_REC_ALL
      };

   struct layoutreturn_file4 {
           offset4         lrf_offset;
           length4         lrf_length;
           stateid4        lrf_stateid;
           /* layouttype4 specific data */
           opaque          lrf_body<>;
   };

   union layoutreturn4 switch(layoutreturn_type4 lr_returntype) {
           case LAYOUTRETURN4_FILE:
                   layoutreturn_file4      lr_layout;
           default:
                   void;
   };

   struct LAYOUTRETURN4args {
           /* CURRENT_FH: file */
           bool                    lora_reclaim;
           layouttype4             lora_layout_type;
           layoutiomode4           lora_iomode;
           layoutreturn4           lora_layoutreturn;
   };
]]></sourcecode>
      </figure>
      <t>If the lora_layout_type layout type is LAYOUT4_FLEX_FILES and the
lr_returntype is LAYOUTRETURN4_FILE, then the lrf_body opaque value
is defined by ff_layoutreturn4 (see <xref target="sec-ff_layoutreturn4"/>).  This allows the
client to report I/O error information or layout usage statistics
back to the metadata server as defined below.  Note that while the
data structures are built on concepts introduced in NFSv4.2, the
effective discriminated union (lora_layout_type combined with
ff_layoutreturn4) allows for an NFSv4.1 metadata server to utilize
the data.</t>
      <section anchor="sec-io-error">
        <name>I/O Error Reporting</name>
        <section anchor="sec-ff_ioerr4">
          <name>ff_ioerr4</name>
          <figure anchor="fig-ff_ioerr4">
            <name>ff_ioerr4</name>
            <sourcecode type="xdr"><![CDATA[
   /// struct ff_ioerr4 {
   ///         offset4        ffie_offset;
   ///         length4        ffie_length;
   ///         stateid4       ffie_stateid;
   ///         device_error4  ffie_errors<>;
   /// };
   ///
]]></sourcecode>
          </figure>
          <t>Recall that <xref target="RFC7862"/> defines device_error4 as in <xref target="fig-device_error4"/>:</t>
          <figure anchor="fig-device_error4">
            <name>device_error4</name>
            <sourcecode type="xdr"><![CDATA[
   struct device_error4 {
           deviceid4       de_deviceid;
           nfsstat4        de_status;
           nfs_opnum4      de_opnum;
   };
]]></sourcecode>
          </figure>
          <t>The ff_ioerr4 structure is used to return error indications for data
files that generated errors during data transfers.  These are hints
to the metadata server that there are problems with that file.  For
each error, ffie_errors.de_deviceid, ffie_offset, and ffie_length
represent the storage device and byte range within the file in which
the error occurred; ffie_errors represents the operation and type of
error.  The use of device_error4 is described in Section 15.6 of
<xref target="RFC7862"/>.</t>
          <t>Even though the storage device might be accessed via NFSv3 and
reports back NFSv3 errors to the client, the client is responsible
for mapping these to appropriate NFSv4 status codes as de_status.
Likewise, the NFSv3 operations need to be mapped to equivalent NFSv4
operations.</t>
        </section>
      </section>
      <section anchor="sec-layout-stats">
        <name>Layout Usage Statistics</name>
        <section anchor="ffiolatency4">
          <name>ff_io_latency4</name>
          <figure anchor="fig-ff_io_latency4">
            <name>ff_io_latency4</name>
            <sourcecode type="xdr"><![CDATA[
   /// struct ff_io_latency4 {
   ///         uint64_t       ffil_ops_requested;
   ///         uint64_t       ffil_bytes_requested;
   ///         uint64_t       ffil_ops_completed;
   ///         uint64_t       ffil_bytes_completed;
   ///         uint64_t       ffil_bytes_not_delivered;
   ///         nfstime4       ffil_total_busy_time;
   ///         nfstime4       ffil_aggregate_completion_time;
   /// };
   ///
]]></sourcecode>
          </figure>
          <t>Both operation counts and bytes transferred are kept in the
ff_io_latency4 (see <xref target="fig-ff_io_latency4"/>.  As seen in ff_layoutupdate4 (see
<xref target="sec-ff_layoutupdate4"/>), READ and WRITE operations are aggregated separately.
READ operations are used for the ff_io_latency4 ffl_read.  Both WRITE
and COMMIT operations are used for the ff_io_latency4 ffl_write.
"Requested" counters track what the client is attempting to do, and
"completed" counters track what was done.  There is no requirement
that the client only report completed results that have matching
requested results from the reported period.</t>
          <t>ffil_bytes_not_delivered is used to track the aggregate number of
bytes requested but not fulfilled due to error conditions.
ffil_total_busy_time is the aggregate time spent with outstanding RPC
calls. ffil_aggregate_completion_time is the sum of all round-trip
times for completed RPC calls.</t>
          <t>In Section 3.3.1 of <xref target="RFC8881"/>, the nfstime4 is defined as the number
of seconds and nanoseconds since midnight or zero hour January 1,
1970 Coordinated Universal Time (UTC).  The use of nfstime4 in
ff_io_latency4 is to store time since the start of the first I/O from
the client after receiving the layout.  In other words, these are to
be decoded as duration and not as a date and time.</t>
          <t>Note that LAYOUTSTATS are cumulative, i.e., not reset each time the
operation is sent.  If two LAYOUTSTATS operations for the same file
and layout stateid originate from the same NFS client and are
processed at the same time by the metadata server, then the one
containing the larger values contains the most recent time series
data.</t>
        </section>
        <section anchor="sec-ff_layoutupdate4">
          <name>ff_layoutupdate4</name>
          <figure anchor="fig-ff_layoutupdate4">
            <name>ff_layoutupdate4</name>
            <sourcecode type="xdr"><![CDATA[
   /// struct ff_layoutupdate4 {
   ///         netaddr4       ffl_addr;
   ///         nfs_fh4        ffl_fhandle;
   ///         ff_io_latency4 ffl_read;
   ///         ff_io_latency4 ffl_write;
   ///         nfstime4       ffl_duration;
   ///         bool           ffl_local;
   /// };
   ///
]]></sourcecode>
          </figure>
          <t>ffl_addr differentiates which network address the client is connected
to on the storage device.  In the case of multipathing, ffl_fhandle
indicates which read-only copy was selected. ffl_read and ffl_write
convey the latencies for both READ and WRITE operations,
respectively.  ffl_duration is used to indicate the time period over
which the statistics were collected.  If true, ffl_local indicates
that the I/O was serviced by the client's cache.  This flag allows
the client to inform the metadata server about "hot" access to a file
it would not normally be allowed to report on.</t>
        </section>
        <section anchor="ffiostats4">
          <name>ff_iostats4</name>
          <figure anchor="fig-ff_iostats4">
            <name>ff_iostats4</name>
            <sourcecode type="xdr"><![CDATA[
   /// struct ff_iostats4 {
   ///         offset4           ffis_offset;
   ///         length4           ffis_length;
   ///         stateid4          ffis_stateid;
   ///         io_info4          ffis_read;
   ///         io_info4          ffis_write;
   ///         deviceid4         ffis_deviceid;
   ///         ff_layoutupdate4  ffis_layoutupdate;
   /// };
   ///
]]></sourcecode>
          </figure>
          <t><xref target="RFC7862"/> defines io_info4 as in <xref target="fig-ff_iostats4"/>.</t>
          <figure anchor="fig-io_info4">
            <name>io_info4</name>
            <sourcecode type="xdr"><![CDATA[
   struct io_info4 {
           uint64_t        ii_count;
           uint64_t        ii_bytes;
   };
]]></sourcecode>
          </figure>
          <t>With pNFS, data transfers are performed directly between the pNFS
client and the storage devices.  Therefore, the metadata server has
no direct knowledge of the I/O operations being done and thus cannot
create on its own statistical information about client I/O to
optimize the data storage location.  ff_iostats4 <bcp14>MAY</bcp14> be used by the
client to report I/O statistics back to the metadata server upon
returning the layout.</t>
          <t>Since it is not feasible for the client to report every I/O that used
the layout, the client <bcp14>MAY</bcp14> identify "hot" byte ranges for which to
report I/O statistics.  The definition and/or configuration mechanism
of what is considered "hot" and the size of the reported byte range
are out of the scope of this document.  For client implementation,
providing reasonable default values and an optional run-time
management interface to control these parameters is suggested.  For
example, a client can define the default byte-range resolution to be
1 MB in size and the thresholds for reporting to be 1 MB/second or 10
I/O operations per second.</t>
          <t>For each byte range, ffis_offset and ffis_length represent the
starting offset of the range and the range length in bytes.
ffis_read.ii_count, ffis_read.ii_bytes, ffis_write.ii_count, and
ffis_write.ii_bytes represent the number of contiguous READ and WRITE
I/Os and the respective aggregate number of bytes transferred within
the reported byte range.</t>
          <t>The combination of ffis_deviceid and ffl_addr uniquely identifies
both the storage path and the network route to it.  Finally,
ffl_fhandle allows the metadata server to differentiate between
multiple read-only copies of the file on the same storage device.</t>
        </section>
      </section>
      <section anchor="sec-ff_layoutreturn4">
        <name>ff_layoutreturn4</name>
        <figure anchor="fig-ff_layoutreturn4">
          <name>ff_layoutreturn4</name>
          <sourcecode type="xdr"><![CDATA[
   /// struct ff_layoutreturn4 {
   ///         ff_ioerr4     fflr_ioerr_report<>;
   ///         ff_iostats4   fflr_iostats_report<>;
   /// };
   ///
]]></sourcecode>
        </figure>
        <t>When data file I/O operations fail, fflr_ioerr_report&lt;&gt; is used to
report these errors to the metadata server as an array of elements of
type ff_ioerr4.  Each element in the array represents an error that
occurred on the data file identified by ffie_errors.de_deviceid.  If
no errors are to be reported, the size of the fflr_ioerr_report&lt;&gt;
array is set to zero.  The client <bcp14>MAY</bcp14> also use fflr_iostats_report&lt;&gt;
to report a list of I/O statistics as an array of elements of type
ff_iostats4.  Each element in the array represents statistics for a
particular byte range.  Byte ranges are not guaranteed to be disjoint
and <bcp14>MAY</bcp14> repeat or intersect.</t>
      </section>
    </section>
    <section anchor="sec-LAYOUTERROR">
      <name>Flexible File Layout Type LAYOUTERROR</name>
      <t>If the client is using NFSv4.2 to communicate with the metadata
server, then instead of waiting for a LAYOUTRETURN to send error
information to the metadata server (see <xref target="sec-io-error"/>), it <bcp14>MAY</bcp14> use
LAYOUTERROR (see Section 15.6 of <xref target="RFC7862"/>) to communicate that
information.  For the flexible file layout type, this means that
LAYOUTERROR4args is treated the same as ff_ioerr4.</t>
    </section>
    <section anchor="flexible-file-layout-type-layoutstats">
      <name>Flexible File Layout Type LAYOUTSTATS</name>
      <t>If the client is using NFSv4.2 to communicate with the metadata
server, then instead of waiting for a LAYOUTRETURN to send I/O
statistics to the metadata server (see <xref target="sec-layout-stats"/>), it <bcp14>MAY</bcp14> use
LAYOUTSTATS (see Section 15.7 of <xref target="RFC7862"/>) to communicate that
information.  For the flexible file layout type, this means that
LAYOUTSTATS4args.lsa_layoutupdate is overloaded with the same
contents as in ffis_layoutupdate.</t>
    </section>
    <section anchor="flexible-file-layout-type-creation-hint">
      <name>Flexible File Layout Type Creation Hint</name>
      <t>The layouthint4 type is defined in the <xref target="RFC8881"/> as in <xref target="fig-layouthint4-v1"/>.</t>
      <figure anchor="fig-layouthint4-v1">
        <name>layouthint4 v1</name>
        <sourcecode type="xdr"><![CDATA[
   struct layouthint4 {
       layouttype4        loh_type;
       opaque             loh_body<>;
   };
]]></sourcecode>
      </figure>
      <artwork><![CDATA[
                          {{fig-layouthint4-v1}}
]]></artwork>
      <t>The layouthint4 structure is used by the client to pass a hint about
the type of layout it would like created for a particular file.  If
the loh_type layout type is LAYOUT4_FLEX_FILES, then the loh_body
opaque value is defined by the ff_layouthint4 type.</t>
    </section>
    <section anchor="fflayouthint4">
      <name>ff_layouthint4</name>
      <figure anchor="fig-ff_layouthint4-v2">
        <name>ff_layouthint4 v2</name>
        <sourcecode type="xdr"><![CDATA[
   /// union ff2_mirrors_hint switch (ff2_protection_type ffmh_type) {
   ///     case FF2_PROTECTION_TYPE_MOJETTE:
   ///         void;
   ///     case FF2_PROTECTION_TYPE_MIRRORED:
   ///         void;
   /// };
   ///
   /// /*
   ///  * We could have this be simply ff2_protection_type
   ///  * for the client to state what protection algorithm
   ///  * it wants.
   ///  */
   /// struct ff2_layouthint4 {
   ///     ff2_protection_type fflh_supported_types<>;
   ///     ff2_mirrors_hint fflh_mirrors_hint;
   /// };

   union ff_mirrors_hint switch (bool ffmc_valid) {
       case TRUE:
           uint32_t    ffmc_mirrors;
       case FALSE:
           void;
   };

   struct ff_layouthint4 {
       ff_mirrors_hint    fflh_mirrors_hint;
   };
]]></sourcecode>
      </figure>
      <t>This type conveys hints for the desired data map.  All parameters are
optional so the client can give values for only the parameter it
cares about.</t>
    </section>
    <section anchor="recalling-a-layout">
      <name>Recalling a Layout</name>
      <t>While Section 12.5.5 of <xref target="RFC8881"/> discusses reasons independent of
layout type for recalling a layout, the flexible file layout type
metadata server should recall outstanding layouts in the following
cases:</t>
      <ul spacing="normal">
        <li>
          <t>When the file's security policy changes, i.e., ACLs or permission
mode bits are set.</t>
        </li>
        <li>
          <t>When the file's layout changes, rendering outstanding layouts
invalid.</t>
        </li>
        <li>
          <t>When existing layouts are inconsistent with the need to enforce
locking constraints.</t>
        </li>
        <li>
          <t>When existing layouts are inconsistent with the requirements
regarding resilvering as described in <xref target="sec-mds-resilvering"/>.</t>
        </li>
      </ul>
      <section anchor="cbrecallany">
        <name>CB_RECALL_ANY</name>
        <t>The metadata server can use the CB_RECALL_ANY callback operation to
notify the client to return some or all of its layouts.  Section 22.3
of <xref target="RFC8881"/> defines the allowed types of the "NFSv4 Recallable
Object Types Registry".</t>
        <figure anchor="fig-new-rca4">
          <name>RCA4 masks for v2</name>
          <sourcecode type="xdr"><![CDATA[
   /// const RCA4_TYPE_MASK_FF2_LAYOUT_MIN     = 20;
   /// const RCA4_TYPE_MASK_FF2_LAYOUT_MAX     = 21;
   ///
]]></sourcecode>
        </figure>
        <figure anchor="fig-CB_RECALL_ANY4args">
          <name>CB_RECALL_ANY4args XDR</name>
          <sourcecode type="xdr"><![CDATA[
   struct  CB_RECALL_ANY4args      {
       uint32_t        craa_layouts_to_keep;
       bitmap4         craa_type_mask;
   };
]]></sourcecode>
        </figure>
        <t>Typically, CB_RECALL_ANY will be used to recall client state when the
server needs to reclaim resources.  The craa_type_mask bitmap
specifies the type of resources that are recalled, and the
craa_layouts_to_keep value specifies how many of the recalled
flexible file layouts the client is allowed to keep.  The mask flags
for the flexible file layout type are defined as in <xref target="fig-mask-flags"/>.</t>
        <figure anchor="fig-mask-flags">
          <name>Recall Mask Flags for v2</name>
          <sourcecode type="xdr"><![CDATA[
   /// enum ff_cb_recall_any_mask {
   ///     PNFS_FF_RCA4_TYPE_MASK_READ = 20,
   ///     PNFS_FF_RCA4_TYPE_MASK_RW   = 21
   /// };
   ///
]]></sourcecode>
        </figure>
        <t>The flags represent the iomode of the recalled layouts.  In response,
the client <bcp14>SHOULD</bcp14> return layouts of the recalled iomode that it needs
the least, keeping at most craa_layouts_to_keep flexible file
layouts.</t>
        <t>The PNFS_FF_RCA4_TYPE_MASK_READ flag notifies the client to return
layouts of iomode LAYOUTIOMODE4_READ.  Similarly, the
PNFS_FF_RCA4_TYPE_MASK_RW flag notifies the client to return layouts
of iomode LAYOUTIOMODE4_RW.  When both mask flags are set, the client
is notified to return layouts of either iomode.</t>
      </section>
    </section>
    <section anchor="client-fencing">
      <name>Client Fencing</name>
      <t>In cases where clients are uncommunicative and their lease has
expired or when clients fail to return recalled layouts within a
lease period, the server <bcp14>MAY</bcp14> revoke client layouts and reassign these
resources to other clients (see Section 12.5.5 of <xref target="RFC8881"/>).  To
avoid data corruption, the metadata server <bcp14>MUST</bcp14> fence off the revoked
clients from the respective data files as described in Section 2.2.</t>
    </section>
    <section anchor="security-considerations">
      <name>Security Considerations</name>
      <t>The combination of components in a pNFS system is required to
preserve the security properties of NFSv4.1+ with respect to an
entity accessing data via a client.  The pNFS feature partitions the
NFSv4.1+ file system protocol into two parts: the control protocol
and the data protocol.  As the control protocol in this document is
NFS, the security properties are equivalent to the version of NFS
being used.  The flexible file layout further divides the data
protocol into metadata and data paths.  The security properties of
the metadata path are equivalent to those of NFSv4.1x (see Sections
1.7.1 and 2.2.1 of <xref target="RFC8881"/>).  And the security properties of the
data path are equivalent to those of the version of NFS used to
access the storage device, with the provision that the metadata
server is responsible for authenticating client access to the data
file.  The metadata server provides appropriate credentials to the
client to access data files on the storage device.  It is also
responsible for revoking access for a client to the storage device.</t>
      <t>The metadata server enforces the file access control policy at
LAYOUTGET time.  The client should use RPC authorization credentials
for getting the layout for the requested iomode ((LAYOUTIOMODE4_READ
or LAYOUTIOMODE4_RW), and the server verifies the permissions and ACL
for these credentials, possibly returning NFS4ERR_ACCESS if the
client is not allowed the requested iomode.  If the LAYOUTGET
operation succeeds, the client receives, as part of the layout, a set
of credentials allowing it I/O access to the specified data files
corresponding to the requested iomode.  When the client acts on I/O
operations on behalf of its local users, it <bcp14>MUST</bcp14> authenticate and
authorize the user by issuing respective OPEN and ACCESS calls to the
metadata server, similar to having NFSv4 data delegations.</t>
      <t>The combination of filehandle, synthetic uid, and gid in the layout
is the way that the metadata server enforces access control to the
data server.  The client only has access to filehandles of file
objects and not directory objects.  Thus, given a filehandle in a
layout, it is not possible to guess the parent directory filehandle.
Further, as the data file permissions only allow the given synthetic
uid read/write permission and the given synthetic gid read
permission, knowing the synthetic ids of one file does not
necessarily allow access to any other data file on the storage
device.</t>
      <t>The metadata server can also deny access at any time by fencing the
data file, which means changing the synthetic ids.  In turn, that
forces the client to return its current layout and get a new layout
if it wants to continue I/O to the data file.</t>
      <t>If access is allowed, the client uses the corresponding (read-only or
read/write) credentials to perform the I/O operations at the data
file's storage devices.  When the metadata server receives a request
to change a file's permissions or ACL, it <bcp14>SHOULD</bcp14> recall all layouts
for that file and then <bcp14>MUST</bcp14> fence off any clients still holding
outstanding layouts for the respective files by implicitly
invalidating the previously distributed credential on all data file
comprising the file in question.  It is <bcp14>REQUIRED</bcp14> that this be done
before committing to the new permissions and/or ACL.  By requesting
new layouts, the clients will reauthorize access against the modified
access control metadata.  Recalling the layouts in this case is
intended to prevent clients from getting an error on I/Os done after
the client was fenced off.</t>
      <section anchor="transport-layer-security">
        <name>Transport Layer Security</name>
      </section>
      <section anchor="rpcsecgss-and-security-services">
        <name>RPCSEC_GSS and Security Services</name>
        <t><em>Why we don't want to support RPCSEC_GSS.</em></t>
        <t>Because of the special use of principals within the loosely coupled
model, the issues are different depending on the coupling model.</t>
        <section anchor="loosely-coupled">
          <name>Loosely Coupled</name>
          <t>RPCSEC_GSS version 3 (RPCSEC_GSSv3) <xref target="RFC7861"/> contains facilities
that would allow it to be used to authorize the client to the storage
device on behalf of the metadata server.  Doing so would require that
each of the metadata server, storage device, and client would need to
implement RPCSEC_GSSv3 using an RPC-application-defined structured
privilege assertion in a manner described in Section 4.9.1 of
<xref target="RFC7862"/>.  The specifics necessary to do so are not described in
this document.  This is principally because any such specification
would require extensive implementation work on a wide range of
storage devices, which would be unlikely to result in a widely usable
specification for a considerable time.</t>
          <t>As a result, the layout type described in this document will not
provide support for use of RPCSEC_GSS together with the loosely
coupled model.  However, future layout types could be specified,
which would allow such support, either through the use of
RPCSEC_GSSv3 or in other ways.</t>
        </section>
        <section anchor="tightly-coupled">
          <name>Tightly Coupled</name>
          <t>With tight coupling, the principal used to access the metadata file
is exactly the same as used to access the data file.  The storage
device can use the control protocol to validate any RPC credentials.
As a result, there are no security issues related to using RPCSEC_GSS
with a tightly coupled system.  For example, if Kerberos V5 Generic
Security Service Application Program Interface (GSS-API) <xref target="RFC4121"/> is
used as the security mechanism, then the storage device could use a
control protocol to validate the RPC credentials to the metadata
server.</t>
        </section>
      </section>
    </section>
    <section anchor="iana-considerations">
      <name>IANA Considerations</name>
      <t><xref target="RFC8881"/> introduced the "pNFS Layout Types Registry"; new layout
type numbers in this registry need to be assigned by IANA.  This
document defines the protocol associated with an existing layout type
number: LAYOUT4_FLEX_FILES (see <xref target="tbl_layout_types"/>).</t>
      <table anchor="tbl_layout_types">
        <name>Layout Type Assignments</name>
        <thead>
          <tr>
            <th align="left">Layout Type Name</th>
            <th align="left">Value</th>
            <th align="left">RFC</th>
            <th align="left">How</th>
            <th align="left">Minor Versions</th>
          </tr>
        </thead>
        <tbody>
          <tr>
            <td align="left">LAYOUT4_FLEX_FILES_V2</td>
            <td align="left">0x6</td>
            <td align="left">RFCTBD10</td>
            <td align="left">L</td>
            <td align="left">1</td>
          </tr>
        </tbody>
      </table>
      <t><xref target="RFC8881"/> also introduced the "NFSv4 Recallable Object Types
Registry".  This document defines new recallable objects for
RCA4_TYPE_MASK_FF2_LAYOUT_MIN and RCA4_TYPE_MASK_FF2_LAYOUT_MAX (see
<xref target="tbl_recallables"/>).</t>
      <table anchor="tbl_recallables">
        <name>Recallable Object Type Assignments</name>
        <thead>
          <tr>
            <th align="left">Recallable Object Type Name</th>
            <th align="left">Value</th>
            <th align="left">RFC</th>
            <th align="left">How</th>
            <th align="left">Minor Versions</th>
          </tr>
        </thead>
        <tbody>
          <tr>
            <td align="left">RCA4_TYPE_MASK_FF2_LAYOUT_MIN</td>
            <td align="left">20</td>
            <td align="left">RFCTBD10</td>
            <td align="left">L</td>
            <td align="left">1</td>
          </tr>
          <tr>
            <td align="left">RCA4_TYPE_MASK_FF2_LAYOUT_MAX</td>
            <td align="left">21</td>
            <td align="left">RFCTBD10</td>
            <td align="left">L</td>
            <td align="left">1</td>
          </tr>
        </tbody>
      </table>
    </section>
    <section numbered="false" anchor="acknowledgments">
      <name>Acknowledgments</name>
      <t>The following from Hammerspace were instrumental in driving Flex
Files v2: David Flynn, Trond Myklebust, Didier Feron, Jean-Pierre
Monchanin, Pierre Evenou, and Brian Pawlowski.</t>
      <t>Christoph Helwig was instrumental in making sure Flex Files v2 was
applicable to more than the Mojette Transformation.</t>
      <t>Pierre Evenou provided the sections for the Mojette Transformation.</t>
    </section>
    <section numbered="false" anchor="open-action-items">
      <name>Open Action Items</name>
      <t>This section is to be removed before publishing as an RFC.</t>
      <ol spacing="normal" type="1"><li>
          <t>How to describe projection header?</t>
        </li>
        <li>
          <t>It is Little Endian, so not good for XDR?</t>
        </li>
        <li>
          <t>If we add XDR, how does v3 handle it?</t>
        </li>
        <li>
          <t>IANA registration for new Protection Types</t>
        </li>
        <li>
          <t>Proxy registration</t>
        </li>
        <li>
          <t>TLS</t>
        </li>
      </ol>
    </section>
  </middle>
  <back>
    <references anchor="sec-combined-references">
      <name>References</name>
      <references anchor="sec-normative-references">
        <name>Normative References</name>
        <reference anchor="RFC2119" xml:base="https://bib.ietf.org/public/rfc/bibxml/reference.RFC.2119.xml">
          <front>
            <title>Key words for use in RFCs to Indicate Requirement Levels</title>
            <author fullname="S. Bradner" initials="S." surname="Bradner"/>
            <date month="March" year="1997"/>
            <abstract>
              <t>In many standards track documents several words are used to signify the requirements in the specification. These words are often capitalized. This document defines these words as they should be interpreted in IETF documents. This document specifies an Internet Best Current Practices for the Internet Community, and requests discussion and suggestions for improvements.</t>
            </abstract>
          </front>
          <seriesInfo name="BCP" value="14"/>
          <seriesInfo name="RFC" value="2119"/>
          <seriesInfo name="DOI" value="10.17487/RFC2119"/>
        </reference>
        <reference anchor="RFC4121" xml:base="https://bib.ietf.org/public/rfc/bibxml/reference.RFC.4121.xml">
          <front>
            <title>The Kerberos Version 5 Generic Security Service Application Program Interface (GSS-API) Mechanism: Version 2</title>
            <author fullname="L. Zhu" initials="L." surname="Zhu"/>
            <author fullname="K. Jaganathan" initials="K." surname="Jaganathan"/>
            <author fullname="S. Hartman" initials="S." surname="Hartman"/>
            <date month="July" year="2005"/>
            <abstract>
              <t>This document defines protocols, procedures, and conventions to be employed by peers implementing the Generic Security Service Application Program Interface (GSS-API) when using the Kerberos Version 5 mechanism.</t>
              <t>RFC 1964 is updated and incremental changes are proposed in response to recent developments such as the introduction of Kerberos cryptosystem framework. These changes support the inclusion of new cryptosystems, by defining new per-message tokens along with their encryption and checksum algorithms based on the cryptosystem profiles. [STANDARDS-TRACK]</t>
            </abstract>
          </front>
          <seriesInfo name="RFC" value="4121"/>
          <seriesInfo name="DOI" value="10.17487/RFC4121"/>
        </reference>
        <reference anchor="RFC4506" xml:base="https://bib.ietf.org/public/rfc/bibxml/reference.RFC.4506.xml">
          <front>
            <title>XDR: External Data Representation Standard</title>
            <author fullname="M. Eisler" initials="M." role="editor" surname="Eisler"/>
            <date month="May" year="2006"/>
            <abstract>
              <t>This document describes the External Data Representation Standard (XDR) protocol as it is currently deployed and accepted. This document obsoletes RFC 1832. [STANDARDS-TRACK]</t>
            </abstract>
          </front>
          <seriesInfo name="STD" value="67"/>
          <seriesInfo name="RFC" value="4506"/>
          <seriesInfo name="DOI" value="10.17487/RFC4506"/>
        </reference>
        <reference anchor="RFC5531" xml:base="https://bib.ietf.org/public/rfc/bibxml/reference.RFC.5531.xml">
          <front>
            <title>RPC: Remote Procedure Call Protocol Specification Version 2</title>
            <author fullname="R. Thurlow" initials="R." surname="Thurlow"/>
            <date month="May" year="2009"/>
            <abstract>
              <t>This document describes the Open Network Computing (ONC) Remote Procedure Call (RPC) version 2 protocol as it is currently deployed and accepted. This document obsoletes RFC 1831. [STANDARDS-TRACK]</t>
            </abstract>
          </front>
          <seriesInfo name="RFC" value="5531"/>
          <seriesInfo name="DOI" value="10.17487/RFC5531"/>
        </reference>
        <reference anchor="RFC5662" xml:base="https://bib.ietf.org/public/rfc/bibxml/reference.RFC.5662.xml">
          <front>
            <title>Network File System (NFS) Version 4 Minor Version 1 External Data Representation Standard (XDR) Description</title>
            <author fullname="S. Shepler" initials="S." role="editor" surname="Shepler"/>
            <author fullname="M. Eisler" initials="M." role="editor" surname="Eisler"/>
            <author fullname="D. Noveck" initials="D." role="editor" surname="Noveck"/>
            <date month="January" year="2010"/>
            <abstract>
              <t>This document provides the External Data Representation Standard (XDR) description for Network File System version 4 (NFSv4) minor version 1. [STANDARDS-TRACK]</t>
            </abstract>
          </front>
          <seriesInfo name="RFC" value="5662"/>
          <seriesInfo name="DOI" value="10.17487/RFC5662"/>
        </reference>
        <reference anchor="RFC7530" xml:base="https://bib.ietf.org/public/rfc/bibxml/reference.RFC.7530.xml">
          <front>
            <title>Network File System (NFS) Version 4 Protocol</title>
            <author fullname="T. Haynes" initials="T." role="editor" surname="Haynes"/>
            <author fullname="D. Noveck" initials="D." role="editor" surname="Noveck"/>
            <date month="March" year="2015"/>
            <abstract>
              <t>The Network File System (NFS) version 4 protocol is a distributed file system protocol that builds on the heritage of NFS protocol version 2 (RFC 1094) and version 3 (RFC 1813). Unlike earlier versions, the NFS version 4 protocol supports traditional file access while integrating support for file locking and the MOUNT protocol. In addition, support for strong security (and its negotiation), COMPOUND operations, client caching, and internationalization has been added. Of course, attention has been applied to making NFS version 4 operate well in an Internet environment.</t>
              <t>This document, together with the companion External Data Representation (XDR) description document, RFC 7531, obsoletes RFC 3530 as the definition of the NFS version 4 protocol.</t>
            </abstract>
          </front>
          <seriesInfo name="RFC" value="7530"/>
          <seriesInfo name="DOI" value="10.17487/RFC7530"/>
        </reference>
        <reference anchor="RFC7861" xml:base="https://bib.ietf.org/public/rfc/bibxml/reference.RFC.7861.xml">
          <front>
            <title>Remote Procedure Call (RPC) Security Version 3</title>
            <author fullname="A. Adamson" initials="A." surname="Adamson"/>
            <author fullname="N. Williams" initials="N." surname="Williams"/>
            <date month="November" year="2016"/>
            <abstract>
              <t>This document specifies version 3 of the Remote Procedure Call (RPC) security protocol (RPCSEC_GSS). This protocol provides support for multi-principal authentication of client hosts and user principals to a server (constructed by generic composition), security label assertions for multi-level security and type enforcement, structured privilege assertions, and channel bindings. This document updates RFC 5403.</t>
            </abstract>
          </front>
          <seriesInfo name="RFC" value="7861"/>
          <seriesInfo name="DOI" value="10.17487/RFC7861"/>
        </reference>
        <reference anchor="RFC7862" xml:base="https://bib.ietf.org/public/rfc/bibxml/reference.RFC.7862.xml">
          <front>
            <title>Network File System (NFS) Version 4 Minor Version 2 Protocol</title>
            <author fullname="T. Haynes" initials="T." surname="Haynes"/>
            <date month="November" year="2016"/>
            <abstract>
              <t>This document describes NFS version 4 minor version 2; it describes the protocol extensions made from NFS version 4 minor version 1. Major extensions introduced in NFS version 4 minor version 2 include the following: Server-Side Copy, Application Input/Output (I/O) Advise, Space Reservations, Sparse Files, Application Data Blocks, and Labeled NFS.</t>
            </abstract>
          </front>
          <seriesInfo name="RFC" value="7862"/>
          <seriesInfo name="DOI" value="10.17487/RFC7862"/>
        </reference>
        <reference anchor="RFC8174" xml:base="https://bib.ietf.org/public/rfc/bibxml/reference.RFC.8174.xml">
          <front>
            <title>Ambiguity of Uppercase vs Lowercase in RFC 2119 Key Words</title>
            <author fullname="B. Leiba" initials="B." surname="Leiba"/>
            <date month="May" year="2017"/>
            <abstract>
              <t>RFC 2119 specifies common key words that may be used in protocol specifications. This document aims to reduce the ambiguity by clarifying that only UPPERCASE usage of the key words have the defined special meanings.</t>
            </abstract>
          </front>
          <seriesInfo name="BCP" value="14"/>
          <seriesInfo name="RFC" value="8174"/>
          <seriesInfo name="DOI" value="10.17487/RFC8174"/>
        </reference>
        <reference anchor="RFC8434" xml:base="https://bib.ietf.org/public/rfc/bibxml/reference.RFC.8434.xml">
          <front>
            <title>Requirements for Parallel NFS (pNFS) Layout Types</title>
            <author fullname="T. Haynes" initials="T." surname="Haynes"/>
            <date month="August" year="2018"/>
            <abstract>
              <t>This document defines the requirements that individual Parallel NFS (pNFS) layout types need to meet in order to work within the pNFS framework as defined in RFC 5661. In so doing, this document aims to clearly distinguish between requirements for pNFS as a whole and those specifically directed to the pNFS file layout. The lack of a clear separation between the two sets of requirements has been troublesome for those specifying and evaluating new layout types. In this regard, this document updates RFC 5661.</t>
            </abstract>
          </front>
          <seriesInfo name="RFC" value="8434"/>
          <seriesInfo name="DOI" value="10.17487/RFC8434"/>
        </reference>
        <reference anchor="RFC8435" xml:base="https://bib.ietf.org/public/rfc/bibxml/reference.RFC.8435.xml">
          <front>
            <title>Parallel NFS (pNFS) Flexible File Layout</title>
            <author fullname="B. Halevy" initials="B." surname="Halevy"/>
            <author fullname="T. Haynes" initials="T." surname="Haynes"/>
            <date month="August" year="2018"/>
            <abstract>
              <t>Parallel NFS (pNFS) allows a separation between the metadata (onto a metadata server) and data (onto a storage device) for a file. The flexible file layout type is defined in this document as an extension to pNFS that allows the use of storage devices that require only a limited degree of interaction with the metadata server and use already-existing protocols. Client-side mirroring is also added to provide replication of files.</t>
            </abstract>
          </front>
          <seriesInfo name="RFC" value="8435"/>
          <seriesInfo name="DOI" value="10.17487/RFC8435"/>
        </reference>
        <reference anchor="RFC8881" xml:base="https://bib.ietf.org/public/rfc/bibxml/reference.RFC.8881.xml">
          <front>
            <title>Network File System (NFS) Version 4 Minor Version 1 Protocol</title>
            <author fullname="D. Noveck" initials="D." role="editor" surname="Noveck"/>
            <author fullname="C. Lever" initials="C." surname="Lever"/>
            <date month="August" year="2020"/>
            <abstract>
              <t>This document describes the Network File System (NFS) version 4 minor version 1, including features retained from the base protocol (NFS version 4 minor version 0, which is specified in RFC 7530) and protocol extensions made subsequently. The later minor version has no dependencies on NFS version 4 minor version 0, and is considered a separate protocol.</t>
              <t>This document obsoletes RFC 5661. It substantially revises the treatment of features relating to multi-server namespace, superseding the description of those features appearing in RFC 5661.</t>
            </abstract>
          </front>
          <seriesInfo name="RFC" value="8881"/>
          <seriesInfo name="DOI" value="10.17487/RFC8881"/>
        </reference>
      </references>
      <references anchor="sec-informative-references">
        <name>Informative References</name>
        <reference anchor="RFC1813" xml:base="https://bib.ietf.org/public/rfc/bibxml/reference.RFC.1813.xml">
          <front>
            <title>NFS Version 3 Protocol Specification</title>
            <author fullname="B. Callaghan" initials="B." surname="Callaghan"/>
            <author fullname="B. Pawlowski" initials="B." surname="Pawlowski"/>
            <author fullname="P. Staubach" initials="P." surname="Staubach"/>
            <date month="June" year="1995"/>
            <abstract>
              <t>This paper describes the NFS version 3 protocol. This paper is provided so that people can write compatible implementations. This memo provides information for the Internet community. This memo does not specify an Internet standard of any kind.</t>
            </abstract>
          </front>
          <seriesInfo name="RFC" value="1813"/>
          <seriesInfo name="DOI" value="10.17487/RFC1813"/>
        </reference>
        <reference anchor="RFC4519" xml:base="https://bib.ietf.org/public/rfc/bibxml/reference.RFC.4519.xml">
          <front>
            <title>Lightweight Directory Access Protocol (LDAP): Schema for User Applications</title>
            <author fullname="A. Sciberras" initials="A." role="editor" surname="Sciberras"/>
            <date month="June" year="2006"/>
            <abstract>
              <t>This document is an integral part of the Lightweight Directory Access Protocol (LDAP) technical specification. It provides a technical specification of attribute types and object classes intended for use by LDAP directory clients for many directory services, such as White Pages. These objects are widely used as a basis for the schema in many LDAP directories. This document does not cover attributes used for the administration of directory servers, nor does it include directory objects defined for specific uses in other documents. [STANDARDS-TRACK]</t>
            </abstract>
          </front>
          <seriesInfo name="RFC" value="4519"/>
          <seriesInfo name="DOI" value="10.17487/RFC4519"/>
        </reference>
      </references>
    </references>
  </back>
  <!-- ##markdown-source:
H4sIAAAAAAAAA929+XIbR5Yv/H8+RY58Iyx1g5BI0UvT4/HQXGzOaLsk5SUm
5kMkUAkgrUIVurJACm1rnuU+y/dkX5wtl6oCKPXMvRP3c0xMi4XKrFzP+jvn
HBwcqNa1pT3Rb0xjytKW+tXljX68fnV580Rflva9m5ZWX7rS6hdmW29a/ZNt
vKsrfaTMdNrYuxN8LXvl7kgV9awyK3uii8bM24Ol2VbWH1Rzf3d8MC/t+7kr
rT+4OzooTWt9q2amtYu62Z5o3xaqMK090b+fn95efFCzuvK28ht/ottmY5Vb
N/gv3x49e/aXZ0fKNNac6B9sZRtTqvu6ebdo6s36RL+yLfxFY7vZ+tauwvCP
1Tu7va+b4kRfVa1tKtsenMNYlfKtqYqJKevKnuit9WrtTvS/tfVspH3dtI2d
+5H22xX9o6hnK7Neu2ox0rN6tbJV6/9dKbNpl3VzovSB0lprV/kTfTvWP+JC
4CNan9tlvTI+fV43C1O5v5nW1dWJ/tGsVrbxazOz+KtdGVee6LJeLLfNPy/g
r/GsXilV1c3KtO7Oniitry/Pjg4P/8L/PD48OpR/fvHsS/7nF188l6dffPnl
Ef/zqy+eP5N/fv3lYfynvPD14VfH8s/j58k/v5B/fv314YlSylXzzpAOvz58
HsYBo1MHBwfaTH3bmFmr1NAZNGVZ33tttLdr0+Ci6Klt762tdLu0emVbU5jW
6Md11dbaqPDA2+bONk+0qQqdvqF9WzdmYXVh79zMPtHzutFGw4kca327tGou
5x6e6ZIOdbtdW+28LuzcVbbQDr4Pf9ezDWy6Nl6bStv3ra3ghKm21jAF3S5N
K9OAEW+81fW8MwpPrzX2rxvXWF1X5VYbVbqVa22hC7toLLZycFbNDJfh3rXL
fA1oyjjjjbfKlI01xfbAvne+ddVCr5u6rWd16cdan0MDeGCpN+e1KX2tTVHY
QsPgm/rOFVbBJxeNa7djrb+v26U+K52t2gPvCqtXrmnqBvqGj8JgXta/2ba1
2pSLunHtcqVNY9XG2wIXusg/O6YzUNWtnbyC/9fWk2trCtt4pc6dn208Xtd6
zqsNV1S35p31el3CpajpHLy6vLk71nDdYTBIADRcDvirdL7Vj5H4/LOz7Xxc
N4snI3W/dLMlzruZLd2dLbRp1b/9++Nl2679ydOn0Jx/Gkuzp/Dgqbf4P9/h
ZZxA999i70/G+qbeNDOrZnVhcUmc9xvraQgw/2QWMwNHWc/rTdX99MK1y80U
LvZT+PLB/YJo59NNNTOzpTXT0j4ZK/Uzz/cHnG+4cXXV7Vx/ROfQ4W0cnvN6
tmkaW7VwFnFp4dCvm3rRWA9H6KrVlbWFh9MytaqwrW1WdDnm+t7qe1O18Nus
Xm/13aFu7Xv8++5I1036zm8b32qjCzefK9xpqyt7r2d11dqqHWt9WTe6cYtl
q6v6fgTtTGOxX5g9vC/dm6pQpijkMXTDP421Pi19PdL2/drOWu3pBOLoV/Wd
hQt8b/V64/FWKbtaL413Xs+beqWBJiA98Hw35N4AXfBjpSanQKV+vnozwg//
cn4NtEDu6sps4ZDrrW1xJccTOvgrVxSlVeoz4EJNXWzwdaWuqiGOPBq87o1t
N03lU0KlfNtsZu2mEcJSWD9r3NTq+6VtmLBhH87rsgbuWxDxa3BpFWyFhb1P
e/V0g8NPQsI8slaP531l22VdeF3PlWkaUy1gJ/BLddWlefjFlIQSbUUyuYcI
Iy1B4gc/HUwN/J0sCE8ZpmFmM+vh542XEwGrGejgCdKN5/r335k9ffgwIlIy
fkYPgR9++DBS9PCQHgKLgzdhwvTDEb/99ZdHHz7APQr0Uxu9KOupKbVvTWv1
qi5sqYHO35nS0gXB8fLBxzOmkvmOtNFTM3t3YKsCr0RTl2EGeoW3Ymq1W61L
C6toC2GQahdz4Mn8eWBDTqvYE1ISBZTEunZpG94f7Vqv63skBevG2dY0W72y
s6WpnF/h1W71rN6UhTQwqjduV2mjb0DSMk3h9W1jZu/CQaCjKNwQRSo4esr0
5w90wK/tzM0dMeVkg3Cus9I08qOiH4+fH9MmLW2/Q7kpAxzeeU2Hra5gCZEC
qqK2Hq+27DfckXdVfV/aYsGc3rTWCZnUa4MnMhFiujJL4KX97bmqfGtNMUp/
ViJCEE2s5m6xaWyh/Wa2pKM1dA6Wxuv5piw13RGF5NXTaWTagDePLrcMqRrs
a+OtV35btUvbupnmucrSzlBe4O/AD64q3J0rNqaM3wESelXptjGFg2NnykS2
qOf4Is+avum8bqxf15UnKlE3qrHr0s1Mi7L4nSldwf+GwTd2bRz2NqvXznq5
b2EEY61/du1SzYbEm2HKy1sOhF8u7NK0at1Yj2cW2pg7ECNgiNRZWGL6Dh91
XiNc4bWbvfPK8PvwOohxxIZgJqAMBcJuylLfN661Xi9qeBce8JfGKukaRUrQ
pVxBJNJSO1jyyjObamu9NHcWjs7M2gI5uUp6pOu2wZ2Ew7PVm3XBzOMKhA6S
bS28PEpmiUQEBF949OL019dvby+ur19f63ptWajHcwHyy9BKJ8IaLlFn57FJ
3GDeWRqzLeKGKyGwY6U++0zrcyBPeNy8CiQKlLhNheeorjIadKLUiaYzSBSH
34ERpLJXXfFxAD2RL/+IbpPMa6RgJyMbxr1cgRg0tckIcjoxRMvVMJ0gTu4S
xQkuDjKZLlXV1syWKcMZqx65jjNfm6Z1s01pmoTk00nscg74mMk49z0wBYXn
oNYra4kyPbjwvMUmE0bGJKuCAIrkdwqcNrIFoMnNyuARreDEt3XNp9TXK4tH
1Y9o9czKJgzVbKEvFDJApRtgOVURVlwewpoN0Q1aOKPn1oA0BjyFTnK8GyMc
fqRsoyHShtcM1ayy3H2+43J7u0BGqpRtDJALbatZXciI2mmh1GM4fk/w/Mn+
mhY3OJVFhAHU099AasZ3YEmMq3wkoMTagE6pukHC0tqKVAS6syLMB7rL/TUG
JYt2aeiIm7Zt3HTTWq+yF8dKpSf/8fnNE1lb1LL5OU1ByLLM4XNPo7wXBpZM
TPFA4K6ItFjj5UrFy2Sf57aahZ3FK9HUyNmm22Rv+6zC3gWe0LmvCkk7dwOb
fPX0NZP7QGkCh6jTh0zLuiKyrExmukhPXtf4AGQ9zP7OGZGUVTj0j721+vff
vZ0dUK9LV7UfPsBuJhIXqJALV8S1IUXcFSDEVpuVbdwMOPPG0k65wlYtSGfI
FWl81IRWX01tWVcLEA7oq7HnlOAaWR0QRHCJSNaI9BVfymfNBCTtx1WzcgMH
J6fvfBDjWiTr+kSJsEZzabfyes/AUwtdCbPRrgZ1QLZr0YAyDPyTxG24TgfI
t2v64ylxbDgeuRAROuR7Lz36zfQAZC3fIcfIIUyb/x5UiekW+6ZXFY2RDsCN
6MzPx8/HR8+QFB4ejY/Gf+kehBGud+PWaHcyLZhYVdqHPnw+ft5vhTP968Z6
5H3b1mpQIm0yQ2KoMsPDo68Ppq7Vf90YWn3ShrGxNhlZ2FTurxtbbjM1M+1V
CEchK8CtyUAYLl0i6gVtAeS19LYF7p5P+Wj8xfiod2e0vtw0sOMjlb3ZXZ74
taCEw8FylV6aqkBTl2h++WKRal7joeJHwW4xQDZ6p39atzndCFcB2WS8cMIP
wqUwi0VjF9SRny3tyoYmpdlqWkebX1Q0kgLPLsG6o7rCjVJlXXtgKJt1GQhx
IO19TdMnVz+h48z+hFDnIs5Hc8A707h643lr1mDtJkbV2NLeGdHvuSfqYgSC
Rb1e17wS+Qq41ttyztSJ1GimS/hi61YoY5fGt2BMCBs1CpvgVnakL15fqnXt
Hf1k29lY9fTMxy8DG8WJ72SkoWFKLaPdPGelge2jPbkrqi/QT0N6WWNnpiwT
Fe2uBnMmn0mSTvs7ALJ03ei1bWAsaF9yjZ21dbON+oRnCTt/EShnfIVUqwUK
s2Tai6cI6apX1DM8IqF1iLXLhikSyOQWodVzSCbTGizdzADnuq5seDeKddp5
xbrViHQekPxY0hNFQZQv8P5UB7RIBjTwTNi7/f5cqbDUYTwJ15nZ+abk3Rgh
/zca/g1mpxHNoUv9mELKWig5Nh3d9gfpHPWRlTWVjyYJZtqoFaDeieI4XIym
3VRAzNtaz0swyJpqq3++vrq98HSWR3pq53VjmdyLcY9Z1vBGjWEV6IR1FwFY
vxgMhFM+MOPQ6etqZvHoCr2czTaNHzYY3LuyRHEfruoafUb4WZifaiyTdFlF
HK0teAB0R7pSTLdLBSuVW1yCzSquEC6Fd+WdbTJR1sxYRZxuXFnQQqVqxuCB
ZlFVgdGr0ou6LrLzvOf0o3kNTANIKaZWF3VlYfdmjQV+bNB+3xvAQK9KNd79
zcaZkFYCpo25bfR0A+xSwys6uKKuL07PgZ3810oUoGaD6lLWMzhpB40tUY/f
J2GMNZhCiVOD/llY7xaVaS2NTUsfbQ3UqyJiNULx6ADFI/waOKNtydzWoyjF
hBQnmZ6auE6taRa2jdI3Ez4aBhE/IJosk0UKudXLuiwU7N2a7MAw0CBd3iRS
z9H4MAoyaHslwk4yjyo6Xj6bCDfB+mF01z7aF+dVl8lnKloir0y3YZptrYt6
gC30zatjrS/MbJlJeWJ4DhufOnajs1WpvaS5Rbk7F2dgXdF9QgaQ3GwwIOEA
E6FTYwsV5DewfsBSD1pYiE2zuYM1DpPa9KPyZwqzxjOcdoyaaGYOymkdSAex
i0CE9lj9ldqkuuPGA3P9CNWRlgZfr+8rL858pUhdWtaljat9/8l0gvgOmQyv
U+PZC1MtNnBC0M76zm7BtVd4/ejl25vbRyP6X/3qNf77+uJ/vr26vjiHf9/8
ePriRfiH4jdufnz99sV5/Fdsefb65cuLV+fU+NXrW509Uo9env76iNjDo9dv
bq9evzp98WgAoNBYVkIRRbBubIs2LpW5PL4/e/P//q/DY/377//AKJIPH/gP
gH58+IAkgL6Guin9CSRBmfXamgZ9OyCsmLVrTelR3PVLYA4gBYyV+tO/wcr8
+4n+x+lsfXj8T/wAJpw9lDXLHuKa9Z/0GtMiDjwa+ExYzex5Z6Xz8Z7+mv0t
6548/MfvSnB+HRx+/d0/KXDx6jO+40AmbphMnLMNSJ0GeTK3osINnS1R5Qm3
FDUgtOdv1qUt2KmIHjKkJr2fHrAiD7mTxjoj2VH3RLsOoa2AXExL8P3OzQw1
fTAa5RMYcvO2xpVeL+v74KTP3L0qU6WLAvAG9GFvGVEBndZxqVg1yqaucOqj
IRMuOL/oKjCcBzyfbC8cGohbgbIGNmWigabMzeh1qjWMUiodKSAJajYD15wo
dTjWK1MZJvSibHs72wDmBjeHHCZwDq9uUZXHVrlf2clOoxM+65CWTUwBgBoB
Hh6WLsjOQ2eK5iK+o4EJodn9jkF19RyMhiNyzle1NlNXshgfvOFqcNTk52Ij
aeDOKFiS654lGeWqRN4ot3QMbIUOtmgEJOJnWm2rO9fU1Yq3xXkVTiEOnzja
Fo+irDk7bYjEsRgXiaeixS3YhZTujVLXdmGaogSpm2UZYevpcg65YsEWFVRl
XLWR3qzRDzKz7o5E8fRjHcvS1+NjMCyp1LAEAyZnYdS5vF2BWDvDAaLCb1CL
goUAdgC9BivU857k9oR8Sx5E9WCWMJmzhuxioBOyG4AkBZpwrrtM8VjC1rE2
h4dtRrCjmRWPXTptcBMHn0K+YGHJg6V8Zd6Bu5I7jj5t51X/u91dmdqFI2Hw
rxvbbAcN92YqFixwhS2sT805IGDNo+aqB8CHunCFAuWtsX5TtnRxaGgTOJL2
r+Akb2t9efXiYnLz66uzUfcY0Ad7ph3kp4A2gPNTiPKStenNJu5X10aRLOvS
ENwBVLVujx0tYxQRA9mQV4zkQhuG31azZVNX7m9IGRUpd77lRcscxP1B8y28
JJ8MQxOZcMDbOZdVCjz8PUrXWYfhS9rBOGxAV3vMIqonlOkCn4nLwz8Jrulw
8sCPbfm8bVwB4ioIrfPe+STAqC1Ssz+r8+L0XYT2ndfJZUBvE19LLZpsIkHq
MZ8XfoJzQFsZ/MWjr8jXM5nPJ9QAjfPJMXBeBV02AFFDV9RL9HgQPbb62q7g
WL0BV1cB9/LMlKV6fP3m7AlRGQAlf/gA+j+K96YkD++gkVkumWx+TehBYR7o
0Fs6exe17e6e4qUV/pftLRgPn4K1Uha5jxcRpt+50S1JJW7m2nJLBiema5sW
keXwves3Z+kcYZdkYRk8IqO6b+pqoeK78UxZ78AowouC3Aw0eYPYpOOL6+vJ
6dnZxc0NoTHGfPZRJdjD6nvGTbrtSCzBGgd0WdVzNJC5ahG47jcA+nJosp3R
bs2ZhM9JGaBdGWv9Y31vEdJBXI/dErJrjQUblFBR2tlRbuVzwa5Bd6EhEGiN
0wdZsq52XWKwNS1Rc9m0bF3aoPWtALclw2fqe7k8wmsSJUD2l3gdKNSg09Ie
iJgpGEhiuR2imBm0wwfxyg6NuKlruKybNZEgwLMHHNXA6UNZvkZTBK6t0Be4
ZmIlB5UZjDcBmRlgrchT6xpFqEbEdpS3wPIh3gAFVgIADM0ApIXiGwoauPXd
tUDCEk7CrGZ6fmcZLwFmLBCwEQ7NMDqA1mqzAP8KGCjgLoGLhRt8o5fQ30i7
VqUALHtnERoRROccY3n69vbHyc2vN8E7kWyqiibpDIPykNaUL74CbgYft8R/
rLmzvmjq9RrIB8vlB646aJf2gIC/USNpzWqNxlgCbYVJrCzBN8gRBAx9ufW4
8uGVx3a8GKfYOK8ICXjAyF4CW0rDciujhMCZJyOwSDVbtO9ks/cjgHIgWIau
aGtnSzR4BlrSUTU/io96SyAI1WEWRGVHeKD01MHBgB9P6Tid8XF6gVD+07MX
T0Q4jPSY2E1A9DB5Cd9nbnHK2i4D09CVAsE6QF5htSI3G7haALFbNvVmsQzf
4UgfDtgJYBAnfo2kiQArei6EK1YfGT6XGwRHQgb3Sr19hC1FH5Cslhp4SPvu
mYfGSghzxwAROF93u0U1IhKAigMBRFB9OuCfQQ+Jqq/R92abqnrio9jqmW3A
pZqYZlWGc8GXopuG+eOsbmg9kJwFfzfx5g7pZfgXwGKN9jNbge82k2pEpV24
OzA6Vwp5J+u1yFO/Pz2f3Nye3l5cnQuCmJoGIBq4QWibcV02fj+IT5lBsG8M
ryDeQHwR8Zu+pZgMYB7VloaIKIAQHFQtdp1i6tYviZqApYFYq0p8KogxFejp
LtJn/Dt2AAMTFa/SKat/69qJHWSIBbMdq8OJ2c2lwuOqliE9Bt4iW47wF75S
riXMSGGrbXqt2bnQYZYB+vmZ1le5lQ3UHXI432Qy/lMQ6hXT5JL1YmLwQRTr
aAVBFUjkX8DwhWtcPaSBwL0F8MwMQ2+qjkGNLGAC4XFIKLUpVq5yEEnX1g1r
WI2lTrRR5CUC7ginH0cZcMm8OC+gzb3FlufBoU70V70J2KcX56dvnujPN674
XM+dLQsS3CGg78MHRjMPDQZcfMAh5ZNs0gTXa0GHhuhlZUn5DjgRQkWJRHaL
1ub7XUj0pSnYHGF9m+gMrH5po24ubk9vb6/5a8iyI0ijL8vnkhQKisSu+ujx
274GFCBVdHaYFasGorIOmnrqKj03foloClFZw1bFHSKz3e0SFgz2Lh87eOmB
2+3U5lg5F+vk/bLmHljXgF9Xwi87qslOZ9jPOVQCQyPRJDATFGOAtQiF64IF
gFIyHKCqMz0gojHTjkYsJsLBtq3KtiWqnPHUZ3c/7hQbQAf984adgonyEqy6
8HGx8gBvSW+9TvTF4L2nccCmFz6zqJBJ06jslRzX30YAN30UTkScLWChrfGt
giWHz8fbjMuJl2gBopZYBoJK8xagVXDvRCJOhE3SA3fRs3qun2nyoeBNvqtB
/Q8WwXa7ZvmS6TvdXNJd4mYQLwetqasZATsBlAi4kUgZAx5cV9sViP+O4dt0
Ghrforw0UiBHiI0qM6HgHhYZf6f76FuxXBMtUolhQQx3rNvbWQ0xRPAhvdqU
rVvTdYw3O4wWz2M6XC+s5uK9AQIO60c7v5PF4O2l2G7GQAjp0o9qiEcczx6J
3WFHiASa3GLbXD56wHjBEk33VtiqhQgvswW+9U6X7h04Lv7jP/5D+6UtS3XQ
3B80B/B/EKF+KBMAaWw+x0df/uUrfW5nWh/rw8OT54eaJwOdqN9P9GdztziA
zx7gDxqTEnz76CUP5HOv7xwo9nNp+OiDUq+HpjCKZIdd28aTW5xUmU21bmzh
Zmhjze/wU768KEvQkRia5oFM8/Avx18caa2Pvj4+/Hp4mjD4ydBc4Yd8ruc7
5/lzhloHZ//aVuJMF6hPVQQA9yja4lCebZutYutIgO70LJOA2XCIPYqCINHn
qgas9yKxdWPXjzGuXoLfZqaxT5BEhkcrBP0Cq2aEP6t8fDqAWJpmSoiW7O7d
5sZmUhY8zVKh1XTOa8+zIvPoGoKPEUdChncXbGkejfDEXxBBoGT6GyfoOsfh
NyD5g2uNPkNby5/BKKxARgWyK6Sne2kKO6MYBGEmGRcjCVwxaRdzcMJEsuA9
b2ncYAwl0roy7WyprGlKJ5PyMUwGD0pV82uI6ghnXtiJADIAwiP9sIQ6fujU
P+dT/5dPPPVEkfNzf4nP9O7jf5tAa8J1XUQlryu0iXIThS6VLSax1arWnLYD
RbIqeI8gLQfuIQrighcacm64sR2PMHqKhQ/+MkbQQf92NbUNWhGHMGt4uEH8
JXx9PWcnydXrl6/PL44nAE+LHpSeg4c/Ruob6LOZOEKOG7qIuHg79OHXYnfM
ljYKmKht54ouWiE2Xvu6tKiY5vhesgPCuuIpjLcSfLe7oAvkxiVTKNOH1OeB
VwZomfIQBG0qlFKHZXZUfE0iwiK5ePbllyeHz778Cj+k6ENEfpJbHT6jH/yM
2vWZr8Jnlnxp5WDBtnSII5jAIb4FZG8X9qunUpIgBY4WOAwL24r1QwU0589L
XOyNX4rAzPZYAQHEfdCzpSXDDeZDSW1hBkyDZJa+t+adrRKfeCpFma3m3CIU
9RVOgMpPQLL54mlNEescThrDmtgchsZFzHNRcGxrx/WkTXlvtr0oOcBx0czu
l7Vynj62Jd950ERWBl8ht3BN3Ks2pfUzG2GOoI2pIPAB3wILC+Avq/AW+x4R
uUk4DTaIvYQj7fXvn/WNYUr1AuwZAIsTQkDKuqy3Enw4fFuUhHeOALhARruq
RtMax6r0fVJVQeZNim2rVJoHgckFKVcFB8XewXcJ3+fW0BKvLxsBVTYgyOSj
g681GDG6CBPKtRATEDyLGEnM1qC1rEOL8iuuQWcFxvipXQkMGHxAu8qkEe6J
1vril7Mff7g6P55cvjj9YfL25mLy5tXlzeT8Rs9LsyA9v8a3Tl/9cDG5OsfD
ycsG8o/WgYVsEyzdnqFqfSlKPe8/dAIsLtrHuCvuhN2kaL+G3TQ+LJUo4Fon
kTAfsxQMlEQyv/K2vAPEJFsH/76FgVOyrluie+UWOpn2zMNEArfMSgCTsy+N
RVtDLzmcaxje5HNrdoJ3GSOGrsMNIkem7ZJMFs7P4CaACNRoU9zZpnXeFqom
cPcAyrYNOKMdIDtR917wMTjjHzOywOJ8Fx2eoI8Jj8MhcMkZWTduZRpXblVh
jWBFhh3bIwmlpFgbxuwlwCkV0GBsyCGv8d1xgHI5H0Kgg57AarKsyGgwvwgw
B8Ld0CfCFANtzeMjE7hhdEAEXsjAooKB2pC+R6xJBJmhCDGJcmuspigE8qfu
Nl8PmLSU+gGt/ugVmRrvZjpk0Rk9tGEYgEdo2nmNub6IJr5+c/GK3nhw09Ig
AKaGwWq4A79AhnhfzxyOKdj8mVFenRO1wOsSHCb42+fooa3sLIXW9Ad0SyRi
aI8FyShOh51+M7ra5IqMaizdRtJkyQUM1v7OIs831YySgbgWyQx+VGzDu0wY
vOyT89c/v/rh+vT8Atfp7MXrmwvxpVCWNRhTqe17O9tI3oZ2cMLU6Wlx5zzY
xLtxF5JsK03Bg31D5O7OPhGZ43LPHx4XSe4iZyFsMPTCB6gzAIzfFjvHnsMC
04de0vUbnOt5DCT5qM5JgACxAcgNfIKiyThgBLLSlI48U/FKsK8UecWruith
Y9bEu7q8s5J8R4gBDfH24uZWXHE4ucvri4v4IO3/4d0gzCqFzGanNxkC8Jey
3B8sEpzFDQUzJYmngmlQ6BXer40fDbUnlF0e5C72FwW3JpqC4QTQX11xGPxE
JPwmYRpXT1+rMHw6hIZpKZGslxCd0Q6cc8ydQyANUK43GKeH93FmOLcLEq3B
mYA0A2l8KBPhxpGiAmuJy4Rgi6Up5+DW1zrITGzyZudI7XlVp2zOYRnoItd1
/JIxrba5413a6zrXOgYsXhALQOcijE7cGgP7KFd0lHdCdqQUd+tY1+dYPxFo
Big39BJ2lzMigS53F4C1JjTHC4EBgruQW6p7INjpjhuZhgiOAMsFh73Nt0tN
bVSVOlGboIXdmwxaLAOrNHJ6OLRL49XUYsYeHCrcIZ92Q54KAjlbn0xPZBcJ
u8uNLyo15HBX4BDZhcqDi1IhCh2BzjA3et2+X7uG46fRk4D0iiUd5xGzQZJR
mjyTcaYKjQUoXWJvfhdNRYcCSADBNY/vq/Tru4QpYbRwxcSFHAEUSoJEA7/A
u0Ib19gSEp8hjhxXURDyFW0dn/+5kbB2CuyD+XQ2DQXbW5Z6/y8RbG+zuIcQ
NL0yhUiJw6Yduuo9+R8iXErXOrFGSBRLPedBwVyqupWAxoEMcv8HZcywqLuP
pEibqHc39j8pbUIvV+f/O6TNkGP0v0DUFIAkaihDdOKeRW9WUmS+u9CABzTG
gxnwpiqS9qoQLpY4aCLfFrbQWHFRjIeEVuiD5FbYHxJU8Vg50rx3uQJRng4k
GVkTGf3x7MtFgEtjFhJYvEuU/v+B1Cu2peAjB164ibyiP8bvWaTJZT25KOJS
TzxTKKvEXG9DI8FjLKciiYzaexQHziGZWz7lKOoeUE7kjMiziaHnt3u8Xx4M
JyCKgsg8Q7aYgPdAVXZfZku20OS0XPrFM0PqQHZszK6TOWKfWed8yNrvOCIP
n484jyzsbBUWCE9BZnjPhJFou9tzQh46HrGTnSfkI44HCeqdE8I4l0CmYgSc
fkzOrqBrjPAUhnWPbx5gUqxhApfcKj6AKXv7mFuV5+sSi0DIrJoRNIg6GMjB
RbwaLjZCONh4Sdh4sFIl+ejEiNTlMkkCICFzMFK0cS04qfl/gwr9ycsbs0Xk
pGuICbPunR3PEwVvHo7130PGQqoqTf99AikDrk7CkDBJjZcizIYS3qWnLx3p
tLHmHcOAq3RIA3YF53daEXaaEAaU4NyKcKBzW0X2kYdY+gBRHdobjrml+8C6
bwBFir7awU5382x9HxX7DCC9riEc20udgci0Sb/tEL2E4jU2CRp+WGsLtl5I
ZJUq7V3VjfyqWOaAPa9TmGmIXAH07cDRdJwBSi6yD4IFIac2TWv9Ng1ZcvMM
4sIW6T2qQjfgW5Mr35DhKIl9TGjdgylfe2KahD3i4mIunYd10pAkHZIAQlot
wAjFJd2lsuoHVFYVoAOUZj/qrvtWCTkpYjHgfKqA8o85kDrI/gFYe24qEbhr
dCSQKeRuB+/doWmrJFRhn6ZtcKaiaA96Fz4Ts8HBDRyEN7GEAOjPnpzJ6Rsv
JYmsUue1vuJyCkvnv8POoMLAuWSdi8lzBov23EJNAJVnYciSuF68hwSNpqSy
INdCN+l2Pf7l/PqJYpj3sy9DMojsuzvz9fON/OX8WqXNnNeAj8FqI/1M65VE
imBeZCz3Aagp1DVUzPgMhULQMUeVFshFbzSVPWnrg1m9WjsS6FYhqTw2mplK
zXG/Ox8WVQjAT5qGS35GADDxdz584CtdbGaUtGwFoaaVPYDeUSiE3fmkVUox
V5/9w9Opq576pVo0dq0//3/0n54+ffq5/h9/0n9oYBKf++/omf7uu8+7z/7H
d999nsGuZHUYccV/jv2S4FUcRMA4NjOFWhgycY/nWELIcNDAA22h0244hqWT
K7/fBGxv4/Z9+yiJDE/2o9nw3aaPG99b+ANLUFp09sb18ksdR6P/Uct39D/h
gk8Qcnw0Acozfj+0NNKvLJEgdrlWT3fFrLbzOZkR0wELP8FqIqU1yFanpane
aSzgJGGBkG7bVXak1+WGMnSDTdYC7OivGwae60dPnz59xMircFHgVOGKLmnV
2A4ExpzNlFq33aPnVTSX9D6EdgXpnYNqBOaZWK3SlG10trHADZOQoRtc2LWt
mFxDplEV4hukkkY198e8JTQnCgD/EitocOxpSMiLOmU194rqkHBuj0LQXscj
lufAHDf3tj0e6dJWi3Z5LPkCsboLsEevVgBEQhOim3FhE2m9cVX7/GhC9hr4
48vjSUvEm1II6FNKRyMlj84ZSLBVCkln7hnqIXHE9dhmeSokZExVXLGMc94M
pxJmQhaAHz9c3J5f/HR1dnH16jLx7ejHaXKQZ92ss8HDP93mo1GNbRsImxeN
JOhbDHiazyc0kAkM85hxTp2nfE+673IWEDZ5SlZpxIZ22tMI09x7CGBkKYi4
FtFPVDkPQqxgvTZ/3ViOTSoMdjeZ1sWW7QcqG1AcC7kZQm2DfFnjvjLheV80
IEIDCaYukqkyhsIf69/lHfkvnC/+bz4v7uT9bz7q5ZWr6uaTWmBmxI979X7w
1Wldlzr5D19lCMyEITCh0Qf5V478HVgcprUDPwGVfWiR+ex1x0qYPdMusSzX
sYzXTCoQ+IrG92Y3NDRsIX/+4z999ORoUL2J4eNH4Uoko+FjKnexrVPr9UDo
ppK8LgDLyrNUI7nBOmMS3I+O1UYjuZN4NXJxdOiM9eNkbDJvzEC4TUvX9VSI
aEuYLWvKiIN0D8CF3M1I44mNf2JRHsmK5NsGyXQeo8YULr0cEtOWX4Co4Qtr
USleK4kiTEJ8Bqipk+pZKs99nef1lMxJmL0yGZvXhGo8xsADlQ0bM/h4/TyR
eNIMPbCN/Vm1tXoWJ9y5aRhRYErfCZXA7trEeTmQeiKrgnX3PKpiSf0rCp5I
56B4DsefOAc6f+RiJBdqdhIgqeMUQ/5zAyR12t8vlU2C6/0FySaUgMq+kYku
94iRbrvRYKFsXYqCrSgJBgcGKDKgh8pKs3o1dVWMjk9Xa9RfC5yhGtpLjmR1
LcXqxRvKBQNFZupETooxxQVffYrw3QV7w5B6Mn+KAQRl1gD5q0BdMi3oJxka
L83/pO4NlP/aVC3kEECQLQoLKUPujzuWNsHVVUb8zZDMfFZC/Mk2A6r2DWgR
U5BkGuChH1AMPNWfCJQD+V68Rvf0ZxNpQlJnh0Zr3rvVZqViQ2qTQF2GBhdz
B3ZOKmWbwXiiWDSPOociKfiPUPikF87X7p8HhjAME2WVTywX4jKozbDXb5Dq
UBcZdsOwQZ0CvOpGohd6sSqgGHUQsztcuBInSxlZeqGdWVJ/DoEVQRgPF4Yw
cHinCs6iPPQnLx8AR4u0uYR69yYfImkSGpiSK1hxx5FoXg3nVwtBd6jtcDwE
RJoMJPfjcNIhHzmr660U8PIh1xsNIp5TFTUMTncnIwJUi6DDKL1tL/Xct9+q
kHQO9haSbwQ8GY/SzCHCb0cH/5B2kCUrhOIhXwV9hNjOkxBNkSZs0y9FnEMz
2O0+C4pcUx9lQIbKSBxHlx1G6Sd894B+OqBsQllHLKOh9WlqquLeFTDhGVlI
MbdLs6neiWIIry3dApNQkGMxmK+hNq9LzZWUb76TJsq4ctMAcUnXIJHG0t29
dxDnhxHTUuVjaKZiTttSohyDicp3NJEIMvseFjUkSYPIBFHi3NoqyNMfwZNL
c8drjqaB2XDpwX7eEKiXydvXHUe6B6OO8BxMmMBSUFit55nU2+mrL/ZKjito
rB5TMmg4Sx014kniShIpu1/DWT3GYMCQmZqha52vPhlF4pf6NXA/MEZ5KMN/
TK2MfBYGEMKTgIv6XpDeHEpGszU7jW/srgF6O8A027JIQ2kveumtAoCfJEQc
KXemks4AVw1HDfPTIdGBZczTMJB3FQGceyIcY7bmpAAKkO7+FFylspNBQ+yM
o98MPUaJd0bBpatFwMrSvnHKF5BqMUU4mkrj933aC6bTQOsnVKQ0jAVgynJ1
DoZp/kPupYSiYs70fiN9dR4CNWSH5w7Ma1UosGKD1DUs9YM7PNmmmFukyi8V
5lXjpacwyI4lBKhM29rVuiVrJw6WiBnXGVcfMVXeoiHyAAL5rPUq1LeUjaOD
FG4i3a+cIqB4V4XZDYeo5LM7+37y6vXt1eWvE57kuX6ci7xLtEQXXJAR2Dym
CQl7A7JBEtSIGfcRb1WIvn117p9Q0psQ7/0xayRacdytlHJdpSo9DIMK0QQX
K28N45N3LAJhDHCZJalioE6xZ9LQNgA8bGZd66WkXUySMiZNw1zunFEk4/xw
cYsvZecKsJGUVLzcks5xXw8Rq0AHO6cWE9Dj1EP1jMixegJ7yKnQcwnHmtEY
CibWamRZzM87Bg3oBgbbkf7jmCmkmNkbvN2vkxkidiCJOzjXkQ6LPPGYdCvM
p5BnLnxCAXFzMcJn4XC3+cB6Yg8PbGmadKm4egNljmzUyvxWNziwear5E0Jd
ciliilAK3awsLB+UhBZxt72vd0pelG/jgf3CE6nCapDHdUOgF6gMz86bUWQw
IVtiyIc7Cnlz+QVMpEbOHZBTuHawJ1v/XicqiBdu4cBbSkceJIZjcB2SQzBu
BDkztEQiOzZFeSpFDaZD6OlAMuzmpmX4z1abVfaR31Vih6XbdDx5dXnz0/Hk
cALi9o3W+lt9OBp67/XN+dHk9ff/cnF2C6/pb/XR4Hvfv3h99q+Tn16/ePvy
gt57Pvje5YuLX8JH8b1jee3DN0r+yZZbmsWEy4B2ZpJOMf2vrGcTePpN+jIb
+Hv/wctg52d77b5hdD7PPqOBHif0U/Z9diwNvU0/fdOfGiV5OM7fpocDb8dV
St7mZ+ncUiN0cpLEAD1wVFG6JUt07q0jO2fwhvSwoIl5B7dJUW6D/lEIZSOo
ZGFeIke2KH5IURIMcCmi4P2IdvcRGzd5r6UMAxie0uIlrOeKLw/z2QuuyGwh
aH+6YV90t9GeihM5eRYdwbP2mtmEg/cIjYSQmwek2nvnw9BxnUBx4UTax9GL
Fs7iZ3mq7Z7zA6hVqy8vMUT7ZvLq9SQzFsDNe/b+Gf13+M3uVlevJ7c/Xr+d
vDzn+x9aHe1pBYlIJlev5SgmrY53tUKTwOT1q4vJyyus8522+jo4UbgxLHph
5z2XlCbvDISh++Nuo6d/Co6cP+lXlx6C5EfRFhSzDBxxgoUyeR+tclS9K4Hz
JOaZcXz5ac8FdYSAgglYxDoeKO6rRxvm87mb8I+ZA6qa+8l8gJZgg/kSjcoD
3qehNbhEA9z9SFJYl6XkUhZIpOcACW4wtVm+AVxmAioMzV52+GhyfsO7fHp2
e/XTxcC56J/BpNXNm9Pr0Gj/GUxaXV+8Ob263nsG9x+jo0nhdxylZGMxdRGJ
OJ2tJVGkv7eYZ1F+zPZ2aByhhZ3P3czZarbN2nQOV94m/JB4JPFHKJp9PCGR
rfshyF2y821OEBnfxr97Iwor1+0cn+4+n8nCUhG/zpr2Vjz0vUqf++5854Tl
oaONAkIcUvxhz715qn+Z/HqiXwEA7Jc//4pmbLClJi+cGQhm81b/ShkLk59u
av0rXJQmPkVBDedZ/2bb1k7mZlO224mYfnt+ajjcL1//y8Xt7cXk8vTti1tR
P28mR5PD5Jgf9hzWe5oeZ02PPq3pUdL0+ac0/Tr76vGnNU2/+sWnNX2eNP3y
05oeJ02/2ndQkm1/xfBQEgCAuP7GxDWgsaio5FNkf70+5MGmwnyv+SEm1Zzs
uI+HDvh8XUwotiI7S+jfgKm+uX59e3EGRcgmt7++uZCpnzyE++j+B9+RQ/xx
yJF9PQwDSvZclF4P67AM4oA1bRf6snsRUPq4OO+tAqQu/Si6lekL0sMuDWA+
LydkF5+AXbxHtIQIdtvQ8z6Z61NdaZET3n27QmMyrZ/M6hICPCdLV/WH1j2K
0rTzfOea7cDF8PLdHaU6CUjenAY2Fbvvjjq4nwTz02PL+pNY8v4F6rNjvVeW
K3wmy+m9chxwSZLjEr304zh298093BoU3XzJMobLx2mI1+5itv0ee6rzp1wD
HsHQDRi4Avqjj7/++45+R3lOZvcR5/QwQrlCs12YRv4d4IxRFQ5pWLlMnUIX
W1IAN6TpzXJdodkVa6O1nVyShA0Db51UiOTiPFBCxVScW1nSaKYFlWE4+NU9
oxB76UoCEhjdTW7zEWVarcloL2Vbyprgq1w1vjNeH1IWguXQvpd6SVwmNMC+
0wXkVJuy4JZUdBUQouIaqxfNRCoZiE8w2J2PwbSR1bgDcOzz8VEPHhvrXQ4U
cP8RK9AoHFByCDD0Zm9DMpeS24+Tdas2Se07qyvwMlSU/ggcOiuqT4fhiAPF
tC8QTZ4nFn15+msCjIvQE7yTmJfRNGYFsKeI+8tubEQlspHDrVFdbqkyrqso
RIb0ZbXzSELPBPIiZyl0g74yKq/hKg9hLoiH56L3bJqJjXiXPRn8c0IVXSD0
vso/UldpAToyVxFyL58roicKi3IIKMh/s00t1icIAI3oH/HS+dnSrnC7QAnw
VkqtUEYIPIqhEgqtOMLqMhA9J3fH+MPNas1ipKRTkRBq1Vt7wcr11zWz4ptZ
AxcQrB5MWZOd5if5Loe7EqjDUEJBlcSxhRfEcCIHgUb0jY6EEA5cWdryAM45
lr5Ca3c8el0izWNL6jvh0yxfKSCZlvW9qudtjkok95wT/yR4XYgAQMDlDcMP
kmTT4D9XRNIo4bsnS3xqmNV/Poj//Tn75Y/035/wC/oWPrHN8C80tj8PjC17
O53CQf6H6n2h28Ofk09kPyUf/7P6g6PW9GGns/DD0Uf39XFDkhn0u5Gp/bHr
vT867+16sfOD+vMff0SgUjrCzg9K//kPiRvx+VTyH1RnyLvnkgot/UuVCi/X
kEQG/NFLtw4RpWD4xaiVG+ujEOM7FCHFulSRMFBwagolhLha5IGpRBHrsO7i
C8CzhLKDxzUKOYiXScREjozLBPpIFKKPmrK1D/hzBQqPcR9JBZBULhdCLPOM
4eUDRR8xo7r0GP3yCeA9hKgSlxuK3WmsYPYUeh9LzPS6ApROzvCSQVLIC1Vt
C3gX+WR3USXcHqcUw7w5Jilmg1MJkvk/BWTWoRhorE+ZxCplkQRzdHEPlDpn
fBoa56PUkm4HdB5+UV0Y9idFHADmbyeOmNB7ALCtOe2S9EbSL1Qgz8K0GSnS
LwqIpQJgSbo13RItMj/R8jCJX5huVQ656BcsgW90gK2GcpijF2SoMm8W0y8p
cgkt1k9yB5LR2jZw9ykrzo4w7J+59Hc32nsYhCKdx7SCgs0O9cI7aYBVyBcf
Uyd2YhEYOB2XNxQGVL1pjdkTmod9SLBu2onU6uFYxBXl+IazTG6m92EtQqK4
E6XylNBKnehsj2Md+qE1VypHTe9r3l0mdZqIiHxgkUWgUmH0ynkqDpGe4pAI
TmVTdxFigpkRRLeE7iiookNOVfo6U9a+8oO54oFqYSaipCpNFy2CYhzFTULs
Aeb3TZIXJalPAoVUFpHIzi8xb0eeRH1HfcOkij3mnBCseUcHT/zmHHUacmHL
QnMFiJZSHwrYpUABf3gd5hu0HSRmBt7ZsJSEhgs4XkhWQV8bSSkIBxdoBrBM
twKh11A4p9ykABftOTxRY9hAWahyK4UbmGw9WCp7Z9GxBAyZVjN4sEJ0t/yY
aMh5eeRdeVd3VIn240iEHypY2q1LOngkBWMXFa5deTa+ESQ1Zg8Z5J1sMRqK
oqUo2KG4A4oEdIsK7ANIknZtBStakgAoKRaztxnlBglxVcarmHM2MUeOEIAK
xhQoGDCCfRTTyhfjv3TMKrQNFHwmwUaIfh51lx2ILSIgOdkSo7UUlNVZmRKI
G0I+eDb3zHYk6h6U+JCfjKt50mRMqQIYhBMRbUPNUeTfZF3gL1OVwoQhwY1X
MTMZpZTA7KWURART+UuRUZwUVPRJ5gQrGSTsDFdPS5CU9+tAXZmZP6fS8Q2W
rBbEfa8UarZAmfXBcRky9a+2meI6wVnIK31Q8hYpNzlQiCjk8VFQMOefi3oF
SXRohiJfR6N6YhQbuCKfe22hW85zHGCvZF+EzeNY+ETX6NhTHErxUGguKWIK
FHiNgGICHYaPw7rAXjEct9jQDtsG9xEYhilGKgvbGA6qo+QCWv9IwFBem8JW
EMLEaNHY7abFvvjUUHqTQNjgrFIGTnwKCYzEKhaJGvCeNLsoW5mw2QG9jYaV
YCOnCoJT167MWgugek84L2l3aXgJ1I6J2aPBOuISWR1WkZJHJ9ZK4BtQNDSX
eYc1NJSPdwCbUN6RHINUhiIGOfbTEXfrBQC+W3WiqobmDSMYxEh9yveZB1OU
WFV0M0fuyXnK8VxprJscCewyxKeqxHZ4j4KYwKOxmgXmVx1KOpEK+yoZJxT1
I8L2nk5iWlh3xzr1YWH/mUWCLkJYierADPHGc3GsTmWsnzHfk2+tKZJqatKT
Cnb3nR3Ad3cFqabT7OLYPmWyaDwO6Qk3a0xnRSHRFJXANpeE+yMGhe+xl5g4
KCCJZOAM0kTRsIM/Q6kUbBn0yNK1bWn1YuMK9P4QTV1yxGhSyKumzHO2kIJW
EaGP8yHvCkrxoRQ2+ZiUVMSmFrfXv744vb24Hun8+dtXpz+dXr04/f7FRQDl
4c/nFy9Of6VaT+LaAoDrezZeJFL2iVI7+xSVKARDC/mPERNsMOdk7hCvUHcJ
gQrXMeGSHU8N51SjAsP8FY7hNBxFoZjpUFzkuDtsWaJ8zBhHhZJ8mp69G0xN
mHeuAxYcH0KjpRgXV5RLc6hyUe3klsAyu2rDldIgWeRwuGoyftyr/+ZhU52u
ZOgKydWuoX8Ws6npK6mLDmwLZ70DFKvURUKFu+wxs9Ls6gG4bmZdQb6OEcmp
+WRvDmysuyS1TzgbIbg/JC0G/CxkJyhCYhxiKtQztawbO0exFhgjmI4l6F3e
g2EFmshP1OuRXtXgznXvQA3EIgI0DaQrjTUezI8Eac55985g3ZA+cYMJFzeS
nMdLfkRIgcI+VSabplkQUL2ts69wnF+9IYyU86oPSB+xW5i/QyZuVJ44Nh90
BcYy9F3DRxD7k7uGebbZqRIritj0sRwcsTFeiW60cmJUycpTqmbIcJ9FIqOe
llhFuSQ2CGGuUgwBeDbShyMQxV5l72aB2IfPx897s9P6Bu42SS5QF4/z7Tly
Mc7KjW+xcFamKHWiiJIUrXXD9Rahm2CECIE5GQjB5sWJk1CgWHhcSkDvigMf
xUQf0cnKtn2VLEVSgcBVlK4sMZ1nC55FHF3NB4MKQ+nwnr1lx1RVWKih+aay
W5WQwW7vKrheBOrhGSjBDD0xpOe2BSY2CZnCT+XpUKQyCOqEaKtLTKNs8FNY
ekA+I7lJsJK3LMHAHnecxl0MQ2DgfNt+hE2DXfiJrXQXJD1R8rGO/X5PhfeM
iA+lO+IkYd2cTnttNL3ylCwxRNG9kySDSy2yVteL/aetCzHbaG7ASh7/ZYln
0vwurP4ilXQrl1hFyXkOpgJaWFicPBg4JyeYZy6PQszWRkmoIK8PWZe2nJM7
ZHCL2WnYL5DkMyE5FUMzY3qhxPv0YMqozGs01votRkfP5xNX2wY2OmwyPUAT
HCNVXKhkTWzm+uL27fWr/iIc99BEfNip2QWGxOSs5ovxl6HRV19j2CAMnwaT
NPvwoScf0xKnF9n5uhT+vm7qaWnBYTsbrvZOcj3mc6RsyFA0E25nNkcuTZ1O
wPkgIL68evX6evLTxfXN5OXVzcvT27Mfyc5d7K4GHlwQCm0UFGcqW/fYjhfj
USpDgbBA1i8qFsd5X55gjGqWV2p3WwjkzVqPj3jxaBESoQrLaANQBAqoLym5
SP8Kgvvn4SPXlQcHj/SgN+02lyMh2RKp6t2QfAVHcIfciuU6gzst8oNkJigo
uFbBq3OYMlTqpJqLItY/bBzA7C2MMIN0KDcEjnrJqCkph8uIKAEAUvKRzPEh
odJoni5sxQgrBluFXAlzVdYLN8PSyxCkSdx3ufXZs3tSMNl9D5duRJ4nDhpW
iTDUu7ej/aIG1dMLeWdMZ4gAcsSkONlTLwJpOvbuMF88oTLHnvIAdWa10/t6
moPQ0qn1Jsf1FEhwRNK+rEFCZOgYwvKGcsKEK52UQOhAv1Cges8gqhcnO+Yq
ooBSP59IQ0yqg8ifn/W3H4kDVOrmJHkVqlJw8mHOUQPd3ehv9c/6T130n1Kv
wrcZRQhvv9Lf6hf6qb7JADcBP8kwmxtqJwf8pWkxP/BnWl9bTAqLnAVUKlFE
n75muYVIbxqvCpZUW83qDagVmvktZoej2hSAEUFxQcAn/cRZnfqp7a6Cd127
B9F9HHHUAGGsNIo0XUJfvVIdpQak5IzgoarFMLy8Y7HzS4K6Nmc5mBwupiyM
LDrHWqs+w2bO1s+db8rt31j2I4pPuSpYSpMlIxOy4iXZpjl9JU9wEzd4ZQtn
JF2TB3YEP1Y0SHzB0aal3sjB8fFKUUIOkP42LYj06M4RmygVNqnvgVrb96CM
uTtOBBY8n2ITIVyolIKTesdcAUvqxxH9FQcYbv/StbljSkQEtv0HJcxTfjVP
aVbSWYBzr7SYv58XkW1lS54P+Wi2aeIQsUMZdPugyws6ABeNx9qbXJYxlbXz
RBpLdvHUa8ljFEtgGJX7x6y4VlqqOlQK+COrwxJsh0leZTwzTb3G8W8qWHsA
oMJOZgefNbaYLEdF/h3C8jM/ATj5WMhkWzalW4uLROlBqiLlysBrG7tA/0xI
Co0m7pqTs/XrlXbNQ7mswblhsjQ/nBM6HQzaO+Oc8qnEi9tNbOYa38ZqtLva
Y0K1obKW83h9wbsHjmHO2pEUTNifKW7JMBbM8U51xV2lKZGRoNpjsEOqtYbk
IW1Dic8RsTiEfEfbg6rqFOBNSQ1BjU5xkuzlS3LtI/pBas0MVC7x3cQnoApJ
vjR0utnVuqy3oOusG84mA9A5W2ziQWchAqaUMg88nlB9p13ag3vXWNU2pvJz
4q5hJTgdBWcflLQ8MnI2pY+CaxauH69aTJsvBQch+0OT15xwFWuFGb5UpevW
sRjI5WcrSgh1AL0jJA3dBUbfxTJIQMqKqAq7kOMdJA/VjasR1wOmdGZkUe7C
/rmTklaBFBBldcG8g+MG0aHccUhBlVglaDWghcitPRMwqoYEGCGgJURDoaQE
wNM7Md4IFaBSsLFaTeFmiACOanD3aj/mQpgQBsPNQDFNAIlPsQotA6swyaLv
YHY6gUKiOqt0Eb1xXfAO7x/ai/nShu3odBlq+gKJobkegH4fr7zgGMnoyVWE
ukae7rlYmXdYjB7kEkr3IwnB0BQGlRdrzEPRsYb1RteqUObMIq5ZSCFnS0Sb
d5fowIiQ00PRX1cSq9/hDAa/fQbYRjAHAzDZ/hW+FXGWg+IKG+gZprPecmbg
vABPRLE4ru2STiVzp/Q8rgR/IrmfQQ3ZTmDRHrSuZi7WhLJ6tc/FGoYQhwvv
cBrLgQ1SOzhTgtDat1LgJ8wPQKAGpzBVW0Vgx6rwB7KdGEUExrpGFIDkl1E6
UDybVHKll1ubEaNQUUflh3THrCQtK5JqkkY4kIP1+Qx7ssfwik7AJCYOz6aU
HHQUtFbarHQTlsUA2jeHNL+uZVSST4moVHcVI3XAADQNFZRuBwhVnG0Xp7SC
YiiBMiPrNeWibly7XOHKzUwJSBg58hRWRkkv5+0Ov3IXcDpCoG0dMGbBLdu4
6UYSeBuwxtC3ODk3ZsD0a2t3BDuMoplgZVd1s03iEyT7Jr/IYhbA4SlwJFiq
QtAGQKAQ3aRUEA3uydMT/LHTUNmukFBWnQ0a+H7EPyVwFQpK7SPodxQIHXdy
m04tWhO9azcUKRnoL9somEezVG2osJ5a4zmKfrudhYTrO2Zw0VoW1jIUT8CO
5mbWCQVo7+uI+Ry6CopENFhAPOFmisqQRjGLoCObFgYzvDxdo8yUsc+xeDuR
25ntBKJsppVtqQA7qLGpmSB1HgPCEhwkxMt7ELQ5KrgmaCBJHB1rzsTGYEsS
QtphxiqwO6piGVCOtqzvmeb8HFu/dKm3J6fh7Ou/IVB6J3/0WyS6XjpQ6iF+
I951nUHoBIArYtoebA+9tofDk+xO2cJAex4NHUDkGLZCYSLkQc3dh7wzmPEU
B8Dld9Fy0OOBaJqh1yIOo/NRrEEQkD7YWZSnej0Sv1MSQRuRg10oQlR2SRlF
CA1/JXdCIgPt9tZlJD1mbhNeHjv73HeNX3jZGawM28agQMpxXUodK5V2TRnV
PReksqmtjpHBO1YbyNjKbBXGSXd4DGbsTEIKBkGctPyBpjmvqMJKB9oi5/u0
LP+zZzxNcK92i7nhRMCJHMwkYHt3PAFUSeh3LtCeSkzJnGzXfKOcV37DC9dS
3ROSLTWomY1NqjKRUAyeCzCSVJaEUeeV3I1Y2QksqaKZAUOoK5sEGvPBTPJS
J4KmzzGyfQvjVIqnDZxu8USx/pTZOjuutRyJO1i5sxVtJywaiw0DZYPmaQSF
EAIh4HJkvW0cJd8VyIHkSW7B1iV5CsB4QwQiWKvZqBhsvKYnCAVhsidrDw4F
jDGspHdHg/VHWMkYqvDMFj7t67Sci4wNhG+2Paejj0I1L6ncsoA6+BkNrBd9
LtRHHPDc6Isg8JJxNrM3DqkOQ1Uz9tkeqBZtgJ1Q2UHFZ89UouNFi37A1GVm
p2W0ZnPKCUpCH6URvg+EdO9OE8yCXrJAo3FQyXcGrXdcRcQFIxCIG96Fuqkq
2JUyWjJQ6wjywqNtLYXecmMVcAH2zkFEHbNgVHQp7o1SyYek+PdABoiKkCDF
hIOJ1z3Y88RvSzU2uMR4xg21C/4ORJ/TR0JPWQuyWATgKlIgrsPDY5c6uQwe
GCj0ElxDnSJGHY2Pt5kA8ulWQ+EEoVSZQdENVKQAzHlMxd606AuQMzMgwmcX
gS3pe2+BCgp07zNZTIwB9wyptrCemcwV6w9nBD+KPumaryxEgDi/imgkku9i
4ggCAhhIcyPzGrinmBa94HwTxpHFCs2k7E1N9HzVIYLBfcyIBCmShLwiZQom
hDOIgyyCzcCU53esa+RK0aVl53P4JrLe9Bt1cEHTrOBK19XB3LSmZJ6mTkGV
kGKj2YCkDIhJnHUdTEXCULMNwmT12+hv6V4WrU+rrep9Lf9UIr2Ga9BVvgIb
p50aICxTpBuQ3jFGs+fMWbWNWywwz9LZ9xMZ1Nnpixe7MBiJj9GxKKGoLE/T
8kgyhWqY/5DEmjMgEVGZA3GhH64MhMg8kG92nAzW7Fj6UpwEHesfSLggp+Th
Srg7ZconrE6qdqfXmcLkMDaNwUZ5gXp2pCkuL2OmIIuw2xBGtk1sclwnUKkr
iR1IfXvolIwuhE01vCojuRPpsAK8gdDB1ZZ8vvW0dAvGyhjxRFLBKRWDBI1U
dK83vtx25sdsIPW1oNnKqKiJ4A4gJcE8viEtvV5tPByUeQgQ47mEjAkUa0zT
iuWYYSmxKc4GX8k4WF4XSrGBOkmGfwWOI66ybSBrIhtd0GYYVti3RpxjOAnD
XCEMH+uEwZs8bImQpiGFdSpOlDoAmpSqkMjBEld4dzvRtA5zIRUCdhFAHI8H
DnvO0Z5w7qeaJ22kpjta3yD3HbiSoL8xDgsvWHKG5uUG48vF4c766TwrSVBH
6x90OYOcslDygyLYpabkk1GsqoX7NeWL26vOi1UBol1gs8b6IXDGUkd0jM+M
hV6aGKirMOk9iRkxuE+uP6wE22Beyhre0OSuo805FJzH2s2fDdmrh31qiHQp
WQmfQbisuJVZJN2hRRpCDAAsRHKykyVtyshF+jCYldFLWEURRqUAJpBjCkgi
QGGvzZ2bkQk9zTrRdzTGgdQxqwvyoDTDGqeAiDESarY0YG60DRTHmXkRpE8z
y1vqgTH3pkkl02Dp2SF/DLbNpNdcoxejwqB3Y7erRzEyRYgPHqq89WOochN1
/eQo9GJ/OFNCGnBG4saOrkM9MkcOcQxbo6hgAbiI22uPIwbEqsJWW8XBbegr
ELj17si1XEhACZNkVcmaIqE97ccCJXcWGNHXBMClC5U4dRXrVQTQxVzgx6Fu
sEvRWWJSCFAKKsd2Z7dqsGB2mtEpI2UhVUUsMJPFxhz3YmP0Y/T9drI4psMi
V3te5uTpn/RZjdELLc9nCH7cE7UoNz3QVEwXL9C064vbyfXFGaLT6Pdv9eE3
+1+9uTqXV4/2vwofllefh+oiSZ0W3p+Bci2hcgrN6ZhG+O3gwEf7G8J4Bxre
XJ3vb4ij7zc8ffGCm3HBlKxYSnbi9tZNKZv5xxRMgdcGKqV0E+7Ca908u3xc
0noxMWwALlk4FDTATqkY6DIpEcOzpYTc6WSPORX344E9LZsJ/Z1k45b/MBF1
f5NPBnLyDS0uj5Fj37JJc4LKwZ5CMut897JxmGbh87HCrXt7fX3x6nZy+eMJ
KX356nXrvoeR142ZNHZWGrcaqF4zVMankYC+fjmf3QVyGrO7RI7s0+BH6Meh
RL8dgCnCeJn6MuH95fwaQLyso3dH3g3hGyiFxGxaZeckvpuei4RTycnMysao
hPZiqbtJPvk8KCn7TdwhiCIfKBnK5o0Axs1LBAdA/saDOAO3kKQXJYEBg3aG
ZLTox0vjqUNQnuh3sdIQJLDZuBJh7ZAA164R6J2W0wq2kSwvCCY+aBw6esFr
gPf4cW/PKNCBCxmp7ko9SfLShXxp48Oh7BiYwONvNpQRk0hQgXXrawmYYv7t
ajIQi5My2iro9wha7tX8iXm3pUU3h32H+s7nzqbUN321Q4Hx1YQCp692qDC+
OlS5RocE7BPLOb3xXdtNZR8zxXcybvO8+BaGB3D7rkXUNG0WEiW1l/IPY0E1
ljWyXz58OBlKKJ+3/j0ns3lK+cL2E8lryvcOi3Ic2+EibXz3rUm9rjar4/AW
/jlEmfJB8aJkD5Os4z0gvAiCMYIri2gKdSkRlUIpzXFxsXIWXh52i7ArhMq0
Mu6TyuQC2huuKpoadyQ0SdMbk2Mfg89CTUnTJvkKKXUoexeSszNOFn2UHusR
h1eFwxvzKA350OHlCCHvRp0E44W4TEDxnCH8pfgmHU6aDhW92RkQnMKA5irx
94mtLN9R1/H6diL/VHLMx5yHIMG7d4sii94bilUCoInC48BMLM4npNb0OAu5
CMCrnZBGRZAaimwhqD9YgJIUDUgnNZ16DB/0xAD4IozVC/fOQiU0+gqNIgGi
V1xOZUqRUfQHIEjuTAnjobyVsYEkGSDG9BYZ001gTJnCdIAJpTOiOwHMVjXb
Hu8ltOGtPrWFIgNfHociA/O5Kyf12k9YibR96jjUAoOTPrENfEXiKT7lK39P
m6puJ4UtHSrSvXbV3IP15Tht19atKSfTjd9O4LePamMWi8aCvU/GiFVu0tZ7
mEbco5RzhKdAJb+HwMF4T9HR6wM98IGwoee6sfqdXbdJ4Hf6jbyaQ/ILYkIS
xGYQK8jyctyJTcp+Q4Q2RmjAmHohGjCksERFSNdRbseKwjryV4PG3ApziMMH
hAea87TGRaHKRKhNk7HxE/tC6+dYPbqWE/xIsxsdV3X2jjCCOVnhOBI2OhU1
haU/CudzuA8y01W247NLQGaq6+tEWBZLtqH3EOEYkwtLEmcVLmJ4KXjLg9ty
bRtXF5hUbfiapMyXxo+wUdnAJECATl/8qMQizzfl3AFeV2LxOwFXfqyG7lpI
3R8+hU/9miyKcAcSs/n1mzNAlJZ+/MAlDHUfNitgY2gbrzdVcQAxFQqLzVON
67DCmLQTu0aHgbC25+PnPQsRMYNAFPLUsm1ey4Fy8VMqP1PV8jdlpFi5okIu
WDeUwWVZbxr9L6baQO3hw5E6/MtXz/RZXTcFqwdvK9gvb0p9C7N8/Pb27EnO
seOwqi4doNRRGKDDqxzyYpATL8TGAEBGculkbnFICpU4wTJwA7hZEBV0XzcF
GWFZ3GprNUWURF3QIhWbRPzAbApgRiVsOQgkYB5XKipeaekDTOCyWSE+9w7c
/GM7HjEuAFD2KJHh9IAURgqKEMhK8Pr3ddZpQkKyjLIY2wtDEksupxNmOGAb
K9bT+0kwLLr2G8zBy/KN4LjhRRwgO0Z2g/DrisAkJobqlKZZxDSN/CM752v0
r81QlnRUBxvgBEHNS0u2CokPSlxO3feIGJ32PWZJ5cwjsywxx8gQU03LTSGO
j9LaDNRaG2QIH/MeEvuHGXo5KYaKsum+BQnehTie8mE+ny9U5PTZc+D1skbR
C+PQ8UHuxk4B+Q5j4qSJtgB1ZldMvXiYDdGINDHQKF15FZEO9O0s788W+ZlE
vI3DPrBKw4utyE7PhxW2wqWpeneKDCOVJuEd57uS8qckWyHfI2JviHhXNG6m
aSJYI/SJa7ME8HBDNY15OyPII7JkIIE0ZXCwRSBBQOfOzGwZHHmISCU7TCfV
/YNoy0fLun2UxD4TfgVwyfchUKEC0xbATKYxD3I0gtXVOFUVUHnYryfQKw+a
ZEje9R9plZG3P84wI2/vss24ul9D1vnhu7/j3eH736+6hy8PlsBlwpJfZp5l
8uxjxH5e8lTmp0dABIZsRGFSqXkoaYd6dt84FJpldqGOxqSdm6DM+s0DL6HE
N2TuCZ/hCcnfMBuMPYTcEKOOHYZsKuRrBGFR0H9poAi0Ux0kXj96GCVqQNYM
+3sB/1jV/AGMUiltsQhO6Ty5LTtgEa5AX9x4TnOj2L0PRIgz9wfKgmQj2pzp
Ms9iooy2VhDHv4KqV2JwDTORaNBxdmcxiCwWzdht8k7I2z6D9mZdV1xgriOx
KXWD8p8LmYfn1kScbk7C+MMWkXfkHjaY/RC9yAGXmzSCeVAs9XzLBC5JhkCx
NUSrazU4LZZr8S44kRafkk4Ri6hGACTI2/fsYcds6wVqNkxb5RjBVgQwF6tH
cVyYB47T7uLrs3pte9mqGGk6mMBhxMUKyJ8PqSURCSDF2Vhu47omlOTBlLrZ
VAdYwmtlKrPgqkISm8XOb6xFQkJ1LH6HYu1msUBlTGySAtyR4ETGDMwJWxsH
A/M+IMtiY31dbsR/PrXqUL/8HquJpRXb2mVjPdQjov0LudXYBgZtnpKKAyrN
4TPVuWWQyZx+59oMKKzH5R+lvEYspcJNdGYpVaiyEIZnzkUhcVNxOjJg+ovb
u4psJ6iJEhsZCxEc6ewZvjdKWEjyIuWjS38QnTg15MZkPpjxdbEBCHcu/MDq
xAj7KP0M6d0DVh8yBqsdJ5mDyXvFkRI2F+Q2lD03lfvrBhKk8qUFzUFq+EVQ
kEGYSZEhnJuaIz4d3gwIkyq3I5XIlQ9kbM/EXuEE6u/OsxnD6TONRzyOHY1H
nI0Pajyh/aDagc4MVhAa+ntC29IpiRxaML0PLaiCYK/NA+qFjKqrXvDzRwKq
jeW/OncSIm9GQ6NOZG4Vcx952zHDD3hQ00pkIesVpAzHIuCyWr0qarST1DKv
3xbx0Er8G2nqD3aGyLllR/OgM4ZCiapaJkHmCQbl4x0a9fjEwNooTrTpJYV6
UnEz4X+IJKL4uYEtTiD6RpdUHKTL2nevJrptVHKSPnY9k97RXawS+GRCP7T+
PmHXhmNHFhuDmbCD36Nw/rfacd0wrtoK1UPQcQccys7ah/BiKTKdbmeaLbIb
bIDHEgg/O9SJPYZUhDEguhPCLuEtlHQfVhDisDC/GaL2unEJmN6fcvQMwMv2
hHQGj/mHJyEtwMZLws2YObPrP0vdxE+6s8LDn4yDZZB2fxphFFskjYlp0yEQ
nAZsgijfFpGMGp9c0o/aOjSi/bfuE6S3Sg72w5uUedkGN4osg92N+ur/5Ebh
EHCjxqXv5BOH8sB3toEkBywMhB2MAWiGsw12lNWHdvUMjgRM+Ee42ShIRBjn
cS/5N1OaFESZKqxJ04O7wx06a9p/rM7eB2OV9TLHX3XxcfxODpFLWWc+HGGc
6fepOrre+9/w1Ppr1cc3dOHwem18qA6MKiTKdOyVDxlRxBYEaewZcl7wjUgI
OEMTOFxWFuthyFcK4+LVUymMS+cwrjatHB8PBZ2r/IeeXEUQp/n8SEq2UrVk
Qivqx/ADl2JHbw7JDCuayJNc+kKD5uXl0eTN9evbi7Pbq9evJre/vrmYvHz9
Lxe3twRaTAWvADN8uAcMCbo4399FFM74wdM/hff/pH+2nBiFMNZLzPpBOdu2
emCeSdO+Bk6la1HBjc1iZpekrYTUjuOzpz2h9mjSu3Eyy+EdKJeTUEYcn/mO
YNvbUWyTPkmXLYJWQ63czkFAw/t8vppN7kzpigSmipt2e/02B6WC6er5EZmu
sBn3+k3W7PL0xU3ebgf0tHO8w8e7wyUBfmCiHbKT93dwd9QT2Zn4HBFICjgz
If/AnO45PDKWpvVYI4qio8waHPlQnyMaBsD3FOwLPiuwAeYASGKQ1tFD/YqD
aagLSqaM6MYp2YsoK6wpKRMacwxJghx45NH4i/EXXVC9lPzyUm+jU7AgS8FK
toX4odS6tJOH9guzSkUlxOENJSIVKFUN2in40jHo7iQGSomW+bkPteX0ui7d
TOquefE/np69wBjlNYRIYrJCOAJYt2nqOEWMx9Q1A33zLEKXDSxLwylGu6OG
fl2FVyLpzb53lGlT5kY1IbjqYHCmk9rOOKUKQ34VMswZxtPE/Ib+7+s7TZ8D
HYMZoynS2BaqHjiUPaSXKYwV+LPvJxQtMTl99etwOJaEGsMQstfRpY/m0TSU
RFU1WiW7Jk4EHWLmAM6aDsW3Wtkfj+WW6ZAfHY2fq+4RZ6N9mxSqpGTgrEc+
IrgZXSFMw/R6+hsYp2/xrWu7cL5tto/GPZZJcRzXZ6fHzJxOb/51AlyLGDnk
rkfa9K0+evbNRzc6/UUaHQ7aGCp7f9DMTLAtQFd6Zfw7ohhEqPqSXL4DpGaQ
yKQGKDUS5saIcOsnbT15Z+06UG0q/Bc9NfgyrOsEhjJEaQe+zzMY+IWx8rfb
NWUrHXUOkAQhRmgqkhOJ/2amzGlW+Dgm8dkYYIDW1U0T3BadOfAUlRQqpSMk
8l9oy0mUMJK4kxpPDa0gC26x15CCMwYzYzdqiKZ2/cyJyxE654ng8DH9jpo/
pORQzHJEygT9ADo5wE66ugEcY4xLms8ns+mEBjwx1ZbWLZNa3ry6vJlcXk46
Bx5NrnAvRh/z8s98H/Yb3+KAw9WgU/ESRnWJP8QbIvWuuhZiLuvX2YyE2FxV
gma1o9SjzDltmF4llQazfrh7iWXkeH2Q763x7Qh3EYlxS8CRwSOUbaaSodGc
9i04+sKRyspx7tJZlYx7Z4FDoLhu5UrTcE4ftXvjHv5k4KB7izIiy6P66OFw
C/seKARDBsfeN2KRQf4UiU+c8uoSUBGQ4/mq4kD/e4QHhhq+AGOsolUBnQJ0
1V2DG4gJoJV9v0YpUFKdhRTomGMqjKl7tkKlBkVdEX4ir0SPlry7+l1YxsD8
MU2S8d4tOIG6SkhUrfNk7Ln5ZEg0RPxarSgwn3NTN81m3YaqI4OpSOeUHXAu
px6GWqgsB3zHqxJz1/cEkMDWx0dcBESkvTN2JHIN1CGfCmAI60oqOxiqiMA5
KZ1Py7kqvP7NHWPugkTZgHDSsmcjpD/v5nqDPDdVCw1iEQWcE+DmxdXHdBnH
MLcGDQ8hAw0FUIUPIIHmgYZMPa4C29l9ja38CR14dj7KS0o8QPh5eZoEhHfe
7xWd57o7o53rABcgwdGzOS+pbg8Igayk2c5s6fNNgyeycLFoFlob8ymHQ4ap
8XFepl0Kwx7eq7yIGvnGBgbOSZp43d9nd8Krw/FX40P8Kpy+fqU+yBpT7Dsw
IR7toQH0lzD4d5Lk1d2ErEG2j9mdeolpQpLYfrovswHDUotUDDQMRnVkZR5C
RI9IFbsKnqWxG7PGosPHlKEgQCT43H9y5XdC49pQpaY7dCQqyCJD2tDoUh+s
2rkjAztrWj56LbnHcFFIpwx2X6xdKzkiusUYQNEBtDKsbN24v3GAQFwOFMUW
tu2VM2UJLYK3mQs+ftxnvSqEr0fu+CQmkOKZgaoW+G3UfYlLnJ69EKnQZ/s1
kpq0Wx2RKZK54PTs7OLmhpMQJSkfECYsQujALGIW7FhZOCp9mKjJFnnebS56
70fAENYJAloMDrCDLYgL6WEzbC0AqQq8dZ2SJSxuF8nh65Tt4DcHJtDNTGdm
lKojK96BT6Z2acp5UFARwbjxFpK4Sf625Oah9KDkwHAqH2/B34dZLFhBF075
+s3FK95B3AvExe+ougHVFVFC01ReODh8aP6FLe0iBD8NIRJCNcuR9tuqXVqo
8rZxrNwsXJFn2lSM7KfS88MJsuKF69wznsLOoiFoBMPKGmFP08KjPFxVo9ru
A3qdMGY1lNmhX0LOeUoVatJaoiR3SR7DcLTTKs2LjRBjyOBVpR+IHY3VJXG2
kYQdRFd4ehNxTlRvp42JlGWl1cahOFc8pRSPsWW46p0WuCXQQsV3R4ivC8VS
wquuwEUDVB2VHuHEVKqysL6mcWFsCfgVNFRi2WFCOfVWe6ktVqUExztkSQn1
KWPiHUQHkAAeTwN8ZST5otENh9a4wSkxoHrTVFSJQCXUvadzwPUMCcmJDuO5
llST4VjnqTBDmeik0nIYKOXf45lF3TyjbRsvI8poz+MIpuEk2LTxT7r8tJsO
Jg2yanOu/bkfQGfuTKIvRBdLjlOie5jvknBbYhLNTnADrGSUpHOVAlNlGXS6
uSTMI/5KZ7fq6gpYgIR1BKpoDWg2UMWG7MORYQbSSMIEkM0VJJpybblVbI2N
OWAlZScktQSTHqbEL5IVlioaYUepyo2LhcI4upcKKFFuYCQV1xf/8+3V9cW5
0D5yKQFuVXGaOc6NljAaOGUd3vyU1hRBHkmdJpWUskpPkyczWGMjC5F7tYDw
E6bCWIzBFqpDdeUEjFP3QaToPqgH6KBxXgFwpCpsWuMrr+4lwk2ABhGP5Joo
GKWUGkwAw49nANAEczYr3wKUDsE3L8zWNkHlo1+v35zdXJxNfrihZBhBH7yh
UACv1OTn5Vbf49p/ThcXvXRc9zG2H0+U+p7z2wuqFGQEYtjwaN24aubWcPGS
AO8SEvIh5g2r3Sqs8USbIomn0prFmrwplOecLz7XrMGWEh7wgrs9425VMlNR
Dp7rx/Hp3fMnAe8AZu4QcjQ3M8iN6CRigtzTUtmN4UFiNs1lj0EZWhKcZdLN
AAUZa32OJT19zd9kFZuIcVriuS+rdJQb2Ns0JZY4R1QA9ep0JRjQYip4epBU
XDsIddDF3V+odePuXGmBqnkP6hrlxQV3XVVhPs0BC8Tx+C+oA2ax9KyDcpIg
CDsn7rmlWFRYCMFnpZ2qLmj5lgtZhQOHuHs6miFLpHwG56XyBbbvW1thtb5O
1TrEgWK2unvIPUew23reLfAoHDYkuN1UAGcot5z8jsq/cy+Q8tajmyQbkqhh
YphBmYkCBk+JqUA3WTUZtD5ny51bIyT7nJLceXKLMVk8XdLkmrT1gioAB9WY
76riu8o3LkmSOd+gJSYZj2dowDRRGUYqXR66SbQnNJ5RrFsVc7PR+FR2TBF8
J9GYZuvl8t9SBe14+alEFMagCrkYMQvjIxJvcDQRhGtFUUpQfMBgEEcKIRto
FwUYPtH5xU+9eD0TUltr5rJ0VjFeN4os497mc/qOqo52E6abjS0N5w6l+xzX
TnFCXi41LsSXrWQDGab/1TZT29Re//SF/gESkbiZ6jILfZqUZnzT1IvGrPRV
QPc//uHm5uD0zRWT2ePDIyCzkPbfx5DiMIcQ7pAgdjqZNWbBTIAV63avI7Tt
rOOO8j9kFb06fXXas4imLtAk4RE6PNEMmQDLEifnN6nwixeUsO5RGmj41TTL
BpmdCYMEo2GipsJVTp2wYc7G+3pGpTAlT2LHp00IAhrByVBGLEYOttMyTcvk
KRWg/iNDz72CG4D//aF/Qh/cH/r68kwe/Vjf6z/0S6z3/RNxXK//UPqPg4MD
7Kr39clPR/oP/ez9l9j++vLs9vvzw2fwKj44zJFpf6CfqjvSTqYwHOgpriY6
60MEmuD3QIHqbmfXf61T/7WK/mvmNL1NgR1vYmtRo6Ecz37fNrDp/Y5szlQB
s45fiNszPGbZqaFt+rG+720SLm7Yp/1D/kMfPdOd/XqR7dYfD3Ry+gt0cvhQ
J7LZybRzr2Rv1p1t/0yfziQ8joAbv5/QTbDFt4/mpvQ2uDEFKUNi+I9mtbKN
XwMVwxBbhyWDUSxAo3/RUJ4AQJqqS1Sd7o5O9Lm5c4W+LLdVNdK3DYQLvdy+
K+10A47Jc1c42+hL24Bh4V+sqQ7eOAh5US/rColfNdL0REOOoXpDctz3jTOV
fmPuIdTknRsrdbaEhLX1eql/tOW9W6Ai0B0jp5LFdMQwTi3jhLcVy3hsmFlx
3Ukiuy/r36CGF2kSEe+rVDY4sV4HG36eZmBnJ5/p15Be+ZTkwqvWrnZuDIYh
zCROWsIZoJZGIcnH15tp6fyS0TcgwF6ejZU6JCEF5UiWj2C8v3FvS2sK23xH
75EC+sK1bWn1RVU4U41A9MSwgLomMOov59fy+hy0I1NgadQRgg/Q/HP3XIsl
rJVXga8wuY8yHlCLNxHpSEQG33/T1O+3WQN6fvviRv1/WTuHBhBAAQA=

-->

</rfc>
