<?xml version="1.0" encoding="US-ASCII"?>
<?xml-stylesheet type="text/xsl" href="rfc2629.xslt" ?>
<!-- generated by https://github.com/cabo/kramdown-rfc2629 version 1.2.13 -->
<!DOCTYPE rfc SYSTEM "rfc2629.dtd" [
<!ENTITY RFC4760 SYSTEM "https://xml2rfc.tools.ietf.org/public/rfc/bibxml/reference.RFC.4760.xml">
<!-- ENTITY USECASES PUBLIC ''
      'https://xml2rfc.tools.ietf.org/public/rfc/bibxml3/reference.I-D.draft-geng-rtgwg-cfn-dyncast-ps-usecase-00.xml'-->
]>
<?rfc compact="yes"?>
<?rfc text-list-symbols="o*+-"?>
<?rfc subcompact="no"?>
<?rfc sortrefs="no"?>
<?rfc symrefs="yes"?>
<?rfc strict="yes"?>
<?rfc toc="yes"?>
<rfc category="info" docName="draft-liu-can-ps-usecases-00" ipr="trust200902"
     submissionType="IETF">
  <front>
    <title
    abbrev="Computing-Aware Networking (CAN) Problem Statement and Use Cases ">Computing-Aware
    Networking (CAN) Problem Statement and Use Cases</title>

    <author fullname="Peng Liu" initials="P." surname="Liu">
      <organization>China Mobile</organization>

      <address>
        <email>liupengyjy@chinamobile.com</email>
      </address>
    </author>

    <author fullname="Philip Eardley" initials="P." surname="Eardley">
      <organization/>

      <address>
        <email>ietf.philip.eardley@gmail.com</email>
      </address>
    </author>

    <author fullname="Dirk Trossen" initials="D." surname="Trossen">
      <organization>Huawei Technologies</organization>

      <address>
        <email>dirk.trossen@huawei.com</email>
      </address>
    </author>

    <author fullname="Mohamed Boucadair" initials="M." surname="Boucadair">
      <organization>Orange</organization>

      <address>
        <email>mohamed.boucadair@orange.com</email>
      </address>
    </author>

    <author fullname="Luis M. Contreras" initials="LM." surname="Contreras">
      <organization>Telefonica</organization>

      <address>
        <email>luismiguel.contrerasmurillo@telefonica.com</email>
      </address>
    </author>

    <author fullname="Cheng Li" initials="C." surname="Li">
      <organization>Huawei Technologies</organization>

      <address>
        <email>c.l@huawei.com</email>
      </address>
    </author>

    <author fullname="Yizhou Li" initials="Y." surname="Li">
      <organization>Huawei Technologies</organization>

      <address>
        <email>liyizhou@huawei.com</email>
      </address>
    </author>

    <date day="23" month="October" year="2022"/>

    <workgroup>rtgwg</workgroup>

    <abstract>
      <t>Many service providers have been exploring distributed computing
      techniques to achieve better service response time and optimized energy
      consumption. Such techniques rely upon the distribution of computing
      services and capabilities over many locations in the network, such as
      its edge, the metro region, virtualized central office, and other
      locations. In such a distributed computing environment, providing
      services by utilizing computing resources hosted in various computing
      facilities (e.g., edges) is being considered, e.g., for computationally
      intensive and delay sensitive services. Ideally, services should be
      computationally balanced using service-specific metrics instead of
      simply dispatching the service requests in a static way or optimizing
      solely connectivity metrics. For example, systematically directing end
      user-originated service requests to the geographically closest edge or
      some small computing units may lead to an unbalanced usage of computing
      resources, which may then degrade both the user experience and the
      overall service performance. We have named this kind of network with
      dynamic sharing of edge compute resources &ldquo;Computing-Aware
      Networking&rdquo; (CAN).</t>

      <t>This document provides the problem statement and the typical
      scenarios of CAN, which is to show the necessity of considering more
      factors when steering the traffic to the appropriate service instance
      based on the basic edge computing deployment to provide the service
      equivalency.</t>
    </abstract>
  </front>

  <middle>
    <section anchor="introduction" title="Introduction">
      <t>Network and Computing convergence has been evolving in the Internet
      for considerable time. With Content Delivery Networks (CDNs)
      &lsquo;frontloading&rsquo; access to many services, over-the-top service
      provisioning has become a driving force for many services, such as
      video, storage and many others. In addition, network operators have
      extended their capabilities by complementing their network
      infrastructure by developing CDN capabilities, particularly in edge
      sites. Compared to a CDN-based content cache capability, more diverse
      computing resource need to be provided for general edge computing in an
      on-demand manner.</t>

      <t>The reason of the fast development of this converged network/compute
      infrastructure is user demand. On the one hand, users want the best
      experience, e.g., expressed in low latency and high reliability, for new
      emerging applications such as high-definition video, AR and VR, live
      broadcast and so on. On the other hand, users want the stable experience
      when moving to different areas.</t>

      <t>Generally, edge computing aims to provide better response times and
      transfer rates compared to Cloud Computing, by moving the computing
      towards the edge of a network. Edge computing can be built on embedded
      systems, gateways, and others, all being located close to end users'
      premises. There are millions of home gateways, thousands of base
      stations, and hundreds of central offices in a city that can serve as
      candidate edges for behaving as service nodes.</t>

      <t>That brings about the key problem of deploying and scheduling traffic
      to the most suitable computing resource in order to meet the users'
      (service-specific) demand.</t>

      <t>Depending on the location of an edge and its capacity, different
      computing resources can be contributed by each edge to deliver a
      service. At peak hours, computing resources attached to a client's
      closest edge may not be sufficient to handle all the incoming service
      requests. Longer response times or even dropping of requests can be
      experienced by users. Increasing the computing resources hosted on each
      edge to the potential maximum capacity is neither feasible nor
      economically viable in many cases. Offloading computation intensive
      processing to the User devices would give the huge pressure of battery,
      and the needed data set (for the computation) that may not exist on the
      user device because of the size of data pool or due to data governance
      reasons.</t>

      <t>While service providers often have their own sites, which in turn
      have been upgraded to the edge sites, a specific service should be
      deployed in multiple edge sites to meet the users' demand. However, only
      the deployment itself might not enough to fully guarantee the quality of
      service. Instead, functional equivalency must be ensured by deploying
      instances for the same service across edge sites for better
      availability. Furthermore, load is to be kept balanced for both static
      and dynamic scenarios. For this, traffic needs to be dynamically steered
      to the "best" service instance. For this, traffic must be delivered to
      optimal edge sites according to information that may need to include,
      e.g., computing information, where the notion of 'best' may highly
      depend on the application demand.</t>

      <t>A particular example is the popular and pervasive 5G MEC service. In
      5G MEC, ULCL UPFs are deployed close to edge sites, which are capable of
      effectively classifying &amp; switching uplink traffic to the suitable
      computing-resources that might be located either in local-area DNs,
      operators' DNs, or even 3rd-party's DNs. Thru possibly using some
      'intelligent' criteria, this could warrant the selection of resources
      with either low&middot;, high-computational power or all-involved
      requirements.</t>

      <t>This document describes sample usage scenarios as well as key areas
      in which current solutions lead to problems that ultimately affect the
      deployment (including the performance) of edge services, and proposes
      the desired features of the CAN system. Those key areas target the
      identification of candidate solution components.</t>
    </section>

    <section anchor="definition-of-terms" title="Definition of Terms">
      <t>This document makes use of the following terms:</t>

      <t><list hangIndent="2" style="hanging">
          <t hangText="CAN:">Aiming at computing and network resource
          optimization by being aware of not only routing metric but also
          computing resource metric in deploying computing and network
          resource, steering traffic to appropriate computing resources,
          etc.</t>

          <t hangText="Service:">A monolithic functionality that is provided
          by an endpoint according to the specification for said service. A
          composite service can be built by orchestrating monolithic
          services.</t>

          <t hangText="Service instance:">Running environment (e.g., a node)
          that makes the functionality of a service available. One service can
          have several instances running at different network locations.</t>

          <t hangText="Service identifier:">Used to uniquely identify a
          service, at the same time identifying the whole set of service
          instances that each represent the same service behavior, no matter
          where those service instances are running.</t>
        </list></t>
    </section>

    <section title="Problem Statement">
      <section title="Multi-deployment of Edge Sites and Service">
        <t>Since edge computing aims at a closer computing service based on
        the shorter network path, there will be more than one edge sites with
        the same application in the city/province/state, a number of
        representative cities have deployed multi-edge sites and the typical
        applications, and there are more edge sites to be deployed in the
        future. Before deploying edge sites, there are some factors need to be
        considered, such as:</t>

        <t>o The exsiting infrastructure capacities, which could be used to
        update to edge sites, e.g. operators' machine room.</t>

        <t>o The amount and frequency of computing resource that is
        needed.</t>

        <t>o The network resource status linked to computing resource.</t>

        <t>When the edge sites are deployed, to improve the effectiveness of
        service deployment, the problem of how to choose optimal edge node to
        deploy services needs to be solved. More stable static information
        should be considered in service deployment, <xref
        target="I-D.contreras-alto-service-edge"/> introduces the
        consideration of depoly applications or functions to the edge, such as
        the type of instance, compute flavor of CPU/GPU, etc, optional storage
        extension, optional hardware acceleration characteristics. Besides
        those, more network and service factors may be considered, such
        as:</t>

        <t>o Network and computing resource topology: the overall
        consideration of network access, connectivity, path protection or
        redundancy. and the location and overall distribution of computing
        resources in network, and the relative position towards network
        topology.</t>

        <t>o Location: the number of users brought, the differentiation of
        service types and number of connections requested by users, etc. For
        edge nodes located in popular area, which with large amount of users
        and service requests, the service duplication can be deployed more
        than other areas.</t>

        <t>o Capacity of multiple edge nodes: not only a single node, but also
        the total number of requests that can be processed by the resource
        pool composed of multiple nodes</t>

        <t>o Service category: For example, whether the business is multi-user
        interaction, such as video conferencing, games, or just resource
        acquisition, such as short video viewing Alto can help to obtain one
        or more of the above information, so as to provide suggestions or
        formulate principles and strategies for service deployment.</t>

        <t>For the collection of those information, seconds level or minutes
        level frequency is enough, while serious real-time processing isn't
        necessary. For example, periodically collecting the total consumption
        of computing resources, or the total number of sessions accessed, to
        notify where to deploy more VMS or containers. Unlike the scheduling
        of request, service deployment should still follow the principle of
        proximity. The more local access, the more resources should be
        deployed. If the resources are insufficient, the operator can be
        informed to increase the hardware resources.</t>
      </section>

      <section title="Traffic Steering among Edges Sites and Service Instances">
        <t>This section shows the necessity of traffic steering among
        different edges in the real city, considering the mobility of the
        people in different time slot, events, etc.</t>

        <t>Traffic needs to be steered to the appropriate edge sites to ensure
        the application demands. Though the computing resource and network
        resource are considered when deploy the edge sites and service, but
        the reference resource information are more static, which can't meet
        the real-time or near real-time service request. That is, in some
        cases, the &lsquo;closest&rsquo; is not the &lsquo;best&rsquo;, there
        will be the variable statues of computing and network could be
        summarized as:</t>

        <t>o Closest site may not have enough resource, the load may
        dynamically change.</t>

        <t>o Closest site may not have related resource, heterogeneous
        hardware in different sites.</t>

        <t>Therefore, more enhancement based on edge computing is need.
        Because for edge computing, the service request always be steered to
        the closest edge site.</t>

        <t>We assume that clients access one or more services with an
        objective to meet a desired user experience. Each participating
        service may be realized at one or more places in the network (called,
        service instances). Such service instances are instantiated and
        deployed as part of the overall service deployment process, e.g.,
        using existing orchestration frameworks, within so-called edge sites,
        which in turn are reachable through a network infrastructure via an
        edge router.</t>

        <t>When a client issues a service request to a required service, the
        request is being steered to one of the available service instances.
        Each service instance may act as a client towards another service,
        thereby seeing its own outbound traffic steered to a suitable service
        instance of the request service and so on, achieving service
        composition and chaining as a result.</t>

        <t>The aforementioned selection of one of candidate service instances
        is done using traffic steering methods , where the steering decision
        may take into account pre-planned policies (assignment of certain
        clients to certain service instances), realize shortest-path to the
        'closest' service instance, or utilize more complex and possibly
        dynamic metric information, such as load of service instances,
        latencies experienced or similar, for a more dynamic selection of a
        suitable service instance.</t>

        <t>It is important to note that clients may move throughout the
        execution of a service, which may, as a result, position other service
        instance 'better' in terms of latency, load, or other metrics. This
        creates a (physical) dynamicity that will need to be catered for.</t>

        <t>Figure 1 shows a common way to deploy edge sites in the metro.
        There is an edge data center for metro area which has high computing
        resource and provides the service to more UEs at the working time.
        Because more office buildings are in the Metro area. And there are
        also some remote edge sites which have limited computing resource and
        provide the service to the UEs closed to them.</t>

        <t>The application such as the AR/VR, video recognition could be
        deployed in both the edge data center in metro area and the remote
        edge sites. In this case, the service request and the resource are
        matched well. Some potential traffic steering may needed just for
        special service request or some small scheduling demand.</t>

        <figure anchor="fig-edge-site-deployment"
                title="Common Deployment of Edge Sites">
          <artwork>
     +----------------+    +---+                  +------------+
   +----------------+ |- - |UE1|                +------------+ |
   | +-----------+  | |    +---+             +--|    Edge    | |
   | |Edge server|  | |    +---+       +- - -|PE|            | | 
   | +-----------+  | |- - |UE2|       |     +--|   Site 1   |-+
   | +-----------+  | |    +---+                +------------+
   | |Edge server|  | |     ...        |            |
   | +-----------+  | +--+         Potencial      +---+ +---+
   | +-----------+  | |PE|- - - - - - -+          |UEa| |UEb|
   | |Edge server|  | +--+         Steering       +---+ +---+
   | +-----------+  | |    +---+       |                  |  
   | +-----------+  | |- - |UE3|                  +------------+          
   | |  ... ...  |  | |    +---+       |        +------------+ | 
   | +-----------+  | |     ...              +--|    Edge    | |
   |                | |    +---+       +- - -|PE|            | |
   |Edge data center|-+- - |UEn|             +--|   Site 2   |-+
   +----------------+      +---+                +------------+         
   High computing resource              Limited computing resource 
   and more UE at Metro area            and less UE at Remote area</artwork>
        </figure>

        <t>Figure 2 shows that when it goes to non working time, for example
        at weekend or daily night, more UEs move to the remote area that are
        close to their house or for some weekend events. So there will be more
        service request at remote but with limited computing resource, while
        the rich computing resource might not be used with less UE in the
        Metro Area. It is possible for so many people request the AR/VR
        service at remote are but with the limited computing resource,
        moreover, as the people move from the metro area to the remote are,
        the edge sites served the common service such as intelligent
        transportation will also change, so it need to steer some traffic back
        to Metro center.</t>

        <figure anchor="fig-edge-mobility"
                title="Steering Traffic among Edge Sites">
          <artwork>
     +----------------+                           +------------+
   +----------------+ |                         +------------+ |
   | +-----------+  | |  Steering traffic    +--|    Edge    | |
   | |Edge server|  | |          +-----------|PE|            | | 
   | +-----------+  | |    +---+ |           +--|   Site 1   |-+
   | +-----------+  | |- - |UEa| |    +----+----+-+----------+
   | |Edge server|  | |    +---+ |    |           |           |
   | +-----------+  | +--+       |  +---+ +---+ +---+ +---+ +---+ 
   | +-----------+  | |PE|-------+  |UE1| |UE2| |UE3| |...| |UEn|
   | |Edge server|  | +--+       |  +---+ +---+ +---+ +---+ +---+
   | +-----------+  | |    +---+ |          |           |  
   | +-----------+  | |- - |UEb| |          +-----+-----+------+          
   | |  ... ...  |  | |    +---+ |              +------------+ | 
   | +-----------+  | |          |           +--|    Edge    | |
   |                | |          +-----------|PE|            | |
   |Edge data center|-+  Steering traffic    +--|   Site 2   |-+
   +----------------+                           +------------+         
   High computing resource              Limited computing resource 
   and less UE at Metro area            and more UE at Remote area</artwork>
        </figure>

        <t>There will also be the common variable of network and computing
        resources, for someone who is not moving but get a poor latency
        sometime. Because of other UEs&rsquo; moving, a large number of
        request for temporary events such as vocal concert, shopping festival
        and so on, and there will also be the normal change of the network and
        computing resource status. So for some fixed UEs, it is also expected
        to steer the traffic to appropriate sites dynamiclly.</t>

        <t>Those problems indicate that traffic needs to be steered among
        different edge sites, because of the mobility of the UE and the common
        variable of network and computing resources. Moreover, some apps in
        the following Section require both low latency and high computing
        resource usage or specific computing HW capabilities (such as local
        GPU); hence joint optimization of network and computing resource is
        needed to guarantee the QoE.</t>
      </section>
    </section>

    <section title="Use Cases">
      <t>This section presents a non-exhaustive list of scenarios which
      require multiple edge sites to interconnect and to coordinate at the
      network layer to meet the service demands and ensure better user
      experience.</t>

      <section title="Computing-Aware AR or VR">
        <t>Cloud VR/AR services are used in some exhibitions, scenic spots,
        and celebration ceremonies. In the future, they might be used in more
        applications, such as industrial internet, medical industry, and meta
        verse.</t>

        <t>Cloud VR/AR introduces the concept of cloud computing to the
        rendering of audiovisual assets in such applications. Here, the edge
        cloud helps encode/decode and render content. The end device usually
        only uploads posture or control information to the edge and then VR/AR
        contents are rendered in the edge cloud. The video and audio outputs
        generated from the edge cloud are encoded, compressed, and transmitted
        back to the end device or further transmitted to central data center
        via high bandwidth networks.</t>

        <t>Edge sites may use CPU or GPU for encode/decode. GPU usually has
        better performance but CPU is simpler and more straightforward to use
        as well as possibly more widespread in deployment. Available remaining
        resources determines if a service instance can be started. The
        instance's CPU, GPU and memory utilization has a high impact on the
        processing delay on encoding, decoding and rendering. At the same
        time, the network path quality to the edge site is a key for user
        experience of quality of audio/ video and input command response
        times.</t>

        <t>A Cloud VR service, such as a mobile gaming service, brings
        challenging requirements to both network and computing so that the
        edge node to serve a service request has to be carefully selected to
        make sure it has sufficient computing resource and good network path.
        For example, for an entry-level Cloud VR (panoramic 8K 2D video) with
        110-degree Field of View (FOV) transmission, the typical network
        requirements are bandwidth 40Mbps, 20ms for motion-to-photon latency,
        packet loss rate is 2.4E-5; the typical computing requirements are 8K
        H.265 real-time decoding, 2K H.264 real-time encoding. We can further
        divide the 20ms latency budget into:</t>

        <t>(i) sensor sampling delay(client), which is considered
        imperceptible by users is less than 1.5ms including an extra 0.5ms for
        digitalization and end device processing.</t>

        <t>(ii) display refresh delay(client), which take 7.9ms based on the
        144Hz display refreshing rate and 1ms extra delay to light up.</t>

        <t>(iii) image/frame rendering delay(server), which could be reduced
        to 5.5ms.</t>

        <t>(iv) network delay(network), which should be bounded to
        20-1.5-5.5-7.9 = 5.1ms.</t>

        <t>So the the budgets for server(computing) delay and network delay
        are almost equivalent, which make sense to consider both of the delay
        for computing and network. And it can&rsquo;t meet the total delay
        requirements or find the best choice by either optimize the network or
        computing resource.</t>

        <t>Based on the analysis, here are some further assumption as figure 3
        shows, the client could request any service instance among 3 edge
        sites. The delay of client could be same, and the differences of
        differente edge sites and corresponding network path has different
        delays:</t>

        <t>o Edge site 1: The computing delay=4ms based on a light load, and
        the corresponding network delay=9ms based on a heavy traffic.</t>

        <t>o Edge site 2: The computing delay=10ms based on a heavy load, and
        the corresponding network delay=4ms based on a light traffic.</t>

        <t>o Edge site 3: The edge site 3's computing delay=5ms based on a
        normal load, and the corresponding network delay=5ms based on a normal
        traffic.</t>

        <t>In this case, we can't get a optimal network and computing total
        delay if choose the resource only based on either of computing or
        network status:</t>

        <t>o If choosing the edge site based on the best computing delay it
        will be the edge site 1, the E2E delay=22.4ms.</t>

        <t>o If choosing the edge site based on the best network delay it will
        be the edge site 2, the E2E delay=23.4ms.</t>

        <t>o If choosing the edge site based on both of the status it will be
        the edge site 3, the E2E delay=19.4ms.</t>

        <t>So, the best choice to ensure the E2E delay is edge site 3, which
        is 19.4ms and is less than 20ms. The differences of the E2E delay is
        only 3~4ms among the three, but some of them will meet the application
        demand while some doesn't.</t>

        <t>The conclusion is that it requires to dynamically steer traffic to
        the appropriate edge to meet the E2E delay requirements considering
        both network and computing resource status. Moreover, the computing
        resources have a big difference in different edges, and the
        &lsquo;closest site&rsquo; may be good for latency but lacks GPU
        support and should therefore not be chosen.</t>

        <figure anchor="Computing-Aware-AR-VR"
                title="Computing-Aware AR or VR">
          <artwork>     Light Load          Heavy Load           Normal load
   +------------+      +------------+       +------------+   
   |    Edge    |      |    Edge    |       |    Edge    |  
   |   Site 1   |      |   Site 2   |       |   Site 3   |
   +-----+------+      +------+-----+       +------+-----+
computing|delay(4ms)          |           computing|delay(5ms)               
         |           computing|delay(10ms)         |
    +----+-----+        +-----+----+         +-----+----+  
    |  Egress  |        |  Egress  |         |  Egress  |            
    | Router 1 |        | Router 2 |         | Router 3 |
    +----+-----+        +-----+----+         +-----+----+   
  newtork|delay(9ms)   newtork|delay(4ms)   newtork|delay(5ms)         
         |                    |                    |
         |           +--------+--------+           |
         +-----------|  Infrastructure |-----------+ 
                     +--------+--------+                    
                              |                   
                         +----+----+                
                         | Ingress |
         +---------------|  Router |--------------+
         |               +----+----+              |
         |                    |                   |
      +--+--+              +--+---+           +---+--+
    +------+|            +------+ |         +------+ |
    |Client|+            |Client|-+         |Client|-+
    +------+             +------+           +------+
                   clien delay=1.5+7.9=9.4ms</artwork>
        </figure>

        <t>Furthermore, specific techniques may be employed to divide the
        overall rendering into base assets that are common across a number of
        clients participating in the service, while the client-specific input
        data is being utilized to render additional assets. When being
        delivered to the client, those two assets are being combined into the
        overall content being consumed by the client. The requirements for
        sending the client input data as well as the requests for the base
        assets may be different in terms of which service instances may serve
        the request, where base assets may be served from any nearby service
        instance (since those base assets may be served without requiring
        cross-request state being maintained), while the client-specific input
        data is being processed by a stateful service instance that changes,
        if at all, only slowly over time due to the stickiness of the service
        that is being created by the client-specific data. Other splits of
        rendering and input tasks can be found in<xref target="TR22.874"/> for
        further reading.</t>

        <t>When it comes to the service instances themselves, those may be
        instantiated on-demand, e.g., driven by network or client demand
        metrics, while resources may also be released, e.g., after an idle
        timeout, to free up resources for other services. Depending on the
        utilized node technologies, the lifetime of such "function as a
        service" may range from many minutes down to millisecond scale.
        Therefore computing resources across participating edges exhibit a
        distributed (in terms of locations) as well as dynamic (in terms of
        resource availability) nature. In order to achieve a satisfying
        service quality to end users, a service request will need to be sent
        to and served by an edge with sufficient computing resource and a good
        network path.</t>
      </section>

      <section title="Computing-Aware Intelligent Transportation">
        <t>For the convenience of transportation, more video capture devices
        are required to be deployed as urban infrastructure, and the better
        video quality is also required to facilitate the content analysis.
        Therefore, the transmission capacity of the network will need to be
        further increased, and the collected video data need to be further
        processed, such as for pedestrian face recognition, vehicle moving
        track recognition, and prediction. This, in turn, also impacts the
        requirements for the video processing capacity of computing nodes.</t>

        <t>In auxiliary driving scenarios, to help overcome the non-line-of-
        sight problem due to blind spot or obstacles, the edge node can
        collect comprehensive road and traffic information around the vehicle
        location and perform data processing, and then vehicles with high
        security risk can be warned accordingly, improving driving safety in
        complicated road conditions, like at intersections. This scenario is
        also called "Electronic Horizon", as explained in<xref
        target="HORITA"/>. For instance, video image information captured by,
        e.g., an in-car, camera is transmitted to the nearest edge node for
        processing. The notion of sending the request to the "nearest" edge
        node is important for being able to collate the video information of
        "nearby" cars, using, for instance, relative location information.
        Furthermore, data privacy may lead to the requirement to process the
        data as close to the source as possible to limit data spread across
        too many network components in the network.</t>

        <t>Nevertheless, load at specific "closest" nodes may greatly vary,
        leading to the possibility for the closest edge node becoming
        overloaded, leading to a higher response time and therefore a delay in
        responding to the auxiliary driving request with the possibility of
        traffic delays or even traffic accidents occurring as a result. Hence,
        in such cases, delay-insensitive services such as in-vehicle
        entertainment should be dispatched to other light loaded nodes instead
        of local edge nodes, so that the delay-sensitive service is
        preferentially processed locally to ensure the service availability
        and user experience.</t>

        <t>In video recognition scenarios, when the number of waiting people
        and vehicles increases, more computing resources are needed to process
        the video content. For rush hour traffic congestion and weekend
        personnel flow from the edge of a city to the city center, efficient
        network and computing capacity scheduling is also required. Those
        would cause the overload of the nearest edge sites if there is no
        extra method used, and some of the service request flow might be
        steered to others edge site except the nearest one.</t>
      </section>

      <section title="Computing-Aware Digital Twin">
        <t>A number of industry associations, such as the Industrial Digital
        Twin Association or the Digital Twin Consortium
        (https://www.digitaltwinconsortium.org/), have been founded to promote
        the concept of the Digital Twin (DT) for a number of use case areas,
        such as smart cities, transportation, industrial control, among
        others. The core concept of the DT is the "administrative shell" <xref
        target="Industry4.0"/>, which serves as a digital representation of
        the information and technical functionality pertaining to the "assets"
        (such as an industrial machinery, a transportation vehicle, an object
        in a smart city or others) that is intended to be managed, controlled,
        and actuated.</t>

        <t>As an example for industrial control, the programmable logic
        controller (PLC) may be virtualized and the functionality aggregated
        across a number of physical assets into a single administrative shell
        for the purpose of managing those assets. PLCs may be virtualized in
        order to move the PLC capabilities from the physical assets to the
        edge cloud. Several PLC instances may exist to enable load balancing
        and fail-over capabilities, while also enabling physical mobility of
        the asset and the connection to a suitable "nearby" PLC instance. With
        this, traffic dynamicity may be similar to that observed in the
        connected car scenario in the previous sub-section. Crucial here is
        high availability and bounded latency since a failure of the (overall)
        PLC functionality may lead to a production line stop, while boundary
        violations of the latency may lead to loosing synchronization with
        other processes and, ultimately, to production faults, tool failures
        or similar.</t>

        <t>Particular attention in Digital Twin scenarios is given to the
        problem of data storage. Here, decentralization, not only driven by
        the scenario (such as outlined in the connected car scenario for cases
        of localized reasoning over data originating from driving vehicles)
        but also through proposed platform solutions, such as those in <xref
        target="GAIA-X"/>, plays an important role. With decentralization,
        endpoint relations between client and (storage) service instances may
        frequently change as a result.</t>
      </section>
    </section>

    <section anchor="Conclusion" title="Conclusion">
      <t>This document presents the problem statement and use cases of CAN in
      which we observe the demand for considering the dynamic nature of
      service requests in terms of requirements on the resources fulfilling
      them in the form of service instances. In addition, those very service
      instances may themselves be dynamic in availability and status, e.g., in
      terms of load or experienced latency.</t>

      <t>As a consequence, we can get two obvious conclusion. One is that the
      traffic needs to be steered among different edge sites, another is that
      when steering traffic, the real-time network and computing resource
      status should be considered at the same time in an effective way. The
      problem of satisfying service-specific metrics to allow for selecting
      the most suitable service instance among the pool of instances available
      to the service throughout the network is a challenge.</t>
    </section>

    <section anchor="security-considerations" title="Security Considerations">
      <t>TBD.</t>
    </section>

    <section anchor="iana-considerations" title="IANA Considerations">
      <t>TBD.</t>
    </section>

    <section title="Contributors">
      <t>The following people have substantially contributed to this
      document:</t>

      <t><figure>
          <artwork>
	Peter Willis
	BT

	Tianji Jiang
	China Mobile
	tianjijiang@chinamobile.com

	Markus Amend
	Deutsche Telekom
	Markus.Amend@telekom.de

	Guangping Huang
	ZTE
	huang.guangping@zte.com.cn</artwork>
        </figure></t>
    </section>
  </middle>

  <back>
    <references title="Informative References">
      <?rfc include="reference.RFC.4786"?>

      <?rfc include="reference.I-D.contreras-alto-service-edge"?>

      <reference anchor="TR22.874">
        <front>
          <title>Study on traffic characteristics and performance requirements
          for AI/ML model transfer in 5GS (Release 18)</title>

          <author fullname="3GPP" surname="">
            <organization>3GPP</organization>
          </author>

          <date year="2021"/>
        </front>
      </reference>

      <reference anchor="TR-466">
        <front>
          <title>TR-466 Metro Compute Networking: Use Cases and High Level
          Requirements</title>

          <author fullname="BBF" surname="">
            <organization>Broadband Forum</organization>
          </author>

          <date year="2021"/>
        </front>
      </reference>

      <reference anchor="HORITA">
        <front>
          <title>Extended electronic horizon for automated driving",
          Proceedings of 14th International Conference on ITS
          Telecommunications (ITST)</title>

          <author fullname=" Y. Horita" surname="">
            <organization/>
          </author>

          <date year="2015"/>
        </front>
      </reference>

      <reference anchor="Industry4.0">
        <front>
          <title>Details of the Asset Administration Shell, Part 1 &amp; Part
          2</title>

          <author fullname="Industry4.0" surname="">
            <organization>Industry4.0</organization>
          </author>

          <date year="2020"/>
        </front>
      </reference>

      <reference anchor="GAIA-X">
        <front>
          <title>"GAIA-X: A Federated Data Infrastructure for Europe"</title>

          <author fullname="Gaia-X" surname="">
            <organization/>
          </author>

          <date year="2021"/>
        </front>
      </reference>

      <reference anchor="MEC">
        <front>
          <title>"Multi-Access Edge Computing (MEC)"</title>

          <author fullname="ETSI" surname="">
            <organization>ETSI</organization>
          </author>

          <date year="2021"/>
        </front>
      </reference>
    </references>

    <section anchor="acknowledgements" numbered="no" title="Acknowledgements">
      <t>The author would like to thank Luigi IANNONE, Christian Jacquenet,
      Kehan Yao and Yuexia Fu for their valuable suggestions to this
      document.</t>
    </section>
  </back>
</rfc>
