hash
stringlengths
32
32
doc_id
stringlengths
5
12
section
stringlengths
4
595
content
stringlengths
0
6.67M
5b09ab2cb31f10c2a9afae8755e36a25
22.011
3.2.2.6 Investigation PLMN Scan
The operator shall be able to control by SIM/USIM configuration whether a UE that is capable shall perform an investigation scan. This investigation scan shall be performed after each successful PLMN selection as well as during limited service state. The investigation scan shall search for a higher prioritised PLMN that does not offer CS voice service. If such a PLMN is available, the user shall be informed. This enables the user to switch to such a PLMN using manual selection if the user so prefers. The investigation scan shall not be performed when no SIM/USIM is inserted.
5b09ab2cb31f10c2a9afae8755e36a25
22.011
3.2.2.7 Void
5b09ab2cb31f10c2a9afae8755e36a25
22.011
3.2.2.8 Steering Of Roaming
Steering to a specific VPLMN It shall be possible for the HPLMN at any time to direct a UE, that is in automatic mode, to search for a specific VPLMN and, if it is available, move to that VPLMN as soon as possible. This VPLMN shall then be regarded as the highest priority VPLMN as defined by the operator, though any EHPLMN or PLMN on the User Controlled PLMN list shall have higher priority. This process shall be done transparently and without inconvenience to the user. If the Operator controlled signal threshold per access technology is set on the USIM, the UE shall only select the VPLMN if the signal quality is equal to or higher than the Operator controlled signal threshold per access technology, otherwise the steering request shall be ignored. If the UE is in manual mode, the steering request shall be ignored. If the UE is registered on a VPLMN that is present on the User Controlled PLMN List, the steering request shall be ignored. PLMNs contained on the User Controlled PLMN List shall have priority over the steered-to-PLMN. The UE shall attempt to register on the specified VPLMN even if the specified VPLMN is present on a Forbidden List. This mechanism shall be available to the HPLMN even if the VPLMN the UE is registered on is compliant to an earlier release of the 3GPP specifications. VPLMN Redirection It shall be possible for the HPLMN to request a UE, that is in automatic mode, to find and register on a different VPLMN from the one it is currently using or trying to register on, if another VPLMN, that is not in a Forbidden List, is available. The original VPLMN shall then be treated as the lowest priority VPLMN and would not be selected by the UE unless it is the only one available to the UE or has been selected in manual mode. This process shall be done transparently and without inconvenience to the user. If the UE is in manual mode, the redirection request shall be ignored. If the UE is registered on a VPLMN that is present on the User Controlled PLMN List, the redirection request shall be ignored. This mechanism shall be available to the HPLMN even if the VPLMN the UE is registered on is compliant to an earlier release of the 3GPP specifications.
5b09ab2cb31f10c2a9afae8755e36a25
22.011
3.2.3 Network selection for Multi-mode terminals with 3GPP Capability
Different types of Multi-mode terminals combining different technologies and systems in one terminal can be produced. It is not possible to foresee all possible configurations and provide a detailed technical specification for network and system selection for all possible multi-mode terminal configurations. The following provides the generic requirements for network and system selection for Multi-mode terminals with 3GPP Capability. These requirements are mandatory for a 3GPP capable multi-mode terminal, unless otherwise is explicitly specified elsewhere in the 3GPP Technical Specifications. - a multi-mode terminal, when in 3GPP mode of operation shall be compliant to the 3GPP specifications, including PLMN selection, cell selection and re-selection, paging reception etc., - As consequence, the multi-mode terminal when entering 3GPP mode of operation shall act as if it were a 3GPP only UE which had just been switched-on. Similarly, when leaving the 3GPP mode of operation the multimode terminal shall act as if it were a 3GPP only UE which had just been switched-off When the multimode terminal is in 3GPP mode, the switching between modes in the multi-mode terminal is considered an overlay functionality selecting mode of operation. For the design of the overlay functionality the following requirements shall be fulfilled: - The overlay functionality shall include a mechanism to avoid ping-pong between systems, e.g., a timer or hysteresis function; - The overlay functionality shall not include network priority mechanisms, which conflict with the network priority mechanisms specified in 3GPP specifications, e.g., the Periodic network selection attempts scanning within 3GPP based systems for PLMNs of higher priority than the current serving PLMN; - Any functionality in the overlay system, such as background scan of other systems, shall not impact the fulfilment of 3GPP protocol requirements (in particular in regard to paging, cell selection, cell re-selection and PLMN selection); - As specified in this technical specification, the 3GPP technical specification provides the capability for the user to set their own 3GPP PLMN selection preferences; as well as the user can manually select any 3GPP PLMN. This has been done to ensure a fair competition environment. These principles shall be maintained in the design of the overlay functionality.
5b09ab2cb31f10c2a9afae8755e36a25
22.011
4 Access control
5b09ab2cb31f10c2a9afae8755e36a25
22.011
4.1 Purpose
Under certain circumstances, it will be desirable to prevent UE users from making access attempts (including emergency call attempts) or responding to pages in specified areas of a PLMN. Such situations may arise during states of emergency, or where 1 of 2 or more co-located PLMNs has failed. Broadcast messages should be available on a cell by cell basis indicating the class(es) or categories of subscribers barred from network access. The use of these facilities allows the network operator to prevent overload of the access channel under critical conditions. It is not intended that access control be used under normal operating conditions. It should be possible to differentiate access control between CS and PS domains. Details are specified in TS23.122[3] and TS25.304 [10]. Not all RATs need to support this functionality.
5b09ab2cb31f10c2a9afae8755e36a25
22.011
4.2 Allocation
All UEs are members of one out of ten randomly allocated mobile populations, defined as Access Classes 0 to 9. The population number is stored in the SIM/USIM. In addition, UEs may be members of one or more out of 5 special categories (Access Classes 11 to 15), also held in the SIM/USIM. These are allocated to specific high priority users as follows. (The enumeration is not meant as a priority sequence): Class 15 - PLMN Staff; -"- 14 - Emergency Services; -"- 13 - Public Utilities (e.g. water/gas suppliers); -"- 12 - Security Services; -"- 11 - For PLMN Use.
5b09ab2cb31f10c2a9afae8755e36a25
22.011
4.3 Operation
5b09ab2cb31f10c2a9afae8755e36a25
22.011
4.3.1 Access Class Barring
If the UE is a member of at least one Access Class which corresponds to the permitted classes as signalled over the radio interface, and the Access Class is applicable in the serving network, access attempts are allowed. Additionally, in the case of the access network being UTRAN the serving network can indicate that UEs are allowed to respond to paging and perform location registration (see, sec 3.1), even if their access class is not permitted. Otherwise access attempts are not allowed. Also, the serving network can indicate that UEs are restricted to perform location registration, although common access is permitted. If the UE responded to paging it shall follow the normal defined procedures and react as specified to any network command. NOTE: The network operator can take the network load into account when allowing UEs access to the network. Access Classes are applicable as follows: Classes 0 - 9 - Home and Visited PLMNs; Classes 11 and 15 - Home PLMN only if the EHPLMN list is not present or any EHPLMN; Classes 12, 13, 14 - Home PLMN and visited PLMNs of home country only. For this purpose the home country is defined as the country of the MCC part of the IMSI. Any number of these classes may be barred at any one time. In the case of multiple core networks sharing the same access network, the access network shall be able to apply Access Class Barring for the different core networks individually. The following are the requirements for enhanced Access control on E-UTRAN. - The serving network shall be able to broadcast mean durations of access control and barring rates (e.g. percentage value) that commonly applied to Access Classes 0-9 to the UE. The same principle as in UMTS is applied for Access Classes 11-15; - E-UTRAN shall be able to support access control based on the type of access attempt (i.e. mobile originating data or mobile originating signalling), in which indications to the UEs are broadcasted to guide the behaviour of UE. E-UTRAN shall be able to form combinations of access control based on the type of access attempt e.g. mobile originating and mobile terminating, mobile originating, or location registration. The ‘mean duration of access control’ and the barring rate are broadcasted for each type of access attempt (i.e. mobile originating data or mobile originating signalling); - The UE determines the barring status with the information provided from the serving network, and perform the access attempt accordingly. The UE draws a uniform random number between 0 and 1 when initiating connection establishment and compares with the current barring rate to determine whether it is barred or not. When the uniform random number is less than the current barring rate and the type of access attempt is indicated allowed, then the access attempt is allowed; otherwise, the access attempt is not allowed. If the access attempt is not allowed, further access attempts of the same type are then barred for a time period that is calculated based on the ‘mean duration of access control’ provided by the network and the random number drawn by the UE; - The serving network shall be able to indicate whether or not a UE shall apply Access Class Barring for SMS access attempts in SMS over SGs, SMS over IMS (SMS over IP), and SMS over S102. This indication is valid for Access Classes 0-9 and 11-15; - The serving network shall be able to indicate whether or not a UE shall apply Access Class Barring for MMTEL voice access attempts. This indication is valid for Access Classes 0-9 and 11-15; - The serving network shall be able to indicate whether or not a UE shall apply Access Class Barring for MMTEL video access attempts. This indication is valid for Access Classes 0-9 and 11-15.
5b09ab2cb31f10c2a9afae8755e36a25
22.011
4.3.2 Service Specific Access Control
Additionally to the above requirements in 4.3.1; - In E-UTRAN it shall be possible to support a capability called Service Specific Access Control (SSAC) to apply independent access control for telephony services (MMTEL) for mobile originating session requests from idle-mode and connected-mode as following: - The serving network shall be able to indicate (as specified in sub-clause 4.3.1) whether or not a UE subject to SSAC shall also apply Access Class Barring. - EPS shall provide a capability to assign a service probability factor [13] and mean duration of access control for each of MMTEL voice and MMTEL video: - assign a barring rate (percentage) commonly applicable for Access Classes 0-9; - assign a flag barring status (barred /unbarred) for each Access Class in the range 11-15; - SSAC shall not apply to Access Class 10; - SSAC can be provided by the VPLMN based on operator policy without accessing the HPLMN; - SSAC shall provide mechanisms to minimize service availability degradation (i.e. radio resource shortage) due to the mass simultaneous mobile originating session requests and maximize the availability of the wireless access resources for non-barred services; - The serving network shall be able to broadcast mean durations of access control, barring rates for Access Classes 0-9, barring status for Access class in the range 11-15 to the UE; - The UE determines the barring status with the information provided from the serving network, and perform the access attempt accordingly. The UE draws a uniform random number between 0 and 1 when initiating connection establishment and compares with the current barring rate to determine whether it is barred or not. When the uniform random number is less than the current barring rate and the type of access attempt is indicated allowed, then the access attempt is allowed; otherwise, the access attempt is not allowed. If the access attempt is not allowed, further access attempts of the same type are then barred for a time period that is calculated based on the ‘mean duration of access control’ provided by the network and the random number drawn by the UE.
5b09ab2cb31f10c2a9afae8755e36a25
22.011
4.3.3 Access Control for CSFB
Access control for CSFB provides a mechanism to prohibit UEs to access E-UTRAN to perform CSFB. It minimizes service availability degradation (i.e. radio resource shortage, congestion of fallback network) caused by mass simultaneous mobile originating requests for CSFB and increases the availability of the E-UTRAN resources for UEs accessing other services. When an operator determines that it is appropriate to apply access control for CSFB, the network may broadcast necessary information to provide access control for CSFB for each class to UEs in a specific area. The network shall be able to separately apply access control for CSFB, SSAC and enhanced Access control on E-UTRAN. The following requirements apply for CSFB to 1xRTT: - In E-UTRAN, the network may apply access control for mobile originating session requests on CSFB from 1xRTT/E-UTRAN UE, The parameters received by the UE are dealt with in accordance with CDMA2000 procedures in 3GPP2 C.S0004-A: "Signaling Link Access Control (LAC) Standard for cdma2000 Spread Spectrum Systems – Addendum 2" [15]. For CSFB to UTRAN or GERAN, the necessary information in the broadcast to provide access control for CSFB is the same as that specified in Clause 4.3.1. In addition to those requirements the following apply: - Access control for CSFB shall apply for Access Class 0-9 and Access Class 11-15.It shall not apply for Access Class 10; - Access control for CSFB shall be applied for idle mode UE; - Access control for CSFB shall apply to all CSFB services; - Access control for CSFB may be provided by the VPLMN based on operator policy without accessing the HPLMN; - If Access control for CSFB, according to the UE's access class, disallows originating session requests for CSFB then a UE shall not send mobile originating session requests for CSFB; - If Access control for CSFB is applied by the UE for a mobile originating session request for CSFB, the UE shall bypass enhanced Access control on E-UTRAN for that session. - The criteria on which a UE determines whether Access control for CSFB allows or disallows originating session requests for CSFB are equivalent to those for enhanced Access control on E-UTRAN, as described in clause 4.3.1; - If access is not granted for the UE, mobile originating session requests for CSFB shall be restricted for a certain period of time to avoid overload of E-UTRAN due to continuous mobile originating session requests from the same UE. The duration of the period shall be determined using the same operation specified in Clause 4.3.1; - In case the network does not provide the Access control for CSFB information, the UE shall be subject to access class barring for Access Classes 0-9 and 11-15 as described in Clause 4.3.1.
5b09ab2cb31f10c2a9afae8755e36a25
22.011
4.3.4 Extended Access Barring
5b09ab2cb31f10c2a9afae8755e36a25
22.011
4.3.4.1 General
Extended Access Barring (EAB) is a mechanism for the operator(s) to control Mobile Originating access attempts from UEs that are configured for EAB in order to prevent overload of the access network and/or the core network. In congestion situations, the operator can restrict access from UEs configured for EAB while permitting access from other UEs. UEs configured for EAB are considered more tolerant to access restrictions than other UEs. When an operator determines that it is appropriate to apply EAB, the network broadcasts necessary information to provide EAB control for UEs in a specific area. The following requirements apply for EAB: • The UE is configured for EAB by the HPLMN; - EAB shall be applicable to all 3GPP Radio Access Technologies used in EPS or UMTS; NOTE: When a UE configured for EAB is located in 5G system, the UE is considered as configured for delay tolerant service. - EAB shall be applicable regardless of whether the UE is in a Home or a Visited PLMN; - A network may broadcast EAB information; - EAB information shall define whether EAB applies to UEs within one of the following categories: a) UEs that are configured for EAB; b) UEs that are configured for EAB and are neither in their HPLMN nor in a PLMN that is equivalent to it; c) UEs that are configured for EAB and are neither in the PLMN listed as most preferred PLMN of the country where the UE is roaming in the operator-defined PLMN selector list on the SIM/USIM, nor in their HPLMN nor in a PLMN that is equivalent to their HPLMN. - EAB information shall also include extended barring information for Access Classes 0-9; - A UE configured for EAB shall use its allocated Access Class(es), as defined in sub-clause 4.2, when evaluating the EAB information that is broadcast by the network, in order to determine if its access to the network is barred; - If a UE that is configured for EAB answers to paging, initiates an emergency call or, is a member of an Access Class in the range 11-15 and according to clause 4.3.1 that Access Class is permitted by the network, then the UE shall ignore any EAB information that is broadcast by the network; - If the network is not broadcasting the EAB information, the UE shall be subject to access barring as described in clause 4.3.1; - If the EAB information that is broadcast by the network does not bar the UE, the UE shall be subject to access barring as described in clause 4.3.1; - In the case of multiple core networks sharing the same access network, the access network shall be able to apply the EAB for the different core networks individually.
5b09ab2cb31f10c2a9afae8755e36a25
22.011
4.3.4.2 Overriding extended access barring
Overriding Extended Access Barring is a mechanism for the operator to allow UEs that are configured for EAB to access the network under EAB conditions. The following requirements apply: - The UE configured with EAB may be configured by the HPLMN with a permission to override EAB; - For a UE configured with the permission to override EAB, the user or application (upper layers in UE) may request the UE to activate PDN connection(s) for which EAB does not apply; - The UE shall override any EAB restriction information that is broadcast by the network as long as it has an active PDN connection for which EAB does not apply.
5b09ab2cb31f10c2a9afae8755e36a25
22.011
4.3.5 Application specific Congestion control for Data Communication (ACDC)
5b09ab2cb31f10c2a9afae8755e36a25
22.011
4.3.5.1 Service description
Application specific Congestion control for Data Communication (ACDC) is an access control mechanism for the operator to allow/prevent new access attempts from particular, operator-identified applications in the UE in idle mode. ACDC does not apply to UEs in connected mode. The network can prevent/mitigate overload of the access network and/or the core network. This feature is optional.
5b09ab2cb31f10c2a9afae8755e36a25
22.011
4.3.5.2 Requirements
5b09ab2cb31f10c2a9afae8755e36a25
22.011
4.3.5.2.1 General
The following requirements apply: - This feature shall be applicable to UTRAN PS Domain and E-UTRAN; - This feature shall be applicable to UEs in idle mode only that are not a member of one or more of Access Classes 11 to 15; - ACDC shall not apply to MMTEL voice, MMTEL video, SMS over IMS (SMS over IP) emergency call and paging response; NOTE 1: Even if any of the above services are initiated by application, ACDC does not apply, but they are subject to other applicable access control methods. - The home network shall be able to configure a UE with at least four and a maximum of sixteen ACDC categories to each of which particular, operator-identified applications are associated. The categories shall be ordered as specified in sub-clause 4.3.5.2.2; NOTE 2: Provisioning of the ACDC categories in the UE is the responsibility of the home network, and the categorization is outside the scope of 3GPP. NOTE 3: A mechanism needs to be provided that enables the UE to verify that the provisioning of the configuration originates from a trusted source. - The serving network shall be able to broadcast, in one or more areas of the RAN, control information, indicating barring information per each ACDC category, and whether a roaming UE shall be subject to ACDC control; NOTE 4: The barring information may be similar to ACB information, and include mean durations of access control (i.e., barring timer) and barring rates (i.e., percentage value). If the barring timer is running due to a previous access attempt from an application in a certain given matched ACDC category, the UE may only allow access attempts from applications in higher ACDC categories (according to the corresponding barring information for those higher categories). If the barring timer is running due to a previous access attempt from an application in a certain given unmatched ACDC category or a uncategorised application, the UE may only allow access attempts from applications in higher ACDC categories than the lowest ACDC category broadcast (according to the corresponding barring information for those higher categories). - The UE shall be able to control whether or not an access attempt for a certain application is allowed, based on this broadcast barring information and the configuration of ACDC categories in the UE; - The serving network shall be able to simultaneously indicate ACDC with other forms of access control; - When both ACDC and ACB controls are indicated, ACDC shall override ACB; - If a UE is configured for both EAB and ACDC, and the serving network simultaneously broadcasts EAB information and ACDC barring information: - If the UE determines as specified in sub-clause 4.3.4.1 that access to the network is not barred or as specified in sub-clause 4.3.4.2 that it is permitted to override an EAB restriction, then access to the network is subject to ACDC; - If the UE determines as specified in sub-clause 4.3.4.1 that access to the network is barred and as specified in sub-clause 4.3.4.2 that it is not permitted to override the EAB restriction, then access to the network is barred. - In the case of multiple core networks sharing the same access network, the access network shall be able to apply ACDC for the different core networks individually. For the mitigation of congestion in a shared RAN, barring rates should be set equal for all Participating Operators.
5b09ab2cb31f10c2a9afae8755e36a25
22.011
4.3.5.2.2 ACDC Categories
When configuring the UE with categories of applications, the home network shall proceed as follows: - Applications whose use is expected to be restricted the least shall be assigned the highest ACDC category; and - Applications whose use is expected to be restricted more than applications in the highest category shall be assigned the second-to-highest ACDC category, and so on; and - Applications whose use is expected to be restricted the most shall either be assigned the lowest ACDC category, or not be categorised at all. For a UE with ACDC categories configured, the applications on the UE that are not assigned to any ACDC category shall be treated by the UE as part of the lowest ACDC category broadcast by the serving network. If the operator requires differentiation with respect to these uncategorized applications, the operator should avoid assigning applications to the lowest ACDC category. When applying ACDC, the serving network broadcasts barring information starting from the highest to the lowest ACDC category. The home network and the serving network may use different categorisation. The serving network decides if ACDC applies to roaming UEs. The number of ACDC categories in the UE may not be the same as the number of ACDC categories broadcast by the serving network. This may happen, e.g. when the UE is roaming and the number of categories broadcast by the serving network is different from the home network. Therefore, the following rules shall apply: - If the serving network broadcasts more ACDC categories than the UE's configuration, the UE shall use barring information for the matching ACDC category and shall bar uncategorised applications using the barring information for the lowest category broadcast by the serving network, and shall ignore barring information for unmatched categories. - If the serving network broadcasts barring information for fewer ACDC categories than the UE's configuration, the UE shall use barring information for the matching ACDC category and shall bar other applications using the barring information for the lowest category broadcast by the serving network. NOTE: A matching ACDC category is an ACDC category for which barring information is broadcast by the serving network and that has the same rank as the rank of a configured ACDC category in the UE. An unmatched ACDC category is either an ACDC category for which barring information is broadcast by the serving network but with no corresponding ACDC category configured in the UE, or an ACDC category configured in the UE but with no corresponding barring information broadcast by the serving network.
5b09ab2cb31f10c2a9afae8755e36a25
22.011
4.3.6 Access Control for Indirect 3GPP Communications
For the case where an Evolved ProSe Remote UE is trying to access the network via an Evolved ProSe UE-to-Network Relay, the following requirements apply: - Access control for CSFB shall not apply to an Indirect 3GPP Communication; - ACB, EAB and ACDC shall apply when the Evolved ProSe UE-to-Network Relay is in IDLE mode and when the access control information for the respective access control is broadcast by the network; - SSAC shall apply when the Evolved ProSe UE-to-Network Relay is either in IDLE or CONNECTED mode and when the access control information for SSAC is broadcast by the network; - The interaction between ACDC, EAB and ACB (defined in paragraph 4.3.5.2.1) shall be based on the Evolved ProSe Remote UE’s configuration; - During the access control procedure, the requirements in subsections of 4.3.1, 4.3.2, 4.3.4 and 4.3.5 shall be performed by Evolved Prose Remote UE using the following parameters: - Access control parameters broadcast by the cell where the Evolved ProSe UE-to-Network Relay is camped on or connected to shall be used in all access control procedures; - Access Class of the Evolved ProSe Remote UE shall be used; - For ACB, the type of access attempt (i.e. mobile originating data, mobile originating signalling, response to paging or emergency call) of the Evolved ProSe Remote UE shall be used; - For EAB, the following information shall be used: - Whether or not the Evolved ProSe Remote UE is configured with EAB; - Whether or not the Evolved ProSe Remote UE is configured with a permission to override EAB; - Whether the PLMN is the Evolved ProSe Remote UE’s Home PLMN, equivalent PLMN or preferred PLMN. The PLMN where the Evolved ProSe UE-to-Network Relay is camped/connected shall be used to make this determination. - For SSAC, the type of access attempt (MMTEL voice or MMTEL video) of the Evolved ProSe Remote UE shall be used; - For ACDC, the ACDC category of the application that triggered the access attempt in the Evolved ProSe Remote UE shall be used.
5b09ab2cb31f10c2a9afae8755e36a25
22.011
4.4 Emergency Calls
An additional control bit known as "Access Class 10" is also signalled over the radio interface to the UE. This indicates whether or not network access for Emergency Calls is allowed for UEs with access classes 0 to 9 or without an IMSI. For UEs with access classes 11 to 15, Emergency Calls are not allowed if both "Access class 10" and the relevant Access Class (11 to 15) are barred. Otherwise, Emergency Calls are allowed. 4.4a Multimedia Priority Service Multimedia Priority Service (TS 22.153 [16]) shall be assigned its own unique access class value (i.e., one of the special access classes 11 to 15). The assigned access class value for Multimedia Priority Service is based on regional/national regulatory requirements and operator policy.
5b09ab2cb31f10c2a9afae8755e36a25
22.011
4.5 Control of UE Capabilities
To protect the user from the effects of a misbehaving UE (e.g. causing additional charges, degraded performance) and to protect the network operator's network capacity, including radio resources and network signalling and processing, means shall be provided for the HPLMN/EHPLMN and the VPLMN to provide an indication to the UE as to which network provided services or functions it is not allowed to use. The Selective UE Capabilities list shall be maintained in the UE and the UE shall not request any services indicated as disabled. At registration the HPLMN/EHPLMN or VPLMN may interrogate the status of the list and provide a new list. The Selective UE Capabilities list shall not be deleted at switch off and will remain valid until a new list is provided by the network. The Selective UE Capabilities list relates to the ME and not to the subscription. It should be ensured that UEs are not maliciously disabled, including malicious disabling by a VPLMN, or accidentally disabled, or kept disabled, and there shall be a mechanism for restoring disabled UEs in all situations (e.g. in the case that the serving network does not support the control of UE Capabilities). The UE should use the indications given in the Selective UE Capabilities list to inform the user of the non-availability of services or functions. There shall be a means for the network to provide an optional customer service number(s) which can be used, by the user, to assist in determining the cause of non-availability of specific services. The specifications should also provide the capability for the network to include an optional text string that will be displayed by the UE. The UE Capabilities list shall take precedence over subscribed services. The services to be included in the list are: - Call Control functions; - Supplementary Services; - Emergency Calls (including the (U)SIM-less case and subject to regional regulatory requirements, i.e. emergency calls shall not be disabled in regions where support of them is required); - SMS, via CS and PS; - LCS, via CS and PS; - GPRS based services; - MBMS; - IMS.
5b09ab2cb31f10c2a9afae8755e36a25
22.011
4.6 Prevention of mobile-originating signalling and/or data traffic
The network shall be able to control the behavior of UEs in E-UTRAN in connected mode to prevent mobile originating signalling and/or data traffic, while the access barring mechanisms specified under Clause 4.3 are being applied to UEs in idle mode.
5b09ab2cb31f10c2a9afae8755e36a25
22.011
5 Support of Localised Service Area (SoLSA)
SoLSA consists of a set of service features that give the operator a basis to offer subscribers different services (e.g. tariffs or access rights) depending on the location of the subscriber. (3GPP TS 22.043 [5]). The following section is only applicable to the support of SoLSA functionality in GERAN.
5b09ab2cb31f10c2a9afae8755e36a25
22.011
5.1 Network selection
The standard automatic and manual network selection procedures will be used. Manual network selection may be required when the PLMN providing the users SoLSA service is not the one on which the user is currently registered. At manual network selection the UE shall provide the means to present the subscribers LSA(s) for each PLMN presented.
5b09ab2cb31f10c2a9afae8755e36a25
22.011
5.2 The Idle-mode operation
The UE shall always select a valid LSA with the highest priority.
5b09ab2cb31f10c2a9afae8755e36a25
22.011
5.2.1 Subscriber moving from a normal environment to his localised service area
The UE shall have the ability to prioritise allowed LSA cells in reselection, making it possible to camp on a LSA cell earlier (the function shall be network controlled).
5b09ab2cb31f10c2a9afae8755e36a25
22.011
5.2.2 Subscriber moving away from his localised service area to a normal environment
The UE shall have the ability to prioritise allowed LSA cells in reselection, making it possible to camp on a LSA cell longer (the function shall be network controlled).
5b09ab2cb31f10c2a9afae8755e36a25
22.011
5.2.3 Subscriber staying in his localised service area
The UE shall have the ability to prioritise allowed LSA cells in reselection by being more persistent (the function shall be network controlled). NOTE: Typically in indoor environments there are occasional reflections and "disturbances" due to macro cells, e.g. near the windows. In such a case LSA cells should be favoured even though there is higher field strength available from the outdoor cells.
5b09ab2cb31f10c2a9afae8755e36a25
22.011
5.3 LSA only access
It shall be possible to allow LSA user to access PLMN only within his LSAs. A LSA user is not allowed to receive and/or originate a call outside the allowed LSA area. When UE is out of the allowed LSA area it shall be registered in PLMN but indicate subscriber/service specific "out of LSA area" notification. It shall be a network-controlled function to prevent terminated or/and originated calls. Emergency calls are however always allowed.
5b09ab2cb31f10c2a9afae8755e36a25
22.011
5.4 Exclusive access
Access to exclusive access cells is restricted to defined LSA subscribers. Non-LSA subscriber shall consider exclusive access cells as not suitable, only allowing to camp for emergency calls (limited service state 3GPP TS 23.122 [3]).
5b09ab2cb31f10c2a9afae8755e36a25
22.011
5.5 Preferential access
As a network-controlled function, it shall be possible in LSA to allocate resources at call setup and during the active mode to LSA users compared to non-LSA users.
5b09ab2cb31f10c2a9afae8755e36a25
22.011
6 Support of 3GPP - WLAN Interworking
Support of 3GPP-WLAN interworking and network selection is captured in TS 22.234 [6] NOTE: The requirement specification 22.234 is no longer maintained from Release 12 onwards.
5b09ab2cb31f10c2a9afae8755e36a25
22.011
6.1 Void
5b09ab2cb31f10c2a9afae8755e36a25
22.011
6.2 Void
5b09ab2cb31f10c2a9afae8755e36a25
22.011
6.3 Void
7 Administrative restriction of subscribers’ access
5b09ab2cb31f10c2a9afae8755e36a25
22.011
7.1 Allowed Location and Routing Area identities access
Means shall be standardised for an administrative restriction of subscribers’ access without the need of having explicit Tracking/Location/Routing Area identities in the individual subscription profiles. To achieve this it shall be possible to indicate per subscriber, in subscription data, allowed categories of Tracking/ Location/Routing Areas. It shall be possible to use this subscription information to restrict subscribers’ access to categories of Tracking/Location/Routing Areas in serving networks accordingly. As a minimum, at least one of the following categories shall be available: a) GERAN; b) UTRAN; c) E-UTRAN; d) NG-RAN. There might be cases where the visited network has not separated the Location/Routing area categories, in which case the administrative restriction of subscribers' access to only GERAN or UTRAN will not be possible. In EPS, an operator may introduce subscriptions supporting the 5G NR Dual Connectivity in E-UTRAN. It shall be possible to indicate in subscription data that a subscriber´s access to the 5G NR Dual Connectivity in E-UTRAN service is restricted. This administrative restriction of subscribers’ access shall be an optional feature.
5b09ab2cb31f10c2a9afae8755e36a25
22.011
7.2 Void
5b09ab2cb31f10c2a9afae8755e36a25
22.011
7.3 UE configured radio technology restriction
A UE shall support a Man Machine Interface setting for the user to disable use of one or more of the UE’s radio technologies, regardless of PLMNs. Radio technologies that individually can be disabled is dependent on supported radio technology of the UE such as GSM/EDGE, WCDMA, E-UTRA, and NR. A UE shall support a Man Machine Interface setting enabling the user to re-enable use of one or more of the ME’s radio technologies for access to a radio access network, regardless of PLMNs. The user can only re-allow a radio technology that the user has previously disallowed. NOTE: The described MMI user setting is a proprietary function of most legacy UE products to allow a user of a UE to change the radio capabilities of the UE. Legacy radio technologies may lack means to mitigate some security attack. If severe enough, the home operator may want to disallow their subscribers to access a radio access network with such radio technology. This configuration of the UE is valid for all PLMNs. A UE shall support a secure mechanism for the home operator to disallow selection of one or more of the UE’s radio technologies for access to a radio access network, regardless of PLMNs. Radio technologies that individually can be disallowed are at least GSM/EDGE, WCDMA, E-UTRA, and NR. A UE shall support a secure mechanism for the home operator to re-allow selection of one or more of the UE’s radio technologies for access to a radio access network, regardless of PLMNs. Radio technologies that individually can be re-allowed are at least GSM/EDGE, WCDMA, E-UTRA, and NR. The home operator can only re-allow a radio technology that the home operator has previously disallowed. For a prioritized service (e.g., Emergency Services, MPS, Mission Critical Services), the UE shall support a mechanism to automatically override user and network disallowed RATs when there are no PLMNs on the allowed radio technologies identified that the UE is able to access. Upon power-cycle or when the USIM is disabled, the UE configuration of enabled/disabled radio technologies configured by the user shall remain as it was before such events happen. The radio technologies disallowed by the HPLMN shall remain as it was before a power cycle. The radio technologies disallowed by the HPLMN shall be bound to the USIM
5b09ab2cb31f10c2a9afae8755e36a25
22.011
8 Support of Home NodeB and Home E-NodeB
The service requirements for Home NodeB and Home eNodeB can be found in [14]. 9 Control of traffic from UE-based applications toward associated server
5b09ab2cb31f10c2a9afae8755e36a25
22.011
9.1 Description
This feature allows operators to control traffic from UEs to an application on a third-party server or the third-party server itself. When an application on a third-party server or the third-party server itself becomes congested or fails the traffic towards that server need to be controlled to avoid/mitigate potential issues caused by resulting unproductive use of 3GPP network resource. This will also make it possible to allow 3GPP network to help third party servers to handle overload and recover from failures (see [7]).
5b09ab2cb31f10c2a9afae8755e36a25
22.011
9.2 Requirements
Under network control, the UE shall be able to control (i.e. block and/or prioritize) traffic from UEs to an application on a third-party server or the third-party server itself without affecting traffic to other applications on the third-party server or to other third-party servers. For this purpose, the UE shall be able to identify the traffic towards the third-party server or the application on the third-party server. When congestion or failure of the application on the third-party server or the third-party server abates, the control shall be applied in a phased manner to gradually restore traffic according to operator policy. This feature and other forms of access control shall be able to be applied simultaneously. In case of simultaneous application of ACDC and this feature, this feature shall take precedence over ACDC to control traffic from applications in the UE, only when such traffic is allowed by ACDC. This feature shall be applicable both to new connections being set up and to existing connections from the UE towards an application on a third-party server or the third-party server. 10 3GPP PS Data Off
5b09ab2cb31f10c2a9afae8755e36a25
22.011
10.1 Description
3GPP PS Data Off is a feature which, when configured by the HPLMN and activated by the user, prevents transport via PDN connections or PDU sessions in 3GPP access networks of all data packets except IP packets required by 3GPP PS Data Off Exempt Services.
5b09ab2cb31f10c2a9afae8755e36a25
22.011
10.2 Requirements
The 3GPP system shall provide a mechanism by which an operator can configure which operator services are defined as 3GPP PS Data Off Exempt Services for their own subscribers. When 3GPP PS Data Off is activated in the UE, in order to preserve charging consistency: - the UE shall inform the network that 3GPP PS Data Off is activated; - the UE shall cease the sending of uplink IP Packets of all services that are not 3GPP PS Data Off Exempt Services via PDN connections in 3GPP access networks; - the network shall cease the sending of downlink IP Packets to the UE for all services that are not 3GPP PS Data Off Exempt Services via PDN connections in 3GPP access networks; - the UE shall cease the sending of uplink traffic over non-IP PDN types via PDN connections in 3GPP access networks; and - the network shall cease the sending of downlink traffic over non-IP PDN types via PDN connections in 3GPP access networks; - the UE shall cease the sending of uplink IP packets of all services that are not 3GPP PS data Off Exempt Services via PDU sessions in 3GPP access networks; - the network shall cease the sending of downlink IP packets to the UE for all services that are not 3GPP PS data Off Exempt Services via PDU sessions in 3GPP access networks; - the UE shall cease the sending of uplink unstructured and ethernet data via PDU sessions in 3GPP access networks; - the network shall cease the sending of downlink unstructured and ethernet data via PDU sessions in 3GPP access networks. NOTE 1: Disabling of traffic on both the uplink and downlink is needed in order to provide consistency of charging between HPLMN and VPLMN, as well as consistency between what the user expects and what the user may be billed for. Each of the following operator services shall be configurable by the HPLMN operator to be part of the 3GPP PS Data Off Exempt Services: - MMTel Voice; - SMS over IMS; - USSD over IMS (USSI); - MMTel Video; - Services over IMS Data Channel; - Particular IMS services not defined by 3GPP, where each such IMS service is identified by an IMS communication service identifier; - Device Management over PS; - Management of USIM files over PS (e.g. via Bearer Independent Protocol); and - IMS Supplementary Service configuration via the Ut interface using XCAP. - User plane location services protocol (LCS-UPP). NOTE 1a: IMS Data Channel is defined in 3GPP TS 26.114[20]. NOTE 1b: Services over IMS Data Channel are treated as a whole to be exempt or not to be exempt. Usage of individual applications on the data channel can be controlled by operators when the PS Data Off Exempt of services over IMS Data Channel is on. NOTE 1c: 3GPP PS Data Off Exempt for LCS-UPP in roaming is subject to agreement between operators. 3GPP PS Data Off may be activated based on roaming status, and the HPLMN may configure up to two sets of 3GPP PS Data Off Exempt Services for its subscribers: one is used when in HPLMN and another when roaming. NOTE 2: The updating to the sets of configured 3GPP Data Off Exempt Services in the VPLMNs and HPLMN is not guaranteed to take effect in real time. The updating to the sets of configured 3GPP Data Off Exempt Services in the UEs is not guaranteed take effect in real time. The user should be made aware of the operator services that are 3GPP PS Data Off Exempt Services. NOTE 3: The system can support falling back to operate over the CS domain in case an operator service is not configured to be a 3GPP PS Data Off Exempt Service and an equivalent CS domain operator service exists for the operator service. Annex A (informative): Change history Change history TSG SA# SA Doc. SA1 Doc Spec CR Rev Rel Cat Subject/Comment Old New Work Item Jun 1999 02.04 Transferred to 3GPP SA1 7.0.0 3.0.0 SA#04 02.04 3.0.0 SP-05 SP-99479 S1-99610 22.011 0001 R99 D Editorial changes for alignment 3.0.0 3.0.1 SP-06 SP-99524 S1-991032 22.011 0002 R99 B COMPACT Cell Selection Part 2 3.0.1 3.1.0 SP-06 SP-99606 22.011 0003 1 R99 B Network Selection 3.0.1 3.1.0 SP-06 SP-99607 22.011 0004 1 R99 B Control of user preference field 3.0.1 3.1.0 SP-07 SP-000055 S1-000138 22.011 0012 R99 F Corrections to 22.011 3.1.0 3.2.0 SP-07 SP-000055 S1-000139 22.011 0013 R99 C Removal of "Home Environment Specific Network Selection Procedure" 3.1.0 3.2.0 SP-07 SP-000071 S1-000161 22.011 0014 R00 B Network Selection 3.1.0 4.0.0 SP-08 SP-000211 S1-000335 22.011 0016 R00 B Reselection attempts of GPRS terminals 4.0.0 4.1.0 SP-09 SP-000372 S1-000549 22.011 0018 R4 F Alignment with 23.122 on selection procedure 4.1.0 4.2.0 SP-11 SP-010036 S1-010245 22.011 0021 Rel-4 A CR to 02.11 on Roaming restrictions for GPRS (Release4) 4.2.0 4.3.0 GPRS SP-11 SP-010151 22.011 0023 Rel-4 A Equivalent handling of PLMNs with different PLMN codes 4.2.0 4.3.0 GSM / UMTS inter-working SP-12 SP-010244 S1-010495 22.011 0025 Rel-4 A Partial PLMN access restriction 4.3.0 4.4.0 SP-12 SP-010244 S1-010514 22.011 0027 Rel-4 A Periodic Network Selection Attempt improvement 4.3.0 4.4.0 SP-12 SP-010244 S1-010517 22.011 0029 Rel-4 A Default value for background scanning timer 4.3.0 4.4.0 SP-14 SP-010685 S1-011341 22.011 0031 Rel-4 A CR to 22.011 R4 'Interaction between equivalent PLMN list and periodic network selection attempts' 4.4.0 4.5.0 SP-14 SP-010684 S1-011334 22.011 0035 Rel-4 A CR to 22.011 R4 'Editorial improvements' 4.4.0 4.5.0 SP-14 SP-010688 S1-011336 22.011 0037 Rel-4 A CR to 22.011 R4 'Clarification on the UE behaviour when receiving a registration rejection' 4.4.0 4.5.0 SP-14 SP-010687 S1-011338 22.011 0039 Rel-4 A CR to 22.011 R4 'Simplification of the procedure for user PLMN reselection' 4.4.0 4.5.0 SP-14 SP-010686 S1-011339 22.011 0041 Rel-4 A CR to 22.011 R4 'Interaction between “equivalent PLMN” list and “Forbidden PLMN” list' 4.4.0 4.5.0 SP-14 SP-010757 --- 22.011 0042 2 Rel-4 A Interaction between ePLMN and manual mode 4.4.0 4.5.0 SP-15 SP-010158 --- 22.011 0045 1 Rel-4 A CR to 22.011 Rel-4: clarification of the term 'country' NOTE: special dispensation was given by SA #15 to allow some leeway on the position of the note. 4.5.0 4.6.0 SP-16 SP-020238 S1-021058 22.011 0046 Rel-4 F Editorial corrections on 22.011 4.6.0 4.7.0 CORRECT SP-16 SP-020267 S1-021043 22.011 Rel-5 Updated from Rel-4 to Rel5 4.7.0 5.0.0 SP-17 SP-020547 S1-021826 22.011 0049 Rel-5 A CR to 22.011 Rel 5 - correction to periodic PLMN scan 5.0.0 5.1.0 TEI4 SP-19 SP-030035 S1-030236 22.011 0050 - Rel-6 B Netshare CR to TS 22.011 5.1.0 6.0.0 NTShar-CR SP-20 SP-030249 S1-030504 22.011 0051 Rel-6 B Network Selection requirements for WLAN Interworking 6.0.0 6.1.0 WLAN-CR SP-20 SP-030248 S1-030532 22.011 0052 Rel-6 B Network selection in shared networks 6.0.0 6.1.0 NTshar-CR SP-22 SP-030774 - 22.011 0053 1 Rel-6 B Administrative restriction of subscribers’ access 6.1.0 6.2.0 TEI SP-23 SP-040088 S1-040200 22.011 0054 - Rel-6 C Periodic network selection attempts enhancement 6.2.0 6.3.0 TEI SP-23 SP-040089 S1-040257 22.011 0056 - Rel-6 F Priority usage of UICC parameters for I-WLAN 6.2.0 6.3.0 WLAN SP-23 SP-040101 S1-040260 22.011 0057 - Rel-6 D Extraction of redundant WLAN network selection information [– now in WLAN TS22.234] 6.2.0 6.3.0 WLAN SP-24 SP-040287 S1-040438 22.011 0058 - Rel-6 F Behaviour of Single Mode mobiles with regards to the use of access technology in the PLMN selector lists 6.3.0 6.4.0 TEI6 SP-24 SP-040287 S1-040442 22.011 0060 - Rel-6 F Use of access technology in Periodic Network Selection attempts 6.3.0 6.4.0 TEI6 SP-24 SP-040287 S1-040443 22.011 0061 - Rel-6 F Clarification on the use of the RAT during network selection 6.3.0 6.4.0 TEI6 SP-24 SP-040295 S1-040526 22.011 0064 - Rel-6 F Priority usage of UICC parameters for I-WLAN 6.3.0 6.4.0 WLAN-CR SP-24 SP-040299 S1-040445 22.011 0062 - Rel-7 F Multimode terminals with 3GPP capability 6.3.0 7.0.0 TEI7 SP-24 SP-040298 S1-040448 22.011 0063 - Rel-7 B Support of multiple HPLMN codes 6.3.0 7.0.0 TEI7 SP-27 SP-050063 S1-050233 22.011 0066 - Rel-7 B Addition of "Network Control of UE Capabilities" 7.0.0 7.1.0 Network Protection SP-27 SP-050063 S1-050245 22.011 0067 - Rel-7 B Reinstate CR016 for GPRS Terminal PLMN Reselection 7.0.0 7.1.0 TEI-4 SP-29 SP-050521 S1-050903 22.011 0068 - Rel-7 C Enhancement of the EHPLMN feature to allow load balancing 7.1.0 7.2.0 TEI7 SP-31 SP-060221 - 22.011 0069 2 Rel-7 C The Last RPLMN 7.2.0 7.3.0 NSP-CR SP-31 SP-060033 S1-060330 22.011 0070 - Rel-7 B Confirmed Roaming 7.2.0 7.3.0 NSP-CR SP-31 SP-060033 S1-060331 22.011 0071 - Rel-7 F Capability of UE to read USIM information for Network Selection 7.2.0 7.3.0 NSP-CR SP-31 SP-060033 S1-060332 22.011 0072 - Rel-7 C Displaying all available EHPLMNs in Manual Mode 7.2.0 7.3.0 NSP-CR SP-31 SP-060033 S1-060333 22.011 0073 - Rel-7 B Requirement for presentation of additional information in manual mode 7.2.0 7.3.0 NSP-CR SP-31 SP-060033 S1-060335 22.011 0074 - Rel-7 B Network selection mode at switch-on 7.2.0 7.3.0 NSP-CR SP-31 SP-060033 S1-060352 22.011 0075 - Rel-7 B Steering of Roaming 7.2.0 7.3.0 NSP-CR SP-32 SP-060311 S1-060630 22.011 0076 - Rel-7 C Clarification on confirmed roaming applicability 7.3.0 7.4.0 NSP-CR SP-32 SP-060311 S1-060632 22.011 0077 - Rel-7 C Incomplete PLMN list for manual mode information 7.3.0 7.4.0 NSP-CR SP-32 SP-060426 - 22.011 0078 1 Rel-7 C Exception in Manual network selection mode when HPLMN is available at power-on 7.3.0 7.4.0 NSP-CR SP-32 SP-060451 S1-060649 22.011 0079 3 Rel-7 C Interaction of steering of roaming with User controlled and forbidden PLMN list 7.3.0 7.4.0 NSP-CR SP-34 SP-060762 S1-061314 22.011 0080 - Rel-7 F Additional Information in Manual Mode 7.4.0 7.5.0 NSP-CR SP-35 SP-070119 S1-070182 22.011 0081 1 Rel-7 F Removal of "Confirmed Roaming" enhancement (section 3.2.2.7) 7.5.0 7.6.0 NSP-CR SP-35 SP-070119 S1-070294 22.011 0083 2 Rel-7 C Single EHPLMN Display Name in Manual Mode (CR to 22.011) 7.5.0 7.6.0 NSP-CR SP-35 SP-070129 S1-070285 22.011 0082 2 Rel-8 C Administrative restriction of subscribers’ access 7.6.0 8.0.0 SAE-R SP-36 SP-070474 22.011 0085 - Rel-8 A Clarification on User Controlled PLMNs in SoR 8.0.0 8.1.0 NSP-CR SP-38 SP-070855 S1-071842 22.011 0091 - Rel-8 C EPS Alignment - References and Definitions 8.1.0 8.2.0 AIPN-SAE-R SP-38 SP-070855 S1-071910 22.011 0092 1 Rel-8 C EPS Alignment - 3.2.2 Procedures 8.1.0 8.2.0 AIPN-SAE SP-38 SP-070855 S1-071911 22.011 0093 1 Rel-8 C EPS Alignment - 7 Administrative restriction of subscribersÆ access 8.1.0 8.2.0 AIPN-SAE SP-38 SP-070855 S1-071925 22.011 0099 1 Rel-8 B Additional requirement for location registration 8.1.0 8.2.0 AIPN-SAE SP-38 SP-070859 S1-071943 22.011 0098 2 Rel-8 F Correction to EHPLMN terminology and editorials 8.1.0 8.2.0 TEI8 SP-38 SP-070913 - 22.011 0090 2 Rel-8 A EHPLMN case added to the optimisation for automatic network selection 8.1.0 8.2.0 NSP-CR SP-38 SP-070934 - 22.011 0095 2 Rel-8 F DSAC requirement in UMTS 8.1.0 8.2.0 TEI8 SP-38 SP-070942 - 22.011 0094 4 Rel-8 B PPAC requirement 8.1.0 8.2.0 PPACR SP-39 SP-080026 S1-080155 22.011 0101 1 Rel-8 A EHPLMN recovery lack of coverage 8.2.0 8.3.0 NSP-CR SP-39 SP-080034 S1-080351 22.011 0102 2 Rel-8 B HNB/HeNB - Closed Subscriber Group (CSG) definitions for UTRA and E-UTRA 8.2.0 8.3.0 HomeNB SP-39 SP-080042 S1-080326 22.011 0107 1 Rel-8 C basic access control enhancement for EPS 8.2.0 8.3.0 AIPN-SAE SP-39 SP-080041 S1-080331 22.011 0108 1 Rel-8 C Barring of inbound roamers of previous release networks from E-UTRAN 8.2.0 8.3.0 TEI8 SP-39 SP-080188 - 22.011 0109 4 Rel-8 B HNB/HeNB - Closed Subscriber Group (CSG) requirements for UTRA and E-UTRA 8.2.0 8.3.0 HomeNB SP-40 SP-080301 S1-080655 22.011 0114 1 Rel-8 D Delete CSGs related access restriction requirements 8.3.0 8.4.0 HomeNB SP-40 SP-080301 S1-080656 22.011 0115 2 Rel-8 F Introduction of length for HNBID 8.3.0 8.4.0 HomeNB SP-40 SP-080301 S1-080767 22.011 0116 3 Rel-8 F HNB/HeNB - Closed Subscriber Group (CSG) requirements for UTRA and E-UTRA 8.3.0 8.4.0 HomeNB SP-40 SP-080432 - 22.011 0117 3 Rel-8 F Relation between EPS Services and GPRS Services 8.3.0 8.4.0 SAE SP-40 SP-080301 S1-080663 22.011 0118 - Rel-8 C Removal of GERAN - E-UTRAN connected mode handover requirement from HeNB 8.3.0 8.4.0 HomeNB SP-40 SP-080453 - 22.011 0119 4 Rel-8 C PPAC restriction to UTRAN 8.3.0 8.4.0 PPACR Corrects typos in previous line of history table. 8.4.0 8.4.1 SP-41 SP-080493 S1-082333 22.011 0122 2 Rel-8 F Correction on definition of length for HNBID 8.4.1 8.5.0 HomeNB SP-41 SP-080493 S1-082354 22.011 0120 2 Rel-8 F Essential correction to manual CSG selection 8.4.1 8.5.0 HomeNB SP-41 SP-080493 S1-082389 22.011 0123 2 Rel-8 F Additional HNB/HeNB operator requirements 8.4.0 8.5.0 HomeNB SP-41 SP-080493 S1-082393 22.011 0128 2 Rel-8 F Terminology alignment and generalization of requirements 8.4.1 8.5.0 HomeNB SP-42 SP-080771 S1-083443 22.011 132 2 Rel-8 F Correction to the maximum limit requirement of CSG members 8.5.0 8.6.0 HomeNB SP-42 SP-080771 S1-083250 22.011 0133 1 Rel-8 F HNB/HeNB and pre-release 8 USIM 8.5.0 8.6.0 HeNB SP-42 SP-080773 S1-083380 22.011 0134 1 Rel-8 F PPAC CR alignment with CT1 8.5.0 8.6.0 PPAC SP-42 SP-080771 S1-084409 22.011 0137 2 Rel-8 F Correction of HNB/HeNB display requirements and manual CSG selection 8.5.0 8.6.0 HomeNB SP-42 SP-080874 S1-084408 22.011 0139 2 Rel-8 C Home network independent support of roamers in CSG 8.5.0 8.6.0 HomeNB SP-42 SP-080771 S1-084410 22.011 0140 1 Rel-8 B HNB mobility 8.5.0 8.6.0 HomeNB SP-42 SP-080785 S1-084389 22.011 0136 2 Rel-9 B SSAC Services Specific Access Control Requirements 8.6.0 9.0.0 SSACR SP-43 SP-090195 S1-090161 22.011 0142 1 Rel-9 D Removing the content of chapter 8 and add a reference instead. 9.0.0 9.1.0 EHNB SP-45 SP-090669 - 22.011 0156 1 Rel-9 F Description Alignment of Access Class Barring for E-UTRAN with Stage 3 9.1.0 9.2.0 TEI9 SP-46 SP-090841 S1-094330 22.011 0158 1 Rel-9 F Alignment between SSAC and Common AC 9.2.0 9.3.0 SSAC SP-48 SP-100442 S1-101221r 22.011 0161 4 Rel-10 B Access control for CSFB 9.3.0 10.0.0 TEI10 SP-49 SP-100580 S1-102153 22.011 0166 - Rel-10 F Clarification on roaming in shared networks 10.0.0 10.1.0 TEI10 SP-49 SP-100580 S1-102178 22.011 0163 1 Rel-10 F Clarification of ACB for CSFB 10.0.0 10.1.0 TEI10 SP-49 SP-100580 S1-102178 22.011 0163 1 Rel-10 F Clarification of ACB for CSFB 10.0.0 10.1.0 TEI10 SP-50 SP-100893 S1-103336 22.011 0168 12 Rel-10 B Access Control for Machine Type Communications 10.1.0 10.2.0 NIMTC SP-50 SP-100805 S1-103328 22.011 0171 - Rel-10 F Clarification of Access Class Barring for CSFB 10.1.0 10.2.0 TEI10 - LTE logo changed into LTE Advanced logo 10.2.0 10.2.1 - SP-51 SP-110162 S1-110409 22.011 0175 3 Rel-10 F Clarifications for Extended Access Barring 10.2.1 10.3.0 NIMTC SP-52 SP-110374 S1-111118 22.011 0181 - Rel-11 F Clarification on applicability of EAB 10.3.0 11.0.0 SIMTC SP-52 SP-110374 S1-111376 22.011 0178 2 Rel-11 B Clarification on how EAB is applied in Shared Network 10.3.0 11.0.0 SIMTC SP-53 SP-110582 S1-112243 22.011 0182 2 Rel-11 F Support of countries with multiple MCC values 11.0.0 11.1.0 TEI11 SP-54 SP-110813 S1-113443 22.011 0186 1 Rel-11 F Applicability of Access Class Barring in a FULL-MOCN configuration 11.1.0 11.2.0 TEI11 SP-59 SP-130123 S1-131287 22.011 0191 2 Rel-11 F Overriding extended access barring 11.2.0 11.3.0 TEI11 SP-59 SP-130110 S1-131279 22.011 0194 2 Rel-12 B Prevention of mobile-originating signalling and/or data traffic of UEs in E-UTRAN in connected mode based on the criteria for the access barring mechanisms 11.2.0 12.0.0 PMOC SP-63 SP-140064 S1-140057 22.011 200 - Rel-12 C Requirement for SSAC in CONNECTED 12.0.0 12.1.0 TEI12 SP-63 SP-140065 S1-140154 22.011 0202 - Rel-12 B Prioritization of MO-SMS for ACB 12.0.0 12.1.0 TEI12 SP-63 SP-140065 S1-140329 22.011 0198 5 Rel-12 B Prioritization of MMTEL for ACB 12.0.0 12.1.0 TEI12 SP-64 SP-140233 S1-141306 22.011 0203 2 Rel-13 B Access control mechanism for operator-identified applications 12.1.0 13.0.0 ACDC SP-65 SP-140507 S1-143622 22.011 0206 2 Rel-13 C Number and usage of ACDC categories 13.0.0 13.1.0 ACDC SP-65 SP-140523 S1-143484 22.011 0205 1 Rel-13 A I-WLAN clean up 13.0.0 13.1.0 TEI12 SP-68 SP-150267 S1-151621 22.011 214 2 Rel-13 F Number of ACDC categories and ACDC in idle mode 13.1.0 13.2.0 ACDC-ST1 SP-68 SP-150270 S1-151546 22.011 213 2 Rel-13 A Access Class for MPS (Rel-13 Mirror) 13.1.0 13.2.0 ePRIOR-St1 SP-69 SP-150535 S1-152405 22.011 217 1 Rel-13 F Handling of uncategorized ACDC applications 13.2.0 13.3.0 ACDC SP-69 SP-150539 S1-152456 22.011 215 1 Rel-14 B Enhancements for control of traffic from UE-based applications toward associated server 13.2.0 14.0.0 CATS SP-70 SP-150747 S1-154542 22.011 0225 1 Rel-14 A Corrections of ACDC Requirements 14.0.0 14.1.0 ACDC-ST1 SP-70 SP-150758 S1-154589 22.011 0224 3 Rel-14 B Requirements to support a defined set of services when 3GPP PS Data Off is configured by the HPLMN 14.0.0 14.1.0 PS_DATA_OFF SP-71 SP-160111 S1-160526 22.011 0224 7 Rel-14 B 3GPP PS Data Off 14.1.0 14.2.0 PS_DATA_OFF SP-71 SP-160091 S1-160283 22.011 0229 1 Rel-14 A Correction of ACDC requirement for IMS services 14.1.0 14.2.0 ACDC-ST1 SP-71 SP-160091 S1-160311 22.011 0233 Rel-14 A Barring of applications in unmatched ACDC categories 14.1.0 14.2.0 ACDC SP-71 SP-160111 S1-160322 22.011 0234 1 Rel-14 F Removal of non-approved CR#0224rev3 on 3GPP PS Data Off 14.1.0 14.2.0 PS_DATA_OFF SP-72 SP-160352 S1-161087 22.011 0236 Rel-14 A Extension of periodic network selection attempts interval 14.2.0 14.3.0 TEI13 SP-74 SP-160894 S1-163439 22.011 0239 2 Rel-14 F Enable CS equivalent services for PS Data Off exempted services 14.3.0 14.4.0 PS_Data_Off SP-75 SP-170151 S1-170350 22.011 0246 1 Rel-14 F PS Data Off handling of services supported by BIP 14.4.0 14.5.0 PS_Data_Off SP-75 SP-170153 S1-170354 22.011 0242 2 Rel-14 B Access Control for Evolved ProSe Remote UE 14.4.0 14.5.0 REAR SP-75 SP-170154 S1-171441 22.011 0251 1 Rel-14 F Access Technology differentiation 14.4.0 14.5.0 TEI14 SP-75 SP-170155 S1-170124 22.011 0244 Rel-15 B Administrative restriction of subscribers’ access to 5G-RAN 14.4.0 15.0.0 SMARTER SP-75 SP-170155 S1-170233 22.011 0243 1 Rel-15 B Add support of 5G in PLMN Selection 14.4.0 15.0.0 SMARTER SP-75 SP-170155 S1-171184 22.011 0249 Rel-15 B Administrative restriction of subscribers’ access to E-UTRAN with NR DC 14.4.0 15.0.0 SMARTER SP-75 SP-170160 S1-171467 22.011 0248 4 Rel-15 C Optimisation of network selection in case of an operator supporting only a subset of specific access technologies 14.4.0 15.0.0 HORNS SP-76 SP-170450 S1-172210 22.011 0257 1 Rel-15 F Adding missing abbreviations and reference 15.0.0 15.1.0 TEI15 SP-76 SP-170446 S1-172379 22.011 0256 1 Rel-15 F Alignment of PLMN selection with stage 3 15.0.0 15.1.0 HORNS SP-76 SP-170442 S1-172423 22.011 0255 2 Rel-15 A Correction to ACDC access control 15.0.0 15.1.0 ACDC-ST1 SP-76 SP-170440 S1-172207 22.011 0260 Rel-15 A Removal of ACB-skip 15.0.0 15.1.0 REAR SP-76 SP-170446 S1-172016 22.011 0253 Rel-15 F Removal of Optimisation of network selection requirements not Agreed at SA#75 15.0.0 15.1.0 HORNS SP-76 SP-170439 S1-172209 22.011 0261 Rel-15 A Removal of VPLMN-specific 3GPP PS Data Off Exempt services 15.0.0 15.1.0 PS_Data_Off SP-77 SP-170689 S1-173273 22.011 0263 1 Rel-15 A CR on REAR 15.1.0 15.2.0 REAR SP-77 SP-170695 S1-173551 22.011 0268 2 Rel-15 F RAT identifiers for PLMN selection in 5G 15.1.0 15.2.0 TEI15 SP-78 SP-170987 S1-174600 22.011 0270 2 Rel-16 B Selective Legacy Radio Disabling 15.2.0 16.0.0 TEI16 SP-79 SP-180142 S1-180635 22.011 0275 3 Rel-16 C Selective Legacy Radio Disabling 16.0.0 16.1.0 TEI16 SP-79 SP-180228 - 22.011 277 Rel-16 A Addition of RLOS availability into Automatic PLMN Selection 16.0.0 16.1.0 PARLOS SP-80 SP-180305 S1-181395 22.011 0283 1 Rel-16 F Alignment of RLOS availability into Automatic PLMN Selection 16.1.0 16.2.0 PARLOS SP-80 SP-180308 S1-181566 22.011 0281 2 Rel-16 A Clarification to forbidden TA 16.1.0 16.2.0 TEI15 SP-80 SP-180308 S1-181678 22.011 0287 1 Rel-16 A IoT periodic network selection attempt timers 16.1.0 16.2.0 TEI15 SP-81 SP-180751 S1-182254 22.011 0296 Rel-16 A Correction to EAB 16.2.0 16.3.0 TEI14 SP-81 SP-180751 S1-182655 22.011 0300 Rel-16 A PS Data Off handling of non-IP PDN types 16.2.0 16.3.0 PS_Data_Off SP-81 SP-180752 S1-182096 22.011 0291 Rel-16 A Clarification to Delay Tolerant 16.2.0 16.3.0 SMARTER SP-81 SP-180756 S1-182633 22.011 0289 1 Rel-16 C Delaying periodic higher priority PLMN searches and reselection attempts when receiving eMBMS service in idle mode 16.2.0 16.3.0 TEI16 SP-81 SP-180907 - 22.011 0297 5 Rel-16 F Clarification on the PLMN selection mode for RLOS 16.2.0 16.3.0 PARLOS Change history Date Meeting TDoc CR Rev Cat Subject/Comment New version 2018-12 SP-82 SP-180999 0301 4 F correction on PLMN selection in RLOS 16.4.0 2018-12 SP-82 SP-180999 0302 3 F Clarification on disabling of legacy radio 16.4.0 2019-12 SP-86 SP-191032 0307 2 B On PLMN selection for MINT 17.0.0 2020-03 SP-87 SP-200123 0310 1 F Clarification on the traffic handling over non-3GPP access for PS DATA Off 17.1.0 2020-09 SA#89e SP-200794 315 1 F Corrections to Service accessibility specification 17.2.0 2020-12 SA#90e SP-201029 319 0 F Satellite access RAT identifier for PLMN selection 17.3.0 2020-12 SA#90e SP-201029 321 1 F Requirement Alignment related to PLMN reselection 17.3.0 2021-09 SA#93e SP-211096 0324 D Correction of 'air interface' terminology 17.4.0 2021-09 SA#93e SP-211068 0322 4 B Signal level Enhanced Network Selection 18.0.0 2021-12 SP#94e SP-211487 0327 1 A Periodic network selection with shared MCC 18.1.0 2022-03 SP#95e SP-220082 0330 D Clean-up of the references for quality improvement 18.2.0 2022-06 SA#96 SP-220433 340 1 A Clarifications on PLMN search for FPLMN Registered UEs 18.3.0 2022-06 SA#96 SP-220428 336 1 A Clarification of Shared MCC definition 18.3.0 2022-12 SA#98 SP-221258 0342 F Clarification of SENSE requirement under disaster roaming condition 18.4.0 2022-12 SA#98 SP-221258 0344 2 F Clarification on the periodic network selection for SENSE 18.4.0 2023-03 SA#99 SP-230228 0348 3 B PS Data Off for IMS Data Channel Service 19.0.0 2023-06 SA#100 SP-230530 0351 1 A Clarification on selection of Forbidden PLMN during Disaster Conditions when Operator controlled signal threshold applies 19.1.0 2023-09 SA#101 SP-231041 0352 D Quality improvements for TS22.011 Release 19 19.2.0 2024-03 SA#103 SP-240202 0359 3 F Clarification on the PS Data Off exemption for services over Data Channel 19.3.0 2024-09 SA#105 SP-241145 0363 1 F Location services user plane protocol and 3GPP PS data off 19.4.0 2024-12 SA#106 SP-241760 0370 2 A Alignment on Satellite E-UTRA 19.5.0 2024-12 SA#106 SP-241762 0364 3 F Clarification on support of 3GPP PS data off for 5G 19.5.0
162c3f0c36a54745dc509a8130708d2d
22.104
1 Scope
The present document provides Stage 1 normative service requirements for 5G systems, in particular service requirements for cyber-physical control applications in vertical domains and requirements for auxiliary applications. In the context of the present document, cyber-physical systems are to be understood as systems that include engineered, interacting networks of physical and computational components; control applications are to be understood as applications that control physical processes. Examples for auxiliary applications are distributed sensing and asset monitoring. Communication services supporting cyber-physical control applications need to be ultra-reliable and, in some cases, the end-to-end latency must be very low. Communication for cyber-physical control applications supports operation in various vertical domains, for instance industrial automation, Smart Grid . The aspects addressed in the present document include: - end-to-end service performance requirements and network performance requirements related to these end-to-end service performance requirements; - support for Ethernet services specific to industrial/high performance use cases. Related Ethernet functionalities include, for example, those in IEEE 802.1Qbv; - direct device connection and indirect network connection for cyber-physical applications.
162c3f0c36a54745dc509a8130708d2d
22.104
2 References
The following documents contain provisions which, through reference in this text, constitute provisions of the present document. - References are either specific (identified by date of publication, edition number, version number, etc.) or non‑specific. - For a specific reference, subsequent revisions do not apply. - For a non-specific reference, the latest version applies. In the case of a reference to a 3GPP document (including a GSM document), a non-specific reference implicitly refers to the latest version of that document in the same Release as the present document. [1] 3GPP TR 21.905: "Vocabulary for 3GPP Specifications". [2] 3GPP TS 22.261: "Service requirements for the 5G system". [3] IEC 61784-3: "Industrial communication networks – profiles – part 3: functional fieldbuses – general rules and profile definitions". [4] BZKI, "Aspects of dependability assessment in ZDKI", June 2017. [5] BZKI, "Requirement Profiles in ZDKI", 2017. [6] IEC 61158: "Industrial communication networks – fieldbus specification", 2014. [7] IEC 61907, "Communication network dependability engineering". [8] Richard C. Dorf and Robert H. Bishop, "Modern Control Systems", Pearson, Harlow, 13th Edition, 2017. [9] Ernie Hayden, Michael Assante, and Tim Conway, "An Abbreviated History of Automation & Industrial Controls Systems and Cybersecurity", SANS Institute, https://ics.sans.org/media/An-Abbreviated-History-of-Automation-and-ICS-Cybersecurity.pdf {accessed: 2017-05-23}, 2014. [10] IEC 61512 "Batch control - Part 1: Models and terminology". [11] RESERVE project, Deliverable D1.3, ICT Requirements, http://www.re-serve.eu/files/reserve/Content/Deliverables/D1.3.pdf, September 2017. [12] RESERVE project, Deliverable D1.2, Energy System Requirements http://www.re-serve.eu/files/reserve/Content/Deliverables/D1.2.pdf, September 2017. [13] G. Garner, "Designing Last Mile Communications Infrastructures for Intelligent Utility Networks (Smart Grids)", IBM Australia Limited, 2010. [14] B. Al-Omar, B., A. R. Al-Ali, R. Ahmed, and T. Landolsi, "Role of Information and Communication Technologies in the Smart Grid", Journal of Emerging Trends in Computing and Information Sciences, Vol. 3, pp. 707-716, 2015. [15] H. Kagermann, W. Wahlster, and J. Helbig, "Recommendations for implementing the strategic initiative INDUSTRIE 4.0", Final report of the Industrie 4.0 working group, acatech – National Academy of Science and Engineering, Munich, April 2013. [16] IEC 62443-3-2: "Security for industrial automation and control systems - Part 3-2: Security risk assessment and system design", in progress. [17] IEC 62657-2: "Industrial communication networks - Wireless communication networks - Part 2: Coexistence management", 2017. [18] IEC 62657-1: "Industrial communication networks – Wireless communication networks – Part 1: Wireless communication requirements and spectrum considerations". [19] IEEE Std 802.1Q: "IEEE Standard for Local and Metropolitan Area Networks---Bridges and Bridged Networks". NOTE: IEEE Std 802.1Qbv-2015 "IEEE Standard for Local and Metropolitan Area Networks--Bridges and Bridges Networks - Amendment 25: Enhancements for Scheduled Traffic" has been included into IEEE Std 802.1Q-2018. [20] IEEE, Use Cases IEC/IEEE 60802, 2018. [21] (void) [22] IEEE Std 802.1AS: "IEEE Standard for Local and Metropolitan Area Networks--Timing and Synchronization for Time-Sensitive Applications". [23] 3GPP TS 22.289: "Mobile Communication System for Railways". [24] IEEE P802.1CS: "IEEE Standard for Local and Metropolitan Area Networks--Link-local Registration Protocol". [25] IEEE P802.1Qdd: "IEEE Draft Standard for Local and Metropolitan Area Networks--Bridges and Bridged Networks -- Amendment: Resource Allocation Protocol (RAP) ". [26] IEC/IEEE 60802: "Time-Sensitive Networking Profile for Industrial Automation". [27] 3GPP TS 22.263: "Service requirements for Video, Imaging and Audio for Professional Applications (VIAPA)". [28] IEC TR 61850-90-1:2010, Communication Networks and Systems for Power Utility automation – Part 90-1: Use of IEC61850 for the communication between substations. [29] 5G DNA White Paper: "5GDN@Smart Grid White Paper: Requirements, Technologies, and Practices" https://www.5gdna.org/ [30] IEC 61850-9-3-2016 – IEC/IEEE International Standard - Communication Networks and Systems for Power Utility automation – Part 9-3: Precision time protocol profile for power utility automation. [31] IEEE Std C37.238-2017 , IEEE Standard Profile for Use of IEEE Std 1588™ Precision Time Protocol in Power System Applications. [32] IEC 61850-90-5:2012, Use of IEC 61850 to transmit Synchrophasors information according to IEEE C37.118. [33] IEEE Std C37.118.2-2011, IEEE Standard for Synchrophasor Data Transfer for Power Systems. [34] IEEE Std 1588-2019: "IEEE Standard for a Precision Clock Synchronization Protocol for Networked Measurement and Control".
162c3f0c36a54745dc509a8130708d2d
22.104
3 Definitions, symbols and abbreviations
162c3f0c36a54745dc509a8130708d2d
22.104
3.1 Definitions
For the purposes of the present document, the terms and definitions given in 3GPP TR 21.905 [1] and the following apply. A term defined in the present document takes precedence over the definition of the same term, if any, in 3GPP TR 21.905 [1]. characteristic parameter: numerical value that can be used for characterising the dynamic behaviour of communication functionality from an application point of view. clock synchronicity: the maximum allowed time offset within a synchronisation domain between the sync master and any sync device. NOTE 1: Clock synchronicity (or synchronicity) is used as KPI of clock synchronisation services. NOTE 2: Clock synchronicity is also referred to as clock (or time) synchronization precision. clock synchronisation service: the service to align otherwise independent user-specific UE clocks. communication service availability: as defined in TS 22.261 [2]. communication service reliability: ability of the communication service to perform as required for a given time interval, under given conditions. NOTE 3: Given conditions would include aspects that affect reliability, such as: mode of operation, stress levels, and environmental conditions. NOTE 4: Reliability may be quantified using appropriate measures such as mean time between failures, or the probability of no failure within a specified period of time. NOTE 5: This definition is based on IEC 61907 [7]. direct device connection: as defined in TS 22.261 [2]. end-to-end latency: as defined in TS 22.261 [2]. error: discrepancy between a computed, observed or measured value or condition and the true, specified or theoretically correct value or condition. NOTE 6: This definition was taken from IEC 61784-3 [3]. factory automation: automation application in industrial automation branches typically with discrete characteristics of the application to be automated with specific requirements for determinism, low latency, reliability, redundancy, cyber security, and functional safety. NOTE 7: Low latency typically means below 10 ms delivery time. NOTE 8: This definition is taken from IEC 62657-1 [18]. global clock: a user-specific synchronization clock set to a reference timescale such as the International Atomic Time. indirect network connection: as defined in TS 22.261 [2]. influence quantity: quantity not essential for the performance of an item but affecting its performance. process automation: automation application in industrial automation branches typically with continuous characteristics of the application to be automated with specific requirements for determinism, reliability, redundancy, cyber security, and functional safety. NOTE 9: This definition is taken from IEC 62657-1 [18]. service area: as defined in TS 22.261 [2]. survival time: as defined in TS 22.261 [2]. sync device: device that synchronizes itself to the master clock of the synchronization domain. sync master: device serving as the master clock of the synchronization domain. transfer interval: time difference between two consecutive transfers of application data from an application via the service interface to 3GPP system. NOTE 10: This definition is based on subclause 3.1.85 in IEC 62657-2 [17]. user experienced data rate: as defined in TS 22.261 [2]. vertical domain: an industry or group of enterprises in which similar products or services are developed, produced, and provided. working clock: a user-specific synchronization clock for a localized set of UEs collaborating on a specific task or work function.
162c3f0c36a54745dc509a8130708d2d
22.104
3.2 Symbols
For the purposes of the present document, the following symbols apply: <symbol> <Explanation>
162c3f0c36a54745dc509a8130708d2d
22.104
3.3 Abbreviations
For the purposes of the present document, the abbreviations given in 3GPP TR 21.905 [1] and the following apply. An abbreviation defined in the present document takes precedence over the definition of the same abbreviation, if any, in 3GPP TR 21.905 [1]. AV Audio-Visual AVPROD AV Production CSIF Communication Service Interface EPON Ethernet Passive Optical Network FIFO First In, First Out GOOSE Generic Object-Oriented Substation Event HCL Higher Communication Layer HMI Human Machine Interface IMU Inertial Measurement Unit LCL Lower Communication Layer PMU Phasor Measurement Unit
162c3f0c36a54745dc509a8130708d2d
22.104
4 Overview
162c3f0c36a54745dc509a8130708d2d
22.104
4.1 Introduction
For the purpose of this document, a vertical domain is a particular industry or group of enterprises in which similar products or services are developed, produced, and provided. Automation refers to the control of processes, devices, or systems in vertical domains by automatic means. The main control functions of automated control systems include taking measurements, comparing results, computing any detected or anticipated errors, and correcting the process to avoid future errors. These functions are performed by sensors, transmitters, controllers, and actuators. In the context of this document, cyber-physical systems are referred to as systems that include engineered, interacting networks of physical and computational components. Cyber-physical control applications are to be understood as applications that control physical processes. Cyber-physical control applications in automation follow certain activity patterns, which are open-loop control, closed-loop control, sequence control, and batch control (see Clause 4.2). Communication services supporting cyber-physical control applications need to be ultra-reliable, dependable with a high communication service availability, and often require low or (in some cases) very low end-to-end latency. Communication in automation in vertical domains follows certain communication patterns. The most well-known is periodic deterministic communication, others are aperiodic deterministic communication and non-deterministic communication (see Clause 4.3). Communication for cyber-physical control applications supports operation in various vertical domains, for instance industrial automation and energy automation. This document addresses service requirements for cyber-physical control applications and supporting communication services from the vertical domains of factories of the future (smart manufacturing), electric power distribution, and central power generation. Service requirements for cyber-physical control applications and supporting communication services for rail-bound mass transit are addressed in TS 22.289 [23].
162c3f0c36a54745dc509a8130708d2d
22.104
4.2 Activity patterns in automation
Open-loop control: The salient aspect of open-loop control is the lack of feedback from the output to the control; when providing commands to an actuator, it is assumed that the output of the influenced process is predetermined and within an acceptable range. This kind of control loop works if the influences of the environment on process and actuator are negligible. Also, this kind of control is applied in case unwanted output can be tolerated [8]. Closed-loop control: Closed-loop control enables the manipulation of processes even if the environment influences the process or the performance of the actuator changes over time. This type of control is realised by sensing the process output and by feeding these measurements back into a controller [8]. Sequence control: Sequence control may either step through a fixed sequence or employ logic that performs different actions based on various system states and system input [8]. Sequence control can be seen as an extension of both open-loop and closed-loop control, but instead of achieving only one output instance, an entire sequence of output instances can be produced [9]. Batch control: Batch processes lead to the production of finite quantities of material (batches) by subjecting input materials to a defined order of processing actions by use of one or more pieces of equipment [10].
162c3f0c36a54745dc509a8130708d2d
22.104
4.3 Communication attributes
Communication in automation can be characterised by two main attributes: periodicity and determinism. Periodicity means that a transmission interval is repeated. For example, a transmission occurs every 15 ms. Reasons for a periodical transmission can be the periodic update of a position or the repeated monitoring of a characteristic parameter. Most periodic intervals in communication for automation are rather short. The transmission is started once and continuous unless a stop command is provided. An aperiodic transmission is, for example, a transmission which is triggered instantaneously by an event, i.e., events are the trigger of the transmission. Events are defined by the control system or by the user. Example events are: - Process events: events that come from the process when thresholds are exceeded or fallen below, e.g., temperature, pressure, level, etc. - Diagnostic events: events that indicate malfunctions of an automation device or module, e.g., power supply defective; short circuit; too high temperature; etc. - Maintenance events: events based on information that indicates necessary maintenance work to prevent the failure of an automation device. Most events, and especially alarms, are confirmed. In this context, alarms are messages that inform a controller or operator that an event has occurred, e.g., an equipment malfunction, process deviation, or other abnormal condition requiring a response. The receipt of the alarm is acknowledged usually within a short time period by the application that received the alarm. If no acknowledgment is received from the target application after a preset time, the so-called monitoring time, the alarm is sent again after a preset time or some failure response action is started. Determinism refers to whether the delay between transmission of a message and receipt of the message at the destination address is stable (within bounds). Usually, communication is called deterministic if it is bounded by a given threshold for the latency/transmission time. In case of a periodic transmission, the variation of the interval is bounded.
162c3f0c36a54745dc509a8130708d2d
22.104
4.4 Control systems and related communication patterns
There are preferences in the mapping between the type of control and the communication pattern. Open-loop control is characterised by one or many messages sent to an actuator. These can be sent in a periodic or an aperiodic pattern. However, the communication means used need to be deterministic since typically an activity response from the receiver and/or the receiving application is expected. Closed-loop control produces both periodic and aperiodic communication patterns. Closed-loop control is often used for the control of continuous processes with tight time-control limits, e.g., the control of a printing press. In this case, one typically relies on periodic communication patterns. Note that in both the aperiodic and periodic case, the communication needs to be deterministic. Logging of device states, measurements, etc. for maintenance purposes and such typically entails aperiodic communication patterns. In case the transmitted logging information can be time-stamped by the respective function, determinism is often not mandatory. In practice, vertical communication networks serve a large number of applications exhibiting a wide range of communication requirements. In order to facilitate efficient modelling of the communication network during engineering and for reducing the complexity of network optimisation, traffic classes or communication patterns have been identified [6]. There are three typical traffic classes or communication patterns in industrial environments [6], i.e., - deterministic periodic communication: periodic communication with stringent requirements on timeliness of the transmission. - deterministic aperiodic communication: communication without a preset sending time. Typical activity patterns for which this kind of communication is suitable are event-driven actions. - non-deterministic communication: subsumes all other types of traffic, including periodic non-real time and aperiodic non-real time traffic. Periodicity is irrelevant in case the communication is not time-critical. Some communication services exhibit traffic patterns that cannot be assigned to one of the above communication patterns exclusively (mixed traffic).
162c3f0c36a54745dc509a8130708d2d
22.104
4.5 Implications for 5G systems
In order to be suitable for automation in vertical domains, 5G systems need to be dependable and flexible to meet specific KPIs to serve specific applications and use cases. They need to come with the system properties of reliability, availability, maintainability, safety, and integrity. What particular requirements each property needs to meet depends on the particularities of the domain and the use case. Annex F discusses the difference between reliability and communication service availability. The requirements in this document provide various sets of performance criteria that need to be met to satisfactorily support different use cases of cyber-physical control applications used by various vertical markets.
162c3f0c36a54745dc509a8130708d2d
22.104
5 Performance requirements
162c3f0c36a54745dc509a8130708d2d
22.104
5.1 Overview
There are two fundamental perspectives concerning dependable communication in 5G systems: the end-to-end perspective of the communication services and the network perspective (see Figure 5.1-1). Figure 5.1-1: Network perspective of 5G system The Communication Service in Figure 5.1-1 may be implemented as a logical communication link between a UE on one side and a network server on the other side, or between a UE on one side and a UE on the other side. In some cases, a local approach (e.g. network edge) is preferred for the communication service on the network side in order to reduce the latency, to increase communication service availability, or to keep sensitive data in a non-public network on the factory site. The tables in Clauses 5.2 through 5.5 below provide sets of requirements where periodicity and determinism are critical to meeting cyber-physical control application needs in various vertical scenarios. While many use cases have similar KPI values, the important distinction is that in order to meet the needs of different verticals and different uses, the 5G system will need to be sufficiently flexible to allow deployment configurations that can meet the different sets of KPIs specific to each use. Communication service availability is considered an important service performance requirement for cyber-physical applications, especially for applications with deterministic traffic. The communication service availability depends on the latency and reliability (in the context of network layer packet transmissions, as defined in TS 22.261 [2]) of the logical communication link, as well as the survival time of the cyber-physical application (see Annex C.3 for further details on these relations). The communication service reliability requirements also depend on the operation characteristics of the corresponding cyber-physical applications. Typically, the communication services critical for the automation application also come with stringent communication service reliability requirements. Note that the communication service reliability requirement has no direct relationship with the communication service availability requirement. The "# of UEs" in the tables in clauses 5.2 to 5.5 is intended to give an indication of the UE density that would need to be served within a given service area. Clock synchronisation is needed in many "vertical" use cases. The requirements and tables in Clause 5.6 provide specific criteria for managing time sensitive communications in an industrial environment. High accuracy positioning is becoming essential for Factories of the Future. The reason for this is that tracking of mobile devices as well as mobile assets is becoming increasingly important in improving processes and increasing flexibility in industrial environments, Clause 5.7 provides positioning requirements for horizontal and vertical accuracy, availability, heading, latency and UE speed in an industrial use case scenario. An example of the relationship between reliability (in the context of network layer packet transmissions, as defined in TS 22.261 [2]), survival time and communication service availability of a logical communication link is illustrated in the following Table 5.1-1. This is done for a special case where packet errors are uncorrelated, which in many cases is an unrealistic assumption. Table 5.1-1: Example of relationship between reliability (as defined in TS 22.261) and communication service availability when the survival time is equal to the transfer interval. Communication service availability Reliability ( as defined in TS 22.261) 1 - p 99.999 9 % 99.9 % 99.999 999 % 99.99 % 99.999 999 99 % 99.999 % 99.999 999 999 9 % 99.999 9 % 99.999 999 999 999 % 99.999 99 %
162c3f0c36a54745dc509a8130708d2d
22.104
5.2 Periodic deterministic communication
Periodic deterministic communication is periodic with stringent requirements on timeliness and availability of the communication service. A transmission occurs every transfer interval. A description of periodic deterministic communication can be found in Clauses 4.3 and 4.4. Additional information on the underlying use cases of the sets of requirements in Table 5.2-1 can be found in Annex A. Further information on characteristic parameters and influence quantities used in Table 5.2-1 can be found in Annex C. The 5G system shall be able to provide periodic deterministic communication with the service performance requirements for individual logical communication links that realise the communication services reported in Table 5.2‑1. Process and asset monitoring using industrial wireless sensors is a special case of periodic deterministic communication with more relaxed requirements on timeliness and availability. These use cases put a slightly different set of requirements on the 5G system due to the specific constraints of industrial wireless sensors. These requirements for individual logical communication links are listed in Table 5.2-2 and additional information on the underlying use cases can be found in Annex A. Smart-Grid use case information can be found in Annex A. Table 5.2-1: Periodic deterministic communication service performance requirements Characteristic parameter Influence quantity Communica­tion service availability: target value (note 1) Communication service reliability: mean time between failures End-to-end latency: maximum (note 2) (note 12a) Service bit rate: user experienced data rate (note 12a) Message size [byte] (note 12a) Transfer interval: target value (note 12a) Survival time (note 12a) UE speed (note 13) # of UEs Service area (note 3) Remarks 99.999 % to 99.999 99 % ~ 10 years < transfer interval value – 50 500 μs 500 μs ≤ 75 km/h ≤ 20 50 m x 10 m x 10 m Motion control (A.2.2.1) 99.999 9 % to 99.999 999 % ~ 10 years < transfer interval value – 40 1 ms 1 ms ≤ 75 km/h ≤ 50 50 m x 10 m x 10 m Motion control (A.2.2.1) 99.999 9 % to 99.999 999 % ~ 10 years < transfer interval value – 20 2 ms 2 ms ≤ 75 km/h ≤ 100 50 m x 10 m x 10 m Motion control (A.2.2.1) 99.999 9 % – < 5 ms 1 kbit/s (steady state) 1.5 Mbit/s (fault case) < 1,500 < 60 s (steady state) ≥ 1 ms (fault case) transfer interval stationary 20 30 km x 20 km Electrical Distribution – Dis­tributed automated switch­ing for isolation and service restoration (A.4.4); (note 5) 99.999 9 % to 99.999 999 % ~ 10 years < transfer interval value 1 k ≤ 10 ms 10 ms - 5 to 10 100 m x 30 m x 10 m Control-to-control in motion control (A.2.2.2); (note 9) 99.999 9 % to 99.999 999 % ~ 10 years < transfer interval value (note 5) 50 Mbit/s ≤ 1 ms 3 x transfer interval stationary 2 to 5 100 m x 30 m x 10 m Wired-2-wireless 100 Mbit/s link replacement (A.2.2.4) 99.999 9 % to 99.999 999 % ~ 10 years < transfer interval value (note 5) 250 Mbit/s ≤ 1 ms 3 x transfer interval stationary 2 to 5 100 m x 30 m x 10 m Wired-2-wireless 1 Gbit/s link replacement (A.2.2.4) 99.999 9 % to 99.999 999 % ~ 10 years < transfer interval value 1 k ≤ 50 ms 50 ms - 5 to 10 1,000 m x 30 m x 10 m Control-to-control in motion control (A.2.2.2); (note 9) > 99.999 9 % ~ 10 years < transfer interval value – 40 to 250 1 ms to 50 ms (note 6) (note 7) transfer interval value ≤ 50 km/h ≤ 2,000 ≤ 1 km2 Mobile robots (A.2.2.3) 99.999 9 % to 99.999 999 % ~ 1 month < transfer interval value – 40 to 250 4 ms to 8 ms (note 7) transfer interval value < 8 km/h (linear movement) TBD 50 m x 10 m x 4 m Mobile control panels – remote control of e.g. assembly robots, milling machines (A.2.4.1); (note 9) 99.999 999 % 1 day < 8 ms (note 14) 250 kbit/s 40 to 250 8 ms 16 ms quasi-static; up to 10 km/h 2 or more 30 m x 30 m Mobile Opera­tion Panel: Emer­gency stop (connectivity availability) (A.2.4.1A) 99.999 99 % 1 day < 10 ms (note 14) < 1 Mbit/s < 1024 10 ms ~10 ms quasi-static; up to 10 km/h 2 or more 30 m x 30 m Mobile Operation Panel: Safety data stream (A.2.4.1A) 99.999 999 % 1 day 10 ms to 100 ms (note 14) 10 kbit/s 10 to 100 10 ms to 100 ms transfer interval stationary 2 or more 100 m² to 2,000 m² Mobile Operation Panel: Control to visualization (A.2.4.1A) 99.999 999 % 1 day < 1 ms (note 14) 12 Mbit/s to 16 Mbit/s 10 to 100 1 ms ~ 1 ms stationary 2 or more 100 m² Mobile Operation Panel: Motion control (A.2.4.1A) 99.999 999 % 1 day < 2 ms (note 14) 16 kbit/s (UL) 2 Mbit/s (DL) 50 2 ms ~ 2 ms stationary 2 or more 100 m² Mobile Operation Panel: Haptic feedback data stream (A.2.4.1A) 99.999 9 % to 99.999 999 % ~ 1 year < transfer interval – 40 to 250 < 12 ms (note 7) 12 ms < 8 km/h (linear movement) TBD typically 40 m x 60 m; maximum 200 m x 300 m Mobile control panels -remote control of e.g. mobile cranes, mobile pumps, fixed portal cranes (A.2.4.1); (note 9) 99.999 9 % to 99.999 999 % ≥ 1 year < transfer interval value – 20 ≥ 10 ms (note 8) 0 typically stationary typically 10 to 20 typically ≤ 100 m x 100 m x 50 m Process automation – closed loop control (A.2.3.1) 99.999 % TBD ~ 50 ms – ~ 100 ~ 50 ms TBD stationary ≤ 100,000 several km2 up to 100,000 km2 Primary frequency control (A.4.2); (note 9) 99.999 % TBD ~ 100 ms – ~ 100 ~ 200 ms TBD stationary ≤ 100,000 several km2 up to 100,000 km2 Distributed Voltage Control (A.4.3) (note 9) > 99.999 9 % ~ 1 year < transfer interval value – 15 k to 250 k 10 ms to 100 ms (note 7) transfer interval value ≤ 50 km/h ≤ 2,000 ≤ 1 km2 Mobile robots – video-operated remote control (A.2.2.3) > 99.999 9 % ~ 1 year < transfer interval value – 40 to 250 40 ms to 500 ms (note 7) transfer interval value ≤ 50 km/h ≤ 2,000 ≤ 1 km2 Mobile robots (A.2.2.3) 99.99 % ≥ 1 week < transfer interval value – 20 to 255 100 ms to 60 s (note 7) ≥ 3 x transfer interval value typically stationary ≤ 10,000 to 100,000 ≤ 10 km x 10 km x 50 m Plant asset management (A.2.3.3) >99.999 999 % > 10 years < 2 ms 2 Mbit/s to 16 Mbit/s 250 to 2,000 1 ms transfer interval value stationary 1 < 100 m2 Robotic Aided Surgery (A.6.2) >99.999 9 % > 1 year < 20 ms 2 Mbit/s to 16 Mbit/s 250 to 2,000 1 ms transfer interval value stationary 2 per 1,000 km2 < 400 km (note 12) Robotic Aided Surgery (A.6.2) >99.999 % >> 1 month (< 1 year) < 20 ms 2 Mbit/s to 16 Mbit/s 80 1 ms transfer interval value stationary 20 per 100 km2 < 50 km (note 12) Robotic Aided Diagnosis (A.6.3) 99.999 9 % to 99.999 999 % ~ 10 years < 0.5 x transfer interval 2.5 Mbit/s 250 500 with localisa­tion informa­tion > 5 ms > 2.5 ms > 1.7 ms (note 10) 0 transfer interval 2 x transfer interval (note 10) ≤ 6 km/h (linear movement) 2 to 8 10 m x 10 m x 5 m; 50 m x 5 m x 5 m (note 11) Cooperative carrying – fragile work pieces; (ProSe communication) (A.2.2.5) 99.999 9 % to 99.999 999 % ~ 10 years < 0.5 x transfer interval 2.5 Mbit/s 250 500 with localisa­tion informa­tion > 5 ms > 2.5 ms > 1.7 ms (note 10) 0 transfer interval 2 x transfer interval (note 10) ≤ 12 km/h (linear movement) 2 to 8 10 m x 10 m x 5 m; 50 m x 5 m x 5 m (note 11) Cooperative carrying – elastic work pieces; (ProSe communication) (A.2.2.5) > 99.9 % DL: < 10 ms UL: < 10 ms UL: > 16 Mbit/s (urban), 640 Mbit/s (rural) DL: > 100 kbit/s (note 15) UL: 800 kbyte UL: 10 ms > 10/km2 (urban), > 100/km2 (rural) (note 16) Distributed energy storage ‒ monitoring (A.4.6) > 99.9 % DL: < 10 ms UL: < 1 s UL: > 128 kbit/s (urban), 10.4 Mbit/s (rural); DL: > 100 kbit/s (note 15) UL: 1.3 Mbyte DL: > 100 kbyte UL: 1000 ms > 10/km2 (urban), > 100/km2 (rural) (note 16) Distributed energy storage ‒ data collection (A.4.6) > 99.99 % General information data collection: < 3 s (note 17) UL: < 2 Mbit/s DL: < 1 Mbit/s < 10,000/km2 (note 18) Advanced metering (A.4.7) 99.999 % < 10 ms 2 Mbit/s to 10 Mbit/s normal: 1 s; fault: 2 ms (note 24) 54/km² (note 19), 78/km2 (note 20) Intelligent distributed feeder automation (A.4.4.3) > 99.99 % 10 ms, 100 ms, 3 s (note 22) > 2 Mbit/s (note 21) 500 in the service area (note 23) Communication distance is from 100 m to 500 m, outdoor, indoor / deep indoor Smart distribution ‒transformer terminal (A.4.8) 99.999 % 5 ms, 10 ms, 15 ms (note 25) 1.2 Mbit/s to 2.5 Mbit/s < 245 byte ≤ 1 ms ≤ 2 ms (note 26) ≤ 100/km2 several km2 High speed current differential protection (note 12a) (A.4.4.4) 99.999 9 % 3 ms 5.4 Mbit/s 140 byte ≤ 1 ms stationary Distributed Energy Resources (DER) and micro-grids (A.4.9) 99.999 9 % 100 ms (note 12a and note 5) < 1 kbit/s per DER stationary Ensuring uninterrupted communication service availability during emergencies (A.4.10) NOTE 1: One or more retransmissions of network layer packets may take place in order to satisfy the communication service availability requirement. NOTE 2: Unless otherwise specified, all communication includes 1 wireless link (UE to network node or network node to UE) rather than two wireless links (UE to UE). NOTE 3: Length x width (x height). NOTE 4: (void) NOTE 5: Communication includes two wireless links (UE to UE). NOTE 6: This covers different transfer intervals for different similar use cases with target values of 1 ms, 1 ms to 10 ms, and 10 ms to 50 ms. NOTE 7: The transfer interval deviates around its target value by < ±25 %. NOTE 8: The transfer interval deviates around its target value by < ±5 %. NOTE 9: Communication may include two wireless links (UE to UE). NOTE 10: The first value is the application requirement, the other values are the requirement with multiple transmission of the same information (two or three times, respectively). NOTE 11: Service Area for direct communication between UEs. The group of UEs with direct communication might move throughout the whole factory site (up to several km²). NOTE 12: Maximum straight-line distance between UEs. NOTE 12a: It applies to both UL and DL unless stated otherwise. NOTE 13: It applies to both linear movement and rotation unless stated otherwise. NOTE 14: The mobile operation panel is connected wirelessly to the 5G system. If the mobile robot/production line is also connected wirelessly to the 5G system, the communication includes two wireless links. NOTE 15: Service bit rate for one energy storage station. NOTE 16: Activity storage nodes/km2. This value is used for deducing the data volume in an area that features multiple energy storage stations. The data volume can be calculated with the following formula (current service bit rate per storage station) x (activity storage nodes/km2) + (video service bit rate per storage station) x (activity storage nodes/km2). NOTE 17: One-way delay from 5G IoT device to backend system. The distance between the two is below 40 km (city range). NOTE 18: Typical connection density in today’s city environment. With the evolution from centralised meters to socket meters in the home, the connection density is expected to increase 5 to 10 times. NOTE 19: When the distributed terminals are deployed along an overhead line, there are about 54 terminals per square kilometre. NOTE 20: When the distributed terminals are deployed in power distribution cabinets, there are about 78 terminals per square kilometre. NOTE 21: Service bit rate of the smart metering application between the smart distribution transformer terminal and the energy end equipment. Once there are multiple smart grid applications, the required service bit rate will be higher. NOTE 22: The end-to-end latency depends on the applications supported by the smart distribution transformer terminal. The lower the end-to-end latency, the more applications can be supported. NOTE 23: The service area is circular with a radius between 100 m and 500 m (0.031 km2 to 0.79 km2). NOTE 24: During the normal working phase of the feeder system, the heartbeat packet is transmitted periodically with a 1 s transfer interval. When a fault occurs, the heartbeat is sent with a 2 ms transfer interval. NOTE 25: The maximum allowed delay between two protection relays would be between 5 ms and 10 ms, depending on the voltage (see IEC 61850-90-1 for more details [aa]). For some legacy systems, the end-to-end latency is usually set to 15 ms. NOTE 26: For a sampling rate of 600 Hz, the transfer interval is 1.7 ms. For 1200 Hz, the transfer interval is 0.83 ms. Table 5.2-2: Communication service performance requirements for industrial wireless sensors Characteristic parameter Influence quantity Communica­tion service availability: target value Communication service reliability: mean time between failure End-to-end latency (note 6) Transfer interval (note 1) (note 7) Service bit rate: user experienced data rate (note 2) (note 7) Battery lifetime [year] (note 3) Message Size [byte] (note 7) Survival time (note 7) UE speed UE density [UE / m²] Range [m] (note 4) Remarks 99.99 % ≥ 1 week < 100 ms 100 ms to 60 s ≤ 1 Mbit/s ≥ 5 20 (note 5) 3 x transfer interval stationary Up to 1 < 500 Process monitoring, e.g. temperature sensor (A.2.3.2) 99.99 % ≥ 1 week < 100 ms ≤ 1 s ≤ 200 kbit/s ≥ 5 25 k 3 x transfer interval stationary Up to 0.05 < 500 Asset monitoring, e.g. vibration sensor (A.2.3.2) 99.99 % ≥ 1 week < 100 ms ≤ 1 s ≤ 2 Mbit/s ≥ 5 250 k 3 x transfer interval stationary Up to 0.05 < 500 Asset monitoring, e.g. thermal camera (A.2.3.2) NOTE 1: The transfer interval deviates around its target value by < ± 25 %. NOTE 2: The traffic is predominantly mobile originated. NOTE 3: Industrial sensors can use a wide variety of batteries depending on the use case, but in general they are highly constrained in terms of battery size. NOTE 4: Distance between the gNB and the UE. NOTE 5: The application-level messages in this use case are typically transferred over Ethernet. For small messages, the minimum Ethernet frame size of 64 bytes applies and dictates the minimum size of the PDU sent over the air interface. NOTE 6: It applies to both UL and DL unless stated otherwise. NOTE 7: It applies to UL.
162c3f0c36a54745dc509a8130708d2d
22.104
5.3 Aperiodic deterministic communication
Aperiodic deterministic communication is without a pre-set sending time, but still with stringent requirements on timeliness and availability of the communication service. A description of aperiodic deterministic communication can be found in Clauses 4.3 and 4.4. Additional information on the underlying use cases of the sets of requirements in Table 5.3-1 can be found in Annex A. Further information on characteristic parameters and influence quantities used in Table 5.3-1 can be found in Annex C. The 5G system shall be able to provide aperiodic deterministic communication with the service performance requirements for individual logical communication links that realise the communication services reported in Table 5.3-1. Table 5.3-1: Aperiodic deterministic communication service performance requirements Characteristic parameter (KPI) Influence quantity Communication service availability Communication service reliability: mean time between failures Max Allowed End-to-end latency (note 1) (note 5) Service bit rate: user-experienced data rate (note 5) Message size [byte] (note 5) Survival time UE speed (note 6) # of UEs Service Area (note 3) Remarks > 99.999 9 % ~ 1 week 10 ms UL: > 10 Mbit/s – – ≤ 50 km/h ≤ 2,000 ≤ 1 km2 Mobile robots – video streaming (A.2.2.3) 99.999 9 % to 99.999 999 % ~ 1 month < 30 ms > 5 Mbit/s – – < 8 km/h (linear movement) TBD TBD Mobile control panels - parallel data transmission (A.2.4.1) 99.999 999 % 1 day <8 ms (note 8) 250 kbit/s 40 to 250 16 ms quasi-static; up to 10 km/h 2 or more 30 m x 30 m Mobile Operation Panel: Emergency stop (emergency stop events) (A.2.4.1A) 99.999 9 % – < 50 ms 0.59 kbit/s 28 kbit/s < 100 – stationary 10 km‑² to 100 km‑² TBD Smart grid millisecond level precise load control (A.4.5) > 99.9 % ~ 1 month < 10 ms – – – < 8 km/h (linear movement) ≥ 3 20 m x 20 m x 4 m Augmented reality; bi-directional transmission to image processing server (A.2.4.2) 99.999 9 % to 99.999 999 % ~ 10 years < 1 ms (note 4) 25 Mbit/s – – stationary 2 to 5 100 m x 30 m x 10 m Wired-2-wireless 100 Mbit/s link replacement (A.2.2.4) 99.999 9 % to 99.999 999 % ~ 10 years < 1 ms (note 4) 500 Mbit/s – – stationary 2 to 5 100 m x 30 m x 10 m Wired-2-wireless 1 Gbit/s link replacement (A.2.2.4) > 99.9 % – DL: < 10 ms UL:<1 s (rural) DL: > 100 kbit/s UL: > 5 Gbit/s (note 9) – – stationary > 100 Distributed energy storage; energy storage station video (A.4.6) > 99.99 % – < 100 ms (note 10); DL:<1 Mbit/s – – – – – Advanced metering (A.4.7) > 99.999 % – 20 ms – < 100 byte – – – several km2 Distributed automated switching for isolation and service restoration (A.4.4.1) (note 7) > 99.999 9 % < 3 ms – 160 byte – – – – Distributed Energy Resources (DERs) and micro-grids (A.4.9) (note 7) NOTE 1: Unless otherwise specified, all communication includes 1 wireless link (UE to network node or network node to UE) rather than two wireless links (UE to UE). NOTE 2: (void) NOTE 3: Length x width x height. NOTE 4: Scheduled aperiodic traffic with transfer interval (max end-to-end allowed latency < transfer interval). NOTE 5: It applies to both UL and DL unless stated otherwise. NOTE 6: It applies to both linear movement and rotation unless stated otherwise. NOTE 7: Communication includes two wireless links (UE to UE). NOTE 8: The mobile operation panel is connected wirelessly to the 5G system. If the mobile robot/production line is also connected wirelessly to the 5G system, the communication includes two wireless links. NOTE 9: The service bit rate in one energy storage station can be calculated as follows:12.5 Mbytes/s x 50 containers x 8 = 5 Gbit/s. NOTE 10: The maximum allowed end-to-end latency is for accuracy fee control. It is the delay for one-way communication between the backend system and the 5G IoT device. The distance between the two is 40 km or lower (city range).
162c3f0c36a54745dc509a8130708d2d
22.104
5.4 Non-deterministic communication
Non-deterministic communication subsumes all other traffic types than periodic/aperiodic deterministic communication. This includes periodic/aperiodic non-real-time traffic. A description of non-deterministic communication can be found in Clauses 4.3 and 4.4. Additional information on the underlying use cases of the sets of requirements in Table 5.4‑1 can be found in Annex A. Further information on characteristic parameters and influence quantities used in Table 5.4-1 can be found in Annex C. The 5G system shall be able to provide non-deterministic communication with the service performance requirements for individual logical communication links that realise the communication services reported in Table 5.4-1. Table 5.4-1: Non-deterministic communication service performance requirements Characteristic parameter (KPI) Influence quantity Communication service reliability: mean time between failures Service bit rate: user-experienced data rate UE speed (note 2) # of UEs Service area (note 1) Remarks ~ 1 month DL: ≥ 1 Mbit/s ~ 0 km/h ≤ 75 km/h ≤ 100 50 m x 10 m x 10 m Motion control - software updates (A.2.2.1) UL: > 10 Mbit/s ≤ 50 km/h (linear movement) ≤ 2,000 ≤ 1 km2 Mobile robots; real-time video stream (A.2.2.3) NOTE 1: Length x width x height NOTE 2: It applies to both linear movement and rotation unless stated otherwise.
162c3f0c36a54745dc509a8130708d2d
22.104
5.5 Mixed traffic
Mixed traffic cannot be assigned to one of the other communication patterns exclusively. Additional information on the underlying use cases of the sets of requirements in Table 5.5-1 can be found in Annex A. Further information on characteristic parameters and influence quantities used in Table 5.5-1 can be found in Annex C. The 5G system shall be able to provide mixed traffic communication with the service performance requirements for individual logical communication links that realise the communication services reported in Table 5.5-1. Table 5.5-1: Mixed traffic communication service performance requirements Characteristic parameter (KPI) Influence quantity Remarks Communication service availability Communication service reliability: mean time between failures Max Allowed End-to-end latency (note 1) (note 3) Service bit rate: aggregate user-experienced data rate Message Size [byte] Survival time UE speed # of UEs Service Area 99.999 999 9 % ~ 10 years 16 ms stationary < 1,000 several km² Wind power plant – control traffic (A.5.2) 99.999 9 % to 99.999 99 % 1 day (note 4) 12 Mbit/s 250 to 1,500 quasi-static; up to 10 km/h 2 or more 30 m x 30 m Mobile Operation Panel: Manufacturing data stream (A.2.4.1A) NOTE 1: Unless otherwise specified, all communication includes 1 wireless link (UE to network node or network node to UE) rather than two wireless links (UE to UE). NOTE 2: (void) NOTE 3: It applies to both UL and DL unless stated otherwise. NOTE 4: The mobile operation panel is connected wirelessly to the 5G system. If the mobile robot/production line is also connected wirelessly to the 5G system, the communication includes two wireless links.
162c3f0c36a54745dc509a8130708d2d
22.104
5.6 Clock synchronisation requirements
162c3f0c36a54745dc509a8130708d2d
22.104
5.6.0 Description
Clock synchronicity, or time synchronization precision, is defined between a sync master and a sync device. The requirement on the synchronicity budget for the 5G system is the time error contribution between ingress and egress of the 5G system on the path of clock synchronization messages. Clock synchronisation requirements specific for direct device connection and indirect network connection are captured in section 7.2.3 and 8.2.3.
162c3f0c36a54745dc509a8130708d2d
22.104
5.6.1 Clock synchronisation service level requirements
The 5G system shall support a mechanism to process and transmit IEEE 1588v2 / Precision Time Protocol messages to support 3rd-party applications which use this protocol. The 5G system shall support a mechanism to synchronise the user-specific time clock of UEs with a global clock. The 5G system shall support a mechanism to synchronize the user-specific time clock of UEs with a working clock. The 5G system shall support two types of synchronization clocks, the global time domain and the working clock domains. The 5G system shall support networks with up to 128 working clock domains (with different synchronization domain identifiers / domain numbers), including for UEs connected through the 5G network. NOTE 1: The domain number (synchronization domain identifier) is defined with one octet in IEEE 802.1AS [22]. The 5G system shall be able to support up to four simultaneous synchronization domains on a UE. NOTE 1A: The four synchronization domains are used, for example, as two synchronization domains for global time and two working clock domains. One pair of global time and working clock is used as redundant synchronization domains for zero failover time. The synchronicity budget for the 5G system within the global time domain shall not exceed 900 ns. NOTE 2: The global time domain requires in general a precision of 1 µs between the sync master and any device of the clock domain. Some use cases require only a precision of ≤ 100 µs for the global time domain if a working clock domain with precision of ≤ 1 µs is available. NOTE 3: (void) The synchronicity budget for the 5G system within a working clock domain shall not exceed 700 ns. NOTE 4: The working clock domains require a precision of ≤ 1 µs between the sync master and any device of the clock domain. NOTE 5: Different working clock domains are independent and can have different precision. NOTE 6: The synchronicity budget for the 5G system is also applicable when the flow of clock synchronization messages traverses the air interface twice. The 5G system shall provide a media-dependent interface for one or multiple IEEE 802.1AS sync domains [22]. The 5G system shall provide an interface to the 5G sync domain which can be used by applications to derive their working clock domain or global time domain (Reference Clock Model). The 5G system shall provide an interface at the UE to determine and to configure the precision and time scale of the working clock domain. The 5G system shall be able to support arbitrary placement of sync master functionality and sync device functionality in integrated 5G / non-3GPP TSN networks. The 5G system shall be able to support clock synchronization through the 5G network if the sync master and the sync devices are served by different UEs. (Flow of clock synchronization messages is in either direction, UL and DL.) The 5G system shall provide a suitable means to support the management of the merging and separation of working clock domains, that is interoperable with the corresponding mechanisms of TSN and IEEE 802.1AS. The 5G system shall provide a suitable means to support precise time distribution, clock synchronization and synchrophasor communication functionalities specific to smart grid applications, e.g.: - IEC 61850-9-3 [30] profile and IEEE Std C37.238-2017 [31]. - at least one of the two profiles for synchrophasor communications: IEC 61850-90-5:2012 [32], or IEEE Std C37.118.2-2011 [33]. 7NOTE 7: The requirement above assumes the ability to support all main clock synchronization functionalities defined in the respective references, unless explicitly indicated. Support of IEC 61850-90-5 can entail supporting multiple profiles (e.g. A-profile, T-profile, KDC-profile and IEEE 802.1Q QoS profiles can be applicable for support of synchrophasor communications).
162c3f0c36a54745dc509a8130708d2d
22.104
5.6.2 Clock synchronisation service performance requirements
Table 5.6.2-1: Clock synchronization service performance requirements for 5G System User-specific clock synchronicity accuracy level Number of devices in one communication group for clock synchronisation 5GS synchronicity budget requirement (note 1) Service area Scenario 1 up to 300 UEs ≤ 700 ns (note 4) ≤ 100 m x 100 m • Motion control (A.2.2.1) • Control-to-control communication for industrial controller (A.2.2.2) 2 up to 300 UEs ≤ 700 ns (note 4) ≤ 1,000 m x 100 m • Control-to-control communication for industrial controller (A.2.2.2) 3 up to 10 UEs < 10 µs ≤ 2,500 m2 • High data rate video streaming 3a up to 100 UEs < 1 µs ≤ 10 km2 • AVPROD synchronisation and packet timing 4 up to 100 UEs < 1 µs < 20 km2 • Smart Grid: synchronicity between PMUs 4a up to 100 UEs < 250 ns to 1 µs < 20 km² Smart Grid: IEC 61850-9-2 Sampled Values 4b up to 100 UEs <10-20 µs < 20 km² Smart Grid: IEC 61850-9-2 Sampled Values – Power system protection in digital substation 4c 54/km² (note 2) 78/km2 (note 3) < 10 µs several km² Smart Grid: Intelligent Distributed Feeder Automation (A.4.4.3) 4d up to 100 UEs <1 ms < 20 km² Smart Grid: IEC 61850-9-2 Sampled Values – Event reporting and Disturbance recording 5 up to 10 UEs < 50 µs 400 km • Telesurgery (A.6.2) and telediagnosis (A.6.3) NOTE 1: The clock synchronicity requirement refers to the clock synchronicity budget for the 5G system, as described in Clause 5.6.1. NOTE 2: When the distributed terminals are deployed along overhead line, about 54 terminals will be distributed along overhead lines in one square kilometre. The resulting power load density is 20 MW/km2. NOTE 3: When the distributed terminals are deployed in power distribution cabinets, there are about 78 terminals in one square kilometre. The resulting power load density is 20 MW/km2. NOTE 4: 5GS synchronicity budget requirement refers to working clock domain for a localized set of UEs callaborating on a specific task or work function. 5.6A Time-sensitive communication requirements The 5G system shall support the fully distributed model for configuration of time-sensitive networking. The 5G system shall support the fully distributed model for configuration of time-sensitive networking that is aligned with Multiple Stream Registration Protocol (MSRP, IEEE 802.1Q [19] clause 35.1), IEEE P802.1CS Link-local Registration Protocol (LRP) [24], and IEEE P802.1Qdd Resource Allocation Protocol (RAP) [25]. The 5G system shall support the user-network / network-network interface for the dynamic configuration of the fully distributed model for time-sensitive networking. 5.6B 5G Timing Resiliency To enable support of many critical services within the 5G network, additional requirements and KPIs that enhance the 5G system with timing resiliency are specified in TS 22.261 [2] clauses 6.36 and 7.8. Those enhancements enable use of the 5G system for time critical services in collaboration with or as a backup to other timing solution such as loss or degradation of GNSS reference timing. 5.6C Support for infrastructure protection of electrical transmission 5.6C.1 Description Transmission infrastructure is a key component of the energy system. Communication enables protection of this infrastructure. The algorithms involved depend on certain constraints must be met, particularly concerning the end-to-end latency. 5.6C.2 Requirements The 5G system shall support an end-to-end latency of less than 5 ms or 10 ms, as requested by the UE initiating the communication. NOTE 1: Whether the end-to-end latency is 5 ms or 10 ms depends on the applied voltage level. NOTE 2: The end-to-end latency is between two UEs, including two wireless links. The 5G system shall support communication channel symmetry in terms of end-to-end latency (latency from UE1 to UE2, and end-to-end latency from UE2 to UE1), with an asymmetry of < 2ms.
162c3f0c36a54745dc509a8130708d2d
22.104
5.7 Positioning performance requirements
162c3f0c36a54745dc509a8130708d2d
22.104
5.7.1 General requirements
High accuracy positioning is becoming essential for Factories of the Future. The reason for this is that tracking of mobile devices as well as mobile assets is becoming increasingly important in improving processes and increasing flexibility in industrial environments. The 5G system shall provide positioning information for a UE that is out of coverage of the network, with accuracy of < [1 m] relative to other UEs that are in proximity and in coverage of the network. Table 5.7.1-1 below lists typical scenarios and the corresponding high positioning requirements for horizontal and vertical accuracy, availability, heading, latency, and UE speed. NOTE: The column on "Corresponding Positioning Service Level in TS 22.261" maps the scenarios listed in Table 5.7.1-1 to the service levels defined in TS 22.261 [2]. Table 5.7.1-1: Positioning performance requirements Scenario Horizontal accuracy Vertical accuracy Availability Heading Latency for position estimation of UE UE speed Corresponding Positioning Service Level in TS 22.261 Mobile control panels with safety functions (non-danger zones) < 5 m < 3 m 90 % n/a < 5 s n/a Service Level 2 Process automation – plant asset management < 1 m < 3 m 90 % n/a < 2 s < 30 km/h Service Level 3 Flexible, modular assembly area in smart factories (for tracking of tools at the work-place location) < 1 m (relative positioning) n/a 99 % n/a 1 s < 30 km/h Service Level 3 Augmented reality in smart factories < 1 m < 3 m 99 % < 0.17 rad < 15 ms < 10 km/h Service Level 4 Mobile control panels with safety functions in smart factories (within factory danger zones) < 1 m < 3 m 99.9 % < 0.54 rad < 1 s n/a Service Level 4 Flexible, modular assembly area in smart factories (for autonomous vehicles, only for monitoring purposes) < 50 cm < 3 m 99 % n/a 1 s < 30 km/h Service Level 5 Inbound logistics for manufacturing (for driving trajectories (if supported by further sensors like camera, GNSS, IMU) of indoor autonomous driving systems)) < 30 cm (if supported by further sensors like camera, GNSS, IMU) < 3 m 99.9 % n/a 10 ms < 30 km/h Service Level 6 Inbound logistics for manufacturing (for storage of goods) < 20 cm < 20 cm 99 % n/a < 1 s < 30 km/h Service Level 7
162c3f0c36a54745dc509a8130708d2d
22.104
5.7.2 Energy efficiency requirements for high accuracy positioning
The 5G system shall support low power high accuracy positioning mechanisms that allow a battery-constrained UE to sustain a long lifetime without changing battery. Some corresponding use cases and example scenarios are captured in Annex A.7.2.
162c3f0c36a54745dc509a8130708d2d
22.104
5.8 Network operation requirements
For use by Industry 4.0, the 5G system needs to meet various operational options that are not typical in a traditional mobile operator setting. Additional system requirements that enable a 5G system to support those options are included in this clause. 5G system shall provide support for reliable communications when a UE serves as a TSN talker or listener so there is no single point of service failure.
162c3f0c36a54745dc509a8130708d2d
22.104
6 Ethernet applications
162c3f0c36a54745dc509a8130708d2d
22.104
6.1 Description
This section lists the requirements applicable to the 5G system for supporting cyber-physical applications using Ethernet. For requirements pertaining to common, fundamental Ethernet transport requirements, and any requirements necessary to support the 5G LAN-type service, see Clause 6.24 in TS 22.261 [2].
162c3f0c36a54745dc509a8130708d2d
22.104
6.2 Requirements
For infrastructure dedicated to high performance Ethernet applications, the 3GPP system shall support clock synchronisation defined by IEEE 802.1AS across 5G-based Ethernet links with PDU-session type Ethernet. For infrastructure dedicated to high performance Ethernet applications, the 3GPP system shall support clock synchronisation defined by IEEE 802.1AS across 5G-based Ethernet links and other ethernet transports such as wired and optical (EPON.) For infrastructure dedicated to high performance Ethernet applications, the accuracy of clock synchronisation should be better than 1µs. For infrastructure dedicated to high performance Ethernet applications, the 3GPP system shall support enhancements for time-sensitive networking as defined by IEEE 802.1Q, e.g. time-aware scheduling with absolute cyclic time boundaries defined by IEEE 802.1Qbv [19], for 5G-based Ethernet links with PDU sessions type Ethernet. For infrastructure dedicated to high performance Ethernet applications, absolute cyclic time boundaries shall be configurable for flows in DL direction and UL direction. For infrastructure dedicated to high performance Ethernet applications, the 3GPP system shall support coexistence of hard-RT traffic following a time-aware schedule and lower priority traffic. The lower priority traffic cannot have a performance degrading impact on the hard-RT traffic. The Ethernet transport service shall support routing based on information extracted from the Ethernet header information created based on IEEE 802.1Qbv.
162c3f0c36a54745dc509a8130708d2d
22.104
7 Direct device connection for cyber-physical control applications
162c3f0c36a54745dc509a8130708d2d
22.104
7.1 Description
This section lists the requirements applicable to the 5G system for supporting cyber-physical control applications using wireless direct device connection.
162c3f0c36a54745dc509a8130708d2d
22.104
7.2 Requirements
162c3f0c36a54745dc509a8130708d2d
22.104
7.2.1 General
The 5G system shall allow UEs to use direct device connection when the UEs are not served by a RAN. The 5G system shall be able to support direct device connection between UEs in close proximity using spectrum different than the spectrum being used for the 5GC-based communication. The 5G system shall be able to support direct device connection for 5G LAN-type private communication. The 5G system shall be able to support multicast communication between the UEs within the group of UEs using direct device connection .
162c3f0c36a54745dc509a8130708d2d
22.104
7.2.2 Network performance
The 5G system shall be able to support direct device connection between a group of UEs for periodic deterministic communication (both unicast and multicast) with respective service performance requirements in Table 5.2-1 related to cooperative carrying. The 5G system shall be able to support mobility of the group of UEs using direct device connections with respective service performance requirements in Table 5.2-1 related to cooperative carrying. The 5G system shall be able to support direct device connection with respective service performance requirements in Table 5.2-1 related to cooperative carrying between UEs up to 50 m distance.
162c3f0c36a54745dc509a8130708d2d
22.104
7.2.3 Clock synchronization
162c3f0c36a54745dc509a8130708d2d
22.104
7.2.3.1 Description
This clause lists the service level requirements and performance requirements applicable to the 5G system for supporting clock synchronization for direct device connection, building on the description already provided in Clause 5.6.0.
162c3f0c36a54745dc509a8130708d2d
22.104
7.2.3.2 Clock synchronization requirements
The 5G system shall be able to support clock synchronization (working clock domain) between the UEs within the group of UEs using direct device connection ProSe communication. The 5G system shall be able to support Precision Time Protocol-based (IEEE 802.1AS [22] or another applicable IEEE Std 1588 [34] profile) or 5G sync domain-based clock synchronization among the group of UEs using direct device connection. The 5G system shall be able to support the sync master of the working clock domain being connected to one of the UEs or being hosted at one of the UEs in the group of UEs using direct device connection. The 5G system shall be able to support up to four simultaneous synchronization domains on a UE using direct device connection. The 5G system using direct device connection shall support one or multiple time domains (IEEE 802.1AS [22] or another applicable IEEE Std 1588 [34] profile or 5G sync domain configuration). The 5G system shall provide a suitable means to support the merging and separation of working clock domains at the UEs within the group of UEs connected by direct device connection, that is interoperable with the corresponding mechanisms of IEEE 802.1AS [22], IEEE Std 1588 [34], or 5G sync domain. For direct device connection, the 5G system shall be able to support a 5GS synchronicity budget for clock synchronization according to Table 7.2.3.2-1. In this case, the sync master and sync device are located at or connected to two UEs which are connected via direct device connection. Table 7.2.3.2-1: Direct device connection clock synchronization service performance requirements for 5G System Number of devices in one communication group for clock synchronization 5GS synchronicity budget requirement (note 1) Service area (note 2) Scenario 2 to 16 ≤ 700 ns 10 m x 10 m x 5 m; 50 m x 5 m x 5 m Cooperative carrying – fragile work pieces; (A.2.2.5) 2 to 16 ≤ 700 ns 10 m x 10 m x 5 m; 50 m x 5 m x 5 m Cooperative carrying – elastic work pieces; (A.2.2.5) NOTE 1: 5G synchronicity budget is the time error between ingress and egress of the 5G system on the path of clock synchronization messages (as described in Clause 5.6.0). For direct device connection, the ingress is one UE and the egress the other directly connected UE. NOTE 2: Service Area for direct device connections between UEs (length x width x height). The group of UEs with direct device connections might move throughout the whole factory site (up to several km²).
162c3f0c36a54745dc509a8130708d2d
22.104
7.2.4 Service Continuity
The 5G system shall be able to minimize service disruption for a group of UEs using direct device connection with respective service performance requirements in Table 5.2-1 related to cooperative carrying, when the group moves between a non-public network and a PLMN (subject to operator policies and agreement between the operators and service providers).
162c3f0c36a54745dc509a8130708d2d
22.104
7.2.5 Direct device connection via UE to UE relay
The 5G system shall be able to support direct device connection via an UE to UE relay with respective service performance requirements in Table 5.2-1 related to cooperative carrying between UEs out of transmission range of each other.
162c3f0c36a54745dc509a8130708d2d
22.104
8 Indirect network connection for cyber-physical control applications
162c3f0c36a54745dc509a8130708d2d
22.104
8.1 Description
This section lists the requirements applicable to the 5G system for supporting cyber-physical control applications using wireless indirect network connection.
162c3f0c36a54745dc509a8130708d2d
22.104
8.2 Requirements
162c3f0c36a54745dc509a8130708d2d
22.104
8.2.1 General
The 5G system shall be able to support multicast communication between the UEs within the group of UEs using indirect network connection.
162c3f0c36a54745dc509a8130708d2d
22.104
8.2.2 Communication via indirect network connection
The 5G system shall be able to provide service to an out-of-coverage UE via indirect network connection using one relay UE while meeting the performance requirements specified for the process automation use cases in Table 5.2-1 (related to Annex A.2.3.1 and A.2.3.2).
162c3f0c36a54745dc509a8130708d2d
22.104
8.2.3 Clock synchronization
162c3f0c36a54745dc509a8130708d2d
22.104
8.2.3.1 Description
This clause lists the service level requirements and performance requirements applicable to the 5G system for supporting clock synchronization for indirect network connection, building on the description already provided in Clause 5.6.0.
162c3f0c36a54745dc509a8130708d2d
22.104
8.2.3.2 Clock synchronization requirements
The 5G system shall be able to support clock synchronization (working clock domain) for UEs using indirect network connection. The 5G system shall be able to support Precision Time Protocol-based (IEEE 802.1AS [22] or another applicable IEEE Std 1588 [34] profile) or 5G sync domain-based clock synchronization for UEs using indirect network connection. The 5G system shall be able to support the sync master of the working clock domain being connected to one of the UEs or being hosted at one of the UEs using indirect network connection. The 5G system shall be able to support up to four simultaneous synchronization domains on a UE using indirect network connection. The 5G system shall provide a suitable means to support the merging and separation of working clock domains at the UEs connected by indirect network connection, that is interoperable with the corresponding mechanisms of IEEE 802.1AS [22] or another applicable IEEE Std 1588 [34] profile or 5G sync domain. For indirect network connection between UEs using one UE-to-network relay, the 5G system shall be able to support a 5GS synchronicity budget for clock synchronization according to Table 8.2.3.2-1. Table 8.2.3.2-1: Indirect network connection clock synchronization service performance requirements for 5G System Number of devices in one communication group for clock synchronization UE density [UE / m²] 5GS synchronicity budget requirement (note 1) Service area (note 2) Scenario 10 to 20 UEs - < 1 ms ≤ 100 m x 100 m x 50 m Process automation – closed loop control (A.2.3.1) ≤ 10,000 to 100,000 - ≤ 1 ms ≤ 10 km x 10 km x 50 m Process and asset monitoring (A.2.3.2) NOTE 1: 5G synchronicity budget is the time error between ingress and egress of the 5G system on the path of clock synchronization messages (as described in Clause 5.6.0). For indirect network connection, 3 cases are considered: 1) If the path of clock synchronization messages is between device and network side, ingress and egress of the 5G system are the remote UE and the corresponding UPF on the network side. 2) If ingress and egress of the 5G system are at the device side, the 5G synchronicity budget is the time error between the involved remote UE and the 5G sync master. 3) If the sync master is inside the 5G system, the 5G synchronicity budget ingress is the sync master in the 5G system and egress is the remote UE. NOTE 2: Service Area for indirect network connections between UEs (length x width x height).
162c3f0c36a54745dc509a8130708d2d
22.104
8.2.4 Service Continuity
The 5G system shall be able to minimize service disruption for a UE using indirect network connection with respective service performance requirements in Table 5.2-1 related to process automation, when the UE moves between a non-public network and a PLMN (subject to operator policies and agreement between the operators and service providers).
162c3f0c36a54745dc509a8130708d2d
22.104
9 Recovery of infrastructure for electrical distribution
162c3f0c36a54745dc509a8130708d2d
22.104
9.1 Description
The robustness of the infrastructure for electrical power distribution may depend upon the possibility to operate telecommunication networks even during an energy system incident, in which electricity cannot be delivered to some network operator facilities. Through coordination between the network operator and the energy system operator, increases in the ability to recover the energy system operation can be achieved.
162c3f0c36a54745dc509a8130708d2d
22.104
9.2 Requirements
Subject to regulatory requirements and operator policy, the 5G system shall support a mechanism by which an MNO can identify the ability of the MNO's infrastructure to continue operation despite a lack of electrical supply service, specifying which physical regions would be affected in terms of physical topology and the remaining time in which operation is possible. NOTE1: This information can facilitate energy distribution system recovery operations. Subject to regulatory requirements, the 5G system shall support a mechanism by which a third party can, in the event of an energy distribution system service interruption, communicate the energy distribution system recovery status in terms of location and time table to the MNO. NOTE2: This information can facilitate MNO operations to facilitate energy system recovery. Annex A (informative): Summary of service performance requirements A.1 About the vertical domains addressed in this Annex A vertical domain is an industry or group of enterprises in which similar products or services are developed, produced, and provided. The vertical domains addressed in this Annex are - Factories of the Future (A.2); - electric-power distribution (A.4); and - central power generation (A.5); and • Connected hospitals or medical facilities (A.6). A.2 Factories of the Future A.2.1 Overview The manufacturing industry is currently subject to a fundamental change, which is often referred to as the "Fourth Industrial Revolution" or simply "Industry 4.0" [15]. The main goals of Industry 4.0 are―among others―the improvement of flexibility, versatility, resource efficiency, cost efficiency, worker support, and quality of industrial production and logistics. These improvements are important for addressing the needs of increasingly volatile and globalised markets. A major enabler for all this is cyber-physical production systems based on a ubiquitous and powerful connectivity, communication, and computing infrastructure. The infrastructure interconnects people, machines, products, and all kinds of other devices in a flexible, secure and consistent manner. Several different application areas can be distinguished: 1) Factory automation: Factory automation deals with the automated control, monitoring and optimisation of processes and workflows within a factory. This includes aspects like closed-loop control applications (e.g., based on programmable logic or motion controllers) and robotics, as well as aspects of computer-integrated manufacturing. Factory automation generally represents a key enabler for industrial mass production with high quality and cost-efficiency. Corresponding applications are often characterised by highest requirements on the underlying communication infrastructure, especially in terms of communication service availability, determinism, and latency. In the Factories of the Future, static sequential production systems will be more and more replaced by novel modular production systems offering a high flexibility and versatility. This involves many increasingly mobile production assets, for which powerful wireless communication and localisation services are required. 2) Process automation: Process automation refers to the control of production and handling of substances like chemicals, food & beverage, pulp, etc. Process automation improves the efficiency of production processes, energy consumption, and safety of the facilities. Sensors measuring process values, such as pressures or temperatures, are working in closed loops via centralised and decentralised controllers. In turn, the controllers interact with actuators, e.g., valves, pumps, heaters. Also, monitoring of attributes such as the filling levels of tanks, quality of material, or environmental data are important, as well as safety warnings or plant shut downs. Workers in the plant are supported by mobile devices. A process automation facility may range from a few 100 m² to several km², and the facility may be geographically distributed. Depending on the size, a production plant may have several 10,000 measurement points and actuators. Autarkic device power supply for years is needed in order to stay flexible and to keep the total costs of ownership low. 3) HMIs and production IT: Human-machine interfaces (HMIs) include all sorts of devices for the interaction between people and production facilities, such as panels attached to a machine or production line, but also standard IT devices, such as laptops, tablet PCs, smartphones, etc. In addition, augmented- and virtual-reality applications are expected to play an increasingly important role in future. 4) Logistics and warehousing: Organisation and control of the flow and storage of materials and goods in the context of industrial production. In this respect, intra-logistics is dealing with logistics within a certain property (e.g., within a factory), for example by ensuring the uninterrupted supply of raw materials on the shop floor level using automated guided vehicles (AGVs), fork-lifts, etc. This is to be seen in contrast to logistics between different sites. Warehousing particularly refers to the storage of materials and goods, which is also getting more and more automated, for example based on conveyors, cranes and automated storage and retrieval systems. 5) Monitoring and maintenance: Monitoring of certain processes and/or assets in the context of industrial production without an immediate impact on the processes themselves (in contrast to a typical closed-loop control system in factory automation, for example). This particularly includes applications such as condition monitoring and predictive maintenance based on sensor data, but also big data analytics for optimising future parameter sets of a certain process, for instance. For these use cases, the data acquisition process is typically not latency-critical. For each of these application areas, a multitude of potential use cases exists, some of which are outlined in the following subclauses. These use cases can be mapped to the given application areas (see Table A.2.1-1). Table A.2.1-1: Mapping of the considered use cases (columns) to application areas (rows) Motion control Control-to-control Mobile control panels with safety Mobile robots Remote access and maintenance Augmented reality Closed-loop process control Process monitoring Plant asset management Factory automation X X X Process automation X X X X HMIs and Production IT X X Logistics and warehousing X X X Monitoring and maintenance X A.2.2 Factory automation A.2.2.1 Motion control A motion control system is responsible for controlling moving and/or rotating parts of machines in a well-defined manner, for example in printing machines, machine tools or packaging machines. A schematic representation of a motion control system is depicted in Figure A.2.2.1-1. A motion controller periodically sends desired set points to one or several actuators (e.g., a linear actuator or a drive) which thereupon perform a corresponding action on one or several processes (in this case usually a movement or rotation of a certain component). At the same time, sensors determine the current state of the process(es), e.g. the current position and/or rotation of one or multiple components, and send the actual values back to the motion controller. This is done in a strictly cyclic and deterministic manner, such that during one application cycle the motion controller sends updated set points to all actuators, and all sensors send their actual values back to the motion controller. Nowadays, typically Industrial Ethernet technologies are used for motion control systems. Figure A.2.2.1-1: Schematic representation of a motion control system Table A.2.2.1-1: Service performance requirements for motion control Use case # Characteristic parameter Influence quantity Communication service availability: target value [%] Communication service reliability: mean time between failures End-to-end latency: maximum Service bitrate: user experienced data rate Message size [byte] Transfer interval: lower bound Transfer interval: upper bound Survival time UE speed # of UEs Service area (note) 1 99.999 to 99.999 99 ~ 10 years < transfer interval value – 50 500 μs – 500 ns 500 μs + 500 ns 500 μs ≤ 72 km/h ≤ 20 50 m x 10 m x 10 m 2 99.999 9 to 99.999 999 ~ 10 years < transfer interval value – 40 1 ms – 500 ns 1 ms + 500 ns 1 ms ≤ 72 km/h ≤ 50 50 m x 10 m x 10 m 3 99.999 9 to 99.999 999 ~ 10 years < transfer interval value – 20 2 ms – 500 ns 2 ms + 500 ns 2 ms ≤ 72 km/h ≤ 100 50 m x 10 m x 10 m NOTE: Length x width x height. Use cases one to three Characteristic parameters and influence quantities for a communication service supporting the cyclic interaction described above. A.2.2.2 Control-to-control communication Control-to-control communication, i.e., the communication between different industrial controllers is already used today for different use cases, such as: - large machines (e.g., newspaper printing machines), where several controls are used to cluster machine functions, which need to communicate with each other; these controls typically need to be synchronised and exchange real-time data; - individual machines that are used for fulfilling a common task (e.g., machines in an assembly line) often need to communicate, for example for controlling and coordinating the handover of work pieces from one machine to another. Typically, a control-to-control network has no fixed configuration of certain controls that need to be present. The control nodes present in the network often vary with the status of machines and the manufacturing plant. Therefore, hot-plugging support for different control nodes is important and often used. Table A.2.2.2-1: Service performance requirements for control-to control communication in motion control Use case # Characteristic parameter Influence quantity Communication service availability: target value [%] Communication service reliability: mean time between failures End-to-end latency: maximum Message size [byte] Transfer interval Survival time UE speed # of UEs Service area (note 1) 1 (note 2) 99.999 9 to 99.999 999 ~ 10 years < transfer interval value 1 k ≤ 10 ms 10 ms stationary 5 to 10 100 m x 30 m x 10 m 2 (note 2) 99.999 9 to 99.999 999 ~ 10 years < transfer interval value 1 k ≤ 50 ms 50 ms stationary 5 to 10 1,000 m x 30 m x 10 m NOTE 1: Length x width x height. NOTE 2: Communication may include two wireless links (UE to UE) Use case one Control-to-control communication between different motion (control) subsystems, as addressed in Subclause A.2.2.1. An exemplary application for this is large printing machines, where it is not possible or desired to control all actuators and sensors by one motion controller only. Use case two Control-to-control communication between different motion (control) subsystems. Exemplary application for this are extra-large machines or individual machines used for fulfilling a common task (e.g., machines in an assembly line). A.2.2.3 Mobile robots Mobile robots and mobile platforms, such as automated guided vehicles, have numerous applications in industrial and intra-logistics environments and will play an increasingly important role in the Factory of the Future. A mobile robot essentially is a programmable machine able to execute multiple operations, following programmed paths to fulfil a large variety of tasks. This means, a mobile robot can perform activities like assistance in work steps, collaboration with other robots, e.g. for car assembly, and transport of goods, materials and other objects. Mobile robot systems are characterised by a maximum flexibility in mobility relative to the environment, with a certain level of autonomy and perception ability, i.e., they can sense and react with their environment. Autonomous guided vehicles (AGVs) are a sub-group of mobile robots. AGVs are driverless and used for moving materials efficiently within a facility. A detailed overview of the state of the art of autonomous-guided-vehicle systems is provided elsewhere in the literature [16]. All mobile robots incorporate all functionalities needed for an AGV. Today, the AGV’s control intelligence is hosted inside the AGV. In the future, centralised fleet control will be hosted in the edge cloud, which will require reliable wireless communication between the control entity and all AGVs. Also, the current paradigm of pre-describing a route for the AGV will be replaced with target-based navigation. This paradigm change will make AGV routes more flexible. Mobile robots are monitored and controlled from a guidance control system. Radio-controlled guidance control is necessary to get up-to-date process information, to avoid collisions between mobile robots, to assign driving jobs to the mobile robots, and to manage the traffic of mobile robots. The mobile robots are track-guided by the infrastructure with markers or wires in the floor or guided by own surround sensors, like cameras and laser scanners. Mobile robot systems can be divided in operation in indoor, outdoor and both indoor and outdoor areas. These environmental conditions have an impact on the requirements of the communication system, e.g., the handover process, to guarantee the required cycle times. Where this document does not explicitly refer to AGVs, the term mobile robots applies to AGVs as well as to mobile robots. Table A.2.2.3-1: Service performance requirements for mobile robots Use case # Characteristic parameter Influence quantity Communication service availability: target value [%] Communication service reliability: mean time between failures End-to-end latency: maximum Service bitrate: user experienced data rate Message size [byte] Transfer interval: lower bound Transfer interval: target value (note) Transfer interval: upper bound Survival time UE speed # of UEs Service area 1 > 99.999 9 ~ 10 years < target transfer interval value – 40 to 250 – < 25 % of target transfer interval value 1 ms to 50 ms + < 25 % of target transfer interval value target transfer interval value ≤ 50 km/h ≤ 2,000 ≤ 1 km2 2 > 99.999 9 ~ 1 year < target transfer interval value – 15 k to 250 k – < 25 % of target transfer interval value 10 ms to 100 ms + < 25 % of target transfer interval value target transfer interval value ≤ 50 km/h ≤ 2,000 ≤ 1 km2 3 > 99.999 9 ~ 1 year < target transfer interval value – 40 to 250 – < 25 % of target transfer interval value 40 ms to 500 ms + < 25 % of target transfer interval value target transfer interval value ≤ 50 km/h ≤ 2,000 ≤ 1 km2 4 > 99.999 9 ~ 1 week 10 ms > 10 Mbit/s – – – – ≤ 50 km/h ≤ 2,000 ≤ 1 km2 NOTE: The transfer interval is not so strictly periodic in these use cases. The transfer interval deviates around its target value within bounds. The mean of the transfer interval is close to the target value. Use case one Periodic communication for the support of precise cooperative robotic motion control (transfer interval: 1 ms), machine control (transfer interval: 1 ms to 10 ms), co-operative driving (10 ms to 50 ms). Use case two Periodic communication for video-operated remote control. Use case three Periodic communication for standard mobile robot operation and traffic management. Use case four Real-time streaming data transmission (video data) from a mobile robot to the guidance control system. Additional information AGVs have the following needs. – The direct-device control is time-critical since the communication involves safety-relevant functions such as emergency stop and the avoidance of obstacles. – For the implementation of swarm intelligence, position and availability information are needed. A possible route change due to a blocked route affects the routes of all other AGVs that will follow. The communication is less time-critical than for safety-relevant functions. – Camera-based navigation requires high data rates. Examples for camera-based navigation are the positioning of boxes/containers, detection of persons and obstacles, as well as creation and administration of a map for flexible navigation. Note that sensor-based navigation requires lower data rates than camera-based navigation. – AGV route control and emergency safety related time critical latency and response can be achieved with an edge cloud where the edge infrastructure is located close to the AGVs Mobile robots have additional needs. – The mobile robot provides an additional service during transport (for instance quality inspection, scanning of surroundings, asset identification, carrying of work pieces). In order to reduce the uplink data rate, pre-compression of data is possible directly on the device. The communication is not or at least less time-critical than the motion control of the mobile robot. – The mobile robot needs to interact with the periphery (for instance intelligent storage racks, stationary robots, and moving machines). This communication is time-critical. Interaction with the periphery can be relevant at the start point, end point, and also at several intermediate stations (for instance the collection of parts from intelligent storage racks). The exact position and orientation can be determined by a centering station and the AGV sensors. Additional scanning by the robot with a video camera may be necessary. – For some mobile robots, their control intelligence might be centralized and hosted in an edge cloud. They require secure communication towards the edge cloud. If the path layout such a mobile robot follows (e.g., including indoor and outdoor) causes it to switch the communication between a private network and a public network, consideration should be given to provide the required level of security for communication. – Mobile robots whose control intelligence is centralized and hosted in the edge cloud needs privacy protection of data stored in the edge cloud. – In order to be able to process time-sensitive data from local sensors and devices (AGVs, robots etc.) in an edge cloud, the edge infrastructure needs to be on-premise or close to the factory A.2.2.4 Wired to wireless link replacement In a traditional factory, the production environment is fixed. Machines that are cooperating are connected via cable, typically using an industrial ethernet technology like PROFINET. In order to increase flexibility in the production setup, the wired links are replaced with wireless links. Figure A.2.2.4-1: Example of four cooperating machines with wireless connections (based on [26]) We assume two or more machines (typically 4 or 5) to be cooperating with each other during production. In order to replace the cables, each machine is equipped with one UE, connected to the controller (shown in Figure A.2.2.4-1). The cooperating machine’s communication can be divided into two types. Periodic traffic and a-periodic traffic. Both types are scheduled, therefore the a-periodic traffic is also adhering to the transfer interval. The traffic requirements are from the point of view of the UE and give the maximum aggregate traffic of all links. Meaning, the traffic per link may change according to the number of cooperating machines but the total traffic at the UE cannot exceed the given values. Table A.2.2.4-1: Service performance requirements for wired to wireless link replacement Use case # Characteristic parameter Influence quantity Communication service availability: target value [%] Communication service reliability: mean time between failures End-to-end latency: maximum Data rate [Mbit/s] Transfer interval Survival time UE speed # of UEs Service area (note 1) 1 (periodic traffic) 99.999 9 to 99.999 999 ~ 10 years < transfer interval value 50 ≤ 1 ms 3 x transfer interval stationary 2 to 5 100 m x 30 m x 10 m 1 (aperiodic traffic) 99.999 9 to 99.999 999 ~ 10 years < transfer interval value 25 ≤ 1 ms (note 2) stationary 2 to 5 100 m x 30 m x 10 m 2 (periodic traffic) 99.999 9 to 99.999 999 ~ 10 years < transfer interval value 250 ≤ 1 ms 3 x transfer interval stationary 2 to 5 100 m x 30 m x 10 m 2 (aperiodic traffic) 99.999 9 to 99.999 999 ~ 10 years < transfer interval value 500 ≤ 1 ms (note 2) stationary 2 to 5 100 m x 30 m x 10 m NOTE 1: Length x width x height. NOTE 2: Transfer interval also applies for scheduled aperiodic traffic Use case one In the case of the 100 Mbit/s link replacement, 50 % periodic traffic and 25 % a-periodic traffic are assumed. Use case two In the case of the 1 Gbit/s link replacement, 25 % periodic traffic and 50 % a-periodic traffic are assumed. A.2.2.5 Cooperative carrying In a smart factory, large or heavy work pieces will be carried from one place to another by multiple mobile robots or AGVs. These mobile robots / AGVs need to work together in order to carry the large or heavy work piece safely. This cooperation is achieved with a cyber-physical control application that controls the drives and movements of the mobile robots / AGVs in a coordinated way, so that large or heavy work pieces are carried smoothly and safely from one place to another (see Figure 5.11.1-1). Figure A.2.2.5-1: Mobile robots / AGVs carrying a large work piece cooperatively The communication between the collaborating mobile robots / AGVs requires high communication service availability and ultra-low latency. The exchange of control commands and control feedback is done with periodic deterministic communication and using time-sensitive networking. There are two distinct use case variants of cooperative carrying: (1) carrying of rigid or fragile work pieces that require very precise coordination between the collaborating mobile robots, and (2) carrying of more flexible or elastic work pieces that allow some tolerance in the coordinated movements of the collaborative mobile robots. The higher tolerance in the coordinated movements allows for either faster movement of the work piece or longer transfer intervals (trade-off between UE speed and transfer interval). Table A.2.2.5-1: Service performance requirements for cooperative carrying Use case # Characteristic parameter Influence quantity Communication service availability: target value [%] Communication service reliability: mean time between failures End-to-end latency: maximum Service bitrate: user experienced data rate Message size [byte] Transfer interval: target value (note 1) Survival time (note 1) UE speed # of UEs Service area (note 2) 1 99.999 9 to 99.999 999 ~ 10 years < 0.5 x transfer interval 2.5 Mbit/s 250; 500 with localisation information > 5 ms > 2.5 ms > 1.7 ms 0 transfer interval 2 x transfer interval ≤ 6 km/h 2 to 8 10 m x 10 m x 5 m; 50 m x 5 m x 5 m 2 99.999 9 to 99.999 999 ~ 10 years < 0.5 x transfer interval 2.5 Mbit/s 250; 500 with localisation information > 5 ms > 2.5 ms > 1.7 ms 0 transfer interval 2 x transfer interval ≤ 12 km/h 2 to 8 10 m x 10 m x 5 m; 50 m x 5 m x 5 m NOTE 1: The first value is the application requirement, the other values are the requirement with multiple transmission of the same information (two or three times respectively). NOTE 2: Service Area for direct communication between UEs (length x width x height). The group of UEs with direct communication might move throughout the whole factory site (up to several km²) Use case one Periodic deterministic communication for cooperative carrying of fragile work pieces (UE to UE / ProSe communication). Use case two Periodic deterministic communication for cooperative carrying of elastic work pieces (UE to UE / ProSe communication). A.2.3 Process automation A.2.3.1 Closed-loop control In the closed-loop control use case for process automation, several sensors are installed in a plant and each sensor performs continuous measurements. The measurement data are transported to a controller, which takes decision to set actuators. The latency and determinism in this use case are crucial. This use case has very stringent requirements in terms of latency and service availability. The required service area is usually bigger than for motion control use cases. Interaction with the public network (e.g., service continuity, roaming) is not required. Table A.2.3.1-1: Service performance requirements for closed-loop control in process automation Use case # Characteristic parameter Influence quantity Communication service availability: target value [%] Communication service reliability: mean time between failures End-to-end latency: maximum Message size [byte] Transfer interval: lower bound Transfer interval: target value Transfer interval: upper bound Survival time UE speed # of UEs Service area (note) 1 99.999 9 to 99.999 999 ≥ 1 year < target transfer interval value 20 -5 % of target transfer interval value ≥ 10 ms +5 % of target transfer interval value 0 typically stationary typically 10 to 20 typically ≤ 100 m x 100 m x 50 m NOTE: Length x width x height. Use case one Several sensors are installed in a plant and each sensor performs continuous measurements. The measurement data are transported to a controller, which takes decision to set actuators. A.2.3.2 Process and asset monitoring For process and asset monitoring in the area of process automation, a large number of industrial wireless sensors are installed in the plant to give insight into process and environmental conditions, asset health and inventory of material. The data are transported to displays for observation and/or to databases for registration and data analysis Examples of sensors are temperature, pressure or flow rate sensors for process monitoring, vibration sensors for health monitoring of e.g. motors, or thermal cameras to detect leakages. Industrial wireless sensors are typically constrained in terms of size, complexity and/or power consumption. The operation for this use case can be in a wide service area, and interaction with the public network (e.g., service continuity, roaming) may be required. Table A.2.3.2-1: Service performance requirements for process and asset monitoring Characteristic parameter Influence quantity Use case Communication service availability: target value [%] Communication service reliability: mean time between failure End-to-end latency Transfer interval (note 1) Bit rate [bits/s] Battery lifetime [year] (note 2) Message Size [byte] Survival time UE speed UE density [UE / m²] Range [m] (note 4) Service area (note 5) 1 99.99 ≥ 1 week < 100 ms 100 ms to 60 s ≤ 1 M ≥ 5 20 (note 3) 3 x transfer interval Stationary Up to 1 < 500 ≤ 10 km x 10 km x 50 m 2 99.99 ≥ 1 week < 100 ms ≤ 1 s ≤ 200 k ≥ 5 25 k 3 x transfer interval Stationary Up to 0.05 < 500 ≤ 10 km x 10 km x 50 m 3 99.99 ≥ 1 week < 100 ms ≤ 1 s ≤ 2 M ≥ 5 250 k 3 x transfer interval Stationary Up to 0.05 < 500 ≤ 10 km x 10 km x 50 m NOTE 1: The transfer interval deviates around its target value by < ± 25 %. NOTE 2: Industrial sensors can use a wide variety of batteries depending on the use case, but in general they are highly constrained in terms of battery size. NOTE 3: The application-level messages in this use case are typically transferred over Ethernet, in which case the minimum Ethernet frame size of 64 bytes applies and dictates the minimum size of the PDU sent over the air interface. NOTE 4: Distance between the gNB and the UE. NOTE 5: Length x width x height. Use case one Sensors generating periodic measurements of a continuous value (e.g. temperature, pressure, flow rate sensors). The traffic is predominantly mobile originated. Use case two Sensors generating waveform measurements (e.g. vibration sensors). Even though the waveform measurement is continuous, it is expected that this type of sensors will buffer and transmit the data periodically (e.g. every second) to save battery by enabling discontinuous transmission. The traffic is predominantly mobile originated. Use case three Cameras (regular or thermal) for asset monitoring (e.g. for leakage detection). Even though the video recording is continuous, it is expected that this type of sensors will buffer and transmit the data periodically (e.g. every second) to save battery by enabling discontinuous transmission. The traffic is predominantly mobile originated. A.2.3.3 Plant asset management To keep a plant running, it is essential that the assets, such as pumps, valves, heaters, instruments, etc., are maintained. Timely recognition of any degradation and continuous self-diagnosis of components are used to support and plan maintenance. Remote software updates enhance and adapt the components to changing conditions and advances in technology. The operation for this use case can be in a wide service area, and interaction with the public network (e.g., service continuity, roaming) may be required. In this use case, the assets themselves are assumed to be connected to the 5G system. The use case where sensors are used to monitor assets is covered in clause A.2.3.2. Table A.2.3.3-1: Service performance requirements for plant asset management Use case # Characteristic parameter Influence quantity Communication service availability: target value [%] Communication service reliability: mean time between failures End-to-end latency: maximum Message size [byte] Transfer interval Survival time UE speed # of UEs Service area (note) 1 99.99 TBD < transfer interval value 20 to 255 several seconds matter of convenience; typically ≥ 3 x transfer interval value typically stationary ≤ 100,000 typically ≤ 10 km x 10 km x 50 m NOTE: Length x width x height. Use case one To keep a plant running, it is essential that the assets, such as pumps, valves, heaters, and instruments are maintained. Timely recognition of any degradation and continuous self-diagnosis of components are used to support and plan maintenance. Remote software updates enhance and adapt the components to changing conditions and advances in technology. A.2.3.4 Inspection in production systems An Edge Computing use case example: Digital twin based production quality maintenance: A digital twin is a virtual representation of a product or production systems. Digital twins are used to simulate, predict and optimize products and production systems. In future, digital twin production control system based on augmented reality based will be used in the factories. In this usecase, a digital twin’s digital and virtual model of a function combined with other physical data to simulate real-time aspects of how a system operates. A digital twin production control system can be automated using machine data and the AI/ML trained data after applying the AI/ML algorithm and further processing. The processed output can be translated to a control command back to the device by a control function running on the edge cloud. In another case, using telemetry data as input, the digital twin model’s output may be fed to an AR server for sending low latency AR streams toward the manual operator in the factory production area. At the same time, it can further be utilized as input by an AI/ML model. A process control function can compare the machine data (example:position, rotation level, speed, sensor data, high-speed photography etc.) and perhaps a high-resolution video from the manufacturing line and if necessary, it can send commands for corrective measure. In this example, the process control functions reside at the edge infrastructure and the inspection related corrective input is sent back to the production system control function. Corrective actions/commands for misalignments from the processed output can be sent in two ways: 1. Manual process with AR server: In this case the service performance requirements should follow the table A.2.4.2-1. 2. Automatic process with AI/ML: This usecase and service performance requirements are described below in table A.2.3.4 The general high-level service requirements of the edge computing usecase for the digital twin based production inspection: • High bandwidth wireless data for both uplink and downlink ◦ exact number depends on video encoding, frame rate and video-resolution requirements • Timing accuracy and low latency ( <=20ms) • High availability of the communication network • Security requirements: Data encryption, connection authentication, user authorization • QoS methods to ensure quality of service performance over different UE to Application connection sessions (video streaming, sensor data, control data) • UE Mobility and session continuity (optional) Table A.2.3.4: Service performance requirements for automated inspection Use case # Characteristic parameter Influence quantity Communication service availability: target value [%] Communication service reliability: mean time between failures End-to-end latency: maximum Message size [byte] Transfer interval: lower bound Transfer interval: target value Transfer interval: upper bound Survival time UE speed # of UEs Service area (note) 1 99.999 ≥ 1 year < target transfer interval value 20 – large packets -20 % of target transfer interval value <=20 ms +20% of target transfer interval value Variable depending upon vertical industry typically stationary < 5 typically typically ≤100 m x 100 m x 50 m NOTE: Length x width x height. Use case one The periodic telemetry data and video images are used from the digital twin in the production system for analysis and then the processed outcome is sent back to the system for any adjustment of the machine components. The following diagram explains the above digital twin usecase steps to manage the production in a factory (both manual and automated operations) • Low-latency AR overlays and incorporation of AI/ML techniques to identify manufacturing issues and improve product quality as well as to enable offline adjustments for optimization, adaptations, and preventive operations on the machines A.2.4 Human machine interfaces A.2.4.1 Mobile control panels Control panels are crucial devices for the interaction between people and production machinery as well as for the interaction with moving devices. These panels are mainly used for configuring, monitoring, debugging, controlling and maintaining machines, robots, cranes or complete production lines. In addition to that, (safety) control panels are typically equipped with an emergency stop button and an enabling device, which an operator can use in case of a safety event in order to avoid damage to humans or machinery. When the emergency stop button is pushed, the controlled equipment immediately comes to a safe stationary position. Likewise, if a machine, robot, etc. is operated in the so-called special ‘enabling device mode’, the operator manually keeps the enabling device switch in a special stationary position. If the operator pushes this switch too much or releases it, the controlled equipment immediately comes to a safe stationary position as well. This way, it can be ensured that the hand(s) of the operator are on the panel (and not under a moulding press, for example), and that the operator does―for instance―not suffer from any electric shock or the like. A common use case for this ‘enabling device mode’ is the installation, testing or maintenance of a machine, during which other safety mechanisms (such as a safety fence) are deactivated. Due to the criticality of these safety functions, safety control panels currently have mostly a wire-bound connection to the equipment they control. In consequence, there tend to be many such panels for the many machines and production units that typically can be found in a factory. With an ultra-reliable low-latency wireless link, it would be possible to connect such mobile control panels with safety functions wirelessly. This would lead to a higher usability and would allow for the flexible and easy re-use of panels for controlling different machines. The cycle times of the control application depends on the process/machinery/equipment whose safety has to be ensured. For a fast-moving robot, for example, end-to-end latencies are lower than for slowly moving linear actuators. Table A.2.4.1-1: Service performance requirements for mobile control panels Use case # Characteristic parameter Influence quantity Communication service availability: target value [%] Communication service reliability: mean time between failures End-to-end latency: maximum Service bitrate: user experienced data rate Message size [byte] Transfer interval: lower bound Transfer interval: target value Transfer interval: upper bound Survival time UE speed # of UEs Service area (note 1) 1 (note 3) 99.999 9 to 99.999 999 ~ 1 month < target transfer interval value – 40 to 250 – < 25 % of target transfer interval value 4 ms to 8 ms + 25 % of target transfer interval value target transfer interval value < 7.2 km/h TBD 50 m x 10 m x 4 m 2 (note 3) 99.999 9 to 99.999 999 ~ 1 month < target transfer interval value > 5 Mbit/s – – < 25 % of target transfer interval value < 30 ms + 25 % of target transfer interval value TBD < 7.2 km/h TBD TBD 3 (note 3) 99.999 9 to 99.999 999 ~ 1 year < target transfer interval – 40 to 250 – < 25 % of target transfer interval value < 12 ms + 25 % of target transfer interval value 12 ms < 7.2 km/h TBD typically 40 m x 60 m; maximum 200 m x 300 m NOTE 1: Length x width (x height). NOTE 2: The transfer interval is not so strictly periodic in these use cases. The transfer interval deviates around its target value within bounds. The mean of the transfer interval is close to the target value. NOTE 3: Communication may include two wireless links (UE to UE) Use case one Periodic, bi-directional communication for remote control. Examples for controlled units: assembly robots; milling machines. Use case two Aperiodic data transmission in parallel to remote control (use case one). Use case three Periodic, bi-directional communication for remote control. Examples for controlled units: mobile cranes, mobile pumps, fixed portal cranes. A.2.4.1A Mobile operation panels Operation and monitoring of machines, mobile robots, or production units via a mobile operation panel provides higher flexibility and comfort for human operators. A single mobile operation panel can be used to manage more than one production system due to its mobility in the factory. The mobile operation panel provides relevant information for configuration, control of industrial machines as well as monitoring of relevant data generated during the construction of a product. The monitoring data is generally considered to be less time-critical subsequently requiring non-real-time communication. On the other hand, the mobile operation panel supports safety-critical functions such as emergency stops or enabling or changing the position of robots and other machines. These functions are generally considered to have strict ultra-low latencies and reliable transmission requirements that must follow strict safety standards making them time-critical (real-time communication). Table A.2.4.1A-1: Service performance requirements for mobile operation panels Use case # Characteristic parameter Influence quantity Communication service availability: target value [%] Communication service reliability: mean time between failures End-to-end latency: maximum Service bitrate: user experienced data rate Direction (note 2) Message size [byte] Transfer interval: target value Survival time UE speed # of UEs Service area (note 1) 1 99.999 999 1 day < 8 ms 250 kbit/s Uplink Downlink 40 to 250 8 ms 16 ms quasi-static; up to 10 km/h 2 or more 30 m x 30 m 2 99.999 99 1 day < 10 ms < 1 Mbit/s Uplink < 1,024 10 ms ~10 ms quasi-static; up to 10 km/h 2 or more 30 m x 30 m 3 99.999 999 1 day 10 ms to 100 ms 10 kbit/s Uplink Downlink 10-100 10-100 ms transfer interval stationary 2 or more 100 m² to 2,000 m² 4 99.999 999 1 day < 1 ms 12 Mbit/s to 16 Mbit/s Downlink 10-100 1 ms ~1 ms stationary 2 or more 100 m² 5 99.999 999 1 day < 2 ms 16 kbit/s (UL) 2 Mbit/s (DL) Uplink Downlink 50 2 ms ~2 ms stationary 2 or more 100 m² 6 99.999 9 to 99.999 99 1 day up to [x] 12 Mbit/s Uplink Downlink 250 to 1,500 quasi-static; up to 10 km/h 2 or more 30 m x 30 m NOTE 1: Length x width. NOTE 2: The mobile operation panel is connected wirelessly to the 5G system. If the mobile robot/production line is also connected wirelessly to the 5G system, the communication includes two wireless links. Use case one Emergency Stop with periodic-deterministic communication for connectivity availability and aperiodic-deterministic communication for emergency stop events. Use case two Safety data stream with periodic deterministic communication. Use case three Visualization of Control with periodic deterministic communication. Use case four Motion Control with periodic deterministic communication. Use case five Haptic feedback data stream with periodic deterministic communication. Use case six Manufacturing data stream with mixed traffic. A.2.4.2 Augmented reality It is envisioned that in future smart factories and production facilities, people will continue to play an important and substantial role. However, due to the envisaged high flexibility and versatility of the Factories of the Future, shop floor workers should be optimally supported in getting quickly prepared for new tasks and activities and in ensuring smooth operations in an efficient and ergonomic manner. To this end, augmented reality may play a crucial role, for example for the following applications: - monitoring of processes and production flows; - step-by-step instructions for specific tasks, for example in manual assembly workplaces; - ad hoc support from a remote expert, for example for maintenance or service tasks. In this respect, especially head-mounted augmented-reality devices with see-through display are very attractive since they allow for a maximum degree of ergonomics, flexibility and mobility and leave the hands of workers free for other tasks. However, if such augmented-reality devices are worn for a longer period (e.g., one work shift), these devices have to be lightweight and highly energy-efficient while at the same time they should not become very warm. A very promising approach is to offload complex (e.g., video) processing tasks to the network (e.g., an edge cloud) and to reduce the augmented-reality head-mounted device’s functionality. This has the additional benefit that the augmented-reality application may have easy access to different context information (e.g., information about the environment, production machinery, the current link state, etc.) if executed in the network. Table A.2.4.2-1: Service performance requirements for augmented reality in human-machine interfaces Use case # Characteristic parameter Influence quantity Communication service availability: target value [%] Communication service reliability: mean time between failures End-to-end latency: maximum UE speed Service area (note) 1 > 99.9 ~ 1 month < 10 ms < 8 km/h 20 m x 20 m x 4 m NOTE: Length x width x height. Use case one Bi-directional message transmission between an augmented-reality device and an image processing server. A.2.5 Monitoring and maintenance A.2.5.1 Remote access and maintenance In factories of the future, there are needs to perform remote access and maintenance to devices and entities, for instance, by remote control centres. The devices and entities might be installed at geographically distributed locations. These devices typically have firmware/software which needs to be updated occasionally. Maintenance information also needs to be collected and distributed from/to these devices periodically. The devices can be both stationary and mobile. Device maintenance may happen in parallel to the actual production process and other communication services performed at the device side without any negative impact on these production communication services. Table A.2.5.1-1: Service performance requirements for remote access and maintenance Use case # Characteristic parameter Influence quantity Communication service availability: target value [%] Communication service reliability: mean time between failures End-to-end latency: maximum Service bitrate: user experienced data rate Message size [byte] Transfer interval: lower bound Transfer interval: upper bound Survival time UE speed # of UEs Service area (note) 1 – ~ 1 month – ≥ 1 Mbit/s – – – – ≤ 72 km/h ≤ 100 50 m x 10 m x 10 m NOTE: Length x width x height. Use case one Transmission of non-deterministic messages in parallel to other interactions. Example applications: software/firmware updates and exchange of maintenance information. A.3 (void) A.4 Electric-power distribution and Smart Grid A.4.1 Overview The energy sector is currently subject to a fundamental change, which is caused by the evolution towards renewable energy, i.e. an increasing number of power plants based on solar and wind power. These changes lead to bi-directional electricity flows and increased dynamics of the power system. New sensors and actuators are being deployed in the power system to efficiently monitor and control the volatile conditions of the grid, requiring real-time information exchange [11][12]. The emerging electric-power distribution grid is also referred to as Smart Grid. The smartness enhances insight into both the grid as a power network and the grid as a system of systems. Enhanced insight improves controllability and predictability, both of which drive improved operation and economic performance and both of which are prerequisites for the sustainable and scalable integration of renewables into the grid and the potential transition to new grid architectures. Smart Grid benefits spread across a broad spectrum but generally include improvements in: power reliability and quality; grid resiliency; power usage optimisation; operational insights; renewable integration; insight into energy usage; and safety and security. Overviews of (future) electric-power distribution can be found elsewhere in the literature [13][14]. A.4.2 Primary frequency control Primary frequency control is among the most challenging and demanding control applications in the utility sector. A primary frequency control system is responsible for controlling the energy supply injected and withheld to ensure that the frequency is not deviating more than 0.02 % from the nominal value (e.g., 50 Hz in Europe). Frequency control is based on having sensors for measuring the features in all parts of the network at all points where energy generation or storage units are connected to the grid. At these points, electronic power converters, also known as inverters, are equipped with communication units to send measurement values to other points in the grid such as a frequency control unit, or receive control commands to inject more, or less, energy into the local network. With the widespread deployment of local generation units, i.e. solar power units, or wind turbines, hundreds of thousands of such units, and their inverters, may have to be connected in a larger power distribution network. Primary frequency control is carried out in one of three ways: 1) Centralised control, all data analysis and corrective actions are determined by a central frequency control unit. 2) Decentralised control, the automatic routine frequency control is performed by the individual local inverter based on local frequency values. Statistics and other information is communicated to the central frequency control unit. 3) Distributed control, the automatic routine frequency control is performed by the individual local inverter based on local and neighbouring frequency values. Statistics and other information are communicated to the central frequency control unit. Table A.4.2-1: Service performance requirements for primary frequency control Use case # Characteristic parameter Influence quantity Communication service availability: target value [%] Communication service reliability: mean time between failures End-to-end latency: maximum Message size [byte] Transfer interval: target value Survival time # of UEs Service area 1 99.999 TBD ~ 50 ms ~ 100 ~ 50 ms TBD ≤ 100,000 several km2 up to 100,000 km2 Use case one Periodic communication service supporting message exchange for primary frequency control. A.4.3 Distributed voltage control In the evolution towards 100 % renewable electric power production, the objective of voltage control is to balance the voltage in future low voltage distribution grids connecting local loads and prosumers, as well as energy storage facilities. The aim is to stabilise the voltage as locally as possible, so that decisions and control commands can be issued as quickly as possible. Distributed voltage control is a challenging and demanding control application. Consumer devices rely on having stable voltage levels to operate successfully. When future energy networks rely on thousands of local energy generation units relying mostly on solar and wind power, then it is crucial to stabilise the voltage levels in all segments of the distribution grid. Inverters, or electronic power converters, measure the voltage and power and change the amount of power injected into the grid, and they connect and disconnect end points from the distribution network. Distributed control means that the automated voltage control shall be performed by the local voltage control units based on local and neighbouring voltage and impedance values. Statistics and other information shall be communicated to the central distribution management system, though. Table A.4.3-1: Service performance requirements for distributed voltage control Use case # Characteristic parameter Influence quantity Communication service availability: target value [%] Communication service reliability: mean time between failures End-to-end latency: maximum Message size [byte] Transfer interval: target value Survival time # of UEs Service area 1 99.999 TBD ~ 100 ms ~ 100 ~ 200 ms TBD ≤ 100,000 several km2 up to 100,000 km2 Use case one Periodic communication service supporting message exchange for distributed voltage control. A.4.4 Distributed energy automation A.4.4.1 Distributed automated switching for isolation and service restoration A power distribution grid fault is a stressful situation. There are self-healing solutions for automated switching, fault isolation, and service restoration. Furthermore, these solutions are ideally suited to handle outages that affect critical power consumers, such as industrial plants or data centres. Supply interruptions must be fixed within less than a second for critical power consumers. Automated solutions are able to restore power supply within a few hundred milliseconds. Figure A.4.4.1-1: Depiction of a distribution ring and a failure (flash of lighting) The FLISR (Fault Location, Isolation & Service Restoration) solution consists of switch controller devices which are especially designed for feeder automation applications that support the self-healing of power distribution grids with overhead lines. They serve as control units for reclosers and disconnectors in overhead line distribution grids. The system is designed for using fully distributed, independent automated devices. The logic resides in each individual feeder automation controller located at the poles in the feeder level. Each feeder section has a controller device. Using peer-to-peer communication among the controller devices, the system operates autonomously without the need of a regional controller or control centre. However, all self-healing steps carried out will be reported immediately to the control centre to keep the grid status up to date. The controllers conduct self-healing of the distribution line in typically 500 ms by isolating the faults. Peer-to-peer communication via IEC 61850 GOOSE (Generic Object Oriented Substation Event) messages provides data as fast as possible (Layer 2 multicast message). They are sent periodically (in steady state, with changing interval time in fault case) by each controller to several or all other controllers of the same feeder and are not acknowledged. The service bit rate per controller is low in steady state, but GOOSE bursts with high service bit rate do occur, especially during fault situations. GOOSE messages are sent by several or all controller units of the feeder nearly at the same point in time during the fault location, isolation and service restoration procedure with a low end-to-end latency. The associated (a)periodic communication KPIs are provided in Table A4.4.1-1. Table A.4.4-1.1: Service performance requirements for distributed automated switching for isolation and service restoration Use case # Characteristic parameter Influence quantity Communication service availability: target value [%] Communication service reliability: mean time between failures End-to-end latency: maximum Service bitrate: user experienced data rate Message size [byte] Transfer interval: target value Survival time UE speed # of UEs Service area (note 1) 1 (note 2) 99.999 9 – < 5 ms 1 kbit/s (steady state) 1.5 Mbit/s (fault case) < 1,500 < 60 s (steady state) ≥ 1 ms (fault case) transfer interval (one frame loss) stationary 20 30 km x 20 km 2 > 99.999 % - 20 ms (note 2) - < 100 - - stationary ≤ 100/km2 several km2 NOTE 1: Length x width NOTE 2: UE to UE communication (two wireless links) Use case one GOOSE (a)periodic deterministic communication service supporting bursty message exchange for fault location, isolation, and service restoration. Use case two Typically event-driven, aperiodic deterministic communication service supporting fault detection and isolation. A.4.4.2 Distributed automation without GOOSE If the control of electrical power distribution components is performed from a central system entity, the controlled entities can be operated in a way that a controlled service restoration is possible without the use of GOOSE. Though this is not as effective as the communication has less strict requirements, this form of distribution automation is nevertheless effective, and it is also compliant with IEC standards and widely deployed in energy systems. The associated KPI is provided in Table A.4.4.2-1. Table A.4.4.2-1: KPI for distributed automation without use of GOOSE Use case # Characteristic parameter Influence quantity Communication service availability: target value [%] Communication service reliability: mean time between failures End-to-end latency: maximum Service bitrate: user experienced data rate Message size [byte] Transfer interval: target value Survival time UE speed # of UEs Service area 1 99.999 – 100 ms to 2 s 9.6 to100 – – – – Concentrated rural: 70.8; Dispersed rural and semi-urban: 7.6; rural support: 0.048; urban: 11.0 several km2 Use case one Distributed automation without use of GOOSE using a centralized architecture. A.4.4.3 Intelligent distributed feeder automation Intelligent distributed feeder automation system which supported by 5G connections is designed to realize intelligent judgment, analysis, fault location, fault isolation and non-fault area power supply restoration operations. As illustrated in the Figure A.4.4.3-1, the distributed feeder automation system is mainly composed of a distribution master station, a distribution terminal, switch stations, and the communication system (UEs in the substations, 5G network, plus the data network). The distribution master station is mainly used for information gathering and human-computer interaction, and the distributed terminals are used for the collection of feeder status information and judgment, fault location, isolation, as well as power supply restoration based on this information. Distributed terminal actions are reported to the distribution master station. The 5G communication system enables communication among the distribution terminals. The distribution master station is usually connected to the 5G system via a data network, which is out of 3GPP scope. Figure A.4.4.3-1: Example of intelligent distributed feeder automation The distribution master station manages multiple distributed terminals. Each distributed terminal is served by a 5G UE to exchange the collected data with other distributed terminals. From an application perspective, the communication between distributed terminals is peer-to-peer. The 5G communication service availability needs to be very high. Therefore, at least two communication links are usually deployed for hot standby or for transmitting data synchronously between two distributed terminals. The associated KPI is provided in Table A.4.4.3-1. Table A.4.4.3-1: KPI for intelligent distributed feeder automation Use case # Characteristic parameter Influence quantity Communication service availability: target value [%] Communication service reliability: mean time between failures End-to-end latency: maximum Service bitrate: user experienced data rate Message size [byte] Transfer interval: target value Survival time UE speed # of UEs Service area 1 99.999 – Normal: 1 s; Fault: 2 ms (note 2) 2 M to10 M (note 1) – Normal: 1 s; Fault: 2 ms (note 2) – – 54/km2 (note 3) 78/km2 (note 4) several km2 NOTE 1: The KPI values are sourced from [29]. NOTE 2: It is the one-way delay from a distributed terminal to 5G network. NOTE 3: When the distributed terminals are deployed along overhead line, about 54 terminals will be distributed along overhead lines in one square kilometre. NOTE 4: When the distributed terminals are deployed in power distribution cabinets, there are about 78 terminals in one square kilometre. Use cases #1: Intelligent distributed feeder automation A.4.4.4 High-speed current differential protection High-speed current differential protection, which is required for sub-millisecond fault detection, is another typical use case of power distribution automation. The approach utilises differential current measurements to significantly reduce fault detection time. The protection relays exchange the current samples via the 5G system. Each relay then compares the sent and received samples to determine if a fault has occurred in a protected area. This is done in order to identify and isolate a fault in the grid. The sampling rate varies and is dependent on the algorithms designed by the manufacturers. A protection relay collects the current samples (with the typical message size of up to 245 bytes) at a frequency of 600 Hz, 1200 Hz, 1600 Hz, or 3000 Hz. The exchange of measurement samples is done in a strictly cyclic and deterministic manner. With the sampling rate of 600 Hz, the transfer interval is 1.7 ms and the required bandwidth 1.2 Mbit/s; for 1200 Hz, the transfer interval is 0.83 ms and the required bandwidth 2.4 Mbit/s. The maximum allowed end-to-end delay between two protection relays is between 5 ms and 10 ms, depending on the voltage (see IEC 61850-90-1 for more details [28]). For some legacy systems, the latency usually is set to 15 ms. The associated KPIs are provided in Table A.4.4.4-1. Table A.4.4.4-1: KPIs for high speed current differential protection Use case # Communication service availability End-to-end latency: maximum (note) Service bitrate: user experienced data rate Message size [byte] Transfer interval: target value Survival time UE speed UE density [#/km2)] Service area 1 > 99.999 % 15 ms 2.5 Mbit/s < 245 ≤ 1 ms transfer interval (one frame loss) stationary ≤ 100/km2 several km2 2 > 99.999 % 15 ms 1.2 Mbit/s < 245 ≤ 2 ms transfer interval (one frame loss) stationary ≤ 100/km2 several km2 3 > 99.999 % 10 ms 2.5 Mbit/s < 245 ≤ 1 ms transfer interval (one frame loss) stationary ≤ 100/km2 several km2 4 > 99.999 % 10 ms 1.2 Mbit/s < 245 ≤ 2 ms transfer interval (one frame loss) stationary ≤ 100/km2 several km2 5 > 99.999 % 5 ms 2.5 Mbit/s < 245 ≤ 1 ms transfer interval (one frame loss) stationary ≤ 100/km2 several km2 6 > 99.999 % 5 ms 1.2 Mbit/s < 245 ≤ 2 ms transfer interval (one frame loss) stationary ≤ 100/km2 several km2 NOTE : UE-to-UE communication. Use case #1: High-speed current differential protection with a sampling rate of 1200 Hz for legacy systems. Use case #2: High-speed current differential protection with a sampling rate of 600 Hz for legacy systems. Use case #3: High-speed current differential protection with a sampling rate of 1200 Hz under voltage condition 1 (see IEC 61850-90-1[28] for more details). Use case #4: High-speed current differential protection with a sampling rate of 600 Hz under voltage condition 1 (see IEC 61850-90-1[28] for more details). Use case #5: High-speed current differential protection with a sampling rate of 1200 Hz under voltage condition 2 (see IEC 61850-90-1[28] for more details). Use case #6: High-speed current differential protection with a sampling rate of 600 Hz under voltage condition 2 (see IEC 61850-90-1[28] for more details). A.4.5 Smart grid millisecond-level precise load control Precise Load Control is the basic application for smart grid. When serious HVDC (high-voltage direct current) transmission fault happens, Millisecond-Level Precise Load Control is used to quickly remove interruptible less-important load, such as electric vehicle charging piles and non-continuous production power supplies in factories. Table A.4.5-1: Service performance requirements for smart grid millisecond-level precise load control Use case # Characteristic parameter Influence quantity Communication service availability: target value [%] Communication service reliability: mean time between failures End-to-end latency: maximum Service bitrate: user experienced data rate Message size [byte] Transfer interval: target value Survival time UE speed # of UEs Service area 1 99.999 9 – < 50 ms 0.59 kbit/s 28 kbit/s < 100 n/a (note) – stationary 10 km-² to 100 km-² TBD NOTE: event-triggered Use case one A non-periodic deterministic communication service between control primary station and load control terminals for removing interruptible less-important load quickly. A.4.6 Distributed energy storage Distributed power generation includes various power sources such as solar, wind, fuel cells, and gas. Distributed power generation typically comes with a low power density and entails thus typicall a decentralised deployment. Decentralisation causes technical problems and challenges Smart Grid operators. When distributed power generation is connected to the elctrical grid, the energy flow becomes more complicated as the user often is both an electricity consumer and producer (a so-called “prosumer”). Therefore, the current in the electricity grid can change direction at different locations of the grid and at different times of the day. The information exchange in a distributed energy grid does not only include power-generation-related data, but also control commands for the distributed energy storage equipment. An example for such a command is “change the load characteristics to realise a flexible electricity grid” etc. Figure A.4.6-1 Example of a distributed-energy storage grid Figure A.4.6-1 shows an example of distributed-energy storage grid. The distributed-energy storage grid needs to exchange information among the distributed-energy storage management platform (DESMP) and distributed-energy devices (DEDs). The DED is a plug-and-play device and periodically collects its energy information, such as battery energy, charge and discharge status, energy alarm information, etc. The DED then transfers this information via 5G UE to the DESMP. The DESMP regularly manages the DEDs, e.g., the DESMP monitors the DEDs working status, controls the DEDs working modes, or configures the DEDs energy parameters etc. The associated KPIs are provided in Table A.4.6-1. Table A.4.6-1: Communication service performance requirements ‒ data for distributed energy storage Use case# Characteristic parameter Influence quantity Communica­tion service availability: target value Communication service reliability: mean time between failures End-to-end latency: maximum Service bit rate: user experienced data rate Message size [byte] Transfer interval: target value Survival time UE speed # of UEs Service area 1 > 99.9 % DL: < 10 ms UL: < 10 ms UL: > 16 Mbit/s (urban); 640 Mbit/s (rural) DL: > 100 kbit/s (note 1) UL: 800 kbyte UL: 10 ms – stationary > 10/km2 (urban); > 100/km2 (rural) (note 2) several km2 2 > 99.9 % DL: < 10 ms UL: < 1 s UL: > 128 kbit/s (urban); 10.4 Mbit/s (rural); DL: > 100 kbit/s (note 1) UL: 1.3 Mbyte DL: > 100 kbyte UL: 1000 ms – stationary > 10/km2 (urban); > 100/km2 (rural) (note 2) several km2 3 > 99.9 % DL: < 10 ms UL:< 1 s (rural) DL: > 100 kbit/s UL: > 5 Gbit/s (note 3) stationary > 100/km2 several km2 NOTE 1: Service bit rate for one energy storage station. NOTE 2: Activity storage nodes/km2. This value is used for deducing the data volume in an area that features multiple energy storage stations. The data volume can be calculated with the following formula (current service bit rate per storage station) x (activity storage nodes/km2) + (video service bit rate per storage station) x (activity storage nodes/km2). NOTE 3: The downlink user experienced data rate is calculated as follows: 12.5 Mbytes/s x 50(containers) x 8 = 5 Gbit/s Use case#1: Distributed energy storage ‒ periodic communication for monitoring Use case#2: Distributed energy storage ‒ periodic communication for data collection Use case#3: Distributed energy storage ‒ aperiodic video communication A.4.7 Advanced metering Instead of recording and sending metering data from a wired electricity meter unit, electricity metering collecting can be executed by a UE-integrated smart meter unit. Smart meter units can send real-time metering data to a server in the utility through mobile networks. In this way, the power enterprise―based on the analysis of the user’s power consumption behavior―gives the user power consumption suggestions, which fosters the user’s power consumption and energy saving habits. The electric smart meters monitor relevant user energy status and deliver the status data to a measurement data management system (MDMS). The MDMS sends control commands according to its policy and the status of the data collected from the smart meters. The MDMS commands include tripping, closing permission, alarm, alarm release, power protection, and power protection release. Accurate-fee control is one of the basic services of advanced metering. When the electric power user doesn’t pay her electric fee on time, the MDMS can cut off the power supply. And when there is a need for temporary power supply for this user, the MDMS can recover the power supply. This operation requires real-time interaction between the electric smart meter and the MDMS. Due to massive number of electricity meters, it is estimated that in the near future, the amount of this kind of interaction will increase 5 to 10 times. The associated KPIs are provided in Table A.4.7-1. Table A.4.7-1: Communication KPI for advanced metering Use case# Characteristic parameter Influence quantity Communica­tion service availability: target value Communication service reliability: mean time between failures End-to-end latency: maximum Service bit rate: user experienced data rate Message size [byte] Transfer interval: target value Survival time UE speed # of UEs Service area 1 > 99.99 Accuracy fee control: < 100 (note 1); General information data collection: < 3000 UL: < 2 M DL: < 1 M – – – stationary < 10 000/km2 (note2) – NOTE 1: One-way delay from 5G IoT device to backend system. The distance between the two is below 40 km (city range). NOTE 2: It is the typical connection density in today city environment. With the evolution from centralised meters to socket meters in the home, the connection density is expected to increase 5 to 10 times. Use case#1: Advanced metering A.4.8 Smart distribution transformer terminal A smart distribution transformer terminal is usually deployed in a distribution transformer area. The terminal can support multiple energy applications simultaneously. Multiple kinds of energy data are collected by the terminal and then delivered to a energy application platform. Figure A.4.8.1-1 illustrates a workflow example for a smart distribution transformer terminal. Figure A.4.8-1: Example of a smart distribution transformer terminal workflow In general, the connections between the smart distribution transformer terminal and the energy application platform are provided by the 5G system. The connections between energy end equipment and smart distribution transformer terminal may be provided by 5G system. In this case, about 300 to 500 energy end devices are connected to one smart distribution transformer terminal. The average service bit rate between the smart distribution transformer terminal and an energy end device is more than 2 Mbit/s in uplink for each application. The related communication distance is between 100 m and 500 m. The associated KPI is provided in Table A.4.8-1. Table A.4.8-1: Key Performance for Smart Distribution Transformer Terminal Use case# Characteristic parameter Influence quantity Communica­tion service availability: target value Communication service reliability: mean time between failures End-to-end latency: maximum Service bit rate: user experienced data rate Message size [byte] Transfer interval: target value Survival time UE speed # of UEs Service area 1 >99.99% – 10 ms, 100 ms, 3 s (note 2) > 2 Mbit/s (note 1) – – – – 500 in the service area (note 3) Communication distance is from 100 m to 500 m, (outdoor, indoor, and deep indoor) NOTE 1: It is the smart metering application data rate between the Smart Distribution Transformer Terminal and energy end equipment. Once there are multiple smart grid applications, it is required more data rate. NOTE 2: It depends on different applications supported by the Smart Distribution Transformer Terminal. The less the latency is, the more applications can be supported. NOTE 3: The distribution area can be calculated as 3.14 x range2 and in general is between 0.031 km2 and 0.79 km2. Use case#1: smart distribution transformer terminal A.4.9 Distributed energy resources and micro-grids Distributed energy resources (DER) become increasingly important. The potentially large number of DERs will have an impact on security, stability, and operation efficiency of the energy grid. The integration of DERs into the energy grid poses many challenges for the involved communication system. To incorporate more renewable and alternative energy sources, the communication infrastructure must have the ability to easily handle an increasing amount of data traffic or service requests and must provide a real-time monitoring and control operation for these distributed energy resources. A reliable communication between the DERs is crucial. When it comes to communications architecture, IEC 61850 is a widely used standard for automation and equipment of power utilities and DER, specifically for defining protocols for IEDs (Intelligent electronic devices) at electrical substations The IEC 61850 standard specifies the timing constraints for messages typically used in substations. GOOSE (Generic Object Oriented Substation Events) and SV (Sampled Values) messages are assumed as time critical messages. They have the tightest deadlines (maximum allowed transfer time) among all IEC 61850 messages, corresponding to 3 ms. While GOOSE is typically used for transfering information related to monitoring and control functions (circuit breaker status etc.), SV is used for transfering measurement samples of current and voltage signals. The SV protocol works on a periodic information transmission model, sending messages at a fixed rate. For protection purposes, the default rate is 4000 or 4800 messages per second for 50 and 60 Hz power systems, respectively. On the other hand, the GOOSE protocol operates in a sporadic information transmission model, where a continuous flow of data is maintained to increase communication reliability. The typical sizes of GOOSE and SV messages are160 and 140 bytes, respectively. GOOSE messages are transmitted in two different modes: (1) safe operation: 1 message per second (service bit rate = 1.28 kbit/s); (2) emergency operation: 32 messages per second (service bit rate = 41.0 kbit/s). SV messages are transmitted at much higher rate, namely 4800 messages per second (service bit rate = 5.4 Mbit/s). The associated KPIs are provided in Table A.4.9-1. Table A.4.9-1: Key Performance for Distributed energy resources (DER): using SV (Sampled Values) message Use case# Characteristic parameter Influence quantity Communica­tion service availability: target value Communication service reliability: mean time between failures End-to-end latency: maximum Service bit rate: user experienced data rate Message size [byte] Transfer interval: target value Survival time UE speed 1 99,9999 % – < 3 ms 4.5 Mbit/s 140 ≤ 1 ms transfer interval stationary 2 99,9999 % – < 3 ms 5.4 Mbit/s 140 ≤ 1 ms transfer interval stationary 3 > 99.9999 % – < 3 ms – 160 – – stationary NOTE: UE to UE communication is assumed. Use case#1: Distributed energy resources and micro-grids: using SV(sample value) message with 50Hz Use case#2: Distributed energy resources and micro-grids: using SV(sample value) message with 60Hz Use case#3: Distributed energy resources and micro-grids: using GOOSE message A.4.10 Ensuring uninterrupted communication service availability during emergencies During emergencies, public mobile land networks (PLMNs) may restrict network access, which may lead to a prohibitevly low communication service availability for machine-type communication (MTC) for Smart Grid applications. An example is communication for microgrids. Microgrids are separate parts of a power grid that can be controlled and operated individually in a so-called island mode, or together with other parts of the power grid. The idea is to prioritise Smart Grid-related communication in order to ensure reliable and available communication for selected devices during emergency conditions. Existing features of a mobile network can be used to differentiate MTC of devices in a microgrid from other kind of MTC traffic or human-to-human communication. These features can help these microgrid devices to have communication service during emergencies. The communication among the microgrid devices enables co-ordination of DERs, which help the DERs can autarkically implement recovery of an islanded microgrid. The associated KPI is provided in Table A.4.10-1. Table A.4.10-1: Key Performance for uninterrupted MTC service availability Characteristic parameter (KPI) Influence quantity Communication service availability: target value Communication service reliability: mean time between failures Max Allowed End-to-end latency (note 1; (note 2) Service bit rate: user-experienced data rate (note 2) Message size [byte] Survival time UE speed # of UEs Service Area 99.999 9 % – 100 ms < 1 kbit/s per DER – – Stationary – – NOTE 1: Unless otherwise specified, all communication includes 1 wireless link (UE to network node or network node to UE) rather than two wireless links (UE to UE). NOTE 2: It applies to both UL and DL unless stated otherwise. A.5 Central power generation A.5.1 Overview This domain comprises all aspects of centralised power generation, i.e. the centralised conversion of chemical energy and other forms of energy into electrical energy. Typical electric-power outputs are 100 MW and more. Examples for pertinent systems are large gas turbines, steam turbines, combined-cycle power plants, and wind farms. The planning and installation of respective equipment and plants as well as the operation, monitoring and maintenance of these plants is encompassed by this vertical domain. A.5.2 Wind power plant network Table A.5.2-1: Service performance requirements for wind power plant network Use case # Characteristic parameter Influence quantity Communication service availability: target value [%] Communication service reliability: mean time between failures End-to-end latency: maximum Packet error ratio UE speed Service area 1 99.999 999 9 ~ 10 years 16 ms < 10-9 stationary several km2 Use case one Communication in support of closed-loop cyber-physical control in a wind farm. The wind farm can be deployed offshore. NOTE: This type of communication service can be provided via a wired connection. A.6 Connected hospitals or medical facilities A.6.1 Overview The traditional value chain for the medical device industry, which historically has been driven by innovation and research and development, is currently witnessing a shift in the landscape. As governments and health insurers worldwide implement measures to control costs, public hospitals are operating on tighter budgets, while private facilities are receiving lower reimbursements. In the developed world, decisions that used to be the sole preserve of doctors are now also made by regulators, hospital administrators, and other non-clinicians. This broader set of influencers comes with different objectives, e.g. the prioritization of cost effectiveness or even just costs. This shift in focus from volume-based healthcare to value-based healthcare has led medical devices companies to move to business models based on providing clinical value with cost efficiency. Technological progress and better infrastructures, in particular high-quality wireless networks, have fed this business model transformation, allowing coordinated therapies, services, and health analytics and enabling efficient outcome measurement solutions. On this matter, 5G enables shifting care location from hospitals to homes and others lower cost facilities which mechanically translates into more savings. Additionally, another example showing that 5G can enable cost savings required by the medical industry can be found inside hospitals where wireless transmission of low latency data streams improves operating room planning, enable streamlining equipment usage and simplifies operating theater implementation. A.6.2 Robotic Aided Surgery Robotic aided surgery is particularly suitable to invasive surgical procedures that require delicate tissue manipulation and access to areas with difficult exposure. It is achieved through complex systems that translate the surgeon’s hand movements into smaller, precise movements of tiny instruments that can generally bend and rotate far more than a human hand is capable of doing inside the patient’s body. In addition, those systems are usually able to filter out hand tremor and therefore allow more consistent outcomes for existing procedures, and more importantly the development of new procedures currently made impractical by the accuracy limits of unaided manipulation. A typical robotic setup for telesurgery can be depicted as follows. Figure A.6.2-1: Typical Robotic Surgery System Setup The robot and the surgeon’s console can be co-located in the same operating room in which case they communicate through a NPN, or, in another deployment option, when specialists and patients are far from each other (hundreds of kilometres) they can exchange data through communication services delivered by PLMNs. The depicted medical application can be instantiated at either side or in the Cloud. Its role consists in: - Generating appropriate haptic feedback based on instrument location, velocity, effort measurements data and images issued by surgical instruments and 3D pre-operative patient body model. This allows to provide tactile guidance by constraining where the instruments (scalpel, etc.) can go. - Filtering motion control commands for better closed loop stability Typical surgery robotic systems can have around 40 actuators and the same number of sensors which allows to compute the data rate requires in each direction in order to execute a given movement. Human sensitivity of touch is very high, tactile sensing has about 400 Hz bandwidth, where bandwidth refers to the frequency of which the stimuli can be sensed. This is why, in general, haptic feedback systems operate at frequencies around 1,000 Hz. This rate naturally applies to the update of all information used in the generation of the haptic feedback, e.g. instruments velocity, position … Therefore, the robot control process involves: - The surgeon console periodically sending a set of points to actuators - Actuators executing a given process - Sensors sampling velocity, forces, positions, … at the very same time and returning that information to the surgeon console at the rate of 1 kHz As opposed to machine to machine communication, robotic aided surgery implies there is a human being in the middle of the control loop, which means that the console generates new commands based on the system state collected in the previous 1 kHz cycle and also on surgeon’s hand movement. Each equipment involved in a robotic telesurgery setup (endoscopes, image processing system, displays, motion controller and haptic feedback systems) is synchronized thanks to a common clock either external or provided by the 5G system. The synchronization is often achieved through dedicated protocols such as e.g. PTP version 2 and allows to e.g. guarantee the consistency of the haptic feedback and displayed images at the master console, or enable the recording and offline replay of the whole procedure. Table A.6.2-1: Service performance requirements for motion control and haptic feedback Use case # Characteristic parameter Influence quantity Communication service availability: target value [%] Communication service reliability: Mean Time Between Failure End-to-end latency: maximum Bit rate Direction Message Size [byte] Transfer Interval Survival time UE speed # of active UEs (note1) Service Area 1 > 99.999 999 > 10 years < 2 ms 2 Mbit/s to 16 Mbit/s network to UE; UE to network 250 to 2,000 1 ms transfer interval stationary 1 room 2 > 99.999 9 > 1 year < 20 ms 2 Mbit/s to 16 Mbit/s network to UE; UE to network 250 to 2,000 1 ms transfer interval stationary < 2 per 1,000 km2 national Note 1: The upper limit of UEs’ density is provided for large service areas to address non-uniform distributions of UEs, while an absolute number of UEs is provided for small service areas. Use case one Periodic communication for the support of precise cooperative robotic motion control and haptic feedback in case of robotic aided surgery where the surgeon console and the robot are collocated in the same operating room Use case two Periodic communication for the support of cooperative robotic motion control and haptic feedback in case of telesurgery. In this case, the surgeon console and the robot are not collocated and communicate with each other through a connection established over a PLMN possibly spanning an entire country. Relaxed requirements imply that much less complex surgical procedures are achievable in use case 2 than in use case 1. It shall be noted that this use case also involves more experienced and trained surgeons, who can cope with longer latencies in the communication system. A.6.3 Robotic Aided Diagnosis Robotic aided diagnosis involves a remote expert in a large central hospital who controls a diagnosis robotic system deployed in a local medical facility. Such robotic systems can be e.g.: - Haptic feedback tool used for palpating and deployed in e.g. a Mobile Specialist Practise facility - Ultrasound probe deployed in an ambulance or a medical facility A typical robotic setup for tele diagnosis can be depicted as follows: Figure A.6.3-1: Typical Robotic Surgery System Setup Specialists and patients are far from each other (typically dozens of kilometres) and can exchange data through communication services delivered by PLMNs. The depicted medical application can be instantiated at either side or in the Cloud. Its role consists in: • Generating appropriate haptic feedback based on instrument location, velocity, effort measurements data and images issued by instruments. • Filtering motion control commands for better closed loop stability Table A.6.3-1: Service performance requirements for motion control and haptic feedback Use case # Characteristic parameter Influence quantity Communication service availability: target value [%] Communication service reliability: Mean Time Between Failure End-to-end latency: maximum Bit rate Direction Message Size [byte] Transfer Interval Survival time UE speed # of active UEs Service Area 1 > 99.999 >> 1 month (< 1 year) < 20 ms 2 Mbit/s to 16 Mbit/s network to UE; UE to network ~80 < 20 ms per 100 km2 transfer interval stationary < 20 per 100 km2 regional Use case one Periodic communication for the support of precise cooperative robotic motion control and haptic feedback in case of robotic aided diagnosis where the expert and the patient are not collocated and communicate with each other through a connection established over a PLMN. A.7 Positioning A.7.1 Overview of positioning in industrial use cases Positioning is particularly important for cyber-physical control applications in vertical domains like factories. The reason for this is that mobile devices and mobile assets are becoming increasingly common in the flexible production and subsequently the need for real-time locations data is increasing. In this context, positioning is especially important for warehousing and logistics processes, autonomous driving systems and fleet management and flexible adaptation in production. In the best case all relevant goods and products are continuously tracked from the moment they are received to the moment they are made available. The tracking process provides the relevant context information that is needed for real-time control and optimization of the material flow and subsequent production processes. In this scenario, autonomous driving systems fetch parts from the warehouse independently and transport them to flexible assembly cells on the shop floor. As part of the flexible fleet management system, these autonomous driving systems are continuously localized and move quickly and in constant interaction with their environment. In the process, production machinery and assembly cells and their given status are monitored seamlessly while relevant objects like tools and workpieces being localized. This makes it possible to adapt quickly to changes in circumstances. The result is flexible, autonomously controlled production that is capable of adapting to new situations at any time. Wireless positioning for human machine interfaces like AR/VR should also be possible. The requirements for positioning vary widley between different use cases. A.7.2 Low Power High Accuracy Positioning Low power high accuracy positioning is an integral part of a considerable number of industrial applications. The total energy needed for a specific operation time for such a low power high accuracy positioning optimized IoT-device is a combination of energy for positioning (varies depending on the used positioning method), energy for communication/‌synchronization and a difficult to predict factor to take additional losses through e.g. security, power management, microcontroller, and self-discharge of batteries into account. Examples of target applications for low power high accuracy positioning are asset tracking in process automation, tracking of vehicles, and tool tracking. Table A.7.2-1 gives an indication of the required operation time of the 5G enabled IoT device and duty cycle of the updated position information for different use cases. Table A.7.2-1: Low power high accuracy positioning use cases Use Case # Horizontal accuracy Corresponding service level (22.261) Positioning interval/ duty cycle battery life time/ minimum operation time 1 10 m Service Level 1 on request 24 months 2 2 m to 3 m Service Level 2 < 4 seconds > 6 months 3 < 1 m Service Level 3 no indication 1 work shift - 8 hours (up to 3 days, 1 month for inventory purposes) 4 < 1 m Service Level 3 1 second 6 - 8 years 5 < 1 m Service Level 3 5 seconds - 15 minutes 18 months 6 < 1 m Service Level 3 15 s to 30 s 6 - 12 months 7 30 cm Service Level 5 250 ms 18 months 8 30 cm Service Level 5 1 second 6 - 8 years (no strong limitation in battery size) 9 10 m Service Level 1 20 minutes 12 years (@20mJ/position fix) Use case one Process automation: Dolly tracking (outdoor). Use case two Process automation: Asset tracking. Use case three Flexible modulare assembly area: Tool tracking in flexible, modular assembly areas in smart factories. Use case four Process automation: Sequence container (Intralogistics). Use case five Process automation: Palette tracking (e.g. in turbine construction). Use case six Flexible modulare assembly area: Tracking of workpiece (in- and outdoor) in assembly area and warehouse. Use case seven Flexible modulare assembly area: Tool assignment (assign tool to vehicles in a production line, left/right) in flexible, modular assembly area in smart factories. Use case eight Flexible modulare assembly area: Positioning of autonomous vehicles for monitoring purposes (vehicles in line, distance 1.5 meter). Use case nine (Intra-)logistics: Asset tracking Annex B (informative): Communication service errors B.1 Introduction IEC 61784-3-3 describes fundamental communication errors that can be identified for applications with functional safety requirements [3]. The description of these communication errors is adjusted to field buses. These errors may however also occur in other communication systems. As explained in Annex C, some of these errors are also used for the assessment of communication services that do not support safety-critical applications. B.2 Corruption Messages may be corrupted due to errors within an application, due to errors on the transmission medium, or due to message interference. NOTE 1: Message error during transfer is a normal event for any standard communication system; such events are detected with high probability at receivers by use of, for instance, hash functions. NOTE 2: Most communication systems include protocols for recovery from transmission errors, so these messages will not be classed as 'loss' until recovery or repetition procedures have failed or are not used. NOTE 3: If the recovery or repetition procedures take longer than a specified deadline, a message is classed as "unacceptable delay". See also the discussion in Clause C.3. NOTE 4: In the very low probability event that multiple errors result in a new message with correct message structure (for example addressing, length, hash function such as CRC, etc.), the message will be accepted and processed further. Evaluations based on a message sequence number or a time stamp can result in fault classifications such as unintended repetition, incorrect sequence, unacceptable delay, insertion [3]. B.3 Unintended repetition Due to an error, fault, or interference, not updated messages are accidentally repeated. NOTE 1: Repetition by the sender is a common procedure when an expected acknowledgment/response is not received from a target station, or when a receiver detects a missing message and asks for it to be resent. In some cases, the lack of response can be detected, and the message repeated with minimal delay and no loss of sequence, in other cases the repetition occurs later and arrives out of sequence with other messages. NOTE 2: Some field buses use redundancy to send the same message multiple times or via multiple alternate routes to increase the probability of good reception [3]. B.4 Incorrect sequence Due to an error, fault, or interference, the predefined sequence (for example natural numbers, time references) associated with messages from a source is incorrect. NOTE 1: Field bus systems can contain elements that store messages (for example FIFOs in switches, bridges, routers) or use protocols that can alter the sequence (for example, by allowing messages with high priority to overtake those with lower priority). NOTE 2: When multiple sequences are active, such as transmission of messages from different source entities or reports relating to different object types, these sequences are monitored separately, and errors can be reported for each sequence [3]. B.5 Loss Due to an error, fault or interference, a message or acknowledgment is not received [3]. B.6 Unacceptable deviation from target end-to-end latency Messages may be delayed or advanced beyond their permitted arrival time window. Causes for this behaviour include errors in the transmission medium, congested transmission lines, interference, and applications sending messages in such a manner that communication services are delayed or denied. Message errors can be recovered in the following ways using scheduled or cyclic scans, for instance, in field buses: a) immediate repetition; b) repetition using spare time at the end of the cycle; c) treating the message as lost and waiting for the next cycle to receive the next value. In case of (a), all subsequent messages in that cycle are slightly delayed, while in case (b) only the resent message is delayed. Cases (a) and (b) are often not classed as an unacceptable deviation from the target end-to-end latency. Case (c) would be classed as an unacceptable delay for cyclic, distributed automation functions, unless the cycle repetition interval is short enough to ensure that delays between cycles are not significant and that the next cyclic value can be accepted as a replacement for the missed previous value before the survival time expiries (see Clause C.3) [3]. B.7 Masquerade Due to a fault or interference, a message is inserted that relates to an apparently valid source entity, so a non-safety related message may be received by a safety-related participant, which then treats it as safety related. NOTE: Communication systems used for safety-related applications can use additional checks to detect masquerade, such as authorised source identities and pass-phrases or cryptography [3]. B.8 Insertion Due to a fault or interference, a message is received that relates to an unexpected or unknown source entity. NOTE: These messages are additional to the expected message stream, and because they do not have expected sources, they cannot be classified as correct, unintended repetition, or incorrect sequence [3]. B.9 Addressing Due to a fault or interference, a safety-related message is delivered to the incorrect safety related participant, which then treats reception as correct [3]. Annex C (informative): Characterising communication services C.1 Modelling of communication in automation C.1.1 Area of consideration For our discussion of communication in automation we apply a definition of the area of consideration for industrial radio communication that is found elsewhere in the literature [4]. This definition is illustrated in Figure C.1.1-1. NOTE: Blue objects: communication system; other objects: automation application system. Figure C.1.1-1: Abstract diagram of the area of consideration for industrial radio communication Here, a distributed automation application system is depicted. This system includes a distributed automation application, which is the aggregation of several automation functions. These can be functions in sensors, measurement devices, drives, switches, I/O devices, encoders etc. All of these functions contribute toward the control of physical objects. Field bus systems, industrial Ethernet systems, or wireless communication systems can be used for connecting the distributed functions. The essential function of these communication systems is the distribution of messages among the distributed automation functions. For cyber-physical control applications, the dependability of the entire communication system and/or of its devices or its links is essential. Communication functions are realised by the respective hardware and software implementation. In order for the automation application system to operate, messages need to be exchanged between spatially distributed application functions. For that process, messages are exchanged at an interface between the automation application system and the communication system. This interface is termed the reference interface. Required and guaranteed values for characteristic parameters, which describe the behavioural properties of the radio communication system, as well as some influence quantities refer to that interface. The conditions that influence the behaviour of wireless communication are framed by the communication requirements of the application (e.g., end-to-end latency), the characteristics of the communication system (e.g., output power of a transmitter), and the transmission conditions of the media (e.g., signal fluctuations caused by multipath propagation). General requirements from the application point of view for the time and failure behaviour of a communication system are mostly related to an end-to-end link. It is assumed in the present document that the behaviour of the link is representative of the communication system as a whole and of the entire scope of the application. C.1.2 Logical link C.1.2.1 Nature and function Starting with the general approach mentioned in Subclause C.1.1, the logical link can be regarded as a possible asset within the area of consideration (see Figure C.1.1-1). The conditions under which its functions are to be performed are vital for the dependability of the automation application system. Figure C.1.2.1-1: The concept of a logical link This is the link between a logical end point in a source device and the logical end point in a target device. Logical end points are elements of the reference interface, which may group several logical end points together. The intended function of the logical link is the transmission of a sequence of messages from a logical source end point to the correct logical target end point. This is achieved by transforming each message into a form that fosters error-free transmission. The transmission process includes certain processes, for instance repetitions, in order to fulfil the intended function. After transmission, the transported package(s) is converted back into a message. The message is to be available and correct at the target within a defined time. The sequence of messages at the target is to be the same as the sequence at the source. The functional units, which are necessary to fulfil this function are shown, in Figure C.1.2.1-1. Figure C.1.2.1-2: The asset "logical link" The required function can be impaired by various influences, which can lead to communication errors. Such errors are described elsewhere in the literature [4][5]. A summary of these errors is provided in Annex A. The occurrence of one of these errors influences the values of the relevant dependability parameters of the logical link. C.1.2.2 Message transformation The present document addresses both OSI-layer-3 (IP) and OSI-layer-2 communication. The model in Figure C.1.2-1 can be used for describing both cases. The implementation of communication functions is split between a higher communication layer (HCL) and a lower communication layer (LCL). The partition of the layer for the two traffic options discussed in the present document is provided in Table C.1.2.2-1. This difference is of importance when discussing the implications of the service performance requirements in Clause 5 and Annex A for the network performance (see Clause C.5). Table C.1.2.2-1: Partition into higher communication layer and lower communication layer OSI level at which the traffic occurs Levels comprised by the higher communication layer Levels comprised by the lower communication layer 3 4 to 6 1 to 3 2 3 to 6 (note) 1 to 2 NOTE: In some vertical application, level 3 to 6 are not implemented. The messages to be transmitted for the intended function of a logical link are defined by strings of characters with a certain semantic. Such a character string is handed over as user data at the reference interface for transmission. If the number of characters in a message is too great for it to be transmitted as a unit, the message is divided for transmission into several packets (fragmentation). C.1.2.3 Communication device The communication devices—together with the physical link—determine the function and thus the dependability of the logical link. The function of the communication devices is the correct sending and correct receipt of sequences of messages. The asset "communication device" is depicted in Figure C.1.2.3-1. Figure C.1.2.3-1: Asset "communication device" C.1.2.4 Communication system The communication system as an asset represents a quantity of logical links whose message transmissions are implemented by wireless devices via one or more media. The communication system function to be provided consists in transmitting messages for all the logical links in the distributed application. This function is to be performed for a defined period, the operating time of the automation application. In an automation application system, it is paramount that requirements pertaining to logical links are fulfilled. These requirements and the conditions can be very different from one case and implementation to the other. The functions (services and protocols) for individual logical links can therefore also be different. Despite these differences, some of the logical links share communication devices and media. C.2 Communication service description C.2.1 Overview Tables C.2.2-1, C.2.3-1 summarise candidate interface parameters for the description of the communication service performance. The lists are grouped according to whether the parameter stands for automation characteristic parameters (Table C.2.2-1) or influence quantities (Table C.2.3-1). The meaning of the columns and rows is explained after each table. NOTE 1: Not all parameters in Table C.2.2-1 and Table C.2.3-1 would be used in a service call. NOTE 2: Ingress and egress in this clause are in reference to the communication service interface between the source application and the communication service interface (ingress) and the communication service and the target application (egress). C.2.2 Characteristic parameters Table C.2.2-1: Candidate characteristic parameters for the dependable communication service interface Parameter name Typical metric (unit) Traffic class (note) Deterministic periodic communication Deterministic aperiodic communication Non-deterministic communication Communication service availability Minimum availability (dimensionless) X X X End-to-end latency Target value and timeliness (ms) X X X Communication service reliability Mean time between failures (days) X X X Service bit rate Target value (bit/s); user experienced data rate (bit/s); time window (s) – X X Update time Target value and timeliness X – – NOTE: – application requirements (KPIs). X: applies; –: does not apply. Parameter description Communication service availability This parameter indicates if the communication system works as contracted ("available"/"unavailable" state). The communication system is in the "available" state as long as the availability criteria for transmitted packets are met. The service is unavailable if the packets received at the target are impaired and/or untimely (e.g. update time > stipulated maximum). If the survival time (see Table C.2.3-1) is larger than zero, consecutive impairments and/or delays are ignored until the respective time has expired. End-to-end latency This parameter indicates the time allotted to the communication system for transmitting a message and the permitted timeliness. Communication service reliability Mean time between failures is one of the typical indicators for communication service reliability. This parameter states the mean value of how long the communication service is available before it becomes unavailable. For instance, a mean time between failures of one month indicates that a communication service runs error-free for one month on average before an error/errors make the communication service unavailable. Usually, an exponential distribution is assumed. This means, there will be several failures where the time between two subsequent errors is below the mean value (1 month in the example). Communication service availability and communication service reliability (mean time between failures) give an indication on the time between failures and the length of the failures. Service bit rate a) deterministic communication The target value indicates committed data rate in bit/s sought from the communication service. This is the minimum data rate the communication system guarantees to provide at any time, i.e. in this case target value = user experienced data rate. b) non-deterministic communication The target value indicates the target data rate in bit/s. This is the information rate the communication system aims at providing on average during a given (moving) time window (unit: s). The user experienced data rate the lower data rate threshold for any of the time windows. Update time Applicable only to periodic communication, the update time indicates the time interval between any two consecutive messages delivered from the egress (of the communication system) to the application. Traffic classes In practice, vertical communication networks serve applications exhibiting a wide range of communication requirements. In order to facilitate efficient modelling of the communication network during engineering, and for reducing the complexity of network optimisation, disjoint QoS sets have been identified. These sets are referred to as traffic classes [6]. Typically, only three traffic classes are needed in industrial environments [6], i.e. - deterministic periodic communication; - deterministic aperiodic communication; and - non-deterministic communication. Deterministic periodic communication stands for periodic communication with stringent requirements on timeliness of the transmission. Deterministic aperiodic communication stands for communication without a pre-set sending time. Typical activity patterns for which this kind of communication is suitable are event-driven actions. Non-deterministic communication subsumes all other types of traffic. Periodic non-real time and aperiodic non-real time traffic are subsumed by the non-deterministic traffic class, since periodicity is irrelevant in case the communication is not time-critical. Usage of the parameters in Table C.2.2-1 Control service request and response; monitoring service response and indication. C.2.3 Influence quantities Table C.2.3-1: Candidate application influencing parameters for the dependable communication service interface Parameter name Typical metric (unit) Traffic class (note) Usage of this parameter Deterministic periodic communication Deterministic aperiodic communication Non-deterministic communication Burst Maximum user data length (byte) and line rate of the communication service interface (bit/s) – X X Service request and response; monitoring service response and indication Message size Maximum or current value (byte) X (X) (X) Service request and response; non-deterministic data transmission; deterministic aperiodic data transmission Service time interval Start (time) and end (time) X X X Service request and response Survival time Maximum (s) X X – Service request and response Transfer interval Target value and timeliness (s) X – – Service request and response NOTE: X: applies; (X): usually does not apply; –: does not apply. Parameter description Burst The transmission of, for instance, program code and configuration data may be handed to the 3GPP system as data burst. In this case, the ingress data rate exceeds the capacity of the network, which implies that some of the data has to be stored within the ingress node of the communication system before it can be transmitted to the egress interface(s). However, the application consuming the communication service requires that the data of such a burst needs to be transmitted completely. This is in contrast to periodic data transmission, where new messages overwrite old ones. Typical metrices for bursts: maximum user data length and line rate of the communication service interface. Message size The user data length indicates the (maximum) size of the user data packet delivered from the application to the ingress of the communication system and from the egress of the communication system to the application. For periodic communication this parameter can be used for calculating the requested user-experienced data rate. If this parameter is not provided, the default is the maximum value supported by the PDU type (e.g. Ethernet PDU: maximum frame length is 1,522 octets, IP PDU: maximum packet length is 65,535 octets). Service time interval Describes the start and end time of a communication service. Note that there are other ways to describe the service time interval numerically, for instance as the tuple [start time, service duration]. Survival time The maximum survival time indicates the time period the communication service may not meet the application's requirement before the communication service is deemed to be in an unavailable state. NOTE 1: The survival time indicates to the communication service the time available to recover from failure. This parameter is thus tightly related to maintainability [7]. Transfer interval Applicable only to periodic communication, the transfer interval indicates the time elapsed between any two consecutive messages delivered by the automation application to the ingress of the communication system. C.3 Up time and up state vs. down state and down time The assessment of periodic deterministic communication services is based on the assessment of successful message transmission over a logical communication link. Message transmission is either: - successful, if it is correctly and timely received, or - unsuccessful, if it is incorrectly received, lost or untimely. Up time and down time can be derived from received messages. As far as timely received messages are correct, the logical communication link status is up. If a message loss or an incorrectly or untimely received message is detected the logical communication link status is down. To denote up and down states the terms “up time interval” and “down time interval”, or alternatively “available” and “unavailable” may be used. An example of the relation between logical communication link status, communication service status and application status is presented in Figure C.3-1. Figure C.3-1: Relation between logical communication link, communication service and application statuses (example with lost messages) The flow of events in Figure C.3-1 is as follows: a) The logical communication link is up and running (blue line is UP). A source device starts sending periodic messages to a target device (orange arrows), on which an automation function (application) is running. The communication service is, from the point of view of the target application, in an up state (violet line is UP) and so is the application (green line is UP). b) The logical communication link status changes to down state if it no longer can support end-to-end transmission of the source device's messages to the target device in agreement with the negotiated communication requirements. Once the application on the target device senses the absence (or unsuccessful reception) of expected messages ("Deadline for expected message" in Figure C.3-1), it will wait a pre-set period before it considers the communication service to be unavailable ; this is the so-called survival time. The survival time can be expressed as - a period or, - especially with cyclic traffic, as maximum number of consecutive incorrectly received or lost messages. c) If the survival time has been exceeded, both the communication service and the application transition into a down state (violet and green lines change to DOWN in Figure C.3-1). The application will usually take corresponding actions for handling such situations of unavailable communication services. For instance, it will commence an emergency shutdown. Note that this does not imply that the target application is "shut off"; rather it transitions into a pre-defined state, e.g. a safe state. In the safe state, the target application might still listen to incoming packets or may try to send messages to the source application. d) Once the logical communication link status is in the up state again (blue line in Figure C.3-1 changes to UP), the communication service state as perceived by the target application will change to the up state. The communication service is thus again perceived as available (violet line changes to UP in Figure C.3-1). The state of the application, however, depends on the counter measures taken by the application. The application might stay in down state if it is in a safe state due to an emergency shutdown. Or, the application may do a recovery and change to up state again. The time needed for the application to return to the up state after the communication service is restored is shown as “Application recovery time” in Figure C.3-1. The availability of the communication service is calculated using the accumulated down time. For instance, in case the communication service is expected to run for a time T, the unavailability U of the communication service can be calculated as Where Δti is the length of the i-th downtime interval of the communication service within the time period T. The communication service availability A can then be calculated as A = 1–U. C.4 Timeliness as an attribute for timing accuracy C.4.1 Overview There are several time parameters in dependability assessment. A required value is specified for every time parameter. This value can be a maximum, mean, modal, minimum etc. Typically, there is a deviation from the desired value to the actual value. Jitter is often used to characterise this variation. Since jitter generally is used for characterising the behaviour of a measured parameter, for instance the scatter of measured end-to-end latencies ("the world as it is"), it can be quite confusing to use it for formulating service performance requirements ("the world as we want it to be"). What is needed is a concept and related parameters that allow for formulating and talking about the end-to-end latency requirements in Clause 5 and Annex A. The most important attribute is timeliness. Timeliness can be formulated a permitted interval for the actual value of the time parameter. Accuracy, earliness and lateness describe the allowed deviation from a target value. Accuracy is the magnitude of deviation. It can be negative (early) or positive (tardy). C.4.2 Network latency requirement formulated by use of timeliness In 5G networks, the end-to-end latency KPI is a critical KPI in order to ensure that the network can deliver the packet within a time limit specified by an application: not too early and not too late. In cyber-physical automation, the arrival time of a specific packet should be strictly inside a prescribed time window. In other words, a strict time boundary applies: [minimum end-to-end latency, maximum end-to-end latency]. Otherwise, the transmission is erroneous. Although most use cases that require timely delivery only specify the maximum end-to-end latency, the minimum latency is also sometimes prescribed. In the latter case, a communication error occurs if the packet is delivered earlier than the minimum end-to-end latency. An example for a related application is putting labels at a specific location on moving objects, and the arrival of a message is interpreted as a trigger for this action. In other words, the application does not keep its own time, but interprets the message arrival as clock signal. Maximum and minimum end-to-end latency alone do not disclose which value is preferred, i.e. target value. The next three subclauses introduce concepts help with relating maximum end-to-end latency, minimum end-to-end latency, and target vale to each other. C.4.3 Timeliness Timeliness is described by a time interval (see Figure C.4.3-1). The interval is restricted by a lower bound (tLB) and an upper bound (tUB). This interval contains all values tA that are within an accepted "distance" to the target value tR. Figure C.4.3-1: Timeliness function A message reception is considered in time, if it is received within the timeliness interval. If it is received outside the timeliness interval, the message reception is considered invalid. This is related to the communication error "unacceptable deviation from target end-to-end latency" (see Subclause B.6). In other words, maximum end-to-end latency = tUB and minimum end-to-end latency = tLB. Timeliness is related to deviation (see Subclause C.4.4), the lower bound tLB is related to earliness (see Subclause C.4.5), and the upper bound tUB is related to lateness (see Subclause C.4.6). C.4.4 Deviation The term deviation describes the discrepancy between an actual value (tA) and a target value (tR). Deviation(tA) = tA – tR. Figure C.4.4-1 shows two examples. The target value is 10 time units (tR = 10) in both cases. In the first case (blue) the actual value measures 12 time units (tA = 12). The difference of both amounts to +2 time units, which means that the deviation is 2 time units [Accurracy(tA) = 2]. The second case (purple) shows the actual value as 9 time units (tA = 9). The difference of both amounts to -1 time unit, which means that the deviation is –1 time units [Accuracy(tA) = –1]. Figure C.4.4-1: Examples for accuracy values Figure C.4.4-2 shows the deviation with respect to the target time (t). The following applies: Deviation(t) < 0 for t < tR; that is, the arrival is early. Deviation(t) = 0 for t = tR; that is, the arrival is as desired, i.e. on time. Deviation(t) > 0 for t > tR; that is, the arrival is late (see also C.4.6) Figure C.4.4-2: Accuracy function C.4.5 Earliness Earliness describes how early the actual value is: earliness is greater than 0 if the actual value is less than the target value (see Figure C.4.5-1). The following applies: Eearliness(tA) = tR – tA = –Deviation(tA) for tA < tR; Eearliness(tA) = 0 for tA  tR. In an example, the target value is 10 time units (tR = 10), and the actual value is 7 time units (tA = 7). The difference of both is 3 time units with respect to being early. That means that the earliness is 3 time units [Eearliness(tA) = 3]. Figure C.4.5-1: Earliness function C.4.6 Lateness Lateness describes how much greater the actual value is than the target value: lateness is greater than 0 if the actual value is greater than the desired value (see Figure C.4.6-1). The following applies: L(tA) = 0 for tA  tR; L(tA) = tA–tR = Deviation(tA) for tA > tR. In an example, the target value is 10 time units (tR = 10), and the actual value measures 14 time units (tA = 14). The difference of both is 4 time units with respect to being late. That means that the lateness is 4 time units [L(tA) = 4]. Figure C.4.6-1: Lateness function C.4.7 Conclusion Using the concepts of earliness and lateness (see Subclauses C.4.5 and C.4.6, respectively), the maximum and minimum end-to-end latency can be rewritten as follows. Maximum end-to-end latency = target end-to-end latency + maximum lateness; Minimum end-to-end latency = target end-to-end latency – maximum earliness. C.5 Communication service terminology w.r.t. 5G network and vertical applications This section clarifies the wording and terminology with respect to communication interfaces that are relevant for vertical applications. Because the 3GPP network does not cover the complete ISO-OSI communication stack, it is important to distinguish between - the vertical applications’ point of view, and - the 3GPP network’s point of view. In this section, the relation between those two is clarified. Figure C.5-1 shows a simplified version of the communication stack. The PHY layer, the MAC layer and some parts of the IP layer are part of the 3GPP network. The layers that are part of the 3GPP network are referred to as lower communication layers (LCL). The communication stack also includes an application. The OSI layers related to providing data to the application are referred to as the higher communication layers (HCL). The interface between LCL and HCL is referred to as communication service interface (CSIF). For the assessment of the overall system performance, it is important to differentiate between the 3GPP network’s performance (i.e., including only the LCL and measured at the CSIF) and the overall system performance including the application layer (i.e., including both, the LCL and the HCL). In Figure C.5-1, the orange arrow depicts the vertical application’s point of view. The blue arrows indicate two options to measure the 3GPP network’s performance, i.e., including and excluding the IP layer. Figure C.5-1: Network performance measurements at different communication system interfaces (CSIF) Figure C.5-2 illustrates how messages are transmitted from a source application device (e.g., a programmable logic controller) to a target application device (e.g. an industrial robot). The source application function (AF) is executed in the source operating system (OS) and hands over a message to the application layer interface of the source communication device. In the higher communication layers (HCL), which are not part of the 3GPP system, the data is processed. From the HCL the data is transferred to the lower communication layers (LCL), which are part of the 3GPP system. After transmission through the physical communication channel and the LCL of the target communication device, the data is passed to the HCL and lastly to the target application device. Characteristic parameters with respect to time are defined in Figure C.5-2. From 3GPP system point of view: - Transfer interval of 5G system: Time between the arrival of two pieces of data at the source CSIF. - End-to-end latency: Time measured from the point when a piece of data received at the CSIF in the source communication device until the same piece of data is passed to the CSIF in the target communication device. From vertical application point of view: - Transfer interval of vertical application: Time between the transmission of two successive pieces of data from the source application. - Transmission time: Time measured from the point when a piece of data is handed from the application layer interface of the source application device, until the same piece of data is received at the application layer interface of the target application device. - Update time: Time between the reception of two consecutive pieces of data at the application layer interface to the target application device. If not stated otherwise, the terms "end-to-end latency" and "transfer interval" refer to the 3GPP system / 5G network parameters in this document. Figure C.5-2: Relation between application device and communication device (downlink example). Annex D (informative): 5G in industrial automation: different and multiple time domains for synchronization D.1 Description The required synchronization precision is usually given as the maximum absolute value of the time difference between sync master and any device in the synchronisation domain (time domain or clock domain). A common example is a synchronisation precision of ≤ 1 µs. This is equivalent to ± the precision value, so ±1 µs between sync master and any device in the synchronisation domain, resulting in two times this value as maximum absolute time difference between any two devices in the synchronisation domain (2 µs in the example). An industrial automation network generally consists of two distinct time domains. First is the global time domain. This is the time used for overall synchronization in the system (e.g. the factory). It is used to align operations and events chronologically. Industrial automation uses the term universal time domain [20] for the global time domain described in this document. Global time is known as a synonym for universal time in industrial automation. Global time is called wall clock in certain areas and standards. The synchronization precision is typically ≤ 1µs [20]. In some areas, a precision of ≤ 100 µs might be sufficient for the global time domain if a working clock with precision of ≤ 1 µs is available. The assigned timescale is usually the International Atomic Time (TAI, temps atomique international), based on the precision time protocol (PTP) epoch (starting from 1 January 1970 00:00:00 TAI) [22]. While there is usually only one global time, multiple global time domains are possible. Clock synchronization in the global time domain usually applies to all UEs within the industrial facility in industrial automation. That is, a global time domain covers usually the industrial facility. Second is the working clock domain. Working clock domains are constrained in size. They often consist of a single machine or a set of neighbouring machines that physically collaborate. The restricted size allows very precise time synchronization (≤1µs) with efficient network components. Synchronisation to a working clock is used to align e.g. production lines, production cells, or machines/functional units. In these cases, the application synchronizes locally within the working clock domains (Figure D.1-1), allowing precise synchronization with more efficient components. A global time domain usually contains multiple working clock domains. The starting point (epoch) is the start of the working clock domain. Figure D.1-1: Global time domain and working clock domains The assigned timescale of a working clock domain is arbitrary (timescale ARB [22]). Therefore, different working clock domains may have different timescales and different synchronisation accuracy and precision. Robots, motion control applications, numeric control, and any kind of clocked / isochronous application rely on the timescale of the working clock domain to make sure that actions are precisely interwoven as needed. Clock synchronization in the working clock domain is constraint in size. A specific working clock domain will contain only a subset of the UEs within the industrial facility. Often, the UEs of the working clock domain are connected to the same gNB. However, it is also possible that a working clock domain contains multiple neighbouring gNBs. This depends on the actual use case and its vertical application. Devices may be part of multiple time domains leading to overlapping working clock domains. The required precision (usually ≤ 1 µs) is between the sync master and any sync devices of the clock domain, both, global time domain and working clock domain. Clock domains might be called sync domains in certain areas and standards. D.2 Merging of working clock domains One key issue of the integration of TSN and 5G wireless networks that has to be handled is mobility. The integration of 5G wireless communication into the industrial communication infrastructure allows for mobility in the manufacturing process. This mobility enhances flexibility in the manufacturing process, e.g. through adding certain manufacturing capabilities on-demand by having a machine move to the respective production line. This means that machines that are synchronized to different working clock domains may need to interact with each other. The following scenario illustrates this. After the mobile machine has arrived at the intended location and is stationary again, the two interacting working clock domains have to be synchronized with each other. Otherwise interaction might not be possible without interfering with ongoing operations. An example is an autonomous mobile handling robot adding parts to an assembly line. Without synchronization between both, correct placement of the parts would be impossible. However, it is not feasible to schedule these interactions beforehand. Therefore, the interaction between different working clock domains requires a concept for handling the communication. TSN provides already mechanisms for this. The 5G systems and the UEs need to provide an interface in order to exchange information of the clock domain. Figure D.2-1: Working clock domain interactions "Merge" and "Separate" When members of different working clock domains interact, there are two possible options (Figure D.2-1). Which option is used depends on the application and its requirements. - Merge: The working clock domains merge into one. This option can be used in applications where synchronization is critical, e.g. high precision robots interacting with each other. - Separate: The members of the different working clock domains interact while keeping their own separate time synchronizations. This option can be used in applications where synchronization is non-critical, e.g. an AGV collecting finished products from a production line. D.3 Time synchronization with 5G networks For the time synchronization with 5G networks, we consider two possible options. The 5G system uses the IEEE 802.1AS sync domains [22]: In this case, the 5G system provides a media dependent interface to the IEEE 802.1AS sync domain, which the application can use to synchronize to the sync domain. In the IEEE 802.1AS standard [22], a similar concept is detailed in the MDSyncSend and MDSyncReceive structures. The 5G system provides the working clock domains and global time domain: In this case, the 5G system has to provide an interface which the application can use to derive their working clock domain or global time domain. A device can belong to multiple working clock domains. An application can use each of these as the reference clock for synchronization (reference clock model). NOTE: The required precision (usually ≤ 1 µs) is between the sync master and any sync device of the clock domain. Annex E (informative): Audio and Video Production E.1 Description AV production includes television and radio studios, outside and remotely controlled broadcasts, live newsgathering, sports events, music festivals, among others. All of these applications require a high degree of reliability, since they are related to the capturing and transmission of data at the beginning of a production chain. This differs drastically when compared to other multimedia services because the communication errors will be propagated to the entire audience that is consuming that content both live and recorded for later distribution. Furthermore, the transmitted data is often post-processed with nonlinear filters which could actually amplify defects that would be otherwise not noticed by humans. Therefore, these applications call for uncompressed or only slightly compressed data, and very low probability of errors. These devices will also be used alongside existing technologies which have a high level of performance and so any new technologies will need to match or improve upon the existing workflows to drive adoption of the technology. The performance aspects that are covered by/in TS 22.263 [27] (Service requirements for Video, Imaging and Audio for professional applications) also target the latency that these services experience. In recent years, production facilities have moved from bespoke unidirectional highly specialised networks to IP based systems and software-based workflows. This migration is expected to continue, and wireless IP connectivity is key to a number of these workflows. Typical set ups require multiple devices such as cameras, microphones and control surfaces that require extremely close synchronisation to maintain consistency of pictures and audio. Such clock synchronization requirements are captured in clause 5.6. Often devices need to communicate directly to each other for instance a camera to a monitor or a microphone to a PA system. Video and audio applications also require extremely high quality of service metrics as the loss of a single packet can cause picture or sound breakup in the downstream processing or distribution. Often this is a legal, regulatory or contractual agreement to maintain a high quality, stable and clear video or audio signal. E.2 Multiple source wireless studio This use case will deploy a multiple camera studio of approximately 1,000 m2 (~5 cameras) where wired and wireless functionalities currently provided by traditional infrastructure technologies are likely to be deployed using a standalone non-public network. A combination of IP enabled wired and wireless cameras working at both HD and UHD resolutions will be deployed in a studio. Associated equipment such as video monitors, prompting systems, camera control will be provided over the 5G network. Camera timing and synchronisation will be provided over the 5G system. As well as video, audio will be sourced from both wired and wireless microphones incl. control/monitoring and combined with the video to produce high quality synchronised AV content. 5G will also be deployed to control lighting and camera robotics. Talkback intercom systems will be deployed using low latency multicast links. Today’s digital AV network transport is typically handled separately for wireless and wired transfers (see figure E.1.2-1). Wireless AV transmissions are implemented with application specific solutions that allow deterministic data transport of a single isolated audio or video link. Wired AV transmissions are Ethernet / IP based. Quality of service in AV IP networks is mainly achieved with IP DiffServ / DSCP based prioritization of packets in network switches. This method is sufficient for most AV use cases since jitter resulting from packet collisions is small, for example in the order of 10 µs per concurring data stream in Gbit Ethernet. Figure E.2-1: Typical AVPROD setup The microphones and cameras can be co-located in a broadcast centre in which case they would communicate through a LAN or NPN. For remote production operations the mixing and production console may be separated by some distance (existing examples are cross continental. In this instance they may communicate via a PLMN or combination of PLMN and WAN networks. Some approaches may also deploy main (leader) equipment at the broadcast centre with secondary (follower) equipment at the location site to reduce latency. Other aspects of this workflow may also include robotic control where both the physical position (height, direction and tilt) and the technical control (focus, zoom, iris, colour) of a device of a camera, microphone or light may be controlled remotely. In this instance round trip latency of < 20ms is required in order for an operator to see a move reflected at his control position as it is made. It is important to note that these are a combination of automated robotics (pre-programmed moves) and manually controlled robotics (following an unpredictable event such a sport). E.3 Timing use in AV production applications Timing of multiple devices such as microphones and camera is also critical. Timing signals are used in 2 separate ways. - To maintain synchronisation between devices so that electronic shutters on cameras operate at the same time and frequency and that when cutting between any two cameras pointed at the same source no discernible jumps can be seen. This requires accuracy within the frame boundary of a given video signal. A single frame of video at 120 Hz would require a clock accurate to within 8 ms. - To timestamp an IP packet carrying a video or audio sample. Existing standards and workflows for AVPROD rely on IEEE 1588 PTP timing with a SMPTE media profile applied. This requires a clock accurate to within 1 µs Figure E.3-1 Typical IP based timing set up for AVPROD Annex F (informative): Relation of reliability and communication service availability Availability and reliability are used both in 3GPP and vertical industries, but with different meanings. Communication service availability addresses the availability of a communication service. This definition follows the vertical standard IEC 61907 [7]. On the other hand, reliability is a 3GPP term and addresses the availability of a communication network. The relation of both terms is depicted in figure F-1 for a mobile network. Figure F-1: Illustration of the concepts reliability and communication service availability. As depicted, reliability covers the communication-related aspects between two nodes (here: end nodes), while communication service availability addresses the communication-related aspects between two communication service interfaces. This might seem to be a small difference, but this difference can lead to situations, where reliability and communication service availability have different values. Example: traffic gets "stuck" The related scenario is depicted in figure F-2. Figure F-2: Example in which reliability and communication service availability have different values. Packets are reliably transmitted from the communication service interface A to end node B, but they are not exposed at the communication service interface B. This scenario addresses unicast communication from application A to B. The packets are handed over at the communication service interface A from the application to the communication network, and the packets are then transmitted to the end node B. In this example, the packets received by end node B are not exposed at the communication service interface B. So, even if all packets that are handed over to end node A are successfully delivered to end node B within the time constraint required by the targeted service (reliability = 100 %), the communication service availability is 0% since no packets arrive at the "end", namely the communication service interface B. Example: packets dropped at the communication service interface The related scenario is depicted in figure F-3. Figure F-3: Example in which reliability and communication service availability have different values. Only half of the packets handed over to the end node A are actually transmitted to end node B and then handed over to application B at the communication service interface B. This scenario describes unicast communication of evenly interspersed packets from application A to B. The packets are handed over at the communication service interface A from the application to the communication network, and the packets are then transmitted to the end node B. However, only every second packet is actually successfully handed over to end node A and then transmitted to end node B. Thus, only half of the packets arrive at application B. Note though that the reliability of the mobile network is 100%, since all packets transmitted by end node A are delivered to end node B within the time constraint required by the targeted service. However, depending on the agreed QoS, the communication service availability can be of the same value as the reliability or much lower. For instance, if the agreed survival time is equal to or larger than the end-to-end latency, reliability and communication service availability are equal. However, if the survival time is smaller, the reliability is two times the communication service availability. Note that the shortest time interval over which the communication service availability should be calculated is the sum of maximum allowed end-to-end latency and survival time. Annex G (informative): Change history Change history Date Meeting TDoc CR Rev Cat Subject/Comment New version 2018-05 SA1#82 S1-181551 – – – Skeleton for TS 22.104 ("Service requirements for cyber-physical control applications in vertical domains") 0.0.0 2018-05 SA1#82 S1-181552 – – – Includes agreements at SA1#82, Dubrovnik, Croatia 0.1.0 2018-08 SA1#83 S1-182344 – – – Includes agreements at SA1#83, West Palm Beach, Florida 0.2.0 2018-11 SA1#84 S1-183276 – – – Includes agreements at SA1#83, Spokane, WA, USA, rapporteur’s clean-up 0.3.0 2018-12 SA#82 SP-181006 Presentation to SA for one-step approval 1.0.0 2018-12 SA#82 SP-181006 Raised to v.16.0.0 following SA approval 16.0.0 2019-03 SA#83 SP-190081 0003 2 F Clarifying UE-to-UE versus UE-to-network 16.1.0 2019-03 SA#83 SP-190081 0002 1 F Moving rail-bound mass transit requirements – shift from cyberCAV 16.1.0 2019-03 SA#83 SP-190081 0001 1 F Clean-up and corrections of TS 22.104 cyberCAV 16.1.0 2019-06 SA#84 SP-190299 0008 F Corrections to TS 22.104 v16.1.0 16.2.0 2019-06 SA#84 SP-190299 0006 2 F Add missing abbreviations to TS 22.104 16.2.0 2019-06 SA#84 SP-190299 0005 2 C Adding edge computing aspect 16.2.0 2019-06 SA#84 SP-190311 0009 1 C Adding vertical positioning requirements to TS 22.104 v16.1.0 17.0.0 2019-09 SA#85 SP-190807 0012 1 B Addition of a new synchronisation performance requirement 17.1.0 2019-09 SA#85 SP-190807 0011 2 B Addition of robotic aided surgery and diagnosis performance requirements 17.1.0 2019-09 SA#85 SP-190800 0023 1 A Correction of a figure number in Annex D.2 17.1.0 2019-09 SA#85 SP-190812 0010 B Add one more case for control-to-control communication 17.1.0 2019-09 SA#85 SP-190812 0013 1 B Network operation requirements 17.1.0 2019-09 SA#85 SP-190812 0015 3 B eCAV – Further 5G service requirements for Positioning 17.1.0 2019-09 SA#85 SP-190812 0018 2 B eCAV – Service performance requirements for Industrial Wireless Sensors 17.1.0 2019-09 SA#85 SP-190812 0016 2 B eCAV – further 5G service requirements for wired to wireless link replacement for smart manufacturing / Industry 4.0 17.1.0 2019-09 SA#85 SP-190812 0020 3 B eCAV – further 5G service requirements for network performance 17.1.0 2019-09 SA#85 SP-190857 0021 3 B ECAV - further 5G service requirements for ProSe communication for CAV 17.1.0 2019-09 SA#85 SP-190856 0019 3 B ECAV - further 5G service requirements for industrial Ethernet integration (clock synchronization, time-sensitive communication) 17.1.0 2019-12 SA#86 SP-191016 0027 5 B Addition of informative annex for AV Prod 17.2.0 2019-12 SA#86 SP-191020 0026 1 F Correction of CMED KPIs tables 17.2.0 2019-12 SA#86 SP-191028 0034 4 A Clarification of clock synchronicity requirements 17.2.0 2019-12 SA#86 SP-191028 0028 F Addition of transmission directions and movement characteristics 17.2.0 2019-12 SA#86 SP-191028 0030 1 A Clarification on communication service reliability 17.2.0 2019-12 SA#86 SP-191028 0035 1 F Derivation of communication service availability and reliability from network performance metrics 17.2.0 2019-12 SA#86 SP-191028 0037 1 D Editorial and minor corrections to TR 22.104 17.2.0 2019-12 SA#86 SP-191028 0032 2 C Network performance requirements for mobile operation panel 17.2.0 2019-12 SA#86 SP-191028 0038 2 F TS 22104 - Annex A for cooperative carrying 17.2.0 2020-07 SA#88e SP-200568 0050 F Correction of service performance requirements in tables of annex A.6 17.3.0 2020-07 SA#88e SP-200567 0049 D 22.104 Miscellaneous editorial corrections 17.3.0 2020-07 SA#88e SP-200562 0041 2 A Clarifications to communication service performance requirements 17.3.0 2020-07 SA#88e SP-200567 0048 2 A Miscellaneous values for further study 17.3.0 2020-07 SA#88e SP-200562 0045 2 A Correcting description of communication service status in Clause C.3 17.3.0 2020-07 SA#88e SP-200562 0047 1 A Clock synchronicity budget for the 5G system 17.3.0 2020-09 SA#89e SP-200790 0055 F Quality improvement – burst definition 17.4.0 2020-09 SA#89e SP-200790 0058 1 D CR 22.104 R17 - Editorial Improvements – Decimal Separator 17.4.0 2021-03 SA#91e SP-210200 0063 D Non-inclusive language replacement 22.104 17.5.0 2021-03 SA#91e SP-210217 0064 1 B Adding energy efficiency use cases for positioning to the ANNEX A 17.5.0 2021-06 SA#92e SP-210564 0075 1 A Quality improvement - update of definition of communication service availability 18.1.0 2021-06 SA#92e SP-210565 0066 1 D Quality improvement - addition of new annex (relationship between reliability and communication service availability) 18.1.0 2021-06 SA#92e SP-210565 0068 1 D 22.104 - V18.0.0 - quality improvement - update of mobile-robots use case description 18.1.0 2021-06 SA#92e SP-210565 0069 1 F Correction of mobile-robot use cases (UE number) 18.1.0 2021-06 SA#92e SP-210565 0070 1 D Quality improvement - service duration 18.1.0 2021-06 SA#92e SP-210517 0065 1 B 5G timing resiliency 18.1.0 2021-06 SA#92e SP-210524 0073 D Alignment of positioning power consumption aspects between 22.261 and 22.104 18.1.0 2021-06 SA#92e SP-210524 0072 1 B Adding LPHAP requirements for Industrial IoT 18.1.0 2021-09 SA#93e SP-211038 0077 2 A Quality improvement: update of reference to IEEE 802.1AS 18.2.0 2021-09 SA#93e SP-211070 0078 1 B Introduction of Smart Energy Infrastructure Requirements 18.2.0 2021-09 SA#93e SP-211070 0080 1 B Annex for smart grid 18.2.0 2021-09 SA#93e SP-211070 0082 1 B Introduction of SEI KPIs 18.2.0 2021-09 SA#93e SP-211070 0083 1 D Adjusting scope clause in TS 22.104 to the specification s content 18.2.0 2021-09 SA#93e SP-211039 0084 1 F Clarification of requirements for clock synchronization with direct device connection and indirect network connection communication 18.2.0 2021-09 SA#93e SP-211070 0087 1 B Inclusion of Smart Energy Infrastructure Requirements 18.2.0 2021-12 SP-94 SP-211497 0089 1 C Correction of references for clause 2 18.3.0 2021-12 SP-94 SP-211497 0090 D Remove editor note for Figure A.4.4.3-1 18.3.0 2021-12 SP-94 SP-211497 0091 1 F Update to Smart Grid normative requirements 18.3.0 2023-03 SA#99 SP-230230 0093 3 B Additional clarification on security, privacy for mobile robots using edge cloud 19.0.0 2023-03 SA#99 SP-230230 0094 4 B An additional usecase for Industrial edge cloud regarding digital twin usage 19.0.0 2023-09 SA#101 SP-231040 0097 1 C Smaller 5GS time sync budget 19.1.0 2024-06 SA#104 SP-240786 0100 1 A Correction of reference to IEEE Std 1588-2019 19.2.0
28a605aed849a8e4682a36cdb0077b63
24.186
1 Scope
The present document provides the protocol details for enhancements to IMS multimedia telephony communication services enabled by supporting the IMS data channel and for AR communication which is one of the applications based on IMS data channel capability, based on stage 1 requirements in 3GPP TS 22.261 [2] and stage 2 requirements in 3GPP TS 23.228 [3]. The present document is applicable to User Equipment (UE), Application Servers (AS)and IP Multimedia (IM) Core Network (CN) subsystem which are intended to support IMS multimedia telephony communication services supporting the IMS data channel and AR communication which is one of the applications based on IMS data channel capability.
28a605aed849a8e4682a36cdb0077b63
24.186
2 References
The following documents contain provisions which, through reference in this text, constitute provisions of the present document. - References are either specific (identified by date of publication, edition number, version number, etc.) or non‑specific. - For a specific reference, subsequent revisions do not apply. - For a non-specific reference, the latest version applies. In the case of a reference to a 3GPP document (including a GSM document), a non-specific reference implicitly refers to the latest version of that document in the same Release as the present document. [1] 3GPP TR 21.905: "Vocabulary for 3GPP Specifications". [2] 3GPP TS 22.261: "Service requirements for the 5G system; Stage 1". [3] 3GPP TS 23.228: "IP Multimedia Subsystem (IMS); Stage 2". [4] 3GPP TS 26.114: "IP Multimedia Subsystem (IMS); Multimedia Telephony; Media handling and interaction". [5] IETF RFC 5688: "A Session Initiation Protocol (SIP) Media Feature Tag for MIME Application Subtype". [6] IETF RFC 6809: "Mechanism to Indicate Support of Features and Capabilities in the Session Initiation Protocol (SIP)". [7] IETF RFC 3264: "An Offer/Answer Model with the Session Description Protocol (SDP)". [8] 3GPP TS 22.173: "IP Multimedia Core Network Subsystem (IMS) Multimedia Telephony Service and supplementary services; Stage 1". [9] 3GPP TS 24.229: "IP multimedia call control protocol based on Session Initiation Protocol (SIP) and Session Description Protocol (SDP); Stage 3". [10] 3GPP TS 24.173: "IMS Multimedia telephony communication service and supplementary services; Stage 3". [11] 3GPP TS 24.275: "Management Object (MO) for Basic Communication Part (BCP) of IMS Multimedia Telephony (MMTEL) communication service". [12] 3GPP TS 24.629: "Explicit Communication Transfer (ECT) using IP Multimedia (IM) Core Network (CN) subsystem; Protocol specification". [13] 3GPP TR 22.873: "Study on evolution of the IP Multimedia Subsystem (IMS) multimedia telephony service". [14] IETF RFC 8864: "Negotiation Data Channels Using the Session Description Protocol (SDP)". [15] 3GPP TS 24.147: "Conferencing using the IP Multimedia (IM) Core Network (CN) subsystem". [16] 3GPP TS 24.604: "Communication Diversion (CDIV) using IP Multimedia (IM) Core Network (CN) subsystem; Protocol specification". [17] 3GPP TS 24.615: "Communication Waiting (CW) using IP Multimedia (IM) Core Network (CN) subsystem; Protocol specification". [18] 3GPP TS 29.175: "IP Multimedia Subsystem; IP Multimedia Subsystem (IMS) Application Server (AS) Services; Stage 3". [19] 3GPP TS 29.176: "IP Multimedia Subsystems (IMS); Media Function (MF) Services; Stage 3". [20] 3GPP TS 32.260: "Telecommunication management; Charging management; IP Multimedia Subsystem (IMS) charging". [21] 3GPP TS 32.255: "Telecommunication management; Charging management; 5G data connectivity domain charging; stage 2". [22] 3GPP TS 24.647: "Advice Of Charge (AOC) using IP Multimedia (IM) Core Network (CN) subsystem". [23] 3GPP TS 24.239: "Flexible Alerting (FA) using IP Multimedia (IM) Core Network (CN) subsystem; Protocol specification". [24] 3GPP TS 24.174: "Support of multi-device and multi-identity in the IP Multimedia Subsystem (IMS); Stage3. [25] 3GPP TS 24.642: " Completion of Communications to Busy Subscriber (CCBS) and Completion of Communications by No Reply (CCNR) using IP Multimedia (IM) Core Network (CN) subsystem; Protocol specification". [26] 3GPP TS 24.183: "IP Multimedia Subsystem (IMS) Customized Ringing Signal (CRS); Protocol specification". [27] 3GPP TS 24.182: "IP Multimedia Subsystem (IMS) Customized Alerting Tones (CAT); Protocol specification". [28] 3GPP TS 24.628: "Common Basic Communication procedures using IP Multimedia (IM) Core Network (CN) subsystem; Protocol specification". [29] 3GPP TS 26.264: "IMS-based AR Real-Time Communication". [30] 3GPP TS 31.103: "Characteristics of the IP multimedia services identity module (ISIM) application". [31] 3GPP TS 31.102: "Characteristics of the Universal Subscriber Identity Module (USIM) application". [32] 3GPP TS 24.610:" Communication HOLD (HOLD) using IP Multimedia (IM) Core Network (CN) subsystem; Protocol specification". [33] 3GPP TS 22.011: "Service accessibility". [34] 3GPP TS 22.156: "Mobile Metaverse Services; Stage 1".
28a605aed849a8e4682a36cdb0077b63
24.186
3 Definitions of terms, symbols and abbreviations
28a605aed849a8e4682a36cdb0077b63
24.186
3.1 Terms
For the purposes of the present document, the terms given in 3GPP TR 21.905 [1] and the following apply. A term defined in the present document takes precedence over the definition of the same term, if any, in 3GPP TR 21.905 [1]. IMS data channel application: An IMS data channel application is an application using IMS data channel capability to provide IMS services. AR anchor: AR anchor is meant to identify a point in the user space to be used to anchoring a visual object. It is kind of metadata allowing accurate overlaying/rendering of text, graphics or video contents. For the purposes of the present document, the following terms and definitions given in 3GPP TS 23.228 [3] apply: Bootstrap data channel Application data channel IMS communication service IMS Communication Service Identifier (ICSI) Standalone IMS Data Channel Session The following terms and definitions given in 3GPP TS 26.264 [29] apply: AR media Split rendering The following terms and definitions given in 3GPP TS 24.229 [9] apply: 3GPP PS data off status The following terms and definitions given in 3GPP TS 22.011 [33] apply: 3GPP PS data off 3GPP PS data off exempt services The following terms and definitions given in 3GPP TS 22.156 [34] apply: Avatar
28a605aed849a8e4682a36cdb0077b63
24.186
3.2 Abbreviations
For the purposes of the present document, the abbreviations given in 3GPP TR 21.905 [1] and the following apply. An abbreviation defined in the present document takes precedence over the definition of the same abbreviation, if any, in 3GPP TR 21.905 [1]. ADC Application Data Channel AR Augmented Reality AOC Advice Of Charge AS Application Server BDC Bootstrap Data Channel CAT Customized Alerting Tones CB Communication Barring CCBS Completion of Communications to Busy Subscriber CCNL Completion of Communications on Not Logged-in CCNR Completion of Communications by No Reply CD Communication Deflection CDIV Communication DIVersion CFB Communication Forwarding Busy CFNL Communication Forwarding on Not Logged-in CFNR Communication Forwarding No Reply CFNRc Communication Forwarding on subscriber Not Reachable CFU Communication Forwarding Unconditional CN Core Network CONF Conference CRS Customized Ringing Signal CW Communication Waiting DC Data Channel DCSF Data Channel Signalling Function CUG Closed User Group eCNAM Enhanced Calling Name ECT Explicit Communication Transfer FA Flexible Alerting HOLD Communication Hold ICSI IMS Communication Service Identifier IM IP Multimedia IMS IP Multimedia Core Network Subsystem ISIM IM Subscriber Identity Module MF Media Function MiD Multi-iDentity MMTel Multimedia Telephony MuD Multi-Device MWI Message Waiting Indication OIP Originating Identification Presentation OIR Originating Identification Restriction OMA-DM Open Mobile Alliance Device Management P2A Person to Application P2A2P Person to Application and Application to Person P2P Person to Person SBI Service Based Interface PSI Public Service Identifier TIP Terminating Identification Presentation TIR Terminating Identification Restriction UE User Equipment UICC Universal Integrated Circuit Card URN Uniform Resource Name USIM Universal Subscriber Identity Module