Chapter 10

LTE URLLC performance

Abstract

This chapter shows the performance of LTE URLLC, comparing the technology to the existing 5G requirements put up by ITU (IMT-2020). Each 5G requirement applicable to URLLC is presented, together with the performance of LTE. Fulfilling the requirements involve both analytical calculations based on the existing design as well as system level and link level simulations. The performance evaluation includes user plane latency. Control plane latency and reliability evaluations. It is concluded that LTE can fulfill the 5G URLLC requirements.

Keywords

3GPP RAN; 4G; 5G; cMTC; Critical MTC; IMT-2020; LTE; mMTC; Third generation partnership project (3GPP); URLLC
In this chapter, the performance of the LTE URLLC design is presented. First the performance requirements on latency and reliability are reviewed. Following the requirements are the tools used to fulfill the requirements where both analytical calculations as well as system level and link level simulations have been used. Finally, link level simulations are provided showing how LTE fulfills the URLLC requirements of IMT-2020 to fulfill the 5G requirements.

10.1. Performance objectives

As mentioned in Chapter 2, the URLLC requirements is a subset of the requirements set up by ITU-R used for a technology to declare itself 5G compliant. This section explains the requirements and how they apply to the LTE technology.
Out of the full set of 5G requirements, the ones of interest for URLLC are user-plane latency, control plane latency and reliability.
To give a better understanding to the reader on the actual requirements, they are described in separate sections below. For the detailed requirement specification, the reader is referred to see Ref. [2].

10.1.1. User plane latency

The requirement on user-plane latency in Ref. [2] is defined as from when a source node sends a packet to where the receiving node receives it. The latency is defined between layer 2 and layer 3 in both nodes. For LTE this would be on top of the PDCP protocol. It is assumed that the device is in an active state, assuming no queuing delays.
A 1   ms latency bound is targeted for URLLC. Both directions (uplink and downlink) have the same requirement.

10.1.2. Control plane latency

The requirement on control plane latency is 20   ms defined from a battery efficient state to when the device is being able to continuously transmit data.

10.1.3. Reliability

The requirement on reliability is defined as the probability of successful transmission of a packet of a certain size, within a certain latency bound, at a given channel condition. In other words, we want the system to guarantee that at a certain SINR we can deliver a certain packet reliably within a maximum time.

10.2. Simulation framework

One could think that allowing 5% of the devices in the network to be out of coverage, while requiring the devices in coverage to experience a reliability of 99.999% is not a consistent reasoning. One should however remember that assuming full traffic load of the network (see Table 10.1), i.e. that all resources in all cells are always occupied, is not realistic. Using instead a more realistic network load, the SINR distribution would shift toward higher SINR (assuming the network is interference limited, which is typically the case), which reduces the ratio of devices falling outside of the target SINR depicted in Fig. 10.1.
The relevant system simulation assumptions to produce the curve in Fig. 10.1 are shown in Table 10.1.

Table 10.1

Urban macro URLLC – Simulation assumptions.
Parameter Setting
Carrier frequency for evaluation 700   MHz
Inter-site distance 500   m
Antennas, eNB Two directional sector antennas with +45°, 45° polarization
Antenna elements, device Two omni antennas with 0°, 90° polarization
Device deployment
80% outdoor, 20% indoor
Randomly and uniformly distributed over the area
Inter-site interference modeling Explicitly modeled
eNB noise figure 5   dB
Device noise figure 9   dB
Device antenna element gain 0   dBi
Traffic model Full buffer
Simulation bandwidth 20   MHz
Device density 10 UEs per Transmission Reception Point
Device antenna height 1.5   m
eNB antenna height 25   m
Sites 3-sector
eNB antenna down-tilt

Table 10.2

Recorded SINR.
Physical channel 5th percentile SINR
Subslot-PDSCH(1) -2.6   dB
Subslot-PUSCH(1) -1.7   dB
Subslot-SPUCCH -1.7   dB

NOTE 1: Same SINR value valid irrespective of transmission duration, i.e. also valid for e.g. slot operation.

The main reason that the downlink SINR is worse than the uplink SINR is the use of power control on the uplink, which is not used on the downlink.

10.3. Evaluation

10.3.1. User plane latency

For data transfer there are three different cases of interest:
  1. • Downlink data transmission
  2. • Uplink data transmission based on a configured grant, and,
  3. • Uplink data transmission after SR
In case downlink data, the network has full control to schedule the packet after it has been prepared for transmission and put in the transmit buffer (assuming no queuing delay). Hence, only the processing delay and the alignment of the data transmission time with the air interface need to be considered.
For the uplink, the situation is somewhat different since the network will not know when the device has data in its buffer. In case of SR-based uplink transmission (see Section 9.3.2.3), there will be additional delay until the uplink grant is received by the device, compared to having a configured grant available when the data has been prepared. This is since the SR needs to be received by the eNB, followed by an uplink grant to the device, before transmission of data can start.
For services with ultra-low latency requirements, it can be assumed that Semi-Persistent Scheduling (SPS), see Section 9.3.2.7.3, is needed to remove the extra delay of sending SR and receiving uplink grant. SPS can also be assumed to be used in case of periodic traffic when the traffic pattern is predictable. Furthermore, subslot operation is required to get down to the IMT-2020 URLLC target of 1   ms latency bound.
To get the full picture of the user-plane latency, the tables below show performance for legacy LTE (subframe based operation with n+4 processing timeline, see Section 9.2.5), short processing time (subframe based operation with n+3 processing timeline, see Section 9.2.5), slot-based and subslot-based transmission.
Two cases are looked at:
  1. • HARQ-based retransmissions: In this case, it is assumed that the data is retransmitted using HARQ. This will result in a spectrally efficient transmission since a packet is only retransmitted in case of error. However, the latency is negatively impacted by additional round-trip delays for each retransmission.
  2. • Blind repetitions: When blind repetitions are used (here assumed consecutive in time), the only additional delay comes from the transmission time of each repetition. Compared to HARQ-based retransmissions however, the spectral efficiency is negatively impacted since extra resources are used without knowing if less resources would have been enough for delivering the packet.
It can be noted that additional delay due to the TDD frame structure has not been considered. This is since the structure of TDD, where downlink and uplink transmissions are multiplexed in time, will inherently increase the delay. Adding that TDD is only defined for slot operation, and not subslot operation, will make the resulting delay relatively far from the IMT-2020 requirements, which is the main focus of this performance chapter.

Table 10.3

Delay components.
Type Comment
L1/L2 processing delay, T L 1 / L 2 image For the L1/L2 processing performed in the eNB, a processing delay equal to the configured transmission time interval is assumed when both transmitting and receiving a packet. The same is assumed in the device. A three symbol subslot is assumed. At the receiver side, it is assumed that data can be delivered to higher layers after L1/L2 processing but before ACK feedback is transmitted.
Alignment delay, T A l i g n image The alignment delay is the wait time required after being ready to transmit until transmission can start on the air interface.
UE/eNB processing delay, T P r o c image
The eNB delay between receiving a SR and transmitting an uplink grant, and between receiving downlink HARQ NACK and the next PDSCH retransmission is assumed to be the same as the timing of the device, i.e. the same as the time between PDSCH reception and downlink HARQ feedback transmission and uplink grant reception and PUSCH transmission.
The shortest processing timeline is assumed to be configured for subslot operation, i.e. n+4, see Section 9.2.5.
When applied, the processing delay includes the L1/L2 processing times.
It could be noted that the values of eNB processing are not specified but vendor specific. The processing delay device on the other hand needs to be strictly followed, see Section 9.2.5.

image

That is, the table and the following latency calculations assume FDD operation.
Fig. 10.2 shows the signal diagram for a downlink data transmission including the associated delays (ignoring propagation delay). It is considered that the packet can be delivered either in the first or the second attempt (depending on packet errors over the air).
The downlink data transmission delay can also be written as an equation, assuming k retransmissions.
T T o t = T L 1 / L 2 , D e v i c e + T L 1 / L 2 , e N B + T A l i g n + 2 k T P r o c + ( 1 + 2 k ) T T x
image (10.1)
For one retransmission, assuming regular timing of n+4 ( T P r o c = 2 T T I image ), and that T L 1 / L 2 = T A l i g n = T T x = 1 TTI image , the delay becomes:
T T o t = 4 + 8 k
image (10.2)
It can be noted that the processing timeline for the HARQ-ACK response, see Section 9.2.5, of the device is set by the specifications, while being implementation dependent for the eNB (in case of asynchronous HARQ). It is however assumed in Eqs. (10.1) and (10.2) that the same processing is assumed for both device and eNB. Furthermore, the L1/L2 processing time is different for different implementations in both network and device, and hence not defined by the specification.
The time to align with the frame structure will be given by the transmission granularity, i.e. how often a packet can be transmitted over the air. In case of LTE, this is based on a fixed structure using either subframe, slot or subslot granularity.
In all cases, it is assumed that the worst-case latency is used. This assumption has no impact on subframe based and slot-based operation due to the symmetry in the radio frame. For subslot operation however, the delay will vary over the subframe due to the irregular subslot structure, see Fig. 10.2. For subslot operation, the delays in Tables 10.4 and 10.5 are considering the largest delay of the five possible subslot starting numbers for the packet arrival. The three different possible PDCCH symbol configurations, one, two, or three, are considered. Using a single PDCCH symbol will decrease the overall latency since the alignment time will be reduced. However, using less symbols for the DL control will have a direct impact on the DL control capacity, and hence this trade-off need to be considered when deciding the size of the DL control region.

Table 10.4

User-plane latency – HARQ-based retransmissions.
Latency (ms) HARQ Rel-14 Rel-15 short processing time (n+3) Rel-15 slot Rel-15 subslot (1 PDCCH symbol) Rel-15 subslot (2 or 3 PDCCH symbol)
DL data 1st transmission 4.0 4.0 2.0 0.86 1.0
1st retransmission 12 10 6.0 2.1 2.4
2nd retransmission 20 16 10 3.4 4.0
3rd retransmission 28 22 14 4.9 5.4
UL data (SR) 1st transmission 12 10 6.0 2.1
1st retransmission 20 16 10 3.4
2nd retransmission 28 22 14 4.9
3rd retransmission 36 28 18 6.1
UL data (SPS) 1st transmission 4.0 4.0 2.0 0.86
1st retransmission 12 10 6.0 2.1
2nd retransmission 20 16 10 3.4
3rd retransmission 28 22 14 4.9

image

Table 10.5

User-plane latency – blind repetitions.
Latency (ms) Blind repetitions Rel-14
Rel-15
Short processing time (n+3)
Rel-15 slot Rel-15 subslot (1 PDCCH symbol) Rel-15 subslot (2 or 3 PDCCH symbol)
DL data 1st transmission 4.0 4.0 2.0 0.86 1.0
1 repetition 5.0 5.0 2.5 1.0 1.1
2 repetitions 6.0 6.0 3.0 1.1 1.3
3 repetitions 7.0 7.0 3.5 1.3 1.4
UL data (SR) 1st transmission 12 10 6.0 2.1
1 repetition 14 12 7.0 2.4
2 repetitions 16 14 8.0 2.9
3 repetitions 18 16 9.0 3.1
UL data (SPS) 1st transmission 4.0 4.0 2.0 0.86
1 repetition 5.0 5.0 2.5 1.0
2 repetitions 6.0 6.0 3.0 1.1
3 repetitions 7.0 7.0 3.5 1.3

image

Another influence on the latency due to the irregular subslot structure is the assumption on 1 TTI latency of L1/L2 processing. It should be noted that this is a simplified assumption and the actual timing need not be related to the over-the-air transmission duration. Still, in this evaluation, it is assumed that the L1/L2 processing assumes the value of 3 symbols, same as the longest subslot duration.
The user plane latency using HARQ-based retransmissions and using blind repetitions is shown in Tables 10.4 and 10.5 respectively using different number of retransmissions and repetitions.
If a 1   ms latency bound is to be respected (marked in bold in Tables 10.4 and 10.5), see Section 10.1.1, one can see that for:
  1. • Downlink data and uplink data using grant-based transmission, either a single transmission is required, or one blind repetition. Using HARQ-based retransmissions will however violate the latency bound. It can be noted that using a blind repetition is only possible in the DL in case 1 PDCCH symbol is configured.
  2. • Uplink SR-based transmission, no configuration fulfills the latency bound

10.3.2. Control plane latency

To summarize, instead of changing existing initial access procedures in how information is transmitted over the air, which could imply implementation cost and backwards compatibility issues in existing networks, a simple approach was taken by shortening the processing time at both the device and the network in the initial access procedure.
The resulting control plane latency is shown in Table 10.6.
As can be seen, the target of 20   ms control plane latency, see Section 10.1.2, is reached.

10.3.3. Reliability

The latency calculations carried out in Section 10.3.1 is an analytical exercise assuming different number of retransmissions and no queuing effects at the scheduler (as assumed in the requirements, see Section 10.1.1). Still, it determines the configurations under which a certain latency bound can be met.
To determine the associated reliability, simulations are needed that model the radio performance under the conditions given by Section 10.3.1.
In Section 10.3.1 it was concluded that for downlink and uplink transmissions, to fulfill the 1   ms target, there is no time for HARQ retransmissions. Instead, blind repetitions need to be used to lower the latency while maintaining a high level of reliability. Furthermore, uplink transmissions need to be carried out by pre-configured uplink grant using SPS (not based on scheduling requests, which increase the overall latency by one round-trip time). Both the use of blind repetitions (instead of HARQ retransmissions) and pre-configured uplink grants (instead of scheduling requests) will increase the usage of radio resources. This is a cost we need to be willing to pay to reach the ultra-reliable radio link performance at a low latency.

Table 10.6

Control plane latency.
Step Description Latency [ms]
Pre-release 15 Release 15
1 Transmission of RACH Preamble (uplink) 1
2 Preamble detection and processing in eNB 2
3 Transmission of RA response (downlink) 1
4 Device processing delay 5 4
5 Transmission of RRC Connection Resume Request (uplink) 1
6 Processing delay in eNB 4 3
7 Transmission of RRC Connection Resume (downlink) 1
8 Processing delay in device of RRC Connection Resume and grant reception 15 7
9 Transmission of RRC Connection Resume Complete and user plane data (uplink)
  Total delay 30 20

image

10.3.3.1. Reliability of physical channels

To understand the reliability performance, we need to look at how the physical channels are designed. This is where most of the description in Chapter 9 comes in. Although it should be noted that the IMT-2020 use case of a latency bound of 1   ms   at a reliability of 99.999% can be seen as an extreme case, this is what much of the LTE URLLC technology was designed and targeted for, and this is also the main use-case we evaluate in this chapter.
The possible physical channels involved in these evaluations are (S)PDCCH, SPUCCH, PDSCH and PUSCH, but as will be seen, not all channels need to be part of the evaluation.
Our task is now to determine an SINR where the overall block error rate, taking all channels involved in the data transfer into account, is below the target of 10 5. To fulfill the overall reliability requirement, this determined SINR should be equal to, or below, the values listed in Table 10.2.
10.3.3.1.1. Downlink
For the downlink, the channels involved in a data transmission are (S)PDCCH and PDSCH. If HARQ retransmissions were to be considered, also SPUCCH would have to be included (i.e. carrying the HARQ-ACK feedback triggering the retransmission). There is however no time within the latency bound for this (see Section 10.1.1) and hence the uplink control channel can be excluded. The reason for writing “(S)PDCCH” is that the PDSCH can be scheduled by either PDCCH or SPDCCH depending on which subslot the data is scheduled on (See Section 9.2.2 and 9.2.3.2.1). However, in this simulation campaign, it is assumed that the downlink data is earliest mapped to symbol index three, implying that SPDCCH is always used. The reason for this choice is to model a worst-case situation where PDCCH is not used for scheduling and less blind transmissions can be performed within a given latency bound (compared to the case where the data can be mapped to symbol index 1).
The channels involved are shown in Fig. 10.3 for a two-antenna port CRS configuration, one symbol CRS-based SPDCCH, and a mapping of the data REs earliest in symbol index three. The five possible starting positions for a blind repetition sequence of two or three transmissions is also shown. Whether two or three transmissions are shown depends on the latency budget of 1   ms (note that Table 10.5 shows only the worst-case latency over all possible starting positions, but in reality, it will vary depending on starting position). A single physical resource block in frequency is shown. Depending on how the mapping of control and data is done, some resource blocks will contain only data, some only control, or, as in the case of Fig. 10.3, both. As shown in Table 10.7, it has been assumed that 32 % of the data allocation overlaps with SPDCCH in the first symbol of the subslot. Hence, in roughly 1/3 of the resource blocks, the PDSCH is rate-matched around SPDCCH.

Table 10.7

Link level simulation assumptions – downlink.
Parameter Setting
Carrier frequency 700   MHz
Bandwidth 20   MHz (100 Resource Blocks)
Channel TDL-C, 363   ns, see Ref. [4]
Device speed 30   km/h
TTI length Subslot (2 or 3 OFDM symbols depending on where transmission is made)
Start of DL data Symbol index 3 (See Section 9.2.2)
Payload 32 bytes
MCS MCS-0 (occupying 55 resource blocks)
Resource allocation Resource allocation type 0 (See Section 9.3.2.9.1)
Transmission mode 2TX, 2RX, 1-layer TX diversity
Reference signal transmission CRS-based
Channel estimation Realistic
SPDCCH
1 symbol CRS-based SPDCCH (See Section 9.2.3.2)
AL8 (8 SCCE in total over bandwidth)
(about 32% of data allocation in first symbol has overlapping SPDCCH)
Transmissions
2 (Blind repetitions without HARQ)
The same redundancy version (RV0), see Section 9.3.2.7.1, used for both transmissions.
DCI size 40 bits payload + 16 bits CRC
The simulations have assumed the worst-case possible configuration where one of the two SPDCCH transmissions are hit by CRS (fourth or fifth configuration in Fig. 10.3). This can be considered the limiting case in terms of performance for the SPDCCH. Regarding PDSCH performance, one of the subslots will have a three-symbol duration resulting in more REs available for data transmission (this does however not improve the SPDCCH performance, which is here configured to one symbol duration).
The simulation assumptions used are shown in Table 10.7.
The simulation results from the above simulation assumptions are shown Fig. 10.4.
As can be seen, the target performance of a reliability of 10 5   at SINR of -2.6   dB (see Table 10.2) is reached assuming a power boosting of 1   dB for the SPDCCH resources. The power boosting is necessary due to the performance imbalance between the data and control channel (the decoding of the data channel is conditioned by the decoding of the control channel). The implication of boosting the power of the control is that the remaining power left for the other resource elements in that OFDM symbol will be more restricted.
10.3.3.1.2. Uplink

Table 10.8

Link level simulation assumptions – uplink.
Parameter Setting
MCS MCS-1 (occupying 85 resource blocks)
Transmission mode 1TX, 2RX
Transmissions
1, 2 or 3 (Automatic repetitions without HARQ)
The same redundancy version (RV0), see Section 9.3.2.7.1, used for all transmissions.
TTI length Subslot (2 OFDM symbols)
From Table 10.5, it can be seen that the 1   ms bound can be fulfilled using two transmissions of PUSCH. The possible starting positions and the following consecutive transmissions are illustrated in Fig. 10.5. Compared to the downlink (see Fig. 10.3), the number of repetitions does not vary depending on where the repetition window starts. Using three transmissions will violate the latency bound.
The worst-case performance would be expected if the two transmissions are performed using three consecutive 2-symbol subslots (second, third and fourth configuration in Fig. 10.5). This is since less resources are available for transmitting the data, and hence, this is what has been simulated (see Table 10.8).
The associated performance is shown in Fig. 10.6, see Ref. [1] for more details.
As can be seen, the PUSCH SINR target of -1.7   dB   at a reliability of 10 5 is achieved with the use of two transmissions. However, it can be noted that the allocation consumes 85 out of 100 resource block in the carrier. Furthermore, it is assumed that this is a pre-configured resource to the device and that the scheduling interval is each subslot. This causes severe restrictions to the network operation (essentially one device taking up almost all of the resource in the uplink).
Using a higher MCS would lower the frequency allocation required. However, although the same overall energy would be transmitted with a smaller allocation (using a higher power spectral density, PSD), some additional degradation would be expected from the increased code rate of the MCS by reducing the resource allocation, less frequency diversity, and possibly from using a higher order modulation.
The results will depend on if the network is interference or noise limited, the fraction of cMTC users etc, but it could be expected that with a slightly relaxed SNR requirement, the impact to the use of the network resources can be alleviated. However, with the requirement of -1.7 dB, the performance margin is very limited for the case of two transmissions, as can be seen in Fig. 10.6.