4683
TRANSCRIPT
-
8/7/2019 4683
1/14
HSDPA Throughput: DoTodays Devices Really
Perform?January 2007
Part Number 79-000781 Rev.0 0107
-
8/7/2019 4683
2/14
Spirent Communications, Inc.15200 Omega Drive
Rockville, MD 20850 USA
Spirent Communications541 Industrial Way West
Eatontown, NJ 07724 USAT: +1 732.544.8700
Email: [email protected]
Web: www.spirent.com
North AmericaT: +1 800.927.2660
Europe, Middle East, AfricaT: +33 1 6137.2250
Asia PacificT: +852 2511-3822
Copyright 2007 Spirent Communications, Inc. All Rights Reserved.
All of the company names and/or brand names and/or product names referred to in this
document, in particular, the name Spirent and its logo device, are either registeredtrademarks or trademarks of Spirent plc and its subsidiaries, pending registration in
accordance with relevant national laws. All other registered trademarks or trademarks are
the property of their respective owners.
The information contained in this document is subject to change without notice and doesnot represent a commitment on the part of Spirent Communications. The information in
this document is believed to be accurate and reliable; however, Spirent Communications
assumes no responsibility or liability for any errors or inaccuracies that may appear in the
document.
-
8/7/2019 4683
3/14
Spirent Communications White Paper 3
HSDPA Throughput: Do Todays
Devices Really Perform?
Contents
Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4
Why Conduct Data Throughput Performance Testing? . . . . . . . . . . . . . . . . . . . . . . . . . . . 5
Physical Layer Data Throughput Testing . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5
Application Layer Data Throughput Testing . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8
Conclusions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12
References . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13
-
8/7/2019 4683
4/14
HSDPA Throughput: Do Todays Devices Really Perform?Introduction
Spirent Communications White Paper 4
Introduction
Todays mobile applications and services require much more from a mobile device than
the ability to make a voice call. From location-based services to video streaming, these
applications generally have one requirement in common: higher bandwidth. To meet theneed for increased bandwidth, most 3G network operators are turning to the High-Speed
Downlink Packet Access (HSDPA) enhancements from Release 5 of the 3GPPs
WCDMA / UMTS network specifications.
These enhancements are based on the concept of an optimized shared data-pipe.
Performance improvements result from the use of adaptive modulation and rate control
techniques that depend heavily on UE interaction and feedback. HSDPA allows more
efficient use of network resources to maximize total aggregate throughput. These
carefully calculated resource trade-offs are intended to enable optimum performance on
not just one but all UEs in the network.
More than ever before, thorough testing of these devices is critical to ensuring they will
not adversely impact either the network or the subscribers quality of experience (QoE)of new applications and services. The 3GPP has developed test standards to establish a
UEs conformance to the requirements of the specifications. However, while
conformance testing may establish a common minimum performance baseline, it does not
provide true real-world characterization metrics that can be used to further optimize boththe network and ultimately the end-users QoE.
The key performance metric for HSDPA is data throughput, which is highly dependent
on the RF multipath and interference environment experienced by a UE. While the 3GPPconformance test specifications include a test that addresses data throughput scenarios,
this test uses a Fixed Reference Channel (FRC) with specified fading and noise. Under
these conditions, the performance of the UE is measured only at the physical layer, and
the test produces a single data point for each set of conditions. While this data point mayenable a simple pass/fail diagnosis, it indicates very little about how well the UE is
performing from an end-users perspective.
To truly understand UE performance, additional data throughput testing needs to be
conducted, including testing at the application layer. This paper uses a Spirent APEX
UMTS Data Performance test system (high-level architecture shown in Figure 1) to
analyze the data throughput performance of multiple Category 6 (CAT 6) UEs in the
presence of varying RF fading and noise conditions at both the physical and application
layers. Comprehensive analysis of the performance test results is carried out to uncover
important differences that can significantly impact the network and subscriber QoE.
Figure 1 Test System Architecture
HSDPA NetworkEmulator Channel
Emulator
UE Under Test
Tx
Rx
Rx/Tx
Ior Io
-
8/7/2019 4683
5/14
HSDPA Throughput: Do Todays Devices Really Perform?Why Conduct Data Throughput Performance Testing?
Spirent Communications White Paper 5
Why Conduct Data Throughput Performance Testing?
When measured in the presence of multipath fading and noise, the data throughput rate is
a fundamental indication of a UEs HSDPA performance. Since throughput performance
is also a key enabler for the successful launch of new data-centric applications, networkoperators and UE manufacturers need to conduct this testing to fully maximize their
return on investment. Failure to find a UE performance flaw prior to launch could end up
costing all involved parties dearly.
Given the highly dynamic configuration of an HSPDA channel, the UE will constantly
compete for resources on that shared channel. The data rate will thus continuously vary
as a function of the channel quality seen by the UE and be will also affected by network
configuration. Given the adaptive nature of the technology, operators and UE
manufacturers should conduct testing under a wide range of varying configurations to
ensure the UE, network and application all work together and perform as designed.
While metrics such as data throughput qualify performance, it is QoE, the subscribers
perception of performance, that ultimately determines acceptance of the device or
application.
3GPP Release 5 specifies twelve categories for HSDPA UEs. Category 12 (CAT 12)
UEs were used on most early HSDPA deployments. CAT 12 UEs support up to 1.8
Mbps using a Quadrature Phase Shift Keying (QPSK) modulation scheme.Comprehensive testing of these devices reveals some performance differences between
UEs. However, the less complex modulation scheme combined with the number of
available HS codes typically offers additional coding protection, resulting in an adequate
performance margin.
Observed performance differences between UEs are bound to increase as potential
throughput rates increase, especially when combined with the more complex 16-
Quadrature Amplitude Modulation (16-QAM) scheme supported by Category 6 (CAT 6)UEs. Factors such as the Channel Quality Indicator (CQI) algorithm and its estimate of
channel quality will ultimately result in a modulation trade-off analysis (QPSK vs. 16-
QAM) which will have a direct impact on the throughput capabilities of the UE. Apoorly implemented CQI algorithm on one model of UE can have a significant adverse
impact on the efficient allocation of network resources, which will affect all other UEs on
the network. As CAT 6 UEs appear commercially in ever-greater numbers, it is the
performance of this UE category that is currently of greatest interest.
Physical Layer Data Throughput Testing
CAT 6 UEs support data rates up to 3.6 Mbps. There are several contributing factors to
this higher data rate capability including a larger available Transport Block Set (TBS)
size, an increase in the number of soft bits and, most significantly, the ability to support
16-QAM modulation. While these improvements may provide real value for the end-
user, they further increase the complexity in developing valid test methodologies.
Testing at the physical layer is a good starting point for initial performance evaluation of
a UE or application. Figure 2 shows the results of physical layer downlink data
throughput tests conducted on four CAT 6 HSDPA UEs under two sets of static channel
conditions.
-
8/7/2019 4683
6/14
HSDPA Throughput: Do Todays Devices Really Perform?Physical Layer Data Throughput Testing
Spirent Communications White Paper 6
0
500
1000
1500
2000
2500
3000
3500
4000
HS-PDSCH = -3 dB / Ior = -60 dBm HS-PDSCH = -6 dB / Ior = -80 dBm
Static Channel Conditions
Configuration = HS-PDSCH = -3 or -6 dB, Ior = -60 or -80 dBm (No Noise)
TTI = 1, # of H-ARQ = 6, CQI = Fixed at 22, TBS = Set per TR 25.214
DataThroughput(Kbps)
UE A
UE B
UE C
UE D
Figure 2 - 6 UE Physical Layer Data Throughput Results (Under StaticConditions)
The data throughput results shown in Figure 2 were obtained under both favorable and
weakened static channel conditions, with the HSDPA channel configured to allow for
maximum throughput (TTI = 1, number of HARQ process = 6, and CQI = Fixed at 22).
The results indicate that under favorable conditions (HS-PDSCH = -3 dB / Ior = -60
dBm) all four UEs perform at close to the maximum expected data throughput rate of 3.6
Mbps. However, under weakened channel conditions (HS-PDSCH = -6 dB / Ior = -80
dBm) there is a significant reduction (> 30% on all of the UEs) in data throughput.
To understand the reasons for this, it is important to remember that the UE must rely on
16-QAM modulation and demodulation to obtain the maximum data throughput rate. 16-
QAM is more complex to successfully decode than the QPSK modulation used by CAT
12 UEs. CAT 6 UE performance is also influenced by code power reduction; these UEs
have little available margin before the median CQI drops below the desired maximum
value of 22.
Significant variation is apparent between UEs under weakened static conditions. While
physical data throughput may provide an indication of UE performance, it does not
isolate the root causes of poor performance. For this, additional metrics such as the
Median CQI or the Mac-HS Statistics are helpful. Figure 3 shows a histogram of a
typical CQI distribution (Median = 16) under a standard fading profile, recorded during astandard data throughput test. A histogram that does not approximate a bell-shaped curve
could indicate an issue with the UEs CQI algorithm or with its receiver.
Reduction in Maximum Throughput @ weakenedchannel conditions (>30% drop on all CAT 6 UEs tested)
Throughput Differences @ weakened channelconditions (>10% when UE B is compared to UE D)
-
8/7/2019 4683
7/14
HSDPA Throughput Testing: Do Todays Devices Really Perform?Physical Layer Data Throughput Testing
Spirent Communications White Paper 7
0.00%
2.00%
4.00%
6.00%
8.00%
10.00%
12.00%
14.00%
0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30
CQI Data (Median 16)
Figure 3 - CAT 6 UE: CQI Histogram Example - Under a Standard Fading Profile(Median CQI = 16)
Mac-HS Statistics can also play a critical role in determining the root cause of poor
performance. Table 1 contains the Mac-HS Statistics for both static tests shown in Figure
2.
Table 1 - Mac-HS Statistics (ACK / NACK / Stat DTX Performance)for CAT 6 UEs
While all the UEs experienced a drop in the physical layer throughput rate underweakened channel conditions, further investigation shows that all the UEs did not report
the same channel quality. Although two UEs obtained a Median CQI of 21, a higher CQI
does not guarantee a higher throughput rate. For example, while UE B reports a higherchannel quality, it is experiencing a higher percentage of NACKs. This results in a
throughput rate more than 10% lower than that of UE D, which reports a Median CQI of
only 20. The conclusion: UE B is requesting network resources that it does not use very
effectively, resulting in an inefficient resource allocation scenario that can adverselyimpact the entire network.
UE Under
Test
Throughput
(Kbps)
Median
CQI
ACK % NACK % Stat DTX
%
Static Conditions HS-PDSCH = -3 dB / Ior = -60 dBm (favorable conditions)
A 3583.76 24 99.99 0.1 0
B 3575.08 24 99.75 0.25 0
C 3583.99 24 100 0 0D 3583.99 24 100 0 0
Static Conditions HS-PDSCH = -6 dB / Ior = -80 dBm (weakened conditions)
A 2402.26 20 67.03 32.97 0
B 2223.86 21 62.05 37.95 0
C 2459.44 21 68.6 31.4 0
D 2507.98 20 69.97 30.03 0
-
8/7/2019 4683
8/14
HSDPA Throughput: Do Todays Devices Really Perform?Application Layer Data Throughput Testing
Spirent Communications White Paper 8
It is important to note that physical layer testing is conducted using a fixed CQI value,
ignoring the actual CQI reported by the UE. This allows an apples-to-apples
comparison across various UEs by removing the CQI algorithm from consideration.
However, since for CAT 6 UEs the CQI value determines the modulation scheme, great
care must be taken when choosing the fixed value used in testing to ensure its potential
impact on the results is clearly understood. For example, under certain channel
conditions, the fixed CQI value used in a test could result in 16-QAM modulation used
for the duration of the test when the actual CQI value reported by the UE indicates the
use of QPSK.
Naturally, this scenario differs from a live network, which would listen and respond to
the CQI reports supplied by the UE and dynamically change the resources allocated to
that UE based on the reported channel conditions. For CAT 6 UEs, this would include a
switch from 16-QAM to QPSK modulation if the reported CQI drops below 15.
Performance on a live network can also be affected by differences in CQI algorithm
implementations between UE manufacturers. A fixed CQI test methodology does not
take into account the dynamic resource allocation nor the contribution that CQI algorithm
variations can have on UE performance in a live network.
It should be clear by now that conducting data throughout testing only at the physical
layer with a standards-based fixed CQI test methodology is insufficient and may lead to
an inaccurate picture of actual UE performance on a live network. To truly determine the
performance of a new application or UE, application layer data throughput tests should beconducted. The result of these tests is also referred to as goodput, the measure of the data
throughput actually available to a mobile application.
Application Layer Data Throughput Testing
Testing at the application layer is more representative of the performance of a UE from
an end-users perspective. However, application layer testing also introduces new testmethodology concerns. For example, what should the Radio Link Control (RLC)
Window size be, what transfer protocol does the application require and how will thatimpact throughput performance?
Figure 4 shows the results of application layer downlink data throughput tests conducted
using a File Transfer Protocol (FTP) for the same four CAT 6 UEs under favorable
channel conditions with four standard fade models applied: ITU Pedestrian A Speed
3km/h (PA3), ITU Pedestrian B Speed 3km/h (PB3), ITU Vehicular A Speed 30km/h
(VA30) and ITU Vehicular A Speed 120km/h (VA120).
-
8/7/2019 4683
9/14
HSDPA Throughput: Do Todays Devices Really Perform?Application Layer Data Throughput Testing
Spirent Communications White Paper 9
0
200
400
600
800
1000
1200
1400
1600
1800
2000
PA3 PB3 VA30 VA120
Fade M odel (Per 34.121 - Appendix D)
Configuration = HS-PDSCH = -3 dB, Ior = -60 dBm, Ior/Ioc = 10 dB
TTI = 1, # of H-ARQ = 6, CQI = Based on UE Reports, TBS = Set per TR 25.214
DataThroughput(K
bps)
UE A
UE B
UE C
UE D
Figure 4 - CAT 6 UE Application Layer (FTP) Data Throughput(Under Favorable Channel Conditions)
As anticipated, testing the UEs at the application layer reveals significant performancedifferences between them. The first observation is that actual downlink data throughput
rates are significantly less than the 3.6 Mbps obtained earlier at the physical layer. The
UEs averaged a 57% drop in data throughput for the PA3 fade model compared with the
maximum throughput allowed at the physical layer, and greater than a 70% drop (on
average) for all of the other fade models (PB3, VA30, and VA120) tested.
It is also evident that strong performance with one fade model is not necessarily
replicated with other fade models (even at the same modeled speed). While UE Cperformed quite well relative to UEs A and B with the PA3 fade model, with all other
fade models tested (including the PB3 fade model which models the same velocity as
PA3) its performance was comparable to that of UEs A and B.
Although actual throughput rates are lower than those anticipated from the physical layer
test results, it is nonetheless clear that UE D has a significant performance advantage
compared with all other UEs tested. Under certain test conditions (ex. PB3 and VA30
fade models), UE D is capable of successfully transferring at least 30% more data than
any of the others under the same channel conditions.
Given the large differences in observed performance, additional statistical data or logs
from these test runs were analyzed. Table 2 contains the Mac-HS Statistics obtained forthe VA30 Fading Profile.
Clear Performance DifferencesUE D shows a 30% increase in data throughputrates when compared to the other UEs tested.
-
8/7/2019 4683
10/14
HSDPA Throughput: Do Todays Devices Really Perform?Application Layer Data Throughput Testing
Spirent Communications White Paper 10
UE Under
Test
Throughput
(Kbps)
Median
CQI
ACK % NACK % Stat DTX
%
A 878.68 14 72.36 27.64 0
B 930.51 15 75.50 24.49 0
C 932.5 15 65.4 34.6 0
D 1344.05 17 71.09 28.91 0
Table 2 - Mac-HS Statistics for CAT 6 UEs under the VA30 Fading Profile
A clear performance difference is apparent in terms of Median CQI even though ACK
and NACK percentages are approximately the same. Under these channel conditions, UE
D outperforms the others in Median CQI level and associated throughput. Results
suggest UE D has a better-performing receiver or a more sophisticated CQI algorithm, or
both.
As previously indicated, available code power and noise level play a significant role in
the data throughput capabilities of a UE or application. Figure 5 shows the results of
application layer downlink data throughput tests using FTP for the four UEs under the
PB3 fade model, with varying code power levels. Figure 6 shows the results of testing
the same four UEs under the PB3 fade model with varying Carrier to Noise (C/N) Ratio.
0
200
400
600
800
1000
1200
1400
1600
HS-PDSCH = -3 dB HS-PDSCH = -4 dB HS-PDSCH = -5 dB HS-PDSCH = -6 dB
Code Power Leve l (HS-PDSCH Level) - For Fade Model PB3 (Per 34.121 Appendix D)
Configuration = HS-PDSCH = -3 to - 6 dB, Ior = -80 dBm, Ior/Ioc = 10 dB
TTI = 1, # of H-ARQ = 6, CQI = Based on UE Reports, TBS = Set per TR 25.214
Da
taThroughput(Kbps)
UE A
UE B
UE C
UE D
Figure 5 - CAT 6 UE Application Layer (FTP) Data Throughput Results atVarying Code Power Levels
Reducing the code power on the UEs clearly results in a reduction in data throughput. A
1 dB drop in available code power yields, on average, a 16% drop in data throughput,while a 2 dB drop in available code power results in an average 29% drop. These results
reinforce the idea that allocated code power plays a significant role in throughput
capability of a UE (or application). They also indicate once again that UE D has a
noticeable performance advantage over the other UEs.
Code Power Reduction yields Throughput Degradation1 dB = 16% (on average) & 2 dB = 29% (on average)
-
8/7/2019 4683
11/14
HSDPA Throughput: Do Todays Devices Really Perform?Application Layer Data Throughput Testing
Spirent Communications White Paper 11
0
100
200
300
400
500
600
700
800
900
1000
10 dB 5 dB 2 dB
Carrier to Noise Ratio (Ior/Ioc) - For Fade Model PB3 (Per 34.121 - Appendix D)
Configuration = HS-PDSCH = - 6 dB, Ior = -80 dBm, Ior/Ioc = 10, 5, & 2 dB
TTI = 1, # of H-ARQ = 6, CQI = Based on UE Reports, TBS = Set per TR 25.214
DataThroughput(K
bps)
UE A
UE B
UE C
UE D
Figure 6 - CAT 6 UE Application Layer (FTP) Data Throughput Results at
Varying Carrier to Noise Ratios
Reducing the C/N Ratio also forces a decrease in overall data throughput rates across all
of the UEs. It is interesting to note the C/N Ratio reduction also impacts UE Ds data
throughput performance advantage over the other UEs. However, even with an 8 dB
reduction in the C/N Ratio, UE D still outperforms the other UEs tested by an average of
28%.
As mentioned earlier, another area of potential impact on data throughput performance is
the data protocol used for the file transfer. All examples so far have used FTP protocol,
typically employed for applications that require acknowledgement of received packets.Some applications, in particular those that are delay sensitive (such as video streaming),
cannot support the overhead required for FTP.
These applications typically employ User Datagram Protocol (UDP). Since UDP is not
subject to acknowledgements, data throughput performance using UDP is, in principle,
higher than with FTP. Figure 7 shows the results of application layer downlink data
throughput tests conducted using UDP, under favorable channel conditions and with the
same standard fade models applied.
C/N Ratio Reduction yields Throughput Degradation(UE D still out performs other UEs by 28% even at 2 dB C/N)
-
8/7/2019 4683
12/14
-
8/7/2019 4683
13/14
HSDPA Throughput: Do Todays Devices Really Perform?References
Spirent Communications White Paper 13
Rather than rely on a simple pass/fail result from a conformance test, network operators
and UE manufacturers should determine the headroom for an application from data
throughput rates obtained during testing. From these UE test results, it is clear there is a
large degradation in maximum application layer data throughput rates under
representative real-world conditions.
The results also indicate that variables such as code power levels, C/N ratios and channel
configurations can all significantly impact the throughput potential of a UE or
application. Other factors also shown to impact the performance include TBS Block
Size, the number of HARQ processes allocated and the number of HS Codes assigned.
For example, the greater the number of codes allocated, the more coding protection
provided to a particular UE. This may result in an improvement in the UEs ability to
absorb errors caused by adverse channel conditions.
One of the more important conclusions stems from the clearly differentiated performance
of UE D. The fact that this UE uses the exact same chipset as UE A begs the question of
whether conformance testing alone adequately evaluates an HSDPA UEs performance.
Should UE manufacturers and application developers rely on the assumption thattechnology providers have adequately tested the performance of their chipsets under a
wide range of representative conditions? Given the complexity of HSDPA devices and
networks and the applications they are expected to support, it appears that the final UE
implementation itself should be more thoroughly tested under the constantly changing
conditions and resource allocations that are part of the real HSDPA environment.
While a thorough test process for an HSDPA device may start with the standards (i.e.
3GPP), it really needs to be extended beyond these minimum conformance-based
requirements to reflect the dynamic WCDMA/HSDPA environment which is built andimplemented based on trade-offs. Failure to extend the scope of testing beyond
conformance can lead to inefficient network resource allocation and performance issues
thus negatively affecting both the network and end-user QoE.
References
1. 3GPP Technical Specification 34.121-1, V7.3.0 (2006-12) 3rd GenerationPartnership Project, Technical Specification Group Radio Access Network, User
Equipment (UE) conformance specification, Radio transmission and reception
(FDD); Part 1: Conformance Specification (Release 7).
2. 3GPP Technical Specification 25.214, V7.3.0 (2006-12) 3rd Generation PartnershipProject, Technical Specification Group Radio Access Network, physical layer
procedures (FDD) (Release 7).
-
8/7/2019 4683
14/14