Ryan Malloy 4ed92efd69 refactor: move spec references out of published site
Cospas-Sarsat specification summaries moved to reference/ for internal
use only. Links updated to point to official cospas-sarsat.int site.

The extracted images remain in public/ for use in other pages.
2026-02-13 05:03:09 -07:00

2861 lines
102 KiB
Markdown
Raw Blame History

This file contains ambiguous Unicode characters

This file contains Unicode characters that might be confused with other characters. If you think that this is intentional, you can safely ignore this warning. Use the Escape button to reveal them.

---
title: "R021: Cospas-Sarsat Meosar System Demonstration And Evaluation Phase I Report C"
description: "Official Cospas-Sarsat R-series document R021"
sidebar:
badge:
text: "R"
variant: "note"
# Extended Cospas-Sarsat metadata
documentId: "R021"
series: "R"
seriesName: "Reports"
documentType: "report"
isLatest: true
issue: 2
revision: 1
documentDate: "December 2015"
originalTitle: "Cospas-Sarsat Meosar System"
---
> **📋 Document Information**
>
> **Series:** R-Series (Reports)
> **Version:** Issue 2 - Revision 1
> **Date:** December 2015
> **Source:** [Cospas-Sarsat Official Documents](https://www.cospas-sarsat.int/en/documents-pro/system-documents)
---
COSPAS-SARSAT MEOSAR SYSTEM
DEMONSTRATION AND EVALUATION
PHASE I REPORT
C/S R.021
Issue 1
COSPAS-SARSAT MEOSAR SYSTEM
DEMONSTRATION AND EVALUATION
PHASE I REPORT
HISTORY
Issue
Revision
Date
Comments
Approved by the Cospas-Sarsat Council (CSC-55)
TABLE OF CONTENTS
Page
History..................................................................................................................................................... i
Table of Contents ................................................................................................................................... ii
List of Figures ....................................................................................................................................... iv
List of Tables ........................................................................................................................................ iv
Document Summary .............................................................................................................................. v
1.
BACKGROUND ................................................................................................................1-6
1.1
MEOSAR System Description ...................................................................................1-6
1.2
The Cospas-Sarsat MEOSAR Demonstration and Evaluation Plan ...........................1-6
1.3
The Phase I of the Cospas-Sarsat MEOSAR D&E ....................................................1-7
1.3.1
Objective of the D&E Phase I ........................................................................1-7
1.3.2
Report of the D&E Phase I ............................................................................1-7
2.
CONDUCT OF PHASE I AND MEOSAR SYSTEM CONFIGURATION .................2-1
2.1
Tests Conducted During the Phase I ..........................................................................2-1
2.2
Participants in the D&E Phase I .................................................................................2-3
2.3
Configuration of the D&E Phase I .............................................................................2-5
2.3.1
Experimental Space Segment ........................................................................2-5
2.3.2
Experimental Ground Segment ......................................................................2-5
2.3.3
Beacon Simulators and Test Beacons ............................................................2-6
2.4
Test Coordination .......................................................................................................2-7
2.5
Operational Issues Encountered during the Phase I Testing ......................................2-7
2.6
Data Collection ...........................................................................................................2-8
3.
RESULTS OF THE TECHNICAL TESTS AND DISCUSSION ..................................3-1
3.1
Test T-1 (Processing Threshold and System Margin) ................................................3-1
3.1.1
Analysis .........................................................................................................3-1
3.1.2
Interpretation ..................................................................................................3-1
3.2
Test T-2 (Impact of Interference) ...............................................................................3-4
3.2.1
Analysis .........................................................................................................3-4
3.2.2
Interpretation ..................................................................................................3-4
3.3
Test T-3 (Valid/Complete Message Acquisition).....................................................3-10
3.3.1
Analysis .......................................................................................................3-10
3.3.2
Interpretation ................................................................................................3-10
3.4
Test T-4 (Independent 2D Location Capability) ......................................................3-14
3.4.1
Analysis .......................................................................................................3-14
3.4.2
Interpretation ................................................................................................3-14
3.5
Test T-5 (Independent 2D Location Capability for Operational Beacons) ..............3-20
3.5.1
Analysis .......................................................................................................3-20
3.5.2
Interpretation ................................................................................................3-21
3.6
Test T-6 (MEOSAR System Capacity) ....................................................................3-29
3.6.1
Analysis .......................................................................................................3-29
3.6.2
Interpretation ................................................................................................3-30
3.7
Test T-7 (Networked MEOLUT Advantage) ...........................................................3-33
3.8
Test T-8 (Combined MEO/GEO Operation Performance (Optional)) .....................3-33
4.
CONCLUSIONS AND RECOMMENDATIONS ...........................................................4-1
4.1
Conclusion ..................................................................................................................4-1
4.1.1
Test T-1 (Processing Threshold and System Margin) ...................................4-1
4.1.2
Test T-2 (Impact of Interference) ..................................................................4-1
4.1.3
Test T-3 (MEOLUT Valid/Complete Message Acquisition) ........................4-1
4.1.4
Test T-4 (Independent Location Capability)..................................................4-2
4.1.5
Test T-5 (Independent 2D Location Capability for Operational Beacons) ....4-3
4.1.6
Test T-6 (MEOSAR System Capacity) .........................................................4-5
4.1.7
Test T-7 (Networked MEOLUT Advantage) ................................................4-5
4.1.8
Test T-8 (Combined MEO/GEO Operation Performance (Optional)) ..........4-5
4.2
Recommendations for the Conduct of Subsequent D&E Phases ...............................4-5
4.2.1
Test T-1 (Processing Threshold and System Margin) ...................................4-5
4.2.2
Test T-2 (Impact of Interference) ..................................................................4-6
4.2.3
Test T-3 (MEOLUT Valid/Complete Message Acquisition) ........................4-6
4.2.4
Test T-4 (Independent Location Capability)..................................................4-6
4.2.5
Test T-5 (Independent 2D Location Capability for Operational Beacons) ....4-6
4.2.6
Test T-6 (MEOSAR System Capacity) .........................................................4-6
4.2.7
Test T-7 (Networked MEOLUT Advantage) ................................................4-6
4.2.8
Test T-8 (Combined MEO/GEO Operation Performance (Optional)) ..........4-7
4.3
Recommendations for the Implementation of the MEOSAR System ........................4-7
LIST OF ANNEXES
ANNEX A
DETAILED LOG OF PHASE I TESTS
LIST OF FIGURES
Figure 1: The MEOSAR System Concept ..........................................................................................1-6
Figure 2: Phase I Test Planning (as Run) ............................................................................................2-3
Figure 3: MEOLUTs Involved in Phase I Testing with 3,000 km Radius Circles .............................2-6
Figure 4: Beacon Simulators Used in the MEOSAR D&E Phase I ....................................................2-7
Figure 5: Spectrum Graphic Sample of T-1 Run 2 .............................................................................3-5
Figure 6: Test T-4 Activations in Red, Some Other Transmission Overlapping Boxed in Green .....3-6
Figure 7: Interference for Test T-5 as Seen by Galileo .......................................................................3-7
Figure 8: Average Throughput for Each Antenna for Test T-5 Observed by the Ottawa MEOLUT .3-8
Figure 9: Test T-6 Beacon Bursts and Overlapping Interference Seen by DASS Satellites. ..............3-9
Figure 10: Location of the 33 Operational Beacons Deployed for Test T-5 ....................................3-21
LIST OF TABLES
Table 1: List of Technical Tests, Test Coordinators and Test Reports ...............................................2-2
Table 2: Participation in MEOSAR D&E Phase I Tests .....................................................................2-4
Table 3: List of Experimental MEOSAR Satellites Used During the MEOSAR D&E Phase I .........2-5
Table 4: MEOLUTs Participating in MEOSAR D&E Phase I Tests..................................................2-6
DOCUMENT SUMMARY
This document provides the report of Phase I of the Cospas-Sarsat MEOSAR Demonstration and
Evaluation (D&E), tests which were conducted from February 2013 to March 2014.
Section 1 provides background on the MEOSAR system and reference material.
Section 2 reviews the planning and conduct of the tests, noting the list of participants, MEOSAR space
and ground assets configuration used during the tests and information of interest on the coordination
of the tests.
Section 3 details, for each D&E test the key results and interpretations as provided by each test
participant that contributed to this Report (Canada, France, Russia, Turkey and USA). The underlying
sub-sections were provided under the responsibility of these administrations and, therefore, were not
reviewed nor commonly agreed by the Correspondence Working Group on the Phase I Report.
Section 4 provides, for each test, the conclusions and recommendations agreed by the Correspondence
Working Group on the Phase I Report, as well as general recommendation regarding the
implementation of the MEOSAR system.
1-6
1.
BACKGROUND
1.1
MEOSAR System Description
Figure 1 provides a graphical summary of the MEOSAR concept. This picture shows the relay of
beacon signals, via multiple satellites, to the MEOLUT. Beacon data is processed by the MEOLUT
that derive the beacon locations, and passed onto the MCC, which in turn notifies the RCC.
Figure 1: The MEOSAR System Concept
1.2
The Cospas-Sarsat MEOSAR Demonstration and Evaluation Plan
The Cospas-Sarsat Council has directed that a demonstration and evaluation (D&E) be performed to
confirm the expected capabilities and benefits of a satellite system in medium-altitude Earth orbit
(MEO) that uses onboard repeater instruments to relay distress alert signals emanating from 406 MHz
distress radiobeacons. The CSC further directed that the D&E should establish the technical and
operational performance characteristics of the MEOSAR system.
The framework for the D&E of the MEOSAR system is provided in document C/S R.018 “Cospas-
Sarsat Demonstration and Evaluation Plan for the 406 MHz MEOSAR System”. In particular,
documents provide guidelines for:
1-7
conducting the D&E of the MEOSAR system in a standard manner among the participants,
collecting a set of results from individual participants, using compatible formats, that can be
consolidated into a final report for review by Cospas-Sarsat participants and other interested
parties,
analysing and translating the results into a set of recommendations for a decision by the
Cospas-Sarsat Council to enter the Initial Operational Capability Phase.
Additional resources regarding the MEOSAR system (e.g, space segment information) are available in
document C/S R.012 “Cospas-Sarsat 406 MHz MEOSAR Implementation Plan”.
CSC-49 agreed to divide the MEOSAR D&E Phase into three phases:
Phase I, during which the participants perform only technical tests,
Phase II, during which the participants perform technical and operational tests,
Phase III, during which the participants replicate the tests of the Phases I and II, when satellites
with L-band downlinks are widely available.
1.3
The Phase I of the Cospas-Sarsat MEOSAR D&E
1.3.1
Objective of the D&E Phase I
In MEOSAR D&E Phase I, participants performed only technical tests (see the detailed definition in
document C/S R.018) to characterise the technical performance of the MEOSAR system. Due to the
limited space segment available, some tests had to be coordinated and the processing be tuned
accordingly, in particular for location accuracy tests.
1.3.2
Report of the D&E Phase I
The D&E Phase I report was produced by a Correspondence Working Group with the support of the
Secretariat, based on:
the reports on the conduct of the tests provided by the test coordinators (see Table 1),
contributions from the test participants, which provided their interpretation of the test results
(see section 3),
agreement among the participants on common conclusions and recommendation for the D&E
Phase I (see section 4).
- END OF SECTION 1 -
2-1
2.
CONDUCT OF PHASE I AND MEOSAR SYSTEM CONFIGURATION
2.1
Tests Conducted During the Phase I
Table 1 provides the list of technical tests planned for the Phase I, their completeness status, the
participants undertaking the role of test coordinator and the reference to the test reports written by the
test coordinators. The detailed conduct of each test can be found in the test coordinators reports.
Planning of the D&E Phase I.
The initial planning of the MEOSAR D&E tests, as proposed at Annex L of document C/S R.018,
showed an estimated Phase I duration of 20 weeks assuming a beginning in January 2013, leading to
completion in May 2013. Due to unexpected delays in the January 2013 commencement, and to the
unavailability of some participants during the EWG-1/2013 meeting, this planning was re-evaluated at
the end of February 2013 and the end of Phase I was delayed until the end of June 2013.
However the test campaign did not progress as anticipated and several additional delays were
encountered during the tests performed by the participants.
At the JC-27 Meeting in June 2013, the D&E participants agreed upon the tests to be conducted during
the summer 2013 period, which included the second run of tests T-1 (Processing Threshold and System
Margin) and T-4 (Independent 2D Location Capability).
At the TG-1/2013 Meeting in September 2013, the D&E participants agreed upon the tests remaining
for the conclusion of the Phase I with a test T-5 conducted in November 2013 and a test T-6 conducted
in December 2013 and beginning of March 2014.
The test participants also agreed to not conduct the runs of test T-7 (Networked MEOLUT Advantage)
because the network configuration was not available and optional test T-8 (Combined MEO/GEO
Operation Performance), due to the time constraints. Consequently, the technical tests of the MEOSAR
D&E Phase I can be considered as completed (see Annex A that provides the detailed log of the Phase I
tests as conducted).
As agreed by the Council (see section 5.2.14 of the CSC-51 report), Phase I will be concluded once
the Phase I report is reviewed by the Joint Committee at its JC-28 session in June 2014. The production
of the Phase I report, for which a draft version was anticipated to be provided at the TG-2/2014
Meeting, was delayed due to the late provision of the Test Coordinators reports.
Table 1 and Figure 2 provide the schedule of the tests conducted during the Phase I testing as run.
2-2
Test
Definition
Run
Status
Test
Coordinator
Test Report Reference
T-1
Processing Threshold and System
Margin
Completed in February - March 2013
USA
Dated 28 February 2014
Completed in July August 2013
T-2
Impact of Interference
Records available only for tests T-1
Run 2, T-4 Run 2, T-5 and T-6 Run 2
Canada
No report available
T-3
Valid/Complete Message Acquisition
France
SAR-RE-DEMEO-783-CNES Iss 1 Rev 3,
dated 14 May 2013
Replaced by a test run at lower
transmission rate conducted in April
T-4
Independent 2D Location Capability
Completed in April 2013
USA
Dated 28 February 2014
Completed in June August 2013
T-5
Independent 2D Location Capability
for Operational Beacons
-
Completed in November 2013
Turkey
T-5 Run1 Beacon Deployment Report -
consolidated v2 - 26.02.2014
T-5 Run1 Test Coordinator Report v1 - 6.05.2014
T-6
MEOSAR System Capacity
Completed in May 2013
France
SAR-RE-DEMEO-788-CNES Iss 1 Rev 1,
dated 23 April 2014
Completed in December 2013 and
T-7
Networked MEOLUT Advantage
Cancelled (network not ready)
Canada
Not applicable
Cancelled (network not ready)
T-8
Combined
MEO/GEO
Operation
Performance
-
Cancelled (optional test)
Turkey
Not applicable
Table 1: List of Technical Tests, Test Coordinators and Test Reports
2-3
Figure 2: Phase I Test Planning (as Run)
2.2
Participants in the D&E Phase I
Table 2 provides the participants in each run of test, which provided at least raw data as per Table J.1
of document C/S R.018 or a test report. Some participants did not provide test results and/or test report.
Table 2 also provides the test during which spectrum of the 406 MHz band was recorded. For test T-5,
the participation in test T-5 is identified either in supplying test beacons or in involving MEOLUTs.
2-4
Test
Definition
Run
T-2: Impact of
Interference
(by Canada)
Australia
Brazil
Canada
France
Russia
Turkey
UK
USA
Hawaii
Maryland
T-1
Processing Threshold and System
Margin
X
(1- channel)
X
X
(1-channel)
X
X
X
X
X
X
X
X
X
X
Under
repair
Under
upgrade
X
X
X
T-3
Valid/Complete Message
Acquisition
X
(1-channel)
X
X
(2-channel)
X
X
X
X
X
X
Modified
script
Under Upgrade
X
Under
Upgrade
X
X
X
X
T-4
Independent 2D Location
Capability
Under Upgrade
Under
Upgrade
X
X
(2-channel)
X
X
X
X
X
Under
upgrade
X
X
T-5
Independent 2D
Location
Capability for
Operational
Beacons
Test
beacon
X
X
X
X
X
X
X
MEOLUT
X
X
X
X
X
X
T-6
MEOSAR System Capacity
X
X
X
X
X
X
(2 channels)
X
X
X (Toulouse run
only)
X
X
X
X
X
Table 2: Participation in MEOSAR D&E Phase I Tests
2-5
2.3
Configuration of the D&E Phase I
2.3.1
Experimental Space Segment
Table 3 provides the list of experimental MEOSAR satellites available for testing during MEOSAR
D&E Phase I.
MEOSAR
Constellation
Satellite
(C/S ID)
Satellite availability status for Phase I or launch date
DASS (GPS-II)
Available
Available
Available
Available
Available
Available
Available
Available
Available
Available
Available subsequent to launch on 4 October 2012
Available subsequent to launch d on 15 May 2013
Galileo
Available for testing from March 2013
Available for testing from March 2013
Glonass
Available with limitations (no ephemeris data available)
Table 3: List of Experimental MEOSAR Satellites Used
During the MEOSAR D&E Phase I
2.3.2
Experimental Ground Segment
The ground segment equipment in place for the Phase I of the MEOSAR D&E consisted of
experimental MEOLUTs located in Brazil, Canada, France, Russia, Turkey, the UK and the USA.
Table 4 provides the MEOLUTs available for testing, their number of antennas, their software
configuration and their availability (note that some participants may have experienced unexpected
down periods for some channels, thus limiting their participation in particular tests, see the Test
Coordinators reports for more detail).
2-6
Country/
Organisation
Location
Number of
Antennas
Configuration
Available for
D&E testing since
Brazil
Brasilia
HGT MEOLUT 600
LP v 1.6 / SP v 1.4 / FP v 1.4
Canada
Ottawa
HGT MEOLUT 600
LP v1.5 / SP v1.3 / FP v1.3
France
Toulouse
HGT MEOLUT 600
LP v1.6 / SP v1.4 / FP v1.4
Russia
Moscow
Not provided
Turkey
Ankara
HGT MEOLUT600
LP v1.6 / SP v1.4 / FP v1.4
UK
Kinloss
HGT MEOLUT 600 (S-band only)
LP v1.7 / SP v1.5 / FP v1.5
USA
Hawaii
Not provided
Maryland
Not provided
Table 4: MEOLUTs Participating in MEOSAR D&E Phase I Tests
Figure 3: MEOLUTs Involved in Phase I Testing with 3,000 km Radius Circles
2.3.3
Beacon Simulators and Test Beacons
Three beacon simulators were used during the Phase I testing, located in Hawaii and Maryland, USA
and Toulouse, France. After each test, the beacon log files were provided by each administration
providing beacon simulators.
2-7
Figure 4: Beacon Simulators Used in the MEOSAR D&E Phase I
(MEOSAR Visibility Circles at Five Degree Elevation)
2.4
Test Coordination
A smooth progression of the D&E planning and tests has been observed thanks to the active
participation of the Test Coordinators and Test Participants. No formal D&E test had to be postponed
or re-scheduled due to a coordination issue.
However, a dry-run test encountered some difficulty because of the work simultaneously being
performed by the Galileo Programme. In addition to this particular case, other tests had been planned
by Test Participants at times similar to those of SAR/Galileo commissioning tests, requiring an active
coordination between France and EC/ESA in order to avoid the simultaneous transmission of beacon
signals.
2.5
Operational Issues Encountered during the Phase I Testing
The Cospas-Sarsat operational community was informed of upcoming D&E tests by a SIT605 message
prior to each test. So far, no major operational issue was encountered during the MESOAR D&E tests,
and at no time did it become necessary to terminate beacon simulator transmissions.
In order to prevent any issues related to the unexpected behaviour of the SARP-3 processors aboard
Sarsat 11, Sarsat 12 and Sarsat 13, as detailed in Attachment 2 of document JC-26/Inf.14, each beacon
test script was verified and cleared in coordination with France. Nonetheless, during the T-3 test run
2-8
using the Hawaii beacon simulator, the memory of two SARP-3 instruments recorded a limited number
of erroneous messages sent by the simulator, and some LEOLUTs produced a few erroneous location
solutions. These solutions were filtered out by the MCCs as the messages were invalid and clearly
related to the on-going tests.
Further investigation confirmed that erroneous messages were recorded in the SARP-3 memory and
that LEOLUT processing should not have produced any location solution given that each erroneous
message was unique in the SARP-3 memory. The number of erroneous messages recorded remains
acceptable in comparison to the size of the SARP-3 memory and no specific mitigation action had to
be undertaken for the particular test T-3. However, additional investigations were conducted and,
taking the most cautious approach, this matter will be carefully monitored during upcoming future tests
to prevent any filling of the SARP-3 memory with erroneous messages. As an additional risk mitigation
measure, not beacon message transmissions were conducted during test T-6 (System Capacity), for
which the number of transmitted bursts is larger than for the other technical tests.
2.6
Data Collection
During the tests, the participants collected the following data:
beacon log data to collect the beacon IDs transmitted (if applicable),
MEOLUT raw data as per csv format defined in Table J.1 of document C/S R.018,
MEOLUT location data as per csv format defined in Table J.2 of document C/S R.018,
MEOLUT pass schedule data as per csv format defined in Table J.3 of document C/S R.018.
All the data provided by the test participants were saved on the MEOSAR D&E FTP server.
- END OF SECTION 2 -
3-1
3.
RESULTS OF THE TECHNICAL TESTS AND DISCUSSION
The following sections provide, for each test:
references to the test participants reports presenting the results of the MEOSAR D&E tests
conducted during the Phase I testing,
a summary of the interpretation of the test analyses, as provided by each administration.
3.1
Test T-1 (Processing Threshold and System Margin)
3.1.1
Analysis
The following test reports were provided by the participants:
Administration
Test report reference
France
“C/S D&E T-1 Test Report Processing Threshold And System Margin”
SAR-RE-DEMEO-762-CNES, v1.3
Russia
Test report available on FTP server
Turkey
TRMEO T-1 Report, dated 10 April 2013
USA
USA Hawaii MEOLUT Report Revision 1.0, dated 15 April 2013
Technical Test T1 Run 1 - USA Maryland Test Report, dated 21 May 2013
Technical Test T1 Run 02 - USA Maryland Test Report, dated 12 September 2013
3.1.2
Interpretation
3.1.2.1 Canada
Data was collected for the first run of test T-1 for the Maryland simulator run only and from one of the
original antennas used during the MEOSAR Proof of Concept. Average C/No calculated at the ground
station was 33.7 dB.Hz. The lowest signal power level received by the old antenna was a 29 dBm burst.
For the second Maryland run, a new four-antenna set was available. The average C/No increased to
35.7 dB.Hz (calculated over all signals received) and all four antennas received signals as lows as
2 dBm.
The results showed that the new antennas performed much better than the older ones. However, for the
second run, while detection rates did reach the 70% threshold for all antennas individually for power
levels above 35 dBm, below this power level, detection rates varied greatly, and these rates were much
less than the 70% threshold. Overall detection rate was much lower than 70% for each antenna.
However, all antennas did detect signals as low as 22 dBm. This fact was a positive outcome
considering the interference present in the entire band, as noted in test T-2 spectrum graphic
collections. Interference impacts the automatic gain control of the repeaters and therefore can impair
3-2
the detection of weaker signals. As well, weaker signals fade out often for certain passes due to the
elevation angle of the beacon simulator from the repeaters view point, which would be one more reason
for the variability and reduction of the detection threshold for signals below 35 dBm.
Canada cannot provide further interpretations at this time, but would consider analysing its data in the
future to help with comparisons with and insights when comparing with the results of Phase II testing.
3.1.2.2 France
Best results were obtained during run 2 with the Maryland transmission. At 37 dBm of transmitted
power, up to 71% of single satellite channels have a valid message throughput higher than 70% (target
value).
It was noted that the results in term of message throughput are significantly dependent on the elevation
angle between the beacon and the satellite due to the beacon antenna pattern.
An issue was observed with the French MEOLUT regarding confirmed messages detection, whose
throughput is too low compared to valid message throughput. Further work is required to investigate
and resolve the issue.
Also to be investigated, low performances with the Toulouse transmission were observed, for which
only 25% of single satellite channels have a valid message throughput higher than 70%.
3.1.2.3 Russia
In rare occasions, the throughput for beacons emitting at 37 dBm was higher than 80%, with lower
throughput values for 35 dBm and further down to 22 dBm. The possible reasons for this behaviour
were investigated and it was found out that the receiver message integration capability was left
unchanged after test T-3 and was operating in such a way that the period deviation of the emitting
beacon was restrained to 50±0.1 seconds to discard other beacon messages emitting at the same
frequency. Unfortunately, the period of the messages in was 48 seconds and that hindered the receiver
from getting bursts integrated. It was believed that this might led to the poor results, especially in the
lower power range.
The results obtained at MEOLUT for run 1 have shown that the throughput performance might have
been better. Considering the burst integration issue at Russian MEOLUT, a retest may be required to
correctly and comprehensively assess the throughput performance.
Unfortunately, the Moscow MEOLUT was unable to participate in run 2 of test T-1 due to an antenna
damage.
3.1.2.4 Turkey
For the Toulouse transmission of 13-14 March 2013, the 2-channel Ankara MEOLUT had a maximum
valid (resp. complete) system message throughput of 60% (resp. 55%), not reaching the 70% threshold
aimed by test T-1 in the transmitted 22 dBm - 37 dBm power range. It took the addition of the GEO
channel for the 3-channel Ankara MEOLUT (2 MEOSAR + 1 GEOSAR channels) to reach a valid
3-3
(resp. complete) system message throughput of 70% at 31 dBm (resp. 35 dBm) with corresponding
average C/No values of 37 dB-Hz (resp. 39 dB-Hz).
For the Maryland transmission of 20-21 February 2013, even the GEO-complemented 3-channel
Ankara MEOLUT had a maximum valid (resp. complete) message throughput of 57% (resp. 49%),
and for the Hawaii transmission of 7-8 March 2013, a maximum valid (resp. complete) message
throughput of 46% (resp. 38%), not reaching the 70% threshold aimed by T-1 in the transmitted
22 dBm - 37 dBm power range.
The results seemed to indicate that the processing threshold would be improved (i.e., get lower) and
consequently the system margin would increase by the addition of more MEOSAR channels as well
as the proximity of the transmission source (i.e., the beacon simulator). Clearly, a 2-channel MEOLUT
was not adequate to have a system throughput above the targeted 70% threshold.
3.1.2.5 USA-Hawaii
Following the analysis of data collected during the test T-1, run 1 the following key observations were
noticed:
Erroneous messages were received by different LEOLUTs tracking SARP-3-equipped
satellites at various geographical locations during the testing period, and this behaviour was
investigated further both by the USA and France,
Single channel throughput statistics for satellites being tracked in the northwest quadrant were
lower by around 10%,
Antenna \#2 had significantly lower overall detections than others,
Antenna \#6 got no detections at all,
The two agreed upon methodologies for obtaining the average C/No values appeared to
generate very similar numbers.
As a conclusion to this run, the data was too inconsistent to obtain a reliable estimate of system margin,
either for the single satellite or the multiple antenna scenarios. In particular, the required detection
threshold of 70% detection rate was generally not achieved for the single channel case.
Following the analysis of data collected during the test T-1, run 2 the following key observations were
noticed:
The rerun of technical test T-1 provided one of the first opportunities at the Hawaii MEOLUT
to successfully compare the performance of DASS satellites to Galileo satellites, which proved
to be a very productive exercise as the detection rates or the Galileo satellites were notably
better,
Overall the Hawaii MEOLUT results for run 2 showed a marked improvement in comparison
to run 1 (even when removing antennas \#2 and \#6 from the run 1 analysis),
As seen before for the Hawaii MEOLUT, detection statistics for satellites being tracked in the
northwest quadrant were lower by around 10% (although not as pronounced with the Galileo
data collected).
Compared to run1, there was considerable improvement in performance, both in consistency and in
general. However, there was still not enough consistency to arrive at any single value for system
margin either from throughput of single satellite/antenna pairings or multiple antenna cases.
3-4
3.1.2.6 USA-Maryland
Single channel results demonstrate variability caused by the effect the beacon antenna pattern, ground
blockage, and differences in the amount of earth noise in the uplink, have on the results. As the
elevation angle to the satellite exceeds 50 degrees, the transmitted EIRP of the beacon decreases and
the probability of reception, therefore, decreases. Similarly, when the footprint of the satellite receive
antenna is over different geographical areas, the earth noise increases and the C/No of the received
beacon messages decreases and the probability of reception, therefore, decreases. More examples of
these effects were contained in document EWG-1/2008/3/9.
The multi antenna results showed that, for four antennas, the message throughput ranges from 97% to
88% for values of beacon output power down to 32 dBm. It is useful to note that the throughput is 96%
at the minimum allowed beacon output power of 35 dBm as specified in document C/S T.001 (see the
table below).
Power
(dBm)
Number of
Transmitted
Bursts
Detection rate of
Valid Message
received
Detection rate of
Valid Complete
received
0.98
0.97
0.98
0.97
0.96
0.96
0.91
0.90
0.91
0.90
0.89
0.88
3.2
Test T-2 (Impact of Interference)
3.2.1
Analysis
The following test reports were provided by the participants:
Administration
Test report reference
Canada
Documents EWG-1/2014/2/2 and TG-1/2013/Inf. 15
3.2.2
Interpretation
3.2.2.1 Canada
All spectrum graphics recorded during the MEOSAR D&E tests are available on the FTP D&E server.
Spectrum graphics can be viewed either using the GIMP tool provided to all participants for download
within the T-2 test folder, or by using any common picture viewer. A very small sample of certain time
periods to enforce an observation and help with the interpretation of the results is provided below, but
to fully appreciate and explore the spectrum graphics it is strongly recommended to view the sample
here in their native image file using GIMP.
3-5
Spectrums recorded during test T-1
Data was collected for test T-1 for the first run from one of the original antennas from the MEOSAR
Proof of Concept. For the second run, all four antennas were available. The results showed that the
new antennas performed better than the older ones. However, while detection rates did reach the
70% throughput threshold for all antennas individually for power levels above 35 dBm, below that,
detection rates varied greatly, dropping below threshold quickly.
Looking at the spectrum graphics produced for test T-1 (see Figure 5), immediate and directly
overlapping interference was not present as often enough, so as to attribute the low detection rate to
interference alone. Looking at the RF levels received, in many instances the ground station LUT
processors should have been able to decode the beacon simulator signals, but no burst was detected,
and hence no beacon message was decoded. Clearly, there were occasions were the actual signal was
received at the ground station and downconverted properly to the LUT signal processors, but for
whatever reason, the signal was not decoded. This lead to the observation that the signal processors
are burdened with dealing with the interference and the multitude of signals that could be considered
as possible beacon bursts. This observation was important to note, as anything from more reference
beacons to more interference or increased number of users just above or below the distress band would
negatively impact detection rates, with the current processing techniques used and capabilities of,
heritage/conventional processors.
Figure 5: Spectrum Graphic Sample of T-1 Run 2
(as presented in document TG-1/2013/Inf.15)
Spectrums recorded during test T-3
Data was collected for test T-3 for the first run from two of the original antennas used during the
MEOSAR Proof of Concept. For the second run only one LUT channel was available and, as such,
3-6
only spectrum graphics for test T-2 was collected. As such, Canada can only comment that detection
rate issues would have been impacted by the interference present at the times of the tests.
Spectrums recorded during test T-4
Data was collected from all beacon simulators active in test T-4, from all four MEOLUT antennas.
Canada did not have time to analyse the complete data sets. From partial analysis of its data, Canada
noticed interference and fade outs (possible scintillation events) when comparing to the spectrum
graphics of test T-2 that would have impacted some of the test burst transmissions, and which could
explain various missed bursts seen by other LUTs as well.
Clearly, as seen in Figure 6, there were occasions were the actual signal was received at the ground
station and downconverted properly to the LUT signal processors, but for whatever reason, the signal
was not decoded, just as in test T-1. This lead to the observation that the signal processors are burdened
with dealing with the interference and the multitude of signals that could be considered possible beacon
bursts.
Figure 6: Test T-4 Activations in Red, Some Other Transmission Overlapping Boxed in Green
Spectrums recorded during test T-5
As seen in Figure 7 (this figure was presented and detailed in document EWG-1/2014/2/2), channels
used by test beacons operating from 406.037 to 406.040 MHz were in the clear for most of the time
for all three test dates, although some interference were observed around 406.045 MHz. Beacons in
the 406.020 to 406.030 MHz channels suffered from interferers most of the time, especially around
406.024 MHz. This would mean that detection rates of the 406.028 MHz beacons would present lower
detection or throughput rates than those in the 406.037 and 406.040 MHz channels.
3-7
Figure 7: Interference for Test T-5 as Seen by Galileo
Note the interferers throughout the band
Ottawas LUT detection rates from test T-5 are provided in Figure 8 as a point of comparison. Antenna
15 had an intermittently active signal processing card, and thus had lower detection rates at certain
times. Note that it was not an antenna or RF path activity issue, just a hardware problem that would
not have been seen and easily determined without the help of the spectrum graphics.
3-8
Figure 8: Average Throughput for Each Antenna for Test T-5
Observed by the Ottawa MEOLUT.
At this stage, the results do not provide clear guidance for the specifications and parameters regarding
the exact coverage areas in which the locations can meet the accuracy requirements within 5 km, 95%
of the time, as proposed currently in documents C/S R.012 and C/S R.018.
Country
Beacon ID
Antenna 14
Antenna 15
Antenna 16
Antenna 17
1 In 4 Antennas 4 In 4 Antennas Average Antenna
Australia
BEFC0 00000 000E3
0.0
0.0
0.2
0.4
0.6
0.0
0.2
Austratlia Average
0.0
0.0
0.2
0.4
0.6
0.0
0.2
279C6 32662 FFBFF
33.2
56.0
38.8
49.1
87.2
4.7
44.3
279C6 360D0 FFBFF
37.0
2.3
24.7
28.9
64.9
0.5
23.3
279C6 7A164 FFBFF
76.6
32.5
51.8
70.1
91.9
12.3
57.7
279C7 4DCFE FFBFF
0.0
0.0
0.0
0.1
0.0
0.0
0.02
279C7 53BAE FFBFF
69.8
62.8
51.8
51.8
92.2
12.6
59.0
279C7 53CA0 FFBFF
45.8
63.2
52.3
57.8
89.2
11.7
54.8
Canada Average
Within Canada
56.4
53.6
48.7
57.2
90.1
10.3
54.0
1C7C0 84B5A FFBFF
21.6
30.9
16.7
21.4
65.8
0.0
22.7
1C7C0 84B5C FFBFF
40.9
4.9
26.2
28.1
70.7
0.0
25.0
9C7E4 3316C 0028C
2.6
7.4
1.5
9.9
19.6
0.0
5.3
9C7E4 333E4 0028C
7.9
2.4
0.6
2.9
12.9
0.0
3.5
9C7E4 C37BA 73590
0.2
0.2
0.1
0.1
0.3
0.0
0.1
9C7E4 C37BA 735D0
7.9
9.7
2.6
7.6
24.9
0.0
6.9
9C7FE 28D29 90CA0
8.9
15.8
4.7
27.5
48.8
0.0
14.2
9C7FE 28D41 52900
3.7
0.2
0.0
0.8
4.2
0.0
1.2
France Average
11.7
9.0
6.5
12.3
30.9
0.0
9.9
21FD0 F9502 FFBFF
5.8
3.7
2.1
4.4
13.8
0.0
4.0
21FD0 F9514 FFBFF
27.4
20.5
24.1
12.7
64.8
0.1
21.2
21FD0 F9520 FFBFF
27.4
3.2
10.6
24.7
56.4
0.0
16.5
21FD0 F9532 FFBFF
9.9
3.9
2.8
7.5
22.0
0.0
6.0
21FD0 F953A FFBFF
11.3
7.1
7.5
3.2
27.2
0.0
7.3
Turkey Average
16.4
7.7
9.4
10.5
36.8
0.0
11.0
1D1C0 007D2 FFBFF
18.5
16.0
8.2
8.4
44.7
0.0
12.8
1D1C0 007D4 FFBFF
32.6
25.8
16.4
15.6
66.7
0.0
22.6
1D1C0 007D6 FFBFF
26.5
21.2
13.9
11.3
57.9
0.0
18.2
1D1D6 10002 FFBFF
28.4
2.6
17.7
18.5
49.8
0.1
16.8
1D1D6 28018 FFBFF
42.7
29.8
21.6
23.8
78.2
0.6
29.5
1D1E0 6C6BF 81FE0
51.0
8.3
35.3
37.6
81.5
0.1
33.1
1D1E1 21ABF 81FE0
34.3
5.9
18.9
26.2
63.1
0.1
21.3
1D1E4 F1BBF 81FE0
47.6
3.7
25.6
33.9
76.1
0.1
27.7
UK Average
35.2
14.2
19.7
21.9
64.8
0.1
22.7
2DDC6 7A0B4 FFBFF
64.0
46.8
49.2
35.3
87.7
5.7
48.8
2DDC6 7A0C4 FFBFF
29.7
32.2
26.3
30.4
73.6
0.5
29.7
2DDC7 52E20 FFBFF
69.3
41.7
50.9
57.3
89.8
9.8
54.8
2DDC7 52E2A FFBFF
2.7
10.2
14.1
21.6
44.9
0.0
12.1
ADDE4 11528 00330
14.5
10.9
9.4
7.4
21.1
1.0
10.6
USA Average
36.0
28.4
30.0
30.4
63.4
3.4
31.2
USA
Canada
France
Turkey
UK
3-9
Spectrums recorded during test T-6
Canada collected data from the run 2 of test T-6 with the Toulouse beacon simulator in December 2013
(see Figure 9). From its analysis the detection rates did not significantly drop with the increase of the
number of transmitted bursts. The 406.070 MHz interferer directly overlapped one to three frequency
channels (runs) in which the test T-6 Run 2 tests were executed for the Toulouse beacon. The interferer
was present more than half the time of the test beacon simulator sequence.
Figure 9: Test T-6 Beacon Bursts and Overlapping Interference Seen by DASS Satellites.
Note fading of bursts in purple box.
As well, the 406.060 MHz interferer was also present in many portions of the test run. This would
mean that a reduction anywhere from 10% to 30% in the number of received bursts would be expected,
due the interference seen during the runs of test T-6. Despite the interference, overall the detection
rates and location accuracies were slightly better than those seen in test T-5.
3-10
Overall summary of test T-2
From the analysis and interpretation provided above for the technical tests, it was noted that persistent
interference was seen by all MEOSAR satellites around 406.070 MHz, as well as intermittent
interference around 406.060 MHz, when in view of north and eastern quadrature of the globe. This
interference was not typically seen in the upper half of the 406 MHz band when satellites do not see
Western Africa and Eurasia. Similar interference and noise behaviour was observed in all spectrum
graphics for all tests.
Canada underlines the importance of spectrum monitoring and recommends to continue monitoring
the 406 MHz band spectrum and analysing the detection rate per channel, or probability of detection
per channel, for all technical tests. This parameter had been assumed to be very high (i.e., 85%), but
was not met yet by the Ottawa ground station. Current tests showed average detection rate to be no
better than 70% per single satellite channel in distances less than 1,000 km away from the MEOLUT,
and to be around 40% for beacons up to 5,000 km away. This would mean that a 6-channel MEOLUT
would be needed to achieve single burst locations better than 95% of the time for distances up to 5,000
km away from the beacon.
3.3
Test T-3 (Valid/Complete Message Acquisition)
3.3.1
Analysis
The following test reports were provided by the participants:
Administration
Test report reference
France
“C/S
D&E
T-3
Test
Report
MEOLUT
Valid/Complete
Message
Acquisition”
SAR-RE-DEMEO-750-CNES v1.1
Russia
Test report is available on FTP D&E server
Turkey
TRMEO T-3 Report v3 dated 2 June 2013
USA
Technical Test T-3 Run1 USA Hawaii MEOLUT Report, Revision 1.0 dated 3 April 2013
Technical Test T3 Run1 - USA Maryland Test Report, dated 15 April 2013
Technical Test T3 ReRun - USA Maryland Test Report, dated 31 May 2013
3.3.2
Interpretation
3.3.2.1 Canada
Data was collected for test T-3 for the first run from two of the original antennas from the MEOSAR
Proof of Concept. For the second run only one MEOLUT channel was available and the signal
processing at the MEOLUT was intermittent due to a hardware issue. As such, only spectrum graphics
for test T-2 were collected. As such Canada can only comment that detection rates could have been
impacted by the interference present at the times of the tests, which would have increased the time to
obtain a valid message.
3-11
3.3.2.2 France
At 37 dBm, for the Toulouse transmission with the modified script (i.e., at a lower beacon transmission
rate), the probability to obtain a valid message after one transmitted burst with the French MEOLUT
was 90% and reached 100% after seven transmitted bursts, and the transfer time was about 10 s.
When the distance of the beacon increased to 6,000 km, the probability to receive a valid message after
seven transmitted bursts decreased to 90% or 80%.
France noted that the Ankara and French MEOLUTs demonstrated improved results with this run of
test T-3 using a modified script, thus showing some limitations on the receiving capacity. This
observation remained to be further investigated.
3.3.2.3 Russia
No consistent results were obtained due to MEOLUT inability to properly integrate beacon bursts
emitted as per T-3 script. However, in order to prepare for run 1 of test T-3, the receiver firmware was
upgraded to retain signal integration capability and ensure compliance with test scenario that required
receiver to process 50 beacons transmitting at the same frequency with a one second transmission rate.
In the view of the above there might have been occasions when beacon message from various beacons
were mixed in one integration sequence. This occurrences being unlikely have not been investigated,
however a certain “cushion” was foreseen to eliminate the probability of taking them into account
while processing the results.
Generally, the results have shown the deterioration of the valid/complete message detection
performance as the emissions locations moved from Toulouse towards Maryland and further to Hawaii,
with the decrease of C/No ratio accordingly. Nonetheless, taking into account the aggressive scenarios
of the tests the results proved the capability of the Russian MEOLUT to process the beacon messages
relayed through MEOSAR satellites with certain level of quality that may be sufficient to meet
expected performance requirements.
3.3.2.4 Turkey
For the run using the Toulouse beacon simulator with the modified script (at a lower beacon
transmission rate), the valid message average detection probability of the Ankara MEOLUT increased
from 88% (resp. 86%) for 1 burst to 99% (resp. 98%) for 7 bursts for a beacon transmission power of
37 dBm (resp. 33 dBm). The Ankara MEOLUT complete message average detection probability
increased from 87% (resp. 84%) for 1 burst to 99% (resp. 97%) for 7 bursts for a beacon transmission
power of 37 dBm (resp. 33 dBm). Valid (resp. complete) message transfer times of 8 seconds (resp.
10 seconds) were obtained at 37 dBm, and valid (resp. complete) message transfer times of 13 seconds
(resp. 18 seconds) were obtained at 33 dBm.
For the run using the Maryland beacon simulator with the modified script, the valid message average
detection probability of the Ankara MEOLUT increased from 68% (resp. 51%) for 1 burst to 99%
(resp. 78%) for 7 bursts for a beacon transmission power of 37 dBm (resp. 33 dBm). The complete
message average detection probability of the Ankara MEOLUT increased from 63% (resp. 40%) for
1 burst to 98% (resp. 78%) for 7 bursts for a beacon transmission power of 37 dBm (resp. 33 dBm).
Valid (resp. complete) message transfer times of 22 seconds (resp. 26 seconds) were obtained at
3-12
37 dBm, and valid (resp. complete) message transfer times of 77 seconds (resp. 92 seconds) were
obtained at 33 dBm.
The results seemed to indicate that average detection probabilities improved, as expected, with the
number of transmitted bursts as well as with the beacon transmission power. Similarly, message
transfer times improved (i.e., got shorter) with the beacon transmission power. On the other hand, for
the test run using the Toulouse beacon simulator with the modified script, 1-burst detection
probabilities well exceeded 70% for both 37 dBm and 33 dBm, in contrast with the lower system
throughputs obtained in test T-1 at the highest beacon transmission power of 37 dBm a point that
needs to be further investigated.
3.3.2.5 USA-Hawaii
Following the analysis of data collected during the test T-3, run 1 the following key observations:
Antennas \#2 and \#6 had significantly lower performance (for detection rate, in particular),
Degradation in the performance to the north and northwest of the MEOLUT was noted and
was more pronounced at the lower transmit power (33 dBm vs. 37 dBm)
In comparing the performance of the Hawaii MEOLUT between the first and the additional
modified run, the detection percentage on 3 of the 4 good antennas showed improvement, but
overall the differences were not significant.
Due to unexplained inconsistencies in the data, there was no definitive conclusion with regard to the
test objectives as a result of the testing. However, it is clear that MEOSAR in general is a very good
system for beacon message detection (both valid and complete), with minimal delays in transfer times
as well.
3.3.2.6 USA-Maryland
Results for 37 dBm are somewhat limited by satellite visibility at the time of the test but are still
excellent.
3-13
Single Burst Valid Message Probability (37 dBm) and
Number of Satellites for the Maryland MEOLUT with the Maryland Simulator
Seven Burst Valid Message Probability (37 dBm) and
Number of Satellites for the Maryland MEOLUT with the Maryland Simulator
Gaps in performance can also be caused by ground based interference degrading satellite performance.
3-14
3.4
Test T-4 (Independent 2D Location Capability)
3.4.1
Analysis
The following test reports were provided by the participants:
Administration
Test report reference
France
“C/S D&E T4 Test Report: Independent 2D Location Capability”
SAR-RE-DEMEO-765-CNES 0100 v2.0
Turkey
TRMEO T-4 Report, dated 29 April 2013
USA
Technical Test T-4 Run1 - USA Hawaii MEOLUT Report, Revision 1.0 dated 24 May 2013
Technical Test T4 - USA Maryland Test Report, dated 31 May 2013
3.4.2
Interpretation
3.4.2.1 Canada
From partial analysis of its data, Canada noticed interference and fade outs (possible scintillation
events), when comparing to the spectrum graphics of test T-2, that would have impacted some of the
beacon bursts transmitted and which could explain various missed bursts seen by other participants
MEOLUTs. Some interference was noted as well on the upper transmission channel from the Toulouse
beacon simulator. Canada could not provide further conclusions at this time, but would consider
analysing this data in the future to help with comparisons with, and insights for, the results of Phase II
testing.
3.4.2.2 France
Run 2 performances improved with respect to run 1 results, thanks to the script modification (i.e., lower
beacon transmission rate), the pass schedule optimization and the technical issue of antenna \#3 fixed.
For the run 2, the computed independent location probability was respectively 82 % for the Toulouse
transmission and 64% for the Maryland transmission (both after 4 bursts), which seemed to be
encouraging, even if the results cannot be directly compared to the requirement of 95% after
12 transmitted bursts.
The probability to obtain an independent location accuracy better than 5 km was respectively 72% for
the Toulouse transmission and 56% for the Maryland transmission. The location error was reduced if
only locations derived from four-satellite measurements were considered.
The probability to obtain single burst independent location accuracy better than 5 km did not meet the
requirement of 95%. A complementary analysis showed that this could be due to inaccurate TOA
calibration of the French MEOLUT. However, the location accuracy of the standalone MEOLUT could
be improved by increasing the number of antennas in order to obtain locations with four or more
tracked satellites.
3-15
The time to obtain the first independent location within 5 km from the actual beacon position was of
about two minutes (in the range from 100s to 150s).
3.4.2.3 Turkey
Whereas the 2-channel Ankara MEOLUT generated a number of locations for the run 2 with the
Toulouse beacon simulator, some of them relatively accurate, we consider that a 2-channel MEOLUT
was not really suitable to be involved in location-related tests such as test T-4.
3.4.2.4 USA-Hawaii
Following the analysis of data collected during the test T-4, run 2 the following key observations were
noticed:
An aspect of performance not directly captured in the required results for test T-4 was location
probability. The following table indicates the percentage of the time that the MEOLUT
computed a single burst location relative to when the mutual visibility of three or more satellites
presented the opportunity to compute one, and while much better at higher power, the overall
numbers are somewhat low.
Transmitted
Power (dBm)
Single Burst
Locations Received
Single Burst
Locations Expected
Percentage
71.5%
29.3%
Both Powers
49.9%
Single Burst Location Probability
The data in the three main result tables required for test T-4 did not appear to have a definitive
pattern which may have been a product of the limited data set, specifically due to the low
number of beacon IDs per slot (which also appeared to skew results for the 95th percentile),
The expected pattern of improvement as more bursts are used to compute a location was
successfully demonstrated,
Recording the number of mutually visible satellites and channels provided useful additional
information in the result tables.
As a conclusion, the percentage of locations generated was low, and location accuracy suffered as well
due to many of available locations having poor geometries. The space segment of 11 satellites used
by the Hawaii MEOLUT during this testing did however provide a limited capability, and matters were
compounded by issues with data collection for the Maryland and Toulouse beacon simulator runs.
The Hawaii MEOLUT had very limited participation in test T-4, run 2. The Hawaii reference beacon
(simulator) did not successfully run due to an unexpected issue with the newest script that had been
agreed at JC-27. Due to the lack of mutual visibility between the space segment, the Hawaii MEOLUT
and the Maryland and Toulouse simulators, the number of locations generated was very limited and no
data analysis was possible.
3-16
3.4.2.5 USA-Maryland
The following three charts plot the probability of calculating a location:
The first chart shows how the average probability of location increases as the number of bursts
(NB) increases.
The second chart shows more detail about how the probability of single burst location varies
across time slots and how it may be affected by the number of satellites that were being tracked.
The probability of single burst location varies quite a bit and there is not as strong a correlation
with the number of satellites being tracked as expected.
The third chart shows more detail about how the probability of multi-burst locations (for cases
of NB = 6 and 7) varies across time slots and how it may be affected by the number of satellites
that were being tracked. The probability of multi-burst locations (for cases of NB = 6 and 7) is
more consistent and the correlation with the number of satellites being tracked is more evident.
3-17
The following three charts plot the accuracy of the locations calculated:
The first chart shows how the average location errors decrease as the number of bursts (NB)
increases. For values of NB greater than 2 the average location error is less than 5 km for the
case of nominal beacon power.
The second chart shows more detail about how the accuracy of single burst location varies
across time slots and how it may be affected by the DOP of the satellites that were being
3-18
tracked. The accuracy of single burst location varies quite a bit and the expectation that as DOP
increases, the location accuracy decreases is not consistent.
The third chart shows more detail about how the accuracy of multi-burst locations (for cases
of NB = 6 and 7) varies across time slots and how it may be affected by the number of satellites
that were being tracked. The probability of multi-burst locations (for cases of NB = 6 and 7) is
more consistent and the correlation with the DOP of satellites being tracked is more evident.
3-19
3-20
3.5
Test T-5 (Independent 2D Location Capability for Operational Beacons)
3.5.1
Analysis
The following test reports were provided by the participants:
Administration
Test report reference
Australia
Beacon deployment report (see annex of the Beacon Deployment Report, Rev. 1, dated 26
February 2014, consolidated by the test coordinator)
Canada
Beacon deployment report (see annex of the Beacon Deployment Report, Rev. 1, dated 26
February 2014, consolidated by the test coordinator)
France
“Operational Beacons Deployment For D&E Test T-5 France Participation”
SAR-RE-DEMEO-833-CN v1.0
“D&E T-5 Test Report: Independent 2d Location Capability For Operational Beacons”
SAR-RE-DEMEO-811-CN v1.0
Turkey
T-5 Run1 Turkey Beacon Deployment Report - 24.02.2014
T-5 Run1 TRMEO Report v1 - 23.02.2014
UK
Beacon deployment report (see annex of the Beacon Deployment Report, Rev. 1, dated 26
February 2014, consolidated by the test coordinator)
USA
Participant Report T5 Run01-Maryland, dated 27 February 2014
US Beacon Deployment Report
Figure 10 below provides the locations of the test beacons used for test T-5. More details on the beacon
models, beacon features and their 24-hour activation periods are available in the Beacon Deployment
Report (Rev. 1, dated 26 February 2014) consolidated by the test coordinator from test participants
reports.
3-21
Figure 10: Location of the 33 Operational Beacons Deployed for Test T-5
3.5.2
Interpretation
3.5.2.1 Canada
The frequency channels of test beacons activated in the range 406.037 to 406.040 MHz, were clear of
interference for most of the time for all three days of testing, except for some interference around
406.045 MHz. Beacons in the 406.020 to 406.030 MHz frequency channels had interferers most of the
time especially around 406.024 MHz. This would mean that detection rates of the 406.028 MHz
beacons would present lower detection or throughput rates than those in the range 406.037 and
406.040 MHz. Detection rates for beacons within 800 km of the Ottawa station was about 54%, and
just under 40% for activations within 5,000 km.
Overall detection rate for each antenna of the Ottawa ground station for T-5 tests.
Comparing the spectrum graphics with the MEOLUT results from Ottawa, which showed average
detection rates to be no better than 60% per single channel in distances less than 1,000 km away from
the MEOLUT; and to be around 0.25 or 25% for beacons up to 5,000 km away, interference is clearly
impacting the detection rates. With this lower rate, it would mean a ten-channel MEOLUT or a network
of ten antennas would be needed to achieve single burst locations better than 95% of the time within
Avg Ottawa LUT
Within 800 km
59.0
51.2
49.1
57.2
90.1
10.2
54.1
3000- 5000 km
34.7
15.6
22.7
23.7
59.0
0.9
24.2
5000- 8000 km
24.5
14.8
14.5
15.5
52.8
0.1
17.3
8000- 12500 km
12.5
10.1
9.5
16.8
41.4
0.0
12.2
Beyond 12500 km
4.3
1.3
0.7
2.3
8.0
0.0
2.2
Average
Within 5000km
46.8
33.4
35.9
40.4
74.5
5.6
39.1
Beyond 5000km
15.4
9.8
9.3
12.5
37.4
0.0
11.7
3-22
an equivalent coverage area with radius or distance up to 5,000 km, or about 15% of the earths surface.
For around 1,000 km (about 1% earths surface), a five-channel MEOLUT would give single burst
locations better than 95% of the time. At this stage, however, the results did not provide clear guidance
for the specifications and parameters regarding the exact coverage areas in which the locations could
meet the accuracy requirements of 5 km, 95% of the time, as proposed currently in documents
C/S R.012 and C/S R.018.
Regarding location accuracy, analysis of the Canadian data showed that for unique single burst
locations, where location solutions converged within 2,000 iterations, the error was better than 5 km
more than 75% of the time. Solutions that needed more than 2,000 iterations to converge, or solutions
using multiple bursts, from multiple antennas, showed a greater variance in errors. In any case, the
errors were not less than 5 km, at least or better than, 95% of the time. One reason for this is the
interference which impacts the C/No of the beacon bursts seen at the satellite. At this stage, the results
did not provide clear enough guidance for specifications and parameters regarding the location
accuracy and coverage areas in which the locations can meet the requirements within 5 km, 95% of the
time as proposed currently. More beacons need to be activated in ranges between 1,000 to 3,000 km
range.
All in all though, considering the in-band interference, MEOSAR has good detection capability and
the results are promising regarding the location accuracy, but further testing is needed with more
antennas “seeing” the satellites so more statistically meaningful results regarding the error and
coverage areas requirements can be deduced.
3.5.2.2 France
The methodology to be followed for the post-processing of test T-5 should be reviewed for future runs.
In fact, the use of a fix window for the locations computation was not fully adapted to the current
processing of the French MEOLUT and could affect the interpretation of the results.
The French MEOLUT was configured in automatic antenna tracking mode and collected data with
only three channels most of the time due to a hardware failure on antenna \#4. As a consequence,
location probability and accuracy were strongly impacted by this failure. On top of that, the number
of locations was too low for some beacons, thus preventing the generation of reliable statistics. The
single channel throughput did not exceed 70% even for beacons near the MEOLUT.
Beyond a certain distance between the beacons and the MEOLUT (~ 13,000 km) the detection
probability is sensibly degraded. The probability to obtain a location with an error lower than 5 km is
higher for multi-burst locations than for single burst locations. The overall accuracy results showed
that about 70% of the locations errors are below 5 km. The 95th percentile of the location error is above
22 km for the Toulouse1 France beacon and equals to 15 km for the Maine-United States beacon signals
(single and multi-burst locations included).
Independent location probability with an error less than 5 km after seven transmitted bursts seems to
be non-compliant with the requirement of 95% after 12 transmitted bursts.
Better results are expected for future runs of test T-5 with a MEOLUT nominal configuration (i,e.,
with four channels available) and with the development of the MEOSAR constellations.
3-23
3.5.2.3 Turkey
Prior to test T-5, the Ankara MEOLUT was upgraded from two channels to six channels.
Regarding the detection of activated beacons, 30 out of the 33 beacons deployed were detected by the
Ankara MEOLUT during testT-5, the remaining three beacons either having transmission issues or
being located too far away from the MEOLUT. Five of the 30 beacons detected were located more
than 10,000 km away from the TRMEO, thus confirming the detection benefit of the MEOSAR system,
even with a limited MEOSAR space segment.
Regarding the System Throughput (i.e. probability of burst detection with at least one satellite), on
average around 60% of the transmitted bursts were detected, increasing to 75% for beacons in the
immediate vicinity of the MEOLUT, thus confirming the “low detection rate” issue observed during
earlier tests. Concerning the detection of bursts by multiple channels, only 7.6% of the transmitted
bursts were detected over three days through at least four satellites, and 24.3% through at least three
satellites, which can be attributed to the aforementioned “low detection rate” issue.
Regarding the location probability (i.e., the ratio of the number of n-burst locations to the expected
number of n-burst locations during at least four-satellite covisibility periods), no correlation was
observed between location probability and the distance between the beacons and the MEOLUT.
Location accuracy was, as expected, better within the geographic region of the MEOLUT (a circle
centered at the MEOLUT with a radius of 3,500 km), with a 50th percentile of 2 km and a 75th percentile
of 5 km. However, at its 95th percentile, the location accuracy went up to the 10-30 km range and
sometimes beyond that range. In addition, the following observations were made:
No significant improvement was noticed due to the integraton of up to 7 bursts. In general,
single-burst locations were almost as accurate as multi-burst locations.
The number of satellites used in the calculation of a location seemed to be the most significant
factor determining location accuracy.
Consequently, Turkey anticipated the definition of “nominal locations” as those locations calculated
with four or more satellites, and “marginal locations” as those locations calculated with three or fewer
satellites, possibly with the use, as well, of the DOP value in those definitions, pointing out the potential
necessity for MEOLUT networking in the real world.
If the current “low detection rate” issue was not significantly improved by the future L-band satellites,
four-channel MEOLUTs might have difficulties in systematically generating locations derived from
four satellites.
Turkey recommended that the following parameters be noted and taken into account in the ongoing
work on MEOLUT Specifications and Design Guidelines:
Burst detection rates (MEOLUT System Throughput),
Impact on location accuracy of the number of satellites used to calculate a location,
Concept of “nominal” and “marginal” locations.
3-24
3.5.2.4 USA-Hawaii
Following the analysis of data collected during the test T-5, the following key observations were
noticed
Two beacons supplied by France provided the closest activations to the Hawaii MEOLUT
during test T-5 test, but both were still quite far away (about 4,440 km). There was significantly
less data received from Papeete1 (75 single burst locations) compared to Papeete2 (538 single
burst locations). This large gap from both systems implied either a problem with the Papeete1
beacon, or a very significant difference in the environment in which it was deployed. For those
locations generated, the accuracy was reasonably good, in particular with the limited
constellation of 12 DASS satellites applied, coming in at 70.7% of single burst location
produced within 5 km, and 100% within 5 km at seven received bursts.
While most other statistics were significantly impacted by distance, the range of the MEOSAR
system was soundly demonstrated by a number of cases.
A fair amount of correlation between DOP and location accuracy can be seen, and although
there are areas that do not coincide, DOP values appear to provide useful information and a
good basis for a potential quality factor for MEOSAR data.
As a conclusion, it is emphasized that this test was performed with the limited space segment of
12 DASS S-Band satellites. In addition, all of the beacon activations were a significant distance from
the Hawaii MEOLUT. Overall the location accuracy could be better, but under these less than ideal
circumstances, many good locations were still generated.
3.5.2.5 USA-Maryland
The performance measured during this test was affected by several factors:
1. the distance between the MEOLUT and the beacon,
2. the amount of mutual visibility of the satellites to the beacon and the MEOLUT, and
3. the resulting satellite geometries, or DOP, used to create the single burst solutions.
As mentioned previously, the Maryland MEOLUT used its automatically generated pass schedule each
day, which included the 12 DASS S-Band satellites only being tracked by four antennas. Therefore,
there was no attempt to optimize the pass schedule for any particular beacon location.
However, the number of available satellites was limited to half the number that will be available in the
operational system, which yielded values of DOP that are worse than what would be achievable with
a full satellite constellation, and, therefore, limited the performance of the system.
In order to illustrate these effects, graphs are presented that show the single burst location accuracy
along with the DOP of the satellites used to generate the location as a function of time. On the same
graph the number of satellites with mutual visibility to both the beacon and the MEOLUT is shown.
These graphs were produced for four beacons on Day 3 of testing that were located within 3,000 miles
of the MEOLUT. It can be seen that location accuracy improves as the DOP value is reduced. It can
also be seen that there are significant portions of the 24-hour period that contained fewer than three
satellites with mutual visibility between the beacon and the MEOLUT. Since it takes three or more
3-25
satellites to produce a single burst location, the amount of available data was limited by satellite
visibility.
Of course, this situation will greatly improve as more satellites are added to the MEOSAR
constellation, but the current situation must be taken into account when interpreting the current results.
Admin
Beacon-id (15 Hex)
Beacon
Type
Freq
(MHz)
GPS
Location
Alt
(m)
Approx distance
(km) from
MEOLUT
Received
raw
packets
Location
results
Canada
279C753BAEFFBFF
EPIRB
406.037
Yes
Leland, Frontenac, Ont., Canada
USA
2DDC752E20FFBFF
EPIRB
406.037
Yes
Maine, United States
USA
ADDE41152800330
EPIRB
406.028
Yes
San Juan, Puerto Rico
6.1
USA
2DDC67A0B4FFBFF
PLB
406.028
Yes
Honduras
3-26
3-27
Method 1 Results
The goal for this method was to see how well the MEOLUT could locate the beacon after 1,2,3,…,7
transmitted bursts. Every merged location reported as “7 burst” location was produced by combining
as many single burst locations that occurred within a 7 transmitted burst window.
The following graph includes all beacons for which the Maryland MEOLUT received data. It,
therefore, includes data from beacons that are as far as 8,600 km away from the MEOLUT.
Nevertheless, there is performance improvement as more opportunity, i.e., larger window size, is made
available.
3-28
The counts are given below:
1 Burst
2 Bursts
3 Bursts
4 Bursts
5 Bursts
6 Bursts
7 Bursts
Error 1 km
Error 5 km
Error 10 km
Total
Method 2 Results
The goal for this method was to determine how the merged locations improve as more single burst
locations are used to produce them. In this method, merged locations were reported using the actual
number of single burst locations used to produce the merged locations, not the number of transmitted
bursts.
The following graph includes all beacons for which the Maryland MEOLUT received data. It,
therefore, includes data from beacons that are as far as 8,600 km away from the MEOLUT.
Nevertheless, there is significant improvement as more data is used to produce the merged location.
The slight dip from 6 to 7 bursts is attributed to the relatively small amount of data available for the
7 burst results.
3-29
The following lists the number of data points for each burst count.
1 Burst
2 Bursts
3 Bursts
4 Bursts
5 Bursts
6 Bursts
7 Bursts
Error 1 km
Error 5 km
Error 10 km
Total
3.6
Test T-6 (MEOSAR System Capacity)
3.6.1
Analysis
The following test reports were provided by the participants:
Administration
Test report reference
France
“C/S D&E T6 Test Report MEOSAR System Capacity”
SAR-RE-DEMEO-788-CNES v1.0
Russia
Test report is available on FTP server
Turkey
TRMEO T-6 Report, dated 30 September 2013
USA
Technical Test T-6 Run2 USA Hawaii MEOLUT Report, Revision 1.0, dated 11 April 2014
Technical Test T-6 Run1 - USA Hawaii MEOLUT Report, Revision 1.0, dated 18 July 2013
Technical Test T6 Run 01 - USA Maryland Test Report, dated 6 September 2013
Technical Test T6 Run 02 - USA Maryland Test Report, dated 21 August 2014
3-30
3.6.2
Interpretation
3.6.2.1 Canada
The 406.070 MHz interferer directly overlapped one to three frequency channels in which the run 2 of
test T-6 was executed for the Toulouse beacon. The interferer was present more than half the time of
the test beacon simulator sequence. As well, the 406.060 MHz interferer was also present in many
portions of the test run. This would mean that a reduction anywhere from 10% to 30% in the number
of received bursts would be expected due the interference seen during the T-6 tests. Despite the
interference, overall the detection rates were better, and location accuracies were slightly better, than
those seen in test T-5.
T-6 detection rate or throughput for Toulouse beacon simulator.
Note that the throughput here was about twice the rate than the average throughput
for the T-5 tests from equivalent distances.
While the overall system detection capability results were very good, the location accuracy was not as
good as initially hypothesised would be achieved with a four-antenna system. One reason for this was
the interference which impacted the C/No level of bursts signals as received by the satellite, which
negatively affected frequency and time of arrival estimation accuracy, and thus, the location accuracy.
Based on the current detection rates seen, by increasing the number of antennas, the time to locate and
the location accuracy would improve.
From the analysis, the results did not provide clear guidance on the system capacity threshold or limit.
Canada would recommend that, in Phase II testing, the number of transmitted bursts (NB) be increased
and two more NB levels of 150 and 200 be added, if the beacon simulators could handle this level of
transmissions for the testing duration. As well, analysis might need to be done in smaller time period
ranges or “chunks”, so one could better compare results when specific interferers were not present, or
when beacon view elevation angles were approximately in the same range. All in all, considering the
in-band interference, the MEOSAR system still has very good capacity and the results from Phase I
were promising.
3.6.2.2 France
The system capacity in test T-6 was evaluated in the range of 25 to 100 simultaneous active beacons
and assessed for two system parameters:
throughput performance (detection probability and time to first Valid/Complete Messages),
location performance.
For both parameters, no drop-off value was observed and the system capacity could not be assessed by
any participants.
Antenna
Antenna
Antenna
Antenna
1 In 4
Antenna
Average
4 In 4
Antennas
Antenna
Antenna
Antenna
Antenna
1 In 4
Antenna
Average
4 In 4
Antennas
46.1
43.0
72.2
46.9
91.3
52.0
10.0
41.4
40.5
69.5
45.0
90.2
49.1
7.0
42.9
34.4
60.9
44.7
83.1
45.7
6.5
39.6
32.4
58.4
43.1
81.9
43.4
4.9
35.7
34.9
56.9
53.9
86.2
45.4
6.5
32.9
32.8
54.7
51.8
85.1
43.1
4.9
29.1
31.6
52.6
58.4
81.8
42.9
6.1
26.7
29.8
50.6
56.4
80.8
40.9
4.7
NB
Valid Messages Detection Probability(%)
Complete Messages Detection Probability(%)
3-31
The computation of the multi-burst locations was specific to each MEOLUT due to the diversity in
manufacturer, software and parameter settings. Participants were then invited to provide as much
information as possible about MEOLUTs features and parameters setting in order to consolidate the
analysis of multi-burst location performances.
In order to quantify the system capacity, France suggested conducting future runs of test T-6 with a
larger range of simultaneous active beacons (from 50 to 200 beacons for example).
3.6.2.3 Russia
No observations were available from Run 1 of test T-6 and only data pertaining to a first portion of run
2 of test T-6 was analysed. Note also that only the system capacity using the MEOLUT throughput
performance was assessed as the Russian MEOLUT featured only one antenna. Results from two slots
were slightly inferior to others.
MEOSAR detection probability requirement is defined in document C/S R.012, Annex E, as the
probability of detecting the transmission of a 406 MHz beacon and recovering at the MEOLUT a valid
beacon message, within ten minutes from the first beacon message transmission shall be a minimum
of 99%.
After 350 s, which is a little less than six minutes, the probabilities of producing first complete and
valid message when 100 beacons were transmitting were slightly lower than 80% and slightly above
90% if two specific slots were discarded from consideration. Assuming that the requirement did not
take into account the number of channels of the MEOLUT, a multichannel MEOLUT would
doubtlessly demonstrate better probabilities, drawing nearer to 100%.
T-6 run 1 results had shown almost the same probabilities with Toulouse transmission approaching
100% and Maryland transmission being around 80%.
It was, therefore, assumed that the system capacity based on the MEOLUT throughput performance
was 100 beacons or higher.
3.6.2.4 Turkey
Valid/complete message detection probabilities and times to first valid/complete messages would
normally be expected to worsen as the number of simultaneously active beacons (NB = 25, 50, 75,
100) increased. This was not always observed in the results as the second-best performance for the
Toulouse transmission, run 1 was obtained for NB=100 and the best performance for the Maryland
transmission, run 1 was obtained for NB=50.
Valid/complete message detection probabilities around 80% to 90% for the Toulouse transmission
were achieved regardless of the number of simultaneously active beacons (NB = 25, 50, 75, 100)
whereas the performance was down to around 40% for the Maryland Tx - a direct impact of the large
distance between the MEOLUT and the Maryland beacon simulator. No significant drop in
valid/complete message detection probabilities was observed when NB increased from 25 to 100.
After 100 seconds, almost a 100% message detection rate was achieved to obtain a valid/complete
message in the case of the Toulouse transmission regardless of the number of simultaneously active
3-32
beacons (NB = 25, 50, 75, 100). This performance was around 70% in the case of the Maryland
transmission after 350 seconds - again, a direct impact of the large distance between the MEOLUT and
the Maryland beacon simulator.
As it could be expected, a reliable determination of system capacity in terms of location capability (i.e.,
the NB value which met the 95% probability of a multi-burst location error lower than 5 km) was not
possible with the results obtained by a two-channel MEOLUT (in other words, a 2-channel MEOLUT
was not really suitable to be involved in location-related tests ).
Note that the Ankara MEOLUT did not participate in run 2 of test T-6.
3.6.2.5 USA-Hawaii
Following the analysis of data collected during the test T-6, run 1 the following key observations were
noticed:
Overall, the detection percentages were low, and the expected behaviour was not demonstrated,
For the Maryland transmission, the results remained low even when taking actual mutual
visibility into account, but for Toulouse the detection rates within mutually visible time periods
were high albeit this was a limited time frame.
As a conclusion, the results did not provide the desired outcome. Specifically, a degradation in
performance as the number of beacons increases would be expected, and was not observed. While this
is could be interpreted as a positive result indicating a high system capacity, the distance between the
Hawaii MEOLUT and both simulators, did not produce enough data to drive the test.
Data from test T-6, run 2 with the Toulouse beacon simulator was collected and minimally analysed,
but no results were uploaded to the MEOSAR D&E FTP server. The reference beacon collocated with
the MEOLUT in Hawaii could not provide transmissions at the short intervals required for this test due
to its inherent design. As the purpose of test T-6 was to determine the system capacity of the MEOSAR
system, the more test data that was processed through the MEOLUT the more likely it would be that
the capacity where performance falls off can be determined. The distances between the Hawaii
MEOLUT and both the Maryland and Toulouse beacon simulators severely limited the usefulness of
T-6 results, and with no co-located simulator, no consistent or identifiable pattern from the results
could be achieved.
3.6.2.6 USA-Maryland
The probability of obtaining a Valid or Complete message from a single beacon burst ranged from
95% to 64% for NB equal 25 to 100, respectively. The range of probabilities change from 96% to 75%
if we compensated for the collisions of beacon bursts resulting from the design of the test script. As
expected, the improvement was greater as NB increased.
The probability of obtaining a Valid or Complete message rapidly increased as more beacon bursts
were used until it reached nearly 100% within 350 seconds, or seven transmitted bursts, for all values
of NB.
3-33
The probability of obtaining a location was affected by the number of satellites being tracked by the
MEOLUT. Of course, this was a function of the number of antennas tracking satellites with mutual
visibility to the beacon.
The probability of obtaining a location from a single beacon burst when four satellites were in clear
view of the beacon ranged from 62% to 39% for NB equal 25 to 100, respectively. The average
probability for time slots when fewer satellites were in clear view was significantly smaller,
demonstrating the significant effect that the number of satellites within view has on location
probability.
The probability of obtaining a multi-burst location when four satellites were in clear view of the beacon
ranged from 1% to 93% for NB equal 25 to 100, respectively. Once again, there was degradation, that
was more pronounced as NB increases, when fewer satellites were used.
Location accuracy was clearly affected by the DOP used to calculate the location. It was also clear that
the typical DOP values used to generate this data was not the same as the typical DOP for a full
constellation of 24 GPS satellites. However, the data showed that, when we approached that value,
location performance improved significantly. Therefore, an overall average for location accuracy did
not provide enough insight into the underlying conditions affecting location accuracy to be useful by
itself.
In an attempt to normalize DOP so that a comparison can be made as NB increases, the location
accuracy for the time slot that had the lowest average DOP for each NB was considered. The
probability that the single burst location error was less than 5 km range from 67% to 56% as NB
increased. The probability that the multi-burst location error was less than 5 km ranged from 92% to
65% as NB increases.
To determine system capacity, a required level of performance needed to be determined for each
parameter. However, at this time the partial satellite constellation available added some uncertainty as
to how to use the data measured during this test. The good news was that this test would be run again
during Phase 2 and those results would add clarity to the determination of system capacity.
3.7
Test T-7 (Networked MEOLUT Advantage)
The D&E participants decided to not conduct test T-7 because the network configuration was not
available.
3.8
Test T-8 (Combined MEO/GEO Operation Performance (Optional))
The D&E participants decided to not conduct test T-8 (optional test) due to time constraints.
- END OF SECTION 3 -
4-1
4.
CONCLUSIONS AND RECOMMENDATIONS
This section provides the conclusions commonly agreed by participants in the MEOSAR D&E tests
and their recommendations for future conduct of the tests in the MEOSAR D&E Phase II.
4.1
Conclusion
4.1.1
Test T-1 (Processing Threshold and System Margin)
The test participants agreed on the following conclusions regarding test T-1.
System margin for single-burst throughput using single-channel results
The detection percentage produced from the single-channel testing varied enough and did not
consistently surpass the 70% threshold defined in document C/S R.018, and therefore it was not
possible to arrive at a system margin for single-burst throughput using single-channel results.
System margin for single-burst throughput using multi-antenna results
Results were improved using multi antennas and a margin above the 70% threshold was achieved when
at least four antennas were used. However, it was the view of the participants that the 70% threshold
value was not the right number to use for the multi-antenna results.
Reference Participants reports for description of underlying causes of the variability of the results.
4.1.2
Test T-2 (Impact of Interference)
The test participants agreed on the following conclusions regarding test T-2.
From Canadas analysis, in its T-2 report, it was noted that persistent interference is seen by all
MEOAR satellites around 406.070 MHz, as well as intermittent interference around 406.060 MHz,
when in view of north and eastern quadrature of the globe. This interference is not typically seen in
the upper half of the 406 MHz band when satellites do not see Western Africa and Eurasia. Similar
interference and noise behaviour was observed in all spectrum graphics for all tests. Locations of these
interferes are provided in the T-2 report.
These interferers directly impact the transmissions of the beacon simulators used for all such tests, and
indirectly on all tests due to the effect of the interference energy on the AGCs of the satellite repeaters;
and other interferers directly impacted T-5. This is one reason for the difficulty in determining and
reaching capacity and system thresholds, and for the negative impact on the detection rates seen in
tests T-1, T-3 and T-6 of Phase I.
4.1.3
Test T-3 (MEOLUT Valid/Complete Message Acquisition)
Two runs of test T-3 were conducted. While not all MEOLUTs participated in a second run referred
as a modified test, whose script was characterized by a lower transmission rate, compared to run 1.
4-2
Most of the participating MEOLUT got improved results with the second run using the modified test
script.
For nominal power of 37 dBm, the results of the test T-3 has shown that the probability of detection of
a valid message is about 90% after 1 burst and higher than 99% after 7 bursts, which is compatible
with expectation for minimum performance at full operational capability (FOC) contained in Annex E
of document C/S R.012 (99% after 10 minutes) for many MEOLUTs.
The expected performances were not reached for some MEOLUTs, and while all causes could not be
explained, many might be explained by:
limited co-visibility conditions (in term of number of channels) with the beacon simulators,
the test script restricted beacon transmission to 7 consecutive bursts per beacon ID (350 sec vs.
600 sec, which is equivalent to the 10 minutes allowed for 99 % probability of detection as per
the MIP),
limited functionality and/or limited participation of some MEOLUTs to some test runs,
interference
The main conclusions drawn from the test were the following:
 the results are compatible with the expectation for minimum performance at full operational
capability (FOC) contained in Annex E of document C/S R.012 (MIP).
 the results have shown the deterioration of the valid/complete message production performance
as the distance between MEOLUT and the beacon simulator increases.
 the average detection probabilities improved with an increase of the number of transmitted bursts
as well as with an increase of the beacon transmission power.
 the transfer time measured for valid messages was around 5 to 10 seconds. Message transfer
times improved (i.e, got shorter) with the beacon transmission power.
 the difference between the first run and the modified script may be explained by some limitations
in the processing capacity of the participating MEOLUTs. It was decided to use the modified
test script for the subsequent runs of tests T-3 and T-4, which was agreed by Participants and
included in document C/S R.018 Issue 2 Revision 1 at CSC-51.
It is expected that the results will improve as the MEOSAR L-band space segment is expanded in the
future.
4.1.4
Test T-4 (Independent Location Capability)
The initial run of test T-4 highlighted issues that necessitated a revision of the test script and a second
run of the test. Run 2 of test T-4 was conducted with a modified transmission script as agreed during
JC-27 in order to reduce the beacon transmission rate (one beacon every two seconds with only one
frequency, instead of one beacon every 0.5 second alternating between two frequencies). The results
summarized here, therefore, result from the analysis of run 2 of test T-4 only.
Independent Location Probability
The probability that a MEOLUT provides an independent 2D location with a location error less than
X km (X = 1, 5 or 10 km) did not reach desired values. Performance for X = 5 km varied but was
always less than 95% but improved as the number of transmitted bursts used was increased.
4-3
Independent Location Accuracy
The 50th percentile, the 75th percentile, and the 95th percentile of the location error of 2D locations did
not reach desired values. Performance for 95th percentile varied but was always greater than 5 km but
improved as the number of transmitted bursts used was increased
Time to First Independent Location
The time elapsed between the first burst transmitted and the first 2D independent location with an error
less than X km (X = 1, 5 or 10 km) was not more than 2 to 3 minutes.
Conclusions
While the results were not as good as expected, the tests results showed that:
The probability of calculating a location was affected by:
o the single channel throughput,
o the number of satellites being tracked with mutual visibility to the beacon,
o the amount of time, or number of beacon bursts transmitted.
The accuracy of calculated locations were impacted by many reasons including:
o the number of satellites used to derive the location (locations derived with additional
satellites were statistically more accurate),
o the geometry of the satellites tracked by MEOLUT antennas (characterized by DOP
for example),
o the signal to noise ratio (C/No) of each channel,
o the accuracy and the calibration of TOA/FOA measurements produced by the
MEOLUTs.
It is expected that the results will improve in the future as:
the probability of location and the location accuracy of the standalone MEOLUT will increase
as the number of available antennas-satellite pairing increases, thus allowing the MEOLUT to
obtain locations with more tracked satellites.
the availability of more satellites on orbit will allow for the typical DOP to be improved which
will increase location accuracy.
4.1.5
Test T-5 (Independent 2D Location Capability for Operational Beacons)
Detection benefit of the MEOSAR system
The tests carried out over 3 days with 33 operational beacons deployed worldwide soundly
demonstrated the vast geographic range of individual MEOLUTs and confirmed the detection benefit
of the MEOSAR system, even with a limited MEOSAR space segment that consisted, at the time of
the tests, of 12 DASS, 2 Galileo and 1 Glonass satellites. Some participants only tracked the DASS
satellites whereas other participants tracked all available satellites.
4-4
Detection rates
The detection probability was gradually degraded as the distance between the beacons and the
MEOLUTs increased. These results, which were corroborated by the detection rates observed during
earlier tests, led the participants to the conclusion that an increase in the number of available satellites
the number of antennas per MEOLUT and an improvement of the single channel detection rate
hopefully brought by the advent of operational L-band satellites would be needed to improve
performance on system detection and meet the expectation on location probability and accuracy.
Independent Location Probability
The location probability did not meet the expectation for minimum performance of 98% at full
operational capability (FOC) contained in Annex E of document C/S R.012. However, some accurate
locations were generated for beacons at extreme distances (e.g., greater than 7,000 km) from the
MEOLUT, which is indicative of the range of MEOSAR.
Independent Location Accuracy
The location accuracy did not meet the expectation for minimum performance of 5 km accuracy, 95%
of the time at full operational capability (FOC) contained in Annex E of document C/S R.012.
However, the location accuracy was, as expected, better within the geographic region of the MEOLUTs
(a circle centered at the MEOLUT with a radius of some 3,000 km), location error was frequenctly
below 5 km wihin that geographic region. Composite locations calculated by the integration of up to
7 bursts offered a higher probability to obtain a location accuracy better than 5 km.
T-5 processing methodology
Run 1 of the T-5 tests has been an opportunity to identify certain missing aspects of the original T-5
processing methodology, and to amend the “windowing methodology” and include MEOLUT System
Throughput and n-Burst Independent Location Probability into the list of data to be analyzed and
reported by Participants.
Conclusion
The test results showed that:
the MEOSAR systems capability to detect beacons is very good, sometimes beyond
expectations,
the location accuracy was not as good during this first run of test T-5 as initially hypothesized,
the results did not provide clear guidance for the specifications and parameters regarding the
exact coverage areas in which the calculated locations could meet the accuracy minimum
performance expectation at full operational capability (FOC) contained in Annex E of
document C/S R.012 (MIP) (5 km, 95% of the time),
one reason for the limitation in the location accuracy performance was the negative impact of
interference on the channel detection rates,
increasing the number of MEOLUT antenna-satellites pairings would improve the location
probability and accuracy, as well as the time to locate,
all in all, MEOSAR has very good detection capability and the location results were promising.
4-5
4.1.6
Test T-6 (MEOSAR System Capacity)
The system capacity in test T-6 (standalone MEOLUT configuration) was evaluated in the range of 25
to 100 simultaneous active beacons and assessed for two system parameters:
throughput performance (detection probability and time to first Valid/Complete Messages),
location performance.
The test did not lead to clear conclusion regarding throughput and location performances and it was
agreed to increase the number of transmitted beacon to try reaching a clear decrease of performance.
The expected outcome in term of location probability and accuracy were not always observed by the
participants even for NB = [25,50,75,100] simultaneous beacons.
Participants at EGW-1/2014 agreed to modify the definition of test T-6 for future runs, which was
subsequently approved in document C/S R.018 Issue 2 Revision 2 at CSC-53.
The modifications to the test T-6 definition were:
an increase of the number of NB transmitted beacons from NB = [25,50,75,100] to NB =
[25,50,75,100,150,200],
modification of the analysis methodology to introduce the computation of the probability to
produce single burst locations.
The computation of the composite locations is specific to each MEOLUT (manufacturer, software and
parameters setting). Therefore, participants are invited to provide as much information as possible
about MEOLUTs features and parameter settings to assist the analyses.
4.1.7
Test T-7 (Networked MEOLUT Advantage)
No conclusions were drawn on test T-7 as this test was not conducted during the MEOSAR D&E
Phase I.
4.1.8
Test T-8 (Combined MEO/GEO Operation Performance (Optional))
No conclusions were drawn on test T-8 as this optional test was not conducted during the MEOSAR
D&E Phase I.
4.2
Recommendations for the Conduct of Subsequent D&E Phases
4.2.1
Test T-1 (Processing Threshold and System Margin)
It is recommended in the analysis of future runs of test T-1 to:
attempt to correlate lower performance cases with occurrence of channel interferers,
attempt to determine a more suitable threshold value for the assessment of the system margin
for single-burst throughput using multi-antenna results.
4-6
4.2.2
Test T-2 (Impact of Interference)
It is recommended to continue to monitor the 406 MHz spectrum and prepare spectrum plots so that
the single-channel throughput can be analysed for any technical tests, to allow the evaluation of the
negative impact interference has on the MEOSAR, and by extension the entire Cospas-Sarsat System.
4.2.3
Test T-3 (MEOLUT Valid/Complete Message Acquisition)
It is recommended that the test be modified to be able to compare the test results with the expectation
for minimum performance at full operational capability (FOC) contained in Annex E of document
C/S R.012 (i.e., 99% probability of valid beacon message detection within ten minutes). Specifically,
beacon transmission could be extended to 13 transmitted bursts (i.e., ten minute beacon transmission)
or the test analysis could aggregate results from multiple beacon IDs to generate the desired results
without changing the test script. The final methodology shall be proposed by the Test Coordinator and
agreed with other participants.
4.2.4
Test T-4 (Independent Location Capability)
It is recommended that further analyses be conducted to evaluate the relationship between location
accuracy and various parameters (e.g., DOP, number of satellites used in location determination, C/No
measurements, etc.).
In order to achieve expected results when T-4 is run as part of Phase II, more satellites than were
available during this Phase I test runs are needed. It is desirable to have as many L-band satellites as
possible.
4.2.5
Test T-5 (Independent 2D Location Capability for Operational Beacons)
It is recommended to continue monitoring the 406 MHz spectrum and it is suggested that participants
optionally analyze the detection rates per channel and per satellite type (L-band vs. S-band).
4.2.6
Test T-6 (MEOSAR System Capacity)
Modifications to the test T-6 definition regarding the number of simultaneous active beacons and the
analysis methodology were agreed at EWG-1/2014 and it is recommended that test participants
conduct future runs of test T-6 accordingly.
Participants are invited:
to provide as much information as possible about MEOLUTs features and parameter settings
to assist the analyses,
to provide details about the methodology applied to compute statistics regarding detection and
location performances and in particular about the mutual visibility conditions of the test periods
selected for the test.
4.2.7
Test T-7 (Networked MEOLUT Advantage)
No recommendations were provided on test T-7 as this test was not conducted during the MEOSAR
D&E Phase I.
4-7
4.2.8
Test T-8 (Combined MEO/GEO Operation Performance (Optional))
No recommendations were provided on test T-8 as this test was not conducted during the MEOSAR
D&E Phase I.
4.3
Recommendations for the Implementation of the MEOSAR System
It has been noted over many MEOSAR-related meetings and tests to date that TOA and FOA
measurement accuracies were amongst crucial factors affecting location accuracy. It is therefore
recommended that Participants revisit and fine tune their MEOLUT TOA/FOA measurements
techniques in order to improve location accuracy
It is recommended that further analyses be conducted to evaluate the relationship between location
accuracy and various parameters (e.g., DOP, number of satellites used in location determination, C/No
measurements, etc.) to provide useful information and a basis for a potential Quality Factor for
MEOSAR data.
- END OF SECTION 4 -
A-1
ANNEX A
DETAILED LOG OF PHASE I TESTS
Week
Nb
Date Start
Test
Test
Run
Time 1st Tx
(yyyy-mm-dd UTC)
Time last Tx
(yyyy-mm-dd UTC)
Beacon
location
Comments
2014-03-05
T-6
2014-03-05 14:45:00
2014-03-07 23:58:00
Maryland
2014-02-05
T-6
2014-02-05 16:05:00
2014-02-07 19:13:00
Maryland
CANCELLED due to weather conditions
2014-01-03
T-6
2014-01-03 18:00:00
2014-01-03 18:41:00
Maryland
Dry run (one sequence)
2013-12-16
T-6
2013-12-16 17:53:00
2013-12-18 22:45:00
Maryland
Retest - DISREGARD due to simulator issue
2013-12-11
T-6
2013-12-11 21:40
2013-12-13 16:00:00
Toulouse
2013-12-09
T-6
2013-12-09
2013-12-11
Maryland
CANCELLED due to weather conditions
2013-11-
14/19/21
T-5
2013-11-14
14:00:00
2013-11-19
14:00:00
2013-11-21 14:00:00
2013-11-15
14:00:00
2013-11-20
14:00:00
2013-11-22 14:00:00
Many
locations
Beacons of various types but in any case test coded and homer device
deactivated
2013-08-29
T-4
2013-08-29 14:00:00
2013-08-30 14:00:00
Toulouse
Retest of the test conducted on 2013-07-25
2013-08-13
T-1
2013-08-13 17:00:00
2013-08-14 17:00:00
Hawaii
With the script agreed at JC-27, except used country code 367 instead of
338.
Transmissions were on whole second intervals (3 & 4 seconds apart)
2013-08-01
T-1
2013-08-01 14:00:00
2013-08-02 14:00:00
Maryland
With the script agreed at JC-27
2013-07-31
T-1
2013-07-31 14:00:00
2013-08-01 14:00:00
Toulouse
With the script agreed at JC-27
2013-07-25
T-4
2013-07-25 14:00:00
2013-07-26 14:00:00
Toulouse
With the script agreed at JC-27
2013-07-24
T-4
2013-07-24 14:00:00
2013-07-25 14:00:00
Hawaii
TRANSMISSION CANCELED due to a simulator issue.
2013-07-23
T-4
2013-07-23 14:00:00
2013-07-24 14:00:00
Maryland
With the script agreed at JC-27
2013-05-23
T-6
2013-05-23 14:00:00
2013-05-24 14:00:00
Maryland
The 24 minute script will run once each hour
2013-05-16
T-6
2013-05-16 12:00:00
2013-05-16 14:00:00
Toulouse
Transmission slot calculated to maximize the satellite visibility for all
the MEOLUTs (except Hawaii). Beginning of each sequence: 12:00,
12:30, 13:00 and 13:30
2013-05-14
T-6
Dry run
2013-05-14 13:00:00
2013-05-14 14:30:00
Toulouse
Each transmission slot with NB=25,50,75,100 beacons
2013-04-25
T3
modified
-
2013-04-25 14:00:00
2013-04-26 14:00:00
Maryland
T-3 at a lower beacon transmission rate (24 beacons in 48 seconds)
2013-04-23
T3
modified
-
2013-04-23 13:00:00
2013-04-24 13:00:00
Toulouse
T-3 at a lower beacon transmission rate (50 beacons in 48 seconds)
2013-04-04
T-4
2013-04-04 13:00:00
2013-04-05 13:00:00
Hawaii
2013-04-02
T-4
2013-04-02 13:00:00
2013-04-03 13:00:00
Toulouse
Transmission stop between April 2, 21:00 UTC and April 3,
07:00 UTC. Transmission extended until April 4, 01:00 UTC
2013-04-01
T-4
Dry
Run
2013-04-01 16:00:00
2013-04-01 18:00:00
Hawaii
Dry run
2013-03-28
T-4
Dry
Run
2013-03-28 15:00:00
2013-03-28 17:00:00
Toulouse
Dry run
2013-03-25
T-4
2013-03-25 14:00:00
2013-03-26 14:00:00
Maryland
2013-03-20
T-4
Dry run
2013-03-20 14:00:00
2013-03-20 20:00:00
Maryland
Dry run
2013-03-13
T-1
2013-03-13 13:00:00
2013-03-14 13:00:00
Maryland
Toulouse
A-2
2013-03-07
T-1
2013-03-07 13:00:00
2013-03-08 13:00:00
Hawaii
2013-03-06
T-1
Dry run
2013-03-06 18:00:00
2013-03-06 19:00:00
Hawaii
1 hour dry run
2013-02-25
T-1
Dry run
2013-02-25 11:00:00
2013-02-25 13:00:00
Toulouse
2013-02-20
T-1
2013-02-20 14:00:00
2013-02-21 14:00:00
Maryland
2013-02-14
T-3
2013-02-14 14:00:00
2013-02-15 14:00:00
Maryland
2013-02-13
T-3
2013-02-13 13:00:00
2013-02-14 13:00:00
Hawaii
24 beacons transmitted per power level (instead of 100)
2013-02-08
T-3
Dry run
2013-02-08 16:00:00
2013-02-08 18:00:00
Hawaii
Subset of test T-3 (24 beacons transmitted per power level, instead of
100). Beacon log file available on the D&E server.
2013-02-05
T-3
2013-02-05 13:00:00
2013-02-06 13:00:00
Toulouse
NOTICE: the test started on 2013-02-05 at 14:00:00 and ended on
2013-02-06 at 14:00:00.
No beacon messages transmitted between 15:00:00 and 15:15:00 on
2013-02-05
2013-01-29
T-3
Dry run
2013-01-29 21:30:00
2013-01-30 21:30:00
Maryland
Subset of test T-3
2013-01-10
T-3
Dry run 2013-01-10 08:30:00
2013-01-11 08:30:00
Toulouse
Beacon log file available on the D&E server
2012-12-20
T-3
Dry run
07:30:00
10:11:00
Toulouse
Subset of test T-3
2012-12-06
T-3
Dry run
09:00:00
10:11:40
Toulouse
Subset of test T-3. Back-up window 12:55:00 14:16:40
- END OF ANNEX A -
- END OF DOCUMENT -
Cospas-Sarsat Secretariat
1250 Boul. René-Lévesque West, Suite 4215, Montreal (Quebec) H3B 4W8 Canada
Telephone: +1 514 500 7999 / Fax: +1 514 500 7996
Email: mail@cospas-sarsat.int
Website: www.cospas-sarsat.int