Loading [MathJax]/extensions/MathZoom.js
A Survey on Technologies, Standards and Open Challenges in Satellite IoT | IEEE Journals & Magazine | IEEE Xplore

A Survey on Technologies, Standards and Open Challenges in Satellite IoT


Abstract:

The Internet of Things (IoT) is expected to bring new opportunities for improving several services for the Society, from transportation to agriculture, from smart cities ...Show More

Abstract:

The Internet of Things (IoT) is expected to bring new opportunities for improving several services for the Society, from transportation to agriculture, from smart cities to fleet management. In this framework, massive connectivity represents one of the key issues. This is especially relevant when IoT systems are expected to cover a large geographical area or a region not reached by terrestrial network connections. In such scenarios, the usage of satellites might represent a viable solution for providing wide area coverage and connectivity in a flexible and affordable manner. Our paper presents a survey on current solutions for the deployment of IoT services in remote/rural areas by exploiting satellites. Several architectures and technical solutions are analyzed, underlining their features and limitations, and real test cases are presented. It has been highlighted that low-orbit satellites offer an efficient solution to support long-range IoT services, with a good trade-off in terms of coverage and latency. Moreover, open issues, new challenges, and innovative technologies have been focused, carefully considering the perimeter that current IoT standardization framework will impose to the practical implementation of future satellite based IoT systems.
Published in: IEEE Communications Surveys & Tutorials ( Volume: 23, Issue: 3, thirdquarter 2021)
Page(s): 1693 - 1720
Date of Publication: 26 May 2021

ISSN Information:

Funding Agency:


CCBY - IEEE is not the copyright holder of this material. Please follow the instructions via https://creativecommons.org/licenses/by/4.0/ to obtain full-text articles and stipulations in the API documentation.
SECTION I.

Introduction

The impact of IoT on many aspects of life, being they industry, logistic or everyday use, is growing quickly. New promising use cases are constantly added, supported by new and smarter technologies, better energy-efficient devices and communication technologies. As a matter of fact, IoT looks set for becoming one of the key technologies in our future. Smart and connected objects are able to massively collect valuable data for supporting informed decisions, reducing operational costs through automation (in industry and home), tracking objects and materials, monitoring assets and environmental parameters, enabling more effective and innovative healthcare solutions. Besides data collection, interconnecting things creates an ecosystem where cooperation and federation are key for optimizing processes and improving reliability. The advances of other emerging technologies such as federated machine learning or edge computing promise to well integrate in this scenario, and bring added-value benefits.

Recent market studies envision that the total number of connected IoT devices will reach 83 billion by 2024, rising from 35 billion connections in 2020. The industrial IoT (IIoT) sector by itself, that includes manufacturing, retail and agriculture, is forecast to account for over 70% of all IoT connections by 2024, with a grow of the number of IIoT units of 180% over the next four years [1].

In the perspective of expanding the benefits brought by IoT to include geographical areas where it is not usually viable, for technical and/or economic reasons, to provide ubiquitous coverage, telecommunication providers and enterprises are looking for integrated IoT-based global coverage solutions. Indeed, IoT operation can become critical in remote areas with low/no cellular connectivity for many different industries such as transportation (maritime, road, rail, air), fleet management, logistics, solar, oil and gas extraction, offshore monitoring, utilities smart metering, farming, environment monitoring, mining, and many others. With this perspective, satellite-based technologies that can be integrated with existing IoT terrestrial networks seem to be the way to go. Satellites are therefore conquering the special and important role of including in this ecosystem remote geographical areas where terrestrial networks are unavailable or out of reach, such as on remote land (think at the case of areas such as, e.g., forests) as well as offshore (e.g., in the oceans).

Analyst firm Omdia forecast that the global satellite IoT connectivity business will more than double its revenues, going from \$ 233 million in 2019 to \$ 544 million in 2025. Cumulative satellite connections are expected to rise fourfold to more than 10 million by 2025 [2]. The installed base of satellite IoT connections is expected to increase by a nearly a factor of four in the coming years, growing at a 25 percent compound annual growth rate (CAGR) from 2.7 million units in 2019 to 10.3 million units in 2025.

In this perspective, satellite operators are exploring new opportunities to enter or strengthen their presence in the IoT market. On the other hand, traditional communication service providers and vendors, but also new and dynamic start-ups, are exploring new opportunities, e.g., by exploiting the existing terrestrial wireless long-range technologies as ground-to-satellite link enablers. Their success will greatly depend on the capacity of industry, standardization bodies, and space agencies in converging their interests and views into viable (complementary) solutions. The work to be done starts from existing IoT technologies, with the aim of enabling seamless integration with incumbent terrestrial infrastructures. This issue represents a key aspect in order for IoT to be quickly accepted and integrated in the current and future communications ecosystem. Indeed, such a scenario evolves very quickly, and it includes both mature technologies as well as new advances.

Existing long-range wireless IoT technologies mainly derive from two different strands: mobile cellular networks and low-power wide area network (LPWAN) emergent technologies. As for mobile networks, which are currently managed by nation-wide mobile network operators (MNOs), various radio access technologies (RATs) are available, spanning from the second-generation (2G) General Packet Radio Service (GPRS) [3], the third-generation (3G) Universal Mobile Telecommunications Service (UMTS) [4], and the fourth-generation (4G) Long-Term Evolution (LTE) standard [5]. Despite mobile networks were historically designed to satisfy human-originated traffic and human-centered applications, in recent years the standardization efforts of Third Generation Partnership Project (3GPP) progressively introduced support for IoT traffic [6]. In particular, massive machine-type communication (mMTC) is expected to support latency/reliability-tolerant IoT traffic with a very high number of service requests per base station (BS), but it was demonstrated that current 4G technologies are not capable to support such IoT traffic due to limitations of the control-plane signaling. The model presented in [7] shows that a BS can accommodate at most few thousand devices to guarantee access latency below 100 ms with high transmission success probability. A lower access latency, in the order of 10 ms, can be achieved only with BSs serving an unrealistically small numbers of devices. However, in the context of the upcoming fifth-generation (5G) of cellular networks, the 3GPP is working on enhancing mMTC traffic support and, relying on the New Radio (NR) air interface [8], relevant attention has been devoted to ultra-reliable low-latency communication (URLLC), that is, latency/reliability-constrained IoT traffic with fairly limited amount of service requests per BS.

An alternative way of supporting mMTC with long-range wireless links is to leverage on ad-hoc RATs solutions, commonly referred to as LPWANs [9], [10]. Those solution mainly comprise unlicensed-spectrum technologies, which are subject to restrictions on the medium access control (MAC) policies (e.g., frequency hopping, duty cycle, listen-before-talk) and transmission policies (e.g., transmission power limits) depending on the region of the world in which they operate. Examples of such networks are LoRaWAN [11] and Sigfox [12]. On the other hand, LPWAN technologies operating on licensed spectrum such as Narrowband IoT (NB-IoT) [5], which are managed by the traditional MNOs, are also available.

Our Contribution: In this survey, we aim at giving an comprehensive overview of the technologies involved in making the satellite IoT paradigm real for mMTC use cases. The survey comes from ongoing discussion among the authors around the different technological approaches to support massive IoT services in extremely wide areas. Indeed, all authors addressed IoT service design issues in their research activities by different perspectives (industry vs academia, but also physical communications vs networking, and terrestrial vs satellite solutions), and the survey represents an effort to provide a holistic view about the problem, including the current scenario, existing industrial efforts and future perspectives. To the best knowledge of the authors, no similar surveys are available in the literature providing such view, useful both for researchers as well as practitioners and innovators in the area of massive IoT.

Our contribution can be summarized as follows:

  • we review the current long-range IoT solutions and identify their main features and limitations;

  • we introduce the concept of satellite IoT, and contextualize it against the recent scientific literature and industrial/commercial initiatives;

  • we overcome the limitations of the scientific literature [6], [10] by focusing our paper on the analysis of the pros and cons of satellite access as far as IoT traffic is concerned;

  • we clarify the distinction between Internet of Remote Things and space-information networks in the context of satellite IoT;

  • we identify and describe the main challenges to be faced in the coming years to fully support the satellite IoT vision;

  • we discuss the latest and future developments of satellite-IoT communications and networking technologies, outlining potential opportunities of further research contributions and taking in due consideration the “5G and beyond” scenarios.

It is worth remarking that, as also stated in the title, the authors want to deal with all the above mentioned items in a holistic and standardization/regulation-oriented fashion. In fact, we will consider not only contributions in the scientific literature, but we complement them with technical papers, reports, and specifications in the standardization community, which are often left aside in this kind of surveys. We also provide a survey of the emerging industrial initiatives in this regards, thus linking the identified innovation fields to the practical activities that are trying to cover the gap between the research world and the real implementation of communication systems for IoT based on satellite links.

In this framework, the closest related work in the recent literature are [13]–​[15], that are three surveys about the emerging trends of satellite communications thanks to dense deployments of miniaturized satellites [13], [14] and 3GPP technologies [15]. Moreover, [16]–​[19] deserve to be mentioned as relevant related work since they provide the background of the definition we will give to satellite IoT. Finally, [20] provides useful information regarding regulations of satellite communications [20, Sec. 2.5] and license-exempt spectrum for short-range devices [20, Ch. 3]. Our work differs from the mentioned papers because i) we define a more general concept of satellite IoT networks, considering a wide range of alternative implementation modes and technological enablers, and ii) our focus is more comprehensive insofar it includes international regulations, standards, and emerging technologies. A comparison table between the present contribution and the related work in terms of covered topics and methodologies is provided in Tab. I.

TABLE I Comparison Between the Present Survey and the Closest Related Work
Table I- 
Comparison Between the Present Survey and the Closest Related Work

Paper Organization: The rest of the paper is organized in three parts. The first part comprises three sections:

  • in Section II, we introduce the current technological enablers of terrestrial long-range IoT, highlighting their limitations in supporting certain IoT use cases;

  • in Section III, we introduce the paradigm of satellite IoT, describing architecture variants, envisioned key performance indicators (KPIs), utilized constellations, and spectrum matters;

  • in Section IV, we focus on 5G and LPWAN enablers of satellite IoT in two separate subsections, respectively.

The second part, which coincides with Section V, features a survey of the most interesting commercial initiatives in the field of satellite IoT, categorizing them according to the technological enabler.

Finally, in the third part we provide an extensive Section VI containing our observations regarding the open challenges and future directions of the research and standardization in this context, and the conclusions of our work in Section VII.

Appendix deals, eventually, with the authorization procedures for satellite communication systems. For the readers’ convenience, the list of used acronyms is also provided at the end of the paper.

SECTION II.

Taxonomy of Long-Range IoT Solutions and Their Limits

Existing terrestrial long-range IoT wireless solutions and standards are considered the most appropriate at driving satellite IoT market expansion for their scalability, reduced costs, greater supplier diversity and easier integration for the customers.

All state-of-the-art long-range IoT wireless solutions share common features that can be summarized as follows:

  • a radio access network (RAN) based on a star topology – each IoT node, hereafter referred to as machine-type device (MTD), is connected directly over-the-air (with a single hop) to an IoT gateway providing connectivity towards an application server, reachable through the Internet;

  • bi-directional communication – both uplink (from the MTD towards the serving gateway) and downlink (from gateway towards the served MTDs) is supported; while the data-plane traffic is mostly carried in uplink, the downlink is required to carry mainly control-plane traffic;

  • a core network infrastructure, which performs radio resource management (RRM) via control-plane signaling, and user-plane traffic routing.

The various solutions differ in several aspects, i.e.,

  • how they address and prioritize constrained devices in terms of energy-efficiency, processing and memory requirements;

  • constraints in terms of airtime usage and bandwidth limitation (LPWANs can only send small blocks of data at a low rate, and therefore are better suited for use cases that do not require high bandwidth and are not time-sensitive);

  • operation in licensed or unlicensed spectrum (with varying degrees of performance in key network factors);

  • standardization;

  • scalability.

A. Cellular Networks (GSM and LTE) for IoT

In the context of cellular networks, the MTDs and the IoT gateway are typically referred to as user equipments (UEs) and the BS, respectively. The management of the network is performed by the mobile core network, comprising the necessary network functions to perform.

At the time of writing, the 4G Evolved Packet System (EPS) is the most popular cellular technology and it leverages the LTE air interface. 5G is currently being deployed and already available only in the major cities. Nonetheless, the 2G Global System for Mobile Communications (GSM) is still operational on precious portions of spectrum below 1 GHz [21], which have favorable propagation properties. The UMTS, which is based on the Wideband Code Division Multiple Access (W-CDMA) technology, is also up and running, but, especially in Europe, it will be dismissed even before 2G networks,1 thus it can be neglected as a long-term wireless technology for mMTC support.

Despite both GSM and LTE were not designed to support mMTC traffic, they were identified as good candidates to support long-range IoT: while the former exploits frequency bands which offer very good coverage areas to UEs, the latter represents the most widespread cellular technology. Clearly, appropriate changes to the original standards had to be made, yielding to Extended Coverage GSM for IoT (EC-GSM-IoT) and Category-M LTE specifications. EC-GSM-IoT systems are based on the GPRS subsystem of GSM [3], [22] and operate on traditional 200-kHz-wide GSM carriers mainly allocated within the frequency band around 900 MHz (Extended GSM 900 Band) [21]. This evolution of the GSM is effectively capable of reducing the UE complexity while supporting energy-efficient operation with extended coverage in both uplink and downlink; the coverage improvement is up to 20 dB, depending on the supported coverage class [3, Sec. 3.3.9]. A similar recipe was adopted for LTE, where a new UE category for bandwidth-limited low-complexity devices targeting mMTC applications was introduced [5, Sec. 23.7]. These devices, referred to as Cat-M1 or Cat-M2 UEs, can operate in any LTE system bandwidth provided that they exploit a minimum channel bandwidth of 6 physical resource blocks (PRBs) (corresponding to the channel bandwidth of a 1.4 MHz LTE system) in both downlink and uplink. The difference between a Cat-M2 UE and a Cat-M1 UE is that the former can supports larger maximum transport block sizes for unicast transmission compared to the latter. It is worth mentioning that Cat-M LTE can be provided also in the unlicensed spectrum according to the MulteFire specifications [23].

Remarks — Both EC-GSM-IoT and LTE Cat-M are amendments of preceding standards to provide long-range wireless connectivity to MTDs. They intrinsically represent temporary solutions, bridging existing technology generations (2G and 4G) towards the next one, i.e., the 5G [24]. In fact, a large body of research has been carried out about the new 5G air interface design, called NR, in order to provide native support to mMTC [25]. The main challenge consists in designing efficient radio access protocols, which are capable of supporting a massive number of devices contending for network access, while allowing for dynamic bandwidth sharing and different approaches at the physical layer (PHY) and MAC layers to ensure multi-service integration under the same network infrastructure (i.e., the so-called network slicing). By fully integrating mMTC in the 5G System (5GS), the 3GPP would enable an extended terrestrial coverage for IoT use cases especially in populated areas, where the mobile networks are widespread. For further considerations about the potential of 5G in supporting IoT, we refer the reader to Section IV-A.

Nevertheless, in the standardization road map, the priority has been given to the design of URLLC, which deals with the more profitable IIoT field, thus at the time of writing the specification of 5G-enabled mMTC is still ongoing – see studies on NR small data transmissions in inactive state [26] and on reduced-capability NR devices serving IoT use cases, but whose requirements are stricter than EC-GSM-IoT and LTE Cat-M but lower than URLLC [27].

B. NB-IoT

Other than revising the default 2G and 4G systems, as seen in the previous subsection, the 3GPP decided to design a brand-new standard for supporting long-range IoT, following the LPWAN paradigm instead of the legacy cellular network one. This initiative gave birth to NB-IoT, that is, a LPWAN standard operating on licensed-based frequency bands.

NB-IoT is derived from the LTE standard, and it provides access to network services using an optimized PHY for very low power consumption (e.g., full carrier bandwidth is 180 kHz for both downlink and uplink) [5]. As indicated in the relevant subclauses of the specification [5], a number of typical LTE protocol functions (e.g., handover, measurement reports, carrier aggregation, dual connectivity) are not supported for NB-IoT. The NB-IoT carrier can substitute a GSM one (whose width is 200 kHz), or can be integrated into a LTE carrier as a PRB (whose width is exactly 180 kHz) that can be either allocated in-band or in the guard band; multi-carrier operation (i.e., utilization of multiple PRBs) is supported, as well. In the downlink, each physical channel occupies the whole NB-IoT carrier, with only one channel per subframe (equal to 1 ms). In the uplink, instead, both single-tone and multi-tone resource units can be scheduled. Single-tone transmissions may exploit a subcarrier spacing of either 3.75 kHz or 15 kHz with a slot duration of 2 ms or 0.5 ms, respectively. Multi-tone transmissions, instead, operate only at 15 kHZ and can group 3, 6, or 12 subcarriers for 4 ms, 2 ms, or 1 ms, respectively.

It is worth noticing that the MulteFire Alliance specified a custom version of the NB-IoT standard (MF NB-IoT) to operate on sub-GHz license-free frequency bands, in particular the 902–928 MHz band in the United States of America (USA), where frequency hopping is adopted, and 863–870 MHz band in the European Union (EU), where duty cycle is adopted [23], [28]. The major changes of MF NB-IoT versus 3GPP NB-IoT are in the physical layer, more specifically in the synchronization signals and broadcast channel, for a better performance and compliance to unlicensed band regulations. On the other hand, the remaining physical channels and procedures of MF NB-IoT as well as the high layer protocols are the same as 3GPP NB-IoT.

Remarks — According to recent insights from mobile industry veterans [29], the main problem on NB-IoT does not reside in the standard or technical appropriateness, rather is a matter of business case. The MNOs are becoming increasingly aware that each NB-IoT connection delivers very little revenue with respect to a flat subscription of a LTE smartphone (0.10 \$ per month vs 20 \$ per month). Thus, they prefer to invest in fully-fledged 5G solutions, which are envisioned to provide a solution to the effective support of mMTC. However, as said in the previous subsection, this creates a gap, since such 5G solutions will not be ready to market still for some time.

C. Sigfox

As far as native unlicensed spectrum technologies are concerned, one of the major players is Sigfox, which aims to be a global service provider of IoT for customers ranging from big manufacturers to start-up companies and mostly single users.2 In Sigfox, MTDs are referred to as end points, while the IoT gateways are defined as radio hub [12]. In conjunction with the Sigfox Cloud, which performs network management functions, the radio hubs create a Sigfox network (SNW) that connects the end points to an application server. The Sigfox communication is referred to as “3D-UNB,” and it implements a ultra-narrowband differential binary phase shift keying (BPSK) modulation scheme in uplink and a Gaussian frequency shift keying (GFSK) modulation scheme in downlink. In fact, various radio configurations are defined to comply with the local regulations, nevertheless all of them contain one uplink macro-channel and one downlink macro-channel which are only 192 kHz wide. When frequency hopping is required by local regulations (e.g., in the USA), 6 contiguous micro-channels of 25 kHz can be implemented in each uplink macro-channel.

Four classes of end points are defined based on the transmission power profile (from “Class 0” to “Class 3”), where Class 0 refers to the maximum transmission power allowed by the local regulations, while increasing class index entails a more aggressive power backoff scheme. The uplink procedure is initiated by an end point wishing to send an uplink message to the Sigfox network; to increase the resiliency of the uplink communication, repeated (up to 3) frame transmissions are possible. On the other hand, the bidirectional procedure is initiated by an end point wishing to send an uplink message and receive an onward downlink message; if a downlink message is sent by the network and received successfully by the end point, a confirmation message is sent by the end point. We finally remark that the reception capability is optional for end points, and that the periodicity of uplink and downlink procedures are pre-defined by Sigfox (up to 140 12-byte uplink messages and 4 8-byte downlink messages per day).

Remarks — Sigfox is both a proprietary technology and a (French) operator. As an operator, Sigfox is using the franchising business model: networks in countries different from France are run by other companies, i.e., the franchisees. The name of the franchisee is not know to the end user, which see the same logo and the same interfaces. Of course, a revenue sharing model is pre-agreed between Sigfox itself and the operating companies in the various countries. Given the particular business model, the service subscribers see a single global worldwide network and no provisions for roaming are technically needed. On the downside, Sigfox is a rather “closed” technology, being it managed by a single company and not by a standardization body or a wide alliance of companies.

D. LoRa and LoRaWAN

Another prominent LPWAN technology in the unlicensed spectrum is LoRaWAN [11]. As far as the PHY is concerned, this RAT relies on both a proprietary modulation called Long-Range (LoRa), which is derived from chirp spread spectrum (CSS) [30], and on frequency-shift keying (FSK); the available bandwidth is either 125 kHz, 250 kHz, or 500 kHz for uplink channels and 500 kHz for downlink channels. From MAC layer upwards, on the contrary to Sigfox, the protocol stack is open and maintained by the LoRa Alliance.3 The RAN comprises radio gateways and and end devices (EDs), with the latter ones divided into three operational classes: all (Class A), beacon (Class B), and continuously listening (Class C). As shown in Fig. 1, each class corresponds to a given energy-saving operating mode based on the up-time of the ED receiver: in Class-A the up-time is minimized because the ED is in listening mode only after an asynchronous uplink transmission, while in Class-B the radio gateways send periodic beacons to synchronize the EDs and make them open receive windows at specific time instants within the radio frame. Finally, in Class-C the ED is always listening to the downlink channel when it is not transmitting in uplink.

Fig. 1. - Graphical description of the three LoRaWAN classes.
Fig. 1.

Graphical description of the three LoRaWAN classes.

Regarding the core network of LoRaWAN, it is commonly referred to as back-end and comprises three servers: a Network Server, which is the mastermind behind the RRM, a Join Server, which manages the security keys, and an Application Server, which is the interface towards external applications. In particular, the LoRaWAN back-end design enables worldwide roaming of MTDs by i) exploiting the Join Server as a trusted third-party owner of secret keys, (thus, smoothing the transition between different network operators), and ii) decoupling the security of the control-plane information (e.g., MAC commands, for which the end point is the Network Server) and the data-plane information (for which the end point is the Application Server) [31].

Remarks — Despite LoRaWAN provides native roaming support to MTDs, some issues are yet unresolved. For example, it has not yet been clarified how a LoRaWAN device can become aware of its location, that is, e.g., realizing that it is out of UE coverage and under USA coverage, where the operating frequency bands and the restrictions are different. This could be possible exploiting an integrated global navigation satellite system (GNSS) or by means of Class-B beacons signaling such information, but clearly both options are not part of the default configuration of a LoRaWAN network – Class A EDs without a GNSS module.

Table II summarizes the features and the limitations of the previously described terrestrial, long-range IoT technologies.

TABLE II Classification of Long-Range IoT Technologies
Table II- 
Classification of Long-Range IoT Technologies

SECTION III.

Satellite IoT: a New Paradigm

The introduction of terrestrial long-range IoT has helped overcoming the limitations of short-range, wireless sensor network technologies [10], broadening the number of application scenarios where IoT can be successfully deployed. In parallel, the relevance of scenarios where the MTD are dispersed over vast and remote areas started to rise in the last few years. The prominent market sectors to which these scenarios belong include transportation, oil and gas, agriculture, and mining [32]. These sectors are growing in relevance since IoT solutions promise to automatize remote maintenance and control operations, reduce their cost, and improve the whole system’s responsiveness and resilience. However, in such scenarios, the terrestrial network infrastructures are often not fully reachable and not even profitable to be deployed. Therefore, connectivity alternatives must be found.

In this context, satellite network support is seen as a viable alternative to overcome the missing terrestrial infrastructure. Satellite service providers have proposed to exploit legacy solutions, providing backhaul network capabilities to sensors gateways [17]. The paradigm of Internet of Remote Things [16], otherwise called also Internet of Everything Everywhere [17], has been introduced to describe the concept of leveraging satellite communication as a cost-effective solution for IoT systems, able to connect remote sites and sensors with the rest of the world. Pushing this concept further, other types of non-terrestrial networks (NTNs) can be considered, opening the scene to a heterogeneous set of technological enablers other than traditional satellites, which can be included in the picture. These include unmanned aerial vehicles (UAVs), drones, high altitude platforms, and airships. The derived networks, the so-called of space-information networks entails the complex management of the interworking among network segments to deliver messages to the intended destination [18].

Other than for communications, satellites are typically employed for remote sensing of Earth and Space physical parameters,4 and equipped with various sensors (e.g., for atmospheric monitoring), thus becoming themselves “special” MTDs. As such, in the authors’ opinion, a comprehensive view of the IoT should also embrace the case in which satellites are considered things themselves, thus pushing further the previous concepts [19]. The joint scenario of terrestrial and space MTD gives rise to a satellite IoT paradigm, capable of serving a variety of applications.

Fig. 2 outlines a sample deployment scenario of an heterogeneous satellite IoT system. On the Earth side, the figure shows a harbor area, in which transport ships and logistic companies exchange pallets and containers of goods, each of which is associated with a MTDs. On the space side, we can notice the presence of a satellite MTD gathering data from the surrounding environment.

Fig. 2. - Reference architecture of a satellite IoT systems. The infrastructure components are labeled in yellow, while the communication architecture components are labeled in black.
Fig. 2.

Reference architecture of a satellite IoT systems. The infrastructure components are labeled in yellow, while the communication architecture components are labeled in black.

In the next subsections, we will introduce the envisioned architectural variants to implement satellite IoT networks, the spectrum matters, the comparison metrics against the above mentioned terrestrial RATs, and the kinds of orbits that are available for satellite communication, focusing on their impact on the end-to-end (E2E) connectivity. We remark that we will primarily focus on the Internet of Remote Things/Internet of Everything Everywhere paradigms, while the space-information network as an IoT component of the overall satellite IoT system will be left aside as it deserves a dedicated treatment.

A. Architecture Matters

We can identify the following infrastructure components of a satellite IoT system:

  • the MTDs, which are located in terrestrial offshore/isolated locations (e.g., a ship in the middle of the sea), but also, as shown in Fig. 2, in the space;

  • the IoT-supporting satellites, forming a constellation that covers even the most remote areas on Earth;

  • the IoT gateways, which, when needed and available, are located in the proximity of a large (yet isolated) cluster of MTDs (e.g., on a ship). Their aim is that of gathering uplink messages generated by the MTDs and send them to the satellite constellation, as well as forwarding downlink messages received from the satellites to the MTDs;

  • ground stations, which are deployed in various Earth spots to collect the aggregate messages gathered by IoT-supporting satellites and providing them with downlink messages to be forwarded to the MTDs;

  • IoT services, comprising core network functionalities to perform RRM as well as external cloud applications for user-plane data management.

The E2E connectivity between MTDs and the IoT services is made up of different stretches. First of all, we have the wireless link between the (terrestrial) MTDs and the serving IoT-supporting satellite. Two configurations are envisioned [16], [32], namely:

  • indirect-to-satellite communication (also called IoT backhauling), where the MTD communicates with a (terrestrial) IoT gateway, which in turn has a bidirectional link towards an IoT-supporting satellite. In indirect mode, the existing protocols can be leveraged, but the area of a deployment is limited by the coverage range of the terrestrial IoT gateway. This configuration is supported by [16] and shown in Fig. 2 as the naval cargo scenario;

  • direct-to-satellite communication (also called direct access), where the serving IoT gateway is deployed on the satellite. A direct access from the MTD to the satellite is a more appealing solution in scenarios which are lacking infrastructure deployments or with a low density of MTDs. This variant is supported by [17], [33] and shown in Fig. 2 as the remote logistics scenario. This option is preferred also by various industrial initiatives – see Section V.

Among the satellites of the constellation, inter-satellite link (ISL), typically based on free-space optics (FSO) (including laser and visible light communication), may be exploited to perform “switching in the sky,” i.e., inter-satellite routing of messages. Once a suitable satellite is reached, a wireless satellite-to-ground link is then used to deliver uplink messages/collect downlink messages from a ground station. It is also possible that uplink (downlink) messages are stored and carried by the serving satellite until a suitable ground station (the target MTD) is in sight [33]. Finally, the IoT services are connected to ground stations thanks to a terrestrial link, e.g., in optical fiber.

It is worth mentioning, finally, the role of edge computing and its impact on the network architecture in the satellite IoT paradigm. Indeed, if user-plane traffic needs to be processed locally while still allowing an IoT island to be remotely connected to remote IoT services then the indirect-to-satellite architecture is necessary, with the presence of an edge computing server co-located with the gateway to treat data. The core network infrastructure must also be tailored to such scenarios, moving the needed network functionalities to the network edge to ensure the E2E system operation.

B. Spectrum Matters

The common naming of frequency bands for satellite communications follows the Institute of Electrical and Electronics Engineers (IEEE) nomenclature: a list of them is reported in Table III. In particular, let us remark that, when considering constellations of small satellites, the spectrum below 5 GHz is technically advantageous because of the low rain fade and reduced antenna design complexity requirements [32].

TABLE III Common Naming of Satellites Frequency Bands in the IEEE Nomenclature
Table III- 
Common Naming of Satellites Frequency Bands in the IEEE Nomenclature

Among the existing systems and spectrum resources, frequency bands below 3 GHz for both geosynchronous equatorial orbit (GEO) and non-GEO satellite networks appear as the more common solutions to address the majority of IoT services belonging to the mMTC category [32]. In Table IV, we provide a simplified spectrum allocation as per [34].

TABLE IV Allocation of Spectrum to Satellite Services as per [34]
Table IV- 
Allocation of Spectrum to Satellite Services as per [34]

C. Comparison Metrics Against Terrestrial IoT RATs

Based on the previous considerations, we shall consider the following KPIs in order to position the new paradigm of satellite IoT in the arena of long-range IoT technologies:

  • coverage area – the major benefit of satellite IoT resides in outperforming the availability of mobile networks and LPWANs thanks to its ubiquitous coverage;

  • uplink capacity – in terms of supported terminals for uplink communication, we may safely state that satellite IoT systems will outperform terrestrial RAT since it provides more flexibility in allocating capacity in areas where it would otherwise be not available;

  • downlink capacity – on the contrary to the previous item, the downlink communication may be challenging due to the extremely long distance between transmitter (the satellite) and the receiver (the terminal) with certain architectural variants (mainly the direct-to-satellite architecture) and because of spectrum allocation issues;

  • spectrum availability – the presence of dedicated frequency bands for satellite communications minimizes the impact of interference typical of LPWANs, thus making this paradigm more similar to licensed-spectrum-based mobile networks;

  • minimum latency – given the complexity of the involved architectures, the E2E latency of a satellite IoT system is expected to be larger than that provided by mobile networks or LPWANs;

  • terminal complexity – especially for the direct-to-satellite architecture, the terminal engineering may be challenging in order to cope with the distance towards the serving satellite;

  • infrastructure complexity – as shown in Fig. 2, various network elements are needed to provide E2E connectivity, thus the complexity of the network infrastructure is expected to be equal/greater than that of mobile networks;

  • roaming support – depending on the international availability of spectrum portions, the roaming capability of satellite IoT systems should outperform that of mobile networks and LPWANs;

  • energy efficiency – apart from the longer distance to be covered, the energy efficiency of satellite IoT in terms of power spent per transmitted bit over surface unit is likely to be somewhat similar to that of mobile networks.

We captured the above mentioned considerations in the radar chart in Fig. 3.

Fig. 3. - Radar chart to compare the envisioned KPIs across different long-range IoT technologies (LPWANs, mobile networks, and satellite IoT).
Fig. 3.

Radar chart to compare the envisioned KPIs across different long-range IoT technologies (LPWANs, mobile networks, and satellite IoT).

D. Orbits and Constellations for IoT-Supporting Satellites

The satellite’s path and what view it will have of Earth are determined by two features: i) the satellite’s height (altitude) and ii) the orbit’s shape (eccentricity and inclination) [35]. Based on the actual values taken by these parameters (altitude, eccentricity, and inclination), we can classify a satellite into three categories [35], [36]:

  • high-Earth orbit satellites, reaching about 36,000 km of altitude. In this spot, the satellite travels moves at the same angular velocity as the Earth, thus following a geosynchronous orbit (GSO). If a GSO satellite circles Earth above the equator (zero inclination), it is referred to as a GEO satellite, and it appears in a fixed position (geostationary), thus being suitable to provide coverage to a specific ground area (e.g., telecommunication satellites);

  • low-Earth orbit (LEO) satellites revolve at an altitude between 160 and 1000 km from the Earth surface. Unlike GSO satellites, LEO satellites fly at a much faster pace because of their proximity to Earth (e.g., 7.8 km/s versus 3 km/s);

  • medium-Earth orbit (MEO) satellites travel a wide range of orbits anywhere between LEO and GEO.

As far as the orbit’s eccentricity is concerned, usually the satellite orbit is circular (zero eccentricity) or has a slight eccentricity. However, in some case, it is useful to set a highly elliptical orbit (HEO), having a perigee below the altitude of 1,000 km and an apogee that can be above 35,000 km. A prominent example of HEO is the Molniya orbit, which is exploited by the homonym Russian satellite series to monitor polar regions that could be difficult to reach with a GEO.

At the time of writing, we identified five major mobile satellite service providers which are active in supporting IoT use cases.

  1. Inmarsat5 – It features 13 GEO satellites that deliver mobile safety and broadband voice and data services around the world by leveraging a nexus of ground stations, which act as traffic gateways directing the satellite signal to terrestrial networks, and data centers located at strategic points around the world. Inmarsat is an investor in Actility,6 that is, the major supplier of Network Server solutions for LoRaWAN. Together with Actility, Inmarsat deployed the first global IoT network,7 with LoRaWAN-based connectivity on the ground and satellite connectivity at the network backbone. The integrated hybrid network has been tested in three application scenarios: i) asset tracking, ii) agribusiness, and iii) oil and gas, deployed in remote regions of Australia and Malaysia [17];

  2. Iridium8 – It consists of a LEO constellation of 66 cross-linked satellites. The global coverage is ensured by a dynamic satellite network that leverages ISLs (each Iridium satellite is linked to up to four others) to route traffic among satellites. As for IoT traffic support, Iridium provides i) a simple and efficient network transport capability for transmitting short data messages between equipment and centralized host computer systems called Short Burst Data, and ii) a cloud-based solution that fosters the inter-working with industry-standard IoT protocols. Experimental prototypes of heterogeneous Iridium-LoRAWAN systems have also been proposed in [37];

  3. Eutelsat9 – With a global fleet of GEO satellites and associated ground infrastructure, Eutelsat is one of the world’s leading satellite operators. It is worth noticing that the so-called “Eutelsat LEO for objects (ELO)” constellation is expected to be operational and integrated with the Sigfox terrestrial connectivity services by 2022, in order to support the upcoming IoT market boom in sectors as diverse as transport, oil and gas, and agriculture.10

  4. Globalstar11 – It features a constellation of LEO satellites and 24 ground stations that constitute an Internet Protocol (IP)-based terrestrial infrastructure to bridge satellites and traditional communications infrastructure on six continents. On the contrary to Iridium, the Globalstar technology does not leverage ISLs, allowing customers to connect to a different satellite or gateway and automatically hand off to another available satellite upon need. Globalstar supports tracking and messaging applications for various industrial IoT use cases without strict latency requirements [16, Sec. 2A].

  5. Thuraya12 – With its two satellites, it contributes to the fleet of five GEO satellites of the Yahsat satellite operator in providing a broad range of fixed and mobile services spanning voice and data communications to both commercial and government sectors. IoT services, both for land and maritime sectors, are supported.

E. Recent Trend: Independent LEO Satellites for IoT

Besides the above mentioned initiatives, a new, interesting one has been gaining momentum, that is, independent LEO satellite constellations. They consists of independent, small (typically, cubic-shaped) satellites, which are expected to give a new impulse to the initiatives surrounding IoT because of their low deployment cost, shorter deployment time, and flexibility. These advantages come at the cost of lower payload capabilities and shorter lifespan, with the latter not being actually a disadvantage, as it gives room to a rapid adaptation to new technology developments. Moreover, being LEO-based, lower propagation delay and smaller propagation loss are experienced as compared to GEO solutions [16], [17]. These unique characteristics have drawn to attention of players such as Google, Space X, OneWeb, and Facebook. Since the weight influences many aspects of a space mission, the common classification of small satellites is based upon it, thus distinguishing among femto (less than 0.1 kg), pico (0.1–1 kg), nano (1–10 kg), micro (10–100 kg), and mini (100–1000 kg) satellites.

Among pico-satellites, CubeSat have emerged for their modular architecture that has become a standard [13]. Becoming cheaper, closer and smaller, they have quickly become enablers of new applications in various fields. Current CubeSat research is mostly focused on remote-sensing applications but they are gaining momentum also in communications and IoT. Indeed, small satellites take advantage of being deployed ad LEO orbit thus providing lower-latency with respect to other solutions. In the CubeSat architecture, a “basic unit,” also called 1U, is defined, and all CubeSat satellites are designed as multiples of 1U, up to 16U. A 1U has a cubic shape with side dimension of 10 cm. This shape is not just a convenient form factor but also provides sufficient surface area for solar power generation and space-thermal stability. Its mass is also limited, equaling 1.33 kg per 1U. The power that a CubeSat can allocate to the communication system is limited by its weight and size constraints: the maximum transmission power is restricted to 1 W (or 30 dBm) in downlink and 100 W (or 50 dBm) in uplink. In the past 20 year, thousands of different CubeSat missions have been launched, being part of the related to telecommunications and IoT. Choices like orbital altitude, in satellite antenna type and constellation are limited by CubeSat specific characteristics. Design choices affect aspects such as the bean coverage and should be taken according to expected application [13].

These recent trends make LEO satellites the preferential solutions for the deployment of satellite IoT systems, as shown in Fig. 4. We remark that this trend is sometimes referred to as “Internet of Space Things” in recent scientific literature [13], [38]–​[40]. This terminology should not create confusion in the reader, because, in our view, the Internet of Space Things is just a subset of the satellite IoT paradigm defined at the beginning of Section III.

Fig. 4. - Classification of satellite orbits. Recently, the typical orbit range considered for IoT-supporting satellites is represented by the shaded area. The displayed orbit inclination is only for graphical purposes and does not necessarily correspond to a real inclination.
Fig. 4.

Classification of satellite orbits. Recently, the typical orbit range considered for IoT-supporting satellites is represented by the shaded area. The displayed orbit inclination is only for graphical purposes and does not necessarily correspond to a real inclination.

In the following, we survey the alternative approaches to implement satellite IoT networks over LEO constellations in terms of link type and ISL configuration.

1) RF Satellites for IoT (With and Without ISL):

In [41], some criteria for the design of RF LEO satellite constellations for IoT have been discussed. In particular two main application scenarios have been introduced, namely: delay tolerant and delay sensitive applications. LEO constellations without ISLs are advisable for delay tolerant applications thanks to low cost and reduced complexity, as they avoid complicated operations of inter-orbital switching. In this type of constellation, the orbit eccentricity plays a key role in determining the coverage area. The use of recursive orbits ensures constant overhead pass times and power level needed for communication. Rosette constellations [42] are recommended in [41] to extend the coverage of LEO constellations without ISL. The 3D coverage diagram of a Rosette constellation with inclination orbit 42 degrees is shown in Fig. 5.

Fig. 5. - Example of 3D coverage diagram of LEO constellation for IoT, without inter-satellite link (courtesy of [41]).
Fig. 5.

Example of 3D coverage diagram of LEO constellation for IoT, without inter-satellite link (courtesy of [41]).

On the other hand, the ISL solution can improve coverage and reduce latency at the price of an increasing complexity. Polar orbital planes are preferred to inclining orbit planes in the design of LEO constellations with ISL [41]. In order to reduce cost and complexity, a minimum number of satellites and ISLs is considered in order to realize the global coverage. In particular, ISLs between orbital planes are ignored by LEO constellations for IoT. The 3D coverage diagram of a LEO constellation of IoT with ISL in shown in Fig. 6. The constellation consists of 40 satellites in five orbit planes. Each satellite has two bidirectional ISLs with its two vertically adjacent satellites in the same plane.

Fig. 6. - Example of 3D coverage diagram of LEO constellation for IoT, with inter-satellite link (courtesy of [41]).
Fig. 6.

Example of 3D coverage diagram of LEO constellation for IoT, with inter-satellite link (courtesy of [41]).

The satellite constellation design should satisfy some specific requirements of throughput and latency. This is what has been addressed in [43], where a satellite system based on an elliptic orbit satellite and a phase-array ground station is proposed for IoT connectivity and Consumer Internet access. The system uses five Molniya satellites and one dimension electronic scanning phased-array terminal. The authors of [43] demonstrated that the elliptic orbit of Molniya well fulfill the requirements of both IoT users and Consumer Internet.

Another remarkable work in the field of RF satellites for IoT is [44], where the authors attempted to transplant IoT LoRaWAN technology to space, by treating the satellite constellation like a set of “things” operating in the sky. Such a configuration may fit very well applications where small satellites (e.g., CubeSats) are by themselves sensors that collect, for instance, Earth observation data. The evaluation and test trials performed in [44] evidenced the known disadvantages of LoRa technologies in terms of reduced data rates. Organizing the satellites into clusters of tenths of nodes, the farthest nodes will achieve a data rate of 293 bps that is unsuitable to support future broadband IoT connections.

2) Optical ISL Solutions for Satellite IoT:

A practical solutions for optical ISL in satellite-based IoT has been proposed in a recent work [45] that is based on the use of laser communication terminals. Such a solution is very effective as it combines very high data rates and resilient performance. Indeed, the use of laser communication terminals guarantees high operational security and immunity to interference source, jamming and eavesdropping. The configuration of [45] considers four ISL link terminals per satellite, providing interconnections to satellites in front, behind, to the right and to the left of the host satellite. Typical requirements are in the 6,000 km distance range, asking for 10 Gbps and more. Such requirements can be fulfilled by an laser communication terminal with mass of 15 kg and a power consumption of 80 W.

Remarks — The utilization of LEO satellite constellations for IoT support is a recent trend, which does not preclude the exploitation of legacy systems based on MEO and GEO constellations. Of course, in such a last case, the issues related to latency should be carefully taken into account. Moreover, the configuration of a direct connection from nodes to satellite has recently gained momentum with respect to the indirect configuration in many recent papers [17], [33] as well as recent commercial initiatives, which will be surveyed in the next pages. Finally, the implementation of satellite backhauling via ISLs (i.e., the so-called switching in the sky) is rather challenging [13], thus it should be carefully exploited, mainly for wide-area IoT network deployments where the message delivery delay is bounded and the serving satellite is not in the line of sight with any ground stations before the delivery deadline has expired.

SECTION IV.

Satellite IoT Enablers

This section describes which contributions might be brought by 5G systems and LPWANs as enablers for satellite IoT. For purpose of clarity, such enablers are first discussed separately in the following two subsections and then the main remarks are drawn in a conclusive final subsection.

A. Satellite IoT Enablers in 5G Systems

Satellites are expected to cover a relevant role in IoT support in 5G systems. Indeed, while terrestrial communications will probably cover the majority of application scenarios, satellite will most likely play a role in the framework of mMTC, for their wide area coverage and relatively short service deployment time. Tab. V, which contains data from [46], summarizes the key satellite mMTC-UE requirements in two distinct time frames: 2018 (i.e., today) and 2023, when 5G is expected to be deployed worldwide. The table clearly outlines two major specific requirements of IoT application scenarios: i) the sparseness of data communication and relaxed latency constraints, and ii) the mean time between maintenance operations. The latter represents an extremely long time, that is measured in years, which is expected to further increase as we look into the future scenarios.

TABLE V Envisioned Requirements for mMTC UEs in the Time Frame 2018–2023, Extracted From [46]
Table V- 
Envisioned Requirements for mMTC UEs in the Time Frame 2018–2023, Extracted From [46]

Various 5G-enabled satellite network architectures are envisioned [47], [48], mainly depending on how the connectivity between UE and BS is implemented (with/without regenerative satellites, with/without ISL). In all cases, two main differences can be identified between 5G NTNs and the traditional terrestrial ones, being i) the long distance between mMTC-UEs and the RAN infrastructure, which introduces long delays in the communication, and ii) the potential amount of devices to be supported in the IoT scenario. The delay component is particularly relevant in the case of GEO satellites, which at an altitude of 36,000 km introduce a minimum one-way latency between mMTC-UEs and RAN of 238 ms. For this reason, the so-called 5G NTNs envision the usage of alternative (lower) orbits and even high-altitude platforms (UAV-enabled/airborne BSs) operating at an altitude of few tens of km to mitigate the latency (down to 2 ms) and Doppler shift issues [49].

A distinctive feature of 5G will be the softwarized and virtualized architecture that, starting from the mobile core network in the form of the revolutionary 5G service-based architecture (SBA), will extend up to the RAN. Indeed, 5G is expected to bring three major advancements that might enable massive IoT connectivity through satellites.

  1. O-RAN – The concept of Open Radio Access Network (O-RAN) is to enable detachment of some functionalities from the BS of the mobile network to move them into another location.13 This enables to implement different strategies to balance complexity and power consumption of the BS. As an example, [50] demonstrates the potential to employ CubeSats as remote processing units to support lightweight BSs that can be mounted on UAVs. This architectural solution is based on the exploitation of the so-called remote radio head (RRH)-baseband unit (BBU) splitting functionality, which allows to separate the protocol stack of the BS in different positions – with corresponding different performance constraints related to the corresponding connection between the two detached components.

  2. PEPs – Performance-enhancing proxies (PEPs) are network agents designed to improve the end-to-end performance of some communication protocols. PEP standards are widely known and used solutions, and they are defined in RFC 3135 (PEPs intended to mitigate link-related degradation) and RFC 3449 (TCP performance implications of network path asymmetry). PEP can be considered as a general approach to improve data transfer performance over satellite connections but also as a potential solution to improve IoT flows performance on satellite connections, e.g., by merging several small flows to reduce the granularity of data connections.

  3. Network slicing – As far as 5G standardization is concerned, the upcoming standard will focus the core network innovation into supporting the concept of network slicing. Network slicing allows to generate virtual self-contained networks that can be adapted in terms of topology, performance and services to fit the constraints of different services. Indeed, the emerging paradigm in 5G seems to point towards configuring and allocating different network slices for different kinds of applications (e.g., mobile broadband, mMTC, URLLC). The mMTC slice will be of specific interest for supporting IoT traffic, as also mentioned in the open challenges later on in the paper.

On a more general level, the direct use of unmodified Internet protocols, such as Hypertext Transfer Protocol (HTTP)/1 and /2 and Transmission Control Protocol (TCP), is challenging in resource limited IoT devices (e.g., CPU, memory, power) and resource constrained networks (e.g., high-latency, low-power, lossy). The Internet Engineering Task Force (IETF) is addressing this issues, and it has developed lightweight application transport protocols that are better suited for the transmission of small messages with low duty cycle in constrained environments, called Constrained Application Protocol (CoAP) [51], to be used in LPWANs for IoT services. CoAP endpoints exchange messages according to a request/response mode, and the resources are accessed through a Uniform Resource Identifier (URI) similar to HTTP. CoAP packets are much smaller (10–20 bytes) than HTTP/TCP flows and run over User Datagram Protocol (UDP).

Remarks – The integration of satellite connections for IoT support in 5G, and in general with IP-native networks, provides specific challenges, including:

  • Compensation of long transmission delay, with the support of Performance Enhancement Proxies, a widely used technique (see discussion above) to split TCP connections in order to better adapt the congestion control strategy to the satellite link while maintaining the terrestrial section un-modified;

  • The need to overcome some of the limitations of HTTP/1 and HTTP/2. The IETF is currently standardizing HTTP/2 support over QUIC, a new UDP-based, stream-multiplexing, always-encrypted transport protocol focused on minimizing application latency. This transition could lead to the standardization of an HTTP/3 supported by QUIC, with the advantage that QUIC provides reliable data transfer and pluggable congestion control, which can be optimized for IoT scenarios. Nevertheless, other solutions supported by IETF and being standardized offer more drastic solutions, such as the development of IoT specific protocols capable of efficiently compressing and adapting data to the needs and requirements of IoT devices.

Additional details and directions of research are described in Section VI related to open challenges.

B. Satellite IoT Enablers in LPWANs

With the introduction of satellite links, LPWAN technologies start to be called low-power global area network (LPGAN). In particular, the recent literature studies have shown the feasibility of using two LPWAN technologies, namely LoRaWAN and NB-IoT, over satellite links.

With respect to the former, the EDs are allowed to employ the frequency under the regulations of their country, including the actual frequency band and maximum transmitting power. In [52], the regional parameters of LoRaWAN, e.g., effective irradiated power (EIRP), carrier frequencies, data rates, are reported for the different regions the Earth Globe. The regulatory adopted for spectrum usage in satellite M2M/IoT applications have been recently detailed in [32] for the different operational scenarios. We recall that the radiofrequency (RF) modulation format of LoRaWAN is proprietary (patented by Semtech) and based on a derivative of CSS modulation [11]; it uses variable orthogonal spreading factors, which allow to trade data rate for range or power to optimize system performance for a constant bandwidth. The applicability of the LoRa modulation to satellite links was investigated by the authors of [53], who investigated the Doppler effect over links lengths of approximately 550 km. Their experiments were carried out in the laboratory and outdoors, using similar velocities and reproducing the Doppler effect over the link with software-defined radios (SDRs) [33]. The experimental results showed that the data rate at such long distances is low, enabling messages of a few bits. In order to improve the performance of the traditional LoRa modulation in this context, Semtech has recently announced a new modulation format called LoRa-E to be used only in uplink, while the current LoRa modulation is still exploited for downlink, thanks to the fact that the radios can switch between the two modulations [54].

These studies confirm that the direct-to-satellite architecture is the preferred one for LoRa-based LPGANs. Such insights are confirmed also in [17], which investigates the technical challenges to be faced for interconnecting satellite and LoRaWAN networks, with a particular focus on: synchronization, gateway selection and replicas cancellation and, finally, cross-layer optimization of network protocol. Related to [17] is the paper authored by Wu et al. [55], where the LoRa adaptability in LEO satellite IoT are reviewed. In particular, network architecture, bandwidth usage, activation mode, and access protocols are redefined considering the satellite scenario and its peculiar constraints.

As for NB-IoT, similar considerations as in Section IV-A, especially regarding the network architecture [56], apply. Nevertheless, in literature various studies regarding enhancements of the NB-IoT air interface to support satellite communications can be found. In [57], a detection algorithm for NB-IoT is designed to minimize the impact of the satellite channel, specifically of the Doppler drift, in the demodulation performance. An uplink scheduling technique for LEO satellite-based NB-IoT systems that mitigates the level of the differential Doppler down to a value tolerable by the IoT devices without increasing their complexity is proposed in [58]. Finally, in [59] the configuration of the NB-IoT air interface parameters is investigated in order to optimize the route of a UAV that has to collect data from scattered terrestrial terminals.

Remarks — The path to bring LPWANs to LPGANs entails several challenges, including:

  • spectrum regulation harmonization, which may lead to harmful interference between satellite links and terrestrial links, especially for LPWANs based on unlicensed spectrum like LoRa;

  • modulation format, demodulation approaches, and Doppler-effect compensation enhancements;

  • architectural enhancements.

Additional details and directions of research are described in Section VI related to open challenges.

C. Final Remarks on Satellite IoT Enablers

It is evident that both 5G and LPWAN technologies are still under development and future solution are expected to ease the support for IoT services. In this framework, it is possible to identify some specific aspects of both technologies that could be considered as enablers for implementing satellite IoT solutions.

For 5G mobile networks, IoT clearly represents a relevant use case. Several technologies are being developed and standardized in such framework, including:

  1. network slicing, which represents the biggest advancement at system level, enabling to define and isolate overlay networks capable of adapting to different use cases (including, of course, IoT);

  2. O-RAN solutions, to better balance BS complexity and support interoperability;

  3. offloading solutions, capable of exploiting edge caches and computational capabilities;

  4. PEPs, to compensate high transmission delay;

  5. IoT-dedicated application- and transport-level protocols support, such as CoAP for reduced overhead communication;

  6. control plane optimization, to optimize signaling overhead in case of massive connections.

On the other hand, the typical LPWAN architecture is already evolving towards supporting longer transmission ranges, with the following features:

  1. potential extension of LoRa or NB-IoT modulations on the satellite link (requiring further adaptations of the standards, e.g., LoRa-E);

  2. flexibility at the physical layer seems to be moving in the direction of introducing SDR technologies, especially at the gateway level to promote interoperability.

SECTION V.

Industrial Focus

After the presentation of the technological enablers, let us now survey the most interesting industrial players in the satellite IoT arena. We will categorize these initiatives by technological enabler, following the presentation order of such enablers we adopted in the previous sections and keeping in mind that, at the time of writing, the standardization of 5G is still ongoing thus no initiatives can be ascribed under the 5G umbrella. Therefore, we will describe independent initiatives leveraging LEO satellites (see Section III-E) and LPWANs (see Section IV-B), distinguishing between LoRa-based initiatives and NB-IoT-based initiatives.

A. Independent LEO Satellite IoT

A number of new IoT initiatives are emerging in the satellite industry, also thanks to the push from the European Space Agency (ESA) and the National Aeronautics and Space Administration (NASA) for small satellite technologies (e.g., CubeSat). This new wave of start-ups is proposing IoT solutions relying on the newer low cost, low power satellite technologies to achieve truly global connectivity coverage. It is worth remarking that initiatives are typically opposed to incumbent (and expensive) satellite connectivity providers, which have already a well-established constellation of satellites on space and are able to provide IoT connectivity mainly targeting mobile and maritime environments or backhaul connectivity services for local area IoT networks relying on terrestrial standards. Most of the new space start-ups, instead, need to bring down their service costs for being able to enter in the market, and have therefore been quick in taking advantage of the new small satellite technologies. These, are intended for operating in the LEO thus allowing the use of low power modems to connect the ground sensors. By being able to provide low cost, low power, low bit-rate connection services, small satellites enable direct-to-satellite architectures, that allow to bypass local area IoT networks and ground IoT gateways.

1) Astrocast Nanosatellite Network [60]:

With a constellation of 80 LEO satellites using proprietary low-power L-band, this Swiss company provides low-latency (<15 min) transmission optimized for direct-to-satellite IoT applications, and supports bidirectional communications enabling acknowledgments, asset commands, and deployment of security patches and software updates (over-the-air software updates). The L-band allows smaller antenna, lower-cost RF components, better propagation (no rain fade, lower power requirements), fewer interference risks than other bands. To the best of the authors’ knowledge, Astrocast is the only initiative with access to L-band spectrum.

2) Myriota [61]:

This Australian company provides low-cost, low-power, secure direct-to-orbit satellite connectivity for the IoT thanks to a constellation of nanosatellites over ultra-high frequency (UHF) or very-high frequency (VHF) frequency bands. On average, four satellite passes per day are granted, with the MTD having approximately 9 minutes to send messages to the satellite in line of sight.

3) Kineis [62]:

It is a French company which uses a proprietary technology and chipset. A constellation of 25 nanosatellites is planned for 2022, with 6 satellites already on orbit for prototypes testing.

4) Kepler Communications [63]:

This Canadian company is developing 140 satellites to provide affordable, low-power, direct-to-satellite IoT connectivity within 2023. They feature bi-directional communications to ensure data acknowledgement and over-the-air firmware updates.

5) Swarm Technologies [64]:

This Silicon Valley startup currently has 9 satellites in orbit, but aims to deploy 150 satellites supporting direct-to-satellite IoT connectivity. Swarm’s hardware can be integrated with third-party devices and supports a variety of communications protocols.

B. LoRa-Based Satellite IoT: Lacuna Space

Lacuna Space [65] proposes a new ecosystem that allows other operators to co-exist, thus implementing a sort of “satellite roaming.” Their technology can be made available also to other partners. In the following, we will focus on the major aspects of the Lacuna Space system.

1) The Physical Layer:

The Lacuna Space system is a unidirectional system, where the ED can transmit a packet in uplink (to the satellites) but there is not a downlink from the satellites to the ED. A new LoRa modulation format is used by Lacuna Space, different from the usual one based on CSS. As pointed out in [66], this new modulation format

“[…] allows up to 1000 simultaneous transmissions/MHz [and it is] Resilient to other industrial, scientific, and medical (ISM) signals using intra-packet fast frequency hopping over a wide bandwidth.”

Such new modulation format is likely LoRa-E, which we mentioned in Section IV-B.

Furthermore, the Lacuna Space ED antennas are not simple dipoles as for usual LoRa devices, rather they are right-hand circular polarized antennas that irradiate toward the sky – see Fig. 7 [66]. It must be pointed out also that the Lacuna Space EDs are using for the uplink exactly the same frequency bands as the “terrestrial ED” (i.e., in Europe the 868 MHz band). In other words, the Lacuna Space EDs pretend to be terrestrial EDs and as such they are subject to the same regulations of terrestrial devices (i.e., Short Range Device according to [67]).

Fig. 7. - Lacuna Space ED disc-shaped antenna pointing to the sky (
${z}$
 direction).
Fig. 7.

Lacuna Space ED disc-shaped antenna pointing to the sky ({z} direction).

2) The Protocol Stack:

The Lacuna Space system leverages the usual LoRaWAN protocol stack; in fact, the identifier of the Lacuna Space network (LoRaWAN NetID) is C00028. Since the downlink is not available, the Lacuna Space ED exploits the Activation by Personalization approach during the Join phase – see [31] for the different methods of joining a LoRaWAN network. All of the roaming characteristics [31] are unchanged, enabling Lacuna Space EDs to interact with other networks as usual. A particular important scenario is the one when an ED associated with a “normal” terrestrial operator is roaming into the Lacuna Space network.

3) Architecture and System Considerations:

From the architectural point of view, the satellites of the Lacuna Space constellations can be considered as “flying gateways,” thus a direct-to-satellite communication is exploited in the uplink. As shown in Fig. 8, a Lacuna Space ED is sending a packet, at a certain time t_{0} , toward the sky and the packet is collected by a satellite. On the other hand, the Lacuna Space system exploits a certain number of Earth stations which act as “terrestrial gateways.” The Earth stations are exploited to receive, at time t_{1} , all the messages a given satellite has collected from the last pass over an Earth station. Each Earth station is the connected to a normal LoRaWAN back-end for further processing of the packets as usual.

Fig. 8. - Representation of typical Lacuna Space time displacement between uplink and downlink.
Fig. 8.

Representation of typical Lacuna Space time displacement between uplink and downlink.

Two important points must be remarked for the Lacuna Space architecture.

  1. The delay t_{1} - t_{0} is likely to be in the order of minutes, thus there may be severe implications in terms of additional latency for the protocol stack (e.g., timeout timers), even in case downlink transmission was possible.

  2. There is no ISL: all the packets collected by a satellite are delivered by the same satellite to an Earth station.

A final consideration should be done regarding the interference management. As shown in Fig. 9, a beam of a satellite con cover a large geographical area where a huge number of terrestrial EDs can be active. As a consequence, a lot of interference can be collected from the terrestrial ED. Furthermore, a significant number of satellite-server EDs could be present in the area covered by the beam. Hence, it is apparent that the Lacuna Space system needs a specific modulation which is resilient to the interference on one side and able to sustain a high capacity on the other. Regarding the capacity, of course, selecting a suitable modulation cannot be the only enabler: we guess some specific other modifications to the receiver must be made and some other techniques such as, quoting [66],

“[…] Lacuna cognitive radio technology to optimize spectrum usage.”

On these techniques, unfortunately, no public material is available at the time of writing, to the authors’ best knowledge.
Fig. 9. - Envisioned CubeSat coverage capability for Northern/Central Italy.
Fig. 9.

Envisioned CubeSat coverage capability for Northern/Central Italy.

C. NB-IoT-Based Satellite IoT

Even if the current 3GPP technologies do not fully support non-terrestrial IoT, various companies have been starting working around those limitations. In particular, companies like, e.g., Sateliot and MediaTek, actively support NB-IoT standardization process with the formalization of its non-terrestrial component expected in Release 17 of the 3GPP standard. The ESA has demonstrated interest as well, supporting initiatives like Sateliot and OQ, and collaborating with the Danish satellite communications waveform and test equipment specialist GateHouse Telecom A/S14 to enable a space-based NB-IoT network.

1) Sateliot [68]:

This Spanish satellite operator has recently signed a partnership agreement with GateHouse Telecom for delivering NB-IoT connectivity technology for a satellite IoT network. The solution is based on a constellation of small LEO satellites which will act as base stations. Sateliot programmed to extend their constellation to 16 satellites in orbit by the end of 2021, and services are expected to start operating in the third quarter of 2022. To this aim, Sateliot activity is also focused in designing and developing of the necessary adaptations to comply with the enhancements of NB-IoT in 5G.

Sateliot’s idea consists of adapting the NB-IoT protocol stack in order to allow MTDs to connect to Sateliot’s LEO satellite network whenever terrestrial coverage is not available or in areas where classic mobile communication infrastructure is not available due to communication challenges. Sateliot is also investigating the interworking of its LEO satellite constellation with mobile core network functionalities.

2) MediaTek [69]:

It is a satellite modem provider and has successfully completed a field trial that transfers NB-IoT data through an Inmarsat Alphasat L-band satellite, which is currently in a GEO. A bi-directional link was set from a MediaTek’s satellite-enabled standard NB-IoT device, to a commercial GEO satellite. MediaTek is an active contributor of 3GPP – see [70].

3) OQ Technology [71]:

Luxembourg-based OQ Technology is building a global LEO CubeSat satellite constellation dedicated for 5G IoT. OQ Technology addresses the satellite IoT solutions deployment on three steps:

  1. uploading OQ software on third-party satellite operator providing SDR-as-a-service connectivity;

  2. getting OQ payload hosted onto other provider’s satellites (through services offered by companies such as Spire Global and NanoAvionics);

  3. building and launching OQ own satellites.

OQ software is designed to be cross-layer compatible with different SDRs, so that it can be uploaded to third partners satellites that offer SDR as a service, standing that they operate in the same frequency and power range.

At the time of writing, OQ has just successfully tested NB-IoT waveforms and synchronization procedures using a third-party small satellite (i.e., the GOMX-4A and GOMX-4B from Danish GomSpace), trials on LEO satellites, working through the Doppler and range issues unique to non-stationary satellites (NB-IoT in LEO satellites is challenging, due to high Doppler and delay environments). The experimentation demonstrated that NB-IoT technology can be used on flying SDR payloads by uploading the waveforms to test the performance.

4) Ligado Networks [72]:

this USA-based company is also working towards adapting standard IoT protocols of LTE Cat-M and NB-IoT for satellite use. Ligado Networks has 40 MHz of spectrum licenses in the USA, nationwide portion of 1500–1700 MHz spectrum (L band). In April 2020, the company received Federal Communications Commission (FCC)’s approval for use of spectrum near the L bands used by Global Positioning System signals for 5G network deployment. Currently, they operates the SkyTerra one GEO satellite, covering all North America, to provide robust connectivity to small, low-power devices that are also served by terrestrial mobile networks, but their plan is to build a satellite-terrestrial network to support the emerging 5G market and IoT applications.

SECTION VI.

Open Challenges and Future Directions

In the previous sections, we identified the technological enablers of the satellite IoT paradigm provided by LPWANs and 5G, as well as the most prominent initiatives which are leveraging them to support a broader adoption of IoT services, thus creating new business opportunities and models. Those represent the current state of the art in terms of standardization and ongoing initiatives. Nevertheless, research is looking more in the future, towards technologies capable of providing further advances and revolutionary solutions. In this section, we focus on the analysis of the gaps that remain both from the technological point of view and standardization/regulation point of view, starting from the remarks provided in Section IV-C, in order to fully unleash the potential of satellite IoT. Indeed, the next subsection provides an overview of potential novel approaches at the different layers of the protocol stack.

A. Innovative Approaches to Tackle Technical Challenges

In the following, we categorize the innovative approaches into PHY design, antenna design, medium access, network and higher layer protocols, and edge computing.

1) Physical Layer Design:

In the recent literature, some remarkable papers have been published about waveform and transmission formats for satellite-aided IoT. Standard LoRa CSS waveform is not suitable for LEO satellite IoT communication due to the high time-domain cross-correlation values between the chirps with different starting frequencies. In order to overcome this issue, [73] introduces a low cross-correlation symmetry chirp signal (SCS) that noticeably improves PHY-layer performance. The acquisition of a SCS is then discussed in [74]. The paper of Qian et al. [75] partially revised the approach of [73], by proposing another typology of chirp signal for satellite IoT transmission, namely: the asymmetry chirp signal (ACS), which is obtained by a modification of SCS. ACS effectively solves the issue noticed for SCS and related to high peaks of spectral cross-correlation in presence of high satellite Doppler shift. Another paper dealing with CSS modulation for satellite IoT is [76]. In particular a new methodology of unipolar coding is proposed in [76] to allow efficient random multiple access to a large number of transmitting devices. A relevant issue in CSS modulation is the vulnerability to the Doppler effect. LEO satellites are characterized by large Doppler shift and this could impair link performance. For this reason, the authors of [77] proposed a folded chirp-rate shift keying modulation, characterized by high resilience to Doppler and other unexpected frequency drifts. As anticipated in Section IV-B, it is worth mentioning that a new modulation format called LoRa-E has been recently announced to overcome the limitation of the default LoRa [54]. It relies on two bit rates (162 and 325 bps), and consists of a frequency hopping spread spectrum (FHSS) modulation. In each transmission, several replicas of the packet header are transmitted over a selected channel, in distinct sub-channels according to the ED’s channel hopping sequence. After the header, the payload is split into fragments which are sent consecutively in each sub-channel as per the hopping sequence, without repetition. This approach increases the complexity at the receiver side, but allows multiple EDs to transmit at the same time without colliding. We recall that LoRa-E will be used only in uplink, while traditional LoRa will be exploited in downlink thanks to the compatibility between these two modulations.

A different transmission waveform has been considered by Doré and Berg in [78], where turbo-coded frequency-shift keying (namely, Turbo-FSK) is employed for narrowband satellite IoT. Turbo-FSK has the main advantage of being a constant-envelope modulation. This fact definitely increases the amplifier efficiency that is a must in power-constrained satellite IoT connection. In the recent paper of Clazzer and Munari [79], two methodologies of random multiple access for IoT via satellite have been analyzed, namely: the Spread Spectrum transmission with ALOHA and the replication of transmitted packets in order to obtain time diversity. In both cases, the receiver employs successive interference cancellation to resolve the collisions. The comparison revealed the tradeoff between spectral efficiency provided by the repetition-based scheme and the interference reduction of Spread Spectrum techniques. The final conclusion of [79] is to consider repetition-based scheme as a valuable solution for higher data rates IoT applications, while Spread Spectrum confirmed its robustness in LoRa-like narrowband use cases. Another solution for spectrally-efficient multiple access in Satellite IoT systems has been very recently proposed by Su et al. in [80], based on random interleaving multiplexing. Substantially, the bit and symbol level interleavers are randomly selected by active users from sets of interleavers. Such a choice allows to differentiate signals of different users sharing a subframe. MAP detectors can be employed to decode multiple superimposed packets. Results shown in [80] evidenced a throughput of 2 bits/symbol, which largely outperforms conventional random access protocols. In the same framework of spectrally-efficient multiplexing methodologies for Satellite IoT, it is worth citing the paper of Hu et al. [81], where different users can be simultaneously transmit their messages over the same bandwidth using constellation coding/decoding. Substantially, the user data are transmitted by means of specific symbols occupying precise positions in higher-order modulation constellations (constellation coding). At the receiver side, constellation decoding should demap the correct symbol to the correct user. The capacity improvement of such a methodology is greatly noticeable, with a price to be paid in terms of non-trivial receiver complexity.

Furthermore, it is worth mentioning the work of Cluzel et al. [82], where a physical layer abstraction is used to evaluate the Bit-Error-Rate (BER) and the Packet-Error-Rate (PER) of a LEO satellite system for IoT, using a time-frequency ALOHA scheme for multiple access management. The PHY-layer abstraction relies on the calculation of an equivalent Signal-to-Noise ratio, based of mutual information, which allows at analytically computing both BER and PER in any collided scenarios.

Finally, we remark that PHY security, exploiting the randomness of the wireless channel to perform authentication, is also an actual research topic in the context of space information networks. For the sake of space constraints, we invite the interested reader to the recent survey about this topic [83].

2) Antenna Design:

The antenna design is a challenging aspect. Clearly, the utilized frequency band and the form factor of the satellite constitute contrasting objectives for the size of the antenna. On the one hand, the lower spectrum requires larger antennas and higher spectrum requires more power and directivity [32]. On the other hand, the antenna has to fit a form factor given by the satellite.

In the authors’ opinion, the discussion related to the antenna design deserve a wider scope, not limited to the satellites for IoT only. Nevertheless, some papers published in the literature explicitly address the target of the efficient antenna design for satellite-IoT systems. In [84], Sanil et al. propose a multi-band circularly polarized microstrip antenna operating at 5.8 GHz, 6.76 GHz, and 8.4 GHz with a fractional bandwidth of 170 MHz, 335 MHz, and 560 MHz respectively. Another recent paper focuses [85] the attention on design and implementation of an X-band phased antenna array. Another work deserving a citation is [86]. Here, the latency involved by mechanical and electronic scanning for ground-station antennas, transmitting from satellite to satellite, is assessed. Results achieved in [86] are clearly in favor of the electronic solution.

3) SDR PHY-Layer Implementation:

It is also worth remarking that future terrestrial and space networks will be characterized by flexibility and dynamic reconfigurability. In such a framework, SDR will play a key role, e.g., by allowing satellites to reconfigure their mission and allocate part of their available frequency bands to new applications [32, Sec. 4.1.2]. Some examples of this have been already shown in the recent literature. In [87], the SDR implementation of a gateway for satellite-based IoT applications was described in detail. GNU Radio open-source platform allowed the effective testing of the overall gateway functionalities at a “proof-of-concept” level. More precisely, the software layer implements the IoT protocols that coordinate the E2E device-to-device or device-to-user communication between the satellite and terrestrial interface, and a multi-band spectrum sensing procedure for detecting the wireless activity in the RF bands specific for each considered short-range terrestrial communication standard.

4) Medium Access:

In order to compete with existing solutions, a bidirectional communication must be ensured by satellite IoT systems.

As far as uplink is concerned, the direct-to-satellite architecture seems to be the preferred one, with various studies investigating novel spectrum access approaches for MTD-to-satellite communication. Typically, these approaches are variations of the Aloha protocol. For example, [88] proposes a distributed method to tune the transmission probability of each MTD according to current traffic load at the serving satellite. The suitability of direct-sequence spread spectrum Aloha to LEO-based satellite IoT system is analyzed in [89]. An irregular repetition slotted Aloha protocol for random access in a power limited scenario is proposed in [90]. Nevertheless, indirect-to-satellite-compliant solutions are also proposed to optimize the energy consumption of the data collection process [91].

On the other hand, we observe that an effective downlink communication (from satellites towards MTDs) may be hard to achieve as well, because of synchronization and interference issues between the transmitter and the receiver. For example, in NB-IoT bidirectional communication is needed for RRM purposes, but the delay introduced by the satellite links will possibly require to modify time advance mechanisms. Even LoRaWAN EDs not requiring acknowledgement for uplink transmissions need bidirectional communication for joining the network, thus it must be supported by satellite IoT systems. In this regards, on the contrary of uplink transmissions, the indirect-to-satellite option seems more suitable, because it can rely on a traditional satellite link between the satellite and the gateway, which is seen as a “peer” of a ground station by the satellite, thus allowing to leverage well-known satellite MAC protocols [16, Sec. III-A]. Another possible solution could be based on grouping of MTDs according to their location, channel characteristics, and traffic features to alleviate the signaling overhead, as envisioned in [16, Sec. V].

It is finally worth mentioning that some recent (at the time of writing) studies appeared in literature have been characterizing the medium access across satellite clusters communicating via ISLs. These studies have been addressing the connection establishment [92], the information aging against the number of hops [93], and the average delay [94].

Interested readers may find further references about the medium access techniques and future challenges for satellite systems in [95], [96].

5) Network and Higher-Layer Protocols:

A relevant technological challenge is represented by the need for IoT-oriented protocol stacks capable of going beyond TCP/IP limitations. In particular.

  • The direct use of unmodified mainstream protocols, such as HTTP/1 and TCP, is challenging in resource limited IoT devices (e.g., CPU, memory, power) and constrained networks (e.g., high-latency, low-power, lossy). As mentioned earlier in the paper, the CoAP can be used in LPWANs for IoT services. However, the protocol is still open to the implementation of a Go-Back-N ARQ to make more efficient use of the available resources and reduce the delivery delay in high-latency networks such as satellite;

  • To overcome some of the limitations of HTTP/1, the IETF has standardized HTTP/2. HTTP/2 runs over TCP, but the IETF is currently standardizing HTTP/2 support over QUIC, a new UDP-based, stream-multiplexing, always-encrypted transport protocol focused on minimizing application latency. This transition could lead to the standardization of an HTTP/3 supported by QUIC, with the advantage that QUIC provides reliable data transfer and pluggable congestion control, which can be optimized for IoT scenarios.

  • More advanced proposals, starting from the assumption that the IP stack is too greedy to be transported by IoT resource limited networks, propose alternative back-compatible solutions capable of efficiently compress and fragment data in order to better support the IoT operating environment. A relevant example in this class of solutions is represented by Static Context Header Compression (SCHC), which promises a technology for compressing and fragmenting messages exchanged on LPWANs to ensure full compatibility of these messages with Internet protocols [97]. SCHC allows the interoperability of the devices to the different LPWANs, and more broadly to the applications deployed on the Internet. In brief, SCHC is a compression mechanism to reduce CoAP/IPv6 headers to just a few bytes, which can then be transported over any LPWAN networks. SCHC prevents synchronization between elements communicating on the network, and this is one of the operations that consumes most bandwidth. This can be achieved since in LPWANs networks, the nature of data flows is highly predictable. SCHC is proposed as an open IETF standard.

  • PEPs represent already a reality in satellite communications. However, their presence might increase the degrees of freedom in designing solutions for satellite IoT, in order to tackle the complexity of managing the massive number of devices and connections to handle.

A new frontier in wireless networking is represented by network coding [98]. Network coding demonstrated to be very helpful for multi-hop networks. In particular, random linear network coding (RLNC) has the attractive and distinctive feature that is the only code that circumvent the need the encoding and decoding in each hop at transport layer [99]. In the paper of Marcano and Jacobsen [100], the advantages of employing RLNC in small satellite constellations for IoT are demonstrated in different application scenarios, namely: event detection/localization, asset tracking and data sensor aggregation. Numerical simulations of [100] show a 50–65% delay reduction for RLNC with respect to the state-of-the-art Reliable Datagram Protocol (RDP), used for comparison in all the considered scenarios. Other papers considers the use of network coding for efficient Hybrid-Automatic Repeat Request (HARQ) implementations in satellite-based IoT systems. For instance, in [101], Zi et al. proposed a novel network-coded HARQ approach for a bidirectional relaying scenario, where the two terrestrial sensors cannot connect for geographical reason and the satellite works as relay node.

Complex systems like satellite-based IoT can profitably exploit software agility provided by software-defined networking (SDN). In [38], Kak et al. assessed the performance of an SDN-based Internet of Space Things network. The system architecture is based on swarms of CubeSat units interconnected by ISLs. SDN orchestration and network-function virtualization (NFV) are used to manage the aggregation of the information flows from ground to satellite and vice-versa in efficient manner, with high throughput and minimized latency. The numerical simulations fully evidenced the potentials of such advanced networking techniques in satellite-based IoT. In another recent paper [102], the seamless integration of terrestrial and satellite networks for IoT is investigated. In particular, a self-organization satellite-terrestrial integrated system (SSTIS) is discussed, composed by a perception layer, a cognition layer and an intelligent layer. Such system, whose architecture is shown in Fig. 10, integrates IoT, SDN, and network functions virtualization technologies to achieve self-monitoring, crisis forecasting, and optimal control.

Fig. 10. - Architecture of the SDN-based self-organization satellite-terrestrial integrated system (SSTIS) of [102] (courtesy by [102]).
Fig. 10.

Architecture of the SDN-based self-organization satellite-terrestrial integrated system (SSTIS) of [102] (courtesy by [102]).

Specific interest and related open challenges are related to the emerging concept of network slicing. Network slicing represents a key technological advancement that will bring unprecedented levels of flexibility and adaptability in network architectures. By leveraging on the diffusion of SDN and NFV implementation, network slicing allows to build virtual self-contained networks on shared pool of computing, storage and network resources. The ultimate goal of network slicing is to reduce the complexity in configuring and managing the networking infrastructure by associating different slices to different types of applications characterized by similar quality of service (QoS) requirements. Indeed, 5G will dedicate a specific slice to mMTC, which could be considered the reference slice to be adapted to satellite IoT solutions described in this paper.

However, while functionalities and interfaces are being standardized, e.g., by the 3GPP, the relevant open challenge is related to understanding the specific QoS requirements of long range IoT services and to define how to configure and provision a proper mMTC slice. Those represent open challenges in this area and promising research topics for the future.

6) Edge Computing:

Research is continuously bringing forward the benefits of edge computing when combined with IoT, allowing local pre-processing the collected data before transmission or at an intermediary stage, such as a gateway or an radio access point. Advanced schemas allow to pre-process the data in conjunction with the data collected from other sensors, thus allowing to exploit existing correlations. Advantages can be various, from diminishing the amount of data transmitted in the backhaul, to the enforcement of privacy-preserving algorithms. Moreover, new approaches to machine learning, such as federated approaches, go a step further in this direction.

Recent research works show that the interest of bringing the edge computing paradigm to space is growing. In space, satellites and IoT nodes operate as a pipeline: data is collected from sensors in the ground, stored, and delivered to the ground station when it is on coverage. From satellite perspective, there are two options of integrating edge computing in this pipeline: at the IoT nodes, or in orbit, at the satellite. For example, satellite constellations be thought operating as computational pipelines: studies in [103] for camera equipped satellites sensing the Earth show advantages both in terms of architecture infrastructure and in processing latency for this approach. But the edge computing paradigm does not only shift processing nearer to the location where the data is generated and/or consumed, but also introduce the application virtualization paradigm in, thus pushing the technology towards more standardized approaches. The work carried on by European Telecommunications Standards Institute (ETSI) multi-access edge computing (MEC) and 3GPP leverages on virtualization for creating a standardized framework supporting edge computing. A body of standardized interfaces has been created for allowing dynamic deployment of applications and allocation of resources. Additionally, an application framework has been defined for service application programming interfaces, envisaged to better support the apps and exploit the edge node characteristics [104].

Traditional satellites are designed and optimized for specific applications, and therefore are highly customized. Embracing the MEC paradigm presupposes therefore a paradigm shift. For the MEC framework being adapted to satellites we have to rethink the role both the satellite node and the ground station, taking into account the limited on-board resources and communications specific characteristics, specially when it comes to small satellites. In Fig. 11, we draw an example on how the ETSI MEC elements can be mapped into the satellite scenario, envisaging a lighter set of functionalities on the satellite, and leaving the control and orchestration on the ground station. The possibility of dynamically loading software in orbit, and of providing services coordinated with the cloud is explored in [105] where the authors propose to adopt a standardized hardware platform combined with a fault-tolerant expandable satellite operation system in order evolve the satellite into space edge computing nodes. Indeed the solution envisioned by the authors aims not only to allow the on-boarding of apps on the satellite, but also the possibility of sharing the free on-board resources with other satellites.

Fig. 11. - Envisioned architecture for the application of edge computing to satellite IoT networks.
Fig. 11.

Envisioned architecture for the application of edge computing to satellite IoT networks.

Discussion — A mapping between the envisioned approaches described in the previous sections and potential impact on the different remote IoT approaches in provided by Tab. VI. The table aims at summarizing the concepts described in the previous sections by tentatively associating the different technologies to different time scaled in the evolution of the Internet of Remote Things. The table addresses three scenarios: independent LEO, 5G-enabled satellite IoT, and LoRa-based satellite IoT solutions. The potential advances are classified between short term and long term, the former representing almost mature/emerging solutions (e.g., that might be deployed by industry), while the latter representing a vision for future research and development. Here are the main points that emerge from such analysis.

  • PHY: in the short term, the different scenarios are expect to develop different approaches to improve spectrum utilization and agility, by either deploying multiple RF interfaces or deploying SDR technology. However, the long term seems to converge towards a more uniform setup consisting of a software-defined RF interface capable of providing multi-standard support and runtime reconfiguration.

  • Medium access: being tightly coupled with physical layer and typically adapted to the application profile, medium access control schemes are expected to maintain a certain level of heterogeneity both in the short as well as in the long term. Relevant upgrades are expected in 5G and beyond mobile networks including in the short term IoT-oriented optimization of the control plane functionalities, and in the long term most advanced mechanisms such as non-orthogonal multiple access schemes for 5G and non-pure-Aloha-based approaches for LoRa.

  • Network and higher layer protocols: the impact of network softwarization and virtualization will most likely impact in the network layer and above that. On one side, emerging solutions targeted to improve network performance in IoT scenarios (SCHC, performance enhancing proxies, network slicing) will become a reality in the short term, depending on the respective reference architectures. In the long term, most likely the Internet of Remote Things will incorporate novel management paradigms such as self-organization networks and SDN. In mobile networks, the 5G SBA will enable higher degrees of freedom in controlling, configuring and optimizing the mMTC network slices.

  • Edge computing: this area is expected to impact and be affected by remote IoT scenarios at different time scales. In the short term, solutions based on independent LEO as well as 5G might integrate edge computing to facilitate computation offloading or enabling deployment of functionalities at the edge of the network (like in the case of 5G and ETSI MEC). This might reduce in the case of LoRa to support flexible placement of the gateway functionality. In the long term, the possibility to host remote containerized solutions will spread across all the scenarios, to enable better placements of key functionalities inside the network infrastructure itself as well as to optimize the allocation of task among network elements (e.g., BS splitting).

TABLE VI Mapping Between the Envisioned Technical Innovations and the Industrial Initiatives of Applicability
Table VI- 
Mapping Between the Envisioned Technical Innovations and the Industrial Initiatives of Applicability

B. Future Standardization and Regulation

As witnessed for the mobile network technology, the key for the success of the satellite IoT systems passes through standardization and regulations. In fact, even if heterogeneity has been well-investigated in the context of traditional communication networks as far as architectures and protocols are concerned, the satellite IoT pledges to be a particularly challenging communication paradigm, thus the integration and interoperability between the involved systems will be a crucial element to achieve coverage and capacity goals [32].

In the following, we identify the main involved bodies and the envisioned challenges regarding spectrum harmonization across continents and the interworking with terrestrial IoT networks.

1) Involved Bodies:

Several standardization and regulatory bodies are related to this field.

  • As far as radio-communications are concerned, the International Telecommunication Union (ITU) represents the framework for the international regulation of spectrum utilization [106]. Two kinds of organizations, that are, regional organizations and specialized international organizations, are involved in spectrum management together with ITU, at either regional or global level. At a regional level, these organizations gather administrations to establish common positions in preparation for ITU decisions, with the aim facilitating the coordinated introduction of new services. This is the case in particular for the European Conference of Postal and Telecommunications Administrations (CEPT), and to a lesser extent for the Inter-American Telecommunication Commission (CITEL), the Asia-Pacific Telecommunity (APT) and the Arab Council of Ministers for Telecommunication and Information. At both global and regional levels, specialized international organizations also exist in sectors of activity that use radio communications and are therefore dependent on spectrum availability, e.g., civil aviation, the maritime sector, meteorology, broadcasting.

  • Of special interest is the International Amateur Radio Union (IARU), the worldwide federation of national amateur radio organizations. IARU is recognized at ITU level and some frequency bands have assigned to the amateur radio via satellite as primary service. We remember that a “primary service” in the ITU terminology is a service which can claim protection fro other services. The allocation of frequency bands to the amateur radio as a primary service has been made in light of the importance of the amateur service in case of natural disasters; historically, the amateur radio services have demonstrated to be the only available telecommunication service surviving during some natural disasters. Many small satellites missions have been launched and are planned as “amateur radio” and in particular, most of them use the “cubesat” satellites; for amateur radio cubesat satellites (usually very few satellites) the authorization procedures by ITU are simplified.

  • The International Organization for Standardization (ISO) is working as well on the standardization of space systems; this standardization is carried out in ISO/TC 20/SC 13 (technical committee 20, sub–committee 13) dealing with “Space data and information transfer systems.”

  • The Consultative Committee for Space Data Systems (CCSDS) is a forum aiming to promote “future data system interoperability.” Members of the CCSDS are all major space agencies (Italy, France, China, Canada, Germany, Europe, Russia, Brazil, Japan, USA, and U.K.) with many other agencies acting as “observers.” The CCSDS in its charter mentions that one it its major purposes is “to maintain cognizance of other international standardization activities that may have direct impact on the design or operation of space mission data systems.” The CCSDS is publishing “recommended standards” as well as many other types of documents. It is worthwhile to note that CCSDS established a liaison with the aforementioned ISO/TC 20/SC 13.

  • The ETSI developed many standards regarding the satellite communications, the most important of which are those related to the Digital Video Broadcasting (DVB). Despite the reference to the “broadcasting,” which is the original target application, those standards have evolved to support interactive services including Internet access. In particular, the DVB-RCS2 [107] standard (Digital Video Broadcasting; Second Generation DVB Interactive Satellite System) includes the so-called “return link” via satellite, i.e., the link from Earth to space. The DVB standards, however, do not have as a specific target the IoT but are directed more towards generic Internet access. The S-MIM standard by ETSI is, quoting [108],

    “[…] especially designed to provide ubiquitous messaging services over S-band GEO satellites using low-power terminals.”

    The S-MIM ETSI standards, however, is reusing the W-CDMA 3G technology and did not achieved commercial success. Anyway, ETSI, via its technical committee Satellite Earth Stations and Systems (SES) is still working on the satellite related standards such as the one of the “Family SL” [109]. Still, this family of standards does not address directly the IoT.

  • The 3GPP has been including significant contributions aimed at including satellite communications techniques to support seamless 5G connectivity [15], [110]. The study item phase on the so-called NTNs have been focusing on integration of both satellites and airborne BSs into the terrestrial 5G network for extending network coverage, improve service continuity, and implement robust multicasting [47]–​[49]. At the time of writing, the 3GPP is working on the Release 17 of the specifications (which is expected to be rolled out by the end of 2021/beginning of 2022) in order to i) provide initial studies results for both NB-IoT and LTE Cat-M for satellites to provide IoT connectivity in remote areas with low or no cellular connectivity [70] and ii) better support NTN services in terms of mobility management and QoS impact [111]. Therefore, it is worth mentioning that we can expect that satellite links in 5G will first be used to support multimedia use cases and broadcasting services, while support for IoT use cases may take a while (in terms of technology advancements and needed investments).

2) Spectrum Harmonization:

A prominent technical challenge to enable a worldwide support of satellite IoT system is related to spectrum. If fact, a device, without knowing its position, should be able to adapt its transceiver to the correct regional regulations (e.g., switching from CEPT to FCC).

In the last ITU World Radiocommunication Conference 2019 [112], the agenda for the next ITU World Radiocommunication Conference 2023 has been set [113]. In that agenda, item 1.18 refers to “[…] studies relating to spectrum needs and potential new allocations to the mobile satellite service in the frequency bands 1695–1710 MHz, 2010–2025 MHz, 3300–3315 MHz and 3385–3400 MHz for future development of narrowband mobile-satellite systems.” We note that the band 2010–2025 MHz is specifically mentioned in [113] for Region 1. The CEPT report [32] timely addresses the work to be done towards ITU World Radiocommunication Conference 2023. Within the UHF range, the 400.15–401 MHz band is available for downlink usage within the CEPT, according to harmonization conditions set out in ERC Decisions (99)05 and (99)06.

In [114], a report on decisions taken at the last ITU World Radiocommunication Conference 2019 is summarized. Quoting [114],

“The required spectrum for TT&C was estimated to be less than 2.5 MHz for downlink and less than 1 MHz for uplink. Consequently, the study groups conducted sharing studies in various bands which yield that no new allocations are suitable for small satellite TT&C [Tracking, Telemetry & Control] on a co-channel sharing basis.”

Quoting again [114],

“The 148–149.9 MHz uplink band has been opened for short-duration missions; however, it is disputable whether this band will be favored by small satellite developers. From a developer’s perspective, the results of WRC-19 for new TT&C allocations are not satisfying. New allocations in UHF would have been favored.”

As a conclusion, the issue of the frequency allocation for small satellites like Cubesat is still open. The 148–149.9 MHz uplink band implies that the antennas for the MTDs would have a dimension too big for the IoT devices.

Let us remark that the CEPT scope is the free market circulation of equipment into the countries that are members of CEPT. The big issue, however, is that non-geostationary satellites (like the CubeSat) fly all over the globe so achieving the “approval” from CEPT is a necessary step but not sufficient. The big step for companies that want to operate a CubeSat network is to achieve the approval from ITU. As explained in [115],

“[…] small satellites generally follow the procedure for non-geostationary satellite networks not subject to coordination, which starts with the submission of the Advanced Public Information (API).”

The whole procedure of authorization takes a minimum of 9 months and a maximum of 7 years, depending on the accuracy of submitted information, on the reactions (e.g., from national regulators) on the pre-publication of the information about the new satellite networks, and so on.

It is worth remarking that, rephrasing from [115]:

  1. there are many new perspective operators and many of them are not fully aware of the procedures for the authorizations;

  2. there is large number of perspective operators and this overwhelms the regulators, since procedure are cumbersome and lengthy;

  3. the operations of new networks involve

    • large constellations and complex systems;

    • regular replenishment/augmentation (the operational life of CubeSat is lower compared to conventional satellites);

    • global coverage (which then involve all national authorities and a large number of stakeholders that may raise concerns on the authorizations).

All of these issues highlight the importance of spectrum coordination and allocation aspects, which are often neglected in the scientific literature. The need for a streamlined procedure to put in operation CubeSat (or, more in general, LEO small satellites) is evident, to support the market growth of this type of networks. The establishment of this new streamlined procedure, however, is not at all simple to achieve and in the Appendix we provide a simplified flow for the current authorization procedure, in order to give the reader a feeling of how complicated this procedure is currently [114].

3) Interworking With Terrestrial IoT Networks:

A reasonable approach to address the integration/interoperability issues in satellite IoT is to distinguish between lower layers (PHY and MAC) and higher layers, because the design of the former ones is strictly dependent on the RAT, while the operation of the latter ones is quite technology-independent [16].

Regarding lower layers, satellite and terrestrial IoT networks may either be keep separate at the terminal side by utilizing two distinct chips operating on different frequency ranges, or they may be fully integrating thanks to a single chip [32, Sec. 4.2]. Clearly, with the former solution, the satellite IoT system would work on two parallel communication paths, while the latter approach allows a more efficient spectrum utilization at the cost of an increased implementation complexity of both the front-end and the core network architecture.

On the other hand, higher-layers design should consider the limitations of MTD capabilities and the constraints of satellite links, thus approaches like the reduction of the levels of encapsulation for IPv6 packets performed by the DVB-RCS2 standard are advised by [16], [33] in the context of satellite IoT.

Efforts to enable interoperability and full integration between the satellite side and the terrestrial side are also encouraged by [17], which refers to the outcomes provided by [41] and [37] regarding the system design of satellite IoT solutions based on NB-IoT and LoRa, respectively. According to [32], hybrid terminals could benefit from further standardization work to ease the interoperability between new IoT technologies.

C. Business Development Opportunities

We started this survey by citing recent market studies that confirm the forecast of a fast growing of the IoT market in the next few years [1]. Nevertheless, while the share of unlicensed spectrum RAT has been constantly increasing, the diffusion of IoT systems operating on licensed frequency bands has not met the original expectation (yet) [29], [116] – at least, waiting the (forecast) boost provided by the upcoming 5GS roll-outs [24].

In this context, the satellite IoT paradigm is expected to guarantee a global terrestrial IoT coverage, thus playing an important role in fulfilling the market demands of various services and vertical industries such as, e.g., smart grid, environmental monitoring, and emergency management [13], [16]. According to [32], a recent market study by the United Kingdom Space Agency estimates that the total amount of devices connected by satellite networks could grow from 3.16 million in 2015 to 5.97 million by 2025. Moreover, from a spectrum utilization perspective, a large majority (i.e., 93%) of the devices will exploit frequency bands below 3 GHz due to the narrowband nature of most satellite IoT applications.

As far as capital expenditure and operational costs are concerned, small LEO satellites will likely be more and more preferred instead of GEO to minimize the costs of deploying IoT-supporting satellites [13]. Moreover, in order to reduce the complexity of a satellite back-hauling based on ISLs, and thus its deployment cost which would curb the adoption of such paradigm, appropriate satellite trajectories and constellations should be designed to ensure line of sight between IoT-supporting satellites and the ground station network. Finally, there is the option of eventually providing an incomplete coverage on the Earth, rather concentrating it on localized areas on Earth according to the specific IoT use case [33], thus reducing the deployment costs.

SECTION VII.

Conclusion

The IoT is gaining momentum due to the market expectations and the rapid development of novel solutions in the field of communications. This paper focused on long-range IoT scenarios, by surveying the potentially available technologies today and providing examples of research and development efforts that are building concrete solutions for satellite supported IoT. Indeed, this area represents a big challenge, both for industry as well as for research, since short term goals and long-term visions are both required to steer the ongoing efforts towards interoperable and affordable solutions, capable of being deployed everywhere in the World. Open challenges are, as well as hints and potential solutions that will affect the future of satellite IoT systems have been discussed in the paper. The viability of satellite technologies for IoT, mainly supported by low-orbit constellations, has been assessed in the framework of the specific requirements imposed by standardization committees, taking the eyes open toward a renewed, future, vision of IoT technologies, driven by emerging standards.

ACKNOWLEDGMENT

The authors would like to thank Dr. Doriana Guiducci and Dr. Peter Farris of the ECO-CEPT Office in Copenhagen, Denmark, for the invaluable support in identifying the latest ITU and CEPT regulations. The research activities presented in this paper fall within the field of interest of the IEEE AESS technical panel on Glue Technologies for Space Systems.

Appendix

ITU Authorization Process

A major milestone in the Satellite Radio Regulation has been the “Legal Framework for Spectrum Access/Use United Nations Outer Space Treaty” dating back to 1967. ITU is recognized in this treaty as the specialized agency responsible for orbits and spectrum as well as for the procedures and the “book keeping” of which systems are allowed to operate. ITU Constitution Articles 44 and 45 [117] form the basis for the Legal Framework for Spectrum Access and Use.

Article 44

Use of the Radio-Frequency Spectrum and of the Geostationary-Satellite and Other Satellite Orbits

  1. Member States shall endeavor to limit the number of frequencies and the spectrum used to the minimum essential to provide in a satisfactory manner the necessary services. To that end, they shall endeavor to apply the latest technical advances as soon as possible.

  2. In using frequency bands for radio services, Member States shall bear in mind that radio frequencies and any associated orbits, including the geostationary-satellite orbit, are limited natural resources and that they must be used rationally, efficiently and economically, in conformity with the provisions of the Radio Regulations, so that countries or groups of countries may have equitable access to those orbits and frequencies, taking into account the special needs of the developing countries and the geographical situation of particular countries.

Article 45

Harmful Interference

  1. All stations, whatever their purpose, must be established and operated in such a manner as not to cause harmful interference to the radio services or communications of other Member States or of recognized operating agencies, or of other duly authorized operating agencies which carry on a radio service, and which operate in accordance with the provisions of the Radio Regulations.

  2. Each Member State undertakes to require the operating agencies which it recognizes and the other operating agencies duly authorized for this purpose to observe the provisions of No. 197 above.

  3. Further, the Member States recognize the necessity of taking all practicable steps to prevent the operation of electrical apparatus and installations of all kinds from causing harmful interference to the radio services or communications mentioned in No. 197 above.

To this end a quite lengthy and cumbersome procedure is in place to put in orbit and start services for a satellite network. With reference to Fig. 12, the process for satellite networks not subject to coordination starts with the prospective satellite operator to disclose and get the approval from the local regulatory agencies (this part is not shown in the figure). After that phase, the prospective operator, through the regional agency, submits a substantial amount of information about the satellite system via a specific ITU format (actually obtained through a software provided by ITU) called API. The first part of the API, after ITU revision for completeness, is then published by ITU. After this publication, the local administrations can file to the ITU possible complaints, observations, requests for clarifications about the prospective system. When the time to file these observation is complete and no roadblocks are found, ITU publish a second part of the API and a new revision cycle starts. If everything goes well the perspective operator get the permission to operate the system. As one can see in Fig. 12, the process does not always proceed smoothly, and it can take up to 7 years to be completed.

Fig. 12. - Graphical explanation of the ITU authorization process of prospective satellite systems.
Fig. 12.

Graphical explanation of the ITU authorization process of prospective satellite systems.

In the context of ITU regulations the World Radiocommunication Conference held in 2019 a key milestone was achieved. The resolution 248 related to “Studies relating to spectrum needs and potential new allocations to the mobile satellite service in the frequency bands 1695–1710 MHz, 2010–2025 MHz, 3300–3315 MHz, and 3385–3400 MHz for future development of narrowband mobile-satellite systems” was adopted. It is of particular importance that these studies “are to be limited to those systems with space stations that have a maximum EIRP of 27 dBW or less, with a beam-width of no more than 120 degrees, and earth stations that individually communicate no more than once every 15 minutes, for no more than 4 seconds at a time, with a maximum EIRP of 7 dBW.” The particular importance of this resolution is that it address the IoT market since the type of communication is infrequent (no more than once every 15 minutes) and the packets are short (4 seconds).

It is clear that further work is needed in order to streamline the procedure to not hinder the booming market of small satellites for IoT.

Acronyms

AbbreviationExpansion
2G

second-generation

3G

third-generation

3GPP

Third Generation Partnership Project

4G

fourth-generation

5G

fifth-generation

5GS

5G System

ACS

asymmetry chirp signal

API

Advanced Public Information

APT

Asia-Pacific Telecommunity

BBU

baseband unit

BPSK

binary phase shift keying

BS

base station

CAGR

compound annual growth rate

CCSDS

Consultative Committee for Space Data Systems

CEPT

European Conference of Postal and Telecommunications Administrations

CITEL

Inter-American Telecommunication Commission

CoAP

Constrained Application Protocol

CSS

chirp spread spectrum

DVB

Digital Video Broadcasting

E2E

end-to-end

EC-GSM-IoT

Extended Coverage GSM for IoT

ED

end device

EIRP

effective irradiated power

ELO

Eutelsat LEO for objects

EPS

Evolved Packet System

ESA

European Space Agency

ETSI

European Telecommunications Standards Institute

EU

European Union

FCC

Federal Communications Commission

FHSS

frequency hopping spread spectrum

FSK

frequency-shift keying

FSO

free-space optics

GEO

geosynchronous equatorial orbit

GFSK

Gaussian frequency shift keying

GNSS

global navigation satellite system

GPRS

General Packet Radio Service

GSM

Global System for Mobile Communications

GSO

geosynchronous orbit

HEO

highly elliptical orbit

HTTP

Hypertext Transfer Protocol

IEEE

Institute of Electrical and Electronics Engineers

IETF

Internet Engineering Task Force

IIoT

industrial IoT

IoT

Internet of Things

IP

Internet Protocol

ISL

inter-satellite link

ISM

industrial, scientific, and medical

ISO

International Organization for Standardization

ITU

International Telecommunication Union

KPI

key performance indicator

LEO

low-Earth orbit

LoRa

Long-Range™

LPGAN

low-power global area network

LPWAN

low-power wide area network

LTE

Long-Term Evolution

MAC

medium access control

MEC

multi-access edge computing

MEO

medium-Earth orbit

mMTC

massive machine-type communication

MNO

mobile network operator

MTD

machine-type device

NASA

National Aeronautics and Space Administration

NB-IoT

Narrowband IoT

NFV

network-function virtualization

NR

New Radio

NTN

non-terrestrial network

O-RAN

Open Radio Access Network

PEP

performance-enhancing proxy

PHY

physical layer

PRB

physical resource block

QoS

quality of service

RAN

radio access network

RAT

radio access technology

RF

radiofrequency

RLNC

random linear network coding

RRH

remote radio head

RRM

radio resource management

SBA

service-based architecture

SCHC

Static Context Header Compression

SCS

symmetry chirp signal

SDN

software-defined networking

SDR

software-defined radio

SES

Satellite Earth Stations and Systems

SNW

Sigfox network

TCP

Transmission Control Protocol

UAV

unmanned aerial vehicle

UDP

User Datagram Protocol

UE

user equipment

UHF

ultra-high frequency

UMTS

Universal Mobile Telecommunications Service

URI

Uniform Resource Identifier

URLLC

ultra-reliable low-latency communication

USA

United States of America

VHF

very-high frequency

W-CDMA

Wideband Code Division Multiple Access

References

References is not available for this document.