
To support the fast growth of IoT and cyber physical systems, as well as the advent of 6G, there is a need for communication and networking models that enable more efficient modes for machine-type communications. This calls for a departure from the assumptions of classical communication theoretic problem formulations as well as the traditional network layers. This new communication paradigm is referred to as goal or task oriented communication, or in a broader sense, is part of the emerging area of semantic communications. Over the past decade, there have been a number of approaches towards novel performance metrics, starting from measures of timeliness such as the Age of Information (AoI), Query Age of Information (QAoI), to those that capture goal oriented nature, tracking or control performance such as Quality of Information (QoI), Value of Information (VoI) and Age of Incorrect Information (AoII), moving toward to more sophisticated end-to-end distortion metrics (e.g. MSE), ML performance, or human perception of the reproduced data, and the application of finite-blocklength information theory in the context of the remote monitoring of stochastic processes, and real-time control. We invite original papers that contribute to the fundamentals, as well as the applications of semantic metrics, and protocols that use them, in IoT or automation scenarios.
To support the fast growth of IoT and cyber physical systems, as well as the advent of 6G, there is a need for communication and networking models that enable more efficient modes for machine-type communications. This calls for a departure from the assumptions of classical communication theoretic problem formulations as well as the traditional network layers. This new paradigm is referred to as goal or task-oriented communication, and is relevant also in part of the emerging area of semantic communications.
In this paper, we consider a remote inference system, where a neural network is used to infer a time-varying target (e.g., robot movement), based on features (e.g., video clips) that are progressively received from a sensing node (e.g., a camera). Each feature is a temporal sequence of sensory data. The inference error is determined by (i) the timeliness and (ii) the sequence length of the feature, where we use Age of Information (AoI) as a metric for timeliness. While a longer feature can typically provide better inference performance, it often requires more channel resources for sending the feature. To minimize the time-averaged inference error, we study a learning and communication co-design problem that jointly optimizes feature length selection and transmission scheduling. When there is a single sensor-predictor pair and a single channel, we develop low-complexity optimal co-designs for both the cases of time-invariant and time-variant feature length. When there are multiple sensor-predictor pairs and multiple channels, the co-design problem becomes a restless multi-arm multi-action bandit problem that is PSPACE-hard. For this setting, we design a low-complexity algorithm to solve the problem. Trace-driven evaluations demonstrate the potential of these co-designs to reduce inference error by up to 10000 times.
Finding an optimal/near-optimal scheduling algorithm to minimize the age of information (AoI) in a multi-source G/G/1 system is well-known to be a hard problem, more so if there is a transmission (energy) cost. In this paper, we consider a multi-source G/G/1 system and the goal is to minimize a weighted sum of the AoI of all sources, subject to an energy cost constraint. We propose a novel doubly randomized non-preemptive scheduling algorithm and show that in the non-preemptive setting, where an update under transmission cannot be preempted, the competitive ratio of the proposed algorithm is at most 3 plus the maximum of the ratio of the variance and the mean of the update inter-generation time distribution of sources. Notably, the competitive ratio is independent of the number of sources, or their service time distributions, and is at most 4 for several common update inter-generation time distributions such as exponential, uniform and Rayleigh. For preemptive setting, where an update under transmission can be preempted, we consider a multi-source G/M/1 system and show that the proposed non-preemptive algorithm has competitive ratio at most 5 plus the maximum of the ratio of the variance and the mean of the update inter-generation time distribution of sources.
We consider a node where packets of fixed size (inbits) are generated at arbitrary intervals. The node is required to maintain the peak age of information (AoI) at the monitor below a threshold by transmitting potentially a subset of the generated packets. At any time, depending on the packet availability and the current AoI, the node can choose which packet to transmit, and at what transmission speed (in bits per second). Power consumption is a monotonically increasing convex function of the transmission speed. In this paper, for any given time horizon, the objective is to find a causal policy that minimizes the total energy consumption while satisfying the peak AoI constraint. We consider competitive ratio as the performance metric, that is defined as the ratio of the expected cost of a causal policy, and the expected cost of an optimal offline policy that knows the input (packet generation times) in advance. We first derive a lower bound on the competitive ratio of all causal policies, in terms of the system parameters (such as power function, packet size and peak AoI threshold), and then propose a particular policy for which we show that its competitive ratio has similar order of dependence on the system parameters as the derived lower bound.
This paper contributes tail bounds of the age-of-information of a general class of parallel systems and explores their potential. Parallel systems arise in relevant cases, such as in multi-band mobile networks, multi-technology wireless access, or multi-path protocols, just to name a few. Typically, control over each communication channel is limited and random service outages and congestion cause buffering that impairs the age-of-information. The parallel use of independent channels promises a remedy, since outages on one channel may be compensated for by another. Surprisingly, for the well-known case of $\text{M}\mid \text{M}\mid 1$ queues we find the opposite: pooling capacity in one channel performs better than a parallel system with the same total capacity. A generalization is not possible since there are no solutions for other types of parallel queues at hand. In this work, we prove a dual representation of age-of-information in min-plus algebra that connects to queueing models known from the theory of effective bandwidth/capacity and the stochastic network calculus. Exploiting these methods, we derive tail bounds of the age-of-information of $\text{G}\mid \text{G}\mid 1$ queues. Tail bounds of the age-of-information of independent parallel queues follow readily. In addition to parallel classical queues, we investigate Markov channels where, depending on the memory of the channel, we show the true advantage of parallel systems. We continue to investigate this new finding and provide insight into when capacity should be pooled in one channel or when independent parallel channels perform better. We complement our analysis with simulation results and evaluate different update policies, scheduling policies, and the use of heterogeneous channels that is most relevant for latest multi-band networks.
We consider a multi-process remote estimation system observing $K$ independent Ornstein-Uhlenbeck processes. In this system, a shared sensor samples the $K$ processes in such a way that the long-term average sum mean square error (MSE) is minimized using signal-independent sampling policies, in which sampling instances are chosen independently from the processes’ values. The sensor operates under a total sampling frequency constraint $f_{\max }$ . The samples from all processes consume random processing delays in a shared queue and then are transmitted over an erasure channel with probability $\epsilon $ . We study two variants of the problem: first, when the samples are scheduled according to a Maximum-Age-First (MAF) policy, and the receiver provides an erasure status feedback; and second, when samples are scheduled according to a Round-Robin (RR) policy, when there is no erasure status feedback from the receiver. Aided by optimal structural results, we show that the optimal sampling policy for both settings, under some conditions, is a threshold policy. We characterize the optimal threshold and the corresponding optimal long-term average sum MSE as a function of $K$ , $f_{\max }$ , $\epsilon $ , and the statistical properties of the observed processes. Our results show that, with an exponentially distributed service rate, the optimal thresho