We consider a setting involving $N$ agents, where each agent interacts with an environment modeled as a Markov Decision Process (MDP). The agents' MDPs differ in their reward functions, capturing heterogeneous objectives/tasks. The collective goal of the agents is to communicate intermittently via a central server to find a policy that maximizes the average of long-term cumulative rewards across environments. The limited existing work on this topic either only provide asymptotic rates, or generate biased policies, or fail to establish any benefits of collaboration. In response, we propose Fast-FedPG - a novel federated policy gradient algorithm with a carefully designed bias-correction mechanism. Under a gradient-domination condition, we prove that our algorithm guarantees (i) fast linear convergence with exact gradients, and (ii) sub-linear rates that enjoy a linear speedup w.r.t. the number of agents with noisy, truncated policy gradients. Notably, in each case, the convergence is to a globally optimal policy with no heterogeneity-induced bias. In the absence of gradient-domination, we establish convergence to a first-order stationary point at a rate that continues to benefit from collaboration.
This paper introduces SideSeeing, a novel initiative that provides tools and datasets for assessing the built environment. We present a framework for street-level data acquisition, loading, and analysis. Using the framework, we collected a novel dataset that integrates synchronized video footaged captured from chest-mounted mobile devices with sensor data (accelerometer, gyroscope, magnetometer, and GPS). Each data sample represents a path traversed by a user filming sidewalks near hospitals in Brazil and the USA. The dataset encompasses three hours of content covering 12 kilometers around nine hospitals, and includes 325,000 video frames with corresponding sensor data. Additionally, we present a novel 68-element taxonomy specifically created for sidewalk scene identification. SideSeeing is a step towards a suite of tools that urban experts can use to perform in-depth sidewalk accessibility evaluations. SideSeeing data and tools are publicly available at https://sites.usp.br/sideseeing/.
Complex systems have motivated continuing interest from the scientific community, leading to new concepts and methods. Growing systems represent a case of particular interest, as their topological, geometrical, and also dynamical properties change along time, as new elements are incorporated into the existing structure. In the present work, an approach is the case in which systems grown radially around some straight axis of reference, such as particle deposition on electrodes, or urban expansion along avenues, roads, coastline, or rivers, among several other possibilities. More specifically, we aim at characterizing the topological properties of simulated growing structures, which are represented as graphs, in terms of a measurement corresponding to the accessibility of each involved node. The incorporation of new elements (nodes and links) is performed preferentially to the angular orientation respectively to the reference axis. Several interesting results are reported, including the tendency of structures grown preferentially to the orientation normal to the axis to have smaller accessibility.
Apr 30 2024
cs.DS arXiv:2404.17882v2
This paper explores the connection between classical isoperimetric inequalities, their directed analogues, and monotonicity testing. We study the setting of real-valued functions $f : [0,1]^d \to \mathbb{R}$ on the solid unit cube, where the goal is to test with respect to the $L^p$ distance. Our goals are twofold: to further understand the relationship between classical and directed isoperimetry, and to give a monotonicity tester with sublinear query complexity in this setting. Our main results are 1) an $L^2$ monotonicity tester for $M$-Lipschitz functions with query complexity $\widetilde O(\sqrt{d} M^2 / \epsilon^2)$ and, behind this result, 2) the directed Poincaré inequality $\mathsf{dist}^{\mathsf{mono}}_2(f)^2 \le C \mathbb{E}[|\nabla^- f|^2]$, where the "directed gradient" operator $\nabla^-$ measures the local violations of monotonicity of $f$. To prove the second result, we introduce a partial differential equation (PDE), the directed heat equation, which takes a one-dimensional function $f$ into a monotone function $f^*$ over time and enjoys many desirable analytic properties. We obtain the directed Poincaré inequality by combining convergence aspects of this PDE with the theory of optimal transport. Crucially for our conceptual motivation, this proof is in complete analogy with the mathematical physics perspective on the classical Poincaré inequality, namely as characterizing the convergence of the standard heat equation toward equilibrium.
Codebook-based beam selection is one approach for configuring millimeter wave communication links. The overhead required to reconfigure the transmit and receive beam pair, though, increases in highly dynamic vehicular communication systems. Location information coupled with machine learning (ML) beam recommendation is one way to reduce the overhead of beam pair selection. In this paper, we develop ML-based location-aided approaches to decouple the beam selection between the user equipment (UE) and the base station (BS). We quantify the performance gaps due to decoupling beam selection and also disaggregating the UE's location information from the BS. Our simulation results show that decoupling beam selection with available location information at the BS performs comparable to joint beam pair selection at the BS. Moreover, decoupled beam selection without location closely approaches the performance of beam pair selection at the BS when sufficient beam pairs are swept.
Bo Peng, Daniel Goldstein, Quentin Anthony, Alon Albalak, Eric Alcaide, Stella Biderman, Eugene Cheah, Xingjian Du, Teddy Ferdinan, Haowen Hou, Przemysław Kazienko, Kranthi Kiran GV, Jan Kocoń, Bartłomiej Koptyra, Satyapriya Krishna, Ronald McClelland Jr., Jiaju Lin, Niklas Muennighoff, Fares Obeid, Atsushi Saito, et al (10) We present Eagle (RWKV-5) and Finch (RWKV-6), sequence models improving upon the RWKV (RWKV-4) architecture. Our architectural design advancements include multi-headed matrix-valued states and a dynamic recurrence mechanism that improve expressivity while maintaining the inference efficiency characteristics of RNNs. We introduce a new multilingual corpus with 1.12 trillion tokens and a fast tokenizer based on greedy matching for enhanced multilinguality. We trained four Eagle models, ranging from 0.46 to 7.5 billion parameters, and two Finch models with 1.6 and 3.1 billion parameters and find that they achieve competitive performance across a wide variety of benchmarks. We release all our models on HuggingFace under the Apache 2.0 license. Models at: https://huggingface.co/RWKV Training code at: https://github.com/RWKV/RWKV-LM Inference code at: https://github.com/RWKV/ChatRWKV Time-parallel training code at: https://github.com/RWKV/RWKV-infctx-trainer
Obtaining accurate and timely channel state information (CSI) is a fundamental challenge for large antenna systems. Mobile systems like 5G use a beam management framework that joins the initial access, beamforming, CSI acquisition, and data transmission. The design of codebooks for these stages, however, is challenging due to their interrelationships, varying array sizes, and site-specific channel and user distributions. Furthermore, beam management is often focused on single-sector operations while ignoring the overarching network- and system-level optimization. In this paper, we proposed an end-to-end learned codebook design algorithm, network beamspace learning (NBL), that captures and optimizes codebooks to mitigate interference while maximizing the achievable performance with extremely large hybrid arrays. The proposed algorithm requires limited shared information yet designs codebooks that outperform traditional codebooks by over 10dB in beam alignment and achieve more than 25% improvements in network spectral efficiency.
The use of one-bit analog-to-digital converter (ADC) has been considered as a viable alternative to high resolution counterparts in realizing and commercializing massive multiple-input multiple-output (MIMO) systems. However, the issue of discarding the amplitude information by one-bit quantizers has to be compensated. Thus, carefully tailored methods need to be developed for one-bit channel estimation and data detection as the conventional ones cannot be used. To address these issues, the problems of one-bit channel estimation and data detection for MIMO orthogonal frequency division multiplexing (OFDM) system that operates over uncorrelated frequency selective channels are investigated here. We first develop channel estimators that exploit Gaussian discriminant analysis (GDA) classifier and approximated versions of it as the so-called weak classifiers in an adaptive boosting (AdaBoost) approach. Particularly, the combination of the approximated GDA classifiers with AdaBoost offers the benefit of scalability with the linear order of computations, which is critical in massive MIMO-OFDM systems. We then take advantage of the same idea for proposing the data detectors. Numerical results validate the efficiency of the proposed channel estimators and data detectors compared to other methods. They show comparable/better performance to that of the state-of-the-art methods, but require dramatically lower computational complexities and run times.
Fluid antenna systems (FASs) can reconfigure their locations freely within a spatially continuous space. To keep favorable antenna positions, the channel state information (CSI) acquisition for FASs is essential. While some techniques have been proposed, most existing FAS channel estimators require several channel assumptions, such as slow variation and angular-domain sparsity. When these assumptions are not reasonable, the model mismatch may lead to unpredictable performance loss. In this paper, we propose the successive Bayesian reconstructor (S-BAR) as a general solution to estimate FAS channels. Unlike model-based estimators, the proposed S-BAR is prior-aided, which builds the experiential kernel for CSI acquisition. Inspired by Bayesian regression, the key idea of S-BAR is to model the FAS channels as a stochastic process, whose uncertainty can be successively eliminated by kernel-based sampling and regression. In this way, the predictive mean of the regressed stochastic process can be viewed as the maximum a posterior (MAP) estimator of FAS channels. Simulation results verify that, in both model-mismatched and model-matched cases, the proposed S-BAR can achieve higher estimation accuracy than the existing schemes.
The increasing demand for wireless data transfer has been the driving force behind the widespread adoption of Massive MIMO (multiple-input multiple-output) technology in 5G. The next-generation MIMO technology is now being developed to cater to the new data traffic and performance expectations generated by new user devices and services in the next decade. The evolution towards "ultra-massive MIMO (UM-MIMO)" is not only about adding more antennas but will also uncover new propagation and hardware phenomena that can only be treated by jointly utilizing insights from the communication, electromagnetic (EM), and circuit theory areas. This article offers a comprehensive overview of the key benefits of the UM-MIMO technology and the associated challenges. It explores massive multiplexing facilitated by radiative near-field effects, characterizes the spatial degrees-of-freedom, and practical channel estimation schemes tailored for massive arrays. Moreover, we provide a tutorial on EM theory and circuit theory, and how it is used to obtain physically consistent antenna and channel models. Subsequently, the article describes different ways to implement massive and dense antenna arrays, and how to co-design antennas with signal processing. The main open research challenges are identified at the end.
Fluid antenna systems (FASs) can reconfigure their antenna locations freely within a spatially continuous space. To keep favorable antenna positions, the channel state information (CSI) acquisition for FASs is essential. While some techniques have been proposed, most existing FAS channel estimators require several channel assumptions, such as slow variation and angular-domain sparsity. When these assumptions are not reasonable, the model mismatch may lead to unpredictable performance loss. In this paper, we propose the successive Bayesian reconstructor (S-BAR) as a general solution to estimate FAS channels. Unlike model-based estimators, the proposed S-BAR is prior-aided, which builds the experiential kernel for CSI acquisition. Inspired by Bayesian regression, the key idea of S-BAR is to model the FAS channels as a stochastic process, whose uncertainty can be successively eliminated by kernel-based sampling and regression. In this way, the predictive mean of the regressed stochastic process can be viewed as the maximum a posterior (MAP) estimator of FAS channels. Simulation results verify that, in both model-mismatched and model-matched cases, the proposed S-BAR can achieve higher estimation accuracy than the existing schemes.
Beam management is a strategy to unify beamforming and channel state information (CSI) acquisition with large antenna arrays in 5G. Codebooks serve multiple uses in beam management including beamforming reference signals, CSI reporting, and analog beam training. In this paper, we propose and evaluate a machine learning-refined codebook design process for extremely large multiple-input multiple-output (X-MIMO) systems. We propose a neural network and beam selection strategy to design the initial access and refinement codebooks using end-to-end learning from beamspace representations. The algorithm, called Extreme-Beam Management (X-BM), can significantly improve the performance of extremely large arrays as envisioned for 6G and capture realistic wireless and physical layer aspects. Our results show an 8dB improvement in initial access and overall effective spectral efficiency improvements compared to traditional codebook methods.
Nov 27 2023
cs.DS arXiv:2311.14247v1
We are interested in testing properties of distributions with systematically mislabeled samples. Our goal is to make decisions about unknown probability distributions, using a sample that has been collected by a confused collector, such as a machine-learning classifier that has not learned to distinguish all elements of the domain. The confused collector holds an unknown clustering of the domain and an input distribution $\mu$, and provides two oracles: a sample oracle which produces a sample from $\mu$ that has been labeled according to the clustering; and a label-query oracle which returns the label of a query point $x$ according to the clustering. Our first set of results shows that identity, uniformity, and equivalence of distributions can be tested efficiently, under the earth-mover distance, with remarkably weak conditions on the confused collector, even when the unknown clustering is adversarial. This requires defining a variant of the distribution testing task (inspired by the recent testable learning framework of Rubinfeld & Vasilyan), where the algorithm should test a joint property of the distribution and its clustering. As an example, we get efficient testers when the distribution tester is allowed to reject if it detects that the confused collector clustering is "far" from being a decision tree. The second set of results shows that we can sometimes do significantly better when the clustering is random instead of adversarial. For certain one-dimensional random clusterings, we show that uniformity can be tested under the TV distance using $\widetilde O\left(\frac{\sqrt n}{\rho^{3/2} \epsilon^2}\right)$ samples and zero queries, where $\rho \in (0,1]$ controls the "resolution" of the clustering. We improve this to $O\left(\frac{\sqrt n}{\rho \epsilon^2}\right)$ when queries are allowed.
Impedance-matching networks affect power transfer from the radio frequency (RF) chains to the antennas. Their design impacts the signal to noise ratio (SNR) and the achievable rate. In this paper, we maximize the information-theoretic achievable rate of a multiple-input-single-output (MISO) system with wideband matching constraints. Using a multiport circuit theory approach with frequency-selective scattering parameters, we propose a general framework for optimizing the MISO achievable rate that incorporates Bode-Fano wideband matching theory. We express the solution to the achievable rate optimization problem in terms of the optimized transmission coefficient and the Lagrangian parameters corresponding to the Bode-Fano inequality constraints. We apply this framework to a single electric Chu's antenna and an array of two electric Chu's antennas. We compare the optimized achievable rate obtained numerically with other benchmarks like the ideal achievable rate computed by disregarding matching constraints and the achievable rate obtained by using sub-optimal matching strategies like conjugate matching and frequency-flat transmission. We also propose a practical methodology to approximate the achievable rate bound by using the optimal transmission coefficient to derive a physically realizable matching network through the ADS software.
With increasing frequencies, bandwidths, and array apertures, the phenomenon of beam squint arises as a serious impairment to beamforming. Fully digital arrays with true time delay per antenna element are a potential solution, but they require downconversion at each element. This paper shows that hybrid arrays can perform essentially as well as digital arrays once the number of radio-frequency chains exceeds a certain threshold that is far below the number of elements. The result is robust, holding also for suboptimum but highly appealing beamspace architectures.
This study addressed the complex task of sentiment analysis on a dataset of 119,988 original tweets from Weibo using a Convolutional Neural Network (CNN), offering a new approach to Natural Language Processing (NLP). The data, sourced from Baidu's PaddlePaddle AI platform, were meticulously preprocessed, tokenized, and categorized based on sentiment labels. A CNN-based model was utilized, leveraging word embeddings for feature extraction, and trained to perform sentiment classification. The model achieved a macro-average F1-score of approximately 0.73 on the test set, showing balanced performance across positive, neutral, and negative sentiments. The findings underscore the effectiveness of CNNs for sentiment analysis tasks, with implications for practical applications in social media analysis, market research, and policy studies. The complete experimental content and code have been made publicly available on the Kaggle data platform for further research and development. Future work may involve exploring different architectures, such as Recurrent Neural Networks (RNN) or transformers, or using more complex pre-trained models like BERT, to further improve the model's ability to understand linguistic nuances and context.
Ongoing research and experiments have enabled quantum memory to realize the storage of qubits. On the other hand, interleaving techniques are used to deal with burst of errors. Effective interleaving techniques for combating burst of errors by using classical error-correcting codes have been proposed in several articles found in the literature, however, to the best of our knowledge, little is known regarding interleaving techniques for combating clusters of errors in topological quantum error-correcting codes. Motivated by that, in this work, we present new three and four-dimensional toric quantum codes which are featured by lattice codes and apply a quantum interleaving method to such new three and four-dimensional toric quantum codes. By applying such a method to these new codes we provide new three and four-dimensional quantum burst-error-correcting codes. As a consequence, new three and four-dimensional toric and burst-error-correcting quantum codes are obtained which have better information rates than those three and four-dimensional toric quantum codes from the literature. In addition to these proposed three and four-dimensional quantum burst-error-correcting codes improve such information rates, they can be used for burst-error-correction in errors which are located, quantum data stored and quantum channels with memory.
Jul 06 2023
cs.DS arXiv:2307.02193v1
We study the connection between directed isoperimetric inequalities and monotonicity testing. In recent years, this connection has unlocked breakthroughs for testing monotonicity of functions defined on discrete domains. Inspired the rich history of isoperimetric inequalities in continuous settings, we propose that studying the relationship between directed isoperimetry and monotonicity in such settings is essential for understanding the full scope of this connection. Hence, we ask whether directed isoperimetric inequalities hold for functions $f : [0,1]^n \to \mathbb{R}$, and whether this question has implications for monotonicity testing. We answer both questions affirmatively. For Lipschitz functions $f : [0,1]^n \to \mathbb{R}$, we show the inequality $d^{\mathsf{mono}}_1(f) \lesssim \mathbb{E}\left[\|\nabla^- f\|_1\right]$, which upper bounds the $L^1$ distance to monotonicity of $f$ by a measure of its "directed gradient". A key ingredient in our proof is the monotone rearrangement of $f$, which generalizes the classical "sorting operator" to continuous settings. We use this inequality to give an $L^1$ monotonicity tester for Lipschitz functions $f : [0,1]^n \to \mathbb{R}$, and this framework also implies similar results for testing real-valued functions on the hypergrid.
Jun 09 2023
cs.RO arXiv:2306.05343v1
The placement of grab bars for elderly users is based largely on ADA building codes and does not reflect the large differences in height, mobility, and muscle power between individual persons. The goal of this study is to see if there are any correlations between an elderly user's preferred handlebar pose and various demographic indicators, self-rated mobility for tasks requiring postural change, and biomechanical markers. For simplicity, we consider only the case where the handlebar is positioned directly in front of the user, as this confines the relevant body kinematics to a 2D sagittal plane. Previous eldercare devices have been constructed to position a handlebar in various poses in space. Our work augments these devices and adds to the body of knowledge by assessing how the handlebar should be positioned based on data on actual elderly people instead of simulations.
We provide a new variational definition for the spread of an orbital under periodic boundary conditions (PBCs) that is continuous with respect to the gauge, consistent in the thermodynamic limit, well-suited to diffuse orbitals, and systematically adaptable to schemes computing localized Wannier functions. Existing definitions do not satisfy all these desiderata, partly because they depend on an "orbital center"-an ill-defined concept under PBCs. Based on this theoretical development, we showcase a robust and efficient (10x-70x fewer iterations) localization scheme across a range of materials.
Most uses of Meta-Learning in visual recognition are very often applied to image classification, with a relative lack of works in other tasks such as segmentation and detection. We propose a generic Meta-Learning framework for few-shot weakly-supervised segmentation in medical imaging domains. We conduct a comparative analysis of meta-learners from distinct paradigms adapted to few-shot image segmentation in different sparsely annotated radiological tasks. The imaging modalities include 2D chest, mammographic and dental X-rays, as well as 2D slices of volumetric tomography and resonance images. Our experiments consider a total of 9 meta-learners, 4 backbones and multiple target organ segmentation tasks. We explore small-data scenarios in radiology with varying weak annotation styles and densities. Our analysis shows that metric-based meta-learning approaches achieve better segmentation results in tasks with smaller domain shifts in comparison to the meta-training datasets, while some gradient- and fusion-based meta-learners are more generalizable to larger domain shifts.
In this paper, we consider the steps to be followed in the analysis and interpretation of the quantization problem related to the $C_{2,8}$ channel, where the Fuchsian differential equations, the generators of the Fuchsian groups, and the tessellations associated with the cases $g=2$ and $g=3$, related to the hyperbolic case, are determined. In order to obtain these results, it is necessary to determine the genus $g$ of each surface on which this channel may be embedded. After that, the procedure is to determine the algebraic structure (Fuchsian group generators) associated with the fundamental region of each surface. To achieve this goal, an associated linear second-order Fuchsian differential equation whose linearly independent solutions provide the generators of this Fuchsian group is devised. In addition, the tessellations associated with each analyzed case are identified. These structures are identified in four situations, divided into two cases $(g=2$ and $g=3)$, obtaining, therefore, both algebraic and geometric characterizations associated with quantizing the $C_{2,8}$ channel.
Distribution testing is a fundamental statistical task with many applications, but we are interested in a variety of problems where systematic mislabelings of the sample prevent us from applying the existing theory. To apply distribution testing to these problems, we introduce distribution testing under the parity trace, where the algorithm receives an ordered sample $S$ that reveals only the least significant bit of each element. This abstraction reveals connections between the following three problems of interest, allowing new upper and lower bounds: 1. In distribution testing with a confused collector, the collector of the sample may be incapable of distinguishing between nearby elements of a domain (e.g. a machine learning classifier). We prove bounds for distribution testing with a confused collector on domains structured as a cycle or a path. 2. Recent work on the fundamental testing vs. learning question established tight lower bounds on distribution-free sample-based property testing by reduction from distribution testing, but the tightness is limited to symmetric properties. The parity trace allows a broader family of equivalences to non-symmetric properties, while recovering and strengthening many of the previous results with a different technique. 3. We give the first results for property testing in the well-studied trace reconstruction model, where the goal is to test whether an unknown string $x$ satisfies some property or is far from satisfying that property, given only independent random traces of $x$. Our main technical result is a tight bound of $\widetilde \Theta\left((n/\epsilon)^{4/5} + \sqrt n/\epsilon^2\right)$ for testing uniformity of distributions over $[n]$ under the parity trace, leading also to results for the problems above.
We propose an adaptive coding approach to achieve linear-quadratic-Gaussian (LQG) control with near-minimum bitrate prefix-free feedback. Our approach combines a recent analysis of a quantizer design for minimum rate LQG control with work on universal lossless source coding for sources on countable alphabets. In the aforementioned quantizer design, it was established that the quantizer outputs are an asymptotically stationary, ergodic process. To enable LQG control with provably near-minimum bitrate, the quantizer outputs must be encoded into binary codewords efficiently. This is possible given knowledge of the probability distributions of the quantizer outputs, or of their limiting distribution. Obtaining such knowledge is challenging; the distributions do not readily admit closed form descriptions. This motivates the application of universal source coding. Our main theoretical contribution in this work is a proof that (after an invertible transformation), the quantizer outputs are random variables that fall within an exponential or power-law envelope class (depending on the plant dimension). Using ideas from universal coding on envelope classes, we develop a practical, zero-delay version of these algorithms that operates with fixed precision arithmetic. We evaluate the performance of this algorithm numerically, and demonstrate competitive results with respect to fundamental tradeoffs between bitrate and LQG control performance.
Beam codebooks are a recent feature to enable high dimension multiple-input multiple-output in 5G. Codebooks comprised of customizable beamforming weights can be used to transmit reference signals and aid the channel state information (CSI) acquisition process. Codebooks are also used for quantizing feedback following CSI acquisition. In this paper, we characterize the role of each codebook used during the beam management process and design a neural network to find codebooks that improve overall system performance. Evaluating a codebook requires considering the system-level dependency between the codebooks, feedback, overhead, and spectral efficiency. The proposed neural network is built on translating codebook and feedback knowledge into a consistent beamspace basis similar to a virtual channel model to generate initial access codebooks. This beamspace codebook algorithm is designed to directly integrate with current 5G beam management standards without changing the feedback format or requiring additional side information. Our simulations show that the neural network codebooks improve over traditional codebooks, even in dispersive sub-6GHz environments. We further use our framework to evaluate CSI feedback formats with regard to multi-user spectral efficiency. Our results suggest that optimizing codebook performance can provide valuable performance improvements, but optimizing the feedback configuration is also important in sub-6GHz bands.
Massive multiple-input multiple-output (MIMO) is an important technology in fifth generation (5G) cellular networks and beyond. To help design the beamforming at the base station, 5G has introduced new support in the form of flexible feedback and configurable antenna array geometries. In this article, we present an overview of MIMO throughout the mobile standards, highlight the new beam-based feedback system in 5G NR, and describe how this feedback system enables massive MIMO through beam management. Finally, we conclude with challenges related to massive MIMO in 5G.
The State of Emergency declaration in Japan due to the COVID-19 pandemic affected many aspects of society in the country, much like the rest of the world. One sector that felt its disruptive impact was education. As educational institutions raced to implement emergency remote teaching (ERT) to continue providing the learning needs of students, some have opened to innovative interventions. This paper describes a case of ERT where Filipino vocabulary was taught to a class of Japanese students taking Philippine Studies in a Japanese university using a cognitive innovation based on virtual reality, an immer-sive technology often researched for immersion and presence. Students were divided into three groups to experience six lessons designed around virtual reality photo-based tours at different immersion levels. While the effect of immersion on satisfaction was not found to be statistically significant, presence and satisfaction were found to be correlated. Despite challenges that were encountered, benefits like enjoyment, increased engagement , and perceived learning were reported by the students. Our findings exemplify how emerging multisensory technologies can be used to enhance affective and cognitive dimensions of human experience while responding to gaps created by the spatial limitations of remote learning.
Jan 06 2023
cs.CY arXiv:2301.01908v1
When educational institutions worldwide scrambled for ways to continue their classes during lockdowns caused by the COVID-19 pandemic, the use of information and communication technology (ICT) for remote teaching has become widely considered to be a potential solution. As universities raced to implement emergency remote teaching (ERT) strategies in Japan, some have explored innovative interventions other than webinar platforms and learning management systems to bridge the gap caused by restricted mobility among teachers and learners. One such innovation is virtual reality (VR). VR has been changing the landscape of higher education because of its ability to "teleport" learners to various places by simulating real-world environments in the virtual world. Some teachers, including the authors of this paper, explored integrating VR into their activities to address issues caused by geographical limitations brought about by the heightened restrictions in 2020. Results were largely encouraging. However, rules started relaxing in the succeeding years as more people got vaccinated. Thus, some fully online classes in Japan shifted to blended learning as they moved toward fully returning to in-person classes prompting educators to modify how they implemented their VR-based interventions. This paper describes how a class of university students in Japan who were taking a Filipino language course experienced a VR-based intervention in blended mode, which was originally prototyped during the peak of the ERT era. Moreover, adjustments and comparisons regarding methodological idiosyncrasies and findings between the fully online iteration and the recently implemented blended one are reported in detail.
Beamforming is a signal processing technique to steer, shape, and focus an electromagnetic wave using an array of sensors toward a desired direction. It has been used in several engineering applications such as radar, sonar, acoustics, astronomy, seismology, medical imaging, and communications. With the advances in multi-antenna technologies largely for radar and communications, there has been a great interest on beamformer design mostly relying on convex/nonconvex optimization. Recently, machine learning is being leveraged for obtaining attractive solutions to more complex beamforming problems. This article captures the evolution of beamforming in the last twenty-five years from convex-to-nonconvex optimization and optimization-to-learning approaches. It provides a glimpse of this important signal processing technique into a variety of transmit-receive architectures, propagation zones, paths, and conventional/emerging applications.
Legal and ethical restrictions on accessing relevant data inhibit data science research in critical domains such as health, finance, and education. Synthetic data generation algorithms with privacy guarantees are emerging as a paradigm to break this data logjam. Existing approaches, however, assume that the data holders supply their raw data to a trusted curator, who uses it as fuel for synthetic data generation. This severely limits the applicability, as much of the valuable data in the world is locked up in silos, controlled by entities who cannot show their data to each other or a central aggregator without raising privacy concerns. To overcome this roadblock, we propose the first solution in which data holders only share encrypted data for differentially private synthetic data generation. Data holders send shares to servers who perform Secure Multiparty Computation (MPC) computations while the original data stays encrypted. We instantiate this idea in an MPC protocol for the Multiplicative Weights with Exponential Mechanism (MWEM) algorithm to generate synthetic data based on real data originating from many data holders without reliance on a single point of failure.
From an information theoretic perspective, joint communication and sensing (JCAS) represents a natural generalization of communication network functionality. However, it requires the re-evaluation of network performance from a multi-objective perspective. We develop a novel mathematical framework for characterizing the sensing and communication coverage probability and ergodic rate in JCAS networks. We employ a formulation of sensing parameter estimation based on mutual information to extend the notions of coverage probability and ergodic rate to the radar setting. We define sensing coverage probability as the probability that the rate of information extracted about the parameters of interest associated with a typical radar target exceeds some threshold, and sensing ergodic rate as the spatial average of the aforementioned rate of information. Using this framework, we analyze the downlink sensing and communication coverage and rate of a mmWave JCAS network employing a shared waveform, directional beamforming, and monostatic sensing. Leveraging tools from stochastic geometry, we derive upper and lower bounds for these quantities. We also develop several general technical results including: i) a generic method for obtaining closed form upper and lower bounds on the Laplace Transform of a shot noise process, ii) a new analog of Hölder's Inequality to the setting of harmonic means, and iii) a relation between the Laplace and Mellin Transforms of a non-negative random variable. We use the derived bounds to numerically investigate the performance of JCAS networks under varying base station and blockage density. Among several insights, our numerical analysis indicates that network densification improves sensing SINR performance -- in contrast to communications.
High-throughput DFT calculations are key to screening existing/novel materials, sampling potential energy surfaces, and generating quantum mechanical data for machine learning. By including a fraction of exact exchange (EXX), hybrid functionals reduce the self-interaction error in semi-local DFT and furnish a more accurate description of the underlying electronic structure, albeit at a high computational cost that often prohibits such high-throughput applications. To address this challenge, we have constructed SeA (SeA=SCDM+exx+ACE), a robust, accurate, and efficient framework for high-throughput condensed-phase hybrid DFT in the PWSCF module of Quantum ESPRESSO (QE) by combining: (1) the non-iterative selected columns of the density matrix (SCDM) orbital localization scheme, (2) a black-box and linear-scaling EXX algorithm (exx), and (3) adaptively compressed exchange (ACE). Across a diverse set of non-equilibrium (H$_2$O)$_{64}$ configurations (with densities spanning 0.4 g/cm$^3$$-$1.7 g/cm$^3$), SeA yields a one$-$two order-of-magnitude speedup (~8X$-$26X) in the overall time-to-solution compared to PWSCF(ACE) in QE (~78X$-$247X speedup compared to the conventional EXX implementation) and yields energies, ionic forces, and other properties with high fidelity. As a proof-of-principle high-throughput application, we trained a deep neural network (DNN) potential for ambient liquid water at the hybrid DFT level using SeA via an actively learned data set with ~8,700 (H$_2$O)$_{64}$ configurations. Using an out-of-sample set of (H$_2$O)$_{512}$ configurations (at non-ambient conditions), we confirmed the accuracy of this SeA-trained potential and showcased the capabilities of SeA by computing the ground-truth ionic forces in this challenging system containing > 1,500 atoms.
Since most of vehicular radar systems are already exploiting millimeter-wave (mmWave) spectra, it would become much more feasible to implement a joint radar and communication system by extending communication frequencies into the mmWave band. In this paper, an IEEE 802.11ad waveform-based radar imaging technique is proposed for vehicular settings. A roadside unit (RSU) transmits the IEEE 802.11ad waveform to a vehicle for communications while the RSU also listens to the echoes of transmitted waveform to perform inverse synthetic aperture radar (ISAR) imaging. To obtain high-resolution images of the vehicle, the RSU needs to accurately estimate round-trip delays, Doppler shifts, and velocity of vehicle. The proposed ISAR imaging first estimates the round-trip delays using a good correlation property of Golay complementary sequences in the IEEE 802.11ad preamble. The Doppler shifts are then obtained using least square estimation from the echo signals and refined to compensate phase wrapping caused by phase rotation. The velocity of vehicle is determined using an equation of motion and the estimated Doppler shifts. Simulation results verify that the proposed technique is able to form high-resolution ISAR images from point scatterer models of realistic vehicular settings with different viewpoints. The proposed ISAR imaging technique can be used for various vehicular applications, e.g., traffic condition analyses or advanced collision warning systems.
In this work we consider discrete-time multiple-input multiple-output (MIMO) linear-quadratic-Gaussian (LQG) control where the feedback consists of variable length binary codewords. To simplify the decoder architecture, we enforce a strict prefix constraint on the codewords. We develop a data compression architecture that provably achieves a near minimum time-average expected bitrate for a fixed constraint on the LQG performance. The architecture conforms to the strict prefix constraint and does not require time-varying lossless source coding, in contrast to the prior art.
In this paper, we develop algorithms for joint user scheduling and three types of mmWave link configuration: relay selection, codebook optimization, and beam tracking in millimeter wave (mmWave) networks. Our goal is to design an online controller that dynamically schedules users and configures their links to minimize the system delay. To solve this complex scheduling problem, we model it as a dynamic decision-making process and develop two reinforcement learning-based solutions. The first solution is based on deep reinforcement learning (DRL), which leverages the proximal policy optimization to train a neural network-based solution. Due to the potential high sample complexity of DRL, we also propose an empirical multi-armed bandit (MAB)-based solution, which decomposes the decision-making process into a sequential of sub-actions and exploits classic maxweight scheduling and Thompson sampling to decide those sub-actions. Our evaluation of the proposed solutions confirms their effectiveness in providing acceptable system delay. It also shows that the DRL-based solution has better delay performance while the MAB-based solution has a faster training process.
The narrowband and far-field assumption in conventional wireless system design leads to a mismatch with the optimal beamforming required for wideband and near-field systems. This discrepancy is exacerbated for larger apertures and bandwidths. To characterize the behavior of near-field and wideband systems, we derive the beamforming gain expression achieved by a frequency-flat phased array designed for plane-wave propagation. To determine the far-field to near-field boundary for a wideband system, we propose a frequency-selective distance metric. The proposed far-field threshold increases for frequencies away from the center frequency. The analysis results in a fundamental upper bound on the product of the array aperture and the system bandwidth. We present numerical results to illustrate how the gain threshold affects the maximum usable bandwidth for the n260 and n261 5G NR bands.
A toric quantum error-correcting code construction procedure is presented in this work. A new class of an infinite family of toric quantum codes is provided by constructing a classical cyclic code on the square lattice $\mathbb{Z}_{q}\times \mathbb{Z}_{q}$ for all odd integers $q\geq 5$ and, consequently, new toric quantum codes are constructed on such square lattices regardless of whether $q$ can be represented as a sum of two squares. Furthermore this work supplies for each $q$ the polyomino shapes that tessellate the corresponding square lattices and, consequently, tile the lattice $\mathbb{Z}^{2}$. The channel without memory to be considered for these constructed toric quantum codes is symmetric, since the $\mathbb{Z}^{2}$-lattice is autodual. Moreover, we propose a quantum interleaving technique by using the constructed toric quantum codes which shows that the code rate and the coding gain of the interleaved toric quantum codes are better than the code rate and the coding gain of Kitaev's toric quantum codes for $q=2n+1$, where $n\geq 2$, and of an infinite class of Bombin and Martin-Delgado's toric quantum codes. In addition to the proposed quantum interleaving technique improves such parameters, it can be used for burst-error correction in errors which are located, quantum data stored and quantum channels with memory.
Recent research in ultra-reliable and low latency communications (URLLC) for future wireless systems has spurred interest in short block-length codes. In this context, we analyze arbitrary harmonic bandwidth (BW) expansions for a class of high-dimension constant curvature curve codes for analog error correction of independent continuous-alphabet uniform sources. In particular, we employ the circumradius function from knot theory to prescribe insulating tubes about the centerline of constant curvature curves. We then use tube packing density within a hypersphere to optimize the curve parameters. The resulting constant curvature curve tube (C3T) codes possess the smallest possible latency, i.e., block-length is unity under BW expansion mapping. Further, the codes perform within $5$ dB signal-to-distortion ratio of the optimal performance theoretically achievable at a signal-to-noise ratio (SNR) $< -5$ dB for BW expansion factor $n \leq 10$. Furthermore, we propose a neural-network-based method to decode C3T codes. We show that, at low SNR, the neural-network-based C3T decoder outperforms the maximum likelihood and minimum mean-squared error decoders for all $n$. The best possible digital codes require two to three orders of magnitude higher latency compared to C3T codes, thereby demonstrating the latter's utility for URLLC.
Apr 13 2022
cs.CL arXiv:2204.05541v1
Languages are classified as low-resource when they lack the quantity of data necessary for training statistical and machine learning tools and models. Causes of resource scarcity vary but can include poor access to technology for developing these resources, a relatively small population of speakers, or a lack of urgency for collecting such resources in bilingual populations where the second language is high-resource. As a result, the languages described as low-resource in the literature are as different as Finnish on the one hand, with millions of speakers using it in every imaginable domain, and Seneca, with only a small-handful of fluent speakers using the language primarily in a restricted domain. While issues stemming from the lack of resources necessary to train models unite this disparate group of languages, many other issues cut across the divide between widely-spoken low resource languages and endangered languages. In this position paper, we discuss the unique technological, cultural, practical, and ethical challenges that researchers and indigenous speech community members face when working together to develop language technology to support endangered language documentation and revitalization. We report the perspectives of language teachers, Master Speakers and elders from indigenous communities, as well as the point of view of academics. We describe an ongoing fruitful collaboration and make recommendations for future partnerships between academic researchers and language community stakeholders.
Motivated by control with communication constraints, in this work we develop a time-invariant data compression architecture for linear-quadratic-Gaussian (LQG) control with minimum bitrate prefix-free feedback. For any fixed control performance, the approach we propose nearly achieves known directed information (DI) lower bounds on the time-average expected codeword length. We refine the analysis of a classical achievability approach, which required quantized plant measurements to be encoded via a time-varying lossless source code. We prove that the sequence of random variables describing the quantizations has a limiting distribution and that the quantizations may be encoded with a fixed source code optimized for this distribution without added time-asymptotic redundancy. Our result follows from analyzing the long-term stochastic behavior of the system, and permits us to additionally guarantee that the time-average codeword length (as opposed to expected length) is almost surely within a few bits of the minimum DI. To our knowledge, this time-invariant achievability result is the first in the literature. The originally published version of the supplementary material included a proof that contained an error that turned out to be inconsequential. This updated preprint corrects this error, which originally appeared under Lemma A.7.
In this letter, we consider a Linear Quadratic Gaussian (LQG) control system where feedback occurs over a noiseless binary channel and derive lower bounds on the minimum communication cost (quantified via the channel bitrate) required to attain a given control performance. We assume that at every time step an encoder can convey a packet containing a variable number of bits over the channel to a decoder at the controller. Our system model provides for the possibility that the encoder and decoder have shared randomness, as is the case in systems using dithered quantizers. We define two extremal prefix-free requirements that may be imposed on the message packets; such constraints are useful in that they allow the decoder, and potentially other agents to uniquely identify the end of a transmission in an online fashion. We then derive a lower bound on the rate of prefix-free coding in terms of directed information; in particular we show that a previously known bound still holds in the case with shared randomness. We generalize the bound for when prefix constraints are relaxed, and conclude with a rate-distortion formulation.
Large-scale analysis of pedestrian infrastructures, particularly sidewalks, is critical to human-centric urban planning and design. Benefiting from the rich data set of planimetric features and high-resolution orthoimages provided through the New York City Open Data portal, we train a computer vision model to detect sidewalks, roads, and buildings from remote-sensing imagery and achieve 83% mIoU over held-out test set. We apply shape analysis techniques to study different attributes of the extracted sidewalks. More specifically, we do a tile-wise analysis of the width, angle, and curvature of sidewalks, which aside from their general impacts on walkability and accessibility of urban areas, are known to have significant roles in the mobility of wheelchair users. The preliminary results are promising, glimpsing the potential of the proposed approach to be adopted in different cities, enabling researchers and practitioners to have a more vivid picture of the pedestrian realm.
Today, various machine learning (ML) applications offer continuous data processing and real-time data analytics at the edge of a wireless network. Distributed real-time ML solutions are highly sensitive to the so-called straggler effect caused by resource heterogeneity and alleviated by various computation offloading mechanisms that seriously challenge the communication efficiency, especially in large-scale scenarios. To decrease the communication overhead, we rely on device-to-device (D2D) connectivity that improves spectrum utilization and allows efficient data exchange between devices in proximity. In particular, we design a novel D2D-aided coded federated learning method (D2D-CFL) for efficient load balancing across devices. The proposed solution captures system dynamics, including data (time-dependent learning model, varied intensity of data arrivals), device (diverse computational resources and volume of training data), and deployment (varied locations and D2D graph connectivity). To minimize the number of communication rounds, we derive an optimal compression rate for achieving minimum processing time and establish its connection with the convergence time. The resulting optimization problem provides suboptimal compression parameters, which improve the total training time. Our proposed method is beneficial for real-time collaborative applications, where the users continuously generate training data resulting in the model drift.
An urban tactical wireless network is considered wherein the base stations are situated on unmanned aerial vehicles (UAVs) that provide connectivity to ground assets such as vehicles located on city streets. The UAVs are assumed to be randomly deployed at a fixed height according to a two-dimensional point process. Millimeter-wave (mmWave) frequencies are used to avail of large available bandwidths and spatial isolation due to beamforming. In urban environments, mmWave signals are prone to blocking of the line-of-sight (LoS) by buildings. While reflections are possible, the desire for consistent connectivity places a strong preference on the existence of an unblocked LoS path. As such, the key performance metric considered in this paper is the connectivity probability, which is the probability of an unblocked LoS path to at least one UAV within some maximum transmission distance. By leveraging tools from stochastic geometry, the connectivity probability is characterized as a function of the city type (e.g., urban, dense urban, suburban), density of UAVs (average number of UAVs per square km), and height of the UAVs. The city streets are modeled as a Manhattan Poisson Line Process (MPLP) and the building heights are randomly distributed. The analysis first finds the connectivity probability conditioned on a particular network realization (location of the UAVs) and then removes the conditioning to uncover the distribution of the connectivity; i.e., the fraction of network realizations that will fail to meet an outage threshold. While related work has applied an MPLP to networks with a single UAV, the contributions of this paper are that it (1) considers networks of multiple UAVs, (2) characterizes the performance by a connectivity distribution, and (3) identifies the optimal altitude for the UAVs.
Aug 10 2021
cs.LG arXiv:2108.03444v1
This paper investigates the possibility of creating a machine learning tool that automatically determines the state of mind and emotion of an individual through a questionnaire, without the aid of a human expert. The state of mind and emotion is defined in this work as pertaining to preference, feelings, or opinion that is not based on logic or reason. It is the case when a person gives out an answer to start by saying, "I feel...". The tool is designed to mimic the expertise of a psychologist and is built without any formal knowledge of psychology. The idea is to build the expertise by purely computational methods through thousands of questions collected from users. It is aimed towards possibly diagnosing substance addiction, alcoholism, sexual attraction, HIV status, degree of commitment, activity inclination, etc. First, the paper presents the related literature and classifies them according to data gathering methods. Another classification is created according to preference, emotion, grouping, and rules to achieve a deeper interpretation and better understanding of the state of mind and emotion. Second, the proposed tool is developed using an online addiction questionnaire with 10 questions and 292 respondents. In addition, an initial investigation on the dimension of addiction is presented through the built machine learning model. Machine learning methods, namely, artificial neural network (ANN) and support vector machine (SVM), are used to determine a true or false or degree of state of a respondent.
We propose a novel graph clustering method guided by additional information on the underlying structure of the clusters (or communities). The problem is formulated as the matching of a graph to a template with smaller dimension, hence matching $n$ vertices of the observed graph (to be clustered) to the $k$ vertices of a template graph, using its edges as support information, and relaxed on the set of orthonormal matrices in order to find a $k$ dimensional embedding. With relevant priors that encode the density of the clusters and their relationships, our method outperforms classical methods, especially for challenging cases.
Jun 29 2021
cs.CL arXiv:2106.14720v1
In the summer of 2020 OpenAI released its GPT-3 autoregressive language model to much fanfare. While the model has shown promise on tasks in several areas, it has not always been clear when the results were cherry-picked or when they were the unvarnished output. We were particularly interested in what benefits GPT-3 could bring to the SemEval 2021 MeasEval task - identifying measurements and their associated attributes in scientific literature. We had already experimented with multi-turn questions answering as a solution to this task. We wanted to see if we could use GPT-3's few-shot learning capabilities to more easily develop a solution that would have better performance than our prior work. Unfortunately, we have not been successful in that effort. This paper discusses the approach we used, challenges we encountered, and results we observed. Some of the problems we encountered were simply due to the state of the art. For example, the limits on the size of the prompt and answer limited the amount of the training signal that could be offered. Others are more fundamental. We are unaware of generative models that excel in retaining factual information. Also, the impact of changes in the prompts is unpredictable, making it hard to reliably improve performance.
The detection and estimation of sinusoids is a fundamental signal processing task for many applications related to sensing and communications. While algorithms have been proposed for this setting, quantization is a critical, but often ignored modeling effect. In wireless communications, estimation with low resolution data converters is relevant for reduced power consumption in wideband receivers. Similarly, low resolution sampling in imaging and spectrum sensing allows for efficient data collection. In this work, we propose SignalNet, a neural network architecture that detects the number of sinusoids and estimates their parameters from quantized in-phase and quadrature samples. We incorporate signal reconstruction internally as domain knowledge within the network to enhance learning and surpass traditional algorithms in mean squared error and Chamfer error. We introduce a worst-case learning threshold for comparing the results of our network relative to the underlying data distributions. This threshold provides insight into why neural networks tend to outperform traditional methods and into the learned relationships between the input and output distributions. In simulation, we find that our algorithm is always able to surpass the threshold for three-bit data but often cannot exceed the threshold for one-bit data. We use the learning threshold to explain, in the one-bit case, how our estimators learn to minimize the distributional loss, rather than learn features from the data.
Millimeter wave vehicular channels exhibit structure that can be exploited for beam alignment with fewer channel measurements compared to exhaustive beam search. With fixed layouts of roadside buildings and regular vehicular moving trajectory, the dominant path directions of channels will likely be among a subset of beam directions instead of distributing randomly over the whole beamspace. In this paper, we propose a deep learning-based technique to design a structured compressed sensing (CS) matrix that is well suited to the underlying channel distribution for mmWave vehicular beam alignment. The proposed approach leverages both sparsity and the particular spatial structure that appears in vehicular channels. We model the compressive channel acquisition by a two-dimensional (2D) convolutional layer followed by dropout. We design fully-connected layers to optimize channel acquisition and beam alignment. We incorporate the low-resolution phase shifter constraint during neural network training by using projected gradient descent for weight updates. Furthermore, we exploit channel spectral structure to optimize the power allocated for different subcarriers. Simulations indicate that our deep learning-based approach achieves better beam alignment than standard CS techniques which use random phase shift-based design. Numerical experiments also show that one single subcarrier is sufficient to provide necessary information for beam alignment.
We consider the problem of determining which classes of functions can be tested more efficiently than they can be learned, in the distribution-free sample-based model that corresponds to the standard PAC learning setting. Our main result shows that while VC dimension by itself does not always provide tight bounds on the number of samples required to test a class of functions in this model, it can be combined with a closely-related variant that we call "lower VC" (or LVC) dimension to obtain strong lower bounds on this sample complexity. We use this result to obtain strong and in many cases nearly optimal lower bounds on the sample complexity for testing unions of intervals, halfspaces, intersections of halfspaces, polynomial threshold functions, and decision trees. Conversely, we show that two natural classes of functions, juntas and monotone functions, can be tested with a number of samples that is polynomially smaller than the number of samples required for PAC learning. Finally, we also use the connection between VC dimension and property testing to establish new lower bounds for testing radius clusterability and testing feasibility of linear constraint systems.
Nonlinear precoding and pulse shaping are jointly considered in multi-user massive multiple-input multiple-output (MIMO) systems with low-resolution D/A-converters (DACs) in terms of algorithmic approach as well as large system performance. Two design criteria are investigated: the mean squared error (MSE) with active constellation extension (ACE) and the symbol error rate (SER). Both formulations are solved based on a modified version of the generalized approximate message passing (GAMP) algorithm. Furthermore, theoretical performance results are derived based on the state evolution analysis of the GAMP algorithm. The MSE based technique is extended to jointly perform over-the-air (OTA) spectral shaping and precoding for frequency-selective channels, in which the spectral performance is characterized at the transmitter and at the receiver. Simulation and analytical results demonstrate that the MSE based approach yields the same performance as the SER based formulation in terms of uncoded SER. The analytical results provide good performance predictions up to medium SNR. Substantial improvements in detection, as well as spectral performance, are obtained from the proposed combined pulse shaping and precoding approach compared to standard linear methods.
Millimeter-wave (mmWave) joint communication-radar (JCR) will enable high data rate communication and high-resolution radar sensing for applications such as autonomous driving. Prior JCR systems that are based on the mmWave communications hardware, however, suffer from a limited angular field-of-view and low estimation accuracy for radars due to the employed directional communication beam. In this paper, we propose an adaptive and fast combined waveform-beamforming design for the mmWave automotive JCR with a phased-array architecture that permits a trade-off between communication and radar performances. To rapidly estimate the mmWave automotive radar channel in the Doppler-angle domain with a wide field-of-view, our JCR design employs a few circulant shifts of the transmit beamformer and apply two-dimensional partial Fourier compressed sensing technique. We optimize these circulant shifts to achieve minimum coherence in compressed sensing. We evaluate the JCR performance trade-offs using a normalized mean square error (MSE) metric for radar estimation and a distortion MSE metric for data communication, which is analogous to the distortion metric in the rate distortion theory. Additionally, we develop a MSE-based weighted average optimization problem for the adaptive JCR combined waveform-beamforming design. Numerical results demonstrate that our proposed JCR design enables the estimation of short- and medium-range radar channels in the Doppler-angle domain with a low normalized MSE, at the expense of a small degradation in the communication distortion MSE.
On-demand deployments of millimeter-wave (mmWave) access points (APs) carried by unmanned aerial vehicles (UAVs) are considered today as a potential solution to enhance the performance of 5G+ networks. The battery lifetime of modern UAVs, though, limits the flight times in such systems. In this letter, we evaluate a feasible deployment alternative for temporary capacity boost in the areas with highly fluctuating user demands. The approach is to land UAV-based mmWave APs on the nearby buildings instead of hovering over the area. Within the developed mathematical framework, we compare the system-level performance of airborne and landed deployments by taking into account the full operation cycle of the employed drones. Our numerical results demonstrate that the choice of the UAV deployment option is determined by an interplay of the separation distance between the service area and the UAV charging station, drone battery lifetime, and the number of aerial APs in use. The presented methodology and results can support efficient on-demand deployments of UAV-based mmWave APs in prospective 5G+ networks.
Phased arrays, commonly used in IEEE 802.11ad and 5G radios, are capable of focusing radio frequency signals in a specific direction or a spatial region. Beamforming achieves such directional or spatial concentration of signals and enables phased array-based radios to achieve high data rates. Designing beams for millimeter wave and terahertz communication using massive phased arrays, however, is challenging due to hardware constraints and the wide bandwidth in these systems. For example, beams which are optimal at the center frequency may perform poor in wideband communication systems where the radio frequencies differ substantially from the center frequency. The poor performance in such systems is due to differences in the optimal beamformers corresponding to distinct radio frequencies within the wide bandwidth. Such a mismatch leads to a misfocus effect in near field systems and the beam squint effect in far field systems. In this paper, we investigate the misfocus effect and propose InFocus, a low complexity technique to construct beams that are well suited for massive wideband phased arrays. The beams are constructed using a carefully designed frequency modulated waveform in the spatial dimension. For the special case of beamforming along the boresight of an array, this waveform is analogous to the frequency modulated continuous wave (FMCW) chirp signal in radar. InFocus mitigates beam misfocus and beam squint when applied to near field and far field systems. Simulation results indicate that InFocus enables massive wideband phased array-based radios to achieve higher data rates than comparable beamforming solutions.
We abstract the core logical functions from applications that require ultra-low-latency wireless communications to provide a novel definition for reliability. Real-time applications -- such as intelligent transportation, remote surgery, and industrial automation -- involve a significant element of control and decision making. Such systems involve three logical components: observers (e.g. sensors) measuring the state of an environment or dynamical system, a centralized executive (e.g. controller) deciding on the state, and agents (e.g. actuators) that implement the executive's decisions. The executive harvests the observers' measurements and decides on the short-term trajectory of the system by instructing its agents to take appropriate actions. All observation packets (typically uplink) and action packets (typically downlink) must be delivered by hard deadlines to ensure the proper functioning of the controlled system. In-full on-time delivery cannot be guaranteed in wireless systems due to inherent uncertainties in the channel such as fading and unpredictable interference; accordingly, the executive will have to drop some packets. We develop a novel framework to formulate the observer selection problem (OSP) through which the executive schedules a sequence of observations that maximize its knowledge about the current state of the system. To solve this problem efficiently yet optimally, we devise a branch-and-bound algorithm that systematically prunes the search space. Our work is different from existing work on real-time communications in that communication reliability is not conveyed by packet loss or error rate, but rather by the extent of the executive's knowledge about the state of the system it controls.
Jan 22 2020
cs.CL arXiv:2001.07209v1
We present a text-based framework for investigating moral sentiment change of the public via longitudinal corpora. Our framework is based on the premise that language use can inform people's moral perception toward right or wrong, and we build our methodology by exploring moral biases learned from diachronic word embeddings. We demonstrate how a parameter-free model supports inference of historical shifts in moral sentiment toward concepts such as slavery and democracy over centuries at three incremental levels: moral relevance, moral polarity, and fine-grained moral dimensions. We apply this methodology to visualizing moral time courses of individual concepts and analyzing the relations between psycholinguistic variables and rates of moral sentiment change at scale. Our work offers opportunities for applying natural language processing toward characterizing moral sentiment change in society.
Vegetation is a relevant feature in the urban scenery and its awareness can be measured in an image by the Green View Index (GVI). Previous approaches to estimate the GVI were based upon heuristics image processing approaches and recently by deep learning networks (DLN). By leveraging some recent DLN architectures tuned to the image segmentation problem and exploiting a weighting strategy in the loss function (LF) we improved previously reported results in similar datasets.
Parkinson's disease (PD) is a common neurological disorder characterized by gait impairment. PD has no cure, and an impediment to developing a treatment is the lack of any accepted method to predict disease progression rate. The primary aim of this study was to develop a model using clinical measures and biomechanical measures of gait and postural stability to predict an individual's PD progression over two years. Data from 160 PD subjects were utilized. Machine learning models, including XGBoost and Feed Forward Neural Networks, were developed using extensive model optimization and cross-validation. The highest performing model was a neural network that used a group of clinical measures, achieved a PPV of 71% in identifying fast progressors, and explained a large portion (37%) of the variance in an individual's progression rate on held-out test data. This demonstrates the potential to predict individual PD progression rate and enrich trials by analyzing clinical and biomechanical measures with machine learning.
Vehicle-to-everything (V2X) communication in the mmWave band is one way to achieve high data-rates for applications like infotainment, cooperative perception, and augmented reality assisted driving etc. MmWave communication relies on large antennas arrays, and configuring these arrays poses high training overhead. In this article, we motivate the use of infrastructure mounted sensors (which will be part of future smart cities) for mmWave communication. We provide numerical and measurement results to demonstrate that information from these infrastructure sensors reduces the mmWave array configuration overhead. Finally, we outline future research directions to help materialize the use of infrastructure sensors for mmWave communication.
Massive MIMO is attractive for wireless information and energy transfer due to its ability to focus energy towards desired spatial locations. In this paper, the overall power transfer efficiency (PTE) and the energy efficiency (EE) of a wireless-powered massive MIMO system is investigated where a multi-antenna base-station (BS) uses wireless energy transfer to charge single-antenna energy harvesting users on the downlink. The users may exploit the harvested energy to transmit information to the BS on the uplink. The overall system performance is analyzed while accounting for the nonlinear nature of practical energy harvesters. First, for wireless energy transfer, the PTE is characterized using a scalable model for the BS circuit power consumption. The PTE-optimal number of BS antennas and users are derived. Then, for wireless energy and information transfer, the EE performance is characterized. The EE-optimal BS transmit power is derived in terms of the key system parameters such as the number of BS antennas and the number of users. As the number of antennas becomes large, increasing the transmit power improves the energy efficiency for moderate to large number of antennas. Simulation results suggest that it is energy efficient to operate the system in the massive antenna regime.
Aug 19 2019
cs.CL arXiv:1908.05760v1
Biomedical Named Entity Recognition (NER) is a challenging problem in biomedical information processing due to the widespread ambiguity of out of context terms and extensive lexical variations. Performance on bioNER benchmarks continues to improve due to advances like BERT, GPT, and XLNet. FLAIR (1) is an alternative embedding model which is less computationally intensive than the others mentioned. We test FLAIR and its pretrained PubMed embeddings (which we term BioFLAIR) on a variety of bio NER tasks and compare those with results from BERT-type networks. We also investigate the effects of a small amount of additional pretraining on PubMed content, and of combining FLAIR and ELMO models. We find that with the provided embeddings, FLAIR performs on-par with the BERT networks - even establishing a new state of the art on one benchmark. Additional pretraining did not provide a clear benefit, although this might change with even more pretraining being done. Stacking the FLAIR embeddings with others typically does provide a boost in the benchmark results.
The capability of smarter networked devices to dynamically select appropriate radio connectivity options is especially important in the emerging millimeter-wave (mmWave) systems to mitigate abrupt link blockage in complex environments. To enrich the levels of diversity, mobile mmWave relays can be employed for improved connection reliability. These are considered by 3GPP for on-demand densification on top of the static mmWave infrastructure. However, performance dynamics of mobile mmWave relaying is not nearly well explored, especially in realistic conditions, such as urban vehicular scenarios. In this paper, we develop a mathematical framework for the performance evaluation of mmWave vehicular relaying in a typical street deployment. We analyze and compare alternative connectivity strategies by quantifying the performance gains made available to smart devices in the presence of mmWave relays. We identify situations where the use of mmWave vehicular relaying is particularly beneficial. Our methodology and results can support further standardization and deployment of mmWave relaying in more intelligent 5G+ "all-mmWave" cellular networks.
Aug 05 2019
cs.CV arXiv:1908.00778v1
The importance of imaging exams, such as Magnetic Resonance Imaging (MRI), for the diagnostic and follow-up of pediatric pathologies and the assessment of anatomical structures' development has been increasingly highlighted in recent times. Manual analysis of MRIs is time-consuming, subjective, and requires significant expertise. To mitigate this, automatic techniques are necessary. Most techniques focus on adult subjects, while pediatric MRI has specific challenges such as the ongoing anatomical and histological changes related to normal development of the organs, reduced signal-to-noise ratio due to the smaller bodies, motion artifacts and cooperation issues, especially in long exams, which can in many cases preclude common analysis methods developed for use in adults. Therefore, the development of a robust technique to aid in pediatric MRI analysis is necessary. This paper presents the current development of a new method based on the learning and matching of structural relational graphs (SRGs). The experiments were performed on liver MRI sequences of one patient from ICr-HC-FMUSP, and preliminary results showcased the viability of the project. Future experiments are expected to culminate with an application for pediatric liver substructure and brain tumor segmentation.
This paper introduces the concept of kernels on fuzzy sets as a similarity measure for $[0,1]$-valued functions, a.k.a. \emphmembership functions of fuzzy sets. We defined the following classes of kernels: the cross product, the intersection, the non-singleton and the distance-based kernels on fuzzy sets. Applicability of those kernels are on machine learning and data science tasks where uncertainty in data has an ontic or epistemistic interpretation.
Millimeter wave (mmWave) communication in typical wearable and data center settings is short range. As the distance between the transmitter and the receiver in short range scenarios can be comparable to the length of the antenna arrays, the common far field approximation for the channel may not be applicable. As a result, dictionaries that result in a sparse channel representation in the far field setting may not be appropriate for short distances. In this paper, we develop a novel framework to exploit the structure in short range mmWave channels. The proposed method splits the channel into several subchannels for which the far field approximation can be applied. Then, the structure within and across different subchannels is leveraged using message passing. We show how information about the antenna array geometry can be used to design message passing factors that incorporate structure across successive subchannels. Simulation results indicate that our framework can be used to achieve better beam alignment with fewer channel measurements when compared to standard compressed sensing-based techniques that do not exploit structure across subchannels.
At present, different deep learning models are presenting high accuracy on popular inference datasets such as SNLI, MNLI, and SciTail. However, there are different indicators that those datasets can be exploited by using some simple linguistic patterns. This fact poses difficulties to our understanding of the actual capacity of machine learning models to solve the complex task of textual inference. We propose a new set of syntactic tasks focused on contradiction detection that require specific capacities over linguistic logical forms such as: Boolean coordination, quantifiers, definite description, and counting operators. We evaluate two kinds of deep learning models that implicitly exploit language structure: recurrent models and the Transformer network BERT. We show that although BERT is clearly more efficient to generalize over most logical forms, there is space for improvement when dealing with counting operators. Since the syntactic tasks can be implemented in different languages, we show a successful case of cross-lingual transfer learning between English and Portuguese.
Joint communication and radar (JCR) waveforms with fully digital baseband generation and processing can now be realized at the millimeter-wave (mmWave) band. Prior work has proposed a mmWave wireless local area network (WLAN)-based JCR that exploits the WLAN preamble for radars. The performance of target velocity estimation, however, was limited. In this paper, we propose a virtual waveform design for an adaptive mmWave JCR. The proposed system transmits a few non-uniformly placed preambles to construct several receive virtual preambles for enhancing velocity estimation accuracy, at the cost of only a small reduction in the communication data rate. We evaluate JCR performance trade-offs using the Cramer-Rao Bound (CRB) metric for radar estimation and a novel distortion minimum mean square error (MMSE) metric for data communication. Additionally, we develop three different MMSE-based optimization problems for the adaptive JCR waveform design. Simulations show that an optimal virtual (non-uniform) waveform achieves a significant performance improvement as compared to a uniform waveform. For a radar CRB constrained optimization, the optimal radar range of operation and the optimal communication distortion MMSE (DMMSE) are improved. For a communication DMMSE constrained optimization with a high DMMSE constraint, the optimal radar CRB is enhanced. For a weighted MMSE average optimization, the advantage of the virtual waveform over the uniform waveform is increased with decreased communication weighting. Comparison of MMSE-based optimization with traditional virtual preamble count-based optimization indicated that the conventional solution converges to the MMSE-based one only for a small number of targets and a high signal-to-noise ratio.
Mar 19 2019
cs.CV arXiv:1903.06949v1
We introduce a framework for dynamic evaluation of the fingers movements: flexion, extension, abduction and adduction. This framework estimates angle measurements from joints computed by a hand pose estimation algorithm using a depth sensor (Realsense SR300). Given depth maps as input, our framework uses Pose-REN, which is a state-of-art hand pose estimation method that estimates 3D hand joint positions using a deep convolutional neural network. The pose estimation algorithm runs in real-time, allowing users to visualise 3D skeleton tracking results at the same time as the depth images are acquired. Once 3D joint poses are obtained, our framework estimates a plane containing the wrist and MCP joints and measures flexion/extension and abduction/aduction angles by applying computational geometry operations with respect to this plane. We analysed flexion and abduction movement patterns using real data, extracting the movement trajectories. Our preliminary results show that this method allows an automatic discrimination of hands with Rheumatoid Arthritis (RA) and healthy patients. The angle between joints can be used as an indicative of current movement capabilities and function. Although the measurements can be noisy and less accurate than those obtained statically through goniometry, the acquisition is much easier, non-invasive and patient-friendly, which shows the potential of our approach. The system can be used with and without orthosis. Our framework allows the acquisition of measurements with minimal intervention and significantly reduces the evaluation time.
Massive MIMO, a key technology for increasing area spectral efficiency in cellular systems, was developed assuming moderately sized apertures. In this paper, we argue that massive MIMO systems behave differently in large-scale regimes due to spatial non-stationarity. In the large-scale regime, with arrays of around fifty wavelengths, the terminals see the whole array but non-stationarities occur because different regions of the array see different propagation paths. At even larger dimensions, which we call the extra-large scale regime, terminals see a portion of the array and inside the first type of non-stationarities might occur. We show that the non-stationarity properties of the massive MIMO channel changes several important MIMO design aspects. In simulations, we demonstrate how non-stationarity is a curse when neglected but a blessing when embraced in terms of computational load and multi-user transceiver design.
Spatial channel covariance information can replace full instantaneous channel state information for the analog precoder design in hybrid analog/digital architectures. Obtaining spatial channel covariance estimation, however, is challenging in the hybrid structure due to the use of fewer radio frequency (RF) chains than the number of antennas. In this paper, we propose a spatial channel covariance estimation method based on higher-order tensor decomposition for spatially sparse time-varying frequency-selective channels. The proposed method leverages the fact that the channel can be represented as a low-rank higher-order tensor. We also derive the Cramér-Rao lower bound on the estimation accuracy of the proposed method. Numerical results and theoretical analysis show that the proposed tensor-based approach achieves higher estimation accuracy in comparison with prior compressive-sensing-based approaches or conventional angle-of-arrival estimation approaches. Simulation results reveal that the proposed approach becomes more beneficial at low signal-to-noise (SNR) region.