Mutual Information Rate of Gaussian and Truncated Gaussian Inputs on Intensity-Driven Signal Transduction Channels

نویسندگان

چکیده

In this letter, we investigate the mutual information rate (MIR) achieved by an independent identically distributed (IID) Gaussian input on intensity-driven signal transduction channel. Specifically, asymptotic expression of continuous-time MIR is given. Next, aiming at low computational complexity, also deduce approximately numerical solution for MIR. Moreover, corresponding lower and upper bounds that can be used to find capacity-achieving distribution parameters are derived in closed-form. Finally, simulation results show accuracy our analysis.

برای دانلود باید عضویت طلایی داشته باشید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

On the Achievable Rate-Regions for the Gaussian Two-way Diamond Channels

In this channel,we study rate region of a Gaussian two-way diamond channel which operates in half-duplex mode. In this channel, two transceiver (TR) nodes exchange their messages with the help of two relay nodes. We consider a special case of the Gaussian two-way diamond channels which is called Compute-and-Forward Multiple Access Channel (CF-MAC). In the CF-MAC, the TR nodes transmit their mes...

متن کامل

Information Capacity of Gaussian Channels

i Information capacity of Gaussian channels is one of the basic problems of information theory. Shannon's results for white Gaussian channels and Fano's "waterfilling" analysis of stationary Gaussian channels are two of the best-known works of early information theory. Results are given here which extend to a general framework these results and others due to Gallager and to Kadota, Zakai, and Z...

متن کامل

The Rate of Entropy for Gaussian Processes

In this paper, we show that in order to obtain the Tsallis entropy rate for stochastic processes, we can use the limit of conditional entropy, as it was done for the case of Shannon and Renyi entropy rates. Using that we can obtain Tsallis entropy rate for stationary Gaussian processes. Finally, we derive the relation between Renyi, Shannon and Tsallis entropy rates for stationary Gaussian proc...

متن کامل

Gaussian Process Optimization with Mutual Information

In this paper, we analyze a generic algorithm scheme for sequential global optimization using Gaussian processes. The upper bounds we derive on the cumulative regret for this generic algorithm improve by an exponential factor the previously known bounds for algorithms like GPUCB. We also introduce the novel Gaussian Process Mutual Information algorithm (GP-MI), which significantly improves furt...

متن کامل

Estimating Mutual Information by Local Gaussian Approximation

Estimating Mutual Information by Local Gaussian Approximation Report Title Estimating mutual information (MI) from samples is a fundamental problem in statistics, machine learning, and data analysis. Recently it was shown that a popular class of non-parametric MI estimators perform very poorly for strongly dependent variables and have sample complexity that scales exponentially with the true MI...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: IEEE Communications Letters

سال: 2023

ISSN: ['1558-2558', '1089-7798', '2373-7891']

DOI: https://doi.org/10.1109/lcomm.2023.3289449