The meaning of the difference of two entropy values

I want to understanding the meaning of the difference of two information entropy values.

I have the following scenario. Let $x$ be a number of hours a user spend on some video sharing websites. Thus, we may have the sets:

$X_{A} = \{x_1,x_2,\cdots,x_{n_A}\}$, and $X_{B} = \{x_1,x_2,\cdots,x_{n_B}\}$ that represent the number of hours the users of $A$ and $B$ spent on the websites $A$ and $B$, respectively.

Now, we can calculate the Cumulative distribution function (CDF) probability values, described here, as follows: For a real-valued random variable X, the CDF is defined as:

$$F_A(x) = P(X\leq x),\ \forall x \in X_A$$

Similarly, we find $F_B(x)$.

Then we calculate the entropy values $E_A$ and $E_B$ for the probabilities $F_A(x)$ and $F_B(x)$ obtained above (as described here):

$$E_A = - \sum\limits_{i=1}^{n_A}P(x_i)log(P(x_i))$$

Similarly, we find $E_B$.

Now my question is, what do $E_A$ and $E_B$ mean in this scenario? and what their difference also means?

Topic information-theory

Category Data Science

About

Geeks Mental is a community that publishes articles and tutorials about Web, Android, Data Science, new techniques and Linux security.