The meaning of the difference of two entropy values
I want to understanding the meaning of the difference of two information entropy values.
I have the following scenario. Let $x$ be a number of hours a user spend on some video sharing websites. Thus, we may have the sets:
$X_{A} = \{x_1,x_2,\cdots,x_{n_A}\}$, and $X_{B} = \{x_1,x_2,\cdots,x_{n_B}\}$ that represent the number of hours the users of $A$ and $B$ spent on the websites $A$ and $B$, respectively.
Now, we can calculate the Cumulative distribution function (CDF) probability values, described here, as follows: For a real-valued random variable X, the CDF is defined as:
$$F_A(x) = P(X\leq x),\ \forall x \in X_A$$
Similarly, we find $F_B(x)$.
Then we calculate the entropy values $E_A$ and $E_B$ for the probabilities $F_A(x)$ and $F_B(x)$ obtained above (as described here):
$$E_A = - \sum\limits_{i=1}^{n_A}P(x_i)log(P(x_i))$$
Similarly, we find $E_B$.
Now my question is, what do $E_A$ and $E_B$ mean in this scenario? and what their difference also means?
Topic information-theory
Category Data Science