2023-9 trading log

18
Entropy is a measure of the amount of disorder or randomness in a system. In thermodynamics, it is a measure of the unavailability of a system's thermal energy for conversion into mechanical work, often interpreted as the degree of disorder or randomness in the system. In information theory, entropy is a measure of the uncertainty or randomness of information content in a message or signal. The greater the entropy in a message or signal, the less predictable it is.

Clause de non-responsabilité

Les informations et les publications ne sont pas destinées à être, et ne constituent pas, des conseils ou des recommandations en matière de finance, d'investissement, de trading ou d'autres types de conseils fournis ou approuvés par TradingView. Pour en savoir plus, consultez les Conditions d'utilisation.