Information Stocks List
|2019-10-18||DHX||Calm After Storm||Range Contraction|
|2019-10-18||DHX||Crossed Above 20 DMA||Bullish|
|2019-10-18||GCT.C||Fell Below 20 DMA||Bearish|
|2019-10-18||GCT.C||Narrow Range Bar||Range Contraction|
|2019-10-18||GVC||Narrow Range Bar||Range Contraction|
|2019-10-18||LIVE||Narrow Range Bar||Range Contraction|
|2019-10-18||LIVE||Non-ADX 1,2,3,4 Bullish||Bullish Swing Setup|
Information is the resolution of uncertainty; it is that which answers the question of "what an entity is" and is thus that which specifies the nature of that entity, as well as the essentiality of its properties. Information is associated with data and knowledge, as data is meaningful information and represents the values attributed to parameters, and knowledge signifies understanding of an abstract or concrete concept. The existence of information is uncoupled with an observer, which refers to that which accesses information to discern that which it specifies; information exists beyond an event horizon for example, and in the case of knowledge, the information itself requires a cognitive observer to be accessed.
In terms of communication, information is expressed either as the content of a message or through direct or indirect observation. That which is perceived can be construed as a message in its own right, and in that sense, information is always conveyed as the content of a message.
Information can be encoded into various forms for transmission and interpretation (for example, information may be encoded into a sequence of signs, or transmitted via a signal). It can also be encrypted for safe storage and communication.
Information reduces uncertainty. The uncertainty of an event is measured by its probability of occurrence and is inversely proportional to that. The more uncertain an event, the more information is required to resolve uncertainty of that event. The bit is a typical unit of information, but other units such as the nat may be used. For example, the information encoded in one "fair" coin flip is log2(2/1) = 1 bit, and in two fair coin flips is
log2(4/1) = 2 bits.
The concept of information has different meanings in different contexts. Thus the concept becomes related to notions of constraint, communication, control, data, form, education, knowledge, meaning, understanding, mental stimuli, pattern, perception, representation, and entropy.