.


Distribution series are series of absolute and relative numbers that characterize the distribution of population units according to a qualitative (attributive) or quantitative attribute. Distribution series built on a quantitative basis are called variational421.


Divisive clustering – see hierarchical clustering422,423.


Documentation generically, any information on the structure, contents, and layout of a data file. Sometimes called «technical documentation» or «a codebook». Documentation may be considered a specialized form of metadata424.


Documented information – information recorded on a material carrier by means of documentation with details that make it possible to determine such information, or, in cases established by the legislation of the Russian Federation, its material carrier425.


Downsampling – overloaded term that can mean either of the following: Reducing the amount of information in a feature in order to train a model more efficiently. For example, before training an image recognition model, downsampling high-resolution images to a lower-resolution format. Training on a disproportionately low percentage of over-represented class examples in order to improve model training on under-represented classes. For example, in a class-imbalanced dataset, models tend to learn a lot about the majority class and not enough about the minority class. Downsampling helps balance the amount of training on the majority and minority classes426.


Driver is computer software that allows other software (the operating system) to access the hardware of a device427.


Drone – unmanned aerial vehicle (unmanned aerial system)428.


Dropout regularization is a form of regularization useful in training neural networks. Dropout regularization works by removing a random selection of a fixed number of the units in a network layer for a single gradient step. The more units dropped out, the stronger the regularization429.


Dynamic epistemic logic (DEL) is a logical framework dealing with knowledge and information change. Typically, DEL focuses on situations involving multiple agents and studies how their knowledge changes when events occur430.


Dynamic model is a model that is trained online in a continuously updating fashion. That is, data is continuously entering the model431,432.

«E»

Eager execution is a TensorFlow programming environment in which operations run immediately. By contrast, operations called in graph execution don’t run until they are explicitly evaluated. Eager execution is an imperative interface, much like the code in most programming languages. Eager execution programs are generally far easier to debug than graph execution programs433.


Eager learning is a learning method in which the system tries to construct a general, input-independent target function during training of the system, as opposed to lazy learning, where generalization beyond the training data is delayed until a query is made to the system434.


Early stopping is a method for regularization that involves ending model training before training loss finishes decreasing. In early stopping, you end model training when the loss on a validation dataset starts to increase, that is, when generalization performance worsens435.


Earth mover’s distance (EMD) is a measure of the relative similarity between two documents. The lower the value, the more similar the documents436.


Ebert test is a test which gauges whether a computer-based synthesized voice can tell a joke with sufficient skill to cause people to laugh. It was proposed by film critic Roger Ebert at the 2011 TED conference as a challenge to software developers to have a computerized voice master the inflections, delivery, timing, and intonations of a speaking human. The test is similar to the Turing test proposed by Alan Turing in 1950 as a way to gauge a computer’s ability to exhibit intelligent behavior by generating performance indistinguishable from a human being