Глоссариум по искусственному интеллекту: 2500 терминов. Том 2 - страница 32
Data set is a set of data that has undergone preliminary preparation (processing) in accordance with the requirements of the legislation of the Russian Federation on information, information technology and information protection and is necessary for the development of software based on artificial intelligence (National strategy for the development of artificial intelligence for the period up to 2030)340.
Data Streaming Accelerator (DSA) is a device that performs a specific task, which in this case is the transfer of data in less time than the CPU would do. What makes DSA special is that it is designed for one of the characteristics that Compute Express Link brings with it over PCI Express 5.0, which is to provide consistent access to RAM for all peripherals connected to a PCI Express port, i.e., they use the same memory addresses.
Data variability describes how far apart data points lie from each other and from the center of a distribution. Along with measures of central tendency, measures of variability give you descriptive statistics that summarize your data341.
Data veracity is the degree of accuracy or truthfulness of a data set. In the context of big data, its not just the quality of the data that is important, but how trustworthy the source, the type, and processing of the data are342.
Data Warehouse is typically an offline copy of production databases and copies of files in a non-production environment343.
Database is a «container» storing data such as numbers, dates or words, which can be reprocessed by computer means to produce information; for example, numbers and names assembled and sorted to form a directory344.
DataFrame is a popular datatype for representing datasets in pandas. A DataFrame is analogous to a table. Each column of the DataFrame has a name (a header), and each row is identified by a number345.
Datalog is a declarative logic programming language that syntactically is a subset of Prolog. It is often used as a query language for deductive databases. In recent years, Datalog has found new application in data integration, information extraction, networking, program analysis, security, and cloud computing346.
Datamining – the discovery, interpretation, and communication of meaningful patterns in data347.
Dataset API (tf. data) is a high-level TensorFlow API for reading data and transforming it into a form that a machine learning algorithm requires. A tf. data. Dataset object represents a sequence of elements, in which each element contains one or more Tensors. A tf.data.Iterator object provides access to the elements of a Dataset. For details about the Dataset API, see Importing Data in the TensorFlow Programmer’s Guide348.
Debugging is the process of finding and resolving bugs (defects or problems that prevent correct operation) within computer programs, software, or systems. Debugging tactics can involve interactive debugging, control flow analysis, unit testing, integration testing, log file analysis, monitoring at the application or system level, memory dumps, and profiling. Many programming languages and software development tools also offer programs to aid in debugging, known as debuggers349.
Decentralized applications (dApps) are digital applications or programs that exist and run on a blockchain or peer-to-peer (P2P) network of computers instead of a single computer. DApps (also called «dapps») are outside the purview and control of a single authority. DApps – which are often built on the Ethereum platform – can be developed for a variety of purposes including gaming, finance, and social media