Data modeling is the process of creating a simplified diagram of a software system and the data elements it contains, using text and symbols to represent the data and how it flows. Data models provide a blueprint for designing a new database or reengineering a legacy application. Overall, data modeling helps an organization use its data effectively to meet business needs for information350
Data portability allows individuals to obtain and reuse their personal data for their own purposes across different services. It allows them to move, copy or transfer personal data easily from one IT environment to another in a safe and secure way, without affecting its usability351.
Data Privacy the assurance that a persons or organizations personal and private information is not inappropriately disclosed. Ensuring Data Privacy requires Access Management, eSecurity, and other data protection efforts352.
Data Processing within the field of information technology, typically means the processing of information by machines. Data processing is defined by procedures designed to make a data collection easier to use, ensure its accuracy, enhance its utility, optimize its format, protect confidentiality, etc. For archival purposes, the process and results of data processing must be systematically and comprehensively captured so that the process applied to the data is transparent to users353.
Data Processor (or Processor) the natural or legal person, or any other body, which processes personal data on behalf of the controller354.
Data Protection Authority monitors and supervises, through investigative and corrective powers, the application of the data protection law. It provides expert advice on data protection issues and handle complaints that may have breached the law355.
Data protection is the process of protecting data and involves the relationship between the collection and dissemination of data and technology, the public perception and expectation of privacy and the political and legal underpinnings surrounding that data. It aims to strike a balance between individual privacy rights while still allowing data to be used for business purposes356.
Data Protection Officer ensures that the organisation processes the personal data of its staff, customers, providers or any other individuals (also referred to as data subjects) in compliance with the applicable data protection rules357.
Data Requestor person or institution that is looking for data and provides the necessary infrastructure, e.g. a publicly available Semantic Container initialized with a semantic description of the data request and intended purpose of the collected data358.
Data Science is a broad grouping of mathematics, statistics, probability, computing, data visualization to extract knowledge from a heterogeneous set of data (images, sound, text, genomic data, social network links, physical measurements, etc.). The methods and tools derived from artificial intelligence are part of this family. Data science is the field of study that combines domain expertise, programming skills, and knowledge of mathematics and statistics to extract meaningful insights from data. Data science practitioners apply machine learning algorithms to numbers, text, images, video, audio, and more to produce artificial intelligence (AI) systems to perform tasks that ordinarily require human intelligence. In turn, these systems generate insights which analysts and business users can translate into tangible business value. Data Science is an interdisciplinary field that uses scientific methods, processes, algorithms and systems to extract knowledge and insights from structured and unstructured data, and apply knowledge and actionable insights from data across a broad range of application domains. Also, Data Science this is an academic/professional field that comprises several components for data analysis and interpretation through mathematics, statistics and information technology. Thus, a data scientist not only collects and analyzes inputs, but also interprets and relates the facts to the context in which they are inserted359,360,361.
Data set is a collection of data. In the case of tabular data, a data set corresponds to one or more database tables, where every column of a table represents a particular variable, and each row corresponds to a given record of the data set in question. The data set lists values for each of the variables, such as for example height and weight of an object, for each member of the data set. Data sets can also consist of a collection of documents or files. Data set a collection of data records. In the SAS statistical software, a «SAS data set» is the internal representation of data. Also, Data set is a set of data that has undergone preliminary preparation (processing) in accordance with the requirements of the legislation of the Russian Federation on information, information technology and information protection and is necessary for the development of software based on artificial intelligence (National strategy for the development of artificial intelligence for the period up to 2030)362,363.
Data Sharing the disclosure of data from one or more organizations to a third party organisation or organizations, or the sharing of data between different parts of an organisation364.
Data Sharing Agreement common set of rules to be adopted by the various organizations involved in a data sharing operation365.
Data sharing governance concept changing «ownership of data-to-data control and data sharing governance366.
Data silos are repositories of fixed data that remain under the control of one group or department and that are isolated from the rest of the organization367.
Data source is the primary location where the data that is being used comes from368.
Data Stakeholders those who use, affect, or are affected by data. Data Stakeholders may be upstream producers, gatherers, or acquirers of information; downstream consumers of information, those who manage, transform, or store data, or those who set policies, standards, architectures, or other requirements or constraints369.
Data Steward is a person with data-related responsibilities as set by a Data Governance or Data Stewardship program. Often, Data Stewards fall into multiple types. Data Quality Stewards, Data Definition Stewards, Data Usage Stewards, etc.370.
Data Subject is the person whose personal data are collected, held or processed. identified or identifiable natural person, who is the subject of personal data371.
Data transfer rate (DTR) is the amount of digital data that is moved from one place to another in a given time. The data transfer rate can be viewed as the speed of travel of a given amount of data from one place to another. In general, the greater the bandwidth of a given path, the higher the data transfer rate372.
Data variability describes how far apart data points lie from each other and from the center of a distribution. Along with measures of central tendency, measures of variability give you descriptive statistics that summarize your data373.
Data veracity is the degree of accuracy or truthfulness of a data set. In the context of big data, its not just the quality of the data that is important, but how trustworthy the source, the type, and processing of the data are374.
Database is an organized collection of structured information, or data, typically stored electronically in a computer system. A database is usually controlled by a database management system (DBMS). Together, the data and the DBMS, along with the applications that are associated with them, are referred to as a database system, often shortened to just database. Data within the most common types of databases in operation today is typically modeled in rows and columns in a series of tables to make processing and data querying efficient. The data can then be easily accessed, managed, modified, updated, controlled, and organized. Most databases use structured query language (SQL) for writing and querying data375.
Database management system (DBMS) is a software package designed to define, manipulate, retrieve and manage data in a database. A DBMS generally manipulates the data itself, the data format, field names, record structure and file structure. It also defines rules to validate and manipulate this data. Database management systems are set up on specific data handling concepts, as the practice of administrating a database evolves. The earliest databases only handled individual single pieces of specially formatted data. Todays more evolved systems can handle different kinds of less formatted data and tie them together in more elaborate ways376.
Databus is a data-centric sharing system where applications exchange information in a virtual, global data space377.
Data-driven decisions are decisions made based on data/information, not experience, hunches, or intuition378.
Dataflow Processing Unit (DPU) is a programmable specialized electronic circuit with hardware accelerated data processing for data-oriented computing.
DDI instance an XML document, marked up according to the DDI DTD. In other words, a codebook or catalog record marked up in DDI-compliant XML379.
Debugging is the process of finding and resolving bugs (defects or problems that prevent correct operation) within computer programs, software, or systems. Debugging tactics can involve interactive debugging, control flow analysis, unit testing, integration testing, log file analysis, monitoring at the application or system level, memory dumps, and profiling. Many programming languages and software development tools also offer programs to aid in debugging, known as debuggers380.
Decentralized applications (dApps) are digital applications or programs that exist and run on a blockchain or peer-to-peer (P2P) network of computers instead of a single computer. DApps (also called «dapps») are outside the purview and control of a single authority. DApps which are often built on the Ethereum platform can be developed for a variety of purposes including gaming, finance, and social media381.
Decentralized control is a process in which a significant number of control actions related to a given object are generated by the object itself on the basis of self-government.
Decentralized finance (DeFi) is an emerging financial technology based on secure distributed ledgers similar to those used by cryptocurrencies. The system removes the control banks and institutions have on money, financial products, and financial services382.
Decision intelligence (DI) is a practical discipline used to improve the decision making process by clearly understanding and programmatically developing how decisions are made and how the outcomes are evaluated, managed and improved through feedback. Also, Decision intelligence is a discipline offers a framework to assist data and analytics practitioners develop, model, align, implement, track, and modify decision models and processes related to business results and performance.
Decision Rights the system of determining who makes a decision, and when, and how, and under what circumstances. Formalizing Decision Rights is a key function of Data Governance383.
Decision support system (DSS) is an information system that supports business or organizational decision-making activities. DSSs serve the management, operations and planning levels of an organization (usually mid and higher management) and help people make decisions about problems that may be rapidly changing and not easily specified in advance i.e. unstructured and semi-structured decision problems. Decision support systems can be either fully computerized or human-powered, or a combination of both. Also, Decision Support Systems is a collection of integrated technologies, software and hardware, that constitute the main support of the organization`s decision making process384.
Decision tree is a tree-and-branch model used to represent decisions and their possible consequences, similar to a flowchart.
Decompression is a feature that is used to restore data to uncompressed form after compression385.
Deep Learning (also known as deep structured learning) is part of a broader family of machine learning methods based on artificial neural networks with representation learning. Learning can be supervised, semi-supervised or unsupervised. Deep-learning architectures such as deep neural networks, deep belief networks, deep reinforcement learning, recurrent neural networks and convolutional neural networks have been applied to fields including computer vision, speech recognition, natural language processing, machine translation, bioinformatics, drug design, medical image analysis, climate science, material inspection and board game programs, where they have produced results comparable to and in some cases surpassing human expert performance. Also, Deep Learning (DL) is a subfield of machine learning concerned with algorithms that are inspired by the human brain that works in a hierarchical way. Deep Learning models, which are mostly based on the (artificial) neural networks, have been applied to different fields, such as speech recognition, computer vision, and natural language processing386.
Deep neural network a multilayer network containing several (many) hidden layers of neurons between the input and output layers, which allows modeling complex nonlinear relationships. GNNs are now increasingly used to solve such artificial intelligence problems as speech recognition, natural language processing, computer vision, etc., including in robotics387.
Deep Technology (DEEP TECH) refers to a startup whose business idea is based on a scientific or otherwise extensive (deep) understanding of technology. The term has been adopted to set certain companies apart from other startups which are also technology driven. A deep tech company may, for instance, base the core of its operations on particularly complex mathematics in the creation of software algorithms. Deep technology companies typically comprise artificial intelligence companies, which try to replicate human thinking, build navigation systems for flying cars and so on388.
DeepMind is an artificial intelligence company founded in 2010 and later acquired by Google in 2014. DeepMind developed AlphaGo program that beat a human professional Go player for the first time.
Default access controls the access controls that apply where a registered individual has not set controls on the registered healthcare provider organizations or nominated representatives who may access the individuals My Health Record. This means that any registered healthcare provider organisation involved in your healthcare can access your record389.