An Active Data Warehouse is a combination of products, features, services, and business partnerships that support the Active Enterprise Intelligence business strategy. This term was coined by Teradata in 2001.
Read more
Within the context of big data, algorithms are the primary means for uncovering insights and detecting patterns. Thus, they are essential to realizing the big data business case.
Read more
An analytics platform is a full-featured technology solution designed to address the needs of large enterprises. Typically, it joins different tools and analytics systems together with an engine to execute, a database or repository to store and manage the data, data mining processes, and techniques and mechanisms for obtaining and preparing data that is not stored.
Read more
Apache Hive is an open-source data warehouse infrastructure that provides tools for data summarization, query and analysis. It is specifically designed to support the analysis of large datasets stored in Hadoop files and compatible file systems, such as Amazon S3. Hive was initially developed by data engineers at Facebook in 2008, but is now used by many other companies.
Read more
Behavioral analytics measure how users engage with digital applications (web, mobile, IoT) and how seemingly unrelated data points can explain or predict outcomes.
Read more
Big data is a group of data sets too large and complex to manipulate or query with standard tools.
Read more
Big data analytics refers to the strategy of analyzing large volumes of data gathered from a wide variety of sources, including social networks, videos, digital images, sensors and sales transaction records.
Read more
A Business Continuity Plan (BCP) is a detailed document that outlines how an organisation will continue to operate through an unplanned disruption in service. The BCP is not just a regulatory requirement within many industries, but should be considered as a guide to reduce the time it takes for operations to return to normal. It plays a critical role in an organisation’s Operational Resilience.
Read more
Business Intelligence (BI) parses business data and presents easy-to-digest reports, performance measures, and trends that drive management decisions.
Read more
Cascading is a platform for developing Big Data applications on Hadoop. It offers a computation engine, systems integration framework, data processing and scheduling capabilities.
Read more
A Customer Data Platform (CDP) is a type of packaged software which creates a persistent, unified customer database that is accessible to other systems.
Read more
Cloud computing refers to the practice of using a network of remote servers to store, manage and process data (rather than an on-premise server or a personal computer) with access to such data provided through the Internet (the cloud).
Read more
Cluster analysis or clustering is a statistical classification technique or activity that involves grouping a set of objects or data so that those in the same group (called a cluster) are similar to each other, but different from those in other clusters.
Read more
Comparative analysis refers to the comparison of two or more processes, documents, data sets or other objects. Pattern analysis, filtering and decision-tree analytics are forms of comparative analysis.
Read more
Concurrency or concurrent computing refers to the form of computing in which multiple computing tasks occur simultaneously or at overlapping times. These tasks can be handled by individual computers, specific applications or across networks.
Read more
Connection analytics is an emerging discipline that helps to discover interrelated connections and influences between people, products, processes, machines and systems within a network by mapping those connections and continuously monitoring interactions between them.
Read more
Correlation analysis refers to the application of statistical analysis and other mathematical techniques to evaluate or measure the relationships between variables.
Read more
Data analysts serve the critical purpose of helping to operationalize big data within specific functions and processes, with a clear focus on performance trends and operational information.
Read more
Data analytics, also known as advanced analytics or big data analytics, is an autonomous or semi-autonomous inspection of data or content using sophisticated techniques and tools beyond those of traditional business intelligence (BI), to uncover deeper insights, make predictions, or produce recommendations. Techniques include data/text mining, machine learning, pattern matching, forecasting, visualization, semantic analysis, sentiment analysis, network and cluster analysis, multivariate statistics, graph analysis, simulation, complex event processing, neural networks.
Read more
Teradata Unified Data Architecture is the first comprehensive big data architecture. This framework harnesses relational and non-relational repositories via SQL and non-SQL analytics.
Read more
Data cleansing, or data scrubbing, is the process of detecting and correcting or removing inaccurate data or records from a database.
Read more
Data gravity appears when the amount of data volume in a repository grows, along with the number of uses, and makes the ability to copy or migrate data onerous and expensive.
Read more
Data lakes complement data warehouses with a design pattern that focuses on original raw data fidelity and long-term storage at a low cost while providing a new form of analytical agility.
Read more
Data latency is the ability to load and update data in near real-time while simultaneously supporting query workloads.
Read more
A data mart is a subject-oriented slice of the data warehouse logical model serving a narrow group of users.
Read more
Data mining is the process of analyzing hidden patterns of data according to different perspectives for categorization into useful information, which is collected and assembled in common areas, such as data warehouses.
Read more
Data models that are tailored to specific industries or business functions can provide a strong foundation or "jump-start" for big data programs and investments.
Read more
Data volume is the storing and processing of petabytes of data natively and in object storage.
Read more
In computing, a data warehouse (DW or DWH), also known as an enterprise data warehouse (EDW), is a system used for reporting and data analysis.
Read more
Deep learning, also known as deep neural learning or deep neural network, is an artificial intelligence (AI) function that mimics how the human brain works to process data and create patterns that facilitate decision making.
Read more
Descriptive analytics are the analysis of historical data to determine what happened, what changed and what patterns can be identified.
Read more
Extract, Transform and Load (ETL) refers to the process in data warehousing that concurrently reads (or extracts) data from source systems; converts (or transforms) the data into the proper format for querying and analysis; and loads it into a data warehouse, operational data store or data mart).
Read more
An extraordinarily large unit of digital data, one Exabyte (EB) is equal to 1,000 Petabytes or one billion gigabytes (GB). Some technologists have estimated that all the words ever spoken by mankind would be equal to five Exabytes.
Read more
Finance analytics, also known as financial analytics, provides differing perspectives on the financial data of a given business, giving insights that can facilitate strategic decisions and actions that improve the overall performance of the business.
Read more
Hadoop is a distributed data management platform or open-source software framework for storing and processing big data. It is sometimes described as a cut-down distributed operating system.
Read more
Hybrid cloud is the combination of on-premises and cloud deployment. Whether an organization’s resources include on-premises, private, public, or multi-cloud, a hybrid cloud ecosystem can deliver the best of all worlds: on-prem when needed and cloud when needed.
Read more
The Internet of Things, also known as IoT, is a concept that describes the connection of everyday physical objects and products to the Internet so that they are recognizable by (through unique identifiers) and can relate to other devices.
Read more
Machine learning is a type of artificial intelligence (AI) that provides computers with the ability to learn without being explicitly programmed. It focuses on the development of computer programs that can teach themselves to grow and change when exposed to new data.
Read more
Master Data Management (MDM) provides a unified view of data across multiple systems to meet the analytic needs of a global business. MDM creates singular views of master and reference data, whether it describes customers, products, suppliers, locations, or any other important attribute.
Read more
Metadata is data that describes other data in a structured, consistent way, so that large amounts of data can be collected, stored, and analyzed over time.
Read more
A mixed workload is an ability to support multiple applications with different SLAs in a single environment.
Read more
MongoDB is a cross-platform, open-source database that uses a document-oriented data model, rather than a traditional table-based relational database structure. This type of model makes the integration of structured and unstructured data easier and faster.
Read more
A branch of artificial intelligence, natural language processing (NLP) deals with making human language (in both written and spoken forms) comprehensible to computers.
Read more
Operational Resilience refers to an organization’s ability to continue functioning in the face of disruption. Considerations for Operational Resilience are multi-faceted. They include, but are not limited to, processes, capabilities, behaviors and systems. For example, Operational Resilience is the capability of an organization to still provide its products in the face of unforeseen supply chain disruptions. Equally, the ability for a company to move their data on-premises while their primary cloud service provider suffered a major outage would factor into that company’s level of Operational Resilience.
Read more
Pattern recognition occurs when an algorithm locates recurrences or regularities within large data sets or across disparate data sets. It is closely linked and even considered synonymous with machine learning and data mining.
Read more
An extremely large unit of digital data, one Petabyte is equal to 1,000 Terabytes. Some estimates hold that a Petabyte is the equivalent of 20 million tall filing cabinets or 500 billion pages of standard printed text.
Read more
Predictive analytics refers to the analysis of big data to make predictions and determine the likelihood of future outcomes, trends or events.
Read more
A type or extension of predictive analytics, prescriptive analytics is used to recommend or prescribe specific actions when certain information states are reached or conditions are met.
Read more
Python is an interpreted, object-oriented, high-level programming language with dynamic semantics. Python has a reputation as a beginner-friendly language, replacing Java as the most widely used introductory language because it handles much of the complexity for the user, allowing beginners to focus on fully grasping programming concepts rather than minute details.
Read more
R is an open-source programming language for statistical analysis. It includes a command line interface and several graphical interfaces. Popular algorithm types include linear and nonlinear modeling, time-series analysis, classification and clustering.
Read more
Retail analytics is the analysis of data generated by retail operations with a goal of making business decisions that drive profitability. The use of retail analytics developed as a response to the retail transformation being driven by unprecedented changes in consumer behavior, intensified pressure on margins, the changing role of stores, and intensified competition for both on and offline channels.
Read more
Risk management, sometimes referred to as risk mitigation, is the process of calculating the maximum acceptable level of overall risk to and from an activity, then using risk assessment techniques to pinpoint the initial level of risk and, if found to be excessive, developing a strategy to mitigate specific individual risks until the collective risk level is pared down to an acceptable level.
Read more
RTIM, also known as Real Time Interaction Manager or Management, uses real-time customer interactions, predictive modeling, and machine learning to deliver consistent, personalized customer experiences across channels.
Read more
Semi-structured data does not follow the format of a tabular data model or relational databases because it does not have a fixed schema. However, the data is not completely raw or unstructured, and does contain some structural elements such as tags and organizational metadata that make it easier to analyze.
Read more
Sentiment analysis is the capture and tracking of opinions, emotions or feelings expressed by consumers engaged in various types of interactions, such as social media posts, customer service calls and surveys.
Read more
A stressed exit is withdrawing from an outsourcing arrangement following the failure or insolvency of the service provider. A non-stressed exit is moving away from an agreement in a more planned and managed way due to strategic, commercial or performance reasons.
Read more
Structured data refers to data sets with strong and consistent organization. Structured data is managed by structured query language (SQL), by which users can easily search and manipulate the data.
Read more
A relatively large unit of digital data, one Terabyte (TB) equals 1,000 Gigabytes. It has been estimated that 10 Terabytes could hold the entire printed collection of the U.S. Library of Congress, while a single TB could hold 1,000 copies of the Encyclopedia Brittanica.
Read more
Unstructured data refers to unfiltered information with no fixed organizing principle. It is often called raw data. Common examples are web logs, XML, JSON, text documents, images, video, and audio files.
Read more
Big data is a collection of data from many different sources and is often describe by five characteristics: volume, value, variety, velocity, and veracity.
Read more
VPC stand for virtual private cloud. VPC is a personal and private virtual network space hosted within a public cloud environment. Each VPC is secure and logically isolated from other virtual networks in the same public cloud.
Read more