If you are new here, AirSage glossary will help to understand main terms and definitions of the Location Intelligence, Business Intelligence and Data Analytics fields.

What is 5G Infrastructure?

5G (acronym of 5th Generation) is the term used to describe the next-generation of mobile networks beyond LTE mobile networks. 5G network infrastructure is made up of macro- and small-cell base stations with edge computing capabilities. In a 5G network, network functions that typically run on hardware become virtualized, working as software.

There are two infrastructure options, standalone infrastructure (NSA) and non-standalone infrastructure (NSA). A non-standalone infrastructure is partly based on the existing 4G LTE infrastructure and brings some new technologies such as 5G New Radio (NR). According to Release 15 of the 3GPP standards body of October 2019, the NSA architecture has the 5G RAN and the 5G NR interface working together with the existing LTE infrastructure and the core network. The 5G standard goes on to say that this means that while only LTE services are supported, the network has the capabilities offered by 5G NR, such as lower latency.

The standalone infrastructure refers to a 5G network that is not based on LTE networks and has its own cloud-native network core that connects to the NR. According to 3GPP release 15, the standalone deployment option consists of user equipment, the RAN – which includes the NR – and the 5G core network. The core 5G network is based on a service-based 5G architecture framework with virtualized network functions.

What is an Analytical Database?

An analytical database is a specialized database management system that is optimized for business analytics applications and services. An analytical database has built-in features to store, manage, and analyze a large volume of data extremely quickly, provide faster query response times, and is more scalable than standard databases.

These features include columnar databases which organize data in columns to reduce the number of data points to be processed; data warehouse applications which include databasing tools in a single platform; hardware usage (in-memory databases) which use system memory to expedite processing; MPP databases which use multiple server clusters operating simultaneously; and online analytical processing databases which keep data cubes which can be analyzed based on multiple parameters.

What is Backtesting?

Backtesting is a term used in modeling, referring to testing a predictive model on historical data. It is a type of prediction and a special type of cross-validation applied to a previous timeframe. In business strategy, investment strategy, or (financial) risk model, backtesting seeks to estimate the performance of a strategy or model during a previous period. This requires a simulation of previous conditions with sufficient detail. This is the first limitation of backtesting: it requires detailed, reliable historical data. Secondly, there is a limit to modeling strategies; they are not to influence historical prices. Finally, backtesting, like other models, is limited by possible over-adjustment. Despite these limitations, backtesting provides information that is not available when models and strategies are tested with synthetic data.

The first step in backtesting is to select the threshold values within a period of time covered by the historical data. Then, for each threshold value, historical data are truncated at the threshold. The forecast model is trained and applied to truncated data. The forecasts thus obtained are compared with the complete original data. Finally, an average forecast error is established for all thresholds. This error can be read as an estimate of the error that will be associated with the model when making actual forecasts (for future data). Choosing the most appropriate threshold values requires a minimum of knowledge. As a rule, increasing the number of threshold values improves resistance to overfitting problems. In stock optimization, since there are hundreds of SKUs to analyze, a few threshold values are enough to decide with almost absolute certainty whether one forecast method is better than the others.

What is Big Data Architecture?

A Big Data architecture is the basis for Big Data Analysis, and it is developed to manage the input, processing, and analysis of data that are too big or complex to be handled by traditional DB systems. Typically, Big Data solutions encompass one or more of the following working load types: batch processing of inactive Big Data sources; real-time processing of Big Data on the go; the interactive exploration of Big Data and predictive Analysis and Machine Learning.

Most architectures for Big Data include some or all of the following components:

  • Data sources: The starting point of all Big Data solutions is one or more data sources. For example, application data archives such as relational databases; application data archives such as relational databases; static files generated by applications such as Web server log files; real-time data sources such as IoT devices.
  • Data storage: Batch processing data is usually placed in a batch file warehouse containing high volumes of large-scale files in multiple formats.
  • Batch processing: As datasets have a considerable size, Big Data solution often has to process data files through time-consuming batch processes to filter, aggregate, and otherwise prepare the data for analysis.
  • Real-time message input: If the solution includes real-time sources, the architecture must include capturing and archiving messages in real-time for flow (stream) processing.
  • Flow processing: After the acquisition of real-time messages, the solution must process them by filtering, aggregating, and otherwise preparing the data for analysis.
  • Analytical data archive: Many Big Data solutions prepare data for analysis and then serve the processed data in a structured format on which it is possible to perform queries with analytical tools.
  • Analysis and report creation: Big Data solutions can provide detailed information about the data through analysis tools and reports.
  • Orchestration: Most Big Data solutions consist of repeated operations (of data processing), encoded in workflows, which transform source data, move data between multiple sources and sinks, load the processed data into an analytical data storage or insert it directly into a dashboard.

What is Business Analytics?

Business analytics is a field that drives practical, data-driven changes in a business by using different processes and methodologies such as data mining, predictive analysis, and statistical analysis to analyze and transform raw data into useful insights, identify and anticipate trends and results, and measure past performance to guide an organization’s business strategy. Business Analytics can also be broken down into multiple components:

  • Data mining describes the process of extracting and analyzing a massive volume of data, using databases, statistics, and machine learning, to identify patterns and establish relationships to solve problems through data analysis.
  • Association & Sequence Identification is the detection of predictable actions performed in combination or sequence with other actions.
  • Text Mining, also known as text analysis, is a process of converting unstructured text data into meaningful and actionable information. Text mining uses different AI technologies to automatically process data and create valuable insights that enable businesses to make data-driven decisions.
  • Forecasting is a method used to make informed predictions by analyzing historical data as the main input to determine the course of (future) trends, events, and behaviors.
  • Predictive Analytics is the use of data, statistical algorithms, and machine learning techniques to determine patterns and make accurate predictions of future outcomes and trends based on historical data.
  • Optimization is a method that uses simulation techniques to identify scenarios that might produce the best results once patterns have been determined and future outcomes and trends have been predicted with the business analysis.
  • Data Visualization is the presentation of data using a graphical format that highlights patterns and trends in data and helps the reader achieve quick insights.

What is Business Forecasting?

Business forecasting is the method employed by companies to make predictions or projections of their future economic conditions, such as sales, potential revenues, spending, by using analytics, data, insights, and experience. Business Forecasting helps to automate and optimize business processes and develop better business strategies. Two approaches can be used to identify patterns and make accurate predictions to drive better decision-making: a qualitative approach and a quantitative approach.

What is Business Intelligence?

Business Intelligence describes a set of processes, technologies, and practices for gathering, storing, analyzing, and interpreting raw data from internal and external sources to convert it into meaningful information that organizations can use to make more tactical and strategic decisions. Business intelligence tools are a suite of software and services used to process (access and analyze) data sets and present the resulting analytical findings. This typically in the form of reports, comprehensive summaries, and visuals (dashboards, graphs, charts, and maps). The objective is to provide users with detailed, self-explanatory intelligence about the state of the business, like a cockpit in a plane.

What is Capacity Management?

Capacity Management is the process of IT monitoring, administration, and planning actions taken to ensure that the IT capacity can handle the data processing requirements and the continuous provision of a consistent and acceptable service level at a known and controlled cost. The capacity management process covers the operating and development environment, including hardware, network equipment, peripherals, software, and human resources. Capacity management assures that IT and resources are planned and programmed to provide a consistent service level appropriate to the company’s current and future needs.

The objectives of capacity management are to:

  • Identify the requirements of IT capacity to meet the current and also future workloads.
  • Generate and maintain an accurate capacity plan.
  • Provide guidance and advice on all issues related to capacity and performance.
  • Make sure that the service performance achievements meet their agreed-upon targets.
  • Measure and monitor the capacity to support service level management.
  • Help with incident management (diagnosis and resolution of incidents and problems).
  • Analyze the impact of changes that they may have had on the capacity plan and take proactive steps to improve the performance wherever it is cost-effective.

What is a Columnar Database?

A Columnar Database is a database management system that stores table data as sections of data columns, compared to most relational databases that store data in rows. This has advantages in data warehouses, customer relationship management (CRM) systems, bibliographic card catalogs, and other ad hoc systems, where aggregates are calculated on a large volume of similar data. A columnar database refers to both a column-oriented structure and a focus on optimization for column-oriented workloads.

This approach contrasts with row-oriented databases or databases stored per row and related databases, which use a value-based storage structure. Columnar databases are designed to return data for a limited number of columns efficiently. All values of a column are stored together. A columnar database excels at reading operations on a limited number of columns. The columnar database systems allow the optimization of the performance of analysis queries by drastically reducing the overall disk I/O requirements and the amount of data to be loaded from the disk.

What is the difference Between CPU and GPU?

CPU and GPU are processor units. A CPU (Central Processing Unit) is considered as the brain of the computer and consists of ALU (Arithmetic Logic Unit), which stores the information, performs calculations, and CU (Control Unit), which in turn is in charge of instructions and branching. GPUs (Graphical Processing Unit) are typically used to enhance images and process videos on computers due to the high performance on specific tasks. It mainly consists of ALUs. The main difference lies in the architecture and its purpose. A CPU is designed for a wide variety of workloads, and it focuses on a smaller number of cores on individual tasks and on getting things done quickly. A GPU is the powerful sibling created for jobs that require high performance (initially developed to accelerate specific 3D rendering tasks, but it has evolved to become more general-purpose parallel processors, handling a growing range of applications).

What is Data Anonymization?

Data Anonymization refers to the process used to protect private or sensitive information by erasing or encrypting identifiers that connect individuals to stored data. Data anonymization aims to protect the confidential activities of an individual or company while maintaining the integrity of the data collected and shared. Data anonymization is carried out by most sectors that deal with sensitive information such as healthcare, finance, and digital media while promoting data sharing integrity. Data anonymization minimizes the risk of unintentional disclosure when data is shared between countries, industries, and even departments in the same company. It also reduces the chances of identity fraud.

What is DBMS?

DBMS (database management system) refers to a software application used to access, create, and manage databases. Organizations use a large amount of data. DBMS makes it possible to organize data in a database, store and transform them into valuable information, and support making strategic decisions. The main functions provided by DBMS are:

  • Data Dictionary Management: DBMS stores definitions of the data elements and their relationships (metadata) in a data dictionary.
  • Data Store Management: DBMS system provides storage not only for the data but also for related data entry forms or screen definitions, report definitions, data validation rules, procedural code, structures to handle video and picture formats, etc.
  • Data Transformation and Presentation: DBMS transforms entered data into the required data structures.
  • Security Management: Security rules to determine the access rights of users and enforce security and data privacy.
  • Backup and Recovery Management: Processes to back-up the data regularly and recover data if a problem occurs to ensure safety and integrity.
  • Multiuser access control: DBMS ensures access to the database to multiple users concurrently, guaranteeing the database’s integrity.
  • Data Integrity Management: DBMS enforces database structure and integrity rules.

What is Data Cleansing?

Data Cleansing refers to the process of modifying data to ensure that it is free of irrelevances and incorrect information and to guarantee, with a certain level of reliability, the accuracy of a large volume of data (database, data warehouse, dataset, etc.). This term has been used in the past to define filtering on the basis of data mining. The process precedes the actual extraction (mining) of a potentially useful and previously unknown amount of information to produce knowledge. When acquiring data, the usage of the cleansing process guarantees a higher level of data quality. A data cleansing system must meet the qualitative criteria:

  • Accuracy;
  • Integrity;
  • Completeness;
  • Validity;
  • Consistency;
  • Uniformity;
  • Density, and
  • Uniqueness.

The following activities are typical for the data cleansing process:

  • fill the attributes that have missing values
  • identify outliers (data extremely different from expected values)
  • eliminate the noise in data
  • correct inconsistency.

What is Data Integration?

Data Integration is a process that uses both technical and business processes of combining data from different sources into a single, unified location, e.g., a data warehouse. Data Integration includes some common elements as a network of data sources, a master server, and a client accessing data from the master server. This process is often a prerequisite of other processes like analysis, reporting, and forecasting. Data Integration allows managing data more efficiently and, by centralizing all data, provides easier access to data for those who need it. Automated updates enable reports to synchronize and run efficiently in real-time whenever needed, reducing errors and rework. In an organized and centralized system, issues are automatically identified, and improvements are applied, resulting in increased quality of business data, providing more accurate data and analysis results.

What is Data Intelligence?

Data Intelligence is the practice of employing artificial intelligence and machine learning tools to analyze and convert massive datasets into valuable insights to allow businesses to make better strategic decisions for future developments. Data intelligence techniques include

  • data orchestration to cleanse, correlate, prepare, and integrate multifaceted data;
  • machine learning to unlock hidden insights and discoveries;
  • metadata management and data cataloging to understand the data and its potential value.

Data Intelligence five main components:

  • descriptive data, to review and examine data to understand and analyze business performance;
  • prescriptive data used to develop and analyze alternative knowledge that can be applied in the courses of action;
  • diagnostic data, for determining the possible causes of specific events;
  • predictive data, to analyze historical data and determine future events;
  • decisive data, to measure the data adequacy and propose future actions.

Organizations can leverage data intelligence to adapt more rapidly to industry trends. By monitoring the analytics that data intelligence provides, they gain insights about patterns, changes, and trends that allow them to develop ideas and directions based on that valuable information. Using big data and AI, data intelligence provides structure to the management and allocation of that data. Besides, it is the leading actor in data transformation, as it transforms a massive amount of data into experienced-based and constantly growing information.

What is Data Management?

Data Management describes the practice of collecting, keeping, and using data securely, efficiently, and cost-effectively. Data management is the development, execution, and supervision of projects, policies, programs, and practices that control, protect, transport, and increase the value of data and information resources. Businesses must handle large amounts of data from non-heterogeneous databases and sources. Data management provides access to heterogeneous information from a central source to manage effective business strategies based on real insight. Data management work has a wide scope with several main activities, such as:

  • Creating, accessing, and updating data in a different data layer;
  • Data storage on several levels;
  • Provide high availability to companies;
  • Using data in a growing variety of applications, analysis, and algorithms;
  • Ensure confidentiality and data security;
  • Data storage and destruction according to retention programs and compliance requirements.

What is Data Validation?

Data Validation ensures that data has been cleansed to guarantee its quality. It employs routines, often called “validation rules,” “validation constraints,” or “control routines,” which check the accuracy, meaningfulness, and reliability of the data being entered into the system. The rules can be implemented through the automated structures of a data dictionary or through the inclusion of an express validating logic for the computer’s application program and its application.

Data validation has been recognized as an essential part of any data management operation. Data must be verified and validated before using it to avoid any inaccurate results. It is a key part of the workflow as it allows to create optimal results.

Validation of the accuracy, transparency, and detail of the data is essential to minimize any project defects. If data validation is not performed, decisions based on data can show imperfections and inaccuracy, not representing the current situation. In addition to verifying data inputs and values, it is required to validate the data model. An unstructured or correctly constructed data model causes problems when using data files in various applications and software. Using validation rules to clean data before use helps mitigate “garbage in = garbage out” scenarios. Ensuring data integrity guarantee the legitimacy of the conclusions.

What is Extrapolation?

Extrapolation is a methodology that involves making statistical forecasts by using historical trends that are projected for a specified period of time into the future. This is a type of estimation of the value of a variable based on its relationship with another variable. In that sense, it resembles interpolation, which does produce estimates between known observations; however, extrapolation is subject to higher uncertainty and a higher risk of producing meaningless results.

Extrapolation can also mean the extension of a method, assuming that similar methods are applicable. It can apply to human experience to project, extend or expand the known experience into an unknown or previously experienced area to gain (usually conjecture) knowledge of the unknown. The extrapolation method can be applied to the problem of internal reconstruction.

What is Geocoding?

Geocoding refers to the process used to transform geographical-administrative location data into a set of geographical coordinates, points. The geocoding is based on the recognition of an address in a specific database and allows an initial identification of the asset on the territory. Geocoding includes two main components: the reference dataset (that is the underlying geographic database contain geographic features that a geocoder uses to generate a geographic output) and the geocoding algorithm. The process generally begins with the input data that are stored in the database. Subsequently, those data are classified into Relative input data or Absolute input data. However, only Absolute input data can be geocoded and transformed into a list of coordinates.

These coordinates are very powerful information for businesses and can be useful in several fields. For example, this information allows businesses to recognize geographical patterns, to develop targeted marketing strategies for specific customers using data management on their geographical location. It is also useful for analyzing address data, monitoring the population growth in a specific area, and better planning events and future projects.

What is Geodata?

Geodata is location information stored in a format that can be used with a geographic information system (GIS). There are different geodata types:

  • vector data, consisting of vertices and paths (three basic types: points, lines, and polygons);
  • raster data consists of pixels or grid cells. Commonly they are square and regularly spaced, but also can be rectangular;
  • geographic database whose purpose is to host vectors and raster;
  • web files (internet type of storage and access to geodata);
  • multitemporal data that links a temporal component to information but also includes a geographical component.

Technologies that can be used to gather geographical data are Global Positioning System (GPS) data, telematics devices, geospatial satellite images, Internet of Things (IoT), and geotagging.

What is Geospatial Analytics?

Geospatial Analysis is the process of using time and position information within traditional data analysis processes. It includes the collection, display, and manipulation of Geographic Information System (GIS) data as images, satellite photographs, historical information, etc.

Geospatial analytics uses geographical coordinates, i.e., longitude and latitude, postal codes, street addresses, and other identifiers to create geographical methods. These models include graphs, statistics, maps, charts, and data views that allow building more understandable complex relationships. Geospatial Analysis allows businesses to simultaneously analyze a large amount of data, knowing what is happening at different places and times and enabling more effective decisions and more accurate results. Maps enable recognizing the models previously indicated in the spreadsheets as the contiguity, proximity, affiliation, and distance. Businesses can gather information from different locations in real-time, using tools like the Internet of Things (IoT), mobile devices, social media, and position sensors. By including time and location analysis, trends can be understood in a geographical or linear context. This means that forecasts can be made at a given site and at a given time in the future.

What Is Geospatial Intelligence?

Geospatial Intelligence as a term has been defined in U.S. Code Title 10, §467: “The term geospatial intelligence means the exploitation and analysis of imagery and geospatial information to describe, assess, and visually depict physical features and geographically referenced activities on the earth. Geospatial intelligence consists of imagery, imagery intelligence, and geospatial information”.

In practice, geospatial intelligence refers to a discipline that includes the extraction and analysis of geospatial images and information to describe, evaluate, and visually represent physical characteristics and geographically related activities on earth.

Geospatial Intelligence combines different fields, such as mapping, cartography, imagery analysis, and imagery intelligence. In addition to its use in a military context, many organizations in sectors such as telecommunication, smart cities, retail, municipalities, transportation, public health and safety and real estate are using geospatial intelligence to improve or optimize everyday life quality.

The main principle of geospatial intelligence is to organize and combine all available data around its geographical location on earth and then leverage it to develop products that can be used by planners and decision-makers.

What is GIS?

GIS stands for “Geographical Information System”. It refers to a software system that allows to acquire, analyze, visualize, and share information derived from geographic data and give a representation of what occurs in it. The technology behind GIS integrates the features of a database that allow to perform searches, store data, draw graphs, with the features of a map that provides spatial data and geographical representations. Thus, GIS is a software that can handle a large amount of geo-referenced information. These data can be expressed through maps or tables and referred to extensive portions of territory, as needed. GIS has many differences from other IT systems, as it offers infinite possibilities of use for all needs related to geographical components. From geo-location of objects to the study of landscape evolution, GIS allows detailed and complex planning of the territory and the actions to be performed on it.

What is GPGPU?

GPGPU is an acronym for General-Purpose computing on Graphics Processing Units. In IT, it refers to the use of a graphics processing unit (GPU) for additional purposes than traditional use in computer graphics. The GPGPU is used for processing extremely demanding in terms of processing power, and for which traditional CPU architectures cannot provide sufficient processing capacity. Due to their nature, these processes are highly parallel and capable of widely benefitting from the typical architecture of the GPU. In addition, this architecture has evolved, offering extreme programmability and an increase in processing power and versatility.

What is GPU Database?

GPU database is a database, relational or non-relational, that uses a GPU (graphics processing unit) to execute specific database operations. Databases on GPUs are usually fast. As a result, GPU databases are more flexible in processing many different data types or massive amounts of data. Leveraging GPU’s processing power, GPU database allows to analyze massive amounts of information and quickly return results.

What is GPU Rendering?

GPU Rendering allows using the graphics card for rendering instead of CPU. This typically accelerates the rendering process as modern GPUs are leveraging higher processing power.

GPU and CPU process data in a similar way, but a GPU focuses on parallel processing. By contrast with CPU technology, GPUs are designed to process instructions simultaneously on multiple cores. GPU Rendering takes a single set of instructions and runs them on multiple cores (32 to hundreds) over multiple data. Compared to a CPU that can simultaneously work on about 24 blocks of data, GPU can handle about 3000 blocks of data.

What is GPU-accelerated Analytics?

GPU-accelerated Analytics involves a set of applications that exploit the massive parallelism of a graphics processing unit (GPU) to accelerate compute-intensive operations for data science, deep learning, machine learning, and other large-scale applications.

What is Information Visualization?

Information Visualization is the study of visual (interactive) representations of abstract data. Abstract data includes both numerical and non-numerical data, as text and geographic information. It is used as a critical component in scientific research, digital libraries, data mining, financial data analysis, market research, etc.

Information visualization assumes that visual representation and interaction techniques take advantage of the path’s width from the human eye to the mind to allow users to see, explore, and understand large amounts of information simultaneously. Information visualization focuses on the study of approaches to communicate abstract information in intuitive ways.

Dashboards and scatter diagrams are common examples of information visualization. By representing an overview and the visualization of relevant connections, the visualization of information allows users to extract insights from abstract data efficiently and effectively.

Information visualization plays an important role in making data accessible and transforming raw data into usable information. It is drawn from the fields of human-machine interaction, visual design, computer science, and cognitive science. Examples include world map-style representations, line graphs, and designs of virtual buildings in 3D or urban plans.

What is Interpolation?

Interpolation is a statistical method uses to estimate values of an unknown function f(x) for specific subjects x in a specific range [a, b] when a number of observed values f(x) are available within that range. This is a type of estimation, a method of constructing new data points within the range of a discrete set of known data points. Interpolation helps to determine the data points among those indicated. Interpolation is required to calculate the value of a function for an intermediate value of the independent function. This is a process of determining unknown values between known data points. It is mainly used to predict unknown values for any related geographical data points, such as noise level, precipitation, altitude, etc.

What is LAN?

LAN stands for Local Area Network. It refers to networks with limited spatial extension. LANs are usually used in private or business premises to configure home or business networks. It supports communication between different devices and the exchange of data.

LAN consists of at least two terminals, but it can also connect several thousand devices. LAN can connect computers, smartphones, printers, scanners, storage devices, servers, and other network devices to each other and connect them to the Internet. However, if wider spatial distances must be covered, MAN and WAN networks are more suitable.

By now, most LANs are made via Ethernet cables. An Ethernet LAN can be divided into several virtual (VLAN) or physical LANs. Switches and routers are used to structure Local Area Networks. As an interface, the hardware controls the connections between individual network users and ensures that data patches reach their destination.

What is Location Intelligence?

Location Intelligence allows to collect, analyze, and organize spatial data using various Geographical Information Systems (GIS) tools. This process allows transforming large amounts of data into color-coded visual representations that enable an easier understanding of trends and generate meaningful insights.

Location intelligence, or spatial intelligence, can also be defined as an extension of traditional Business Intelligence (BI). It refers to a process of deriving meaningful insights from geospatial data, organizing, and understanding the technologies, applications, and practices that allow to relate spatial data processed by GIS with business data processed by BI applications. The insights acquired can be harnessed from organizations to better understand spatial patterns, consumers’ behaviors, interests, and preferences, and make more effective decisions.

What is Real-Time Analytics?

Real-time Analytics refers to the analysis of big data, including technologies and processes used to measure, manage and analyze data in real-time as soon as it enters the system, thus allowing organizations to visualize and understand the data immediately.

Real-time analytics applied in business contexts allows organizations to immediately obtain insights, operate on them, understand customer needs, and prevent potential issues before they occur. Businesses can leverage the power of real-time analysis and big data to optimize internal operations, as well as improve workflows, support sales, and apply more effective marketing strategies. These tools also provide insight into customer behavior and market trends in real-time, enabling immediate response, and staying ahead of the competition.

What is SQL Engine?

SQL, the acronym of Structure Query Language, can be defined as a server that allows managing databases, based on a relational model, that contain specific information. SQL engine is a program that identifies and enables SQL commands to enter a relational DB and process data.

What is Spatiotemporal Data Analysis?

Spatiotemporal Data Analysis is a new research area that, as a result of the development and application of new computational techniques. It allows the analysis of massive space-time databases. Data collected that includes at least one spatial and one temporal property lead to spatiotemporal models. An event in a space-time data set represents a spatial and temporal circumstance that exists at a specific time t and a specific position x.

What is Visual Analytics?

Visual Analytics describes the science of analytical reasoning supported by interactive visual interfaces. Despite the existence of different methods of automated data analysis, visual analytics can provide a more effective response in a context where data is produced at an increasing rate and the ability to collect and store it is higher than the ability to analyze it. It is also essential to include it to approach the complex nature of the many problems at an early stage in the data analysis process.

Visual analysis methods allow decision-makers to combine their human flexibility, creativity, and basic knowledge with the massive storage and processing capabilities of today’s computers to get an overview of complex problems. Using advanced visual interfaces, users can interact directly with data analysis capabilities, enabling them to make well-informed decisions in complex situations. Visual Analytics combines business intelligence and analysis tools in a single system representing reality in a “data-driven” format. Thus, millions of data are analyzed in a few seconds and displayed in a graphical interface, logically and not in a pre-constituted format.