Data Integration and Operations

Data is not always centralized or stored in the form that is required to make quality business decisions

data sources connection

Multi Data Source, Load, Connect and Transform

Moving data between different data sources and automating the data flow between systems is always a challenge. FlexRule’s Information Requirement Diagram allows you to visually model the flow while the logic is decoupled from the data sources.

data quality

Building Data Quality and Validation Pipeline

Ensuring data quality requires validating business rules. It also requires different perspectives (e.g. quality for safety, compliance, operation etc.). Using FlexRule, different Data Quality and Validation abstractions can be defined and executed against the same source of data.

data exchange

In-Memory Operation and Manipulation

Data operations can be modelled visually or by data operation expressions (monadic operators). Connect multiple data sources together to enrich and manipulate the data. Then compile the information by applying analytics and AI to build the results all in-memory so privacy is not compromised.

ETL

Data Virtualisation – Single View of Things

Build data virtualization based on multiple data sources. Transform the results based on different abstractions by applying business rules. Put together the end results and expose them as a new list or view for your consumers, customers, users, devices and more.

Database and API Connectors: A Large, Growing Library of Connectors

Connect to Anything, Anywhere!

Database: Wide range of databases such as MS SQL Server, Oracle, PostgreSQL, MySQL, MS Access as well as NoSQL databases or event even custom database.

API: Generic REST API connector to connect to any REST API endpoint. Communicate with any VERB and even attach contents and files if needed.

Applications: Pre-built integration to services and apps such as Gmail, Twitter, LinkedIn, Google Calendar, Salesforce, Dynamic 365, and so on…

Files: Any location such as local computer, FTP, web servers, etc., and any format such as CSV, Excel, XML, JSON, PDF, and more.

API integrations

Standardize Data Exchange: Contextually Validate Data using Rules and Constraints

Dynamic Data Validation is Crucial

With data located all over the place, and with many different processes within the organization needing some part of it, Dynamic Data Validation is a vital way of establishing Data Standard Exchange based on scenarios.

Fact Concept is an advanced approach which allows you to define a data structure based on individual scenarios and business processes needing the data. It allows you to define different view points and abstractions of data structure, constraints and rules for different cases rather than a fixed structure and format of data for all. One size, does not fit all!

Data Operations: Transform, Enrich, and Validate

Deal with Complex Data Scenarios

Once you connect to the data sources using our rich API and Applications connectors, you need to query the data. Once the data is available for your model, you need to do data operations such as: Filter, Join, Lookup, Group, and so on.

FlexRule provides a strong Data operation that allows you to build a model either visually (i.e. Information Requirement Diagram), or Syntax language (Monadic operator) to deal with load data. Both options for processing and dealing with data are in-memory and very quick.

data virtualization

Data Virtualization: Build a Single View of Your Data

Pull Your Data Together

Data Virtualization allows you to pull data from disparate data sources, then run it through the model that creates outputs based on the input of data sources, and finally Expose it to the data consumers.

Our Data Virtualization capability allows the linking of data from multiple data sources, the matching and finding of relationships between them using an exact match or a fuzzy logic match with a confidence score, and exposes the result as a REST API end point to the data consumers.

Data Integration: Connectors, ETL, Virtualization, Dynamic Validation and Analytics.