Technology Capabilities
Technology CapabilitiesEvery big or mid-sized company has a proliferation of sites, edge devices, apps, and di...
Over the past decade, cars have undergone a significant transformation to provide a mor...
GlobalLogic provides unique experience and expertise at the intersection of data, design, and engineering.
Get in touchA pipeline created for a particular use case may not be reusable for a different one and will require additional development effort to change. As a result, there is a need for frameworks that build new pipelines, adding additional data sources or data sinks with minimal time and development effort. Ideally, the framework should also be flexible in customizing and extending it to easily adapt to suit enterprise-specific requirements.
A number of low code and no-code solutions exist that allow for visually creating the data pipelines across a variety of sources and sinks. However, they do not provide the flexibility and modularity typically required to customize the pipelines for a given scenario.
Using a low code framework consisting of reusable, modular components that can be stitched together to compose the required pipelines is a better approach.
In this post, you’ll learn about the requirements for the low code framework and the approach to designing this framework.
Creating and maintaining pipelines to move data in and out of the platform is a major consideration. A data platform framework that allows its users to perform the different operations in a consistent way, irrespective of the underlying technology, will greatly reduce time and effort.
What do you look for in a low code framework? Here are some suggested requirements.
Modular: The framework should be modular in design. Each component of the framework can be used, managed, and enhanced independently.
Out-of-the-Box Functionality: Support integration with common data sources and sinks, and perform transformations out of the box. The components should be easy to implement for common use cases.
Flexible: The framework should be able to integrate with different services/systems across clouds or from on-premises.
Extensible: Allow extending existing components to customize as per specific requirements or add new custom components to implement new functionalities.
Code First: Provide a programmable way of defining and managing pipelines. API and/or SDK support should be available to programmatically create and access the pipelines.
Cross Cloud Support: Support for data sources, sinks, and services across different cloud services. You should be able to migrate pipelines using the framework for one cloud or on-premises to another cloud environment.
Reusable: Provides common reusable templates that allow for creating jobs in an easy way.
Scalable: Ability to scale workers dynamically or by configuration to handle high performance. The framework should automatically scale the underlying compute in response to changing workloads.
Managed Service: The framework should be deployable on a fully managed cloud service. Provisioning the infrastructure capacity, managing, configuring, and scaling the environment should be managed automatically. Minor version upgrades and patches are automatically updated and support is provided for major version updates.
GUI-based Definition: An intuitive GUI for creating and maintaining the data pipelines will be useful. The job runs and logs from execution should be accessible through a job monitoring and management portal.
Security: Out-of-the-box integration with an enterprise-level IAM tool for authentication and role-based access control.
The data platform framework provides the base foundation upon which you can build specific accelerators or tools for data integration and data quality/validation use cases.
While designing the framework, it is important to consider the following points:
Here are the building blocks for such a framework:
At GlobalLogic, we are working on a similar approach as part of the Data Platform Accelerator (DPA). Our DPA consists of a suite of micro-accelerators built on top of a platform framework based on cloud PaaS technologies.
We regularly work with our clients to help them with their data journeys. Share your needs with us using the contact form below and we are happy to discuss your next steps.