Despite efforts to enforce business rules on data within the existing architecture, the team faced challenges. Frequent rule changes and the constant addition of new business attributes resulted in flawed data passing through, causing significant issues for end-users.
Additionally, the process of incorporating new data was time-consuming, taking approximately 25 hours, while updating existing data required 3 hours. This delay was largely due to a monolithic pipeline generating large files encompassing entire datasets, sometimes reaching gigabytes in size.
Debugging and resolving data movement issues were time-consuming and laborious, leading to customer dissatisfaction. The sequential processing of data also strained server resources, with memory and CPU usage peaking at 90%, further delaying data publication.

Integrated an additional data validation layer just before the enterprise data exchange to effectively filter out flawed data and promptly notify the support team for swift resolution.
Fragmented the existing monolithic file into multiple smaller files, enabling parallel processing and significantly decreasing outbound file processing time from approximately 10 hours to just around 1 hour.
Transitioned from generating physical JSON files to directly querying data from the MDM system and transmitting it into Kafka every hour, eliminating the time required for file creation, publication, and retrieval.
Deployed a new correlation ID framework to meticulously track the flow of data from inbound to outbound, encompassing third-party systems, thereby furnishing comprehensive visibility and context for each data batch.
Enhanced efficiency by refining query logic to execute only a single query encompassing all IDs, resulting in a significant reduction in CPU usage to a mere 20%.
Transitioned from a 24-hour schedule to near real-time availability of master data for end consumers, ensuring they have access to up-to-date information whenever needed. Successfully reduced data errors, guaranteeing the prompt availability of accurate data to end consumers, thereby enhancing decision-making processes.
Implemented architectural and design changes that effectively reduced TechDepth in the MDM system, resulting in improved data quality and optimized infrastructure utilization. These enhancements have led to improved efficiency and reliability in data processing and transmission.
Implemented measures to enhance the quality of data transmitted to downstream systems, thereby facilitating accurate business decisions, and streamlining workflow processes. These measures have significantly contributed to the overall improvement of data integrity and reliability across the organization.
Continued efforts are aimed at establishing a scalable, efficient, and reliable enterprise data-as-a-service system, with plans for further enhancements and adaptations to meet evolving business needs. These ongoing initiatives demonstrate our commitment to long-term success and innovation in data management and utilization.
Altimetrik is committed to protecting your personal information. To apply for a position, you will need to provide your email address and create a login. Your information will be used in accordance with applicable data privacy laws, our Privacy Policy, and our Privacy Notice.