Advertisement

Data Integration Tools

Data integration tools are used to collect data from external (and internal) sources, and to reformat, cleanse, and organize the collected data. The ultimate goal of data integration tools is to combine data from a variety of different sources, and provide their users with a single, standardized flow of data. Use of these tools helps […]

10 Advantages of Real-Time Data Streaming in Commerce

While early science fiction shows like “Buck Rogers” (1939) and “The Fly” (1950) depicted teleportation technology, it was Star Trek’s transporter room that made real-time living matter transfer a classical sci-fi trope. While we haven’t built technology that enables real-time matter transfer yet, modern science is pursuing concepts like superposition and quantum teleportation to facilitate information transfer across any distance […]

Informatica Launches New Databricks-Validated Unity Catalog Integrations

According to a new press release, Informatica, a leading enterprise cloud data management company, has strengthened its strategic partnerships by launching enhanced Databricks-validated Unity Catalog integrations. These integrations enable no-code data ingestion and transformation pipelines to run natively on Databricks, providing a best-in-class solution for onboarding data from over 300 sources. The joint offering facilitates […]

Testing and Monitoring Data Pipelines: Part Two

In part one of this article, we discussed how data testing can specifically test a data object (e.g., table, column, metadata) at one particular point in the data pipeline. While this technique is practical for in-database verifications – as tests are embedded directly in their data modeling efforts – it is tedious and time-consuming when end-to-end data […]

Testing and Monitoring Data Pipelines: Part One

Suppose you’re in charge of maintaining a large set of data pipelines from cloud storage or streaming data into a data warehouse. How can you ensure that your data meets expectations after every transformation? That’s where data quality testing comes in. Data testing uses a set of rules to check if the data conforms to […]

Data Observability vs. Monitoring vs. Testing

Companies are spending a lot of money on data and analytics capabilities, creating more and more data products for people inside and outside the company. These products rely on a tangle of data pipelines, each a choreography of software executions transporting data from one place to another. As these pipelines become more complex, it’s important […]