Online transaction processing (OLTP) enables rapid, accurate data processing for most of today’s business transactions, such as through ATMs, online banking, e-commerce, and other types of daily services. With OLTP, the common, defining characteristic of any transaction is its atomicity, or indivisibility. A transaction either succeeds as a whole, fails, or is canceled. It cannot […]
A Brief History of Data Modeling
Data Modeling is the “act” of creating a data model (physical, logical, conceptual, etc.) and includes defining and determining an organization’s data needs and goals. The act of Data Modeling defines not just data elements, but also the structures they form and the relationships between them. Developing a data model requires the data modelers to work […]
Common Types of Cloud Computing
Before the cloud era, businesses had to rely on in-house data centers and internal hardware and software infrastructures to conduct online business. Organizations had to make substantial investments to set up their websites and networks. Additionally, each business had to hire the right people to manage and monitor their infrastructures. This approach not only added to the […]
How Kubernetes Can Help You Weather Regional Disaster Recovery
If you’re lucky, you’ve only had to worry about managing a data disaster recovery effort once or twice in your career, if at all. However, as the rate and number of natural disasters have increased, the chances of needing to navigate through a worst-case scenario have risen. As of April 11, 2023, the U.S. had […]
Controlling SAP HANA Data Sprawl
Enterprises running large SAP HANA instances in the cloud are seeing a new challenge appear as their databases continue to grow. Since SAP HANA has a simplified data layout and structure compared to a more complex legacy database, it was assumed this would result in less data sprawl and duplication. But does the data stay […]
Modeling Modern Knowledge Graphs
In the buzzing world of data architectures, one term seems to unite some previously contending buzzy paradigms. That term is “knowledge graphs.” In this post, we will dive into the scope of knowledge graphs, which is maturing as we speak. First, let us look back. “Knowledge graph” is not a new term; see for yourself […]
Spark vs. Flink: Key Differences and How to Choose
Apache Spark is an open-source, distributed computing system that provides a fast and scalable framework for big data processing and analytics. The Spark architecture is designed to handle data processing tasks across large clusters of computers, offering fault tolerance, parallel processing, and in-memory data storage capabilities. Spark supports various programming languages, such as Python (via […]
Semantic Technology and Integration 101: What It Is and Why It Matters
New technologies like ChatGPT are all the rage, as they aim to answer questions and provide information that makes our lives easier. Yet, the validity of the results generated has come under scrutiny and, as a result, much emphasis has been made on how organizations can get relevant and trustworthy data into the hands of […]
A Brief History of the Data Warehouse
A data warehouse stores data from in-house systems and various outside sources. Data warehouses are designed to support the decision-making process through data collection, consolidation, analytics, and research. They can be used in analyzing a specific subject area, such as “sales,” and are an important part of modern business intelligence. The architecture for data warehouses was […]
Leveraging Data Stream Processing to Improve Real-Time Data Analysis
Data stream processing is rapidly emerging as a critical technology for modernizing enterprise applications and improving real-time data analysis for data-driven applications. As businesses become more reliant on real-time data analysis, data stream processing enables them to analyze and process large amounts of data in real time, providing timely insights and enabling informed decision-making. Traditionally, […]