All the data you need.

Tag: Data Pipeline

Hewlett Packard Enterprise Acquires Pachyderm to  Expand AI-at-Scale Capabilities with Reproducible AI
Hewlett Packard Enterprise (NYSE: HPE) today announced an expansion to its AI-at-scale offerings with the acquisition of Pachyderm, a startup that delivers software, based on open-source technology, to automate reproducible machine learning pipelines that target large-scale AI applications.
How to Ensure an Effective Data Pipeline Process
In this contributed article, Rajkumar Sen, Founder and CTO at Arcion, discusses how the business data in a modern enterprise is spread across various platforms and formats. Data could belong to an operational database, cloud warehouses, data lakes and lakehouses, or even external public sources. Data pipelines connecting this variety …
eBook: Unlock Complex and Streaming Data with Declarative Data Pipelines
Our friend, Ori Rafael, CEO of Upsolver and advocate for engineers everywhere, released his new book "Unlock Complex and Streaming Data with Declarative Data Pipelines." Ori discusses why declarative pipelines are necessary for data-driven businesses and how they help with engineering productivity, and the ability for businesses to unlock more …
Hevo Data and Databricks Partner to Automate Data Integration for the Lakehouse
Businesses today have a wealth of information siloed across databases like MongoDB, PostgreSQL, and MySQL, and SaaS applications such as Salesforce, Zendesk, Intercom,… The post Hevo Data and Databricks Partner to Automate Data Integration for the Lakehouse appeared first on Databricks.
Databricks Announces General Availability of Delta Live Tables
Databricks, the Data and AI company and pioneer of the data lakehouse paradigm, announced the general availability of Delta Live Tables (DLT), the first ETL framework to use a simple declarative approach to build reliable data pipelines and to automatically manage data infrastructure at scale. Turning SQL queries into production …
5 Easy Steps to Make Your Data Business Ready
In this contributed article, Ayush Parashar Vice President of Engineering at Boomi, discusses five core components to a strong data strategy so businesses can derive insights from and act on the data. As the uses for data continue to grow, businesses must ensure their data is actually usable.
How Governing Observability Data is Critical to ESG Success
In this contributed article, Nick Heudecker, Senior Director of Market Strategy at Cribl, discusses how observability data comprises the logs, events, metrics, and traces that make things like security, performance management, and monitoring possible. While often overlooked, governing these data sources is critical in today’s enterprises. The current state of …
How Yum! Brands Uses Location Data from Foursquare to Make Smarter Decisions
Join this virtual event with a compelling panel of technology leaders to discuss to discover how Yum! Brands and other organizations are leveraging location-based data to boost in-app location accuracy, increase in-store foot traffic, and expand e-commerce business.
How Yum! Brands Uses Location Data from Foursquare to Make Smarter Decisions
Join this virtual event with a compelling panel of technology leaders to discuss to discover how Yum! Brands and other organizations are leveraging location-based data to boost in-app location accuracy, increase in-store foot traffic, and expand e-commerce business.
How to build a data extraction pipeline with Apache Airflow
Data extraction pipelines might be hard to build and manage, so it's a good idea to use a tool that can help you with these tasks. Apache Airflow (https://airflow.apache.org/) is a popular...
Learn the Comcast Architecture for Enterprise Metadata and Security
Comcast will present a live session on their architecture for metadata and security at our upcoming Databricks AWS Cloud Data Lake DevDay. The event includes a hands-on lab with Databricks notebooks that integrate with Amazon Web Services (AWS) Services like AWS Glue and Amazon Redshift. Our partner Privacera will also …
Global Study Sponsored by Qlik Finds Strong Relationship Between Optimizing Data Pipelines and Business Value
Qlik® announced a global study that shows organizations that strategically invest in creating data-to-insights capabilities through modern data and analytics pipelines are seeing significant bottom line impact. The global IDC survey, sponsored by Qlik, of 1,200 business leaders* shows that companies with a higher ability to identify, gather, transform, and …
How to make data lakes reliable
Data professionals across industries recognize they must effectively harness data for their businesses to innovate and gain competitive advantage. High quality, reliable data forms the backbone for all successful data endeavors, from reporting and analytics to machine learning. Delta Lake is an open-source storage layer that solves many concerns around …
Ascend Creates Automated and Intelligent Dataflows to Power Successful Digital Transformations
Ascend, provider of the Autonomous Dataflow Service, emerged from stealth with $19M in funding to de-risk big data projects and accelerate digital transformations. Ascend operates the only solution with which data engineering teams can quickly build, scale, and operate continuously optimized, Apache Spark-based pipelines. By combining declarative configurations and deep …
Loan Risk Analysis with XGBoost and Databricks Runtime for Machine Learning
For companies that make money off of interest on loans held by their customer, it’s always about increasing the bottom line. Being able to assess the risk of loan applications can save a lender the cost of holding too many risky assets. It is the data scientist’s job to run …