News

With Apache Spark Declarative Pipelines, engineers describe what their pipeline should do using SQL or Python, and Apache Spark handles the execution.
The no-code ETL tool works by combining a generative AI assistant for pipeline creation and Unity Catalog for governance.
AI and multimodal data are reshaping analytics. Success requires architectural flexibility: matching tools to tasks in a ...
Databricks, the Data and AI company, today extends its leadership in the unified governance category with powerful ...
Lakeflow Enters GA: Today, Lakeflow became generally available, providing a unified data engineering solution from ingestion to transformation and orchestration. Notably, the new Lakeflow Declarative ...