Data ingestion using Informatica Cloud Data Integration into a Databricks Delta Lake enables intelligent ingestion of high volumes of data from multiple sources into a data lake. Delta Lake automatically provides high reliability and performance.
Make it easier to create high-volume data pipelines for data at scale. Drag-and-drop pipelines created with Informatica Big Data Management can be pushed down to Databricks for processing in an optimized Apache Spark implementation.
With the right datasets, data scientists can build more accurate models and verify the lineage of data used to create AI models and fuel analytics. End-to-end lineage also addresses compliance with GDPR and other regulations.
“Seamless integration between Databricks and Informatica enables data engineers to easily discover the right datasets and ingest high volumes of data from multiple sources into Delta Lakes. This means joint customers can use the reliability and performance at scale from Databricks to make data ready for analytics and machine learning–and get intelligent governance to find, track and audit that data from end to end.”
— Ali Ghodsi, Co-founder & CEO, Databricks
Find out how you can accelerate your AI and analytics initiatives.