Sitemap
Data Science Collective

Advice, insights, and ideas from the Medium data science community

Member-only story

A Practical Guide on Leveraging Ollama in AI Pipelines

--

Photo by on

Since I started working in data engineering 10 years ago, implementing data transformations has always looked to be the most emphasized and significant stage in the development cycle. However, well-formatted code for data transformations is ineffective without proper validation of its outputs to ensure reliability in data pipelines.

When it comes to ensuring reliability in data pipelines, two key topics typically emerge: unit tests and data quality checks. I’m not suggesting that these are the only considerations, but I believe that if you have a solid set of unit tests and data quality checks, your data pipelines will be significantly more reliable than those lacking them.

At the same time, I must admit that working on unit tests and data quality checks is a time-consuming task. Sometimes, we do not have the capacity needed to ensure them due to various reasons, such as budget limitations, high demand for quick market entry, and others.

Thus, I want to present how you can improve the reliability of your data pipeline with AI, and even better, at a low cost. As you know, AI is becoming more integrated into code development, automating various tasks and accelerating the software development life cycle.

Data Science Collective
Data Science Collective

Published in Data Science Collective

Advice, insights, and ideas from the Medium data science community

Danilo Pinto
Danilo Pinto

Written by Danilo Pinto

I'm a guy working in the data area over the last 10 years, and also known as a "Data Engineering Team Manager" on LinkedIn xD

Responses (5)