Logo Taplio

Taplio

Paul Iusztin's Linkedin Analytics

Get the Linkedin stats of Paul Iusztin and many LinkedIn Influencers by Taplio.

Want detailed analytics of your Linkedin Account? Try Taplio for free.

Paul Iusztin

open on linkedin

I am a senior machine learning engineer and contractor with ๐Ÿฒ+ ๐˜†๐—ฒ๐—ฎ๐—ฟ๐˜€ ๐—ผ๐—ณ ๐—ฒ๐˜…๐—ฝ๐—ฒ๐—ฟ๐—ถ๐—ฒ๐—ป๐—ฐ๐—ฒ. I design and implement modular, scalable, and production-ready ML systems for startups worldwide. My central mission is to build data-intensive AI/ML products that serve the world. Since training my first neural network in 2017, I have 2 passions that fuel my mission: โ†’ Designing and implementing production AI/ML systems using MLOps best practices. โ†’ Teaching people about the process. . I currently develop production-ready Deep Learning products at Metaphysic, a leading GenAI platform. In the past, I built Computer Vision and MLOps solutions for CoreAI, Everseen, and Continental. Also, I am the Founder of Decoding ML, a channel for battle-tested content on learning how to design, code, and deploy production-grade ML and MLOps systems. I am writing articles and posts each week on: - ๐˜“๐˜ช๐˜ฏ๐˜ฌ๐˜ฆ๐˜ฅ๐˜๐˜ฏ: 29k+ followers - ๐˜”๐˜ฆ๐˜ฅ๐˜ช๐˜ถ๐˜ฎ: 2.5k+ followers ~ ๐Ÿ”— https://medium.com/@pauliusztin - ๐˜š๐˜ถ๐˜ฃ๐˜ด๐˜ต๐˜ข๐˜ค๐˜ฌ (๐˜ฏ๐˜ฆ๐˜ธ๐˜ด๐˜ญ๐˜ฆ๐˜ต๐˜ต๐˜ฆ๐˜ณ): 6k+ followers ~ ๐Ÿ”— https://decodingml.substack.com/ . If you want to learn how to build an end-to-end production-ready LLM & RAG system using MLOps best practices, you can take Decoding MLโ€™s self-guided free course: โ†’ ๐˜“๐˜“๐˜” ๐˜›๐˜ธ๐˜ช๐˜ฏ ๐˜Š๐˜ฐ๐˜ถ๐˜ณ๐˜ด๐˜ฆ: ๐˜‰๐˜ถ๐˜ช๐˜ญ๐˜ฅ๐˜ช๐˜ฏ๐˜จ ๐˜ ๐˜ฐ๐˜ถ๐˜ณ ๐˜—๐˜ณ๐˜ฐ๐˜ฅ๐˜ถ๐˜ค๐˜ต๐˜ช๐˜ฐ๐˜ฏ-๐˜™๐˜ฆ๐˜ข๐˜ฅ๐˜บ ๐˜ˆ๐˜ ๐˜™๐˜ฆ๐˜ฑ๐˜ญ๐˜ช๐˜ค๐˜ข ~ ๐Ÿ”— https://github.com/decodingml/llm-twin-course . ๐Ÿ’ฌ If you need machine learning solutions for your business, letโ€™s discuss! ๐ŸŒŽ Only open to full remote positions as a contractor. . Contact: ๐Ÿ“ฑ Phone: +40 732 509 516 โœ‰๏ธ Email: p.b.iusztin@gmail.com ๐Ÿ’ป Decoding ML: https://linktr.ee/decodingml ๐Ÿ•ต๐Ÿปโ€โ™‚๏ธ Personal site & Socials: https://www.pauliusztin.me/

Check out Paul Iusztin's verified LinkedIn stats (last 30 days)

Followers
54,582
Posts
20
Engagements
10,036
Likes
8,662

What is Paul talking about?

nicheai
  • frequency
  • engagement

Who is engaging with Paul

Banias Baabe profile picture
Samanth Koduru profile picture
Jason Quick profile picture
Piotr Malicki profile picture
Alexandru Razvant profile picture
Saba Hesaraki profile picture
Alex Vesa profile picture
Tales Marra profile picture
Ashok Kumar Murugesan profile picture
Daniel Svonava profile picture
Silvestre Pรฉrez O. profile picture
Priyanka Kamath profile picture
Hendy Fergus Atheri Hura profile picture
ORLANDO MONTENEGRO REYES profile picture
Sujoy Chabri profile picture
Maria Vechtomova profile picture
Achutha Subhash profile picture
Jai Singh profile picture
Ankit Srivastava profile picture
Joshua Rio-Ross profile picture
Ajeet Singh  profile picture
Prashant K Dhingra profile picture
Mattheus Chediak profile picture
Ray Florez profile picture
Anirban Datta profile picture
Rutuja Surve profile picture
Kaan Kabalak profile picture
Abby Morgan profile picture
Raphaรซl Hoogvliets profile picture
zahra sadeghi profile picture
Li Yin profile picture
Ali Alemi Matin Pour profile picture
Bharat Raghunathan profile picture
bรนi xuรขn hรนng profile picture
Victor Robles profile picture
Shamal De Silva profile picture
Leon Jose profile picture
Koyelia Ghosh Roy profile picture
Aniruddh Durga Naga Gilakamsetty profile picture
Pascal Biese profile picture
Diego Marinho profile picture
Diego Miranda profile picture
Thiago Costa profile picture
Bruna Lemberck profile picture

Paul Iusztin's Best Posts (last 30 days)

Use Taplio to search all-time best posts


Iโ€™ve been working with GenAI for 3+ years. Hereโ€™s something all engineers must come to terms with: If youโ€™re building LLM-powered applications, at some point, youโ€™ll need to generate high-quality datasets to fine-tune SLMs. Why? โ†’ Fine-tuning SLMs reduces costs, latency, and throughput while maintaining high accuracy for specific tasks. โ†’ Some domains require specialized fine-tuning for better domain adaptation. โ†’ Fine-tuned models give you more control over AI behavior and response generation. Thatโ€™s exactly what weโ€™re tackling with our ๐—ฆ๐—ฒ๐—ฐ๐—ผ๐—ป๐—ฑ ๐—•๐—ฟ๐—ฎ๐—ถ๐—ป ๐—”๐—œ ๐—”๐˜€๐˜€๐—ถ๐˜€๐˜๐—ฎ๐—ป๐˜. ... and today, Iโ€™m breaking down the dataset generation feature pipeline we built for fine-tuning our summarization SLM. The input to our generation pipeline will be raw documents from MongoDB (Notion & crawled resources). And the output is a high-quality summarization dataset published to Hugging Faceโ€™s dataset registry. Since this pipeline generates features used to train an LLM, itโ€™s called a feature pipeline. Hereโ€™s how it works, step by step: ๐Ÿญ. ๐——๐—ฎ๐˜๐—ฎ ๐—˜๐˜…๐˜๐—ฟ๐—ฎ๐—ฐ๐˜๐—ถ๐—ผ๐—ป โ†’ Pulls raw documents from MongoDB and standardizes formatting. ๐Ÿฎ. ๐——๐—ผ๐—ฐ๐˜‚๐—บ๐—ฒ๐—ป๐˜ ๐—˜๐˜…๐—ฝ๐—น๐—ผ๐—ฟ๐—ฎ๐˜๐—ถ๐—ผ๐—ป โ†’ Analyzes length & quality scores distributions to make informed decisions. ๐Ÿฏ. ๐——๐—ฎ๐˜๐—ฎ ๐—™๐—ถ๐—น๐˜๐—ฒ๐—ฟ๐—ถ๐—ป๐—ด โ†’ Removes low-value content, keeping only high-quality documents. ๐Ÿฐ. ๐—ฆ๐˜‚๐—บ๐—บ๐—ฎ๐—ฟ๐—ถ๐˜‡๐—ฎ๐˜๐—ถ๐—ผ๐—ป โ†’ We use a more powerful LLM (e.g., `gpt-4o`) to generate multiple summaries per document by varying temperature and sampling parameters (a process known as distillation) ๐Ÿฑ. ๐—ค๐˜‚๐—ฎ๐—น๐—ถ๐˜๐˜† ๐—–๐—ผ๐—ป๐˜๐—ฟ๐—ผ๐—น โ†’ Filters out poor-quality summaries. ๐Ÿฒ. ๐——๐—ฎ๐˜๐—ฎ๐˜€๐—ฒ๐˜ ๐—ฆ๐—ฝ๐—น๐—ถ๐˜๐˜๐—ถ๐—ป๐—ด โ†’ Divides data into training, evaluation, and test sets (done before storing the dataset and not at training time!) ๐Ÿณ. ๐—ฉ๐—ฒ๐—ฟ๐˜€๐—ถ๐—ผ๐—ป๐—ถ๐—ป๐—ด & ๐——๐—ฒ๐—ฝ๐—น๐—ผ๐˜†๐—บ๐—ฒ๐—ป๐˜ โ†’ Publishes the final dataset to Hugging Face. To keep the pipeline reproducible, trackable, and scalable, we manage it using ZenML, which: โ†’ Orchestrates the entire workflow from extraction to deployment. โ†’ Ensures traceability & versioning of pipeline runs & datasets. โ†’ Allows dynamic configuration for different filtering, summarization & structuring techniques. Even if youโ€™re not deep into fine-tuning, at some point, youโ€™ll need a structured way to generate datasets for specialized AI applications. This is one of the most critical components of your pipeline. Want to learn more? Check out the link in the comments.


283

The unthinkable happened... (And it wouldn't have been possible without you) The LLM Engineers Handbook GitHub repo has just crossed 2,800 stars. But that's not all... It's now less than 100 stars away from being the most popular repository in the Packt GitHub organization. Honestly, I'm flattered. This milestone wouldn't have been possible without your support. Iโ€™m truly grateful for every one of you who has contributed, starred, or even shared the repo. Weโ€™ve been working hard to make this repository a powerful resource for anyone looking to master LLMs. Along with the book, you'll learn how to implement production-ready, industry-level LLM & RAG applications. (Of course, following best practices in software engineering, system design and MLOps) And we've made some changes: - Weโ€™ve accepted PRs from contributors to fix bugs - Docs have been updated based on reader feedback If you haven't already, check it out. Contribute to the repo and help us hit that No. 1 spot. LFG๐Ÿ”ฅ (Link in the comments)


244

Over the past 6 months, Iโ€™ve seen the word โ€œagentโ€ thrown around a lot. But letโ€™s be honest, most of the time it's misleading... In reality, what people often refer to as an "agent" is just Python scripts wrapped around LLM calls. (Iโ€™ve fallen into that trap myself early on.) But agents (and agentic systems) are fundamentally different from classic LLM workflows. Hereโ€™s how I break it down: ๐—ช๐—ผ๐—ฟ๐—ธ๐—ณ๐—น๐—ผ๐˜„๐˜€ โ†’ Follow a fixed, predefined sequence โ†’ Every step is planned in advance โ†’ Think โ€œassembly lineโ€ logicโ€”reliable, repeatable, predictable โ†’ Great for structured, stable tasks ๐—”๐—ด๐—ฒ๐—ป๐˜๐˜€ โ†’ Decide what to do next on the fly โ†’ Use reasoning to pave their path dynamically โ†’ Think of them like an intelligent assistant, dynamically adapting its plan โ†’ Ideal for unstructured, evolving tasks ๐—ช๐—ผ๐—ฟ๐—ธ๐—ณ๐—น๐—ผ๐˜„๐˜€ ๐—ฎ๐—ฟ๐—ฒ ๐—ฑ๐—ฒ๐˜๐—ฒ๐—ฟ๐—บ๐—ถ๐—ป๐—ถ๐˜€๐˜๐—ถ๐—ฐ. ๐—”๐—ด๐—ฒ๐—ป๐˜๐˜€ ๐—ฎ๐—ฟ๐—ฒ ๐—ฎ๐—ฑ๐—ฎ๐—ฝ๐˜๐—ถ๐˜ƒ๐—ฒ. Thatโ€™s the trade-off: Workflows = stability Agents = flexibility And thatโ€™s why true agents shine in complex GenAI apps where step-by-step logic wonโ€™t cut it. So next time you hear someone say โ€œagent,โ€ ask: Does it just call an LLM? Or does it reason, adapt, and act? Letโ€™s raise the bar for what we call agentic systems.

  • diagram

205

AI/ML engineers, listen up! If you want to be in the top 1%, mastering ML frameworks isnโ€™t enough... You need elite software engineering skills. Why? Because AI/ML systems donโ€™t live in Jupyter notebooks. In production, they need: โ†’ Efficient, scalable code โ†’ Optimized cloud architectures โ†’ High-performance systems that can handle real-world demands One of my clients can attest to this first-hand... We rewrote parts of their **RAG ingestion pipeline in Rust** and reduced latency from **minutes to seconds.** That kind of performance boost doesn't come from just knowing how to prompt a model. โ€ฆ it comes from understanding how systems behave **under load**, how to optimize **I/O and memory**, and how to **build real software.** And thatโ€™s where ๐—–๐—ผ๐—ฑ๐—ฒ๐—–๐—ฟ๐—ฎ๐—ณ๐˜๐—ฒ๐—ฟ๐˜€ comes in... CodeCrafters is one of the best platforms for leveling up real-world engineering skills. (I'm even planning to use it to learn rust) Here's why I say that: - Itโ€™s not just theory โ†ณ Want to understand how Git, Redis, Docker, Kafka, or Shell actually work? โ†ณYouโ€™ll build them yourself, piece by piece. - It forces you to write production-grade code. โ†ณ No hand-holding. No shortcuts. โ†ณ Just hands-on, real-world problem-solving that makes you a better engineer. - Itโ€™s perfect for AI/ML engineers who want to stand out. โ†ณ If youโ€™re building LLM apps, RAG systems, or agentic workflows, you need strong SWE fundamentals. โ†ณ This platform helps you build that foundation. Interested? You'll get a ๐Ÿฐ๐Ÿฌ% ๐—ฑ๐—ถ๐˜€๐—ฐ๐—ผ๐˜‚๐—ป๐˜ if you use my affiliate link. (The link is in the comments)


178

I promised I'd make some bold moves this year. So, here goes nothing... On April 7th, I'll be giving my ๐—™๐—œ๐—ฅ๐—ฆ๐—ง ๐—ถ๐—ป-๐—ฝ๐—ฒ๐—ฟ๐˜€๐—ผ๐—ป ๐˜๐—ฎ๐—น๐—ธ at the QCon Software Development Conferences in London. Talk about biting the bullet! As an introvert (with a sprinkle of social anxiety thrown in), I must admit... I'm actually quite excited. ... and a little nervous - in the best way possible. The opportunity to share insights with an incredible community of engineers and data scientists doesn't come around often. Thus, I'll be sure to give it my all. In my talk, Iโ€™ll be diving into the data dimension of designing LLM and RAG applications. Namely, we will cover: - The data flow and features pipeline that powers LLM and RAG systems - LLMOps best practices for ensuring data is shared, versioned, processed, and analyzed for LLM training and inference. - Concrete use cases like building an LLM Twin and a Second Brain AI Assistant. The goal: To give you a clear, practical framework for architecting the data layer of LLM systems that scales with accuracy, reliability, and efficiency. If you'd like to attend, I'll leave a registration link in the comments (with ยฃ60 off the price code) See you there ๐Ÿ‘Š


176

If you're an aspiring AI engineer, listen up... There are 2 basic skills you must master before touching any ML model: ๐Ÿญ. ๐—ฃ๐—ฟ๐—ผ๐—ด๐—ฟ๐—ฎ๐—บ๐—บ๐—ถ๐—ป๐—ด (Python is a good start) ๐Ÿฎ. ๐—–๐—น๐—ผ๐˜‚๐—ฑ ๐—ฒ๐—ป๐—ด๐—ถ๐—ป๐—ฒ๐—ฒ๐—ฟ๐—ถ๐—ป๐—ด Without these skills, you will not even be able to deploy or scale your AI/ML models effectively in production. Fortunately, ๐ŸŽง Eric Riddoch noticed this problem and filled the need - Enter ๐—–๐—น๐—ผ๐˜‚๐—ฑ ๐—˜๐—ป๐—ด๐—ถ๐—ป๐—ฒ๐—ฒ๐—ฟ๐—ถ๐—ป๐—ด ๐—ณ๐—ผ๐—ฟ ๐—ฃ๐˜†๐˜๐—ต๐—ผ๐—ป ๐——๐—ฒ๐˜ƒ๐—ฒ๐—น๐—ผ๐—ฝ๐—ฒ๐—ฟ๐˜€ live course. Iโ€™ve had the pleasure of knowing Eric for almost a year, and I can confidently say heโ€™s a brilliant cloud, DevOps, and MLOps engineer. Heโ€™s also an incredible teacher with a knack for making complex topics approachable and fun to learn. Hereโ€™s what youโ€™ll learn in the course: - Enterprise-level AWS account management - Fundamentals of cloud engineering - Designing cloud-native RESTful APIs that scale from 4 to 4 million requests/day - Writing, testing, locally mocking, and deploying code using AWS SDK and OpenAI - Advanced observability and monitoring techniques: Logs, metrics, traces, and alerts If you're concerned that the course is not with your desired cloud vendor, don't worry... The principles learned during the course can easily be transferred to other platforms (e.g., GCP and Azure) I strongly recommend this course to anyone looking to lay the foundation for a career in AI, ML, and MLOps engineering. The next cohort starts on March 31st and runs until May 16th! Using the code DECODINGML will get you 10% off your registration. And if that's not enough, Eric offers a scholarship program that can significantly reduce the price - depending on your use case. Check out the link in the comments to enroll.


173

If you think you โ€œknowโ€ prompt engineering... Think again. Iโ€™ve been following Nir Diamant for a while now - his GitHub repos and Substack have become go-to resources for AI practitioners. He has a rare gift: The ability to break down complex GenAI topics like heโ€™s teaching a 7-year-old (without dumbing anything down). ... And now heโ€™s done it again with a new eBook: ๐—ฃ๐—ฟ๐—ผ๐—บ๐—ฝ๐˜ ๐—˜๐—ป๐—ด๐—ถ๐—ป๐—ฒ๐—ฒ๐—ฟ๐—ถ๐—ป๐—ด ๐—ณ๐—ฟ๐—ผ๐—บ ๐—ญ๐—ฒ๐—ฟ๐—ผ ๐˜๐—ผ ๐—›๐—ฒ๐—ฟ๐—ผ โ€“ ๐— ๐—ฎ๐˜€๐˜๐—ฒ๐—ฟ ๐˜๐—ต๐—ฒ ๐—”๐—ฟ๐˜ ๐—ผ๐—ณ ๐—”๐—œ ๐—œ๐—ป๐˜๐—ฒ๐—ฟ๐—ฎ๐—ฐ๐˜๐—ถ๐—ผ๐—ป This isnโ€™t just another โ€œuse more bullet points in your promptโ€ kind of guide. Itโ€™s a practical deep dive with: โ†’ Code examples โ†’ Real-world exercises โ†’ Clear explanations of common mistakes โ†’ And the subtle mechanics behind great AI interaction One reader put it best: โ€œ๐˜ž๐˜ฐ๐˜ธ... ๐˜ ๐˜ต๐˜ฉ๐˜ฐ๐˜ถ๐˜จ๐˜ฉ๐˜ต ๐˜ ๐˜ฌ๐˜ฏ๐˜ฆ๐˜ธ ๐˜ฑ๐˜ณ๐˜ฐ๐˜ฎ๐˜ฑ๐˜ต ๐˜ฆ๐˜ฏ๐˜จ๐˜ช๐˜ฏ๐˜ฆ๐˜ฆ๐˜ณ๐˜ช๐˜ฏ๐˜จ ๐˜ฃ๐˜ฆ๐˜ง๐˜ฐ๐˜ณ๐˜ฆ ๐˜ณ๐˜ฆ๐˜ข๐˜ฅ๐˜ช๐˜ฏ๐˜จ ๐˜ต๐˜ฉ๐˜ช๐˜ด ๐˜ฃ๐˜ฐ๐˜ฐ๐˜ฌ, ๐˜ฃ๐˜ถ๐˜ต ๐˜ ๐˜ธ๐˜ข๐˜ด ๐˜ญ๐˜ช๐˜จ๐˜ฉ๐˜ต๐˜บ๐˜ฆ๐˜ข๐˜ณ๐˜ด ๐˜ข๐˜ธ๐˜ข๐˜บ ๐˜ง๐˜ณ๐˜ฐ๐˜ฎ ๐˜ฆ๐˜ท๐˜ฆ๐˜ฏ ๐˜ฃ๐˜ฆ๐˜ช๐˜ฏ๐˜จ ๐˜ข๐˜ธ๐˜ข๐˜ณ๐˜ฆ ๐˜ฐ๐˜ง ๐˜ต๐˜ฉ๐˜ฆ ๐˜ฏ๐˜ถ๐˜ข๐˜ฏ๐˜ค๐˜ฆ ๐˜ข๐˜ฏ๐˜ฅ ๐˜ค๐˜ฐ๐˜ฎ๐˜ฑ๐˜ญ๐˜ฆ๐˜น๐˜ช๐˜ต๐˜ช๐˜ฆ๐˜ด... ๐˜›๐˜ฉ๐˜ช๐˜ด ๐˜ช๐˜ด ๐˜ต๐˜ฉ๐˜ฆ ๐˜ฃ๐˜ฐ๐˜ฐ๐˜ฌ ๐˜ต๐˜ฐ ๐˜ฑ๐˜ช๐˜ค๐˜ฌ ๐˜ถ๐˜ฑ ๐˜ช๐˜ง ๐˜บ๐˜ฐ๐˜ถ ๐˜ธ๐˜ข๐˜ฏ๐˜ต ๐˜ต๐˜ฐ ๐˜จ๐˜ฆ๐˜ต ๐˜ฑ๐˜ข๐˜ด๐˜ต ๐˜ซ๐˜ถ๐˜ด๐˜ต ๐˜ฌ๐˜ฏ๐˜ฐ๐˜ธ๐˜ช๐˜ฏ๐˜จ ๐˜ธ๐˜ฉ๐˜ข๐˜ต ๐˜ฑ๐˜ณ๐˜ฐ๐˜ฎ๐˜ฑ๐˜ต ๐˜ฆ๐˜ฏ๐˜จ๐˜ช๐˜ฏ๐˜ฆ๐˜ฆ๐˜ณ๐˜ช๐˜ฏ๐˜จ ๐˜ช๐˜ด, ๐˜ข๐˜ฏ๐˜ฅ ๐˜ด๐˜ต๐˜ข๐˜ณ๐˜ต ๐˜ฎ๐˜ข๐˜ด๐˜ต๐˜ฆ๐˜ณ๐˜ช๐˜ฏ๐˜จ ๐˜ต๐˜ฉ๐˜ฆ ๐˜ค๐˜ณ๐˜ข๐˜ง๐˜ต.โ€ I couldnโ€™t agree more. If youโ€™re building LLM apps, agentic systems, or even just playing with ChatGPT prompts, check this eBook out. Youโ€™ll walk away with a new toolkit that goes far beyond theory. Want a copy? The link is in the comments.

  • No alternative text description for this image

180

Here's something most people overlook: The RAG feature pipeline is the most important part of the entire AI stack. Not the LLM. Not the prompt. Not even the fancy agent framework. Why? Because if your retrieval sucks, your generation will too - no matter how good your model is. Let me walk you through how we architected our RAG feature pipeline for the ๐—ฆ๐—ฒ๐—ฐ๐—ผ๐—ป๐—ฑ ๐—•๐—ฟ๐—ฎ๐—ถ๐—ป ๐—”๐—œ ๐—”๐˜€๐˜€๐—ถ๐˜€๐˜๐—ฎ๐—ป๐˜: The pipeline ingests raw documents from MongoDB, where both Notion and crawled content are stored in a single standardized collection. We donโ€™t care about the source - just that the data is clean and usable. The output? โ†’ Chunked + embedded documents โ†’ Stored in a dedicated vector store โ†’ Indexed and ready for semantic search Where does this pipeline fit? Itโ€™s an offline batch pipeline, decoupled from the live user experience. At query time, the pipeline does not run. All processing is done beforehand, so retrieval is fast, stable, and cost-efficient. Meanwhile, the online pipeline (our agentic RAG module) fetches the chunks, reasons over them, and generates the answer. The architecture is made up of 7 key components: ๐Ÿญ/ ๐——๐—ฎ๐˜๐—ฎ ๐—˜๐˜…๐˜๐—ฟ๐—ฎ๐—ฐ๐˜๐—ถ๐—ผ๐—ป Pulls all raw documents from MongoDB, regardless of source. ๐Ÿฎ/ ๐——๐—ผ๐—ฐ๐˜‚๐—บ๐—ฒ๐—ป๐˜ ๐—™๐—ถ๐—น๐˜๐—ฒ๐—ฟ๐—ถ๐—ป๐—ด Applies quality scores to drop noisy or low-value docs. ๐Ÿฏ/ ๐—–๐—ต๐˜‚๐—ป๐—ธ๐—ถ๐—ป๐—ด Splits documents into manageable segments for vectorization. ๐Ÿฐ/ ๐—ฃ๐—ผ๐˜€๐˜-๐—ฃ๐—ฟ๐—ผ๐—ฐ๐—ฒ๐˜€๐˜€๐—ถ๐—ป๐—ด Applies one of two retrieval strategies (chosen from our YAML configs): Parent Retrieval โ†’ Links each chunk to its full doc Contextual Retrieval โ†’ Adds summaries to enrich semantic relevance ๐Ÿฑ/ ๐—˜๐—บ๐—ฏ๐—ฒ๐—ฑ๐—ฑ๐—ถ๐—ป๐—ด All chunks (for both strategies) are vectorized using a configurable embedding model (OpenAI or Hugging Face) ๐Ÿฒ/ ๐—œ๐—ป๐—ฑ๐—ฒ๐˜…๐—ถ๐—ป๐—ด Embeddings are stored and indexed in MongoDB for fast lookups. ๐Ÿณ/ ๐—™๐—ถ๐—ป๐—ฎ๐—น ๐—ข๐˜‚๐˜๐—ฝ๐˜‚๐˜ A structured, searchable knowledge baseโ€”ready for RAG-based generation. The entire pipeline is managed by ZenML, which is: โ†’ Reproducible โ†’ Configurable โ†’ Versioned โ†’ Traceable If youโ€™re serious about building production-grade GenAI systems, this is where you focus. ... because 90% of your generation's quality is determined before the LLM even gets involved. Letโ€™s stop treating pipelines like afterthoughts. They are the product. Want to learn more? Check the link in the comments.

  • No alternative text description for this image

242

RAG has kept โ€œdyingโ€ for the past 4 years. But hereโ€™s why that will never happen: All the LLMs (even the most advanced ones) struggle without the right context. It doesnโ€™t matter if your model has 128k+ token windows or cutting-edge fine-tuning... If it doesnโ€™t retrieve the right data or the context is full of noise or formatted incorrectly, it wonโ€™t generate the right answers. Thatโ€™s why retrieval is the hardest part of RAG. Most RAG failures arenโ€™t about generation - they happen before the LLM even sees the data. If the retrieval step is weak, your AI assistant will: - Fetch irrelevant information - Miss critical details - Hallucinate confidently wrong responses But more context isnโ€™t the answer... Better context is. Lesson 5 of the Second Brain AI Assistant course is all about fixing retrieval with a production-ready RAG feature pipeline that. (And it's now live!) Namely, in this lesson, you will learn: - The fundamentals of RAG. - How to design and implement a production-ready RAG pipeline - Implement contextual retrieval (an advanced RAG technique) from scratch. - Implement parent retrieval (another advanced RAG technique) using LangChain. - Extend LangChain to add custom behavior using OOP. - The critical role of chunk size in optimizing retrieval quality - Write a configuration layer to switch between different algorithms and models dynamically - How to manage everything with an MLOps framework (we use ZenML) By the end of this lesson, youโ€™ll be equipped with the skills to build a flexible, modular RAG feature pipeline. This pipeline gives access to our AI assistant to our Second Brain and provides reliable context to generate meaningful answers. Sounds interesting? Pick up lesson 5 today. (The link is in the comments) Thank you, Anca Ioana Muscalagiu, for contributing with another fantastic lesson to @Decoding ML !


263

Over the past 2 years, I've been on a mission... (And it's finally coming together) My goal has been to give as much back to the AI community as possible. I've done this by creating a total of 6 open-source courses designed to help others build real-world, production-grade AI systems. Now, I'm thrilled to announce that I've brought all my FREE courses together in one place. These are not your typical tutorials showing isolated concepts in scattered Notebooks... They're designed for those looking to master production and industry-level AI. It covers everything from: - End-to-end GenAI, LLM, RAG apps - MLOps/LLMOps for AI scalability - AI Systems for real-world applications I've been sure to constantly refine and improve these courses to keep them at the cutting edge. Our latest project, the ๐—ฆ๐—ฒ๐—ฐ๐—ผ๐—ป๐—ฑ ๐—•๐—ฟ๐—ฎ๐—ถ๐—ป ๐—”๐—œ ๐—”๐˜€๐˜€๐—ถ๐˜€๐˜๐—ฎ๐—ป๐˜ course, is our crown jewel so far - It's an end-to-end project that takes you through the full process of building a production-level AI assistant. The world of AI is moving incredibly fast. ... I believe that knowledge-sharing is key to pushing the field forward. Iโ€™m currently working on a new open-source course focused on AI agents that I know youโ€™ll love. (More on that soon) All in all, I'm incredibly proud of what we've built so far. I'm excited to see how these resources help others in their AI journey. Check it out, share it, and let's keep building the future together. The link is in the comments.


256

There's a major problem with using general-purpose LLMs to build apps. But not enough people are talking about it... Of course, I'm talking about: โ†’ Skyrocketing API bills โ†’ Vendor lock-in โ†’ Degrading performance we can't do anything about โ†’ Lack of control of your data Thankfully, I've found a solution. In Lesson 4 of the ๐—ฆ๐—ฒ๐—ฐ๐—ผ๐—ป๐—ฑ ๐—•๐—ฟ๐—ฎ๐—ถ๐—ป ๐—”๐—œ ๐—”๐˜€๐˜€๐—ถ๐˜€๐˜๐—ฎ๐—ป๐˜ course, we tackle these challenges head-on. Specifically, you'll learn: - How to fine-tune open-source small language models (SLMs) using LoRA and QLoRA to specialize them on your tasks. - Efficient fine-tuning with tools like TRL, Unsloth AI, and Comet - Architecting modular, scalable pipelines with MLOps in mind for production-ready applications. - Deploying your own specialized models on Hugging Face Inference Endpoints as a real-time API endpoint. - Evaluating your models using vLLM to ensure theyโ€™re production-ready. By the end of this lesson, you'll have the tools and knowledge to take back control of your AI systems (without breaking the bank or being tied to a single provider) Ready to get started? The link is in the comments.


276

AI engineers must understand this to build any successful system: (Especially when working with RAG, LLMs, or agent-based apps) How to architect data pipelines. We all know data is the lifeblood of the systems we build. You can have the most advanced algorithms, but without clean data, you're just spinning your wheels. To understand how data pipelines for LLM apps should look, letโ€™s architect one that powers our Second Brain AI Assistant. Here are the core components: ๐Ÿญ/ ๐——๐—ฎ๐˜๐—ฎ ๐—ฐ๐—ผ๐—น๐—น๐—ฒ๐—ฐ๐˜๐—ถ๐—ผ๐—ป Data sources can be many things - Notion, crawled links, custom APIs... you name it. In this case, we'll use Notionโ€™s API to collect personal data, extract all the links, and standardize everything into Markdown format. Why Markdown? Because the internet is full of it. Thus, LLMs are heavily trained using this format and know how to work with it. To replicate steps down the line, we make a snapshot of the raw data into a data lake like S3. ๐Ÿฎ/ ๐—˜๐—ง๐—Ÿ ๐—ฑ๐—ฎ๐˜๐—ฎ ๐—ฝ๐—ถ๐—ฝ๐—ฒ๐—น๐—ถ๐—ป๐—ฒ This is where the magic happens. After collecting the data, we perform several transformations: 1. Download the Notion docs from S3 2. Load them in memory into Pydantic objects (content + metadata) 3. Crawl all the links inside the Notion docs (when building the AI assistant, we want to generate answers based on the content of the links as well โ†’ this is where the power of our lists of saved resources kicks in) 4. Compute a quality score per document using a combination of heuristics and LLMs (as LLM calls are expensive, you first want to squeeze the most out of it using heuristics) 5. Store all the documents and the quality score inside a document database (with this, we can compute statistics and filter documents down the line based on our needs) For our setup, weโ€™ve used MongoDB: a Swiss knife to store our unstructured documents and, later on, the embeddings for RAG. We can think about it as part of our logical feature store. ๐Ÿฏ/ ๐— ๐—Ÿ๐—ข๐—ฝ๐˜€ ๐—™๐—ฟ๐—ฎ๐—บ๐—ฒ๐˜„๐—ผ๐—ฟ๐—ธ To keep everything organized and maintainable, we use ZenML This helps us easily manage, version, and deploy our offline data pipelines. ๐—ง๐—ฎ๐—ธ๐—ฒ๐—ฎ๐˜„๐—ฎ๐˜†: Architecting and managing these pipelines is crucial for creating LLM/RAG systems. If you're an AI/ML engineer stepping into the GenAI world, get comfortable building and managing data pipelines that crawl unstructured data and process it for LLMs. (Yes! Along with the 1000+ things you must know, AI Engineers must also dig into data engineering.) Want to see how to do this? Check the link in the comments.


341

Less than 3 years ago, I graduated with my masterโ€™s in ML. Fast forward to today, Iโ€™m: - Contracting - Consulting - Creating content Globally! So, what helped me accelerate my career? ๐Ÿญ. ๐—œ ๐˜€๐˜๐—ฎ๐—ฟ๐˜๐—ฒ๐—ฑ ๐—ฒ๐—ฎ๐—ฟ๐—น๐˜† While still in university, I didnโ€™t wait for the โ€œperfectโ€ opportunity. By my third year, I had already landed a junior SWE role and was working like crazy. ๐Ÿฎ. ๐—œ ๐—ฝ๐˜‚๐˜ ๐—ถ๐—ป ๐˜๐—ต๐—ฒ ๐˜๐—ถ๐—บ๐—ฒ - ๐—ป๐—ผ ๐˜€๐—ต๐—ผ๐—ฟ๐˜๐—ฐ๐˜‚๐˜๐˜€. For years, I spent 10-12 hours a day researching, building, and creating- mostly alone. Monday to Saturday. That helped me transform 3 years of experience into 6. ๐Ÿฏ. ๐—œ ๐—ฎ๐—ฐ๐—ฐ๐—ฒ๐—ฝ๐˜๐—ฒ๐—ฑ ๐˜๐—ต๐—ฒ ๐˜๐—ฟ๐—ฎ๐—ฑ๐—ฒ-๐—ผ๐—ณ๐—ณ๐˜€ I knew that accelerating my career meant sacrifices in other areas. I had less time for social life. ... but this meant more time for deep work and learning There was only one exception: My girlfriend. No matter how much I worked, I made time for her - she kept me sane. ๐Ÿฐ. ๐—›๐—ฒ๐—ฎ๐—น๐˜๐—ต > ๐—›๐˜‚๐˜€๐˜๐—น๐—ฒ โ†’ Workouts: 3-4x per week โ†’ Food: Eat clean, avoid junk โ†’ Sleep: 7-8 hours, no compromise โ†’ Stress: Keep it under control (no toxic jobs, no endless pressure) ๐Ÿฑ. ๐—ก๐—ผ ๐—ฎ๐—น๐—ฐ๐—ผ๐—ต๐—ผ๐—น. ๐—ก๐—ผ ๐˜€๐—บ๐—ผ๐—ธ๐—ถ๐—ป๐—ด. Clarity is everything. You canโ€™t afford to be foggy-minded when making big moves. (I still enjoy exceptions here and there, but nothing that slows me down.) ๐Ÿฒ. ๐—œ ๐—ณ๐—ผ๐—น๐—น๐—ผ๐˜„๐—ฒ๐—ฑ ๐—บ๐˜† ๐—ถ๐—ป๐˜๐˜‚๐—ถ๐˜๐—ถ๐—ผ๐—ป - ๐—ฎ๐—น๐˜„๐—ฎ๐˜†๐˜€. I ignored the โ€œsafeโ€ paths. Regardless of what โ€œmore experiencedโ€ people told me, I always made hard decisions. Most people project their fears onto you, but they donโ€™t live your life. ๐Ÿณ. ๐—œ ๐˜๐—ผ๐—ผ๐—ธ ๐˜๐—ต๐—ฒ ๐—บ๐—ถ๐—ป๐—ฑ๐˜€๐—ฒ๐˜ ๐—ผ๐—ณ ๐—ฎ๐—ป ๐—ฒ๐˜…๐—ฝ๐—น๐—ผ๐—ฟ๐—ฒ๐—ฟ If you follow paved paths, youโ€™re in high competition. If you carve your own path, you move faster. I never aimed to be top 1% in existing tracks. I created my own track instead. ๐Ÿด. ๐—œ ๐—ป๐—ฒ๐˜ƒ๐—ฒ๐—ฟ ๐˜€๐—ฒ๐˜๐˜๐—น๐—ฒ๐—ฑ When my managers said: - โ€œThe job market is tough.โ€ - โ€œThis is what others get paid.โ€ - โ€œYouโ€™re asking for too much.โ€ - โ€œHave patience.โ€ I didnโ€™t listen. I doubled down and kept pushing forward. Itโ€™s exhausting at times, but curiosity > comfort. And thatโ€™s what kept me going. So, if youโ€™re early in your career, hereโ€™s my advice: โ†’ Be intentional about how you spend your time. โ†’ Build skills relentlessly - donโ€™t wait for permission. โ†’ Challenge conventional wisdom - not all โ€œadviceโ€ is useful. โ†’ Prioritize health - burnout kills long-term success. โ†’ Bet on yourself - no one else will do it for you. Happy Sunday!

  • No alternative text description for this image

290

Many ML projects fail to transition from POC to production-ready. Here's one simple reason why: The initial focus was never on scalability or production constraints. Fraud detection presents a perfect use case for building production-first ML systems. It combines the need for real-time and batch processing, low-latency predictions, and high-accuracy models. Here's a look into what that system could look like: ๐——๐—ฎ๐˜๐—ฎ ๐˜€๐—ผ๐˜‚๐—ฟ๐—ฐ๐—ฒ๐˜€ In fraud detection, you deal with real-time transactions, streaming data and historical records. Real-time and streaming data require you to instantly compute features, while historical records help track user profiles and spending patterns. In our architecture, real-time features are computed through HTTP requests, streaming data will flow through Kafka topics, with historical data stored in a data warehouse for batch processing. ๐—™๐—ฒ๐—ฎ๐˜๐˜‚๐—ฟ๐—ฒ ๐—ฝ๐—น๐—ฎ๐˜๐—ณ๐—ผ๐—ฟ๐—บ At the heart of the system is the feature platform, like Tecton ... This centralizes all the features. More specifically, it allows us to manage features in an offline store (for training - high throughput) and online store (for serving - low latency). Using the same feature engineering logic during training and inference avoids training-serving skew. ๐—™๐—ฒ๐—ฎ๐˜๐˜‚๐—ฟ๐—ฒ ๐—ฝ๐—ถ๐—ฝ๐—ฒ๐—น๐—ถ๐—ป๐—ฒ๐˜€ These convert raw data into meaningful features. By centralizing your features into a feature store like Tecton, you can leverage their feature views to define features once and reuse them across models and pipelines. A feature view is defined as a data source(s) + a function that maps raw data into features. Next, using a process known as materialization, you sync the raw data sources with the online/offline stores while applying the transformations. ๐—ง๐—ฟ๐—ฎ๐—ถ๐—ป๐—ถ๐—ป๐—ด ๐—ฝ๐—ถ๐—ฝ๐—ฒ๐—น๐—ถ๐—ป๐—ฒ๐˜€ It ingests features and labels to train models (stored in a model registry). Leveraging the feature store, you can easily apply time-traveling strategies to version your dataset. ๐—œ๐—ป๐—ณ๐—ฒ๐—ฟ๐—ฒ๐—ป๐—ฐ๐—ฒ ๐—ฝ๐—ถ๐—ฝ๐—ฒ๐—น๐—ถ๐—ป๐—ฒ๐˜€ It takes new transaction data, enriches it with features from Tecton's feature platform, and applies the trained model to generate predictions. Online stores are crucial at serving time. They provide low-latency access to up-to-date features. When a transaction occurs... The pipeline quickly retrieves pre-computed features, combines them with real-time features, and computes predictions. ๐—ข๐—ฏ๐˜€๐—ฒ๐—ฟ๐˜ƒ๐—ฎ๐—ฏ๐—ถ๐—น๐—ถ๐˜๐˜† ๐—ฝ๐—ถ๐—ฝ๐—ฒ๐—น๐—ถ๐—ป๐—ฒ Lastly, an observability pipeline is essential for monitoring the system's health and detecting drifts. The final touch is an alarm system that sends emails or SMS or denies transactions if fraud is detected. Want to dive deeper into building such systems? Check out the link in the comments.


287

The course is finally complete! You can learn for free how to build your Second Brain AI assistant with LLMs, agents, RAG, fine-tuning, LLMOps, and AI systems techniques. Thanks for sharing Andrew Davidson is the crown jewel of Decoding ML

Profile picture of Andrew Davidson

Andrew Davidson


Looking for weekend fun? Build a second brain AI assistant with Paul Iusztin's course! https://lnkd.in/eUztq49A The 6th and final lesson just dropped.


293

RAG is evolving. Standard retrieval is no longer enough... If youโ€™re building LLM-powered applications, you need something more powerful: ๐—”๐—ด๐—ฒ๐—ป๐˜๐—ถ๐—ฐ ๐—ฅ๐—”๐—š. Thatโ€™s exactly what weโ€™re tackling in Lesson 6 of the Second Brain AI Assistant course. ... and it's now LIVE! Most RAG systems passively retrieve context and hope the LLM generates the right response. But this approach is limited. โ†’ What if your assistant needs to reason about multiple sources? โ†’ What if retrieval alone doesnโ€™t fully align with the userโ€™s intent? โ†’ What if the retrieved context isn't enough, and the system needs to iterate? Agentic RAG bridges that gap. Instead of just fetching documents, agents dynamically plan, retrieve, and refine their responses, making AI assistants: โ†’ Smarter โ†’ More interactive โ†’ More accurate And in Lesson 6, weโ€™re building one from scratch... Specifically, youโ€™ll learn: โ†’ How to build an Agentic RAG module that goes beyond simple retrieval โ†’ Integrating retrieval with AI agents for dynamic, multi-step reasoning โ†’ Adding LLMOps, such as prompt monitoring to optimize retrieval workflows for cost, latency, and throughput โ†’ Evaluating long and complex LLM chains to ensure reliability (moderation, hallucination, response performance) โ†’ Scaling retrieval architectures to handle real-world AI assistant demands By the end of this lesson, youโ€™ll understand what it takes to build stable, efficient, and intelligent RAG-powered assistants. Huge thanks to Anca Ioana Muscalagiu for contributing another fantastic lesson to Decoding ML! The link for Lesson 6 is in the comments :)


1k

If you're building GenAI apps, you need to understand one thing: LLMs donโ€™t โ€œthinkโ€ like humans. They generate responses based on probabilities. ... and how you craft your prompts can make or break your application. Thatโ€™s why Iโ€™ve been diving deep into "๐—ฃ๐—ฟ๐—ผ๐—บ๐—ฝ๐˜ ๐—˜๐—ป๐—ด๐—ถ๐—ป๐—ฒ๐—ฒ๐—ฟ๐—ถ๐—ป๐—ด ๐—ณ๐—ผ๐—ฟ ๐—Ÿ๐—Ÿ๐— ๐˜€." I must say, itโ€™s one of the best resources out there for practitioners building with LLMs. โ†’ Itโ€™s written with character โ†’ It focuses on how to actually build GenAI products (practical theory) โ†’ Itโ€™s packed with fundamentals that every AI/ML engineer should know I especially loved how they teach prompt engineering beyond basic techniques you can find everywhere. They teach you how to manage and process the context injected into the prompts, which is required to scale LLM apps beyond PoCs. Iโ€™m still working my way through the book, but after finishing the first 7 chapters, I can confidently say it's a must-read. Specifically for: โ†’ Anyone building GenAI applications โ†’ Engineers looking to optimize their prompt engineering skills โ†’ Experienced AI practitioners looking to build modular LLM apps. I'd also like to say a huge thanks to John Berryman for sending me a copy - I'm absolutely loving it. If youโ€™re serious about building with LLMs, this book is worth your time. (I'll leave a link to it in the comments)


435

Iโ€™ve just started working at a new AI start-up (still in shadow mode) on building cutting-edge products using GenAI, LLMs, RAG, and Agents. All the good stuff. This will help me improve my content with even more real-world battle-tested content and stories. So excited to start providing more in-depth insights from this move ๐ŸคŸ


    1k

    Cursor + Claude 3.7 keeps disappointing me. I am not a vibe coder. I strongly articulate what should be implemented, how it should be implemented, and the issues within the codebase. But when things get tricky and require extensive understanding of the code, infrastructure and data, it just doesnโ€™t work. I lost 2 hours of my life talking to Claude (Extended) when a GitHub Issues thread solved it in 5 minutes. My take is that LLMs are great for quickly ramping up your project (instead of digging into the documentation), but you still have to: - design - plan - know the best engineering practices - be open to use Google What is your experience?


    1k

    I canโ€™t get enough of this meme ๐Ÿ˜‚. Are you down that path yet?


    1k

    Want to drive more opportunities from LinkedIn?

    Content Inspiration, AI, scheduling, automation, analytics, CRM.

    Get all of that and more in Taplio.

    Try Taplio for free

    Famous LinkedIn Creators to Check Out

    Austin Belcak

    @abelcak

    CultivatedCulture.com // I teach people how to land jobs they love in today's market without traditi...

    1m

    Followers

    Matt Gray

    @mattgray1

    Over the last decade, Iโ€™ve built 4 successful companies and a community of over 14 million people. ...

    1m

    Followers

    Richard Moore

    @richardjamesmoore

    โฉYou know how all the clients you'll ever work with are on LinkedIn, right? But you struggle to gene...

    103k

    Followers

    Shlomo Genchin

    @shlomogenchin

    Hey! Here are 3 ways I can help you: 1๏ธโƒฃ Talks and Workshops: I'll show your team, or students, how...

    49k

    Followers

    Sam G. Winsbury

    @sam-g-winsbury

    We turn entrepreneurs into credible thought leaders through personal branding so they can scale thei...

    45k

    Followers

    Hi! Iโ€™m Daniel. Iโ€™m the creator of The Marketing Millennials and the founder of Authority, a B2B Lin...

    147k

    Followers

    Sabeeka Ashraf

    @sabeekaashraf

    You know whatโ€™s crazy? This next line youโ€™re about to read... Kiss. Marry. Kill: Elon Musk? That ...

    20k

    Followers

    Ash Rathod

    @ashrathod

    You already know storytelling is essential for your business and brand. But storytelling is much m...

    73k

    Followers

    Izzy Prior

    @izzyprior

    No matter how outrageously amazing your mission is, it's likely you're not seeing the results you ne...

    81k

    Followers

    Vaibhav Sisinty โ†—๏ธ

    @vaibhavsisinty

    I'm an engineer turned marketer, now a founder. I've worked at Uber and Klook, focusing on marketi...

    446k

    Followers

    Wes Kao

    @weskao

    Wes Kao is an entrepreneur, coach, and advisor who writes at newsletter.weskao.com. She is co-founde...

    107k

    Followers

    Luke Matthews

    @lukematthws

    LinkedIn has changed. You need to change too. Hey I'm Luke, I've been marketing for 5+ years on ...

    186k

    Followers

    Justin Welsh

    @justinwelsh

    Over the last decade, I helped build two companies past a $1B valuation and raise over $300M in vent...

    1m

    Followers

    Amelia Sordell ๐Ÿ”ฅ

    @ameliasordell

    Klowt builds personal brands. I founded the business after realising that the best leads came throu...

    228k

    Followers

    Tibo Louis-Lucas

    @thibaultll

    Founder Prev Taplio & Tweet Hunter (sold) Building Typeframes & revid.ai Invested in animstats.com ...

    6k

    Followers

    Andy Mewborn

    @amewborn

    I use to be young & cool. Now I do b2b SaaS. Husband. Dad. Ironman. Founder of Distribute // Co-fo...

    212k

    Followers

    Guillaume Moubeche

    @-g-

    If youโ€™re here, that's because you know that your personal growth will drive your business growth ๐Ÿš€...

    80k

    Followers

    Sahil Bloom

    @sahilbloom

    Sahil Bloom is the New York Times Bestselling author of The 5 Types of Wealth: A Transformative Guid...

    1m

    Followers