Get the Linkedin stats of Paul Iusztin and many LinkedIn Influencers by Taplio.
open on linkedin
I am a senior machine learning engineer and contractor with ๐ฒ+ ๐๐ฒ๐ฎ๐ฟ๐ ๐ผ๐ณ ๐ฒ๐ ๐ฝ๐ฒ๐ฟ๐ถ๐ฒ๐ป๐ฐ๐ฒ. I design and implement modular, scalable, and production-ready ML systems for startups worldwide. My central mission is to build data-intensive AI/ML products that serve the world. Since training my first neural network in 2017, I have 2 passions that fuel my mission: โ Designing and implementing production AI/ML systems using MLOps best practices. โ Teaching people about the process. . I currently develop production-ready Deep Learning products at Metaphysic, a leading GenAI platform. In the past, I built Computer Vision and MLOps solutions for CoreAI, Everseen, and Continental. Also, I am the Founder of Decoding ML, a channel for battle-tested content on learning how to design, code, and deploy production-grade ML and MLOps systems. I am writing articles and posts each week on: - ๐๐ช๐ฏ๐ฌ๐ฆ๐ฅ๐๐ฏ: 29k+ followers - ๐๐ฆ๐ฅ๐ช๐ถ๐ฎ: 2.5k+ followers ~ ๐ https://medium.com/@pauliusztin - ๐๐ถ๐ฃ๐ด๐ต๐ข๐ค๐ฌ (๐ฏ๐ฆ๐ธ๐ด๐ญ๐ฆ๐ต๐ต๐ฆ๐ณ): 6k+ followers ~ ๐ https://decodingml.substack.com/ . If you want to learn how to build an end-to-end production-ready LLM & RAG system using MLOps best practices, you can take Decoding MLโs self-guided free course: โ ๐๐๐ ๐๐ธ๐ช๐ฏ ๐๐ฐ๐ถ๐ณ๐ด๐ฆ: ๐๐ถ๐ช๐ญ๐ฅ๐ช๐ฏ๐จ ๐ ๐ฐ๐ถ๐ณ ๐๐ณ๐ฐ๐ฅ๐ถ๐ค๐ต๐ช๐ฐ๐ฏ-๐๐ฆ๐ข๐ฅ๐บ ๐๐ ๐๐ฆ๐ฑ๐ญ๐ช๐ค๐ข ~ ๐ https://github.com/decodingml/llm-twin-course . ๐ฌ If you need machine learning solutions for your business, letโs discuss! ๐ Only open to full remote positions as a contractor. . Contact: ๐ฑ Phone: +40 732 509 516 โ๏ธ Email: p.b.iusztin@gmail.com ๐ป Decoding ML: https://linktr.ee/decodingml ๐ต๐ปโโ๏ธ Personal site & Socials: https://www.pauliusztin.me/
Check out Paul Iusztin's verified LinkedIn stats (last 30 days)
Use Taplio to search all-time best posts
Iโve been working with GenAI for 3+ years. Hereโs something all engineers must come to terms with: If youโre building LLM-powered applications, at some point, youโll need to generate high-quality datasets to fine-tune SLMs. Why? โ Fine-tuning SLMs reduces costs, latency, and throughput while maintaining high accuracy for specific tasks. โ Some domains require specialized fine-tuning for better domain adaptation. โ Fine-tuned models give you more control over AI behavior and response generation. Thatโs exactly what weโre tackling with our ๐ฆ๐ฒ๐ฐ๐ผ๐ป๐ฑ ๐๐ฟ๐ฎ๐ถ๐ป ๐๐ ๐๐๐๐ถ๐๐๐ฎ๐ป๐. ... and today, Iโm breaking down the dataset generation feature pipeline we built for fine-tuning our summarization SLM. The input to our generation pipeline will be raw documents from MongoDB (Notion & crawled resources). And the output is a high-quality summarization dataset published to Hugging Faceโs dataset registry. Since this pipeline generates features used to train an LLM, itโs called a feature pipeline. Hereโs how it works, step by step: ๐ญ. ๐๐ฎ๐๐ฎ ๐๐ ๐๐ฟ๐ฎ๐ฐ๐๐ถ๐ผ๐ป โ Pulls raw documents from MongoDB and standardizes formatting. ๐ฎ. ๐๐ผ๐ฐ๐๐บ๐ฒ๐ป๐ ๐๐ ๐ฝ๐น๐ผ๐ฟ๐ฎ๐๐ถ๐ผ๐ป โ Analyzes length & quality scores distributions to make informed decisions. ๐ฏ. ๐๐ฎ๐๐ฎ ๐๐ถ๐น๐๐ฒ๐ฟ๐ถ๐ป๐ด โ Removes low-value content, keeping only high-quality documents. ๐ฐ. ๐ฆ๐๐บ๐บ๐ฎ๐ฟ๐ถ๐๐ฎ๐๐ถ๐ผ๐ป โ We use a more powerful LLM (e.g., `gpt-4o`) to generate multiple summaries per document by varying temperature and sampling parameters (a process known as distillation) ๐ฑ. ๐ค๐๐ฎ๐น๐ถ๐๐ ๐๐ผ๐ป๐๐ฟ๐ผ๐น โ Filters out poor-quality summaries. ๐ฒ. ๐๐ฎ๐๐ฎ๐๐ฒ๐ ๐ฆ๐ฝ๐น๐ถ๐๐๐ถ๐ป๐ด โ Divides data into training, evaluation, and test sets (done before storing the dataset and not at training time!) ๐ณ. ๐ฉ๐ฒ๐ฟ๐๐ถ๐ผ๐ป๐ถ๐ป๐ด & ๐๐ฒ๐ฝ๐น๐ผ๐๐บ๐ฒ๐ป๐ โ Publishes the final dataset to Hugging Face. To keep the pipeline reproducible, trackable, and scalable, we manage it using ZenML, which: โ Orchestrates the entire workflow from extraction to deployment. โ Ensures traceability & versioning of pipeline runs & datasets. โ Allows dynamic configuration for different filtering, summarization & structuring techniques. Even if youโre not deep into fine-tuning, at some point, youโll need a structured way to generate datasets for specialized AI applications. This is one of the most critical components of your pipeline. Want to learn more? Check out the link in the comments.
The unthinkable happened... (And it wouldn't have been possible without you) The LLM Engineers Handbook GitHub repo has just crossed 2,800 stars. But that's not all... It's now less than 100 stars away from being the most popular repository in the Packt GitHub organization. Honestly, I'm flattered. This milestone wouldn't have been possible without your support. Iโm truly grateful for every one of you who has contributed, starred, or even shared the repo. Weโve been working hard to make this repository a powerful resource for anyone looking to master LLMs. Along with the book, you'll learn how to implement production-ready, industry-level LLM & RAG applications. (Of course, following best practices in software engineering, system design and MLOps) And we've made some changes: - Weโve accepted PRs from contributors to fix bugs - Docs have been updated based on reader feedback If you haven't already, check it out. Contribute to the repo and help us hit that No. 1 spot. LFG๐ฅ (Link in the comments)
Over the past 6 months, Iโve seen the word โagentโ thrown around a lot. But letโs be honest, most of the time it's misleading... In reality, what people often refer to as an "agent" is just Python scripts wrapped around LLM calls. (Iโve fallen into that trap myself early on.) But agents (and agentic systems) are fundamentally different from classic LLM workflows. Hereโs how I break it down: ๐ช๐ผ๐ฟ๐ธ๐ณ๐น๐ผ๐๐ โ Follow a fixed, predefined sequence โ Every step is planned in advance โ Think โassembly lineโ logicโreliable, repeatable, predictable โ Great for structured, stable tasks ๐๐ด๐ฒ๐ป๐๐ โ Decide what to do next on the fly โ Use reasoning to pave their path dynamically โ Think of them like an intelligent assistant, dynamically adapting its plan โ Ideal for unstructured, evolving tasks ๐ช๐ผ๐ฟ๐ธ๐ณ๐น๐ผ๐๐ ๐ฎ๐ฟ๐ฒ ๐ฑ๐ฒ๐๐ฒ๐ฟ๐บ๐ถ๐ป๐ถ๐๐๐ถ๐ฐ. ๐๐ด๐ฒ๐ป๐๐ ๐ฎ๐ฟ๐ฒ ๐ฎ๐ฑ๐ฎ๐ฝ๐๐ถ๐๐ฒ. Thatโs the trade-off: Workflows = stability Agents = flexibility And thatโs why true agents shine in complex GenAI apps where step-by-step logic wonโt cut it. So next time you hear someone say โagent,โ ask: Does it just call an LLM? Or does it reason, adapt, and act? Letโs raise the bar for what we call agentic systems.
AI/ML engineers, listen up! If you want to be in the top 1%, mastering ML frameworks isnโt enough... You need elite software engineering skills. Why? Because AI/ML systems donโt live in Jupyter notebooks. In production, they need: โ Efficient, scalable code โ Optimized cloud architectures โ High-performance systems that can handle real-world demands One of my clients can attest to this first-hand... We rewrote parts of their **RAG ingestion pipeline in Rust** and reduced latency from **minutes to seconds.** That kind of performance boost doesn't come from just knowing how to prompt a model. โฆ it comes from understanding how systems behave **under load**, how to optimize **I/O and memory**, and how to **build real software.** And thatโs where ๐๐ผ๐ฑ๐ฒ๐๐ฟ๐ฎ๐ณ๐๐ฒ๐ฟ๐ comes in... CodeCrafters is one of the best platforms for leveling up real-world engineering skills. (I'm even planning to use it to learn rust) Here's why I say that: - Itโs not just theory โณ Want to understand how Git, Redis, Docker, Kafka, or Shell actually work? โณYouโll build them yourself, piece by piece. - It forces you to write production-grade code. โณ No hand-holding. No shortcuts. โณ Just hands-on, real-world problem-solving that makes you a better engineer. - Itโs perfect for AI/ML engineers who want to stand out. โณ If youโre building LLM apps, RAG systems, or agentic workflows, you need strong SWE fundamentals. โณ This platform helps you build that foundation. Interested? You'll get a ๐ฐ๐ฌ% ๐ฑ๐ถ๐๐ฐ๐ผ๐๐ป๐ if you use my affiliate link. (The link is in the comments)
I promised I'd make some bold moves this year. So, here goes nothing... On April 7th, I'll be giving my ๐๐๐ฅ๐ฆ๐ง ๐ถ๐ป-๐ฝ๐ฒ๐ฟ๐๐ผ๐ป ๐๐ฎ๐น๐ธ at the QCon Software Development Conferences in London. Talk about biting the bullet! As an introvert (with a sprinkle of social anxiety thrown in), I must admit... I'm actually quite excited. ... and a little nervous - in the best way possible. The opportunity to share insights with an incredible community of engineers and data scientists doesn't come around often. Thus, I'll be sure to give it my all. In my talk, Iโll be diving into the data dimension of designing LLM and RAG applications. Namely, we will cover: - The data flow and features pipeline that powers LLM and RAG systems - LLMOps best practices for ensuring data is shared, versioned, processed, and analyzed for LLM training and inference. - Concrete use cases like building an LLM Twin and a Second Brain AI Assistant. The goal: To give you a clear, practical framework for architecting the data layer of LLM systems that scales with accuracy, reliability, and efficiency. If you'd like to attend, I'll leave a registration link in the comments (with ยฃ60 off the price code) See you there ๐
If you're an aspiring AI engineer, listen up... There are 2 basic skills you must master before touching any ML model: ๐ญ. ๐ฃ๐ฟ๐ผ๐ด๐ฟ๐ฎ๐บ๐บ๐ถ๐ป๐ด (Python is a good start) ๐ฎ. ๐๐น๐ผ๐๐ฑ ๐ฒ๐ป๐ด๐ถ๐ป๐ฒ๐ฒ๐ฟ๐ถ๐ป๐ด Without these skills, you will not even be able to deploy or scale your AI/ML models effectively in production. Fortunately, ๐ง Eric Riddoch noticed this problem and filled the need - Enter ๐๐น๐ผ๐๐ฑ ๐๐ป๐ด๐ถ๐ป๐ฒ๐ฒ๐ฟ๐ถ๐ป๐ด ๐ณ๐ผ๐ฟ ๐ฃ๐๐๐ต๐ผ๐ป ๐๐ฒ๐๐ฒ๐น๐ผ๐ฝ๐ฒ๐ฟ๐ live course. Iโve had the pleasure of knowing Eric for almost a year, and I can confidently say heโs a brilliant cloud, DevOps, and MLOps engineer. Heโs also an incredible teacher with a knack for making complex topics approachable and fun to learn. Hereโs what youโll learn in the course: - Enterprise-level AWS account management - Fundamentals of cloud engineering - Designing cloud-native RESTful APIs that scale from 4 to 4 million requests/day - Writing, testing, locally mocking, and deploying code using AWS SDK and OpenAI - Advanced observability and monitoring techniques: Logs, metrics, traces, and alerts If you're concerned that the course is not with your desired cloud vendor, don't worry... The principles learned during the course can easily be transferred to other platforms (e.g., GCP and Azure) I strongly recommend this course to anyone looking to lay the foundation for a career in AI, ML, and MLOps engineering. The next cohort starts on March 31st and runs until May 16th! Using the code DECODINGML will get you 10% off your registration. And if that's not enough, Eric offers a scholarship program that can significantly reduce the price - depending on your use case. Check out the link in the comments to enroll.
If you think you โknowโ prompt engineering... Think again. Iโve been following Nir Diamant for a while now - his GitHub repos and Substack have become go-to resources for AI practitioners. He has a rare gift: The ability to break down complex GenAI topics like heโs teaching a 7-year-old (without dumbing anything down). ... And now heโs done it again with a new eBook: ๐ฃ๐ฟ๐ผ๐บ๐ฝ๐ ๐๐ป๐ด๐ถ๐ป๐ฒ๐ฒ๐ฟ๐ถ๐ป๐ด ๐ณ๐ฟ๐ผ๐บ ๐ญ๐ฒ๐ฟ๐ผ ๐๐ผ ๐๐ฒ๐ฟ๐ผ โ ๐ ๐ฎ๐๐๐ฒ๐ฟ ๐๐ต๐ฒ ๐๐ฟ๐ ๐ผ๐ณ ๐๐ ๐๐ป๐๐ฒ๐ฟ๐ฎ๐ฐ๐๐ถ๐ผ๐ป This isnโt just another โuse more bullet points in your promptโ kind of guide. Itโs a practical deep dive with: โ Code examples โ Real-world exercises โ Clear explanations of common mistakes โ And the subtle mechanics behind great AI interaction One reader put it best: โ๐๐ฐ๐ธ... ๐ ๐ต๐ฉ๐ฐ๐ถ๐จ๐ฉ๐ต ๐ ๐ฌ๐ฏ๐ฆ๐ธ ๐ฑ๐ณ๐ฐ๐ฎ๐ฑ๐ต ๐ฆ๐ฏ๐จ๐ช๐ฏ๐ฆ๐ฆ๐ณ๐ช๐ฏ๐จ ๐ฃ๐ฆ๐ง๐ฐ๐ณ๐ฆ ๐ณ๐ฆ๐ข๐ฅ๐ช๐ฏ๐จ ๐ต๐ฉ๐ช๐ด ๐ฃ๐ฐ๐ฐ๐ฌ, ๐ฃ๐ถ๐ต ๐ ๐ธ๐ข๐ด ๐ญ๐ช๐จ๐ฉ๐ต๐บ๐ฆ๐ข๐ณ๐ด ๐ข๐ธ๐ข๐บ ๐ง๐ณ๐ฐ๐ฎ ๐ฆ๐ท๐ฆ๐ฏ ๐ฃ๐ฆ๐ช๐ฏ๐จ ๐ข๐ธ๐ข๐ณ๐ฆ ๐ฐ๐ง ๐ต๐ฉ๐ฆ ๐ฏ๐ถ๐ข๐ฏ๐ค๐ฆ ๐ข๐ฏ๐ฅ ๐ค๐ฐ๐ฎ๐ฑ๐ญ๐ฆ๐น๐ช๐ต๐ช๐ฆ๐ด... ๐๐ฉ๐ช๐ด ๐ช๐ด ๐ต๐ฉ๐ฆ ๐ฃ๐ฐ๐ฐ๐ฌ ๐ต๐ฐ ๐ฑ๐ช๐ค๐ฌ ๐ถ๐ฑ ๐ช๐ง ๐บ๐ฐ๐ถ ๐ธ๐ข๐ฏ๐ต ๐ต๐ฐ ๐จ๐ฆ๐ต ๐ฑ๐ข๐ด๐ต ๐ซ๐ถ๐ด๐ต ๐ฌ๐ฏ๐ฐ๐ธ๐ช๐ฏ๐จ ๐ธ๐ฉ๐ข๐ต ๐ฑ๐ณ๐ฐ๐ฎ๐ฑ๐ต ๐ฆ๐ฏ๐จ๐ช๐ฏ๐ฆ๐ฆ๐ณ๐ช๐ฏ๐จ ๐ช๐ด, ๐ข๐ฏ๐ฅ ๐ด๐ต๐ข๐ณ๐ต ๐ฎ๐ข๐ด๐ต๐ฆ๐ณ๐ช๐ฏ๐จ ๐ต๐ฉ๐ฆ ๐ค๐ณ๐ข๐ง๐ต.โ I couldnโt agree more. If youโre building LLM apps, agentic systems, or even just playing with ChatGPT prompts, check this eBook out. Youโll walk away with a new toolkit that goes far beyond theory. Want a copy? The link is in the comments.
Here's something most people overlook: The RAG feature pipeline is the most important part of the entire AI stack. Not the LLM. Not the prompt. Not even the fancy agent framework. Why? Because if your retrieval sucks, your generation will too - no matter how good your model is. Let me walk you through how we architected our RAG feature pipeline for the ๐ฆ๐ฒ๐ฐ๐ผ๐ป๐ฑ ๐๐ฟ๐ฎ๐ถ๐ป ๐๐ ๐๐๐๐ถ๐๐๐ฎ๐ป๐: The pipeline ingests raw documents from MongoDB, where both Notion and crawled content are stored in a single standardized collection. We donโt care about the source - just that the data is clean and usable. The output? โ Chunked + embedded documents โ Stored in a dedicated vector store โ Indexed and ready for semantic search Where does this pipeline fit? Itโs an offline batch pipeline, decoupled from the live user experience. At query time, the pipeline does not run. All processing is done beforehand, so retrieval is fast, stable, and cost-efficient. Meanwhile, the online pipeline (our agentic RAG module) fetches the chunks, reasons over them, and generates the answer. The architecture is made up of 7 key components: ๐ญ/ ๐๐ฎ๐๐ฎ ๐๐ ๐๐ฟ๐ฎ๐ฐ๐๐ถ๐ผ๐ป Pulls all raw documents from MongoDB, regardless of source. ๐ฎ/ ๐๐ผ๐ฐ๐๐บ๐ฒ๐ป๐ ๐๐ถ๐น๐๐ฒ๐ฟ๐ถ๐ป๐ด Applies quality scores to drop noisy or low-value docs. ๐ฏ/ ๐๐ต๐๐ป๐ธ๐ถ๐ป๐ด Splits documents into manageable segments for vectorization. ๐ฐ/ ๐ฃ๐ผ๐๐-๐ฃ๐ฟ๐ผ๐ฐ๐ฒ๐๐๐ถ๐ป๐ด Applies one of two retrieval strategies (chosen from our YAML configs): Parent Retrieval โ Links each chunk to its full doc Contextual Retrieval โ Adds summaries to enrich semantic relevance ๐ฑ/ ๐๐บ๐ฏ๐ฒ๐ฑ๐ฑ๐ถ๐ป๐ด All chunks (for both strategies) are vectorized using a configurable embedding model (OpenAI or Hugging Face) ๐ฒ/ ๐๐ป๐ฑ๐ฒ๐ ๐ถ๐ป๐ด Embeddings are stored and indexed in MongoDB for fast lookups. ๐ณ/ ๐๐ถ๐ป๐ฎ๐น ๐ข๐๐๐ฝ๐๐ A structured, searchable knowledge baseโready for RAG-based generation. The entire pipeline is managed by ZenML, which is: โ Reproducible โ Configurable โ Versioned โ Traceable If youโre serious about building production-grade GenAI systems, this is where you focus. ... because 90% of your generation's quality is determined before the LLM even gets involved. Letโs stop treating pipelines like afterthoughts. They are the product. Want to learn more? Check the link in the comments.
RAG has kept โdyingโ for the past 4 years. But hereโs why that will never happen: All the LLMs (even the most advanced ones) struggle without the right context. It doesnโt matter if your model has 128k+ token windows or cutting-edge fine-tuning... If it doesnโt retrieve the right data or the context is full of noise or formatted incorrectly, it wonโt generate the right answers. Thatโs why retrieval is the hardest part of RAG. Most RAG failures arenโt about generation - they happen before the LLM even sees the data. If the retrieval step is weak, your AI assistant will: - Fetch irrelevant information - Miss critical details - Hallucinate confidently wrong responses But more context isnโt the answer... Better context is. Lesson 5 of the Second Brain AI Assistant course is all about fixing retrieval with a production-ready RAG feature pipeline that. (And it's now live!) Namely, in this lesson, you will learn: - The fundamentals of RAG. - How to design and implement a production-ready RAG pipeline - Implement contextual retrieval (an advanced RAG technique) from scratch. - Implement parent retrieval (another advanced RAG technique) using LangChain. - Extend LangChain to add custom behavior using OOP. - The critical role of chunk size in optimizing retrieval quality - Write a configuration layer to switch between different algorithms and models dynamically - How to manage everything with an MLOps framework (we use ZenML) By the end of this lesson, youโll be equipped with the skills to build a flexible, modular RAG feature pipeline. This pipeline gives access to our AI assistant to our Second Brain and provides reliable context to generate meaningful answers. Sounds interesting? Pick up lesson 5 today. (The link is in the comments) Thank you, Anca Ioana Muscalagiu, for contributing with another fantastic lesson to @Decoding ML !
Over the past 2 years, I've been on a mission... (And it's finally coming together) My goal has been to give as much back to the AI community as possible. I've done this by creating a total of 6 open-source courses designed to help others build real-world, production-grade AI systems. Now, I'm thrilled to announce that I've brought all my FREE courses together in one place. These are not your typical tutorials showing isolated concepts in scattered Notebooks... They're designed for those looking to master production and industry-level AI. It covers everything from: - End-to-end GenAI, LLM, RAG apps - MLOps/LLMOps for AI scalability - AI Systems for real-world applications I've been sure to constantly refine and improve these courses to keep them at the cutting edge. Our latest project, the ๐ฆ๐ฒ๐ฐ๐ผ๐ป๐ฑ ๐๐ฟ๐ฎ๐ถ๐ป ๐๐ ๐๐๐๐ถ๐๐๐ฎ๐ป๐ course, is our crown jewel so far - It's an end-to-end project that takes you through the full process of building a production-level AI assistant. The world of AI is moving incredibly fast. ... I believe that knowledge-sharing is key to pushing the field forward. Iโm currently working on a new open-source course focused on AI agents that I know youโll love. (More on that soon) All in all, I'm incredibly proud of what we've built so far. I'm excited to see how these resources help others in their AI journey. Check it out, share it, and let's keep building the future together. The link is in the comments.
There's a major problem with using general-purpose LLMs to build apps. But not enough people are talking about it... Of course, I'm talking about: โ Skyrocketing API bills โ Vendor lock-in โ Degrading performance we can't do anything about โ Lack of control of your data Thankfully, I've found a solution. In Lesson 4 of the ๐ฆ๐ฒ๐ฐ๐ผ๐ป๐ฑ ๐๐ฟ๐ฎ๐ถ๐ป ๐๐ ๐๐๐๐ถ๐๐๐ฎ๐ป๐ course, we tackle these challenges head-on. Specifically, you'll learn: - How to fine-tune open-source small language models (SLMs) using LoRA and QLoRA to specialize them on your tasks. - Efficient fine-tuning with tools like TRL, Unsloth AI, and Comet - Architecting modular, scalable pipelines with MLOps in mind for production-ready applications. - Deploying your own specialized models on Hugging Face Inference Endpoints as a real-time API endpoint. - Evaluating your models using vLLM to ensure theyโre production-ready. By the end of this lesson, you'll have the tools and knowledge to take back control of your AI systems (without breaking the bank or being tied to a single provider) Ready to get started? The link is in the comments.
AI engineers must understand this to build any successful system: (Especially when working with RAG, LLMs, or agent-based apps) How to architect data pipelines. We all know data is the lifeblood of the systems we build. You can have the most advanced algorithms, but without clean data, you're just spinning your wheels. To understand how data pipelines for LLM apps should look, letโs architect one that powers our Second Brain AI Assistant. Here are the core components: ๐ญ/ ๐๐ฎ๐๐ฎ ๐ฐ๐ผ๐น๐น๐ฒ๐ฐ๐๐ถ๐ผ๐ป Data sources can be many things - Notion, crawled links, custom APIs... you name it. In this case, we'll use Notionโs API to collect personal data, extract all the links, and standardize everything into Markdown format. Why Markdown? Because the internet is full of it. Thus, LLMs are heavily trained using this format and know how to work with it. To replicate steps down the line, we make a snapshot of the raw data into a data lake like S3. ๐ฎ/ ๐๐ง๐ ๐ฑ๐ฎ๐๐ฎ ๐ฝ๐ถ๐ฝ๐ฒ๐น๐ถ๐ป๐ฒ This is where the magic happens. After collecting the data, we perform several transformations: 1. Download the Notion docs from S3 2. Load them in memory into Pydantic objects (content + metadata) 3. Crawl all the links inside the Notion docs (when building the AI assistant, we want to generate answers based on the content of the links as well โ this is where the power of our lists of saved resources kicks in) 4. Compute a quality score per document using a combination of heuristics and LLMs (as LLM calls are expensive, you first want to squeeze the most out of it using heuristics) 5. Store all the documents and the quality score inside a document database (with this, we can compute statistics and filter documents down the line based on our needs) For our setup, weโve used MongoDB: a Swiss knife to store our unstructured documents and, later on, the embeddings for RAG. We can think about it as part of our logical feature store. ๐ฏ/ ๐ ๐๐ข๐ฝ๐ ๐๐ฟ๐ฎ๐บ๐ฒ๐๐ผ๐ฟ๐ธ To keep everything organized and maintainable, we use ZenML This helps us easily manage, version, and deploy our offline data pipelines. ๐ง๐ฎ๐ธ๐ฒ๐ฎ๐๐ฎ๐: Architecting and managing these pipelines is crucial for creating LLM/RAG systems. If you're an AI/ML engineer stepping into the GenAI world, get comfortable building and managing data pipelines that crawl unstructured data and process it for LLMs. (Yes! Along with the 1000+ things you must know, AI Engineers must also dig into data engineering.) Want to see how to do this? Check the link in the comments.
Less than 3 years ago, I graduated with my masterโs in ML. Fast forward to today, Iโm: - Contracting - Consulting - Creating content Globally! So, what helped me accelerate my career? ๐ญ. ๐ ๐๐๐ฎ๐ฟ๐๐ฒ๐ฑ ๐ฒ๐ฎ๐ฟ๐น๐ While still in university, I didnโt wait for the โperfectโ opportunity. By my third year, I had already landed a junior SWE role and was working like crazy. ๐ฎ. ๐ ๐ฝ๐๐ ๐ถ๐ป ๐๐ต๐ฒ ๐๐ถ๐บ๐ฒ - ๐ป๐ผ ๐๐ต๐ผ๐ฟ๐๐ฐ๐๐๐. For years, I spent 10-12 hours a day researching, building, and creating- mostly alone. Monday to Saturday. That helped me transform 3 years of experience into 6. ๐ฏ. ๐ ๐ฎ๐ฐ๐ฐ๐ฒ๐ฝ๐๐ฒ๐ฑ ๐๐ต๐ฒ ๐๐ฟ๐ฎ๐ฑ๐ฒ-๐ผ๐ณ๐ณ๐ I knew that accelerating my career meant sacrifices in other areas. I had less time for social life. ... but this meant more time for deep work and learning There was only one exception: My girlfriend. No matter how much I worked, I made time for her - she kept me sane. ๐ฐ. ๐๐ฒ๐ฎ๐น๐๐ต > ๐๐๐๐๐น๐ฒ โ Workouts: 3-4x per week โ Food: Eat clean, avoid junk โ Sleep: 7-8 hours, no compromise โ Stress: Keep it under control (no toxic jobs, no endless pressure) ๐ฑ. ๐ก๐ผ ๐ฎ๐น๐ฐ๐ผ๐ต๐ผ๐น. ๐ก๐ผ ๐๐บ๐ผ๐ธ๐ถ๐ป๐ด. Clarity is everything. You canโt afford to be foggy-minded when making big moves. (I still enjoy exceptions here and there, but nothing that slows me down.) ๐ฒ. ๐ ๐ณ๐ผ๐น๐น๐ผ๐๐ฒ๐ฑ ๐บ๐ ๐ถ๐ป๐๐๐ถ๐๐ถ๐ผ๐ป - ๐ฎ๐น๐๐ฎ๐๐. I ignored the โsafeโ paths. Regardless of what โmore experiencedโ people told me, I always made hard decisions. Most people project their fears onto you, but they donโt live your life. ๐ณ. ๐ ๐๐ผ๐ผ๐ธ ๐๐ต๐ฒ ๐บ๐ถ๐ป๐ฑ๐๐ฒ๐ ๐ผ๐ณ ๐ฎ๐ป ๐ฒ๐ ๐ฝ๐น๐ผ๐ฟ๐ฒ๐ฟ If you follow paved paths, youโre in high competition. If you carve your own path, you move faster. I never aimed to be top 1% in existing tracks. I created my own track instead. ๐ด. ๐ ๐ป๐ฒ๐๐ฒ๐ฟ ๐๐ฒ๐๐๐น๐ฒ๐ฑ When my managers said: - โThe job market is tough.โ - โThis is what others get paid.โ - โYouโre asking for too much.โ - โHave patience.โ I didnโt listen. I doubled down and kept pushing forward. Itโs exhausting at times, but curiosity > comfort. And thatโs what kept me going. So, if youโre early in your career, hereโs my advice: โ Be intentional about how you spend your time. โ Build skills relentlessly - donโt wait for permission. โ Challenge conventional wisdom - not all โadviceโ is useful. โ Prioritize health - burnout kills long-term success. โ Bet on yourself - no one else will do it for you. Happy Sunday!
Many ML projects fail to transition from POC to production-ready. Here's one simple reason why: The initial focus was never on scalability or production constraints. Fraud detection presents a perfect use case for building production-first ML systems. It combines the need for real-time and batch processing, low-latency predictions, and high-accuracy models. Here's a look into what that system could look like: ๐๐ฎ๐๐ฎ ๐๐ผ๐๐ฟ๐ฐ๐ฒ๐ In fraud detection, you deal with real-time transactions, streaming data and historical records. Real-time and streaming data require you to instantly compute features, while historical records help track user profiles and spending patterns. In our architecture, real-time features are computed through HTTP requests, streaming data will flow through Kafka topics, with historical data stored in a data warehouse for batch processing. ๐๐ฒ๐ฎ๐๐๐ฟ๐ฒ ๐ฝ๐น๐ฎ๐๐ณ๐ผ๐ฟ๐บ At the heart of the system is the feature platform, like Tecton ... This centralizes all the features. More specifically, it allows us to manage features in an offline store (for training - high throughput) and online store (for serving - low latency). Using the same feature engineering logic during training and inference avoids training-serving skew. ๐๐ฒ๐ฎ๐๐๐ฟ๐ฒ ๐ฝ๐ถ๐ฝ๐ฒ๐น๐ถ๐ป๐ฒ๐ These convert raw data into meaningful features. By centralizing your features into a feature store like Tecton, you can leverage their feature views to define features once and reuse them across models and pipelines. A feature view is defined as a data source(s) + a function that maps raw data into features. Next, using a process known as materialization, you sync the raw data sources with the online/offline stores while applying the transformations. ๐ง๐ฟ๐ฎ๐ถ๐ป๐ถ๐ป๐ด ๐ฝ๐ถ๐ฝ๐ฒ๐น๐ถ๐ป๐ฒ๐ It ingests features and labels to train models (stored in a model registry). Leveraging the feature store, you can easily apply time-traveling strategies to version your dataset. ๐๐ป๐ณ๐ฒ๐ฟ๐ฒ๐ป๐ฐ๐ฒ ๐ฝ๐ถ๐ฝ๐ฒ๐น๐ถ๐ป๐ฒ๐ It takes new transaction data, enriches it with features from Tecton's feature platform, and applies the trained model to generate predictions. Online stores are crucial at serving time. They provide low-latency access to up-to-date features. When a transaction occurs... The pipeline quickly retrieves pre-computed features, combines them with real-time features, and computes predictions. ๐ข๐ฏ๐๐ฒ๐ฟ๐๐ฎ๐ฏ๐ถ๐น๐ถ๐๐ ๐ฝ๐ถ๐ฝ๐ฒ๐น๐ถ๐ป๐ฒ Lastly, an observability pipeline is essential for monitoring the system's health and detecting drifts. The final touch is an alarm system that sends emails or SMS or denies transactions if fraud is detected. Want to dive deeper into building such systems? Check out the link in the comments.
The course is finally complete! You can learn for free how to build your Second Brain AI assistant with LLMs, agents, RAG, fine-tuning, LLMOps, and AI systems techniques. Thanks for sharing Andrew Davidson is the crown jewel of Decoding ML
Andrew Davidson
Looking for weekend fun? Build a second brain AI assistant with Paul Iusztin's course! https://lnkd.in/eUztq49A The 6th and final lesson just dropped.
RAG is evolving. Standard retrieval is no longer enough... If youโre building LLM-powered applications, you need something more powerful: ๐๐ด๐ฒ๐ป๐๐ถ๐ฐ ๐ฅ๐๐. Thatโs exactly what weโre tackling in Lesson 6 of the Second Brain AI Assistant course. ... and it's now LIVE! Most RAG systems passively retrieve context and hope the LLM generates the right response. But this approach is limited. โ What if your assistant needs to reason about multiple sources? โ What if retrieval alone doesnโt fully align with the userโs intent? โ What if the retrieved context isn't enough, and the system needs to iterate? Agentic RAG bridges that gap. Instead of just fetching documents, agents dynamically plan, retrieve, and refine their responses, making AI assistants: โ Smarter โ More interactive โ More accurate And in Lesson 6, weโre building one from scratch... Specifically, youโll learn: โ How to build an Agentic RAG module that goes beyond simple retrieval โ Integrating retrieval with AI agents for dynamic, multi-step reasoning โ Adding LLMOps, such as prompt monitoring to optimize retrieval workflows for cost, latency, and throughput โ Evaluating long and complex LLM chains to ensure reliability (moderation, hallucination, response performance) โ Scaling retrieval architectures to handle real-world AI assistant demands By the end of this lesson, youโll understand what it takes to build stable, efficient, and intelligent RAG-powered assistants. Huge thanks to Anca Ioana Muscalagiu for contributing another fantastic lesson to Decoding ML! The link for Lesson 6 is in the comments :)
If you're building GenAI apps, you need to understand one thing: LLMs donโt โthinkโ like humans. They generate responses based on probabilities. ... and how you craft your prompts can make or break your application. Thatโs why Iโve been diving deep into "๐ฃ๐ฟ๐ผ๐บ๐ฝ๐ ๐๐ป๐ด๐ถ๐ป๐ฒ๐ฒ๐ฟ๐ถ๐ป๐ด ๐ณ๐ผ๐ฟ ๐๐๐ ๐." I must say, itโs one of the best resources out there for practitioners building with LLMs. โ Itโs written with character โ It focuses on how to actually build GenAI products (practical theory) โ Itโs packed with fundamentals that every AI/ML engineer should know I especially loved how they teach prompt engineering beyond basic techniques you can find everywhere. They teach you how to manage and process the context injected into the prompts, which is required to scale LLM apps beyond PoCs. Iโm still working my way through the book, but after finishing the first 7 chapters, I can confidently say it's a must-read. Specifically for: โ Anyone building GenAI applications โ Engineers looking to optimize their prompt engineering skills โ Experienced AI practitioners looking to build modular LLM apps. I'd also like to say a huge thanks to John Berryman for sending me a copy - I'm absolutely loving it. If youโre serious about building with LLMs, this book is worth your time. (I'll leave a link to it in the comments)
Iโve just started working at a new AI start-up (still in shadow mode) on building cutting-edge products using GenAI, LLMs, RAG, and Agents. All the good stuff. This will help me improve my content with even more real-world battle-tested content and stories. So excited to start providing more in-depth insights from this move ๐ค
Cursor + Claude 3.7 keeps disappointing me. I am not a vibe coder. I strongly articulate what should be implemented, how it should be implemented, and the issues within the codebase. But when things get tricky and require extensive understanding of the code, infrastructure and data, it just doesnโt work. I lost 2 hours of my life talking to Claude (Extended) when a GitHub Issues thread solved it in 5 minutes. My take is that LLMs are great for quickly ramping up your project (instead of digging into the documentation), but you still have to: - design - plan - know the best engineering practices - be open to use Google What is your experience?
Content Inspiration, AI, scheduling, automation, analytics, CRM.
Get all of that and more in Taplio.
Try Taplio for free
Austin Belcak
@abelcak
1m
Followers
Matt Gray
@mattgray1
1m
Followers
Richard Moore
@richardjamesmoore
103k
Followers
Shlomo Genchin
@shlomogenchin
49k
Followers
Sam G. Winsbury
@sam-g-winsbury
45k
Followers
Daniel Murray
@daniel-murray-marketing
147k
Followers
Sabeeka Ashraf
@sabeekaashraf
20k
Followers
Ash Rathod
@ashrathod
73k
Followers
Izzy Prior
@izzyprior
81k
Followers
Vaibhav Sisinty โ๏ธ
@vaibhavsisinty
446k
Followers
Wes Kao
@weskao
107k
Followers
Luke Matthews
@lukematthws
186k
Followers
Justin Welsh
@justinwelsh
1m
Followers
Amelia Sordell ๐ฅ
@ameliasordell
228k
Followers
Tibo Louis-Lucas
@thibaultll
6k
Followers
Andy Mewborn
@amewborn
212k
Followers