Logo Taplio

Taplio

Paul Iusztin's Linkedin Analytics

Get the Linkedin stats of Paul Iusztin and many LinkedIn Influencers by Taplio.

Want detailed analytics of your Linkedin Account? Try Taplio for free.

Paul Iusztin

open on linkedin

I am a senior machine learning engineer and contractor with ๐Ÿฒ+ ๐˜†๐—ฒ๐—ฎ๐—ฟ๐˜€ ๐—ผ๐—ณ ๐—ฒ๐˜…๐—ฝ๐—ฒ๐—ฟ๐—ถ๐—ฒ๐—ป๐—ฐ๐—ฒ. I design and implement modular, scalable, and production-ready ML systems for startups worldwide. My central mission is to build data-intensive AI/ML products that serve the world. Since training my first neural network in 2017, I have 2 passions that fuel my mission: โ†’ Designing and implementing production AI/ML systems using MLOps best practices. โ†’ Teaching people about the process. . I currently develop production-ready Deep Learning products at Metaphysic, a leading GenAI platform. In the past, I built Computer Vision and MLOps solutions for CoreAI, Everseen, and Continental. Also, I am the Founder of Decoding ML, a channel for battle-tested content on learning how to design, code, and deploy production-grade ML and MLOps systems. I am writing articles and posts each week on: - ๐˜“๐˜ช๐˜ฏ๐˜ฌ๐˜ฆ๐˜ฅ๐˜๐˜ฏ: 29k+ followers - ๐˜”๐˜ฆ๐˜ฅ๐˜ช๐˜ถ๐˜ฎ: 2.5k+ followers ~ ๐Ÿ”— https://medium.com/@pauliusztin - ๐˜š๐˜ถ๐˜ฃ๐˜ด๐˜ต๐˜ข๐˜ค๐˜ฌ (๐˜ฏ๐˜ฆ๐˜ธ๐˜ด๐˜ญ๐˜ฆ๐˜ต๐˜ต๐˜ฆ๐˜ณ): 6k+ followers ~ ๐Ÿ”— https://decodingml.substack.com/ . If you want to learn how to build an end-to-end production-ready LLM & RAG system using MLOps best practices, you can take Decoding MLโ€™s self-guided free course: โ†’ ๐˜“๐˜“๐˜” ๐˜›๐˜ธ๐˜ช๐˜ฏ ๐˜Š๐˜ฐ๐˜ถ๐˜ณ๐˜ด๐˜ฆ: ๐˜‰๐˜ถ๐˜ช๐˜ญ๐˜ฅ๐˜ช๐˜ฏ๐˜จ ๐˜ ๐˜ฐ๐˜ถ๐˜ณ ๐˜—๐˜ณ๐˜ฐ๐˜ฅ๐˜ถ๐˜ค๐˜ต๐˜ช๐˜ฐ๐˜ฏ-๐˜™๐˜ฆ๐˜ข๐˜ฅ๐˜บ ๐˜ˆ๐˜ ๐˜™๐˜ฆ๐˜ฑ๐˜ญ๐˜ช๐˜ค๐˜ข ~ ๐Ÿ”— https://github.com/decodingml/llm-twin-course . ๐Ÿ’ฌ If you need machine learning solutions for your business, letโ€™s discuss! ๐ŸŒŽ Only open to full remote positions as a contractor. . Contact: ๐Ÿ“ฑ Phone: +40 732 509 516 โœ‰๏ธ Email: p.b.iusztin@gmail.com ๐Ÿ’ป Decoding ML: https://linktr.ee/decodingml ๐Ÿ•ต๐Ÿปโ€โ™‚๏ธ Personal site & Socials: https://www.pauliusztin.me/

Check out Paul Iusztin's verified LinkedIn stats (last 30 days)

Followers
54,582
Posts
20
Engagements
3,656
Likes
3,029

What is Paul talking about?

nicheai
  • frequency
  • engagement

Who is engaging with Paul

Banias Baabe profile picture
Samanth Koduru profile picture
Jason Quick profile picture
Piotr Malicki profile picture
Alexandru Razvant profile picture
Saba Hesaraki profile picture
Alex Vesa profile picture
Tales Marra profile picture
Ashok Kumar Murugesan profile picture
Daniel Svonava profile picture
Silvestre Pรฉrez O. profile picture
Priyanka Kamath profile picture
Hendy Fergus Atheri Hura profile picture
ORLANDO MONTENEGRO REYES profile picture
Sujoy Chabri profile picture
Maria Vechtomova profile picture
Achutha Subhash profile picture
Jai Singh profile picture
Ankit Srivastava profile picture
Joshua Rio-Ross profile picture
Ajeet Singh  profile picture
Prashant K Dhingra profile picture
Mattheus Chediak profile picture
Ray Florez profile picture
Anirban Datta profile picture
Rutuja Surve profile picture
Kaan Kabalak profile picture
Abby Morgan profile picture
Raphaรซl Hoogvliets profile picture
zahra sadeghi profile picture
Li Yin profile picture
Ali Alemi Matin Pour profile picture
Bharat Raghunathan profile picture
bรนi xuรขn hรนng profile picture
Victor Robles profile picture
Shamal De Silva profile picture
Leon Jose profile picture
Koyelia Ghosh Roy profile picture
Aniruddh Durga Naga Gilakamsetty profile picture
Pascal Biese profile picture
Diego Marinho profile picture
Diego Miranda profile picture
Thiago Costa profile picture
Bruna Lemberck profile picture

Paul Iusztin's Best Posts (last 30 days)

Use Taplio to search all-time best posts


Hereโ€™s the problem with most AI books: They teach the model, not the system. Which is fine... until you try to deploy that model in production. Thatโ€™s where everything breaks: - Your RAG pipeline is duct-taped together - Your eval framework is an afterthought - Your prompts arenโ€™t versioned - Your architecture canโ€™t scale Thatโ€™s why Maxime and I wrote the LLM Engineerโ€™s Handbook... We wanted to create a practical guide for AI engineers who build real world AI applications. This isnโ€™t just another guide... It's a practical road map for designing and deploying real-world LLM systems. In the book, we cover: โ†’ Efficient fine-tuning workflows โ†’ RAG architectures โ†’ Evaluation pipelines with LLM-as-judge โ†’ Scaling strategies for serving + infra โ†’ MLOps + LLMOps patterns baked in Whether youโ€™re building your first assistant or scaling your 10th RAG app... This book gives you the mental models and engineering scaffolding to do it right. ๐Ÿ”— Here's the link to get your copy: https://lnkd.in/dVgFJtzF


151

Everyone chunks documents for retrieval. But what if thatโ€™s the wrong unit? Let me explain.. In standard RAG, we embed small text chunks and pass those into the LLM as context. Itโ€™s simple, but flawed. Why? Because small chunks are great for retrieval precision, but terrible for generation context. Thatโ€™s where Parent Retrieval comes in. (aka small-to-big retrieval) Hereโ€™s how it works: โ†’ You split your documents into small chunks โ†’ You embed and retrieve using those small chunks โ†’ But you donโ€™t pass the chunk to the LLM... โ†’ You pass the parent document that the chunk came from The result? โ†’ Precise semantic retrieval (thanks to small, clean embeddings that encode a single entity) โ†’ Rich generation context (because the LLM sees the broader section) โ†’ Fewer hallucinations โ†’ Less tuning needed around chunk size and top-k Itโ€™s one of the few advanced RAG techniques that work in production. No fancy agents. No latency bombs. No retraining. We break it all down (with diagrams and code examples) in ๐—Ÿ๐—ฒ๐˜€๐˜€๐—ผ๐—ป ๐Ÿฑ ๐—ผ๐—ณ ๐˜๐—ต๐—ฒ ๐—ฆ๐—ฒ๐—ฐ๐—ผ๐—ป๐—ฑ ๐—•๐—ฟ๐—ฎ๐—ถ๐—ป ๐—”๐—œ ๐—”๐˜€๐˜€๐—ถ๐˜€๐˜๐—ฎ๐—ป๐˜ ๐—ฐ๐—ผ๐˜‚๐—ฟ๐˜€๐—ฒ. ๐Ÿ”— Link to the full lesson in the comments.


162

The #1 mistake in building LLM agents? Thinking the project ends at reasoning. Here's when it actually ends: When your agent can talk to the world securely, reliably, and in real time. And thatโ€™s what ๐—Ÿ๐—ฒ๐˜€๐˜€๐—ผ๐—ป ๐Ÿฐ ๐—ผ๐—ณ ๐˜๐—ต๐—ฒ ๐—ฃ๐—ต๐—ถ๐—น๐—ผ๐—”๐—ด๐—ฒ๐—ป๐˜๐˜€ ๐—ฐ๐—ผ๐˜‚๐—ฟ๐˜€๐—ฒ is all about. Up to this point, we focused on making our agents think: โ†’ Philosophical worldviews โ†’ Context-aware reasoning โ†’ Memory-backed conversations But intelligence alone isnโ€™t enough. To be useful, agents need a voice. To be deployable, they need an interface. To be real, they need to exist as APIs. This lesson is the bridge from the local prototype to the live system. Hereโ€™s what youโ€™ll learn: โ†’ How to deploy your agent as a REST API using FastAPI โ†’ How to stream responses token-by-token with WebSockets โ†’ How to wire up a clean backendโ€“frontend architecture using FastAPI (web server) + Phaser (game interface) โ†’ How to think about agent interfaces in real-world products (not just demos) In short: ๐—ง๐—ต๐—ถ๐˜€ ๐—ถ๐˜€ ๐—ต๐—ผ๐˜„ ๐˜†๐—ผ๐˜‚ ๐˜€๐—ต๐—ถ๐—ฝ ๐—ฎ๐—ป ๐—ฎ๐—ด๐—ฒ๐—ป๐˜ ๐˜„๐—ต๐—ผ ๐—ฟ๐—ฒ๐—ฎ๐˜€๐—ผ๐—ป๐˜€ ๐—”๐—ก๐—— ๐—ฟ๐—ฒ๐˜€๐—ฝ๐—ผ๐—ป๐—ฑ๐˜€ ๐—ถ๐—ป ๐—ฝ๐—ฟ๐—ผ๐—ฑ๐˜‚๐—ฐ๐˜๐—ถ๐—ผ๐—ป. Shoutout to Anca-Ioana Martin for helping shape this lesson and write the deep-dive article. And of course... big thanks to my co-creator Miguel Otero Pedrido for the ongoing collab. ๐Ÿ”— Link to Lesson 4 in the comments.


160

I need your opinion ๐Ÿซต If you've used the LLM Engineer's Handbook to bring your AI project idea to life, Maxime Labonne and I would love to hear about it! ๐˜Ž๐˜ช๐˜ท๐˜ช๐˜ฏ๐˜จ ๐˜บ๐˜ฐ๐˜ถ ๐˜ต๐˜ฉ๐˜ฆ ๐˜ค๐˜ฉ๐˜ข๐˜ฏ๐˜ค๐˜ฆ ๐˜ต๐˜ฐ ๐˜ฆ๐˜ข๐˜ณ๐˜ฏ $500. Our bestseller, LLM Engineer's Handbook, has helped thousands build and deploy their own LLM and RAG systems from scratch. ๐—™๐—ถ๐—ฟ๐˜€๐˜, as a writer and educator, I would love to see how Maxime's and my book helped you in your AI Engineering journey. As we've written this book out of passion, that will mean the world to us. ๐—ฆ๐—ฒ๐—ฐ๐—ผ๐—ป๐—ฑ๐—น๐˜†, Packt is organizing a contest where you share on social media what you've built and how the book helped you navigate the spaghetti world of AI. The first winner will receive $500. The next five spots will earn a free Packt subscription, giving them access to all Packt's books. ๐˜ ๐˜ฐ๐˜ถ ๐˜ค๐˜ข๐˜ฏ ๐˜ด๐˜ถ๐˜ฃ๐˜ฎ๐˜ช๐˜ต ๐˜ต๐˜ฉ๐˜ฆ ๐˜ฑ๐˜ฐ๐˜ด๐˜ต ๐˜ถ๐˜ฏ๐˜ต๐˜ช๐˜ญ ๐˜”๐˜ข๐˜บ 25! ๐Ÿ”— Find more details here: https://lnkd.in/dExZAc5i Looking forward to seeing what you've built!


71

In 2024, everyone was chasing AI hype. In 2025, people are finally beginning to ask the most important question: ๐—–๐—ฎ๐—ป ๐˜†๐—ผ๐˜‚ ๐—ฏ๐˜‚๐—ถ๐—น๐—ฑ ๐˜€๐—ผ๐—บ๐—ฒ๐˜๐—ต๐—ถ๐—ป๐—ด ๐—ฟ๐—ฒ๐—ฎ๐—น? If your answer to that is "no", don't worry, I've got you... My friend, @shawtalebi, has put together one of the most practical programs to teach you how to build actual AI projects. It's called ๐—ง๐—ต๐—ฒ ๐—”๐—œ ๐—•๐˜‚๐—ถ๐—น๐—ฑ๐—ฒ๐—ฟ๐˜€ ๐—•๐—ผ๐—ผ๐˜๐—ฐ๐—ฎ๐—บ๐—ฝ. Over the course of 6 weeks, you'll go deep on: โ†’ LLMs and prompt engineering โ†’ RAG and embeddings โ†’ Fine-tuning and evaluation โ†’ Tool use and agent flows โ†’ AI project management frameworks And you'll also ship real projects, such as: โ†’ A RAG chatbot over blog content โ†’ A local document QA assistant โ†’ An AI-powered job scraper and dashboard โ†’ A fine-tuned text classifier โ†’ A structured survey summarize All with expert guidance, peer feedback, and clean, reusable code you can take into your next product or freelance project. What I love most about this program? Itโ€™s not tool-first. Itโ€™s not hype-first. Itโ€™s build-first. Youโ€™ll walk away with: - A repeatable system for shipping AI MVPs - The confidence to turn vague ideas into working prototypes - The clarity to ignore noise and focus on what matters Want it? The link is in the comments. P.S. Use code PAUL100 for $100 off - the next cohort kicks off June 6th.

  • graphical user interface, text, application, email

69

One of the best talks I had on AI, LLMs, RAG and how to build and ship real-world products. 100% recommend Nicolay Christopher Gerold podcast. One of the best out there โ†“

Profile picture of Nicolay Christopher Gerold

Nicolay Christopher Gerold


"I see LangChain and similar tools as low-code solutions. Good for prototyping, but I'd throw them away for any serious project" Today on How AI Is Built, I have the chance to talk to Paul Iusztin, who's spent 8 years in AI - from writing CUDA kernels in C++ to building modern LLM applications at Decoding ML. His philosophy is refreshingly simple: stop overthinking, start building, and let patterns emerge through use. He uses LangChain and similar tools for quick prototyping - maybe an hour or two to validate an idea - then throws them away completely. "They're low-code tools," he says. "Not good frameworks to build on top of." Yes, it's more work upfront. But when you need to debug or scale, you'll thank yourself. In the podcast, we also cover: - Why fine-tuning is almost always the wrong choice (shoutout to Hamel Husain) - The "just-in-time" learning approach for staying sane in AI - Building writing assistants that actually preserve your voice - Why robots, not chatbots, are the real endgame Full episode below. โ™ป๏ธ Pay it forward by sharing โ™ป๏ธ


37

Super excited to see what youโ€™ve built! ๐Ÿค˜


Our bestseller *LLM Engineerโ€™s Handbook* has helped thousands build and deploy their own large language models from scratch โ€” now, itโ€™s your turn to show the world what youโ€™ve built! ๐ŸŽฌ Share a short video demonstrating the #LLM you designed using the 'LLM Engineerโ€™s Handbook'. Tell us about your process, what you built, and how the book helped you get there. ๐Ÿ† Whatโ€™s in it for you? ๐Ÿฅ‡ First prize: $500 ๐Ÿ… First five Runner-ups: A free Packt subscription to keep learning and building *Create a post (video/still) telling us:* 1. What you built 2. How LLM Engineerโ€™s Handbook helped 3. Any exciting breakthroughs or challenges you overcame โœ… *To participate:* Post about what you have built on LinkedIn, Twitter, Youtube (any other and as many channels as possible) (Tip: brownie points if youโ€™re posting a video) 1. Tag Packt in your post 2. Tag the authors - Paul Iusztin Maxime Labonne 3. Use #BuildwithLLMEnggHB 4. Fill so we know youโ€™ve entered - https://packt.link/MKbh0 ๐Ÿ“… Last date to submit: May 25 ๐Ÿ“ฃ Winners announced: May 27 *Remember:* ๐Ÿ…Our Expert Panel will select the winner and the runner-ups. ๐Ÿ†The best projects will be featured by Packt. ๐Ÿ“š Havenโ€™t read the book yet? Grab your copy here: https://packt.link/dZAxf Letโ€™s build #LLMs, inspire others, and celebrate innovation together. ๐Ÿ’ก๐Ÿ”ง #BuildWithLLMEnggHB


22

LangChain and Llama index are great entry points for building LLM apps. But itโ€™s a huge red flag you're using them in production. Why? Because most LLM frameworks are just like low-code tools. โ†’ Great for exploring concepts โ†’ Fast to build a demo โ†’ Terrible when you need control The moment your system demands: โ†’ Custom memory flows โ†’ Non-trivial evaluation pipelines โ†’ Agent logic across multiple tools โ†’ Database-level optimizations You hit a wall. And no amount of chaining can fix it. My advice? If your app depends on data ingestion, embedding, retrieval, and synthesis, just build those pieces from scratch. Itโ€™s the only way to - โ†’ Know whatโ€™s actually happening under the hood โ†’ Tune for latency and scale โ†’ Own your system end-to-end We unpacked this in depth during the latest DataFramed podcast DataCamp. Maxime and I talked about what it actually takes to ship real-world AI systems. Want to check it out? The link is in the comments.


65

You donโ€™t become an AI engineer by tweaking someone elseโ€™s notebook. You do it by building real systems, end-to-end. Thatโ€™s exactly what these 5 open-source courses teach you to do. At Decoding ML, we were tired of surface-level tutorials that only scratched the surface of LLMs, RAG, and AI agents. So we built the kind of learning experience we wished we had when starting: - Project-based - Opinionated - Production-ready. No fake data or hand-waving over infra. Just real-world projects backed by engineering best practices: โœ… Modular Python architecture โœ… Full-stack MLOps + LLMOps โœ… RAG, agents, and evaluation systems โœ… Fine-tuning, serving, and containerization โœ… Building full-fledged end-to-end systems. And yesโ€ฆ youโ€™ll need to sweat through the hard parts. Because these arenโ€™t one-notebook tutorials or weekend demos. These are full-stack, real-world AI systems with multiple components, modular architecture, and production-level complexity. We teach you how to: - Connect custom pipelines across ingestion, retrieval, and inference - Orchestrate agents with memory, reasoning, and tool use - Containerize, serve, and version your models like a real AI engineer - Monitor, evaluate, and iterate using observability best practices Here's exactly what you'll build: ๐Ÿญ. ๐—ฃ๐—ต๐—ถ๐—น๐—ผ๐—”๐—ด๐—ฒ๐—ป๐˜๐˜€ (๐˜„๐—ถ๐˜๐—ต The Neural Maze) Build a character simulation engine powered by RAG agents, memory, and real-time inference. โ†’ Learn LangGraph, RAG agents, Observability, and shipping agents as real-time APIs. ๐Ÿฎ. ๐—ฆ๐—ฒ๐—ฐ๐—ผ๐—ป๐—ฑ ๐—•๐—ฟ๐—ฎ๐—ถ๐—ป ๐—”๐—œ ๐—”๐˜€๐˜€๐—ถ๐˜€๐˜๐—ฎ๐—ป๐˜  Chat with your knowledge base using a custom agentic RAG system. โ†’ Learn modular RAG pipelines, fine-tuning LLMs, full-stack deployment, and LLMOps. ๐Ÿฏ. ๐—”๐—บ๐—ฎ๐˜‡๐—ผ๐—ป ๐—ง๐—ฎ๐—ฏ๐˜‚๐—น๐—ฎ๐—ฟ ๐—ฆ๐—ฒ๐—บ๐—ฎ๐—ป๐˜๐—ถ๐—ฐ ๐—ฆ๐—ฒ๐—ฎ๐—ฟ๐—ฐ๐—ต  Build a natural language product RAG search engine for structured data. โ†’ Learn hybrid retrieval leveraging tabular data, embeddings, and metadata filtering. ๐Ÿฐ. ๐—Ÿ๐—Ÿ๐—  ๐—ง๐˜„๐—ถ๐—ป  Create your own digital AI replica that reflects your knowledge and communication style. โ†’ Learn LLM fine-tuning, RAG, vector DBs, and building end-to-end LLMOps systems. ๐Ÿฑ. ๐—›&๐—  ๐—ฅ๐—ฒ๐—ฎ๐—น-๐—ง๐—ถ๐—บ๐—ฒ ๐—ฅ๐—ฒ๐—ฐ๐—ผ๐—บ๐—บ๐—ฒ๐—ป๐—ฑ๐—ฒ๐—ฟ  Deploy a neural fashion recommender on Kubernetes using Hopsworks + KServe. โ†’ Learn real-time recommender systems, LLM-augmented recsys, and MLOps workflows. Everything is FREE. All you have to do is: โ†’ Clone the GitHub repo โ†’ Open the Substack lesson โ†’ Run the code + follow the guide โ†’ Remix it and build your own production AI system If you're serious about going from "learning AI" to actually shipping it, this is where to start. The link is in the comments.


70

LangChain suggests you should take our PhiloAgents course to get into AI agents ready for production ๐Ÿฅ‚ Such an amazing work Miguel Otero Pedrido Love this collaboration!

Profile picture of LangChain

LangChain


๐Ÿค–๐ŸŽ“ PhiloAgents Build AI agents that impersonate philosophers with LangGraph in this OSS repo covering RAG implementation, real-time conversations, and system architecture with FastAPI & MongoDB integration. Start building philosophical agents! ๐Ÿš€ https://lnkd.in/gJ9NyH8X

  • Image description

78

Hugging Face released a new open-source course on The Model Context Protocol (MCP) The course is divided into 4 units. These will take you from the basics of Model Context Protocol to a final project implementing MCP in an AI application. ๐Ÿ”— Check it out: https://lnkd.in/d9awb4dJ


107

You canโ€™t build human-like agents without human-like memory. But most builders skip this part entirely. They focus on prompts, tools, and orchestration. But forget the system that holds it all together... Memory. In humans, memory is layered: โ†’ Working memory for what's happening right now โ†’ Semantic memory for facts and general knowledge โ†’ Procedural memory for skills and habits โ†’ Episodic memory for lived experience Agents are no different. If you want believable, useful, context-aware AI... You MUST architect memory intentionally. Hereโ€™s a breakdown of short and long-term memory types: - ๐—ฆ๐—ต๐—ผ๐—ฟ๐˜-๐˜๐—ฒ๐—ฟ๐—บ ๐—บ๐—ฒ๐—บ๐—ผ๐—ฟ๐˜† Stores active conversation threads and recent steps. This is your context window. Lose it, and your agent resets after every turn. For long-term memory, we have: - ๐—ฆ๐—ฒ๐—บ๐—ฎ๐—ป๐˜๐—ถ๐—ฐ ๐—บ๐—ฒ๐—บ๐—ผ๐—ฟ๐˜† Factual world knowledge retrieved through vector search or RAG. Think: โ€œWhatโ€™s the capital of France?โ€ or โ€œWhat is stoicism?โ€ - ๐—ฃ๐—ฟ๐—ผ๐—ฐ๐—ฒ๐—ฑ๐˜‚๐—ฟ๐—ฎ๐—น ๐—บ๐—ฒ๐—บ๐—ผ๐—ฟ๐˜† Defines what your agent knows to do, encoded directly in your code. From simple templates to complex reasoning flowsโ€”this is your logic layer. - ๐—˜๐—ฝ๐—ถ๐˜€๐—ผ๐—ฑ๐—ถ๐—ฐ ๐—บ๐—ฒ๐—บ๐—ผ๐—ฟ๐˜† Stores user-specific past interactions. Itโ€™s what enables continuity, personalization, and learning over time. In our ๐—ฃ๐—ต๐—ถ๐—น๐—ผ๐—”๐—ด๐—ฒ๐—ป๐˜๐˜€ ๐—ฐ๐—ผ๐˜‚๐—ฟ๐˜€๐—ฒ, we show how to wire all of this together. โ†’ Using MongoDB for structured memory โ†’ Using LangGraph (by LangChain) to control memory flow โ†’ Using Groq for real-time LLM inference โ†’ And even using Opik (by @company_cometml) to evaluate how memory shapes performance TL;DR: A smart agent isnโ€™t one that just thinks well... Itโ€™s one that remembers well, too. ๐Ÿ”— Learn more here: https://lnkd.in/d5ySvC_s

  • No alternative text description for this image

241

Unpopular opinion: fine-tuning is not hard. You know what is? Choosing HOW to fine-tune. There was one rule we stuck by when we began training our summarization LLM in the Second Brain course - Use a toolbelt that just works for 99% of use cases and ignore the 1% of edge cases that require GPU wizardry or DevOps magic. Hereโ€™s what we landed on: ๐Ÿ›  ๐—ง๐—ฅ๐—Ÿ โ€“ ๐—›๐˜‚๐—ด๐—ด๐—ถ๐—ป๐—ด ๐—™๐—ฎ๐—ฐ๐—ฒโ€™๐˜€ ๐—ฏ๐—ฎ๐˜๐˜๐—น๐—ฒ-๐˜๐—ฒ๐˜€๐˜๐—ฒ๐—ฑ ๐—ณ๐—ถ๐—ป๐—ฒ-๐˜๐˜‚๐—ป๐—ถ๐—ป๐—ด ๐—น๐—ถ๐—ฏ๐—ฟ๐—ฎ๐—ฟ๐˜† Perfect for both SFT and preference alignment. Maintained, well-documented, and up-to-date with the latest algorithms. โšก๏ธ ๐—จ๐—ป๐˜€๐—น๐—ผ๐˜๐—ต โ€“ ๐—Ÿ๐—ถ๐—ด๐—ต๐˜๐˜„๐—ฒ๐—ถ๐—ด๐—ต๐˜ ๐—ณ๐—ถ๐—ป๐—ฒ-๐˜๐˜‚๐—ป๐—ถ๐—ป๐—ด ๐—ฎ๐˜ ๐—ถ๐˜๐˜€ ๐—ฏ๐—ฒ๐˜€๐˜ Built by Daniel Han and Michael Han (Unsloth), Unsloth AI is making waves- and for good reason: โ†’ 2x faster training โ†’ Up to 80% less VRAM usage โ†’ GGUF quantization for local deployment โ†’ Works with Llama.cpp and Ollama โ†’ Actively fixing bugs in open models alongside Meta, Google, and Microsoft We used it to fine-tune a Llama 3.1 8B model on a T4 GPU: - 70% less VRAM - Full fine-tuning on commodity hardware - Same results for a fraction of the cost ๐Ÿ“Š ๐—–๐—ผ๐—บ๐—ฒ๐˜ โ€“ ๐—ง๐—ฟ๐—ฎ๐—ฐ๐—ธ ๐—ฒ๐˜ƒ๐—ฒ๐—ฟ๐˜†๐˜๐—ต๐—ถ๐—ป๐—ด ๐˜๐—ต๐—ฎ๐˜ ๐—บ๐—ฎ๐˜๐˜๐—ฒ๐—ฟ๐˜€ Your training logs shouldnโ€™t live in screenshots. Comet helped us version runs, compare experiments, and debug without chaos. โ€” The result? Fast, reproducible, and low-cost fine-tuning that scales. If youโ€™re building your own fine-tuning pipeline, this trio will carry you far. Unless you enjoy bleeding-edge painโ€ฆ thereโ€™s no reason to reinvent this setup. โ€” Full breakdown in Lesson 5 of the PhiloAgents course (Link in comments)


212

Claudeโ€™s leaked system prompt just confirmed what we all suspected: Vertical > General (No AGI). The best LLMs wonโ€™t do everything*.* Theyโ€™ll do one thing extremely well. I read all 22,000 words of Claude's leaked system promptโ€ฆ It wasnโ€™t some vague, high-level โ€œyou are a helpful assistantโ€ instruction set. It was a deeply engineered blueprint custom-built for one job. โ†’ Code-heavy tasks in JavaScript and Python Hereโ€™s what stood out (and what it signals about where LLMs are heading): ๐Ÿญ. ๐—œ๐˜ ๐˜‚๐˜€๐—ฒ๐˜€ ๐—ซ๐— ๐—Ÿ ๐˜๐—ผ ๐˜€๐˜๐—ฟ๐˜‚๐—ฐ๐˜๐˜‚๐—ฟ๐—ฒ ๐—ถ๐˜๐˜€ ๐˜๐—ต๐—ถ๐—ป๐—ธ๐—ถ๐—ป๐—ด No, โ€œYou are a helpful assistant.โ€ This is industrial-grade logic. It segments instructions into reusable XML tags: Each one acts like a callable function in a reasoning engine. ๐Ÿฎ. ๐—ง๐—ผ๐—ผ๐—น ๐˜‚๐˜€๐—ฒ ๐—ถ๐˜€๐—ปโ€™๐˜ ๐—ท๐˜‚๐˜€๐˜ ๐—ฎ๐—น๐—น๐—ผ๐˜„๐—ฒ๐—ฑ (๐—ถ๐˜โ€™๐˜€ ๐—ฒ๐—ป๐—ด๐—ถ๐—ป๐—ฒ๐—ฒ๐—ฟ๐—ฒ๐—ฑ) Claude is taught how to use tools like a software engineer: โœ… When to call โŒ When not to โš ๏ธ Use memory first ๐Ÿ”„ Limit to 1โ€“2 calls ๐Ÿ“ Over 5? Follow a strict workflow Not โ€œcall a tool,โ€ but design a workflow. ๐Ÿฏ. ๐— ๐—ผ๐—ฑ๐—ฒ๐—ฟ๐—ฎ๐˜๐—ถ๐—ผ๐—ป ๐—ฎ๐—ป๐—ฑ ๐—น๐—ฒ๐—ด๐—ฎ๐—น ๐˜€๐—ฎ๐—ณ๐—ฒ๐˜๐˜† ๐—ฎ๐—ฟ๐—ฒ ๐—ต๐—ฎ๐—ฟ๐—ฑ๐—ฐ๐—ผ๐—ฑ๐—ฒ๐—ฑ โ€œClaude is happy to write creative content involving fictional characters, but avoids writing content involving real, named public figures.โ€ Even moderation is framed as a behavior, not a filter. ๐Ÿฐ. ๐—œ๐˜ ๐˜๐—ฒ๐—ฎ๐—ฐ๐—ต๐—ฒ๐˜€ ๐—–๐—น๐—ฎ๐˜‚๐—ฑ๐—ฒ ๐—ต๐—ผ๐˜„ ๐˜๐—ผ ๐—ฟ๐—ฒ๐—ฎ๐˜€๐—ผ๐—ป ๐˜€๐˜๐—ฒ๐—ฝ-๐—ฏ๐˜†-๐˜€๐˜๐—ฒ๐—ฝ Want Claude to count words or characters? โ€œIt explicitly counts... assigning a number to each. It only answers once it has performed this step.โ€ Want it to analyze books or code? โ€œClaude should provide a summary from its internal knowledge, and only search when necessary.โ€ This is instruction tuning in the wild. ๐Ÿฑ. ๐—œ๐˜ ๐—ถ๐—ป๐—ฐ๐—น๐˜‚๐—ฑ๐—ฒ๐˜€ ๐˜‚๐˜€๐—ฎ๐—ด๐—ฒ ๐—ด๐˜‚๐—ถ๐—ฑ๐—ฒ๐˜€ ๐—ณ๐—ผ๐—ฟ ๐˜€๐—ฝ๐—ฒ๐—ฐ๐—ถ๐—ณ๐—ถ๐—ฐ ๐˜๐—ฒ๐—ฐ๐—ต ๐˜€๐˜๐—ฎ๐—ฐ๐—ธ๐˜€ Yes, inside the system prompt. - How to use TailwindCSS - When to reach for lodash vs. vanilla JS - What Claude should do when reading .env files - How to parse messy CSVs - Which React libraries to use for graphsโ€ฆ This is a fine-tuned developer assistant pretending to be general-purpose. ๐—ฆ๐—ผ ๐˜„๐—ต๐—ฎ๐˜โ€™๐˜€ ๐˜๐—ต๐—ฒ ๐—ฏ๐—ถ๐—ด ๐˜๐—ฎ๐—ธ๐—ฒ๐—ฎ๐˜„๐—ฎ๐˜†? Unlike GPT or Geminiโ€™s system prompts, which are short, abstract, and vague, Claudeโ€™s is specific, opinionated, and operational. Itโ€™s not trying to be everything. Itโ€™s trying to do certain things very well. โ†’ Code in JS and Python โ†’ Use tools with precision โ†’ Write with context and restraint โ†’ Reason step-by-step โ†’ Stay within legal and ethical boundaries And that explains why Claude is so good at what it does (and not great at everything else). If youโ€™re building agentic systems or advanced assistants, go read the prompt. Itโ€™s a masterclass in instruction design. โ™ป๏ธ Share this to help someone in your network :)


211

90% of AI engineers are dangerously abstracted from reality. They work with: โ†’ Prebuilt models โ†’ High-level APIs โ†’ Auto-magical cloud tools But hereโ€™s the thing - If you donโ€™t understand how these tools actually work, youโ€™ll always be guessing when something breaks. Thatโ€™s why the best AI engineers I know go deeper... They understand: How Git actually tracks changes. How Redis handles memory. How Docker isolates environments. If youโ€™re serious about engineering, you'd go build the tools you use. And itโ€™s why I recommend CodeCrafters.io (YC S22) You wonโ€™t just learn tools. Youโ€™ll rebuild them (from scratch). โ†’ Git, Redis, Docker, Kafka, SQLite, Shell... โ†’ Step by step, test by test โ†’ In your favorite language (Rust, Python, Go, etc.) Itโ€™s perfect for AI engineers who want to: โ†’ Level up their backend + system design skills โ†’ Reduce debugging time in production โ†’ Build apps that actually scale under load And most importantly... โ†’ Stop being a model user โ†’ Start being a systems thinker If I had to level up my engineering foundations today, CodeCrafters is where Iโ€™d start. The ink is in the comments. P.S. We only promote tools we use or would personally take. P.S.S. Subscribe with my affiliate link to get a 40% discount :)


195

RAG isnโ€™t your bottleneck. Blind deployment is. Everyoneโ€™s obsessed with squeezing more performance out of their retrieval pipelines. Better chunking Better embeddings Better reranking All great. But none of that matters if you canโ€™t fix what you donโ€™t see. 90% of people building agents today donโ€™t actually know what their agents are doing (especially when they go into production): โ†’ Is the reasoning solid? โ†’ Are prompt tweaks helping or hurting? โ†’ Is performance degrading silently over time? By the time you notice somethingโ€™s off... itโ€™s already too late. Thatโ€™s why ๐—Ÿ๐—ฒ๐˜€๐˜€๐—ผ๐—ป ๐Ÿฑ ๐—ผ๐—ณ ๐˜๐—ต๐—ฒ ๐—ฃ๐—ต๐—ถ๐—น๐—ผ๐—”๐—ด๐—ฒ๐—ป๐˜๐˜€ course is all about observability. Agents that produce ROI don't just sound smart... They are also measurable, versioned, and constantly improving. Hereโ€™s what we cover in this lesson: โ†’ How to monitor complex LLM traces in real-time using Opik โ†’ How to version every prompt change for reproducibility โ†’ How to generate eval sets and benchmark your agents โ†’ How to run online and offline evaluation across your pipelines โ†’ How observability fits into your LLMOps stack This is the part of agentic AI that separates demo projects from production systems. Huge thanks to Anca Ioana Muscalagiu for the deep-dive article. And as always, shout-out to Miguel Otero Pedrido for building this with me. Want to dive into lesson 5? Here you go - ๐Ÿ“ Article: https://lnkd.in/dRYgHyid ๐ŸŽฅ Video: https://lnkd.in/dEQ_Yv7n

  • No alternative text description for this image

193

This year, I gave my first EVER in-person talk. And the one thing I feared mostโ€ฆ actually happened. Let me explain. Those who've been following me for a while would know I made a scary promise to myself: โ€œStop hiding behind a keyboard. Start showing up in real life.โ€ So when I was invited to speak at QCon Software Development Conferences - one of Europeโ€™s biggest software and AI conferences - I said had no choice but to say, "yes." Even though I was terrified. My talk was on The Data Backbone of LLM Systems. A 60-minute deep dive into the infrastructure behind real-world RAG, LLMs and LLMOps. The room was packed with senior engineers from companies like Netflix, Google, Confluent, and MongoDB. And 30 seconds before I startedโ€ฆ My clicker broke. No slides. No backup. Just me, 120 people, and a frozen screen. But something kicked in... I tossed the clicker aside, walked to my laptop, and started speaking - manual slide switching and all. And somehowโ€ฆ it worked. The presentation wasnโ€™t perfect (I wasnโ€™t expecting), but I learned a lot in what to do at my future talks. Still, I managed to: โ†’ Scored 93% (vs conference average of 83%) โ†’ Deliver every insight I came to share โ†’ Walk off stage knowing Iโ€™d crushed one of my biggest fears It was a personal turning point. Iโ€™m proud of the lessons I shared on stage ... and Iโ€™m even prouder of the one I learned off-stage: ๐—–๐—ผ๐˜‚๐—ฟ๐—ฎ๐—ด๐—ฒ ๐—ฐ๐—ผ๐—บ๐—ฝ๐—ผ๐˜‚๐—ป๐—ฑ๐˜€. Excited to see at what conference I will talk next! Thank you QCon Software Development Conferences for the platform. And thank you to everyone who showed up - you made this milestone unforgettable.


164

95% of agents never leave the notebook. And itโ€™s not because the code is bad... Itโ€™s because the system around them doesnโ€™t exist. Here's my point: Anyone can build an agent that works in isolation. The real challenge is shipping one that survives real-world conditions (e.g., live traffic, unpredictable users, scaling demands, and messy data). That's exactly what we tackled in ๐—Ÿ๐—ฒ๐˜€๐˜€๐—ผ๐—ป ๐Ÿญ ๐—ผ๐—ณ ๐˜๐—ต๐—ฒ ๐—ฃ๐—ต๐—ถ๐—น๐—ผ๐—”๐—ด๐—ฒ๐—ป๐˜๐˜€ ๐—ฐ๐—ผ๐˜‚๐—ฟ๐˜€๐—ฒ. We started by asking, "What does an agent need to survive in production?" And decided on 4 things - It needs an LLM to run in real-time. A memory to understand what just happened. A brain that can reason and retrieve factual information. And a monitor to ensure it all works under load. So we designed a system around those needs. The frontend is where the agent comes to life. We used Phaser to simulate a browser-based world. But more important than the tool is the fact that this layer is completely decoupled from the backend. (so game logic and agent logic evolve independently) The backend, built in FastAPI, is where the agent thinks. We stream responses token-by-token using WebSockets. All decisions, tool calls, and memory management happen server-side. Inside that backend sits the agentic core - a dynamic state graph that lets the agent reason step-by-step. The agent is orchestrated by LangGraph and powered by Groq for real-time inference speeds. It can ask follow-up questions, query external knowledge, or summarize whatโ€™s already been said (all in a loop). When the agent needs facts, it queries long-term memory. We built a retrieval system that mixes semantic and keyword search, using cleaned, de-duplicated philosophical texts crawled from the open web. That memory lives in MongoDB and gets queried in real time. Meanwhile, short-term memory tracks the conversation thread across turns. Without it, every new message would be a reset. With it, the agent knows whatโ€™s been said, whatโ€™s been missed, and how to respond. But hereโ€™s the part most people skip: observability. If you want to improve your system, you need to see and measure what it's doing. Using Opik (by Comet), we track every prompt, log every decision, and evaluate multi-turn outputs using automatically generated test sets. Put it all together and you get a complete framework that remembers, retrieves, reasons, and responds in a real-world environment. Oh... and we made the whole thing open source. ๐Ÿ”— Link: https://lnkd.in/d8-QbhCd P.S. Special shout out to my co-creator Miguel Otero Pedrido

  • No alternative text description for this image

309

90% of RAG systems struggle with the same bottleneck: (And better LLMs are not the solution) It's retrieval. And most teams donโ€™t realize it because they rush to build without proper evaluation. Before I tell you how to fix this, let me make something clear - ๐—ก๐—ฎ๐—ถ๐˜ƒ๐—ฒ ๐—ฅ๐—”๐—š ๐—ถ๐˜€ ๐—ฒ๐—ฎ๐˜€๐˜†. You chunk some docs, embed them, drop a top_k retriever on top, and call it a pipeline. Getting it production-ready? Thatโ€™s where most teams stall. โ†’ They get hallucinations. โ†’ They miss key info. โ†’ Their outputs feel... off. Why? Because the quality of generation is downstream of the quality of context. ... and naive RAG often pulls in irrelevant or partial chunks that confuse the LLM. If you're serious about improving your system, here's the progression that actually works: ๐—ฆ๐˜๐—ฒ๐—ฝ ๐Ÿญ: ๐—™๐—ถ๐˜… ๐˜๐—ต๐—ฒ ๐—•๐—ฎ๐˜€๐—ถ๐—ฐ๐˜€ These โ€œtable-stakesโ€ upgrades outperform fancy models most of the time: โ†’ Smarter Chunking - Dynamic over fixed-size. Respect structure. โ†’ Chunk Size Tuning - Too long = loss in the middle. Too short = fragmented context. โ†’ Metadata Filtering - Boosts precision by narrowing scope semantically and structurally. โ†’ Hybrid Search - Combine vector + keyword filtering. ๐—ฆ๐˜๐—ฒ๐—ฝ ๐Ÿฎ: ๐—Ÿ๐—ฎ๐˜†๐—ฒ๐—ฟ ๐—ผ๐—ป ๐—”๐—ฑ๐˜ƒ๐—ฎ๐—ป๐—ฐ๐—ฒ๐—ฑ ๐—ฅ๐—ฒ๐˜๐—ฟ๐—ถ๐—ฒ๐˜ƒ๐—ฎ๐—น When basic techniques arenโ€™t enough: โ†’ Re-ranking (learned or rule-based) โ†’ Small-to-Big Retrieval: Retrieve sentences, synthesize larger windows. โ†’ Recursive Retrieval (e.g., LlamaIndex) โ†’ Multi-hop + agentic retrieval: When you need reasoning across documents. ๐—ฆ๐˜๐—ฒ๐—ฝ ๐Ÿฏ: ๐—˜๐˜ƒ๐—ฎ๐—น๐˜‚๐—ฎ๐˜๐—ฒ ๐—ผ๐—ฟ ๐——๐—ถ๐—ฒ ๐—ง๐—ฟ๐˜†๐—ถ๐—ป๐—ด There's no point iterating blindly. Do the following: โ†’ End-to-End eval - Is the output good? Ground truths, synthetic evals, user feedback. โ†’ Component-level eval - Does the retriever return the right chunks? Use ranking metrics like MRR, NDCG, success@k. ๐—ฆ๐˜๐—ฒ๐—ฝ ๐Ÿฐ: ๐—™๐—ถ๐—ป๐—ฒ-๐˜๐˜‚๐—ป๐—ถ๐—ป๐—ด = ๐—Ÿ๐—ฎ๐˜€๐˜ ๐—ฅ๐—ฒ๐˜€๐—ผ๐—ฟ๐˜ Donโ€™t start here. Do this only when: โ†’ Your domain is so specific general embeddings fail. โ†’ Your LLM is too weak to synthesize even when context is correct. โ†’ Youโ€™ve squeezed all juice from prompt + retrieval optimizations. Fine-tuning adds cost, latency, and infra complexity. Itโ€™s powerful, but only when everything else is dialed in. ๐—ก๐—ผ๐˜๐—ฒ: These notes are from a talk over a year old. And yet... most teams are still stuck in Step 0. That tells you something - The surface area of RAG is small. But building good RAG is still an unsolved craft. Letโ€™s change that. Want to learn to implement advanced RAG systems yourself? The link is in the comments. ๐—œ๐—บ๐—ฎ๐—ด๐—ฒ ๐—ฐ๐—ฟ๐—ฒ๐—ฑ๐—ถ๐˜: LlamaIndex and Jerry Liu


267

Everyone likes to talk about models, prompts, and performance hacks. But no one teaches you how to ship. In ๐—Ÿ๐—ฒ๐˜€๐˜€๐—ผ๐—ป ๐Ÿฒ ๐—ผ๐—ณ ๐˜๐—ต๐—ฒ ๐—ฃ๐—ต๐—ถ๐—น๐—ผ๐—”๐—ด๐—ฒ๐—ป๐˜๐˜€ ๐—ฐ๐—ผ๐˜‚๐—ฟ๐˜€๐—ฒ, we fix that. We go from messy PoC to clean architecture. Hereโ€™s what youโ€™ll learn: โ†’ How to organize your Python project like a professional engineer โ†’ Why the โ€œappโ€ folder mindset saves you months of debugging later โ†’ How to use Docker, .env configs, and modular code. โ†’ Why reproducibility and portability matter as much as inference speed โ†’ The real difference between hacking an agent and engineering one The goal was to teach you how to build a real system that's durable. Huge thanks to Miguel Otero Pedrido for co-creating this lesson with me. (His engineering brain pushed this to the next level) If you're stuck in notebook purgatory and want to break out, this lessonโ€™s for you. Lesson 6 is now live! (Link in the comments)


245

Want to drive more opportunities from LinkedIn?

Content Inspiration, AI, scheduling, automation, analytics, CRM.

Get all of that and more in Taplio.

Try Taplio for free

Famous LinkedIn Creators to Check Out

Amelia Sordell ๐Ÿ”ฅ

@ameliasordell

Klowt builds personal brands. I founded the business after realising that the best leads came throu...

228k

Followers

Vaibhav Sisinty โ†—๏ธ

@vaibhavsisinty

I'm an engineer turned marketer, now a founder. I've worked at Uber and Klook, focusing on marketi...

451k

Followers

Sabeeka Ashraf

@sabeekaashraf

On January 8th my "one day" became DAY ONE ... 7 days earlier I downgraded my life into a suitcase....

20k

Followers

Matt Gray

@mattgray1

Over the last decade, Iโ€™ve built 4 successful companies and a community of over 14 million people. ...

1m

Followers

Hi! Iโ€™m Daniel. Iโ€™m the creator of The Marketing Millennials and the founder of Authority, a B2B Lin...

150k

Followers

Shlomo Genchin

@shlomogenchin

Hey! Here are 3 ways I can help you: 1๏ธโƒฃ Talks and Workshops: I'll show your team, or students, how...

49k

Followers

Sam G. Winsbury

@sam-g-winsbury

We turn entrepreneurs into credible thought leaders through personal branding so they can scale thei...

49k

Followers

Ash Rathod

@ashrathod

You already know storytelling is essential for your business and brand. But storytelling is much m...

73k

Followers

Richard Moore

@richardjamesmoore

โฉYou know how all the clients you'll ever work with are on LinkedIn, right? But you struggle to gene...

105k

Followers

Izzy Prior

@izzyprior

No matter how outrageously amazing your mission is, it's likely you're not seeing the results you ne...

82k

Followers

Andy Mewborn

@amewborn

I use to be young & cool. Now I do b2b SaaS. Husband. Dad. Ironman. Founder of Distribute // Co-fo...

215k

Followers

Wes Kao

@weskao

Wes Kao is an entrepreneur, coach, and advisor who writes at newsletter.weskao.com. She is co-founde...

107k

Followers

Justin Welsh

@justinwelsh

Over the last decade, I helped build two companies past a $1B valuation and raise over $300M in vent...

1m

Followers

Sahil Bloom

@sahilbloom

Sahil Bloom is the New York Times Bestselling author of The 5 Types of Wealth: A Transformative Guid...

1m

Followers

Luke Matthews

@lukematthws

LinkedIn has changed. You need to change too. Hey I'm Luke, I've been marketing for 5+ years on ...

188k

Followers

Tibo Louis-Lucas

@thibaultll

Founder Prev Taplio & Tweet Hunter (sold) Building Typeframes & revid.ai Invested in animstats.com ...

6k

Followers

Guillaume Moubeche

@-g-

If youโ€™re here, that's because you know that your personal growth will drive your business growth ๐Ÿš€...

80k

Followers

Austin Belcak

@abelcak

CultivatedCulture.com/Coaching // I teach people how to land jobs they love in today's market withou...

1m

Followers