Skip to content

Writing and mumblings

I write about a mix of consulting, open source, personal work, and applying llms. I won't email you more than twice a month, not every post I write is worth sharing but I'll do my best to share the most interesting stuff including my own writing, thoughts, and experiences.

Subscribe to my Newsletter Follow me on X

For posts about RAG (Retrieval-Augmented Generation) or LLMs (Large Language Models), check out the category labels in the sidebar. Here are some of my best posts on these topics:

Personal Stories

RAG and LLM Insights

Consulting and Tech Advice

Talks and Interviews

Creating Content That Converts: My Guide for AI Consultants

This is some of the notes I've taken for learnindieconsulting.com

Why I Prioritize Content (And Why You Should Too)

Let me share something I wish I'd understood sooner: consistent content creation isn't just a marketing tactic—it's the foundation of a thriving consulting business.

When I started my consulting journey, I was stuck in the time-for-money trap. I'd jump on Zoom calls with prospects, explain the same concepts repeatedly, and wonder why scaling was so difficult. Then I had a realization that changed everything: what if I could have these conversations at scale?

Now I extract blog post ideas from every client call. Every Friday, I review about 17 potential topics from the week's conversations. I test them with social posts, see which ones get traction (some get 700 views, others 200,000), and develop the winners into comprehensive content.

Here's why this approach has transformed my business:

My Content is My 24/7 Sales Team

Last year, I wrote a post called "RAG is more than embeddings" that generated around 30,000 views on Hacker News. This single piece became one of the biggest catalysts for my consulting business.

The real magic happens when a prospect asks a question, and I can send them a blog post I wrote months ago. It immediately signals that I've already thought deeply about their problem. They think, "This person has already solved the exact challenge I'm facing."

Breaking Free From Hourly Billing

Content helped me escape the limitations of hourly billing. After getting tired of repeating the same RAG implementation advice at $1,000 per hour, I created a course that's more accessible while freeing my time for higher-value work.

This ladder of offerings at different price points lets clients engage with me at whatever level makes sense for them—from free blog posts to high-end consulting. Without content, this business model wouldn't be possible.

My Content is My Lab for Testing Ideas

I don't try to perfectly identify my audience before creating. Instead, I use content as a sensor to discover what resonates.

When I started on Twitter, I had about 500 followers. I noticed my follower-to-tweet ratio was around 1.7, meaning if I wanted 10,000 followers, I'd need to write roughly 7,000 tweets. So I committed to consistent creation—it's like losing weight; it depends on how often you're working out.

Volume negates luck, and these efforts compound quickly. Don't wait months—focus on creating volume and consistency.

Building Long-Term Assets That Appreciate

Every piece of content I create becomes part of a growing library that continuously generates value. Unlike client projects that end, my content:

  • Continues attracting leads while I sleep
  • Builds upon itself as pieces reference each other
  • Serves as the foundation for courses and other products
  • Allows me to work with better clients on more interesting problems

Now let me walk you through exactly how I approach content creation, from foundational frameworks to practical execution.

The Value Equation: The Heart of All Great Content

Everything I create revolves around a simple but powerful formula:

Value = (Dream Outcome × Probability of Success) ÷ (Time × Effort)

This equation isn't just about demonstrating value as a consultant—it's about creating content that resonates deeply with readers.

When I write about RAG implementation, I don't just explain technical details. I show readers:

  • The dream outcome: "Imagine a system that consistently returns precisely relevant information, dramatically improving user trust and reducing churn"
  • How I increase probability: "By implementing these three specific retrieval strategies, you can improve precision by 37%"
  • How I reduce time: "This approach cuts development time from months to weeks"
  • How I minimize effort: "My framework eliminates the need to manually tune parameters"

This value equation shapes everything—my blog posts, social content, course materials, and sales conversations. It's not just theory; it directly improves how I communicate with clients.

The AIDA Framework: My Structure for Every Piece

Every piece of content I create follows this proven structure:

  • **A**ttention: I grab interest with a hook that speaks directly to a pain point
  • **I**nterest: I build credibility with relevant insights and information
  • **D**esire: I create a vivid picture of the outcome readers want
  • **A**ction: I provide a clear next step

This isn't just marketing-speak. Practicing this framework has made me more articulate during sales conversations and client calls. It's like "wax on, wax off"—you're training communication muscles that serve you across contexts.

Let me share an example of how I applied this to a recent blog post:

Attention: "Most RAG systems fail because engineers focus exclusively on embeddings while ignoring three critical components."

Interest: "After implementing 20+ RAG systems across industries, I've identified patterns that consistently separate high-performing systems from failures."

Desire: "By addressing these overlooked elements, you can build systems that deliver consistently relevant results and earn genuine user trust."

Action: "Use my evaluation framework (linked here) to identify the specific weaknesses in your current implementation."

Crafting Titles That Demand Attention

I've learned that your title isn't just important—it's about 80% of the work. A mediocre post with a compelling title will outperform a brilliant post with a boring title every time.

Strong vs. Weak Titles: A Tale of Two Approaches

Here's an example from my own experience:

When I first started writing, I'd use titles like: "I spent the weekend playing around with ChatGPT"

This attracted almost no interest because: - It focused on what I did, not what readers could gain - It gave no reason to click through - It didn't specify any outcome or benefit

When I shifted to titles like: "3 ChatGPT Prompts That Doubled My Content Creation Speed" - Engagement skyrocketed because: - It quantified the benefit (doubled speed) - It promised specific, actionable advice (3 prompts) - It addressed a common pain point (slow content creation)

My Title Testing Strategy

I never commit to a full article without testing the title first. Here's my process:

  1. I create 3-5 different social posts with the same core idea but different framing
  2. I track which gets the most engagement
  3. I use the winning framing for my full content piece

For example, when writing about pricing strategies, I tested: - "3 mistakes I made in my AI pricing strategy" - "How I increased my consulting rates by 200% in 6 months" - "The psychological barrier preventing consultants from charging what they're worth" - "Why your AI services are underpriced (and what to do about it)"

The last one generated 4x more engagement than the others, so that became my focus for the full article.

Finding Your Audience Through Content

One question I hear constantly is: "How do I find my audience?"

My answer might surprise you: I don't try to perfectly identify my audience before creating content. Instead, I create content, see what resonates, and refine my understanding through this feedback loop.

When I built my open-source tool Instructor, I started with a message about wanting a simple tool that does one thing well, without unnecessary complexity. The audience came secondary to the message and dream.

That said, I do have a mental checklist I run through before creating content:

  • What specific role or position is my ideal reader in?
  • What business stage are they in?
  • What specific pain are they experiencing?
  • What dream outcome are they pursuing?
  • What has prevented them from achieving this outcome?
  • How technically sophisticated are they?

But I hold these ideas loosely, allowing real-world feedback to refine my understanding.

Platform Selection: Where to Share Your Thinking

Each platform attracts different audiences and serves different purposes in my content ecosystem:

Twitter/X: My Testing Ground

I use Twitter as my primary testing ground. With a follower-to-tweet ratio between 0.5-4, I know that generating 10,000 followers requires 2,500-20,000 tweets.

Twitter has been especially valuable for connecting with founders, investors, and technical audiences. Most of my industry relationships—including collaborations with other prominent consultants—came through Twitter.

LinkedIn: My Corporate Channel

When targeting enterprise clients, VPs, Directors, and corporate decision-makers, I shift to LinkedIn. The content is more formal and business-focused, emphasizing outcomes over technical details.

I've found that LinkedIn works better for corporate/enterprise VP-level content, while Twitter is ideal for founders and investors.

My Website: The Hub of My Content Ecosystem

I strongly believe in owning your distribution channel. While platforms like Medium or Towards Data Science might seem appealing, they diminish your personal brand.

On my own website, I can publish blog content while also offering email capture and a services page. This creates a journey where people might read my posts over several months before they're ready to explore my services.

I use a simple framework—markdown docs with MkDocs Material—making it easy to edit and maintain content without relying on third-party platforms.

Email Newsletter: My Most Valuable Asset

The quality of subscribers matters more than quantity. "If you offered me a hundred thousand AI news newsletter subscribers versus 200 CTOs in Silicon Valley, I would probably pay $80,000 for that CTO email list."

I started my newsletter relatively late, after my Twitter following had grown significantly. If you're starting now, make email capture a priority from the beginning.

My Content Creation Process

Over time, I've developed a systematic approach to creating content that consistently delivers results:

1. Idea Generation: Mining for Gold

I extract content ideas from multiple sources:

  • Client conversations: Every call generates potential topics
  • Social media testing: I post shorter ideas to see which ones resonate
  • Content gaps: I identify questions that aren't well answered
  • Personal experience: I share lessons from my own mistakes and successes

I've automated much of this process. After client calls, I use AI tools to extract potential blog ideas from the transcripts. Every Friday, I review these ideas, identifying patterns and high-potential topics.

2. Structuring for Impact

For each piece, I follow this proven structure:

The Hook (Attention) I open with something that immediately grabs attention—a provocative question, surprising statistic, or challenge to conventional wisdom.

For example, instead of starting with "RAG systems are important," I might open with: "90% of RAG implementations fail because they focus on the wrong problem entirely."

The Value Promise (Interest) I clearly state what the reader will learn and why it matters to them specifically.

Example: "In this guide, you'll discover the three overlooked components that determine RAG success—insights I've gathered from implementing systems for companies like X, Y, and Z."

The Main Content (Desire) I deliver actionable insights using specific examples and case studies. I demonstrate expertise without overwhelming with technical details.

The key is focusing on the hierarchy of value. Technical details are often the "lowest pole of value." I've learned that clients care much more about business outcomes than technical implementations.

The Call to Action (Action) I provide a clear next step that aligns with my business goals while offering genuine value.

3. Optimization for Different Readers

Before publishing, I optimize for different reading styles:

For Skimmers: - I use short paragraphs (2-3 sentences max) - I include clear subheadings - I incorporate bullet points and numbered lists - I bold key points

For Deep Readers: - I include personal stories or examples - I add visuals where relevant - I ask thought-provoking questions throughout - I provide evidence and results

Distribution: The 80/20 Rule

I spend 20% of my time creating content and 80% on distribution. Here's my approach:

1. Repurpose Aggressively

I turn blog posts into tweet threads, convert videos into blog posts, and extract key points for LinkedIn posts. This maximizes the return on every piece of content I create.

When my hands were hurting from typing too much, I started recording videos. This created an efficient workflow: 1. Test ideas with tweets to see what resonates 2. Record short videos about high-performing topics 3. Convert the video transcripts into blog posts

2. Cross-Platform Promotion

I share blog posts across all social channels, mention content in relevant communities, and reference past pieces in new content. This creates a network effect where each piece strengthens the others.

3. Engagement Amplification

I respond to all comments quickly, tag relevant people who might find my content valuable, and follow up with engaged readers. This turns content into conversations, which often lead to client relationships.

Measuring What Matters

I track several metrics to evaluate content effectiveness:

  1. Visibility metrics: Views, impressions, time on page
  2. Engagement metrics: Shares, comments, saves
  3. Conversion metrics: Email signups, call bookings
  4. Business impact: Content referenced in sales calls, deal close rates

The most important question I ask new clients: "How did you find me?" Their answers inform my content strategy.

TAM Assessment: Who Are You Really Reaching?

I'm strategic about estimating the Total Addressable Market (TAM) for my content:

Specific vs. General Content Highly specific titles reach smaller audiences but convert better. General titles reach larger audiences but convert poorly.

For example, "How to implement a RAG system" reaches a broad audience of developers, while "How to create RAG systems for financial compliance" reaches a tiny audience but with much higher conversion potential.

Balance TAM with Conversion Potential I aim for the sweet spot: content specific enough to convert well but broad enough to reach sufficient potential clients.

My Content Creation Checklist

Before publishing anything, I run through this mental checklist:

Pre-Creation

  • Have I tested this topic with smaller content first?
  • Is this aligned with my business goals and services?
  • What platform is most appropriate for this content?

Creation

  • Does my title follow the strong title formula?
  • Does my opening hook generate attention?
  • Have I used the AIDA framework throughout?
  • Does this content reflect the value equation?
  • Is my expertise clearly demonstrated?

Distribution

  • Have I planned how to promote this across platforms?
  • Can I repurpose this content into multiple formats?
  • Are there specific people I should notify about this content?

Measurement

  • What metrics will indicate success for this content?
  • How does this content fit into my overall strategy?
  • What follow-up content might make sense based on this piece?

Final Thoughts: Content as a Flywheel

The greatest benefit of consistent content creation is the flywheel effect it creates:

  1. Client work generates insights for content
  2. Content attracts new clients
  3. New clients provide more insights
  4. The cycle continues and accelerates

This has allowed me to work with increasingly better clients on more interesting problems at higher rates. My content acts as a filter, attracting people who value my specific expertise and approach.

Remember: Great content doesn't just attract an audience—it attracts the right audience who values your expertise enough to pay for it. And that makes all the difference in building a sustainable, fulfilling consulting business.


I hope this guide helps you create content that not only resonates with your audience but also drives real business results. If you have questions or want to share your content journey, reach out—I'm always interested in hearing how these approaches work for others.

Version Control for the Vibe Coder (Part 1)

Imagine this: you open Cursor, ask it to build a feature in YOLO-mode, and let it rip. You flip back to Slack, reply to a few messages, check your emails, and return...

It's still running.

What the hell is going on? .sh files appear, there's a fresh Makefile, and a mysterious .gitignore. Anxiety creeps in. Should you interrupt it? Could you accidentally trash something critical?

Relax—you're not alone. This anxiety is common, especially among developers newer to powerful agents like Cursor's. Fortunately, Git is here to save the day.

Fine-Tuning Embedding Models for Enterprise RAG: Lessons from Glean

Systematically improving RAG systems

This transcript is based off of a guest lecture given in my course, Systematically Improving RAG Applications

Retrieval-Augmented Generation (RAG) systems have become essential tools for enterprises looking to harness their vast repositories of internal knowledge. While the theoretical foundations of RAG are well-understood, implementing these systems effectively in enterprise environments presents unique challenges that aren't addressed in academic literature or consumer applications. This article delves into advanced techniques for fine-tuning embedding models in enterprise RAG systems, based on insights from Manav Rathod, a software engineer at Glean who specializes in semantic search and ML systems for search ranking and assistant quality.

The discussion focuses on a critical yet often overlooked component of RAG systems: custom-trained embedding models that understand company-specific language, terminology, and document relationships. As Jason Liu aptly noted during the session, "If you're not fine-tuning your embeddings, you're more like a Blockbuster than a Netflix." This perspective highlights how critical embedding fine-tuning has become for competitive enterprise AI systems.

Hard Truths From the AI Trenches

I never planned to become a consultant. But somewhere between building machine learning systems and growing my Twitter following, companies started sliding into my DMs with the same message: "Help, our AI isn't working."

So I started charging to join their stand-ups. Sometimes I didn't even code. I just asked uncomfortable questions.

Here's what I've learned watching companies burn millions on AI.

Your AI problems aren't AI problems

The head of AI at a major company once showed me their evaluation framework. They had 23 evals. One of them tested if their AI could "talk like a pirate."

I shit you not.

Meanwhile, 75% of their users were just trying to make their AI read Excel spreadsheets. But they had three engineers focused on a fancy document integration that less than 10% of users touched.

This is the pattern everywhere I go. Companies obsess over sophisticated AI capabilities while ignoring what users actually do with their product.

The truth? Your AI problem is almost always a data problem. Or a process problem. Or a people problem. But rarely is it actually an AI problem.

Stop waiting for the model to save you

"When GPT-5 comes out, all our problems will be solved."

I've heard this exact sentiment in different words from CTOs who should know better. It's magical thinking, and it's everywhere.

Here's a hard truth: capabilities are monotonically increasing. Each new model will be better than the last. But if your recall sucks now, ChatGPT 47.0 isn't going to save you.

Fix your retrieval. Fix your data pipeline. Fix your evaluation process. These are things you control today.

When I tell clients this, they often look disappointed. They wanted me to tweak some prompt magic or recommend a fancy new model. Instead, I'm telling them to do the unsexy work of: - Building better test sets - Implementing proper feedback loops - Actually looking at their data

One client went from 27% recall to 85% in four hours once they stopped praying to the model gods and started making targeted improvements to their data preprocessing. It wasn't glamorous, but it worked.

Measure what matters (and it's not what you think)

Most AI teams are measuring the wrong things. They obsess over model perplexity or ROUGE scores while their business burns.

At a sales tech company, engineers spent weeks trying to perfect their call transcript summarization accuracy. Meanwhile, sales managers didn't care about perfect summaries — they wanted to know which prospects had buying objections.

Classification first, extraction second. Always.

When we shifted from "how accurately can we summarize this call?" to "can we identify sales objections with 95% recall?", suddenly the business value was obvious. The sales team could follow up on objections, coach their reps better, and close more deals.

The best metric isn't technical. It's: "Does this solve a real problem someone will pay for?"

Experimentation speed is your only moat

Ask most AI teams how long it takes to run an experiment, and you'll get embarrassed silence. Four days? A week? "Well, we have to coordinate with three teams and..."

The companies winning with AI can run 10 experiments before lunch.

One client went from spending weeks fine-tuning prompts to implementing a system where they could: 1. Sample real-time traffic 2. Test multiple approaches 3. Compare metrics 4. Deploy winners

All in the same day.

Your ability to learn fast is everything. Build for learning velocity above all else.

As I tell my clients: how do you get better at jiu-jitsu? Mat time. More experiments, more data, more learning.

Testimonials > Metrics

Nobody cares that your RAG system has 78.9% recall. They care that FancyCorp saved $2M using your product.

I was advising a construction tech company where we dramatically improved their document retrieval. But the testimonial that closed their next funding round wasn't "They improved our retrieval precision by 40%."

It was: "Their system automatically identifies contractors who haven't signed liability waivers 4 days before they're scheduled on site. This prevented $80,000 in construction delays last month alone."

When marketing your AI, translate your metrics into outcomes people actually care about.

Find the hidden value patterns

The highest-leverage consulting I do isn't technical at all. It's identifying where the actual value is.

At a document management company, we discovered that 60% of questions weren't about document content at all. Users were asking: "Who last modified this document?" They needed contact information to verify decisions.

So we modified every text chunk to include metadata about creation date, modification date, and owner contact info. Suddenly, the most valuable feature wasn't fancy RAG—it was helping people find who to call.

At another client, we found that what users really wanted wasn't answering questions about blueprints. They wanted to count rooms. We could have spent months improving general QA, or we could build a specialized counter. We built the counter.

Your users are telling you what they value. But you have to listen.

Network effects are everything (and AI doesn't have them yet)

Here's the biggest challenge facing AI companies: getting more customers doesn't make your product better. It just creates more opportunities to disappoint people.

Netflix gets better when more people use it. Spotify gets better. Doordash gets better.

But most AI apps? Adding users just adds load without improving quality.

What if having 1,000 users made your recall go up by 20%? You'd have a flywheel. Right now, most companies are just pushing a boulder.

This is why few AI companies have achieved escape velocity. The ones that do figure out how to create virtuous loops where more usage makes their systems smarter in ways users can feel.

The 'Oh shit' moment

You know you've found product-market fit when user growth starts hurting. One client went viral with their AI assistant, hitting 400,000 MRR almost overnight.

The next week? 340,000. Week after? 280,000.

Why? Their product wasn't ready for scale. They blew their first impression with thousands of users who will never come back.

AI products often focus so much on capabilities that they forget fundamentals like stability, speed, and reliability. The most elegant RAG system means nothing if it takes 10 seconds to respond or breaks under load.

Most consultants solve the wrong problem (including me sometimes)

I've fired clients when I realized I couldn't help them. Not because their AI challenges were too hard, but because their organizational challenges were too deep.

When a client books meetings with me at 5am while I'm about to board a flight, that's not an AI problem. When a team is rebuilding the same database integration for the third time, that's not an AI problem.

Sometimes the highest-value thing I can do is tell clients: "Your issue isn't that your AI is bad. It's that your company's processes are broken."

And sometimes the best consulting is just giving smart people permission to trust their instincts. Like when an engineer tells me: "The only way forward is to try random permutations of these parameters to see what works," and I say, "Yes, that's called hyperparameter optimization, and it's exactly what you should be doing."

Lessons learned

The pattern I see across every client is this: companies spend too much time on capabilities and not enough on connections. They build amazing AI features without connecting them to business outcomes.

The best AI isn't the most impressive in a demo. It's the one that solves a real problem customers will pay for.

When a client asks how they're doing, I don't start with technical metrics. I ask: "Are pilots converting to paid? Are usage numbers climbing? Are customers telling their friends?"

Those are the only metrics that matter.

Remember: - Start with the business outcome, work backwards - Measure what matters (hint: it's rarely model performance) - Optimize for learning velocity - Connect your AI directly to problems people will pay to solve

And if all else fails, remember what I tell prospective clients: "If we meet at a party six months from now, I don't want you to say 'Jason helped us improve our AI answers.' I want you to say 'Jason helped us close our Series A.'"

That's the difference between solving AI problems and solving business problems.


If you found this useful, follow me on Twitter @jxnlco for more AI insights. And if you're struggling with your AI strategy, DM me. I might be able to help.

How to Systematically Improve RAG Applications

Retrieval-Augmented Generation (RAG) is a simple, powerful idea: attach a large language model (LLM) to external data, and harness better, domain-specific outputs. Yet behind that simplicity lurks a maze of hidden pitfalls: no metrics, no data instrumentation, not even clarity about what exactly we’re trying to improve.

In this mega-long post, I’ll lay out everything I know about systematically improving RAG apps—from fundamental retrieval metrics, to segmentation and classification, to structured extraction, multimodality, fine-tuned embeddings, query routing, and closing the loop with real user feedback. It’s the end-to-end blueprint for building and iterating a RAG system that actually works in production.

I’ve spent years consulting on applied AI—spanning recommendation systems, spam detection, generative search, and RAG. That includes building ML pipelines for large-scale recommendation frameworks, doing vision-based detection, curation of specialized datasets, and more. In short, I’ve seen many “AI fails” up close. Over time, I’ve realized that gluing an LLM to your data is just the first step. The real magic is how you measure, iterate, and keep your system from sliding backward.

We’ll break everything down in a systematic, user-centric way. If you’re tired of random prompt hacks and single-number “accuracy” illusions, you’re in the right place.

10 “Foot Guns" for Fine-Tuning and Few-Shots

Let me share a story that might sound familiar.

A few months back, I was helping a Series A startup with their LLM deployment. Their CTO pulled me aside and said, "Jason, we're burning through our OpenAI credits like crazy, and our responses are still inconsistent. We thought fine-tuning would solve everything, but now we're knee-deep in training data issues."

Fast forward to today, and I’ve been diving deep into these challenges as an advisor to Zenbase, a production level version of DSPY. We’re on a mission to help companies get the most out of their AI investments. Think of them as your AI optimization guides, they've been through the trenches, made the mistakes, and now we’re here to help you avoid them.

In this post, I’ll walk you through some of the biggest pitfalls. I’ll share real stories, practical solutions, and lessons learned from working with dozens of companies.

Making Money is Negative Margin

In 2020 I had a hand injury that ended my career for 2-3 years. I've only managed to bounce back into being an indie consultant and educator. On the way back to being a productive member of society I've learned a few things:

  1. I have what it takes to be successful, whether that's the feeling of never wanting to be poor again, or some internal motivation, or the 'cares a lot' or the 'chip on the shoulder' - whatever it is, I believe I will be successful
  2. The gift of being enough is the greatest gift I can give myself
  3. I will likely make too many sacrifices by default, not too few, and it will reflect in my regrets later in life

No One Has Potential But Yourself

I had a conversation with my friend today that shook something loose in my head: no one has potential. Like most of the lies I tell myself, this is obviously false - and yet, sometimes we need these extreme statements to see a deeper truth.

We often combat excess pessimism with excess optimism. We see potential in others and believe they can change. But this is just a projection of our own potential and values and beliefs.

Let me explain.

I want to invite my lawyer, Luke, to talk a little bit about the legal side of consulting. If you're new you should also checkout our consulting stack post.

In August, Luke officially launched Virgil. Their goal at Virgil is to be a one-stop shop for a startup’s back office, combining legal with related services that founders often prefer to outsource, such as bookkeeping, compliance, tax, and people operations. We primarily operate on flat monthly subscriptions, allowing startups to focus on what truly moves the needle.

He launched Virgil with Eric Ries, author of The Lean Startup, and Jeremy Howard, CEO of Answer AI. He's able to rely on the Answer AI team to build tools and help him stay informed about AI. He's licensed to practice in Illinois, and they have a national presence. That's his background and the essence of what we're building at Virgil.