Saturday, October 4, 2025

From Plan to PR: The Agentic Data Mesh Spec is Ready for Review

First Commit

​Yesterday, I laid out the plan. Today, it's in version control.

​The first official pull request for the Agents Running a Data Mesh Demo is now open. This PR is focused entirely on one thing: getting the initial specification right.

​Titled "docs(spec): Create initial spec for agentic data mesh demo," this PR introduces the foundational document that will guide the project. Before a single piece of infrastructure is provisioned, this is our chance to align on the "what" and the "why."

​Inside, you'll find the high-level plan, which establishes two core concepts:

  1. The Contextual Environment: The entire demo will run on a foundation of standard Confluent Cloud topics, providing a realistic, event-driven source for our agents to work with.
  2. The High-Level Agent Workflow: The spec outlines the three key phases for the agents: first they discover potential data products within the topics, then they analyze them for value and structure, and finally they define them as official, usable products in the mesh.

​This workflow will be carried out by our team of specialized agents—Discovery, Ideation, Creation, and Communication—with the critical final step of access control managed by a human in the loop.

​Your Feedback is Needed

​I'm opening this PR to get your feedback on the spec itself. This is about building in the open and making sure the foundation is solid before we start laying bricks.

  • ​Is the scope clear?
  • ​Does the discover, analyze, and define workflow cover the necessary steps?
  • ​Are we starting with the right contextual environment?

​Head over to the PR, review the initial spec, and help shape the direction of this project.

​➡️ Review the spec and leave your thoughts here: PR #1: The Initial Spec

Friday, October 3, 2025

From Chaos to Clarity: Using AI Agents to Build a Data Mesh

In today's enterprise, we're surrounded by a strange paradox. We are drowning in data yet starving for insights. The promise of being "data-driven" is often blocked by a simple reality: our data is siloed in different systems, its quality is questionable, and accessing it is a high-friction process that slows everyone down. We have the raw materials, but we can't seem to build anything meaningful with them.

​For years, we tried to solve this by centralizing everything into massive data lakes or warehouses. While well-intentioned, this often created a new bottleneck: a single, overwhelmed central team responsible for a universe of data they couldn't possibly be experts in.

​There has to be a better way.

​What is a Data Mesh?

​A few years ago, a new paradigm emerged that flipped the old model on its head: the Data Mesh.

​The core idea, pioneered by Zhamak Dehghani, is simple but profound: stop treating data as a byproduct shuffled into a central repository. Instead, treat data as a product.

​This means the teams that are closest to the data—the domain experts who actually create and understand it—are responsible for owning it and serving it to the rest of the organization. The "Sales" team owns and serves "Customer Account" data. The "Logistics" team owns and serves "Shipment Status" data. Each domain provides its data as a reliable, well-documented, and easy-to-use product. This approach is built on four key principles:

  1. Domain Ownership: Responsibility for data shifts from a central team to the teams that know it best.
  2. Data as a Product: Data is treated with the same care as a software product, with a focus on user experience, quality, and lifecycle management.
  3. Self-Serve Data Platform: A common platform empowers domain teams to build and share their data products easily.
  4. Federated Computational Governance: A set of global rules and standards ensures the entire mesh is secure, interoperable, and trustworthy.

​The Data Mesh is a brilliant sociotechnical concept. But in practice, the technical execution can place a heavy burden on domain teams. Asking every team to become expert data engineers is a tall order. So, how do we lower the barrier to entry?

​Where Do AI Agents Fit In?

​This is where the next frontier of software development comes into play: Agentic AI.

​Instead of asking our developers to manually handle all the repetitive and complex tasks of creating a data product, we can delegate that work to autonomous AI agents. These agents can act as tireless assistants, automating the undifferentiated heavy lifting and allowing humans to focus on what truly matters.

​Imagine what this looks like:

  • The Data Scout Agent: A developer points the agent at a database. The agent automatically profiles the schemas, identifies sensitive PII, analyzes query patterns, and suggests potential "data products" based on usage.
  • The Quality Guardian Agent: Once a data product is defined, this agent automatically generates data quality tests based on the data's statistical profile. It continuously monitors the data for anomalies or drift, filing a ticket or alerting the owners when standards slip.
  • The Product Manager Agent: This agent handles the "product" aspects. It generates clear documentation, creates sample queries, and even builds the self-service access layer, like a secure GraphQL API endpoint for the data product.

​In this world, the developer's role shifts from being a data plumber to being a manager of agents. Their job becomes about defining the goals and constraints, while the agents handle the meticulous implementation.

​This is Just the Beginning

​By combining the organizational clarity of a Data Mesh with the technical leverage of AI agents, we can finally start to resolve the paradox we began with. We can empower teams to share their knowledge as high-quality data products, creating an ecosystem where discovering and using data is a creative, low-friction experience.

​In my first post, I talked about a "Contribution First" mindset. For me, that means not just exploring ideas but also building and sharing the tools that bring those ideas to life. Talking about this is one thing, but contributing a solution is another.

​This is a topic I'm incredibly passionate about, and this post is just the beginning. Keep an eye out for a formal PR tomorrow, where I'm excited to take the next step and share more about this work with the community.

Thursday, October 2, 2025

From Prototype to Production: The "Contribution First" Workflow

How often does this happen? A product team validates an exciting new concept with a sleek prototype. Users love it. Stakeholders are bought in. The validated design is handed over to engineering. Weeks later, the team tries to merge the "finished" feature back into the main repository, only to be met with a wall of merge conflicts and architectural drift.

​The disconnect between a validated idea and its technical implementation can derail even the best-laid plans. This is because we often treat contribution as the last step in the development process.

​We can fix this by bridging that gap with a "Contribution First" workflow.

​The principle is simple: Once a prototype is validated, the very first engineering step should be opening a pull request against the repository you intend to contribute to.

The Old Way: The Design-to-Engineering Chasm

  1. ​Prototype an idea and get user validation.
  2. ​Hand off specs to engineering.
  3. ​Engineers fork the repository or create a long-lived branch.
  4. ​Develop the feature in isolation.
  5. ​Open a massive Pull Request weeks later.
  6. ​Spend days resolving conflicts and responding to feedback that requires major changes, losing the momentum from the prototype.

The New Way: Contribution First

  1. ​Prototype an idea and get user validation.
  2. Immediately open a draft Pull Request against the main branch. This PR's description links directly to the prototype and user feedback. The initial commit might just be a placeholder or a failing test that represents the new feature.
  3. ​This PR now becomes the single source of truth for the work, connecting the why (the prototype) with the how (the implementation).
  4. ​The team iterates on the feature inside this contribution framework, with constant visibility.

Why This Changes Everything

​Adopting a "Contribution First" mindset directly after prototyping brings incredible benefits:

  • Connects Product and Engineering: The PR becomes a living document that directly ties the implementation back to the validated design goals.
  • Eliminates Big-Bang Merges: Integration happens continuously, not as a painful final step.
  • Fosters Early Technical Feedback: Maintainers can provide feedback on architecture in the context of the prototype's goals from day one.
  • Maintains Momentum: The excitement from a successful prototype flows directly into implementation, without the "dead air" of siloed development.

​It’s a small shift in process that creates a seamless path from idea to production. You're no longer just building something for the project; you're building it with the project, from the very first commit.

Wednesday, October 1, 2025

Starting a Content Journey with a 'Contribution First' Mindset

Welcome to the start of my content journey!

For my first post, I want to share the strategy that's going to guide this blog and my social media presence. Many people think of content creation as a linear process: draft a long-form article, then try to chop it up for Twitter and LinkedIn. But I'm taking a different approach—one I'm calling the "Contribution First" workflow.

The idea is simple: the most valuable content isn't "created," it's "documented." As I work on my projects, build a new demo, or just solve a tricky bug, I'm going to capture those genuine insights in a private "thought bank." This could be a simple text file, a markdown doc, or even a personal Git repo.

From there, the content is translated for each platform:
- Twitter gets the atomic idea: a quick tip, a single line of code, or a surprising tool discovery that can be shared instantly.
- LinkedIn gets the professional lesson: a more structured reflection on a challenge or a broader insight about a technical pattern.
- This blog becomes the long-form archive: where I'll combine a series of these small insights into a comprehensive article, like a deep dive into a project or a guide to a new workflow.

This feels more authentic and sustainable than a traditional content schedule. It ensures that everything I share is a direct result of my real-world work.

So, this is my commitment: to share the raw, unedited, and valuable parts of my process, not just the polished final products. I hope you'll follow along with this journey.

From Plan to PR: The Agentic Data Mesh Spec is Ready for Review

First Commit ​Yesterday, I laid out the plan. Today, it's in version control. ​The first official pull request for the Agents Running a...