Codifying Expertise: Turning Developer Knowledge into AI-Ready Intelligence
For decades, developer education has relied on a familiar triad: books, articles, and tutorials. Structured content authored by experts and consumed by learners. This model, reinforced by hands-on guides and community discussion, created a shared knowledge base that shaped generations of software engineers. Developer advocacy, too, thrived in this world: blog posts, conference talks, GitHub repositories, and open forums drove peer-to-peer knowledge exchange and the spread of best practices.
But with the rise of AI-enhanced development environments, especially the integration of Large Language Models (LLMs) into the software creation process, the landscape of knowledge sharing is undergoing a seismic shift.
The LLM as a Learning Interface
Today, many developers are learning through their tools, not about them. Rather than reading a tutorial, a developer might prompt an AI assistant embedded in their IDE: “Show me how to implement optimistic locking in Hibernate”. The assistant responds in seconds with runnable code, context-specific advice, and even links to relevant APIs or documentation. The friction of discovery is gone. So is much of the need to seek out traditional long-form content.
This marks a fundamental change: knowledge is no longer just external and declarative—it's operationalized, on-demand, and personalized.
What Becomes of Books, Blogs, and Talks?
These formats won’t vanish, but their function is evolving. They are now part of the training substrate for LLMs. The insight from a developer blog, once consumed by a few thousand readers, may now shape the responses of AI tools used by millions. The voice of the expert becomes diffuse, embedded, and algorithmically interpreted.
Yet there’s a tradeoff. When the answer is “just there,” surfaced by a model trained on years of community effort, we risk disconnection from the why and how behind solutions. Tutorials encourage understanding; AI assistants encourage execution.
Developer Skills: Augmented or Atrophied?
It’s fair to ask: are we dumbing down? In the short term, many developers are offloading foundational thinking to LLMs. Syntax recall, boilerplate generation, even architectural patterns—these are now “autocomplete on steroids.” Junior developers may never learn the basics the hard way. Mid-level engineers might stagnate. Senior developers, if not vigilant, may lose their sharpness.
But the long-term view is more complex. AI assistants can accelerate learning if used consciously. Developers who engage with the model—questioning, refining, and understanding its output—stand to grow faster. The critical skill shifts from knowing to evaluating: Can I trust this suggestion? Is this code correct, secure, maintainable?
Knowledge Sharing in the New Paradigm
If AI assistants are the new interface to knowledge, how do we continue sharing best practices?
Custom LLMs: From Hype to Necessity
Enterprises and teams already ask: Should we train our own LLM? The better question is: How can we control the knowledge environment our developers rely on?
For some, this means fine-tuning models on internal documentation, past incidents, and legacy systems. For others, it’s aligning off-the-shelf models with internal best practices using RAG (Retrieval-Augmented Generation). Either way, knowledge becomes an asset to embed—not just publish.
The more developers rely on AI tools, the more important it is that these tools reflect the right knowledge, not just what’s publicly available.
An emerging open source initiative that could bridge the gap between developer expertise and AI model behavior is InstructLab. Unlike passive ingestion pipelines, InstructLab enables developers to directly shape open-source LLMs through structured, version-controlled contributions. By allowing domain experts to submit skills, concepts, and evaluations in plain text or YAML—without retraining the base model—InstructLab provides a powerful framework for encoding and refining organizational knowledge. This aligns perfectly with the evolving role of GitHub repositories: not just as code storage, but as structured, testable knowledge units for both developers and machines. In this new model, a pull request is not just a contribution to a project—it's an update to the memory and reasoning capability of AI systems used across the organization.
GitHub Repositories: From Human-Readable to Machine-Consumable Knowledge
For years, GitHub repositories have served as informal documentation platforms, how-to guides, and code-sharing ecosystems. A well-maintained repo often includes:
Recommended by LinkedIn
This worked well for human readers. But in an AI-driven world, the next evolution is clear: repositories must also become first-class machine-readable knowledge sources.
Gitingest and GitMCP: Signposts for the Future
Tools like Gitingest and GitMCP illustrate this transition. These frameworks aim to structure and contextualize repositories for ingestion into LLMs or developer assistants.
These tools shift the role of the repo from “passive archive” to “active participant” in the knowledge delivery process.
What Should the Ideal AI-Ready Repository Look Like?
To serve both humans and LLMs effectively, we need to design repositories that are structured, annotated, and purposeful. Below are key components of a machine-optimized knowledge repo:
1. Structured Content Layout
2. Semantic Metadata
3. LLM Ingestion Markers
4. Behavioral Definitions (via MCP)
5. Executable Knowledge
Final Thoughts
We are at an inflection point. Developer knowledge is no longer just written, read, and remembered—it is encoded, embedded, and executed. This unlocks massive productivity gains but demands a rethinking of how we learn, share, and preserve expertise.
As developer advocates, educators, and tool builders, we must embrace this shift not with nostalgia for old formats, but with intention. The mission remains: enable developers to build better software. The mediums and methods, however, are evolving—and fast.
We are no longer just writing for developers. We’re writing for the models that will teach them.
Ignacio Riesgo 👆
DevX Specialist Solution Architect at Red Hat
1moThanks Markus for this inspiring article. In contradiction to this, I was lately thinking about what will actually happen when AI models get more and more trained by content generated with AI. Compared to the shrinking amount of new content produced by humans, it might simply end in a kind of stagnation for getting smarter models. Your insights offers a way out of this circle.