User Tools

Site Tools


ai:developer-in-the-ai-decade

The 3 Ways to Develop in the AI Decade

The software development landscape has fundamentally shifted. AI is no longer a buzzword — it's the infrastructure. For developers who want to stay relevant and thrive in this decade, there are three clear paths to pursue. Each builds on the last, and together they form a complete strategy for the modern developer.


Way 1: LLM Models — Build, Fine-Tune, and Leverage Open Source

At the foundation of everything happening in AI today are Large Language Models. Understanding how they work — and even building or customizing your own — is the first major skill path.

Build Your Own (From Scratch or Fine-Tuned)

Learn the architecture: Transformers, attention mechanisms, tokenization, and training loops. Fine-tune existing models on domain-specific data — your company's codebase, legal documents, medical records — using frameworks like PyTorch and Hugging Face Transformers. Tools like LoRA and QLoRA make efficient fine-tuning possible on consumer hardware.

This path is ideal for developers who want to work at AI companies, build AI products, or deeply understand what's under the hood.

Leverage Open Source Models

The open-source LLM ecosystem is exploding. Models like LLaMA, Mistral, Gemma, Qwen, DeepSeek, and Phi are freely available and increasingly competitive with proprietary options. You can run them locally using tools like Ollama, LM Studio, or vLLM — no cloud costs, full data privacy.

On the API side, you can access Claude (Anthropic), GPT (OpenAI), Gemini (Google), and DeepSeek, switching between them based on cost, quality, and task requirements.

Why This Matters

Understanding LLMs at this level lets you make informed decisions about which models to use, when to fine-tune vs. use off-the-shelf, and how to optimize cost and performance. You stop being a consumer and start being a builder.


Way 2: AI Agents + RAG + Memory — The New Development Stack

The real power of LLMs is unlocked when you combine them with Agents, Retrieval-Augmented Generation (RAG), and Memory systems. This is where AI moves from “chatbot” to “autonomous worker.”

AI Agents

An AI Agent is an LLM that can take actions: read files, write code, run commands, call APIs, browse the web, and make decisions in a loop. The tooling here is maturing fast, and your choice depends on what you're building:

  • Claude Code (Anthropic) — A command-line agentic coding tool. It reads your codebase, writes code, runs tests, and handles complex multi-file changes autonomously. Best for developers who want a powerful out-of-the-box agent for coding workflows.
  • GitHub Copilot — Integrated into your IDE with real-time code suggestions and chat-based assistance. Best for developers who want AI embedded directly in their editor with minimal setup.
  • Agentic frameworks (LangChain, CrewAI, AutoGen, OpenAI Agents SDK) — For building custom agents that chain multiple LLM calls with tool usage. Best when you need agents tailored to non-coding workflows or complex multi-step pipelines.

A practical decision framework: if you're coding, start with Claude Code or Copilot. If you're building agents for end users or automating business processes, reach for an agentic framework.

RAG (Retrieval-Augmented Generation)

LLMs have a knowledge cutoff and can hallucinate. RAG solves this by connecting the LLM to your own data sources. Your documents are split into chunks, converted into vector embeddings, and stored in a vector database. When the user asks a question, relevant chunks are retrieved and injected into the LLM's prompt as context.

For vector databases, the landscape breaks down like this:

  • Pinecone — Fully managed, minimal ops overhead. Best for teams that want to move fast without managing infrastructure.
  • Weaviate — Open source with hybrid search (vector + keyword). Best for projects that need flexibility and self-hosting.
  • Chroma — Lightweight, developer-friendly. Best for prototyping and smaller projects.
  • pgvector — Postgres extension. Best when you already run Postgres and want to avoid adding another database to your stack.

With RAG, your AI agent can answer questions about internal docs, your codebase, your database — anything you feed it.

Memory

By default, LLMs are stateless — every conversation starts from zero. Memory systems fix this by adding short-term memory (conversation history within a session) and long-term memory (persisted preferences, past decisions, and learned context across sessions).

Tools like Mem0, LangChain Memory, and custom database-backed solutions enable agents that remember your project, your preferences, and your codebase over time.

Why This Matters

This is the layer where AI becomes genuinely useful for software development. An agent with RAG and memory doesn't just autocomplete your code — it understands your project, retrieves the right context, and makes decisions that align with your architecture.


Way 3: Using LLMs and AI Agents to Cut Costs in Software Development

The most practical path for most developers: use AI as a productivity multiplier. The goal here isn't to build AI — it's to use AI to ship software faster, cheaper, and with fewer people.

Accelerated Development Cycles

Tasks that took hours now take minutes: boilerplate generation, CRUD APIs, database migrations, test writing, documentation. AI agents can handle entire workflows — read a spec, create the API endpoints, write the tests, and generate the documentation.

Reduced Team Size for the Same Output

A single developer with Claude Code or Copilot can produce the output that previously required a small team. This is particularly impactful for startups and small companies: you can build production-grade software with 1–3 developers instead of 5–10.

Lower Error Rates and Faster Debugging

AI catches bugs, suggests fixes, and can scan your codebase for inconsistencies. Test generation alone can save dozens of hours per sprint.

Documentation and Knowledge Transfer

AI generates documentation from code, creates onboarding materials, and keeps READMEs up to date — tasks that teams traditionally neglect but that compound in value over time.

Why This Matters

This path has the highest immediate ROI. You don't need to understand transformer architecture. You need to know how to prompt effectively, choose the right tools, and integrate AI into your daily workflow.


What Actually Drives Developer Productivity in the AI Era

After months of hands-on experience building with AI tools across real projects, I've developed a mental model for what actually determines results. This isn't based on formal research — it's a framework drawn from building, shipping, and iterating with these tools daily.

The Productivity Model

Factor Weight Why
Developer Skill Level ~60% Your fundamentals still matter most. Architecture, design patterns, debugging, system thinking — AI amplifies skill, it doesn't replace it. A senior developer with AI tools will outperform a junior developer with the same tools by a wide margin.
Tools You Use ~30% Whether you use AI tools at all matters enormously. A developer using Claude Code, Copilot, or any capable AI coding tool will dramatically outperform one who doesn't.
Skill in Using AI ~10% This is the good news. Learning to use AI effectively — prompting, context management, workflow integration — is not hard. Most developers can reach competency in 2–3 weeks. There are many official guides, and the learning curve is gentle.

The key insight: your existing developer skills are the foundation. AI tools are the multiplier. The “AI skill” itself is the easiest part to learn — the bottleneck is your engineering fundamentals and your tool selection.


How to Improve: A Practical Strategy

1. Use Multiple LLM Providers

Don't lock yourself into one provider. Each has distinct strengths, and knowing when to use which one is a real competitive advantage:

  • Claude (Anthropic) — Excellent for long-context reasoning, careful analysis, and code generation with nuance. My go-to for complex refactoring and architecture decisions.
  • GPT (OpenAI) — Strong general-purpose model with the widest ecosystem of integrations and plugins.
  • Gemini (Google) — Large context windows and strong multimodal capabilities. Useful for tasks involving images, audio, or very long documents.
  • DeepSeek — Competitive open-weight model, particularly strong at coding tasks. A cost-effective alternative for many use cases.
  • Local/Open Source (Ollama + Gemma, Qwen, LLaMA) — Free, private, and increasingly capable. Run locally for sensitive projects or to eliminate API costs entirely.

The rule of thumb: use the cheapest model that gets the job done, and switch providers based on the task.

2. Build Your Own Tools

Relying solely on commercial tools limits you. Two approaches:

Build from scratch — Create custom AI agents tailored to your specific workflow. You get full control, deep understanding, and zero vendor dependency. The upfront cost is higher, but the investment compounds.

Build from open source — Projects like OpenCode (an open-source Claude Code alternative), Aider, and Continue.dev give you community-driven foundations to fork, customize, and extend. You get speed-to-start with the ability to diverge as your needs grow.

Either way, building your own tooling means you understand exactly what's happening, you can optimize for your use cases, and you're not locked into any single vendor's roadmap.


3. Demo: AI Agent Built from Scratch

To put this into practice, I built a custom AI agent that handles three core development tasks. Rather than describe it, I'll show you:

[Video Demo]

The video walks through the agent performing all three tasks on a real project — you'll see exactly how it works, where it excels, and where it still has rough edges.

What the Agent Does

Document Creation — The agent reads project requirements, existing code, and architecture decisions, then generates comprehensive documentation: API docs, README files, architecture overviews, and onboarding guides. It keeps documentation in sync as the codebase evolves.

Coding — Given a specification, the agent generates complete API endpoints with routing, validation, error handling, and database integration. It follows the project's existing patterns and conventions, learned through RAG over the codebase.

Test Writing — The agent analyzes existing code and generates unit tests, integration tests, and edge case tests. It detects the testing framework already in use (Jest, Pytest, etc.) and follows the project's testing conventions to produce meaningful coverage — not just superficial tests.

This demo is proof that you don't need expensive commercial tools to get real productivity gains. A well-built custom agent, powered by open-source models or API access, can automate significant portions of the development workflow.


Conclusion

The three paths for developers are clear:

  1. Understand LLMs — Know what's under the hood so you can make smart decisions about models, cost, and architecture.
  2. Build with Agents + RAG + Memory — This is the new application layer, and it's where the most interesting engineering problems live.
  3. Use AI to ship faster and cheaper — This is where the immediate value is, and it's accessible to every developer today.

The developers who combine strong fundamentals with the right AI tools — and who invest in building their own tooling rather than depending entirely on commercial products — will have an enormous advantage in the years ahead.

ai/developer-in-the-ai-decade.txt · Last modified: by phong2018