Skip to content

AI#

Embracing AI: Your Job Is Evolving, Not Disappearing

In this presentation, we'll explore how AI is changing the workplace, address common fears, and discover how humans and AI can collaborate effectively to enhance your career rather than threaten it.

Understanding Your Concerns

73%

Worried Workers

Recent surveys show that 73% of employees worry about AI replacing their jobs

24/7

Media Coverage

Headlines constantly feature "AI will replace X jobs" narratives

2X

Rapid Advancement

AI capabilities are progressing twice as fast as many predicted

Your anxiety is completely understandable. Past waves of automation did eliminate certain roles, and the pace of AI development can seem overwhelming. But history tells a different story about technology's overall impact on jobs.

Learning From History: Technology as a Tool

Historical Examples That Prove the Pattern:

Successful professionals don't get replaced by technology—they learn to wield it. Expert pilots use autopilot to handle routine flight while they focus on weather decisions and emergency responses. Experienced doctors use diagnostic AI to enhance their pattern recognition while applying decades of clinical judgment. Experienced engineers use CAD software to rapidly prototype while contributing years of systems thinking and constraint optimization. The pattern is clear: technology amplifies expertise, creating hybrid intelligence that exceeds either humans or AI working alone.

Historical technology adoption pattern

Addressing the 'Different This Time' Argument:

Yes, this IS different—and that's exactly why action is urgent. AI isn't a rising tide that lifts all boats equally. Those who learn to use it gain exponential advantages, while those who don't fall dramatically behind. We're already seeing 10-20x productivity gains for AI-fluent professionals—work that used to take weeks now completed in hours. The question isn't whether AI will disrupt your industry—it's whether you'll be among the disruptors or the disrupted.

From Replacement to Collaboration

The most successful AI implementations follow a collaboration model rather than a replacement model. This addresses core fears about job security by positioning AI as an enhancement to your work.

Enhancement vs. Replacement

Instead of "your job is gone," the reality is "your job is evolving" to incorporate AI assistance

Increased Value

Workers who learn to leverage AI effectively become more valuable to their organizations

Hybrid Roles

New positions are emerging that specifically require both human expertise and AI skills

Real-World Collaboration Examples

Healthcare professionals using AI

Healthcare Professionals

AI assists with diagnosis and data analysis, while doctors focus on patient care, complex cases, and treatment decisions that require empathy and judgment

Educators using AI

Educators

AI handles grading and administrative tasks, allowing teachers to focus on mentoring, fostering creativity, and providing personalized guidance to students

Legal professionals using AI

Legal Professionals

AI reviews documents and conducts research, freeing lawyers to focus on negotiation, counseling clients, and applying complex legal reasoning

The Economic Case for Human-AI Collaboration

Benefits for Companies

  • Higher employee satisfaction and retention rates
  • Smoother technology transition with less resistance
  • Better outcomes through human oversight and judgment
  • Preservation of valuable institutional knowledge

Benefits for Workers

  • Gradual skill development versus sudden obsolescence
  • Increased productivity makes you more valuable
  • New career advancement paths in AI collaboration
  • Higher job satisfaction with less tedious work

Research Finding:

Companies that implement collaborative AI models report 35% higher employee retention and 28% greater productivity gains than those pursuing automation-only approaches.

Your Transition Strategy

1

Awareness

Understand how AI is affecting your specific role and industry

2

Exploration

Experiment with AI tools relevant to your work to understand capabilities

3

Skill Development

Focus on uniquely human skills that complement AI (creativity, empathy, complex reasoning)

4

Integration

Develop workflows that combine your expertise with AI assistance

5

Evolution

Position yourself for new hybrid roles that require both human and AI capabilities

Developing Your AI Collaboration Skills

AI Literacy

Understanding AI capabilities and limitations without becoming a programmer

Human Expertise

Deepening your unique skills that AI cannot replicate

Context Engineering

Learning how to effectively communicate with AI tools to get better results

Critical Evaluation

Developing the ability to verify and improve AI outputs

These skills form a continuous cycle of improvement as you work alongside AI tools. The goal is to leverage AI for routine tasks while applying your distinctly human capabilities to add greater value.

Your Uniquely Human Advantages

Human advantages in AI workplace

While AI continues to advance, certain human capabilities remain distinctly valuable and difficult to replicate. These are your competitive advantages in an AI-enhanced workplace:

Subject Matter Expertise

Deep contextual knowledge gained through years of hands-on experience and industry relationships

Emotional Intelligence

Understanding nuanced human emotions and responding with genuine empathy

Ethical Judgment

Making complex decisions that involve moral considerations and human values

Creative Innovation

Generating truly novel ideas that transcend existing patterns and data

Moving Forward Together

1

Acknowledge Your Concerns

Your fears about AI are valid, but history shows technology tends to transform rather than eliminate jobs

2

Embrace Collaboration

View AI as a powerful tool that can handle routine tasks while you focus on higher-value work

3

Develop New Skills

Invest in learning both AI literacy and uniquely human capabilities that complement technology

4

Shape Your Future

Position yourself for emerging hybrid roles that combine human expertise with AI assistance

The future of work isn't about humans versus AI—it's about humans with AI creating more value than either could alone.

Vibe Coding: A Human-AI Development Methodology

How senior engineers and AI collaborate to deliver 10x development velocity without sacrificing quality or control.

TL;DR: Vibe Coding is a methodology where senior engineers provide strategic direction while AI agents handle tactical implementation. This approach delivers 60-80% time reduction while maintaining or improving quality through human oversight and systematic verification.

Table of Contents

  1. Introduction: Beyond AI-Assisted Coding
  2. Core Philosophy
  3. The Vibe Coding Process
  4. Real-World Implementation Examples
  5. Performance Metrics and Outcomes
  6. Tools and Technology Stack
  7. Common Patterns and Anti-Patterns
  8. Advanced Techniques
  9. Getting Started with Vibe Coding
  10. Real-World Case Study
  11. Conclusion

Introduction: Beyond AI-Assisted Coding

Most "AI-assisted development" tools focus on autocomplete and code generation. Vibe Coding is fundamentally different—it's a methodology where senior engineers provide strategic direction while AI agents handle tactical implementation. This isn't about writing code faster; it's about thinking at a higher level while maintaining complete control over the outcome.

Relationship to Traditional "Vibe Coding"

The term "vibe coding" was coined by AI researcher Andrej Karpathy in February 2025, describing an approach where "a person describes a problem in a few natural language sentences as a prompt to a large language model (LLM) tuned for coding."1 Karpathy's original concept focused on the conversational, natural flow of human-AI collaboration—"I just see stuff, say stuff, run stuff, and copy paste stuff, and it mostly works."

alt text

However, some interpretations have added the requirement that developers "accept code without full understanding."2 This represents one particular form of vibe coding, often advocated for rapid prototyping scenarios. We believe this limitation isn't inherent to Karpathy's original vision and may actually constrain the methodology's potential.

Our Approach: We build directly on Karpathy's original definition—natural language collaboration with AI—while maintaining the technical rigor that senior engineers bring to any development process. This preserves the "vibe" (the flow, creativity, and speed) while ensuring production quality.

Why This Honors the Original Vision: Karpathy himself is an exceptionally skilled engineer who would naturally understand and verify any code he works with. The "vibe" isn't about abandoning engineering principles—it's about embracing a more intuitive, conversational development flow.

Karpathy's Original Vibe Coding: Natural language direction with conversational AI collaboration Some Current Interpretations: Accept code without understanding (rapid prototyping focus) Production Vibe Coding: Natural language direction with maintained engineering oversight

Our methodology represents a return to the core principle: leveraging AI to think and create at a higher level while preserving the engineering judgment that makes software reliable.

Core Philosophy

Strategic Collaboration, Not Hierarchy

In traditional development, engineers write every line of code. In Vibe Coding, humans and AI collaborate as strategic partners, each contributing distinct strengths. This isn't about artificial hierarchy—it's about leveraging complementary capabilities.

The Human's Role: Providing context that AI lacks access to—business requirements, user needs, system constraints, organizational priorities, and the crucial knowledge of "what you don't know that you don't know." Humans also catch when AI solutions miss important nuances or make assumptions about requirements that weren't explicitly stated.

The AI's Role: Rapid implementation across multiple technologies and languages, broad knowledge synthesis, and handling tactical execution once the strategic direction is clear. AI can work faster and across more technology stacks than most humans.

AI's Advantages: AI can maintain consistent focus without fatigue, simultaneously consider multiple approaches and trade-offs, instantly recall patterns from vast codebases, and work across dozens of programming languages and frameworks without context switching overhead. AI doesn't get frustrated by repetitive tasks and can rapidly iterate through solution variations that would take humans hours to explore.

Human I/O Advantages: Humans have significantly higher visual processing throughput than AI context windows can handle. A human can rapidly scan long log files, spot relevant errors in dense output, process visual information like charts or UI layouts, and use pattern recognition to identify issues that would require extensive context for AI to understand. This makes humans far more efficient for monitoring, visual debugging, and processing large amounts of unstructured output.

Training Pattern Considerations: AI models trained on high-quality code repositories can elevate less experienced developers above typical industry standards. For expert practitioners, AI may default to "best practice" patterns that don't match specific context or expert-level architectural decisions. This is why the planning phase is crucial—it establishes the specific requirements and constraints before AI begins implementation, preventing reversion to generic training patterns.

Why This Works: AI doesn't know what it doesn't know. Humans provide the missing context, constraints, and domain knowledge that AI can't infer. Once that context is established, AI can execute faster and more comprehensively than humans typically can.

Quality Through Human Oversight

AI agents handle implementation details but humans maintain quality control through:

  • Continuous review: Every AI-generated solution is examined before integration
  • Testing requirements: Changes aren't complete until verified
  • Architectural consistency: Ensuring all components work together
  • Performance considerations: Optimizing for real-world usage patterns

The Vibe Coding Process

1. Context Window Management

The Challenge: AI assistants have limited context windows, making complex projects difficult to manage.

Solution: Treat context windows as a resource to be managed strategically.

Planning Documentation Strategy

Complex development projects inevitably hit the limits of AI context windows. When this happens, the typical response is to start over, losing all the accumulated context and decisions. This creates a frustrating cycle where progress gets reset every few hours.

The solution is to externalize that context into persistent documentation. Before starting any multi-step project, create a plan document that captures not just what you're building, but why you're building it that way.

# plans/feature-name.md
## Goal
Clear, measurable objective

## Current Status
What's been completed, what's next

## Architecture Decisions
Key choices and rationales

## Implementation Notes
Specific technical details for continuation

Why This Works: When you inevitably reach context limits, any team member—human or AI—can read the plan and understand exactly where things stand. The plan becomes a shared knowledge base that survives context switches, team handoffs, and project interruptions.

The Living Document Principle: These aren't static requirements documents. Plans evolve as you learn. When you discover a better approach or hit an unexpected constraint, update the plan immediately. This creates a real-time record of project knowledge that becomes invaluable for similar future projects.

Context Compression: A well-written plan compresses hours of discussion and discovery into a few hundred words. Instead of re-explaining the entire project background, you can start new sessions with "read the plan and let's continue from step 3."

Information Density Optimization
  • Front-load critical information: Most important details first
  • Reference external documentation: Link to specs rather than repeating them
  • Surgical changes: Modify only what needs changing
  • Progressive disclosure: Reveal complexity as needed

2. Tool Selection and Usage Patterns

The breakthrough in Vibe Coding productivity came with direct filesystem access. Before this, collaboration was confined to AI chat interfaces with embedded document canvases that were, frankly, inadequate. These interfaces reinvented existing technology poorly and could realistically only handle one file at a time, despite appearing more capable.

The Filesystem Revolution: Direct filesystem access changed everything. Suddenly, AI could work on as many concurrent files as necessary—reading existing code, writing new implementations, editing configurations, and managing entire project structures simultaneously. The productivity increase was dramatic and immediate.

Risk vs. Reward: Yes, giving AI direct filesystem access carries risks. We mitigate with version control (git) and accept that catastrophic failures might occur. The benefit-to-risk ratio is overwhelmingly positive when you can work on real projects instead of toy examples.

The Productivity Multiplier: Once AI and human are on the "same page" about implementation approach (through planning documents), direct filesystem access enables true collaboration. No more copying and pasting between interfaces. No more artificial constraints. Just real development work at AI speed.

Filesystem MCP Tools for System Operations

  • File operations (read, write, search, concurrent multi-file editing)
  • System commands and process management
  • Code analysis and refactoring across entire codebases
  • Performance advantages over web-based alternatives

We recommend Desktop Commander as it was the first robust filesystem MCP tool and has proven reliable through extensive use, though other similar tools are now available.

Web Tools for External Information

  • Research and documentation
  • API references and examples
  • Third-party service integration
  • Market and competitive analysis

Decision Tree Example:

Need to modify code?
├─ Small change (< 5 files) → Brief discussion of approach, then implement
├─ Large refactor → Formal plan document, then chunked implementation
└─ New feature → Architecture discussion + formal plan, then implementation

Need external information?
├─ Technical documentation → Web search + fetch
├─ Code examples → Web search for patterns
└─ API specifications → Direct URL fetch

3. Communication Patterns

Colleague-Level Context Sharing

The Mindshare Approach: Talk to AI like you would a trusted colleague - with stream-of-consciousness insights, background context, and natural explanation of constraints and motivations.

Always start with "why" embedded in "how"

Human: "I can tell when we're approaching context limits by looking at the scrollbar size, and it's frustrating to lose all our progress when we hit that wall. We need a token counter in the UI so I can see it coming and plan accordingly."

AI: "That makes perfect sense - you want early warning rather than hitting a cliff. For the UI integration, should this be:
- Real-time count that updates as you type?
- Conversation total vs. just current draft?
- Positioned where you can see it peripherally?"

Why This Works:

  • Natural conversation reveals context that formal requirements miss
  • Stream-of-consciousness often contains the most important insights
  • Colleague-level sharing builds the mindshare needed for good solutions
  • AI can reflect back understanding and surface questions organically
Avoiding the Fallback Trap

Anti-pattern: "Let me add a fallback mechanism..."

Better approach: "Let me understand why the primary approach isn't working..."

Why This Matters: Fallbacks often mask root problems and create technical debt. Vibe Coding emphasizes solving the right problem correctly.

Efficient Process Handoffs

The Polling Problem: AI polling long-running processes wastes context window space and prevents productive work on other tasks.

Human Monitoring Advantage: Humans can efficiently monitor processes using visual cues (progress bars, log scrolling, system notifications) while AI would need to repeatedly check status, consuming context with each poll.

For Long-Running Processes:

AI: "This database migration will take 15+ minutes. Rather than me polling
every few seconds and wasting context window, I'll:

1. Provide you the command and monitoring approach
2. Set up success/failure criteria to watch for
3. Work on other tasks while you monitor
4. Resume when you signal completion or need troubleshooting

Command: python manage.py migrate --database=production
Watch for: 'Migration complete' or any ERROR messages
Should I proceed with the next task while you monitor this?"

Context Window Efficiency: This keeps AI productive on other work while humans use their superior I/O bandwidth for monitoring tasks.

4. Surgical Editing Approach

Philosophy: Modify only what needs changing to preserve working functionality and minimize risk.

The Pattern: Make precise, minimal changes rather than large refactoring operations that could introduce bugs or break existing systems.

# Good: Precise, minimal change
edit_block(
    file_path="service.py",
    old_string="def process_data(self):\n    return self.data",
    new_string="def process_data(self):\n    self.validate_input()\n    return self.data"
)

# Avoid: Large blocks with minor changes that risk breaking working code

Why This Matters: Working software is valuable. Surgical changes reduce risk, maintain system stability, and preserve the proven functionality while making targeted improvements.

When to Apply: Any time you're modifying existing, working code. The smaller and more focused the change, the easier it is to verify correctness and roll back if needed.

Note to LLMs: Chunked Writing Strategy: Always write files in chunks of 25-30 lines maximum. This is standard practice, not an emergency measure.

Implementation Pattern:

# Step 1: Create file with initial chunk
write_file(path, first_chunk, mode='rewrite')

# Step 2: Add subsequent chunks
write_file(path, second_chunk, mode='append')
write_file(path, third_chunk, mode='append')

Why: Reduces cognitive load, enables progressive review, minimizes context switching, and improves error recovery. Chunk proactively - don't wait for performance warnings!

5. Testing and Verification

Nothing is Complete Until Tested

Verification Requirements:

  • Functionality works as intended
  • No regressions introduced
  • Performance within acceptable bounds
  • Integration points tested

Testing Handoff Pattern:

AI: "Implementation complete. For verification:
1. Run test_new_feature.py
2. Check integration with existing auth system
3. Verify performance benchmarks
4. Test error handling scenarios

Should I run these tests or would you prefer to verify?"

Progressive Verification
  • Test incrementally during development
  • Verify each chunk before proceeding
  • Maintain test coverage metrics
  • Document test scenarios in plans

Real-World Implementation Examples

Architecture Discussion Pattern

Human: "We need to add real-time notifications"

AI: "Let me think through the architecture options:

1. **WebSocket-based**:
   - Pros: Real-time, bidirectional
   - Cons: Connection management complexity
   - Best for: High-frequency updates

2. **Server-Sent Events**:
   - Pros: Simpler, automatic reconnection
   - Cons: Unidirectional only
   - Best for: One-way notifications

3. **Polling with smart intervals**:
   - Pros: Simple implementation
   - Cons: Higher latency, more requests
   - Best for: Low-frequency updates

Given your current Flask stack and notification frequency requirements,
I'd recommend SSE. Should we proceed with this approach?"

Result: Human makes informed decision with full context of trade-offs.

Problem Resolution Pattern

Human: "The Docker build is failing"

AI: "Let me investigate systematically:
1. Check the build log for specific errors
2. Verify Dockerfile syntax and layer dependencies
3. Check for base image compatibility issues
4. Examine any recent changes to dependencies

Rather than adding fallback mechanisms, I'll identify the root cause first."

Process:

  1. Gather information before proposing solutions
  2. Test hypotheses systematically
  3. Fix root cause, not symptoms
  4. Document resolution for future reference

Integration Challenge Example

Human: "The new auth service isn't working with our existing API"

AI: "Let me trace through the integration points:
1. Authentication flow and token format
2. API endpoint expectations vs. new service output
3. Error handling between services
4. Timing and timeout configurations

I see the issue - the token format changed. Instead of adding a
compatibility layer, let's align the services properly."

Domain Expert Collaboration Pattern

Vibe Coding isn't limited to software engineers. Any domain expert can leverage their specialized knowledge to create tools they've always needed but couldn't build themselves.

Real-World Example: A veteran HR professional with decades of recruiting experience collaborated with AI to create a sophisticated interview assessment application. The human brought invaluable domain expertise—understanding what questions reveal candidate quality, how to structure evaluations, and the nuances of effective interviewing. The AI handled form design, user interface creation, and systematic organization of assessment criteria.

Result: A professional-grade interview tool created in hours that would have taken months to develop traditionally, combining lifetime expertise with rapid AI implementation.

Key Pattern:

  • Domain Expert provides: Years of specialized knowledge, understanding of real-world requirements, insight into what actually works in practice
  • AI provides: Technical implementation, interface design, systematic organization
  • Outcome: Tools that perfectly match expert needs because they're built by experts

Performance Metrics and Outcomes

Real-World Productivity Gains

Our Experience: When we plan development work in one-week phases, we consistently complete approximately two full weeks worth of planned work in a single six-hour focused session.

This means: 14 days of traditional development compressed into 6 hours - roughly a 56x time compression for planned, collaborative work.

Why This Works: The combination of thorough planning, immediate AI implementation, and continuous human oversight eliminates most of the typical development friction:

  • No research delays (AI has broad knowledge)
  • No context switching between tasks
  • No waiting for code reviews or approvals
  • No debugging cycles from misunderstood requirements
  • No time lost to repetitive coding tasks

Quality Improvements

  • Fewer bugs: Human oversight catches issues early
  • Better architecture: More time for design thinking
  • Consistent code style: AI follows established patterns
  • Complete documentation: Plans and decisions preserved

Knowledge Transfer

  • Reproducible process: New team members can follow methodology
  • Preserved context: Plans survive team changes
  • Continuous learning: Both human and AI improve over time
  • Scalable expertise: Senior engineers can guide multiple projects

Tools and Technology Stack

Primary Development Environment

  • Desktop Commander: File operations, system commands, code analysis
  • Claude Sonnet 4: Strategic thinking, architecture decisions, code review
  • Git: Version control with detailed commit messages
  • Docker: Containerization and deployment
  • Python: Primary development language with extensive AI tooling

Workflow Integration

  • MkDocs: Documentation and knowledge management
  • GitHub: Code hosting and collaboration
  • Plans folder: Context preservation across sessions
  • Testing frameworks: Automated verification of changes

Context Management Structure

/project-root/
├── plans/               # Development plans and status
├── docs/               # Documentation and guides
├── src/                # Source code
├── tests/              # Test suites
└── docker/             # Deployment configurations

Common Patterns and Anti-Patterns

Successful Patterns

1. Plan → Discuss → Implement → Verify

1. Create plan in plans/ folder
2. Discuss approach and alternatives
3. Implement in small, verifiable chunks
4. Test each component before integration

2. Progressive Disclosure

- Start with high-level architecture
- Add detail as needed for implementation
- Preserve decisions for future reference
- Update plans with lessons learned

3. Human-AI Handoffs

AI: "This requires domain knowledge about your business rules.
Could you provide guidance on how customer tiers should affect pricing?"

Anti-Patterns to Avoid

1. The Fallback Trap

❌ "Let me add error handling to catch this edge case"
✅ "Let me understand why this edge case occurs and fix the root cause"

2. Over-Engineering

❌ "I'll build a generic framework that handles all possible scenarios"
✅ "I'll solve the immediate problem and refactor when patterns emerge"

3. Context Amnesia

❌ Starting fresh each session without reading existing plans
✅ Always begin by reviewing current state and previous decisions

4. Tool Misuse

❌ Using web search for file operations
✅ Desktop Commander for local operations, web tools for external information

Advanced Techniques

Multi-Agent Coordination

For complex projects, different AI agents can handle different aspects:

  • Architecture Agent: High-level design and system integration
  • Implementation Agent: Code generation and refactoring
  • Testing Agent: Test creation and verification
  • Documentation Agent: Technical writing and knowledge capture

Dynamic Context Switching

# Large refactor example
1. Create detailed plan with checkpoint strategy
2. Implement core changes in chunks
3. Test each checkpoint before proceeding
4. Update plan with lessons learned
5. Continue or adjust based on results

Knowledge Preservation Strategies

# In plans/lessons-learned.md
## Problem: Authentication integration complexity
## Solution: Standardized token format across services
## Impact: 3 hours saved on future auth integrations
## Pattern: Define service contracts before implementation

Battle-Tested Best Practices

Through hundreds of hours of real development work, we've identified these critical practices that separate successful Vibe Coding from common pitfalls:

The "Plan Before Code" Discovery

The Key Rule: No code gets written until we agree on the plan.

Impact: This single rule eliminated nearly all wasted effort in our collaboration. When AI understands the full context upfront, implementation proceeds smoothly. When AI starts coding without complete context, it optimizes for the wrong requirements and creates work that must be discarded.

Why It Works: AI doesn't know what it doesn't know. Planning forces the human to surface all the context, constraints, and requirements that AI can't infer. Once established, execution becomes efficient and accurate.

The "Less is More" Philosophy

Pattern: Always choose the simplest solution that solves the actual problem.

❌ "Let me build a comprehensive framework that handles all edge cases"
✅ "Let me solve this specific problem and refactor when patterns emerge"

Why This Works: Complex solutions create technical debt and make future changes harder. Simple, targeted changes preserve architectural flexibility.

Surgical Changes Over Refactoring

Pattern: Modify only what needs changing, preserve working functionality.

# Good: Minimal, focused change
edit_block(
    file="service.py",
    old="def process():\n    return data",
    new="def process():\n    validate_input()\n    return data"
)

# Avoid: Large refactoring that could break existing functionality

Why This Matters: Working software is valuable. Surgical changes reduce risk and maintain system stability.

Long-Running Process Handoffs

Critical Pattern: AI should never manage lengthy operations directly.

AI: "This database migration will take 15+ minutes. I'll provide the command
and monitor approach rather than running it myself:

Command: python manage.py migrate --database=production
Monitor: Check logs every 2 minutes for progress
Rollback: python manage.py migrate --database=production 0001_initial

Would you prefer to run this yourself for better control?"

Human Advantage: Human oversight of long processes prevents resource waste and enables real-time decision making.

The Fallback Mechanism Trap

Anti-Pattern: "Let me add error handling to catch this edge case..." Better Approach: "Let me understand why this edge case occurs and fix the root cause..."

❌ Fallback: Add try/catch to hide the real problem
✅ Root Cause: Investigate why the error happens and fix the source

Time Savings: Solving root problems prevents future debugging sessions and creates more maintainable code.

Verification-First Completion

Rule: Nothing is complete until it's been tested and verified.

AI Implementation → Human/AI Testing → Verification Complete
   ↑                                           ↓
   ←——————— Fix Issues If Found ←———————————————

Testing Handoff Options:

  • "Should I run the tests or would you prefer to verify?"
  • "Here's what needs testing: [specific scenarios]"
  • "Implementation ready for verification: [verification checklist]"

Getting Started with Vibe Coding

For Individual Developers

  1. Set up tools: Desktop Commander, AI assistant, documentation system
  2. Start small: Choose a well-defined feature or bug fix
  3. Practice patterns: Plan → Discuss → Implement → Verify
  4. Document learnings: Build your pattern library

For Teams

  1. Establish standards: File organization, documentation formats, handoff protocols
  2. Train together: Practice the methodology on shared projects
  3. Create templates: Standard plan formats, common decision trees
  4. Measure outcomes: Track speed and quality improvements

Success Metrics

  • Development velocity: Features delivered per sprint
  • Code quality: Bug rates, review feedback, maintainability
  • Knowledge retention: How quickly new team members become productive
  • Context preservation: Ability to resume work after interruptions

Conclusion: The Future of Development

Vibe Coding represents a fundamental shift from human-as-coder to human-as-conductor. By treating AI as a powerful but directed tool rather than a replacement, we achieve:

  • 10x velocity improvements without sacrificing quality
  • Better architecture through increased design thinking time
  • Preserved expertise via documentation and knowledge transfer
  • Scalable impact where senior engineers can guide multiple projects

This methodology isn't about replacing developers—it's about amplifying their impact. Senior engineers become force multipliers, able to execute complex projects that would normally require entire teams.

The companies that master human-AI collaboration today will dominate tomorrow's software landscape. Vibe Coding provides a proven framework for making that transition successfully.

Ready to Start? Begin with a small, well-defined feature in your current project. Apply the Plan → Discuss → Implement → Verify cycle. Set up a /plans folder in your project and create your first development plan using the methodology outlined here.


Want to learn more? Contact us to discuss how Vibe Coding can transform your development process, or explore our Services for enterprise implementations.

Real-World Example: Token Counter Implementation

To demonstrate this methodology in practice, here's an actual session from our development work. The context: we were working in Claude Desktop with system prompts providing project structure information and collaboration preferences.

The Initial Request

Human: "we could also add a token counter somewhere in the lit-desktop UI ... that's the kind of project that is super clear to a reader what we're after"

AI: "Perfect! A token counter in lit-desktop is an excellent demonstration project. It's:

  • Crystal clear goal: Everyone understands what a token counter does
  • Visual result: Immediate, tangible outcome users can see
  • Right scope: Substantial enough to show methodology, small enough to complete
  • Real value: Users actually want to track token usage"

Planning Phase (5 minutes)

Human: "i think it's per conversation. if you recall how you and I work here in claude desktop there's always this point about an hour into any project and we can both tell when it's approaching. i don't know how you know but I can tell by the size of the scroll bar..."

AI: Created detailed implementation plan covering:

  • Technical decisions (client-side approximation vs server-side)
  • UI placement (near text input for pre-send awareness)
  • Update frequency (200ms debounce)
  • Data flow (conversation total + live draft)

Architecture Research (5 minutes)

Before writing any code, we examined the existing lit-desktop structure:

  • Located Angular chat components
  • Found message input textarea with [(ngModel)]="currentInput"
  • Identified messages: ChatMessage[] array for conversation data
  • Determined optimal display location in status hint area

Implementation (14 minutes)

Service Creation: Built TokenCounterService with approximation algorithm

countTokens(text: string): number {
  const charBasedTokens = Math.ceil(text.length / 4);
  const wordBasedTokens = Math.ceil(text.trim().split(/\s+/).length / 0.75);
  return Math.round((charBasedTokens + wordBasedTokens) / 2);
}

Component Integration: Added token counting to chat page component

  • Injected service into existing constructor
  • Added properties for conversation and draft token counts
  • Integrated with message loading and session switching

Template Updates: Modified HTML to display counter

<span *ngIf="!isLoading">
  <span>Lit can make mistakes</span>
  <span *ngIf="tokenCountDisplay" class="token-counter"> • {{ tokenCountDisplay }}</span>
</span>

Problem Resolution (4 minutes)

Browser Compatibility Issue: Initial tokenizer library required Node.js modules

  • Problem: gpt-3-encoder needed fs/path modules unavailable in browser
  • Solution: Replaced with custom approximation algorithm

Duplicate Method: Accidentally created conflicting function - Problem: Created duplicate onInputChange method - Solution: Integrated token counting into existing method

Quality Assurance

Human: "are we capturing the event for switching sessions?"

This caught a gap in our implementation - token count wasn't updating when users switched conversations. We added the missing updates to selectChat() and createNewChat() methods.

Human: "instead + tokens how about what the new total would be"

UX improvement suggestion led to cleaner display format: - Before: "~4,847 tokens (+127 in draft)" - After: "~4,847 tokens (4,974 if sent)"

Final Result

Development Outcome: Successfully implemented token counter feature across 4 files with clean integration into existing Angular architecture.

Timeline (based on development session):

  • 14:30 - Planning and architecture discussion complete
  • 14:32 - Service implementation with token approximation algorithm
  • 14:34 - CSS styling for visual integration
  • 14:36 - HTML template updates for display
  • 14:40 - Component integration and event handling
  • 14:40-15:00 - Human stuck in meeting; AI waits
  • 15:03 - UX refinement and improved display format
  • 15:04 - Final verification and testing

Total active development time: 14 minutes (including planning, implementation, and verification)

What This Demonstrates

The session shows core Vibe Coding principles in action:

  • Human strategic direction: Clear problem definition and UX decisions
  • AI tactical execution: Architecture research and rapid implementation
  • Continuous verification: Testing and validation at each step
  • Quality collaboration: Human oversight caught integration gaps and suggested UX improvements
  • Surgical changes: Modified existing architecture rather than building from scratch
  • Context preservation: Detailed planning enabled seamless execution

The development session demonstrates how Vibe Coding enables complex feature development in minutes rather than hours, while maintaining high code quality through human oversight and systematic verification.


Want to experience Vibe Coding yourself? Contact our team to discuss how Vibe Coding can transform your development process.

About the Author

Ben Vierck is a senior software architect and CTO at JOV.AI, with over a decade of experience leading engineering teams at Fortune 100 companies. He developed the Vibe Coding methodology through hundreds of hours of real-world collaboration with AI development assistants while building production systems and open source projects.


Tags: AI Development, Human-AI Collaboration, Software Engineering, Development Methodology, Productivity, AI Tools, Enterprise Development

References


  1. Wikipedia contributors. "Vibe coding." Wikipedia, The Free Encyclopedia, https://en.wikipedia.org/wiki/Vibe_coding. Accessed June 20, 2025. 

  2. Merriam-Webster Dictionary. "Vibe coding." Slang & Trending, https://www.merriam-webster.com/slang/vibe-coding. Accessed June 20, 2025.