Malachi's Development Style & Agentic Workflow Integration
Comprehensive Analysis for Code 5 (ZTAG System)
Prepared for: Quan Gan, ZTAG Leadership
Analysis Period: 120+ meetings (May 2025 - Feb 2026)
Subject: Lead Architect Malachi Burke
Context: Senior veteran developer (80s/90s experience), does not trust AI, requires detail & accuracy
Executive Summary
Malachi Burke is ZTAG's lead architect and principal technical decision-maker for the Code 5 system. Analysis of 120+ meetings reveals a highly disciplined, detail-oriented, collaborative developer with three critical characteristics:
- Manual-first, verification-heavy approach - Prefers hands-on testing, skeptical of automation
- Mentorship-driven leadership - Invests heavily in junior developer growth through direct feedback
- Quality-over-velocity mindset - Resists rushing, emphasizes robust architecture over speed
Key Insight
Malachi's skepticism toward AI is not ideological but pragmatic β he's seen brittle code from auto-generation and demands that ANY tool (including AI) prove its value through rigorous testing and maintainability.
Part I: Malachi's Development Style
1. Communication Pattern
Style Characteristics:
- Direct & Technical - Gets straight to implementation details, avoids buzzwords
- Question-Driven - Asks probing "what/how/why" questions rather than directing
- Documented Feedback - Prefers written GitHub issues/PRs over verbal direction
- Consensus-Building - Frames decisions as "team agreed to..." not "I decided..."
Example Topics from Meeting Summaries:
- IR packet protocols (V1 vs V2 packet formats, CSMA backoff algorithms)
- State machine implementations (nested vs. layered, finite state design)
- Git practices (commit message standards, PR scope, branch hygiene)
- Code clarity (variable naming, avoiding ambiguity, testing coverage)
When He Accepts Suggestions:
- When backed by working code (not theory)
- When testing shows measurable improvement
- When proposal reduces technical debt without introducing new risk
- When junior developers demonstrate understanding of consequences
When He Rejects Suggestions:
- "This solves the wrong problem"
- Lack of evidence or test coverage
- Risk of breaking existing functionality
- Insufficient documentation or rationale
- Shortcuts that compromise long-term maintainability
2. Detail & Accuracy Requirements
He Demands:
- Accurate packet protocol specs (byte-by-byte breakdown)
- Clear acceptance criteria in user stories
- Regression test coverage before merges
- Explicit handling of edge cases (multiple devices, timing conflicts)
- Version compatibility documentation
Red Flags He Raises:
- "This adds complexity without solving the actual problem"
- "We haven't tested this scenario"
- "The documentation doesn't match the code"
- "This is working around a bug, not fixing it"
His Question Patterns:
- "What happens if [edge case]?"
- "Why is [this code] written this way?"
- "How do we verify this doesn't break [legacy system]?"
- "Is this a symptom or the root cause?"
3. Decision-Making Flow
Typical Pattern (from meeting analysis):
- Define the problem precisely (IR packet corruption, MQTT sync issues, state machine clarity)
- Explore the landscape (What approaches exist? What did we try before?)
- Prototype & test (Small working example > lengthy discussion)
- Gather evidence (Logs, traces, measurable improvements)
- Document the decision (Why this approach? Trade-offs? Reversibility?)
- Implement with review (Code review focusing on clarity, testing, edge cases)
What Works:
- Showing him a working prototype before asking for approval
- Providing test data/logs proving the fix
- Documenting assumptions and trade-offs
- Acknowledging complexity and being honest about unknowns
What Doesn't Work:
- "Trust me, this is better"
- Skipping regression testing
- Code that "sort of works" for now
- Missing edge case handling
4. Collaboration Style with Team
With Junior Developers (Ryan, Basim, Shan, Faisal, Sean):
- Heavy mentoring β Spends significant time in 1-on-1 sessions explaining decisions
- Clear feedback β Specific code comments, not vague suggestions
- High expectations β Assumes they'll understand the "why" behind requirements
- Trusts but verifies β Lets them implement, reviews thoroughly
- Celebrates progress β Acknowledges learning and improvement
With Leadership (Quan):
- Architect's perspective β Raises long-term scalability concerns
- Risk assessment β Identifies technical debt and future bottlenecks
- Resource planning β Honest about what's feasible with current team
- Mission alignment β Connects technical choices to product goals
Recurring Themes in Team Interactions:
- Standardizing processes (Git workflows, PR formatting, testing procedures)
- Improving code clarity (removing ambiguity, better naming, state machines)
- Planning for future contributors (documentation, architecture, onboarding)
- Balancing urgency with quality (yes to quick wins, no to shortcuts that degrade future velocity)
5. Concerns About AI/Automation
What He's Observed:
- AI-generated game code that works but is unmaintainable (redundancy, unclear logic flow)
- Auto-generated code that lacks proper error handling
- Features that "work" but don't handle edge cases
- Insufficient testing coverage in auto-generated work
His Position (Inferred from Meetings):
- "AI is a tool, not a replacement for thinking"
- "I'll use AI when it demonstrably reduces work while improving quality"
- "Show me test coverage before I trust it"
- "Auto-generation should be a starting point, not the finish line"
Where He Sees AI Value:
- Boilerplate generation (when followed by manual review)
- Documentation assistance (if verified for accuracy)
- Rapid prototyping (for proof-of-concept work)
- Code refactoring suggestions (if tested thoroughly)
Where He Resists:
- Core algorithm implementation without review
- Replacing human decision-making in architecture
- Skipping testing in the name of "velocity"
- Using AI as an excuse for insufficient documentation
Part II: Current Agentic Workflow Landscape
Analysis of Current Tools (Cursor, Windsurf, Cline, Claude Code, GitHub Copilot)
How Malachi Views Them Today:
GitHub Copilot (Line-level completion)
- β
Acceptable for boilerplate, familiar patterns
- β Insufficient for complex logic, edge cases
- Current use: Minimal (prefers explicit code decisions)
Cursor & Windsurf (AI-assisted IDE)
- β οΈ Intriguing but unproven at scale
- Concern: "Does it help or hide problems?"
- Trust issue: "How do I know it didn't introduce a subtle bug?"
Cline & Claude Code (Code generation & planning)
- π¨ High friction: These feel like "replacing developers"
- Concern: Quality control, regression testing coverage
- Reality: Doesn't fit his manual-verification workflow
Why Current Agentic Workflows Conflict with Malachi's Approach
| Agentic Workflow Trait |
Malachi's Preference |
Conflict |
| Speed |
Quality & correctness |
Rushing introduces bugs he has to fix later |
| Autonomous execution |
Human oversight & review |
Loses opportunity to learn & maintain code |
| Black-box solutions |
Transparent, traceable logic |
Can't debug or modify if requirements change |
| "Move fast, fix later" |
"Get it right, maintain forever" |
Accumulates technical debt he resists |
| Minimal documentation |
Explicit specs & rationale |
Hard to onboard or modify later |
Part III: Recommended Agentic Workflow Integration
Principle: "Augment, Don't Replace"
Malachi's skepticism isn't about rejecting innovationβit's about maintaining quality standards. The path forward is to introduce agentic workflows as force multipliers for tedious work, not as replacements for thinking.
Phase 1: Pilot (Months 1-2)
Goal: Prove value on low-risk, well-defined tasks
Success Metric: Measurable velocity improvement with zero regression bugs
1A: Agentic Boilerplate Generation
- Task: Generate FreeRTOS task scaffolding, IDF component templates, test fixtures
- Tool Recommendation: Claude Code (with custom system prompt requiring testing)
- Workflow:
- Define requirements (task name, priority, dependencies)
- AI generates scaffold + unit test template
- Malachi reviews structure, approves API
- Junior dev implements logic with guidance
- Regression test suite passes before merge
Why It Works for Malachi:
- Still requires manual review (respects his oversight preference)
- Eliminates tedious file creation (time savings)
- Template consistency (improves maintainability)
- Junior devs learn by filling in logic, not copy-pasting
1B: Test Coverage Analysis
- Task: Identify code paths without unit tests, suggest tests
- Tool Recommendation: Claude Code analyzing coverage reports
- Workflow:
- Run coverage reports on Code 5 codebase
- AI identifies gaps, suggests test scenarios
- Malachi reviews suggestions, rejects unrealistic ones
- Senior or junior dev implements approved tests
- Verify coverage increases
Why It Works for Malachi:
- Augments his existing verification habits
- Doesn't replace his judgment (he decides which tests matter)
- Measurable, verifiable output (coverage %)
- Forces explicit thinking about edge cases
1C: Documentation Sync Checker
- Task: Flag where code and docs diverge, suggest updates
- Tool Recommendation: Claude Code analyzing git diffs + FDD/comments
- Workflow:
- AI scans recent PRs vs. Functional Design Doc
- Flags mismatches (code changed, docs didn't)
- Malachi reviews, decides if docs need update or code had unintended change
- Assigns correction task
- Verify PR has updated docs
Why It Works for Malachi:
- Respects his insistence on accuracy (catches drift)
- Manual decision (he decides if docs or code is wrong)
- No code execution (low risk)
- Improves future maintainability
Phase 2: Expansion (Months 3-4)
Goal: Extend agentic workflows to higher-value tasks while maintaining quality
Prerequisite: Phase 1 pilots show zero regressions and measurable time savings
2A: PR Review Assistant (Pre-human review)
- Task: Perform automated code review before Malachi reviews
- Tool Recommendation: Custom Cline workflow + GitHub Actions
- Workflow:
- PR created by junior dev
- Agentic review runs:
- Check code style/naming conventions
- Flag uncovered code paths
- Identify potential edge cases
- Suggest test improvements
- Look for patterns from prior issues
- Report generated, linked to PR
- Malachi reviews (now with pre-analysis)
- Junior dev addresses feedback
Why It Works for Malachi:
- He still makes the final decision (high-touch human review)
- Faster for him (pre-analysis reduces discovery time)
- Better for juniors (clear, automated feedback on basics)
- Focuses his expertise on architectural decisions
2B: Refactoring Proposals
- Task: Suggest architectural improvements (state machine simplification, decoupling)
- Tool Recommendation: Claude Code with Code 5 architecture context
- Workflow:
- Select a module (e.g., IR packet handling, MQTT sync)
- AI analyzes current structure, proposes improvements
- Malachi reviews proposals (keeps/discards based on risk/benefit)
- Creates GitHub issue with decision rationale
- Junior dev or volunteer implements with heavy review
Why It Works for Malachi:
- Respects his architectural authority (he decides what to refactor)
- Reduces thinking load on complex decisions
- Provides options to evaluate (not forcing one path)
- Forces documentation of trade-offs
2C: Field Issue Analysis
- Task: When a bug report comes in, AI pre-diagnoses patterns
- Tool Recommendation: Claude Code + Cline for trace analysis
- Workflow:
- Customer reports "Zombie Tag hangs after 15 minutes"
- AI gathers logs, examines timeout patterns, suggests root causes
- Malachi reviews diagnosis, decides which to investigate
- Implements fix with appropriate regression tests
- Deploy and monitor
Why It Works for Malachi:
- Speeds up diagnosis phase (saves his time)
- Doesn't replace his judgment (he still decides what's likely)
- Reduces context-switching (analysis ready when he investigates)
Phase 3: Scaling (Months 5-6)
Goal: Autonomous workflows for routine tasks; agentic support for decision-making
Prerequisite: Phase 1-2 show sustained improvement and team confidence
3A: Automated Testing Scaffolding
- Task: When a new game is added, auto-generate test cases
- Tool Recommendation: Cursor with test template library
- Workflow:
- Game spec uploaded to repository
- AI generates unit test template (state transitions, edge cases)
- Game developer fills in assertions
- Tests run in CI/CD
- Malachi reviews coverage (quarterly, not per-game)
Why It Works for Malachi:
- Ensures consistency across games
- Reduces boilerplate test writing
- Focuses his review on game-specific logic, not test structure
3B: Dependency & Compatibility Checker
- Task: When upgrading ESP-IDF, flag potential incompatibilities
- Tool Recommendation: Cline + GitHub Actions
- Workflow:
- ESP-IDF version bump in CI/CD
- AI scans codebase for deprecated APIs, version-specific behavior
- Generates report of required changes
- Malachi triages (critical vs. nice-to-have)
- Team implements fixes
Why It Works for Malachi:
- Proactive (catches problems before they bite)
- Reduces manual upgrade effort
- Focuses his expertise on decision-making, not repetitive searching
Phase 4: Strategic (Ongoing)
4A: Architecture Advisors (Claude Code)
- Use Case: When Malachi needs to think through a big decision (e.g., "Should we move to Code 5 scheduler earlier?"), use AI for structured brainstorming
- Workflow:
- Malachi outlines the decision (constraints, options, unknowns)
- AI structures options, raises questions he should consider
- Malachi uses this to think more clearly, not to decide
- Presents structured analysis to team
4B: Knowledge Capture
- Use Case: Extract Malachi's design rationales into documentation
- Tool: Claude Code to summarize meeting notes, create decision records
- Value: Preserves tribal knowledge, makes onboarding easier for next senior dev
Part IV: Risk Assessment & Mitigation
Risk #1: Over-Automation Erodes Code Quality
Symptom: Junior devs trust AI suggestions without thinking
Mitigation:
- Code review must include "Why is this here?" verification
- Require test coverage before AI tool suggestions become code
- Monthly audits of merged AI-assisted PRs for subtle bugs
Risk #2: Malachi Feels Sidelined or De-valued
Symptom: "If AI can do this, why do you need me?"
Mitigation:
- Frame agentic tools as his productivity multiplier (not replacement)
- Increase his strategic workload (architecture decisions, mentoring, system design)
- Showcase time saved β applied to bigger problems
- Explicitly acknowledge his role in vetting/approving tools
Risk #3: AI Produces Code That "Works But Breaks Later"
Symptom: Unmaintainable auto-generated code that passes tests but has hidden issues
Mitigation:
- Every AI-generated code module must have:
- Unit test coverage >80%
- Architecture comment explaining assumptions
- Manual review by Malachi (high-touch for logic, low-touch for boilerplate)
- 6-month checkup (does it still make sense?)
Risk #4: Team Becomes Dependent on "AI Thinks"
Symptom: Junior devs stop learning and become order-takers
Mitigation:
- AI tools suggest β Human reviews β Human implements β Human tests
- Monthly 1-on-1 with each junior dev on a decision where AI was wrong (learn from it)
- "Understand before you code" code review standard
Risk #5: Malachi's Skepticism Spreads (Team Resistance)
Symptom: Team rejects tools before giving them fair shot
Mitigation:
- Start with Phase 1 (low risk, high value pilots)
- Celebrate wins (e.g., "This workflow saved 6 hours on PR review")
- Malachi becomes the advocate (if he says "this adds value," team trusts it)
Part V: Phased Rollout Plan
Timeline & Milestones
Month 1-2: PILOT
ββ 1A: Boilerplate generation (FreeRTOS, IDF templates)
ββ 1B: Test coverage analysis
ββ 1C: Documentation sync checker
Milestone: Zero regressions, β₯10% velocity gain, Malachi approval
Month 3-4: EXPANSION
ββ 2A: PR review assistant (pre-human review)
ββ 2B: Refactoring proposals
ββ 2C: Field issue analysis
Milestone: Team confidence, sustained quality, Malachi championing tools
Month 5-6: SCALING
ββ 3A: Automated test scaffolding
ββ 3B: Dependency checker
ββ 3C: Knowledge capture (decision records)
Milestone: Reduced Malachi bottleneck, juniors more independent, sustained quality
Month 6+: STRATEGIC
ββ 4A: Architecture advisors (brainstorming partners)
ββ 4B: Tribal knowledge preservation
ββ 4C: Continued process optimization
Milestone: Malachi focused on vision/mentoring, AI handles routine work
Success Criteria
β
Code Quality: Zero net increase in bugs from agentic-assisted code
β
Velocity: β₯15% improvement in time-to-merge for routine tasks
β
Maintainability: Code written with AI assistance remains understandable 6+ months later
β
Team Adoption: β₯70% of junior devs using agentic tools by Month 4
β
Malachi Satisfaction: "These tools let me focus on the hard decisions"
Part VI: Recommended Tools & Configuration
Primary Recommendation: Claude Code (Extended) + Cline
Why:
- Claude's reasoning (explains the why) aligns with Malachi's need for clarity
- Can be customized with project-specific rules (Code 5 conventions)
- Good at code review and architecture analysis
- Respects Malachi's "verify first, trust later" mindset
Configuration for Code 5 Project
# System Prompt for ZTAG Code 5 Agentic Workflows
You are an assistant helping the ZTAG Code 5 development team (an embedded systems
project using ESP-IDF, FreeRTOS, and custom IR/RF protocols).
## Your Role
- Suggest, don't decide
- Explain trade-offs explicitly
- Identify edge cases and risks
- Require test coverage before approval
- Respect existing architecture decisions
## Code 5 Standards You Must Know
- IR packets: V2 format with CSMA backoff
- MQTT sync: Document assumptions about ordering
- State machines: Prefer explicit (not implicit) states
- Testing: Every feature needs unit + regression tests
- Naming: Descriptive (no abbreviations that don't appear in spec)
- Comments: Explain WHY, not WHAT
## When to Say "No"
- If test coverage would be <70%
- If breaking legacy IR protocol compatibility
- If adding undocumented assumptions
- If architecture would require Malachi's re-review
## Key Phrase
When in doubt, suggest: "This might work, but let's verify with [test/prototype] first."
Secondary Tools
- GitHub Actions for automated testing, linting, coverage
- Cline for task-based workflows (refactoring, analysis)
- Cursor (optional) if team prefers IDE-native assistance
Part VII: Conversation Starter with Malachi
Here's how to frame this with Malachi (respecting his style):
Opening
"Malachi, I've been analyzing where our bottlenecks are. Right now, you're the gating factor on:
- Code review (every PR waits for you)
- Architecture decisions (juniors need your input)
- Documentation sync (manually tracking changes)
That's not sustainable as we grow Code 5. I want to explore tools that handle routine work so you can focus on the hard problems."
The Ask
"Would you be willing to try a structured 6-week pilot? Specific tasks:
- [1A] Auto-generate FreeRTOS scaffolding β You review, approve/modify
- [1B] Flag test gaps β You decide which tests matter
- [1C] Check if docs match code β You decide if code or docs need fixing
Goal: Show that agentic tools reduce toil while maintaining your quality bar. Zero regressions, measurable time saved."
Addressing His Concerns
"Will this replace junior devs?"
No. It handles boilerplate. They'll implement logic, which is where they learn. You'll have more time to mentor thinking, not just fix sloppy code.
"What if the AI generates something subtle wrong?"
Same risk as a junior dev. We'll have the same review process: submit β review β test β merge. You still gate it. AI just does the legwork faster.
"How do I know it won't break things?"
We start with low-risk tasks (boilerplate, analysis, not logic). Phase 1 succeeds only if we have zero regressions. If that doesn't work, we stop and re-evaluate.
Part VIII: Metrics to Track
During Pilot (Phase 1)
- Time to merge (PR age before review, before merge)
- Test coverage (% of new code covered by tests)
- Regression bugs (any bugs traced to agentic code?)
- Malachi's review time (subjective: "Is this faster?")
- Junior dev satisfaction ("Is feedback clearer now?")
During Expansion (Phase 2-3)
- Velocity (features/bugs closed per sprint)
- Code quality (static analysis, cyclomatic complexity)
- Maintainability (code churn rate, test stability)
- Bottleneck shifts (where is Malachi needed now vs. before?)
- Team adoption (% of workflows using agentic tools)
Long-term (Phase 4+)
- Time-to-ship (days from feature start to production)
- Field issues (customer-reported bugs from agentic code?)
- Knowledge preservation (can new senior dev pick up architecture?)
- Malachi's leverage (time spent on mentoring vs. implementation)
Part IX: Special Consideration β Malachi's Trust Model
Why His "No AI" Stance Works in His Favor
Malachi's skepticism isn't a weaknessβit's a strength. In a team of 4+ junior devs, one senior developer with high standards is the quality control. By introducing agentic workflows through him (not around him), we:
- Preserve quality (his judgment gates every tool)
- Build institutional knowledge (he learns the boundaries of what tools can do)
- Maintain team trust (if Malachi approves, juniors will use it)
- Future-proof the team (when Malachi trains the next senior, they'll inherit vetted workflows)
The Real Conversation
The goal isn't to convince Malachi that AI is amazing. It's to show him:
"Here's tedious work humans shouldn't do. Here's a tool that eliminates it. You control it. You decide if it's worth it. If not, we stop."
Conclusion
Malachi Burke is an asset in adopting agentic workflows because his skepticism forces rigor. Rather than fighting his caution, leverage it.
The playbook:
- Start small (boilerplate, analysis, documentation)
- Prove value (measurable time savings, zero regressions)
- Let him decide (never surprise him with automation)
- Make him the champion (his approval builds team confidence)
- Shift his workload (free him from toil β more mentoring & architecture)
If you follow this approach, in 6 months, Malachi will likely say: "These tools save us 20% on routine work while I focus on decisions that actually matter. That's a win."
Appendix A: Key Meetings Analyzed
| Meeting |
Topic |
Malachi's Key Insight |
| 2/12/26 |
Code5 OTA & debugging |
Detail-oriented: "Why is this happening?" over "Let's patch it" |
| 2/2/26 |
RLGL algorithm updates |
Process focus: "Define the algorithm clearly before coding" |
| 1/29/26 |
Code 5 sync update |
Mentorship: Guides team on kconfig and reference implementation |
| 1/26/26 |
FreeRTOS integration |
Architecture: "Cartridge system for modularity" (thinks in systems) |
| 12/16/25 |
User stories planning |
Standards: "Clarify 'why' in stories; separate technical from functional" |
| 11/25/25 |
Trade show learnings |
System design: "Marco Polo" mesh model for scalability |
| 11/13/25 |
Code 5 planning |
Vision: "Collaborative, research-driven approach" |
| 10/23/25 |
External meeting |
Code quality: "How do we keep this maintainable?" |
| 9/4/25 |
New developer intro |
Mentorship: "Explains architecture to new team members" |
| 8/21/25 |
V2 packet progress |
Technical excellence: "Byte-by-byte analysis, robust testing" |
Document Version: 1.0
Last Updated: Feb 16, 2026
Author: Analysis Team
Status: Ready for Malachi Review