Project Management

Project governance, engineering process, stakeholder management, and quality assurance strategy for a team of 5 engineers over 9 months with AI-first approach.

1. Project Management Framework

Why Not Pure Scrum?

FrameworkFit?Reasoning
Pure Scrum5 people don't need heavy ceremonies. 4-hour sprint planning for 5 people = waste
Pure Kanban⚠️Good for flow, but lacks checkpoints for migration milestones
Shape Up (modified)6-week cycles + cooldown. Fits migration phases. Appetite-based (fixed time, variable scope)
SAFeOverkill for 5 people. SAFe is for 50+ engineers

Decision: Shape Up modified + Kanban within cycles. 6-week cycles aligned with migration phases. Appetite-based estimation — fixed time, variable scope. No story points.

Project Rhythm — 6 Cycles × 6 Weeks

Each 6-week cycle: Week 1 → Shaping (define appetite, scope bets) | Week 2–5 → Building (Kanban flow, daily standups) | Week 6 → Cooldown (retro, tech debt, learning)

Estimation — Appetite-Based (No Story Points)

AppetiteDurationExample
Small Batch≤ 1 weekCommunications service extraction (AI handles most)
Big Batch2–4 weeksTravel Booking full extraction + tests + React pages
Epic1 cycle (6 weeks)Event Management + Reporting + related React modules

Rule: If a task exceeds 6 weeks → must be broken down further. No "ongoing" tasks.

Task Board

Backlog
Unshaped ideas
Shaped
Scoped, estimated, assigned
Building
In active dev (WIP ≤ 5)
Review
CodeRabbit + human review
Staging
QA on staging env
Done
In production

WIP Limit: 5 (1 per engineer max). No multitasking. If blocked → swarm (help each other unblock).

2. Ceremonies & Cadence

CeremonyFrequencyDurationPurposeWho
Daily StandupDaily10 minBlockers only. No status reporting — use the boardAll 5
Cycle ShapingEvery 6 weeks2 hoursDefine appetite, scope bets, assign pitchesTech Lead + team
Weekly DemoWeekly30 minShow working software to stakeholdersRotating presenter
Cycle RetroEvery 6 weeks1 hourWhat worked, what didn't, AI effectiveness reviewAll 5
Architecture ReviewBi-weekly1 hourReview ADRs, service boundaries, tech decisionsTech Lead + senior
AI Workflow CheckWeekly15 minAI metrics review, prompt calibration, governanceTech Lead

Total ceremony time: ~3.5 hours/week — under 10% of working time. The rest = build.

Weekly Cadence

Daily Schedule

09:00 Standup (10 min) — blockers only
09:10 Deep work — NO meetings until 12:00
14:00 Open for ad-hoc pairing, reviews
16:00 Async review — PRs, CodeRabbit comments
17:00 AI batch runs scheduled (overnight migration)

Weekly Schedule

Monday AM: Sprint planning (30 min)
Tue–Thu: Heads-down development (async standups)
Wednesday AM: Architecture review (bi-weekly, 1 hr)
Friday AM: Code review session (60 min cross-service)
Friday PM: Demo (30 min) + AI metrics check (15 min)

3. Team Structure & Roles

Bus Factor Mitigation

With 5 people, 1 person leaving = 20% capacity lost. Every mitigation strategy is critical:

Pair on critical modules

2 people know each service. No single ownership.

Bus factor ≥ 2 for every module

Rotate reviewer

Code review rotates — everyone reviews everyone's code

Cross-knowledge across codebase

AI code walkthrough

Weekly rotation: Senior explains Travel, Backend explains Event...

Everyone understands every service

DevOps cross-train

Every backend eng knows how to deploy their own service

DevOps doesn't become a bottleneck

Service Ownership Map

ServicePrimary OwnerSecondaryFrontend
Travel BookingD2 (Sr Backend)D1 (Tech Lead)D4 + D5
Event ManagementD3 (Backend)D2 (Sr Backend)D4
WorkforceD4 (Fullstack)D1 (Tech Lead)D4 + D5
CommunicationsD4 (Fullstack)D3 (Backend)D4
ReportingD5 (FE/DevOps)D3 (Backend)D5 + D4
Payment ACLD1 (Tech Lead)D2 (Sr Backend)
API GatewayD1 (Tech Lead)D5 (FE/DevOps)
CI/CD + InfraD5 (FE/DevOps)D1 (Tech Lead)

4. Stakeholder Management

Communication Plan

StakeholderChannelFrequencyContent
C-Level / SponsorExecutive summary (1-page)Bi-weeklyRisk status, milestone progress, AI ROI metrics, budget burn
Product OwnerDemo + written updateWeeklyWorking features, migration progress, upcoming changes
Business UsersChange notificationPer migration phaseWhat's changing, what's not, who to contact
External API ConsumersAPI deprecation notice30 days aheadBreaking changes, migration guides, new endpoints
Engineering TeamStandup + boardDailyIn-progress work, blockers, decisions needed
Security / ComplianceAudit reportMonthlySAST results, AI governance, payment module status

The "No" Framework — Expectation Management

"Can we add feature X?""Yes, if we defer [Y]. Here's the trade-off."
"Can we speed up?""We're at 2x AI capacity. Adding people adds coordination cost. We can re-scope instead."
"Why isn't Payment modernized?""By design. Constraint: Payment frozen Phase 1. Plan exists for Phase 2. Here's the ACL keeping it safe."
"Can we skip testing?""No. With 75% AI-generated code, testing IS the quality gate. Non-negotiable."
"Competitor launched feature Z""Noted. Added to backlog. Current priority: foundation first."

Escalation Path

P4 (Low)

Engineer fixes → PR → merge. No escalation.

P3 (Medium)

Engineer + Tech Lead discuss. Fix within cycle. Mention in weekly update.

P2 (High)

Tech Lead decides → immediate fix. Notify PO same day. Include in exec summary.

P1 (Critical)

War room (all hands). Tech Lead → Sponsor within 1hr. Hourly updates. Post-mortem 48h.

5. Development Lifecycle

Definition of Done

Code

Feature implemented and builds successfully
AI-generated code reviewed by human (mandatory)
Follows Clean Architecture structure
No TODO/HACK comments left untracked

Testing

Unit tests pass (≥80% coverage for new code)
Contract tests pass (Pact — for API changes)
Integration tests pass (DB, event bus)
No regression in existing tests

Security

SAST scan clean (CodeQL)
No secrets in code
Payment-related: 2 human reviewers approved

Observability

Structured logging for key operations
OpenTelemetry trace spans for cross-service calls
Health check endpoint working

Documentation

API changes reflected in OpenAPI spec
ADR created for architecture decisions
README updated if setup instructions changed

Deployment

Docker image builds successfully
Deployed to staging and tested
Monitoring/alerting configured for new endpoints

6. Code Review Process

Git Workflow

Branch Naming

feature/{module}-{description}
fix/{module}-{description}
infra/{description}

Rules

  • • PR required for main (no direct push)
  • • CodeRabbit auto-review on PR create
  • • ≥1 human approval required
  • • Payment-related: ≥2 human approvals
  • • CI must pass (build + test + security)
  • • Squash merge to main (clean history)

ADR Process

Triggers: Any decision affecting service boundaries, database choice, communication patterns, technology selection, security model, or AI governance rules.

Flow: Engineer drafts ADR (AI-assisted) → Tech Lead reviews (24h) → Team review in Architecture Review → Accepted/Rejected/Amended → Stored in /docs/adrs/ADR-NNN-title.md

7. Release Management

Service Deployment

2–3× per week

Every merged PR → auto-deploy to staging. Manual approval for production. Rolling update (zero downtime). Feature flags for incomplete features.

Module Go-Live

Once per phase

Full module cutover: traffic routes from legacy → new. Canary: 5% → 25% → 50% → 100%. Rollback via YARP < 5 min. Stakeholder notified 1 week before.

Database Migration

1–2 total

Per-service DB cutover. CDC running weeks before. Blue-green: new DB + old DB fallback. Data verification scripts mandatory.

Canary Release Process

Feature Flags

Flag NamePurpose
travel.new-serviceRoute traffic to new Travel service
event.new-serviceRoute traffic to new Event service
react.travel-uiShow new React UI for Travel
react.event-uiShow new React UI for Events
ai.smart-routingEnable AI-based API routing
ai.anomaly-detectionEnable AI monitoring alerts
reporting.cqrs-modeUse CQRS read models vs legacy

Tool: Azure App Configuration. Rules: All new services behind flags. Per-tenant and per-region capable. Kill switch → fallback to legacy instantly. Flags cleaned up every cycle.

8. Quality Assurance

Testing Pyramid

LevelToolAI ContributionCoverage Target
Unit TestsxUnit / NUnit80% AI-generated≥80% on new code
Contract TestsPact70% AI-generatedEvery service boundary
Integration TestsDocker Compose in CI50% AI-generatedDB, event bus, APIs
E2E TestsCypress / Playwright30% AI-assistedCritical paths only
Manual TestingHuman exploratory0% AIFriday demo sessions

Quality Gates

PR Level

Build passes
Unit tests (≥80% changed files)
Contract tests pass
SAST clean
CodeRabbit: no critical findings
Human review: approved

Staging Level

Integration tests pass
E2E critical paths pass
Performance baseline not degraded
No new security vulnerabilities

Production Level

All staging gates pass
Feature flag ready (kill switch)
Monitoring/alerting configured
Rollback plan documented
Tech Lead approval

9. Metrics & Reporting

Delivery Metrics

Services migrated5/5 (60% by M6)
API endpoints migrated120 total
React modules live3–4/5
Cycle progress6 cycles on time

Quality Metrics

Test coverage (new code)≥85%
Contract test pass rate100%
Production incidents (P1/P2)0
Zero downtime maintained✅ Yes

AI Metrics

AI-generated code ratio~68%
AI code bug rate vs human0.8x (lower)
AI PR rejection rate~12%
Effective multiplier (measured)1.9x
AI tool cost (monthly)~$1,050

Team Health

Sprint velocity trend↗ ↗ → → (stabilizing)
Team satisfaction (retro)4.2/5
Overtime hours this cycle≤2 (acceptable)
Bus factor per service≥2 (met)

Reporting Cadence

ReportAudienceFrequencyContent
Health DashboardTeamReal-timeAll metrics above (live board)
Weekly DemoProduct Owner + BusinessWeeklyWorking features + metrics
Exec SummaryC-Level / SponsorBi-weekly1-page: milestones, risks, decisions
AI ROI ReportSponsorMonthlyAI cost vs productivity, quality comparison
Cycle ReportAll stakeholdersEvery 6 weeksFull review: delivered, deferred, learnings
Post-MortemTeam + stakeholdersPer P1/P2 incidentRCA, prevention, action items

10. Knowledge Management

Documentation Hierarchy

/docs/
├── adrs/ ← Architecture Decision Records
├── runbooks/ ← Operational runbooks
├── api/ ← OpenAPI specs (auto-generated)
├── onboarding/ ← New member guides
└── migration/ ← Migration-specific docs

Rule: Docs live with code (in repo). No separate wiki — prevents doc drift. AI generates first draft, human reviews.

New Engineer Onboarding

Day 1

Dev env setup (AI-assisted), read architecture + AI workflow guides, access repos/CI/Azure

Day 2–3

Pair with Senior, run test suite locally, deploy to staging, review 3 recent PRs

Day 4–5

First small task (bug fix), full PR flow, first AI-assisted dev task

Week 2

Own small feature end-to-end, attend architecture review, AI code walkthrough

Target: Productive contributor by Day 10. AI tools reduce onboarding time by ~40%.

11. Continuous Improvement

Retrospective Framework (Cycle-end, 1 hour)

START

  • • What should we start doing?
  • • New AI tools/prompts to try?
  • • What process is missing?

STOP

  • • What's wasting our time?
  • • Which AI patterns aren't working?
  • • What ceremonies are useless?

CONTINUE

  • • What's working well?
  • • Highest ROI AI workflows?
  • • What should we NOT change?

AI-Specific Questions (added 15 min)

  • • Where did AI help most this cycle?
  • • Where did AI cause rework? (hallucination tracking)
  • • Is our 2x multiplier holding? Actual measurement?
  • • Any prompt library updates needed?
  • • AI governance: any near-misses?

Learning Budget

Per engineer, per cycle (6 weeks): 4 hours intentional learning + 2 hours AI experimentation. Cooldown week (Week 6): focus on tech debt, learning, experimentation. Investment: ~6 hours per cycle = 1.5% of working time.