Purpose of the Integrated Baseline Review
The Integrated Baseline Review is the government’s opportunity to verify that the contractor’s Performance Measurement Baseline (PMB) is realistic, comprehensive, and executable before significant work begins. It is not an audit — it is a joint assessment. The IBR answers one question: does this plan have a reasonable chance of being executed within the proposed cost and schedule?
Without an IBR, problems hide in the baseline for months. A control account budgeted at 2,000 hours may actually require 3,500 hours — but nobody asks until the variance report shows –40% at month 12. By then, the overrun is irreversible. The IBR is designed to catch these issues at month 3, when corrective action is still possible.
The IBR also establishes a shared understanding between the government and contractor about what the program will deliver, when, and for how much. This shared baseline becomes the foundation for every future performance discussion, variance analysis, and EAC assessment.
IBR Preparation Checklist
Preparation determines IBR success. Programs that invest 4–6 weeks in structured preparation pass cleanly. Programs that treat the IBR as a briefing exercise generate dozens of action items and risk losing government confidence in the baseline.
| Category | Preparation Item | Ready Criteria |
|---|---|---|
| Baseline | PMB fully loaded in the EVMS tool | All CAs have time-phased budgets; total equals Contract Budget Base |
| Baseline | Management Reserve (MR) identified | MR is reasonable (typically 5–15% of PMB); documented rationale exists |
| Schedule | IMS loaded and statused | All tasks linked; critical path identified; no open-ended activities; resource-loaded |
| Schedule | Schedule aligns with cost baseline | Task durations × resource rates = CA budgets (within 5% tolerance) |
| Risk | Risk register populated | Each CA has identified risks; handling strategies defined; cost/schedule impacts quantified |
| CAMs | CAM training complete | Every CAM can explain their scope, schedule, budget, EV method, and risks |
| CAMs | Work packages defined | Near-term work (6 months) has discrete work packages; far-term has planning packages |
| Documentation | System Description current | Describes EVMS processes, tools, thresholds, and organizational structure |
💡 The Pre-IBR Dry Run
Conduct an internal dry run 2–3 weeks before the IBR. Assign senior staff to role-play government reviewers. Interview every CAM using the same questions the government will ask. This serves two purposes: it identifies problems while there is still time to fix them, and it gives CAMs practice articulating their plans under questioning. Programs that skip the dry run are consistently surprised during the real IBR — and surprises during an IBR are never good.
What Reviewers Look For: Three Dimensions
Government reviewers assess the baseline across three integrated dimensions. A plan that is strong in one dimension but weak in another will not pass. The three dimensions must be internally consistent — the schedule must support the technical approach, the budget must support the schedule, and the risks must be reflected in both.
Technical Risk
- Is the technical approach feasible?
- Are technology readiness levels (TRLs) realistic?
- Have key technical risks been identified with mitigation plans?
- Are test and verification plans adequate?
- Does the SOW scope trace to WBS elements completely?
Schedule Logic
- Is the IMS fully networked with valid logic?
- Are constraints justified (not artificial)?
- Does the critical path make sense technically?
- Is schedule margin adequate for high-risk paths?
- Are external dependencies (GFE, subcontractors) identified?
Resource Loading
- Are labor hours and rates realistic for the skill mix?
- Are resources available when the schedule requires them?
- Are overhead and indirect rates current and justified?
- Is material lead time reflected in the schedule?
- Is subcontractor budget consistent with proposals/quotes?
Conducting the Review
The IBR typically spans 3–5 days for a major program. The structure follows a pattern: program-level overview first, then detailed control account interviews, followed by cross-cutting assessments and a closeout session.
| Day | Activity | Participants | Focus |
|---|---|---|---|
| Day 1 | Program Overview Briefing | Full team | Technical approach, WBS, OBS, master schedule, PMB summary, risk overview |
| Day 2–3 | Control Account Deep Dives | Reviewers + individual CAMs | Scope, schedule, budget, EV methods, risks for each CA; 30–60 min each |
| Day 4 | Cross-Cutting Reviews | Reviewers + functional leads | Schedule network integrity, resource conflicts across CAs, subcontract management, indirect rates |
| Day 5 | Findings and Closeout | Full team | Preliminary findings, action item review, severity ratings, closure plan |
Government reviewers will ask each CAM a standard set of questions. Every CAM should be able to answer these without hesitation:
| # | Question | What They Are Really Assessing |
|---|---|---|
| 1 | “Walk me through your scope of work.” | Does the CAM understand their deliverables? |
| 2 | “What is your basis of estimate for hours and dollars?” | Is the budget grounded in data or guesswork? |
| 3 | “Show me your schedule and critical path.” | Is the schedule realistic and logic-driven? |
| 4 | “What are your top 3 risks?” | Has the CAM thought about what could go wrong? |
| 5 | “How will you measure earned value?” | Is the EV method objective and auditable? |
| 6 | “Where is your budget most likely to be insufficient?” | Honesty and self-awareness — the most revealing question |
The last question is critical. A CAM who says “I’m confident in the entire budget” either does not understand their risks or is not being candid. Experienced reviewers trust CAMs who can articulate where the plan is weakest.
Post-IBR: Action Items and the Risk-Adjusted Baseline
The IBR produces a list of findings categorized by severity. Each finding becomes an action item with an owner, due date, and closure criteria.
| Severity | Definition | Example | Typical Resolution |
|---|---|---|---|
| Critical | Baseline is unreliable in this area; execution risk is high | Schedule has no logic linking integration tasks to test; budget assumes rates 30% below actuals | Baseline replanning required; may delay PMB acceptance |
| Major | Significant gap that must be resolved but baseline is usable | Risk register missing for 3 of 12 CAs; material lead times not in schedule | 30-day corrective action; tracked to closure |
| Minor | Documentation or process gap; low execution risk | EV method descriptions incomplete; work authorization documents missing signatures | 60-day administrative closure |
After action items are resolved, the program establishes a risk-adjusted baseline. This is the PMB supplemented by a quantified risk assessment: the program knows the baseline cost is $50M, but the risk-adjusted range is $52M–$58M based on identified risks and their probabilities. This range becomes the context for every future EAC discussion.
⚠️ Do Not Treat the IBR as a Checkbox
Some programs view the IBR as a compliance hurdle — something to “get through” so they can start executing. This attitude guarantees a painful review and a weak baseline. The IBR is the last opportunity to fix the plan before spending real money. Every problem found during the IBR is a problem that will not appear as a surprise variance 12 months from now. Programs that embrace the IBR as a genuine planning validation consistently outperform those that treat it as overhead.
🎯 The Bottom Line
The IBR validates that the Performance Measurement Baseline is realistic, comprehensive, and executable before significant spending begins. Prepare rigorously — train every CAM, conduct a dry run, and ensure technical risk, schedule logic, and resource loading are internally consistent. Treat findings as opportunities to strengthen the plan, not as failures. The risk-adjusted baseline that emerges from a thorough IBR is the foundation for credible performance measurement throughout the program. Next: Monthly CPR/IPMR Generation — turning performance data into the deliverable report.
Stop reading, start modeling
Model your process flow, run simulations, optimize staffing with TOC math, and test your knowledge with 107 interactive checks — all in one platform.