What Boards Still Miss About Assessing Senior Leaders

14 min read

How much of the evidence behind your last senior appointment would survive a serious review?
The claim hidden inside every senior appointment

In my experience, every senior appointment contains a claim. Not simply that one candidate looked stronger on the day, but that the board had enough evidence to say this named person was ready for this named role, on this horizon, under these conditions.

That is rarely how the claim is stated in the room. Senior candidates are impressive. The discussions are serious. The process can look robust. But the research base behind serious assessment is older and more exacting than many appointment records reflect, and it does not support several habits that still shape senior appointment decisions.

Executive appointments are not recruitment. At board level, they are governance judgments. They are expensive to correct, slow to reverse, and hard to explain after the fact. Once the decision is made, the process becomes protected by the outcome. Good hire. the assessment worked. Disappointing hire, a one-off judgment error. Neither tells the board whether the process itself was sound.

I worked with a board that spent four months choosing a divisional CEO. Three credible candidates. Competency framework. Panel interviews. Reference checks. On paper, the process looked serious. Within eighteen months, the appointment had failed.

The role required a turnaround operator. The process never truly tested for it. That was not because the board ignored evidence. It was because the process looked structured while comparability kept drifting. The role brief was generic. The interviews were not genuinely comparable. Different panellists weighted different qualities. The references confirmed reputation more readily than role-specific capability. Additional assessment input arrived after the board had effectively formed its view. What looked like rigour was a sequence of impressions held together by confident language.

The research base that would have tightened that process is not new. It is decades old. It does not remove judgment. It does answer a narrower and more useful question: what claim is the board actually entitled to make, on this evidence, before it appoints?

What the research supports

For more than four decades, the personnel selection literature has not been asking whether assessment matters. It has been asking which methods justify stronger claims about later job performance, under what conditions, and with what loss of confidence once design discipline falls away.

On that point, the research is firmer than many senior appointment records suggest. Structure matters. Interviews become more defensible when role requirements are explicit, candidates face comparable questions, answers are evaluated against pre-declared criteria, and multiple assessors work from the same frame (Levashina et al., 2014). Without that discipline, charm, chemistry, seniority bias, and retrospective rationalisation start doing the real work while the process continues to present itself as objective.

The classic selection literature reported meaningful criterion-related validity across a range of methods, with stronger support for some than others (Schmidt & Hunter, 1998). Assessment centres also show useful validity, but their value still depends on design discipline, assessor quality, and role relevance (Gaugler et al., 1987). A polished method is not the same thing as a well-supported claim.

The evidence has also become more demanding over time. Sackett and colleagues revisited influential validity estimates and argued that some widely cited corrections had systematically overstated effect sizes (Sackett et al., 2022). That revision matters. It does not weaken the case for rigour. It raises the bar. Old headline figures are not permanent law. The proper question is narrower, what is supported now, for this method, in this context, and with what limits?

There is a second layer of discipline that matters even more at board level. Messick argued that validity is not a property sitting inside a tool. It is an argument about the interpretation and use of the evidence a process produces (Messick, 1989). Kane extended that logic by insisting that claims require an explicit interpretive framework (Kane, 2013). A board can therefore use a method with real evidential support and still overstate what it learned from it. A score generated for one purpose can be stretched into another. A well-run interview can be treated as though it settled readiness for every future role. The question is not whether a tool sounds serious. It is whether the claim the board is making is proportionate to the evidence it actually has.

What the research does not support

The research does not support a universal method for choosing senior leaders. The evidence base for broad personnel selection is much richer than the evidence base for CEO, ExCo, and board-level appointments specifically. Senior roles are more context-loaded, more political, and less standardised than mid-level hiring. Publicly available evidence rarely provides a board with a clear formula for choosing between two credible finalists.

The research does not support the use of potential as a stable, context-free label. Work on high-potential identification has shown repeatedly how organisations collapse multiple questions into one word: potential for which role, on what time horizon, under what conditions, and on what evidence? (Silzer & Church, 2009; Church & Silzer, 2014). Once that collapse happens, the label starts doing more work than the evidence beneath it.

Nor does the literature support the quiet promotion of soft signals into evidence simply because the candidates are senior. Polish, fluency, sponsor enthusiasm, prestige, and narrative confidence may all matter inside a board process. That is different from saying the research supports them as indicators of later role performance. A board may still choose to weight them. It should record that choice as board judgment, not let it pass as evidence by default.

This is where I abstain from a stronger claim. The literature gives boards better discipline. It does not produce an automatic answer between two credible senior finalists. It tells the board what it is entitled to say with confidence, and where confidence should stop.

Where senior assessment still drifts

The difficulty is not just technical. It is social.

At the top of the house, candidates are usually impressive. Strong biographies. Polished interviews. Persuasive strategic narratives. In that setting, a process can appear robust while becoming less comparable at every stage. Criteria shift midstream. Different assessors privilege different qualities. References confirm prior preferences. A decision emerges. Then a story is written around it.

There is a status trap inside that pattern. I once heard a NomCo Chair say, almost verbatim, “We do not need a process for people at this level. We know quality when we see it.” The confidence was sincere. The risk was real. Structured assessment gets misread as bureaucratic and therefore beneath the most senior decisions. Informal judgment gets mistaken for seasoned wisdom. That is backwards. The higher the stakes, the stronger the case for explicit criteria, explicit evidence classes, and explicit uncertainty.

Category confusion compounds the problem. Succession, selection, development, and cultural fit are not the same decision. A person may be a strong internal successor for one role and a weak candidate for another. A leader may have the biography the board likes and still lack evidence for a turnaround brief, an international expansion mandate, or a regulated-risk environment. Once different decisions collapse into one, broad language starts doing work that role-specific evidence should have done.

That is why boards can spend months on a senior appointment and still end up comparing unlike with unlike. The discussion is lengthy. The notes are full. The process sounds serious. But the record still does not show a stable claim.

Role-Conditional Readiness

Every board-level assessment conversation eventually arrives at the same sentence: this candidate is ready. The room nods. But the sentence only sounds decisive because its missing terms stay implicit. Ready for what, exactly? In what context? Over what horizon? On what evidence? With which gaps still open?

Role-Conditional Readiness starts from one correction. Readiness is not a portable label attached to a person. It is a bounded board claim about the fit between a named person, a named role, a named time horizon, and a named set of decision conditions.

If Messick is right that validity concerns the interpretation and use of evidence rather than the tool in isolation, if Kane is right that claims require an explicit interpretive framework, and if Sackett is right that the inherited headline numbers are less generous than many readers assumed, then readiness cannot be treated as a general quality the board simply sees. It has to be argued, bounded, and recorded. The board is not choosing between evidenced and unevidenced candidates. It is deciding what claim it is entitled to make, and what risk it is consciously accepting.

That is why this is more than assessment hygiene. It changes what the board is accountable for. Under ordinary practice, the verdict is treated as the work. Under Role-Conditional Readiness, the work is to show what the verdict rests on, where the evidence is thinner than the language being used, and which risks are being accepted as board judgment rather than allowed to pass as evidence by default.

Four requirements make it operational.

First, the role is defined before candidates are discussed. Not a vague leadership ideal. A live brief: what outcomes matter, what constraints apply, what inherited context will the person walk into, which capabilities are essential on day one, and which can develop in role.

Second, the evidence classes are separated. A documented track record in a comparable setting is not the same as sponsor confidence. A structured interview is not the same as an unstructured conversation. Direct evidence of prior role-relevant behaviour is not the same as board judgment about how a candidate may grow in role. All may enter the decision. They should not be collapsed into one undifferentiated impression.

Third, the abstentions are named. If the evidence is weak on a crucial criterion, that gap should remain visible in the final discussion. Not every gap should be filled with confidence. Some should remain unresolved and govern the risk conversation.

And finally, the board documents why the final decision still made sense. Not to turn judgment into science. To make judgment auditable. To give the board something it can revisit, learn from, and defend when the question comes back twelve months later. 

Picture a NomCo assessing three candidates for a group CFO role. Before any verdict, the committee agrees the brief: the financial restructuring ahead, the regulatory complexity, the investor-relations burden, and the team the new CFO will inherit. Each candidate is then assessed against that brief, with evidence separated by class and gaps named openly. The committee may still choose the candidate with the thinnest evidence on one dimension. But it does so as a documented risk decision rather than an unexamined assumption.

Role-Conditional Readiness does not promise certainty. It gives the board something better. A cleaner claim.

The appointment record the board should be able to defend

This argument would weaken if boards could show that unstructured, narrative-heavy senior assessment produces outcomes at least as reliable and at least as defensible as structured, role-specific alternatives. It would weaken further if boards routinely kept a clear appointment record: role criteria agreed in advance, evidence separated by class, unresolved gaps left visible, and the final risk decision stated plainly.

Boards do not need less judgment in senior appointments. They need better conditions for judgment. Explicit criteria. Evidence classes that do not blur into one another. Honest abstention when the evidence is thin. A record that shows where the claim stops.

Before the next senior appointment, that is what the board should require. Not a longer process. A more disciplined one.

This is where the board moves from preference to accountable judgment. Before it later attributes outcomes after the fact, or asks where a system’s claim stops, it first has to decide what claim it is entitled to make about readiness at the point of appointment.

Twelve months later, the board should be able to defend not that the candidate interviewed brilliantly, or that the room felt aligned, but that it knew what it had evidence for, what it did not, and what risk it consciously accepted when it made the appointment. That is what Role-Conditional Readiness gives it.

Written by James Nash.

First published on Substack, co-published on inBeta.io. March 2026
Series: The Seven™, by James Nash © inBeta™ Ltd 2026. All rights reserved.

The Author

James Nash

James is the founder of inBeta. He has spent fifteen years working with boards and senior leadership teams at global and publicly listed companies on succession, talent, capability, and leadership governance. He holds executive education from Saïd Business School, University of Oxford, in Artificial Intelligence (including Audit and Ethics), Executive Leadership, Strategic Innovation, and Executive Finance. He founded inBeta because he kept watching boards make their most important decisions on instinct, narrative, and incomplete information, and believed the evidence base existed to do it differently. James is a certified AI Auditor, AI Ethicist, and AI Professional (CAIA, CAIE, CAIP; Oxethica), and a certified practitioner in CliftonStrengths (Gallup), Hogan (including PBC 360), FIRO-B, and Cultural Intelligence (CQC).

Methods appendix. This article forms part of my thinking on evidence-led leadership decisions, a series of pieces I am surfacing through Spring and Summer 2026 about a governance standard for senior appointment decisions, not a single assessment method. The appendix discloses the principles behind that standard at a level appropriate for board review. It does not disclose controlled assessment design, scoring rules, thresholds, or proprietary parameters.

Construct. Role-Conditional Readiness. A documented board discipline for stating what a board is entitled to claim about a named person’s readiness for a named role, on a named horizon, under stated decision conditions. A governance discipline, not a hiring formula.

My intended use. To help boards, NomCo Chairs, CHROs, and General Counsel evaluate the evidential discipline of senior appointment processes and keep an appointment record that can be defended after the fact.

My excluded uses. My writing and thought leadership are my own and do not evaluate any specific board’s past appointment decision. It does not prescribe a universal selection method. It does not treat readiness as a stable, context-free quality. It does not provide legal, hiring, or employment advice.

Abstention conditions. The standard I have written about applies to high-stakes senior appointments where the board is making a consequential readiness claim about an individual. It may not apply in the same form to lower-stakes hiring, internal development conversations, or cases where the evidence base is too thin to support a role-specific claim.

Source classes. Three classes of evidence. First, peer-reviewed research on interview structure, assessment centre validity, revised validity estimates, validity theory, interpretive use, and high-potential labelling: Levashina et al. (2014), Schmidt and Hunter (1998), Gaugler et al. (1987), Messick (1989), Kane (2013), Sackett et al. (2022), Silzer and Church (2009), and Church and Silzer (2014). Second, practitioner observation from my own board and senior-appointment work, where I have seen recurring breakdowns in role definition, evidence comparability, and appointment record-keeping. Third, the governance standard developed across this series, including the Claim Boundary set out more fully in a companion article, which establishes the principle that leadership decisions should state what they do claim, do not claim, and cannot yet claim on the evidence available.

Bibliography

Relevant to: What Boards Still Miss About Assessing Senior Leaders
Published: March 2026

Church, A. H., & Silzer, R. (2014). Going Behind the Corporate Curtain with a Blueprint for Leadership Potential. Human Resource Development Quarterly, 25(1), 3–37. https://doi.org/10.1002/hrdq.21186

Gaugler, B. B., Rosenthal, D. B., Thornton, G. C., & Bentson, C. (1987). Meta-Analysis of Assessment Center Validity. Journal of Applied Psychology, 72(3), 493–511. https://doi.org/10.1037/0021-9010.72.3.493

Kane, M. T. (2013). Validating the Interpretation and Uses of Test Scores. Journal of Educational Measurement, 50(1), 1–73. https://doi.org/10.1111/jedm.12000

Levashina, J., Hartwell, C. J., Morgeson, F. P., & Campion, M. A. (2014). The Structured Employment Interview: Narrative and Quantitative Review of the Research Literature. Personnel Psychology, 67(1), 241–293. https://doi.org/10.1111/peps.12052

Messick, S. (1989). Validity. In R. L. Linn (Ed.), Educational Measurement (3rd ed., pp. 13–103). American Council on Education and Macmillan.

Sackett, P. R., Zhang, C., Berry, C. M., & Lievens, F. (2022). Revisiting Meta-Analytic Estimates of Validity in Personnel Selection: Addressing Systematic Overcorrection for Restriction of Range. Journal of Applied Psychology, 107(11), 2040–2068. https://doi.org/10.1037/apl0000994

Schmidt, F. L., & Hunter, J. E. (1998). The Validity and Utility of Selection Methods in Personnel Psychology. Psychological Bulletin, 124(2), 262–274. https://doi.org/10.1037/0033-2909.124.2.262

Silzer, R., & Church, A. H. (2009). The Pearls and Perils of Identifying Potential. Industrial and Organizational Psychology, 2(4), 377–412. https://doi.org/10.1111/j.1754-9434.2009.01163.x

Academic Journal

Our research.
Published.

We are preparing our first academic publication on leadership intelligence, board governance, and evidence-based succession. Register below for early access when it publishes.

Summer 2026

Academic Journal

FTSE100, Fortune 500, and ASX boards already
use it.

We are preparing our first academic publication on leadership intelligence, board governance, and evidence-based succession. Register below for early access when it publishes.

Summer 2026

Academic Journal

Our research.
Published.

We are preparing our first academic publication on leadership intelligence, board governance, and evidence-based succession. Register below for early access when it publishes.

Summer 2026

Our events

The Breakfast Table

By invitation. Where talent leaders come to think out loud. Evidence, AI, and the future of human capital. Informal. Connected. The conversations that shape how the best teams operate.

The Working Session

By invitation. Board chairs, CEOs, and senior directors exploring the leadership questions that don't have easy answers. Small rooms. Real evidence. Frank conversation.

inBeta Labs

By application. Leadership intelligence for the teams building it every day. What's changing, what's coming, and what the evidence says about how organisations will find, develop, and keep their best people.

The Breakfast Table

By invitation. Where talent leaders come to think out loud. Evidence, AI, and the future of human capital. Informal. Connected. The conversations that shape how the best teams operate.

The Working Session

By invitation. Board chairs, CEOs, and senior directors exploring the leadership questions that don't have easy answers. Small rooms. Real evidence. Frank conversation.

inBeta Labs

By application. Leadership intelligence for the teams building it every day. What's changing, what's coming, and what the evidence says about how organisations will find, develop, and keep their best people.

Apply to come to Labs.

Solutions

Pricing

Company

Legal

Resources

Social

© Copyright 2026 inBeta. inBeta, Optics, Divergence and The Seven are all trademarks of inBeta Ltd