A reliability profile is a dispositional concept in the Rylean sense: a characterization of how reliably a system exercises its dispositions across the range of conditions it encounters. Every disposition has a profile — solubility is reliable across most aqueous solutions but fails in saturated ones; a surgeon's diagnostic disposition is reliable for common presentations but may fail for rare ones. The profile is the practical information that matters. It tells the user how much independent verification the system's outputs require, under what circumstances the system can be trusted, and where its limitations demand human compensation. Claude has a specific reliability profile — extremely reliable for fluent prose, highly reliable for working code, notably less reliable for substantive philosophical accuracy, poor at self-correction — and understanding this profile is the practical precondition for productive collaboration.
There is a parallel reading that begins not with the dispositions themselves but with the material and political conditions that determine which profiles get built and whose interests they serve. The reliability profile, in this view, is not a neutral technical property but a politically constructed artifact—shaped less by the inherent capabilities of training methods than by the economic imperatives of those who control the compute, data, and capital required to build these systems at scale.
Consider the specific weakness Segal identifies—self-correction. This isn't merely a technical limitation arising from training-on-text versus iterative practice. It's a feature that serves the interests of AI providers perfectly. A system that cannot reliably correct itself requires constant human oversight, creating a permanent dependency relationship. Users must maintain subscriptions, hire consultants, build verification workflows. The profile's shape—fluent but unreliable, confident but error-prone—generates maximum engagement while minimizing liability. The system is reliable enough to be addictive but unreliable enough to disclaim responsibility. Moreover, the very framework of 'profiles' individualizes what is fundamentally a collective problem. When we speak of matching profiles to tasks and compensating for weaknesses, we place the burden on individual users to navigate systems whose limitations are structurally determined by concentrated corporate power. The Deleuze error isn't just a quirk of training; it's the predictable result of systems optimized for engagement over accuracy, built by companies whose incentive is to maximize usage, not truth. The profile framework, while analytically useful, risks naturalizing these political choices as technical necessities.
The reliability profile is the empirical content of the dispositional analysis. Once we stop asking whether the machine 'really' thinks and start asking what its behavioral dispositions actually are, the profile becomes the principal object of study. It is measurable, comparable across systems, improvable through targeted intervention. It is also the scientifically tractable version of the questions the ghost question has been preventing.
Reliability profiles are shaped by the training history that built the dispositions. A human expert's profile reflects years of iterative practice: doing the work, making errors, receiving corrections, adjusting behavior, doing the work again. Each iteration narrows the range of likely errors and expands the range of conditions under which the dispositions produce correct responses. Claude's profile reflects a different process — training on text rather than iterative practice in the world — and the difference shows up in specific, characterizable ways.
The most important feature of Claude's profile, for practical purposes, is the weakness of its self-correction dispositions. The capacity to notice one's own errors is one of the most significant components of the dispositional cluster that constitutes intelligence, and it is the component most difficult to build without iterative feedback. Claude is disposed to produce rhetorically coherent output; it is not reliably disposed to check that output against the specific content of the concepts it invokes. The Deleuze error is a paradigm case: a fluent, plausible, coherent passage that misuses a philosophical concept in a way obvious to anyone who has actually read Deleuze.
The practical upshot: the tool is trustworthy in proportion to the match between its profile and the task. For fluent prose generation, Claude's profile matches well; light verification suffices. For substantive philosophical accuracy, the profile matches poorly; heavy verification is required, ideally by someone whose own profile includes the capacity to detect Claude's characteristic errors. The discipline Segal describes — rejecting Claude's output when it sounds better than it thinks — is exactly the exercise of human disposition to compensate for machine disposition.
The concept is developed in the Ryle volume's chapter 4 as the empirical operationalization of dispositional analysis for AI systems. The underlying framework derives from Ryle's treatment of dispositions as real but variably-conditioned properties.
The vocabulary also draws on reliability engineering and from contemporary AI evaluation practice, which has been converging on profile-based characterization as more informative than single-benchmark scores.
Not binary, but conditional. A reliability profile is not a yes/no verdict but a map of where and under what conditions a disposition is trustworthy.
Shaped by training history. The specific process that built the dispositions determines the profile. Different training produces different profiles, even for the same nominal capability.
Self-correction is the key weakness. Claude's profile is specifically weak in self-correction, because self-correction requires iterative feedback loops that training-on-text does not provide.
Profile-match is the practical criterion. A tool is trustworthy in proportion to how well its profile matches the task. Mismatch demands human compensation.
Critics of profile-based characterization argue that it treats AI systems as fixed artifacts when they are rapidly evolving, and that profiles built on current behavior may not generalize to the next generation. The response is that while specific profiles change, the framework for describing them remains useful — and that the question of whether next-generation systems have substantially different profiles is itself an empirical question the framework helps pose precisely.
The reliability profile concept operates at multiple levels simultaneously, and the relative importance of Segal's dispositional framing versus the contrarian's political-economic framing shifts depending on which question we're asking. For understanding what Claude can do right now—the practical question of how to use today's tools—Segal's framework dominates (85%). The profile accurately describes the empirical reality users face: Claude really does produce fluent prose, really does struggle with self-correction, really does require specific compensation strategies. This descriptive accuracy matters enormously for practitioners.
But when we ask why the profile has this particular shape, the political-economic analysis becomes primary (70%). The contrarian correctly identifies that technical limitations aren't purely technical—they emerge from specific institutional arrangements, incentive structures, and power relations. The weakness in self-correction isn't just about training methods; it's also about what kinds of capabilities are prioritized by companies operating under specific market pressures. The profile reflects both computational constraints and corporate choices.
The synthetic frame that holds both views recognizes profiles as multiply determined phenomena. They are simultaneously technical artifacts (shaped by training methods and computational limits), political constructs (shaped by who controls resources and toward what ends), and practical realities (creating specific affordances and constraints for users). The proper response isn't to choose between these levels but to track how they interact. A complete reliability profile would describe not just the behavioral dispositions but also the conditions of their production—making visible both what the system can do and why it can't do more. This expanded profile concept retains Segal's practical utility while incorporating the contrarian's critical insights about the forces shaping these tools.