FirmAdapt
FirmAdapt
Back to Blog
healthcareICD-10medical codingAIautomation

AI for ICD-10 Code Suggestion: How It Actually Works in Practice

By Basel IsmailApril 2, 2026

ICD-10 has 72,000 diagnosis codes, up from the roughly 14,000 in ICD-9. The explosion in specificity was designed to improve data quality for research, public health, and outcomes tracking, but it created a practical problem for coders and physicians. Finding the most accurate code in a system with 72,000 options takes time, and the default human behavior is to fall back on familiar codes that are close enough rather than searching for the precise match.

The Manual Lookup Problem

Traditional ICD-10 coding involves reading a clinical note, identifying the diagnoses, and looking up the appropriate codes using either the tabular index, an electronic code search tool, or memory. Experienced coders can code common conditions from memory quickly, but the specificity requirements of ICD-10 mean that even common conditions have dozens of variations.

Take type 2 diabetes. The category E11 alone has over 80 sub-codes depending on complications: with kidney complications, with ophthalmic complications, with neurological complications, with circulatory complications, and combinations thereof. A patient with type 2 diabetes who has both diabetic neuropathy and diabetic chronic kidney disease stage 3 needs E11.22 and E11.65 at minimum, and possibly additional codes for the specific manifestations.

When coders are processing 40 to 60 charts per day, the temptation to assign E11.9 (type 2 diabetes without complications) instead of the three or four more specific codes is understandable. The claim will still pay. But the coding inaccuracy affects risk adjustment scores, quality metrics, and the practice's data integrity.

How AI Code Suggestion Works

AI-driven ICD-10 suggestion tools use natural language processing to read clinical documentation and extract diagnostic information. The AI does not just look for diagnosis keywords. It understands clinical context well enough to infer diagnoses from the combination of symptoms, test results, and treatment decisions documented in the note.

When a note describes elevated creatinine of 2.4, proteinuria on urinalysis, and a medication list that includes lisinopril adjusted for renal dosing in a patient with documented type 2 diabetes, the AI recognizes the pattern of diabetic chronic kidney disease and suggests the appropriate ICD-10 codes, including the stage based on the GFR calculation it derives from the creatinine value.

The suggestion appears in the coding interface with a confidence score and links to the specific documentation elements that support it. The coder can accept the suggestion with one click, modify it, or reject it. Each interaction feeds back into the model, improving future suggestions for that practice's documentation patterns.

Specificity Improvements

The most measurable impact of AI code suggestion is on code specificity. Studies from 3M Health Information Systems show that AI-assisted coding increases the average specificity level of assigned codes by 15% to 20%. This means fewer unspecified codes (the ones ending in .9 that indicate incomplete specificity) and more codes that capture the full clinical picture.

For risk-adjusted payment models like Medicare Advantage, this specificity improvement directly affects revenue. HCC (Hierarchical Condition Category) risk scores depend on specific diagnosis codes being reported. A patient with E11.22 (type 2 diabetes with diabetic chronic kidney disease) generates a higher and more accurate risk score than one coded with just E11.9 (unspecified diabetes). The revenue difference can be $2,000 to $5,000 per patient per year in capitated payment models.

Even in fee-for-service environments, specificity matters. Some payers adjust reimbursement based on the severity reflected in diagnosis codes. And from a compliance perspective, specific codes supported by documentation are far safer than unspecified codes that might trigger audit questions about documentation adequacy.

Workflow Integration Patterns

The most successful implementations integrate AI code suggestions directly into the coding workflow rather than requiring coders to switch between systems. When the AI suggestion appears alongside the coder's working screen, the review-and-accept workflow adds seconds per chart, not minutes.

Some practices use AI coding at the physician level, embedding suggestions into the EHR's assessment and plan section. As the physician documents a diagnosis, the system suggests the most specific ICD-10 code with a single click to accept. This physician-level coding works well for simple encounters but is less practical for complex cases where professional coders add significant value. Healthcare operations platforms that support both physician-level and coder-level AI integration give practices flexibility to match the tool to the encounter complexity.

For facilities with a coding backlog, AI can be used in pre-coding mode, where it processes notes overnight and presents a coded chart for human review the next morning. This approach can reduce the coding backlog significantly because the human review step is faster than coding from scratch.

Training and Adaptation

AI code suggestion systems perform best when they have been trained on a practice's specific documentation style, specialty focus, and payer mix. A cardiology practice documents differently than a primary care practice, and the AI needs to understand those differences to suggest appropriate codes.

Most vendors offer a calibration period of four to eight weeks where the AI processes historical charts and the coding team provides feedback on suggestion accuracy. After calibration, the system typically achieves 85% to 92% agreement with human coders on the primary diagnosis code and 75% to 85% agreement on the full code set for a given encounter.

The gap between AI and human agreement is not necessarily AI error. In many cases, the AI suggests more specific codes that are technically more accurate than what the human coder assigned, particularly for secondary diagnoses that coders might skip under time pressure. The calibration process helps the coding team and the AI converge on the right level of coding thoroughness.

Limitations Worth Knowing

AI code suggestion is not a replacement for coding expertise. The technology works well for straightforward encounters but still struggles with certain scenarios. Complex surgical cases with multiple procedures, unusual diagnosis combinations, and documentation that is ambiguous or contradictory all challenge AI systems.

The quality of AI suggestions is directly tied to the quality of documentation. If a physician's note is vague or incomplete, the AI cannot infer codes that the documentation does not support. In fact, one of the secondary benefits of AI coding tools is that they highlight documentation gaps by producing low-confidence suggestions or flagging notes where the assessment does not match the documented findings.

For practices considering AI code suggestion, the most realistic expectation is that it will handle 70% to 80% of coding volume with high accuracy, freeing experienced coders to focus their expertise on the 20% to 30% of cases that genuinely require human judgment. That division of labor is where the productivity and accuracy gains compound.

Ready to uncover operational inefficiencies and learn how to fix them with AI?
Try FirmAdapt free with 10 analysis credits. No credit card required.
Get Started Free
AI for ICD-10 Code Suggestion: How It Actually Works in Practice | FirmAdapt | FirmAdapt