Beyond Single-Specialty Answers: Why Cross-System Reasoning Matters
The Patient Who Crossed Three Specialties
A 58-year-old woman with type 2 diabetes, moderate CKD (eGFR 38 mL/min), and newly diagnosed atrial fibrillation presents to her primary care physician. The clinical question seems straightforward: what is the best anticoagulation strategy?
It is not straightforward. The question spans three domains — cardiology (stroke prevention in AF), nephrology (anticoagulant dosing and bleeding risk in CKD), and endocrinology (diabetes medications that interact with renal function and coagulation pathways). The optimal answer requires synthesizing evidence from all three, and the evidence does not agree.
RE-LY established dabigatran's efficacy in AF but excluded patients with eGFR below 30 and provided limited subgroup data for the 30-50 range. ROCKET AF included patients with moderate renal impairment but used a reduced rivaroxaban dose that has since been debated. ARISTOTLE, studying apixaban, showed preserved efficacy in moderate CKD — but the subgroup with concurrent diabetes showed a non-significant trend toward higher bleeding rates that most guidelines do not address.
A cardiologist recommends apixaban based on ARISTOTLE. A nephrologist flags the bleeding trend in patients with eGFR below 40. An endocrinologist notes that the patient's SGLT2 inhibitor — empagliflozin — has renal hemodynamic effects that could shift her eGFR trajectory, potentially moving her across the dose-adjustment threshold within months.
No single specialist holds all three pieces. And the primary care physician, who needs to write the prescription, may not have time to consult all three or to reconcile contradictory subspecialty recommendations into a coherent plan. This is not an edge case. This is Tuesday.
What is cross-system reasoning?
Cross-system reasoning is the ability to synthesize clinical evidence across multiple organ systems and medical specialties simultaneously, rather than answering from a single-specialty perspective. When a patient's clinical question spans cardiology, nephrology, and endocrinology, cross-system reasoning considers how evidence from each domain interacts — identifying conflicts between specialty guidelines, surfacing subgroup data from trials that enrolled patients with similar comorbidity profiles, and tracing drug-disease interactions across systems. A 2023 BMJ analysis found that among hospitalized adults over 65, 62% had at least one clinically significant drug-disease interaction that crossed specialty boundaries.
How Clinical Decision Support Typically Fails Here
Most clinical decision support tools are structured around single-specialty knowledge domains. Ask about anticoagulation in AF, you get a cardiology answer. Ask about drug dosing in CKD, you get a nephrology answer. The tool does not connect these domains, because its knowledge is organized in the same specialty silos that medical education and hospital departments are organized in.
This is not a critique of any specific tool — it is a structural limitation inherited from how clinical knowledge has been organized for a century. Textbooks divided by organ system. Journals divided by specialty. Guidelines written by specialty societies. The entire infrastructure of medical knowledge production assumes a single-system frame.
But patients do not present within a single system. A 2023 analysis in The BMJ found that among hospitalized adults over 65, the median number of active diagnoses spanning at least two organ systems was 4.2. Among these patients, 62% had at least one clinically significant drug-disease interaction that crossed specialty boundaries. The authors estimated that 23% of adverse drug events in this population involved interactions that no single-specialty guideline would have flagged. The silos are not just an inconvenience — they are a patient safety problem.
The Evidence Fragmentation Problem
Cross-system clinical questions are difficult not just because they span specialties, but because the relevant evidence is scattered across journals, databases, and guideline documents that rarely reference each other.
Take metformin use in patients with both heart failure and CKD. For years, metformin carried a black-box warning against use in heart failure based on theoretical lactic acidosis risk — a warning that originated from phenformin data in the 1970s. By 2020, multiple observational studies and meta-analyses — published across diabetes journals, cardiology journals, and nephrology journals — showed metformin was not only safe in stable heart failure but potentially beneficial.
The ADA updated its guidance. The ACC/AHA heart failure guidelines acknowledged the data but with different emphasis. The KDIGO CKD guidelines addressed the eGFR threshold for metformin but did not specifically discuss heart failure comorbidity. A physician treating a patient with all three conditions needed to cross-reference three separate guideline documents, each updated on different cycles, to arrive at a coherent answer. That is not a reasonable expectation for a 15-minute visit.
In 2024, Bergqvist and colleagues published a systematic review in Annals of Internal Medicine mapping cross-referencing between major specialty guidelines. They examined 47 guidelines across 8 specialties and found that only 12% of recommendations with implications for other organ systems included any reference to the relevant cross-specialty evidence. Cardiology guidelines referenced nephrology data 18% of the time when relevant. Endocrinology referenced cardiology 14% of the time. Nephrology referenced endocrinology 9% of the time. The specialties are barely talking to each other in their own documents.
What Gets Lost Between the Silos
The practical consequence: clinically significant connections get discovered late or not at all. Three examples:
- SGLT2 inhibitors and gout. By 2023, multiple studies showed SGLT2 inhibitors reduce serum uric acid by 10-15% through increased renal urate excretion. Published in nephrology and endocrinology journals. Rheumatology guidelines for gout management did not mention SGLT2 inhibitors as a relevant concomitant therapy until late 2025 — a two-year gap during which patients with diabetes, CKD, and gout could have benefited from a drug that addressed all three conditions simultaneously.
- GLP-1 receptor agonists and NASH. The hepatoprotective effects of semaglutide were demonstrated in the STEP trials and dedicated hepatology studies by 2022. Endocrinologists prescribing semaglutide for diabetes knew. Gastroenterologists managing NASH patients without diabetes were slower to incorporate this evidence, and cross-specialty referral patterns did not consistently connect these findings until 2024.
- Beta-blockers and COPD. For decades, beta-blockers were avoided in COPD patients due to bronchospasm concerns. By 2019, multiple studies including a Cochrane review established that cardioselective beta-blockers were safe in COPD and that withholding them in patients with concurrent heart failure increased mortality. Yet a 2024 survey in Chest found that 34% of pulmonologists still routinely recommended against beta-blocker use in COPD patients with heart failure. The evidence existed for five years. A third of specialists had not adopted it.
What Cross-System Reasoning Looks Like
A clinical tool that reasons across systems would handle the original question — anticoagulation in AF with CKD and diabetes — differently from a single-specialty tool. Instead of a cardiology answer with a footnote about renal dosing, it would simultaneously consider:
- The primary efficacy data for each anticoagulant in AF (cardiology evidence)
- Subgroup analyses for patients with moderate CKD, specifically looking for differential efficacy or bleeding signals in the eGFR 30-45 range (nephrology evidence)
- The effect of concurrent SGLT2 inhibitor therapy on renal hemodynamics and anticipated eGFR trajectory (endocrinology + nephrology evidence)
- Drug-drug interactions between the anticoagulant, the SGLT2 inhibitor, and other diabetes medications (pharmacology evidence)
- Whether any landmark trials enrolled patients with a similar comorbidity profile and what those subgroups showed (cross-referencing trial populations)
The output would not be a single recommendation but a synthesis: here is what the cardiology data shows, here is how the CKD data modifies it, here is the interaction with the patient's diabetes regimen, and here are the specific subgroup data most relevant to this patient's profile. With every citation verified and linked to the original paper.
The Scale Problem
A physician could perform this cross-specialty synthesis manually. Pull up the AF guidelines. Cross-reference the CKD dosing tables. Search PubMed for the specific drug interaction. Read the subgroup analyses from the relevant trials. Academic physicians at tertiary centers do this for complex cases. It takes 45 minutes to two hours per question.
In primary care, where physicians manage panels of 2,000+ patients and average 15-20 minutes per visit, this level of synthesis is not feasible for every clinical question. The evidence exists. The connections exist. But the human capacity to hold all relevant cross-specialty data in working memory while making a time-pressured clinical decision is fundamentally limited. This is not about physician competence — it is about arithmetic. There are more connections than any brain can track in real time.
This is where computation has a genuine role. Not replacing clinical judgment, but performing the cross-referencing, subgroup identification, and evidence synthesis that a physician would do given unlimited time. The physician still interprets the synthesis. The physician still makes the decision. But the physician starts from a complete cross-system picture rather than a single-specialty fragment.
Measuring What Matters
How do you tell whether a clinical tool actually performs cross-system reasoning versus simply retrieving text from multiple specialty databases? The distinction matters: assembling paragraphs from a cardiology source and a nephrology source is not the same as reasoning about how the two domains interact for a specific patient.
Three markers distinguish real cross-system reasoning from superficial multi-specialty retrieval:
- Contradiction identification. When guidelines from two specialties conflict — one recommends a drug, another flags a concern — the tool should surface the conflict explicitly rather than presenting one specialty's recommendation and ignoring the other.
- Subgroup specificity. When a trial enrolled patients with a similar comorbidity profile, the tool should identify and present that subgroup data rather than reporting only the overall trial result. Your patient with AF, CKD, and diabetes deserves to know what happened to patients like her in the trials, not just what happened on average.
- Interaction reasoning. The tool should trace pathways between conditions — such as how an SGLT2 inhibitor's renal hemodynamic effects might influence anticoagulant dosing thresholds — rather than treating each condition as independent.
Ailva traces connections across specialties that no single specialist would see in isolation. When evidence from one domain modifies the answer from another, that connection surfaces — with the specific papers, specific data points, and specific patient populations that make it relevant. See how cross-system reasoning works in practice.
For a concrete example of cross-system evidence synthesis in action, see our deep dive into SGLT2 inhibitors in HFpEF with CKD — a case spanning cardiology, nephrology, and endocrinology. And to understand why citations matter as much as the reasoning itself, read why clinical AI tools hallucinate citations and how to verify them.
Want to try Ailva?
Ailva is a clinical intelligence platform that delivers evidence-based answers with verified citations and cross-system reasoning. Free for all NPI holders.