This diagnostic study evaluated the performance of an AI algorithm vs. radiology residents in interpreting anteroposterior (AP) frontal chest radiographs in emergency and inpatient settings. The AI model, trained on 342,126 frontal chest X-rays from multiple hospitals, was tested against five U.S.-based third-year radiology residents using a separate dataset of 1,998 X-rays with ground truth labels established via triple consensus adjudication. The AI achieved a mean AUC of 0.807 (training) and 0.772 (field testing), with PPV (0.730) and specificity (0.980) significantly exceeding those of radiology residents (PPV: 0.682, specificity: 0.973; P < .001), while sensitivity remained comparable (AI: 0.716, Residents: 0.720; P = .66). These findings suggest that AI algorithms can match or outperform radiology residents in preliminary CXR reads, particularly in routine cases, but expert oversight is still required for complex findings. Integrating AI into radiology workflows could enhance efficiency, improve diagnostic accuracy, and address resource shortages in emergency care.
