{"id":1088394,"date":"2024-09-30T12:00:00","date_gmt":"2024-09-30T19:00:00","guid":{"rendered":"https:\/\/www.microsoft.com\/en-us\/research\/?p=1088394"},"modified":"2024-11-05T06:42:27","modified_gmt":"2024-11-05T14:42:27","slug":"stress-testing-biomedical-vision-models-with-radedit-a-synthetic-data-approach-for-robust-model-deployment","status":"publish","type":"post","link":"https:\/\/www.microsoft.com\/en-us\/research\/blog\/stress-testing-biomedical-vision-models-with-radedit-a-synthetic-data-approach-for-robust-model-deployment\/","title":{"rendered":"Stress-testing biomedical vision models with RadEdit: A synthetic data approach for robust model deployment"},"content":{"rendered":"\n<p class=\"has-text-align-center\"><strong><em>This paper has been accepted at the <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" target=\"_blank\" href=\"https:\/\/eccv2024.ecva.net\/\">18th European Conference on Computer Vision (ECCV 2024)<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>, the premier gathering on computer vision and machine learning.<\/em><\/strong><\/p>\n\n\n\n<figure class=\"wp-block-image size-full\"><img loading=\"lazy\" decoding=\"async\" width=\"1400\" height=\"788\" src=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/09\/RadEdit-Socials-2024-BlogHeroFeature-1400x788-1.jpg\" alt=\" On the left is a simple drawing of the lungs. The drawing shows the borders of the left and right lung as well as the trachea and the left and right main stem bronchi. The text under the drawing reads: Original image. To the right of the drawing are the 3 additional inputs of RadEdit. They are arranged vertically. On top there is an example editing prompt. It reads \"Consolidation\". Below there is the same drawing of the lung again but this time the left lung is shaded blue. The text reads: Edit mask according to prompt. Lastly, on the bottom, there is the same drawing of the lung but this time the right lung is shaded red. The text reads: \"Do not edit mask\". On the right of the 3 additional inputs there is a box saying \u201cRadEdit\u201d. Finally, on the right of the figure, there is the drawing of the lung again. The upper part of the left lung is shaded grey. The text reads: Edited image. Between all the elements, the drawing of the lung, the 3 additional inputs, the box that says \u201cRadEdit\u201d, and the edited image, there are arrows pointing to the next element from left to right. \" class=\"wp-image-1089042\" srcset=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/09\/RadEdit-Socials-2024-BlogHeroFeature-1400x788-1.jpg 1400w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/09\/RadEdit-Socials-2024-BlogHeroFeature-1400x788-1-300x169.jpg 300w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/09\/RadEdit-Socials-2024-BlogHeroFeature-1400x788-1-1024x576.jpg 1024w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/09\/RadEdit-Socials-2024-BlogHeroFeature-1400x788-1-768x432.jpg 768w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/09\/RadEdit-Socials-2024-BlogHeroFeature-1400x788-1-1066x600.jpg 1066w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/09\/RadEdit-Socials-2024-BlogHeroFeature-1400x788-1-655x368.jpg 655w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/09\/RadEdit-Socials-2024-BlogHeroFeature-1400x788-1-240x135.jpg 240w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/09\/RadEdit-Socials-2024-BlogHeroFeature-1400x788-1-640x360.jpg 640w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/09\/RadEdit-Socials-2024-BlogHeroFeature-1400x788-1-960x540.jpg 960w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/09\/RadEdit-Socials-2024-BlogHeroFeature-1400x788-1-1280x720.jpg 1280w\" sizes=\"auto, (max-width: 1400px) 100vw, 1400px\" \/><\/figure>\n\n\n\n<p>Biomedical vision models are computational tools that analyze medical images, like X-rays, MRIs, and CT scans, and are used to predict medical conditions and outcomes. These models assist medical practitioners in disease diagnosis, treatment planning, disease monitoring, and risk assessment. However, datasets used to train these models can be small and not representative of real-world conditions, which often leads to these models performing worse in actual medical settings. To avoid misdiagnoses and other errors, these models must be rigorously tested and adjusted to perform reliably across different conditions.<\/p>\n\n\n\n<p>To mitigate the dataset challenge of not having enough diverse data and to improve the testing of biomedical vision models, we developed \u201c<a href=\"https:\/\/www.microsoft.com\/en-us\/research\/publication\/radedit-stress-testing-biomedical-vision-models-via-diffusion-image-editing\/\">RadEdit: Stress-testing biomedical vision models via diffusion image editing<\/a>,\u201d presented at ECCV 2024. Aligned with the <a href=\"https:\/\/www.microsoft.com\/en-us\/ai\/principles-and-approach\/?msockid=37f5423dfa656a721ad151fefb9e6b78\">Microsoft Responsible AI principles<\/a> of reliability and safety, RadEdit helps researchers identify when and how models might fail before they are deployed in a medical setting. RadEdit uses generative image editing to simulate different dataset shifts (e.g., a shift in the patients&#8217; demographics), helping researchers to identify weaknesses in the model. By employing text-to-image diffusion models trained on a wide array of chest X-ray datasets, RadEdit can generate synthetic yet realistic X-rays.<\/p>\n\n\n\n<p>RadEdit\u2019s approach involves using multiple image masks (binary images representing designated regions of a reference image), as illustrated in Figure 1, to limit changes to specific areas of the image, therefore preserving their integrity. It generates synthetic datasets free from spurious correlations and artifacts, addressing shortcomings in existing editing techniques. Traditional editing techniques often overlook biases within the generative model, leading to synthetic data that perpetuate these biases. Alternatively, these other editing techniques restrict edits to the point of unrealistic outputs.<\/p>\n\n\n\n\t<div class=\"border-bottom border-top border-gray-300 mt-5 mb-5 msr-promo text-center text-md-left alignwide\" data-bi-aN=\"promo\" data-bi-id=\"1160910\">\n\t\t\n\n\t\t<p class=\"msr-promo__label text-gray-800 text-center text-uppercase\">\n\t\t<span class=\"px-4 bg-white display-inline-block font-weight-semibold small\">video series<\/span>\n\t<\/p>\n\t\n\t<div class=\"row pt-3 pb-4 align-items-center\">\n\t\t\t\t\t\t<div class=\"msr-promo__media col-12 col-md-5\">\n\t\t\t\t<a class=\"bg-gray-300 display-block\" href=\"https:\/\/www.microsoft.com\/en-us\/research\/story\/on-second-thought\/\" aria-label=\"On Second Thought\" data-bi-cN=\"On Second Thought\" target=\"_blank\">\n\t\t\t\t\t<img decoding=\"async\" class=\"w-100 display-block\" src=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2026\/01\/MFST_feature_SecondThought_1400x788.jpg\" alt=\"On Second Thought with Sinead Bovell\" \/>\n\t\t\t\t<\/a>\n\t\t\t<\/div>\n\t\t\t\n\t\t\t<div class=\"msr-promo__content p-3 px-5 col-12 col-md\">\n\n\t\t\t\t\t\t\t\t\t<h2 class=\"h4\">On Second Thought<\/h2>\n\t\t\t\t\n\t\t\t\t\t\t\t\t<p id=\"on-second-thought\" class=\"large\">A video series with Sinead Bovell built around the questions everyone\u2019s asking about AI. With expert voices from across Microsoft, we break down the tension and promise of this rapidly changing technology, exploring what\u2019s evolving and what\u2019s possible.<\/p>\n\t\t\t\t\n\t\t\t\t\t\t\t\t<div class=\"wp-block-buttons justify-content-center justify-content-md-start\">\n\t\t\t\t\t<div class=\"wp-block-button\">\n\t\t\t\t\t\t<a href=\"https:\/\/www.microsoft.com\/en-us\/research\/story\/on-second-thought\/\" aria-describedby=\"on-second-thought\" class=\"btn btn-brand glyph-append glyph-append-chevron-right\" data-bi-cN=\"On Second Thought\" target=\"_blank\">\n\t\t\t\t\t\t\tExplore the series\t\t\t\t\t\t<\/a>\n\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t\t\t\t<\/div><!--\/.msr-promo__content-->\n\t<\/div><!--\/.msr-promo__inner-wrap-->\n\t<\/div><!--\/.msr-promo-->\n\t\n\n\n<h2 class=\"wp-block-heading\" id=\"how-radedit-works\">How RadEdit works<\/h2>\n\n\n\n<p>RadEdit improves biomedical image editing using three key inputs, as illustrated in Figure 1:<\/p>\n\n\n\n<ul class=\"wp-block-list\">\n<li><strong>Text prompt<\/strong>: Defines the desired modifications. For example, a disease can be added with a description like \u201cConsolidation\u201d<\/li>\n\n\n\n<li><strong>Edit mask<\/strong>: A binary mask indicating the main area to be modified, such as the \u201cright lung\u201d<\/li>\n\n\n\n<li><strong>Keep mask<\/strong>: A binary mask outlining parts of the original image to be preserved, like the \u201cleft lung\u201d<\/li>\n<\/ul>\n\n\n\n<figure class=\"wp-block-image aligncenter size-full\"><img loading=\"lazy\" decoding=\"async\" width=\"1400\" height=\"788\" src=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/09\/RadEdit-2024-Consolidation-BlogHeroFeature.jpg\" alt=\" On the left is a simple drawing of the lungs. The drawing shows the borders of the left and right lung as well as the trachea and the left and right main stem bronchi. The text under the drawing reads: Original image. To the right of the drawing are the 3 additional inputs of RadEdit. They are arranged vertically. On top there is an example editing prompt. It reads \"Consolidation\". Below there is the same drawing of the lung again but this time the left lung is shaded blue. The text reads: Edit mask according to prompt. Lastly, on the bottom, there is the same drawing of the lung but this time the right lung is shaded red. The text reads: \"Do not edit mask\". On the right of the 3 additional inputs there is a box saying \u201cRadEdit\u201d. Finally, on the right of the figure, there is the drawing of the lung again. The upper part of the left lung is shaded grey. The text reads: Edited image. Between all the elements, the drawing of the lung, the 3 additional inputs, the box that says \u201cRadEdit\u201d, and the edited image, there are arrows pointing to the next element from left to right. \" class=\"wp-image-1088415\" srcset=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/09\/RadEdit-2024-Consolidation-BlogHeroFeature.jpg 1400w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/09\/RadEdit-2024-Consolidation-BlogHeroFeature-300x169.jpg 300w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/09\/RadEdit-2024-Consolidation-BlogHeroFeature-1024x576.jpg 1024w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/09\/RadEdit-2024-Consolidation-BlogHeroFeature-768x432.jpg 768w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/09\/RadEdit-2024-Consolidation-BlogHeroFeature-1066x600.jpg 1066w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/09\/RadEdit-2024-Consolidation-BlogHeroFeature-655x368.jpg 655w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/09\/RadEdit-2024-Consolidation-BlogHeroFeature-240x135.jpg 240w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/09\/RadEdit-2024-Consolidation-BlogHeroFeature-640x360.jpg 640w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/09\/RadEdit-2024-Consolidation-BlogHeroFeature-960x540.jpg 960w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/09\/RadEdit-2024-Consolidation-BlogHeroFeature-1280x720.jpg 1280w\" sizes=\"auto, (max-width: 1400px) 100vw, 1400px\" \/><figcaption class=\"wp-element-caption\">Figure 1: RadEdit\u2019s inputs and outputs. By using separate \u201cedit\u201d and \u201ckeep\u201d masks, RadEdit can make the desired modifications to an image with precise spatial control and realistic output.<\/figcaption><\/figure>\n\n\n\n<p>RadEdit depends on a diffusion model for image editing, where the image is first converted to a latent noise representation by inverting the diffusion generative process. The noise representation is then iteratively denoised over multiple time steps. During each step, RadEdit:<\/p>\n\n\n\n<ol class=\"wp-block-list\">\n<li>Uses the text prompt to conditionally generate pixels within the edit mask with classifier-free guidance.<\/li>\n\n\n\n<li>Generates the remaining pixels based on the original image and edited area.<\/li>\n\n\n\n<li>Replicates the content of the original image within the \u201ckeep\u201d mask, ensuring that this area remains unaltered.<\/li>\n<\/ol>\n\n\n\n<p>Finally, a quality check ensures that the edited image is faithful to the editing prompt. RadEdit uses Microsoft&#8217;s <a href=\"https:\/\/www.microsoft.com\/en-us\/research\/publication\/learning-to-exploit-temporal-structure-for-biomedical-vision-language-processing\/?msockid=2d32e418b7106f871a12f664b6ba6e2e\">BioViL-T<\/a> to compute an image-text alignment score that we can then use to filter out low-quality and unfaithful edits.<\/p>\n\n\n\n<h2 class=\"wp-block-heading\" id=\"simulating-dataset-shifts\">Simulating dataset shifts<\/h2>\n\n\n\n<p>A key feature of RadEdit is its ability to simulate dataset shifts with precise spatial control for comprehensive model performance evaluation. This includes differences in image acquisition, the appearance of underlying pathologies, and population characteristics.<\/p>\n\n\n\n<p>Particularly notable is RadEdit&#8217;s ability to simulate image variations from different sources (e.g., different hospitals), helping researchers identify potential biases in models trained solely on data from one source. For example, in a COVID-19 study, if all positive cases in a dataset come from a single hospital and all negative cases come from a different hospital, a model trained on detecting COVID-19 might over-rely on hospital-specific indicators from the X-ray images. Among others, we considered the laterality markers in the corners of an X-ray (e.g., a highly visible letter &#8220;L&#8221; on the left side of the X-ray) as well as the amount of black space on the image edges to be hospital-specific indicators. To test if a model relies too much on differences in image acquisition, we created synthetic data using RadEdit, where we removed COVID-19 features while retaining hospital-specific indicators. After creating the synthetic dataset with the COVID-19 features no longer present, we can test if the COVID-10 detection model still predicts COVID-19. This would indicate that the model is biased with respect to hospital-specific indicators.<\/p>\n\n\n\n<p>RadEdit can also remove specific diseases, like pneumothorax (collapsed lung), from an image while keeping treatment features like chest drains. This helps researchers understand how models detect and understand \u201cvisual shortcuts.\u201d Because RadEdit maintains the size and location of the main anatomical structures (like lungs, ribs, and heart), it can also be used to stress-test segmentation models. For example, RadEdit can add rare abnormalities or medical devices to lung images to test how well segmentation models handle new variations, ensuring they generalize accurately across different populations. Figure 2 illustrates these three examples of stress-testing scenarios.<\/p>\n\n\n\n<figure class=\"wp-block-image aligncenter size-full\"><img loading=\"lazy\" decoding=\"async\" width=\"1195\" height=\"788\" src=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/09\/NEW_RadEdit-2024-Figure-1-BlogHeroFeature-1195x788-15Aug24.jpg\" alt=\"All drawings of lungs are the same as in Figure 1. The drawing shows the borders of the left and right lung as well as the trachea and the left and right main stem bronchi. In the first row on the left there are two drawings of a lung. The first drawing of a lung labelled \"COVID-19\" has an arrow pointing upwards in the upper right corner of the drawing, outside of the lung. The arrow is meant to indicate a marking on an X-ray. In addition, the lungs are shaded light grey. The light grey shading is meant to represent the effects of covid on the lung. The second drawing of a lung on the right labelled \"No COVID-19\" has no arrow marking and the lungs are shaded in a dark grey. On top of these two drawings the text says, \"Biased training datasets\". Right next to the two drawings of lungs is a big arrow pointing to the right. The arrow is labelled \"acquisition shift\". On the right end of the arrow is another drawing of a lung. The lungs in the drawing are shaded dark grey, in addition the drawing features an arrow marking in the upper right corner. The drawing is labelled \"No COVID-19\", next to the drawing is a smaller arrow pointing to the right. At the right end of the arrow there is a text reading \"False positive\". \n \nIn the second row on the left there are again two drawings of a lung. The first drawing of a lung labelled \"PTX\" has a tube in the left lung indicated by a white line. In addition, the top of the left lung is black indicating a pneumothorax. The second drawing of a lung on the right labelled \"No PTX\" has no tube and pneumothorax. On top of these two drawings the text says, \"Biased training datasets\". Right next to the two drawings of lungs is a big arrow pointing to the right. The arrow is labelled \"manifestation shift\". On the right end of the arrow is another drawing of a lung. The left lung in the drawing features a tube but no pneumothorax. The drawing is labelled \"No PTX\", next to the drawing is a smaller arrow pointing to the right. At the right end of the arrow there is a text reading \"False positive\". \n\nIn the third row on the left there are again two drawings of a lung. The first drawing of a lung labelled \"Healthy\" without any changes. The second drawing of a lung on the right labelled \"Lung segmentation\" shows a segmentation mask overlayed with the left and right lung in green. On top of these two drawings the text says, \"Biased training datasets\". Right next to the two drawings of lungs is a big arrow pointing to the right. The arrow is labelled \"population shift\". On the right end of the arrow is another drawing of a lung. The right lung in the drawing features a pacemaker in the upper right lung lobe. The drawing is labelled \"Abnormality\". On the right is another drawing of a lung. The lung shows the pacemaker in the same location. The lungs in the drawing are overlayed with a segmentation mask of the lungs. However, the pacemaker is not included in the segmentation mask. \" class=\"wp-image-1088418\" srcset=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/09\/NEW_RadEdit-2024-Figure-1-BlogHeroFeature-1195x788-15Aug24.jpg 1195w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/09\/NEW_RadEdit-2024-Figure-1-BlogHeroFeature-1195x788-15Aug24-300x198.jpg 300w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/09\/NEW_RadEdit-2024-Figure-1-BlogHeroFeature-1195x788-15Aug24-1024x675.jpg 1024w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/09\/NEW_RadEdit-2024-Figure-1-BlogHeroFeature-1195x788-15Aug24-768x506.jpg 768w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/09\/NEW_RadEdit-2024-Figure-1-BlogHeroFeature-1195x788-15Aug24-240x158.jpg 240w\" sizes=\"auto, (max-width: 1195px) 100vw, 1195px\" \/><figcaption class=\"wp-element-caption\">Figure 2: Stress-testing models by simulating dataset shifts via image editing.<\/figcaption><\/figure>\n\n\n\n<h2 class=\"wp-block-heading\" id=\"stress-testing-multimodal-models\">Stress-testing multimodal models<\/h2>\n\n\n\n<p>We have used RadEdit to stress-test image classification and segmentation models, and we see potential for future applications in complex multimodal tasks like generating radiology reports. RadEdit can help identify limitations in multimodal large language models (MLLMs) like Microsoft&#8217;s <a href=\"https:\/\/www.microsoft.com\/en-us\/research\/publication\/maira-1-a-specialised-large-multimodal-model-for-radiology-report-generation\/\">MAIRA-1<\/a> and <a href=\"https:\/\/www.microsoft.com\/en-us\/research\/publication\/maira-2-grounded-radiology-report-generation\/\">MAIRA-2<\/a>, especially when dealing with rare conditions or unusual combinations of findings not well-represented in the training data. These MLLMs take one or more radiological images and relevant clinical information as input to produce detailed text reports.<\/p>\n\n\n\n<p>RadEdit can generate synthetic image-report pairs for challenging scenarios. For example, manually editing a report to describe a rare combination of findings and then using RadEdit to edit the corresponding image, creates a valuable test case for the MLLM. This approach allows us to stress-test MLLM with diverse synthetic data, identifying weaknesses or biases and ensuring the model is more robust in real-world scenarios. This is a crucial step for using these models safely and effectively in clinical settings.<\/p>\n\n\n\n<h2 class=\"wp-block-heading\" id=\"implications-and-looking-forward\">Implications and looking forward<\/h2>\n\n\n\n<p>RadEdit offers significant advantages for the biomedical research community. It helps identify biases and blind spots before deployment, helping to ensure that biomedical vision models perform reliably in clinical settings. By simulating dataset shifts, RadEdit reduces the need to collect additional evaluation data, saving time and resources.<\/p>\n\n\n\n<p>RadEdit is applicable to a wide range of settings and can be used to stress-test state-of-the-art foundation models like Microsoft&#8217;s <a href=\"https:\/\/www.microsoft.com\/en-us\/research\/publication\/rad-dino-exploring-scalable-medical-image-encoders-beyond-text-supervision\/\">Rad-DINO<\/a> and <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" target=\"_blank\" href=\"https:\/\/microsoft.github.io\/BiomedParse\/\">BiomedParse<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>. By integrating RadEdit into their research workflow, researchers can validate that their biomedical vision models are not only state-of-the-art but also more prepared for the complexities of real-world deployment. In the future, we envision RadEdit being applied to more complex multimodal tasks, such as generating radiology reports.<\/p>\n\n\n\n<p>The code for RadEdit as well as the weights of the diffusion model we used can be found under <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" target=\"_blank\" href=\"https:\/\/huggingface.co\/microsoft\/radedit\">https:\/\/huggingface.co\/microsoft\/radedit<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>.<\/p>\n\n\n\n<h2 class=\"wp-block-heading\" id=\"acknowledgments\">Acknowledgments<\/h2>\n\n\n\n<p>We would like to thank our paper coauthors: <a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/fperezgarcia\/\">Fernando P\u00e9rez-Garc\u00eda<\/a>, <a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/sbondtaylor\/\">Sam Bond-Taylor<\/a>, Pedro P. Sanchez, Boris van Breugel, <a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/harssharma\/\">Harshita Sharma<\/a>, <a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/vsalvatelli\/\">Valentina Salvatelli<\/a>, Maria T. A. Wetscherek, <a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/hamurfet\/\">Hannah Richardson<\/a>, <a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/mlungren\/\">Matthew P. Lungren<\/a>, <a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/adityan\/\">Aditya Nori,<\/a> and Ozan Oktay, as well as all our collaborators across <a href=\"https:\/\/www.microsoft.com\/en-us\/industry\/health\/microsoft-cloud-for-healthcare?msockid=191ee027a78e6e430e59f339a60a6f7a\">Microsoft Cloud for Healthcare<\/a> and <a href=\"https:\/\/www.microsoft.com\/en-us\/research\/lab\/microsoft-health-futures\/\">Microsoft Health Futures<\/a>.<\/p>\n\n\n\n<p><em>RadEdit is intended for research purposes only and not for any commercial or clinical use.<\/em><\/p>\n","protected":false},"excerpt":{"rendered":"<p>RadEdit stress-tests biomedical vision models by simulating dataset shifts through precise image editing. It uses diffusion models to create realistic, synthetic datasets, helping to identify model weaknesses and evaluate robustness.<\/p>\n","protected":false},"author":43518,"featured_media":1088988,"comment_status":"closed","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"msr-url-field":"","msr-podcast-episode":"","msrModifiedDate":"","msrModifiedDateEnabled":false,"ep_exclude_from_search":false,"_classifai_error":"","msr-author-ordering":[{"type":"user_nicename","value":"Max Ilse","user_id":"41095"},{"type":"user_nicename","value":"Daniel Coelho de Castro","user_id":"39811"},{"type":"user_nicename","value":"Javier Alvarez-Valle","user_id":"32137"}],"msr_hide_image_in_river":null,"footnotes":""},"categories":[1],"tags":[],"research-area":[13556,13562,13553],"msr-region":[],"msr-event-type":[],"msr-locale":[268875],"msr-post-option":[269148,243984,269142],"msr-impact-theme":[],"msr-promo-type":[],"msr-podcast-series":[],"class_list":["post-1088394","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-research-blog","msr-research-area-artificial-intelligence","msr-research-area-computer-vision","msr-research-area-medical-health-genomics","msr-locale-en_us","msr-post-option-approved-for-river","msr-post-option-blog-homepage-featured","msr-post-option-include-in-river"],"msr_event_details":{"start":"","end":"","location":""},"podcast_url":"","podcast_episode":"","msr_research_lab":[849856],"msr_impact_theme":[],"related-publications":[],"related-downloads":[],"related-videos":[],"related-academic-programs":[],"related-groups":[780706,1143270],"related-projects":[978063],"related-events":[],"related-researchers":[{"type":"user_nicename","value":"Max Ilse","user_id":41095,"display_name":"Max Ilse","author_link":"<a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/maxilse\/\" aria-label=\"Visit the profile page for Max Ilse\">Max Ilse<\/a>","is_active":false,"last_first":"Ilse, Max","people_section":0,"alias":"maxilse"},{"type":"user_nicename","value":"Daniel Coelho de Castro","user_id":39811,"display_name":"Daniel Coelho de Castro","author_link":"<a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/dacoelh\/\" aria-label=\"Visit the profile page for Daniel Coelho de Castro\">Daniel Coelho de Castro<\/a>","is_active":false,"last_first":"Coelho de Castro, Daniel","people_section":0,"alias":"dacoelh"}],"msr_type":"Post","featured_image_thumbnail":"<img width=\"960\" height=\"540\" src=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/09\/RadEdit-Socials-2024-BlogHeroFeature-1400x788-1-960x540.jpg\" class=\"img-object-cover\" alt=\"On the left is a simple drawing of the lungs. The drawing shows the borders of the left and right lung as well as the trachea and the left and right main stem bronchi. The text under the drawing reads: Original image. To the right of the drawing are the 3 additional inputs of RadEdit. They are arranged vertically. On top there is an example editing prompt. It reads &quot;Consolidation&quot;. Below there is the same drawing of the lung again but this time the left lung is shaded blue. The text reads: Edit mask according to prompt. Lastly, on the bottom, there is the same drawing of the lung but this time the right lung is shaded red. The text reads: &quot;Do not edit mask&quot;. On the right of the 3 additional inputs there is a box saying \u201cRadEdit\u201d. Finally, on the right of the figure, there is the drawing of the lung again. The upper part of the left lung is shaded grey. The text reads: Edited image. Between all the elements, the drawing of the lung, the 3 additional inputs, the box that says \u201cRadEdit\u201d, and the edited image, there are arrows pointing to the next element from left to right.\" decoding=\"async\" loading=\"lazy\" srcset=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/09\/RadEdit-Socials-2024-BlogHeroFeature-1400x788-1-960x540.jpg 960w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/09\/RadEdit-Socials-2024-BlogHeroFeature-1400x788-1-300x169.jpg 300w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/09\/RadEdit-Socials-2024-BlogHeroFeature-1400x788-1-1024x576.jpg 1024w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/09\/RadEdit-Socials-2024-BlogHeroFeature-1400x788-1-768x432.jpg 768w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/09\/RadEdit-Socials-2024-BlogHeroFeature-1400x788-1-1066x600.jpg 1066w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/09\/RadEdit-Socials-2024-BlogHeroFeature-1400x788-1-655x368.jpg 655w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/09\/RadEdit-Socials-2024-BlogHeroFeature-1400x788-1-240x135.jpg 240w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/09\/RadEdit-Socials-2024-BlogHeroFeature-1400x788-1-640x360.jpg 640w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/09\/RadEdit-Socials-2024-BlogHeroFeature-1400x788-1-1280x720.jpg 1280w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/09\/RadEdit-Socials-2024-BlogHeroFeature-1400x788-1.jpg 1400w\" sizes=\"auto, (max-width: 960px) 100vw, 960px\" \/>","byline":"<a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/maxilse\/\" title=\"Go to researcher profile for Max Ilse\" aria-label=\"Go to researcher profile for Max Ilse\" data-bi-type=\"byline author\" data-bi-cN=\"Max Ilse\">Max Ilse<\/a>, <a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/dacoelh\/\" title=\"Go to researcher profile for Daniel Coelho de Castro\" aria-label=\"Go to researcher profile for Daniel Coelho de Castro\" data-bi-type=\"byline author\" data-bi-cN=\"Daniel Coelho de Castro\">Daniel Coelho de Castro<\/a>, and Javier Alvarez-Valle","formattedDate":"September 30, 2024","formattedExcerpt":"RadEdit stress-tests biomedical vision models by simulating dataset shifts through precise image editing. It uses diffusion models to create realistic, synthetic datasets, helping to identify model weaknesses and evaluate robustness.","locale":{"slug":"en_us","name":"English","native":"","english":"English"},"_links":{"self":[{"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/posts\/1088394","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/users\/43518"}],"replies":[{"embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/comments?post=1088394"}],"version-history":[{"count":25,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/posts\/1088394\/revisions"}],"predecessor-version":[{"id":1089267,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/posts\/1088394\/revisions\/1089267"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/media\/1088988"}],"wp:attachment":[{"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/media?parent=1088394"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/categories?post=1088394"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/tags?post=1088394"},{"taxonomy":"msr-research-area","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/research-area?post=1088394"},{"taxonomy":"msr-region","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-region?post=1088394"},{"taxonomy":"msr-event-type","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-event-type?post=1088394"},{"taxonomy":"msr-locale","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-locale?post=1088394"},{"taxonomy":"msr-post-option","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-post-option?post=1088394"},{"taxonomy":"msr-impact-theme","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-impact-theme?post=1088394"},{"taxonomy":"msr-promo-type","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-promo-type?post=1088394"},{"taxonomy":"msr-podcast-series","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-podcast-series?post=1088394"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}