METHOD AND APPARATUS FOR PROVIDING INFORMATION ASSOCIATED WITH IMMUNE PHENOTYPES FOR PATHOLOGY SLIDE IMAGE
The present disclosure relates to a method, performed by at least one computing device, for providing information associated with immune phenotype for pathology slide image. The method may include obtaining information associated with immune phenotype for one or more regions of interest (ROIs) in a pathology slide image, generating, based on the information associated with the immune phenotype for one or more ROIs, an image indicative of the information associated with the immune phenotype, and outputting the image indicative of the information associated with immune phenotype.
Latest LUNIT Inc. Patents:
- METHOD AND SYSTEM FOR ANALYSING PATHOLOGY IMAGE
- METHOD AND APPARATUS FOR PROVIDING EXAMINATION-RELATED GUIDE ON BASIS OF TUMOR CONTENT PREDICTED FROM PATHOLOGY SLIDE IMAGES
- Method and system for parallel processing for medical image
- METHOD AND APPARATUS FOR ANALYZING PATHOLOGICAL SLIDE IMAGE
- Method for managing annotation job, apparatus and system supporting the same
This application is a continuation of International Application No. PCT/KR2021/005772 filed on May 7, 2021 which claims priority to Korean Patent Application No. 10-2020-0055483 filed on May 8, 2020, Korean Patent Application No. 10-2021-0054206 filed on Apr. 27, 2021, and Korean Patent Application No. 10-2021-0059519 filed on May 7, 2021, the entire contents of which are herein incorporated by reference.
TECHNICAL FIELDThe present disclosure relates to a method and a device for providing information associated with immune phenotype for pathology slide image, and more specifically, to a method and a device for generating and outputting an image indicative of information associated with immune phenotype for one or more regions of interest (ROIs) in a pathology slide image.
BACKGROUND ARTRecently, a third-generation anticancer drug for cancer treatment, that is, the immune checkpoint inhibitor that utilize the immune system of the patient's body has gained attention. By the immune anticancer drug, it may refer to any drug that prevents cancer cells from evading the body's immune system or makes immune cells better recognize and attack cancer cells. Since it acts through the body's immune system, there are few side effects from the anticancer drugs, and the survival period of cancer patients treated with the immune checkpoint inhibitor may be longer than when treated with other anticancer drugs. However, these immune checkpoint inhibitor are not always effective for all cancer patients. Therefore, it is important to predict the response rate of the immune checkpoint inhibitor in order to predict the effect of the immune checkpoint inhibitor on the current cancer patient.
Meanwhile, in order to predict the response to the immune checkpoint inhibitor, a user (e.g., a doctor, a patient, and the like) may be provided with immune response information generated through a pathology slide image of a patient's tissue. According to the related art, the user (e.g., doctor, patient, and the like) may be provided with information on the immune response (e.g., immune cell expression information, and the like) to each of the plurality of patches included in the pathology slide image in order to predict the reactivity of the immune checkpoint inhibitor. In this case, it may be difficult for the user to intuitively recognize immune response information for each of numerous patches included in the pathology slide image. In addition, the immune response information may also be generated for a patch among a plurality of patches that is substantially unnecessary for predicting the response to the immune checkpoint inhibitor.
SUMMARY Technical ProblemThe present disclosure provides a method and a device for providing information associated with immune phenotype for pathology slide image to solve the problems described above.
Technical SolutionThe present disclosure may be implemented in various ways, including a method, a device (system), a computer readable storage medium storing instructions, or a computer program.
According to an embodiment of the present disclosure, a method, performed by at least one computing device, for providing information associated with immune phenotype for pathology slide image includes obtaining information associated with immune phenotype for one or more regions of interest (ROIs) in a pathology slide image, generating, based on the information associated with the immune phenotype for one or more ROIs, an image indicative of the information associated with the immune phenotype, and outputting the image indicative of the information associated with immune phenotype.
According to an embodiment, the one or more ROIs are determined based on a detection result for one or more target items for the pathology slide image.
According to an embodiment, the one or more ROIs include at least some regions in the pathology slide image that satisfy a condition associated with the one or more target items.
According to an embodiment, the one or more ROIs are regions being output upon input of the detection result for the one or more target items for the pathology slide image or the pathology slide image to an ROI extraction model, and the ROI extraction model is trained to output a reference ROI upon input of the detection result for one or more target items for a reference pathology slide image or a reference pathology slide image.
According to an embodiment, the obtaining includes obtaining the immune phenotype of the one or more ROIs, the generating includes generating an image including a visual representation corresponding to the immune phenotype of the one or more ROIs, and the immune phenotype includes at least one of immune inflamed, immune excluded, or immune desert.
According to an embodiment, the obtaining includes obtaining one or more immune phenotype scores for the one or more ROIs, the generating includes generating an image including a visual representation corresponding to the one or more immune phenotype scores, and the one or more immune phenotype scores include at least one of a score for immune inflamed, a score for immune excluded, or a score for immune desert.
According to an embodiment, the obtaining includes obtaining a feature associated with one or more immune phenotypes for the one or more ROIs, the generating includes generating an image including a visual representation corresponding to the feature associated with the one or more immune phenotypes, and the feature associated with the one or more immune phenotypes includes at least one of a statistical value or a vector associated with the immune phenotype.
According to an embodiment, the outputting includes outputting one or more ROIs in the pathology slide image together with an image including a visual representation.
According to an embodiment, the outputting may include overlaying the image including the visual representation on one or more ROIs in the pathology slide image.
According to an embodiment, the method further includes obtaining a detection result for one or more target items from the pathology slide image, generating an image indicative of the detection result for one or more target items, and outputting the image indicative of the detection result for one or more target items.
There may be provided a computer program stored in a computer-readable recording medium for executing, on a computer, the method for providing the information associated with the immune phenotype for the pathology slide image described above according to an embodiment of the present disclosure.
A computing device according to an embodiment may include a memory storing one or more instructions, and a processor configured to execute the stored one or more instructions to obtain information associated with immune phenotype for one or more ROIs in the pathology slide image, generate, based on the information associated with the immune phenotype for one or more ROIs, an image indicative of the information associated with the immune phenotype, and output the image indicative of the information associated with the immune phenotype.
Advantageous EffectsAccording to some embodiments of the present disclosure, by providing the user with an image visually representing the information associated with immune phenotype, it is possible to enable the user to intuitively recognize the information associated with immune phenotype for each region. In addition, by providing a visual representation indicative of the information associated with immune phenotype by overlaying it on a corresponding ROI in a pathology slide image, it is possible to enable the user to recognize at a glance which region of the pathology slide image corresponds to the information indicated by the corresponding visual representation.
According to some embodiments of the present disclosure, in order to determine the immune phenotype and/or response or non-response to the immune checkpoint inhibitor, the ROI in the pathology slide image which actually requires analysis may be determined. That is, instead of analyzing the entire pathology slide image, the information processing system and/or the user terminal may perform processing (e.g., determining an immune phenotype and/or calculating an immune phenotype score, and the like) only on the ROIs while excluding the regions where analysis is unnecessary, such that computer resources, processing costs, and the like can be minimized.
According to some embodiments of the present disclosure, more accurate results can be provided by processing (e.g., determining an immune phenotype and/or calculating an immune phenotype score, and the like) only on the significant region when determining the immune phenotype and/or determining response or non-response to the immune checkpoint inhibitor.
The effects of the present disclosure are not limited to the effects described above, and other effects not described will be able to be clearly understood by those of ordinary skill in the art (hereinafter, referred to as “ordinary technician”) from the description of the claims.
Embodiments of the present disclosure will be described with reference to the accompanying drawings described below, in which like reference numerals denote like elements, but are not limited thereto.
Hereinafter, specific details for the practice of the present disclosure will be described in detail with reference to the accompanying drawings. However, in the following description, detailed descriptions of well-known functions or configurations will be omitted when it may make the subject matter of the present disclosure rather unclear.
In the accompanying drawings, the same or corresponding elements are assigned the same reference numerals. In addition, in the following description of the embodiments, duplicate descriptions of the same or corresponding components may be omitted. However, even if descriptions of elements are omitted, it is not intended that such elements are not included in any embodiment.
Advantages and features of the disclosed embodiments and methods of accomplishing the same will be apparent by referring to embodiments described below in connection with the accompanying drawings. However, the present disclosure is not limited to the embodiments disclosed below, and may be implemented in various different forms, and the present embodiments are merely provided to make the present disclosure complete, and to fully disclose the scope of the invention to those skilled in the art to which the present disclosure pertains.
The terms used herein will be briefly described prior to describing the disclosed embodiments in detail. The terms used herein have been selected as general terms which are widely used at present in consideration of the functions of the present disclosure, and this may be altered according to the intent of an operator skilled in the art, conventional practice, or introduction of new technology. In addition, in a specific case, a term is arbitrarily selected by the applicant, and the meaning of the term will be described in detail in a corresponding description of the embodiments. Therefore, the terms used in the present disclosure should be defined based on the meaning of the terms and the overall contents of the present disclosure rather than a simple name of each of the terms.
As used herein, the singular forms “a,” “an,” and “the” are intended to include the plural forms as well, unless the context clearly indicates the singular forms. Further, the plural forms are intended to include the singular forms as well, unless the context clearly indicates the plural forms. As used throughout the description, when one part is referred to as “comprising” (or “including” or “having”) other elements, the part can comprise (or include or have) only those elements or other elements as well as those elements unless specifically described otherwise.
Further, the term “module” or “unit” used herein refers to a software or hardware component, and “module” or “unit” performs certain roles. However, the meaning of the “module” or “unit” is not limited to software or hardware. The “module” or “unit” may be configured to be in an addressable storage medium or configured to reproduce one or more processors. Accordingly, as an example, the “module” or “unit” may include components such as software components, object-oriented software components, class components, and task components, and at least one of processes, functions, attributes, procedures, subroutines, program code segments of program code, drivers, firmware, micro-codes, circuits, data, database, data structures, tables, arrays, and variables. Furthermore, functions provided in the components and the “modules” or “units” may be combined into a smaller number of components and “modules” or “units”, or further divided into additional components and “modules” or “units.”
According to an embodiment, the “module” or “unit” may be implemented as a processor and a memory. The “processor” should be interpreted broadly to encompass a general-purpose processor, a central processing unit (CPU), a microprocessor, a digital signal processor (DSP), a controller, a microcontroller, a state machine, and so forth. Under some circumstances, the “processor” may refer to an application-specific integrated circuit (ASIC), a programmable logic device (PLD), a field-programmable gate array (FPGA), and so on. The “processor” may refer to a combination of processing devices, e.g., a combination of a DSP and a microprocessor, a plurality of microprocessors, one or more microprocessors in conjunction with a DSP core, or any other combination of such configurations. In addition, the “memory” should be interpreted broadly to encompass any electronic component capable of storing electronic information. The “memory” may refer to various types of processor-readable media such as random access memory (RAM), read-only memory (ROM), non-volatile random access memory (NVRAM), programmable read-only memory (PROM), erasable programmable read-only memory (EPROM), electrically erasable PROM (EEPROM), flash memory, magnetic or optical data storage, registers, and so on. The memory is said to be in electronic communication with a processor if the processor can read information from and/or write information to the memory. The memory integrated with a processor is in electronic communication with the processor.
In the present disclosure, the “system” may refer to at least one of a server device and a cloud device, but not limited thereto. For example, the system may include one or more server devices. As another example, the system may include one or more cloud devices. As another example, the system may be configured together with both a server device and a cloud device and operated.
In the present disclosure, “target data” may refer to any data or data item that can be used for training of a machine learning model, and may include, for example, data indicative of an image, data indicative of voice or voice characteristics, and the like, but is not limited thereto. In the present disclosure, the whole pathology slide image and/or at least one patch (or region) included in the pathology slide image are explained as the target data, but is not limited thereto, and any data that can be used for training a machine learning model may correspond to the target data. In addition, the target data may be tagged with label information through an annotation task.
In the present disclosure, the “pathology slide image” refers to an image obtained by capturing a pathological slide fixed and stained through a series of chemical treatments in order to observe a tissue removed from a human body with a microscope. For example, the pathology slide image may refer to a digital image captured with a microscope, and may include information on cells, tissues, and/or structures in the human body. In addition, the pathology slide image may include one or more patches, and the one or more patches may be tagged with label information (e.g., information on immune phenotype) through the annotation work. For example, the “pathology slide image” may include H&E-stained tissue slides and/or IHC-stained tissue slides, but is not limited thereto, and tissue slides applied with various staining methods (e.g., chromogenic in situ hybridization (CISH), Fluorescent in situ hybridization (FISH), Multiplex IHC, and the like), or unstained tissue slides may also be included. As another example, the “pathology slide image” may be a patient's tissue slide generated to predict a response to immune checkpoint inhibitor, and it may include a tissue slide of a patient before treatment with immune checkpoint inhibitor and/or a tissue slide of a patient after treatment with immune checkpoint inhibitor.
In the present disclosure, a “biomarker” may be defined as a marker that can objectively measure a normal or pathology state, a degree of response to a drug, and the like. For example, the biomarker may include immune checkpoint inhibitor and PD-L1 as a biomarker, but is not limited thereto, and may include Tumor Mutation Burden (TMB) value, Microsatellite Instability (MSI) value, Homologous Recombination Deficiency (HRD) value, CD3, CD8, CD68, FOXP3, CD20, CD4, CD45, CD163, and other various biomarkers related to immune cells.
In the present disclosure, the “patch” may refer to a small region within the pathology slide image. For example, the patch may include a region corresponding to a semantic object extracted by performing segmentation on the pathology slide image. As another example, the patch may refer to a combination of pixels associated with the label information generated by analyzing the pathology slide image.
In the present disclosure, the “regions of interest (ROIs)” may refer to at least some regions to be analyzed in the pathology slide image. For example, the ROIs may refer to at least some regions in the pathology slide image that include a target item. As another example, the ROIs may refer to at least some of a plurality of patches generated by segmenting the pathology slide image.
In the present disclosure, a “machine learning model” and/or an “artificial neural network model” may include any model that is used for inferring an answer to a given input. According to an embodiment, the machine learning model may include an artificial neural network model including an input layer (layer), a plurality of hidden layers, and output layers. In an example, each layer may include a plurality of nodes. For example, the machine learning model may be trained to infer label information for pathology slide images and/or at least one patch included in the pathology slides. In this case, the label information generated through the annotation task may be used to train the machine learning model. In addition, the machine learning model may include weights associated with a plurality of nodes included in the machine learning model. In an example, the weight may include an any parameter associated with the machine learning model.
In the present disclosure, “training” may refer to any process of changing a weight associated with the machine learning model using at least one patch and the label information. According to an embodiment, the training may refer to a process of changing or updating weights associated with the machine learning model through one or more of forward propagation and backward propagation of the machine learning model using at least one patch and the label information.
In the present disclosure, the “label information” is correct answer information of the data sample information, which is obtained as a result of the annotation task. The label or label information may be used interchangeably with terms such as annotation, tag, and so on as used in the art. In the present disclosure, the “annotation” may refer to an annotation work and/or annotation information (e.g., label information, and the like) determined by performing the annotation work. In the present disclosure, the “annotation information” may refer to information for the annotation work and/or information generated by the annotation work (e.g., label information).
In the present disclosure, the “target item” may refer to data/information, an image region, an object, and the like to be detected in the pathology slide image. According to an embodiment, the target item may include a target to be detected from the pathology slide image for diagnosis, treatment, prevention, or the like of a disease (e.g., cancer). For example, the “target item” may include a target item in units of cells and a target item in units of areas.
In the present disclosure, “each of a plurality of A” and/or “respective ones of a plurality of A” may refer to each of all components included in the plurality of A, or may refer to each of some of the components included in a plurality of A. For example, each of the plurality of ROIs may refer to each of all ROIs included in the plurality of ROIs or may refer to each of some ROIs included in the plurality of ROIs.
In the present disclosure, “instructions” may refer to one or more instructions grouped based on functions, which are the components of a computer program and executed by the processor.
In the present disclosure, a “user” may refer to a person who uses a user terminal. For example, the user may include an annotator who performs an annotation work. As another example, the user may include a doctor, a patient, and the like who is provided with the information associated with immune phenotype and/or a prediction result of a response to immune checkpoint inhibitor (e.g., a prediction result as to whether or not the patient responds to immune checkpoint inhibitor). In addition, the user may refer to the user terminal, or conversely, the user terminal may refer to the user. That is, the user and the user terminal may be interchangeably used herein.
The information processing system 100 and the user terminal 110 are any computing devices used to generate and provide the information associated with immune phenotype for pathology slide image. In an example, the computing device may refer to any type of device equipped with a computing function, and may be a notebook, a desktop, a laptop, a server, a cloud system, and the like, for example, but is not limited thereto.
The information processing system 100 may receive a pathology slide image. For example, the information processing system 100 may receive the pathology slide image from the storage system 120 and/or the user terminal 110. The information processing system 100 may generate information associated with immune phenotype of the pathology slide image and provide it to the user 130 through the user terminal 110. In an embodiment, the information processing system 100 may determine one or more ROIs in the pathology slide image and generate information associated with immune phenotype for the one or more ROIs. In this example, the information associated with immune phenotype may include at least one of an immune phenotype of one or more ROIs, an immune phenotype score of one or more ROIs, and a feature associated with the immune phenotype of one or more ROIs.
In an embodiment, the information processing system 100 may determine one or more ROIs based on a detection result for one or more target items for the pathology slide image. For example, the information processing system 100 may determine one or more ROIs including at least some regions in the pathology slide image that satisfy a condition associated with the one or more target items. Additionally or alternatively, upon input of the detection result for one or more target items for a pathology slide image (e.g., a pathology slide image including a detection result for target item) to a region-of-interest (ROI) extraction model, the information processing system 100 may determine regions being output as one or more ROIs. In this example, the ROI extraction model may correspond to a model trained to output a reference ROI when the detection result for one or more target items for reference pathology slide image (e.g., a reference pathology slide image including the detection result for target item) is input.
The user terminal 110 may obtain from the information processing system 100 the information associated with immune phenotype for one or more ROIs in the pathology slide image. For example, the information associated with immune phenotype for one or more ROIs may include an immune phenotype (e.g., at least one of immune inflamed, immune excluded, or immune desert) for one or more ROIs. Additionally or alternatively, the information associated with immune phenotype for one or more ROIs may include one or more immune phenotype scores for the one or more ROIs (e.g., the immune phenotype score is at least one of a score for immune inflamed, a score for immune excluded or a score for immune desert). Additionally or alternatively, the information associated with immune phenotype for one or more ROIs may include a feature associated with one or more immune phenotypes for one or more ROIs (e.g., at least one of statistical value or vector associated with the immune phenotype).
Then, the user terminal 110 may generate an image indicative of the information associated with immune phenotype based on the information associated with immune phenotype for one or more ROIs. In an embodiment, the user terminal 110 may generate an image including a visual representation corresponding to the immune phenotype of one or more ROIs. In another embodiment, the user terminal 110 may generate an image including a visual representation corresponding to the one or more immune phenotype scores. For example, the visual representation may include color (e.g., color, brightness, saturation, and the like), text, image, mark, figure, and the like.
The user terminal 110 may output an image indicative of the information associated with the generated immune phenotype. In an embodiment, the user terminal 110 may output one or more ROIs in the pathology slide image together with an image including a visual representation. That is, the one or more ROIs in the pathology slide image and the image including the visual representation may be displayed together on the display device associated with the user terminal 110. In another embodiment, the user terminal 110 may overlay the image including the visual representation on one or more ROIs in the pathology slide image. That is, one or more ROIs in the pathology slide image overlaid with the image including the visual representation may be displayed on the display device associated with the user terminal 110. Accordingly, the user 130 (e.g., doctor, patient, and the like) may be provided with an image indicative of information associated with immune phenotype through the user terminal 110.
The storage system 120 is a device or a cloud system that stores and manages pathology slide images associated with a target patient and various data associated with a machine learning model to provide information associated with immune phenotype for the pathology slide image. For efficient data management, the storage system 120 may store and manage various types of data using a database. In this example, the various data may include any data associated with the machine learning model, and include, for example, a file of the target data, meta information of the target data, label information for the target data that is the result of the annotation work, data related to the annotation work, a machine learning model (e.g., an artificial neural network model), and the like, but are not limited thereto. While
According to some embodiments of the present disclosure, by providing the user 130 with an image visually representing the information associated with immune phenotype, it is possible to enable the user 130 to intuitively recognize the information associated with immune phenotype for each region. In addition, according to some embodiments of the present disclosure, in order to determine the immune phenotype and/or response or non-response to the immune checkpoint inhibitor, the ROI in the pathology slide image which actually requires analysis may be determined. That is, instead of analyzing the entire pathology slide image, the information processing system 100 and/or the user terminal 110 may perform processing (e.g., determining an immune phenotype and/or calculating an immune phenotype score, and the like) only on the ROIs while excluding the regions where analysis is unnecessary, such that computer resources, processing costs, and the like can be minimized. In addition, more accurate prediction results may be provided by processing (e.g., determining an immune phenotype and/or calculating an immune phenotype score, and the like) only on the significant region when determining the immune phenotype and/or determining response or non-response to the immune checkpoint inhibitor.
The target item detection unit 210 may receive the pathology slide image (e.g., H&E-stained pathology slide image, IHC-stained pathology slide image, and the like), and detect one or more target items in the received pathology slide image. In an embodiment, the target item detection unit 210 may use the artificial neural network model for target item detection to detect one or more target items in the pathology slide image. In this example, the artificial neural network model for target item detection may correspond to a model trained to detect one or more reference target items from the reference pathology slide image. For example, the target item detection unit 210 may detect a target item in units of cells and/or a target item in units of regions in the pathology slide image. That is, the target item detection unit 210 may detect tumor cell, lymphocyte, macrophages, dendritic cell, fibroblast, endothelial cell, blood vessel, cancer stroma, cancer epithelium, cancer area, normal area (e.g., normal lymph node architecture region), and the like, as the target item in the pathology slide image.
The ROI determination unit 220 may determine one or more ROIs in the pathology slide image. In this example, the ROI may include a region in which one or more target items are detected in the pathology slide image. For example, the ROI determination unit 220 may determine, as the region of interest, a patch including one or more target items from among a plurality of patches forming the pathology slide image. In an embodiment, the ROI determination unit 220 may determine one or more ROIs based on a detection result for one or more target items for a pathology slide image. For example, the ROI determination unit 220 may determine one or more ROIs including at least some regions in the pathology slide image that satisfy a condition associated with the one or more target items. Additionally or alternatively, upon input of the detection result for one or more target items for a pathology slide image and/or the pathology slide image to an ROI extraction model, the ROI determination unit 220 may determine the regions being output as one or more ROIs.
The immune phenotype determination unit 230 may generate the information associated with immune phenotype for one or more ROIs in the pathology slide image. In an embodiment, the immune phenotype determination unit 230 may determine the immune phenotype of one or more ROIs based on the detection result for one or more target items. For example, the immune phenotype determination unit 230 may determine whether the immune phenotype of one or more ROIs is immune inflamed, immune excluded, or immune desert, based on the detection result for one or more target items. In another embodiment, the immune phenotype determination unit 230 may calculate the immune phenotype score of one or more ROIs based on the detection result for one or more target items. For example, the immune phenotype determination unit 230 may calculate a score for immune inflamed of one or more ROIs, a score for immune excluded, and/or a score for immune desert, based on the detection result for one or more target items. To this end, the immune phenotype determination unit 230 may calculate a score indicating a probability that the immune phenotype of one or more ROIs is immune inflamed, a score indicating a probability that it is immune excluded, and/or a score indicating a probability that it is immune desert.
In another embodiment, the immune phenotype determination unit 230 may generate a feature associated with one or more immune phenotypes for one or more ROIs. In this example, the feature associated with one or more immune phenotypes may include at least one of a statistical value or a vector associated with the immune phenotype. For example, the feature associated with one or more immune phenotypes may include score values related to the immune phenotype output from an artificial neural network model or a machine learning model. That is, it may include score values output in the process of determining the immune phenotype for one or more ROIs. As another example, the feature associated with one or more immune phenotypes may include a density value, number, or various statistics of immune cells corresponding to a threshold (or cut-off) for an immune phenotype or a vector value or the like expressing the distribution of immune cells.
As another example, the feature associated with one or more immune phenotypes may include a scalar value or a vector value including a relative relationship (e.g., a histogram vector or a graph expression vector considering the direction and distance) or relative statistics (e.g., the ratio of the number of immune cells to the number of specific cells, and the like) between an immune cell or cancer cell and a specific cell (e.g., cancer cells, immune cells, fibroblasts, lymphocytes, plasma cells, macrophage, endothelial cells, and the like). As another example, the feature associated with one or more immune phenotypes may include scalar values or vector values including statistics (e.g., the ratio of the number of immune cells to the cancer stromal region, and the like) or distributions (e.g., histogram vector or graph representation vector, and the like) of immune cells or cancer cells in a specific region (e.g., cancer area, cancer stromal region, tertiary lymphoid structure, normal region, necrosis, fat, blood vessel, high endothelial venule, lymphatic vessel, nerve, and the like).
As another example, the feature associated with one or more immune phenotypes may include scalar values or vector values including relative relationship (e.g., histogram vector or graph expression vector considering the direction and distance) or relative statistics (e.g., the ratio of the number of immune cells to the number of specific cells, and the like) between positive/negative cells according to the expression amount of the biomarker and the specific cells (e.g., cancer cells, immune cells, fibroblasts, lymphocytes, plasma cells, macrophage, endothelial cells, and the like). As another example, the feature associated with one or more immune phenotypes may include scalar values or vector values including statistics (e.g., the ratio of the number of immune cells to the cancer stromal region, and the like) or distributions (e.g., histogram vector or graph representation vector, and the like) of positive/negative cells according to the expression amount of the biomarker in a specific region (e.g., cancer area, cancer stromal region, tertiary lymphoid structure, normal region, necrosis, fat, blood vessel, high endothelial venule, lymphatic vessel, nerve, and the like).
In
The image generation unit 310 may obtain the information associated with immune phenotype for one or more ROIs in the pathology slide image. For example, the image generation unit 310 may receive the information associated with immune phenotype for one or more ROIs generated by the information processing system. Additionally or alternatively, as the user terminal 110 generates the information associated with immune phenotype for one or more ROIs, the image generation unit 310 may obtain the information associated with immune phenotype for one or more ROIs. Additionally or alternatively, the image generation unit 310 may receive the information associated with immune phenotype for one or more ROIs stored in an internal and/or external device of the user terminal 110.
The image generation unit 310 may generate an image indicative of the information associated with immune phenotype based on the information associated with immune phenotype for one or more ROIs. In an embodiment, when obtaining the immune phenotype of one or more ROIs, the image generation unit 310 may generate an image including a visual representation corresponding to the immune phenotype of one or more ROIs. In another embodiment, when receiving one or more immune phenotype scores for one or more ROIs, the image generation unit 310 may generate an image including a visual representation corresponding to one or more immune phenotype scores.
In another embodiment, the image generation unit 310 may generate an image including a visual representation corresponding to a feature (e.g., a value of a feature) associated with one or more immune phenotypes. For example, it is possible to generate an image including a heatmap according to the expression rate of a biomarker, a classification map classified based on a specific threshold (e.g., cut-off), and the like. In this example, the classification map may include a map visualizing a result of classifying a tumor proportion score (TPS) and/or a combined proportion score (CPS), which are information related to the expression of PD-L1, based on a specific threshold.
The image output unit 320 may output through the display device an image indicative of the information associated with immune phenotype. In an embodiment, the image output unit 320 may output through the display device one or more ROIs in the pathology slide image and an image including a visual representation. Alternatively, the image output unit 320 may overlay an image including a visual representation on one or more ROIs in the pathology slide image.
In addition,
In an embodiment, the one or more ROIs may be determined based on a detection result for one or more target items for a pathology slide image. For example, the one or more ROIs may include at least some regions in the pathology slide image that satisfy a condition associated with the one or more target items. As another example, the one or more ROIs may be the regions being output upon input of the detection result for one or more target items for the pathology slide image and/or the pathology slide image to the ROI extraction model. In this case, the ROI extraction model may correspond to a model that is trained to output a reference ROI upon input of the detection result for one or more target items for reference pathology slide image (e.g., a reference pathology slide image including the detection result for target item), and/or a reference pathology slide image.
The processor may generate an image indicative of information associated with immune phenotype based on the information associated with immune phenotype for one or more ROIs (S420). In an embodiment, the processor may generate an image including a visual representation corresponding to immune phenotype of one or more ROIs. In another embodiment, the processor may generate an image including a visual representation corresponding to one or more immune phenotype scores. For example, the processor may generate an image including a visual representation corresponding to a score for immune inflamed. As another example, the processor may generate an image including a visual representation corresponding to a score for immune excluded. As another example, the processor may generate an image including a visual representation corresponding to a score for immune desert. In still another embodiment, the processor may generate an image including a visual representation corresponding to a feature associated with one or more immune phenotypes. For example, the processor may generate an image including a visual representation corresponding to a value of a feature associated with one or more immune phenotypes.
Then, the processor may output an image indicative of the information associated with immune phenotype (S430). In an embodiment, the processor may output one or more ROIs in the pathology slide image together with an image including a visual representation. In another embodiment, the processor may overlay an image including a visual representation on one or more ROIs in the pathology slide image.
In an embodiment, the processor may obtain a detection result for one or more target items from the pathology slide image, and generate an image indicative of the detection result for one or more target items. Then, the processor may output the image indicative of the detection result for one or more target items.
The ROI determination unit 220 of the information processing system may determine one or more ROIs in the pathology slide image. In this example, the ROIs may correspond to the regions having various shapes, such as circle, square, rectangle, polygon, contour, and the like. In an embodiment, the ROI determination unit 220 may determine one or more ROIs based on the detection result for one or more target items for a pathology slide image. For example, among a plurality of patches (e.g., 1 mm2 sized patches) generated by dividing the pathology slide image into N grids (where N is any natural number), the ROI determination unit 220 may determine a patch, from which one or more target items (e.g., items and/or immune cells associated with a cancer) are detected, as the ROI.
In an embodiment, the ROI determination unit 220 may determine one or more ROIs such that the ROIs include at least some regions in the pathology slide image that satisfy a condition associated with one or more target items. For example, the ROI determination unit 220 may determine a region, which has the number and/or area of target items (e.g., tumor cells, immune cells, cancer area, cancer stroma, and the like) equal to or greater than a reference value in the pathology slide image, as the ROI. Additionally or alternatively, the ROI determination unit 220 may determine a region, which has a numerical value such as a ratio, a density, and the like of the target item equal to or greater than a reference value in the pathology slide image, as the ROI. In this example, the target item may correspond to a cell and/or region detected to determine the immune phenotype. In addition, the reference value may refer to a numerical value set for the target item to define a statistically significant immune phenotype and/or a clinically significant immune phenotype.
In this case, the ROI determination unit 220 may determine a region, which has any area in the pathology slide image, as the ROI. For example, the area of the ROI may be dynamically determined to satisfy the condition associated with one or more target items described above. That is, the area of the ROI is not fixedly predetermined, and may be dynamically determined as the ROI determination unit 200 determines the region, which has the number, area, ratio, and/or density and the like of the target items equal to or greater than the reference value, as the ROI. Alternatively, the ROI determination unit 220 may determine one or more ROIs such that the area of the ROIs has a statically predetermined value.
In another embodiment, upon input of the detection result for one or more target items for the pathology slide image and/or the pathology slide image to an ROI extraction model, the ROI determination unit 220 may determine the regions being output as one or more ROIs. In this example, the ROI extraction model may correspond to a machine learning model (e.g., Neural Network, CNN, SVM, and the like) trained to output a reference ROI upon input of a detection result for one or more target items for reference pathology slide image and/or a reference pathology slide image. Even in this case, the area of the ROI may be determined dynamically and/or statically.
In an embodiment, when a plurality of ROIs are determined for the pathology slide image, the ROI determination unit 220 may determine the plurality of ROIs such that at least some of the plurality of ROIs overlap with each other. For example, when the ROI determination unit 220 determines a first ROI and a second ROI for the pathology slide image, at least some regions of the first ROI and at least some regions of the second ROI may be regions overlapping with each other. In another embodiment, when a plurality of ROIs for one pathology slide image is determined, the ROI determination unit 220 may determine the plurality of ROIs such that at least some of the plurality of ROIs do not overlap with each other.
As illustrated, the ROI determination unit 220 may receive the pathology slide image 510 (e.g., pathology slide image including a detection result for target item), and determine one or more ROIs in the pathology slide images 522_1, 522_2, 522_3, 522_4, 524, and 526. For example, the ROI determination unit 220 may determine the regions 522_1, 522_2, 522_3 and 522_4, which have a specific area (e.g., 1 mm2 as a predetermined area) that satisfies the condition associated with the target item, as the ROIs. As another example, the ROI determination unit 220 may determine a region 524, which satisfies the condition associated with the target item, as the ROI, and the area of the ROI may be dynamically determined. As still another example, the ROI determination unit 220 may determine an elliptical region 526, which satisfies the condition associated with the target item, as the ROI.
In another embodiment, the immune phenotype determination unit 230 may calculate an immune phenotype score for one or more ROIs based on the detection result for one or more target items in one or more ROIs. For example, the immune phenotype determination unit 230 may calculate at least one of a score for immune inflamed, a score for immune excluded, or a score for immune desert in the corresponding ROI based on the detection result for one or more target items in one or more ROIs. In this case, the immune phenotype determination unit 230 may determine an immune phenotype of the corresponding ROI based on at least one of a score for immune inflamed, a score for immune excluded, and a score for immune desert for one or more ROIs. For example, when the score for immune inflamed of a specific ROI is equal to or greater than a threshold value, the immune phenotype determination unit 230 may determine the immune phenotype of the corresponding ROI as immune inflamed.
In an embodiment, the immune phenotype determination unit 230 may calculate at least one of the number, distribution, or density of target items in one or more ROIs, and may determine an immune phenotype and/or an immune phenotype score of one or more ROIs based on at least one of the calculated number, distribution, or density of the immune cells. For example, the immune phenotype determination unit 230 may calculate, within one or more ROIs, a density of lymphocytes in the cancer area and a density of lymphocytes in the cancer stroma area, and determine the immune phenotype of the one or more ROIs based on at least one of the density of immune cells in the cancer area or the density of immune cells in the cancer stroma. Additionally or alternatively, the immune phenotype determination unit 230 may determine the immune phenotype of one or more ROIs as one of immune inflamed, immune excluded, or immune desert, by referring to the number of immune cells included in the specific region in the cancer area.
For example, the immune phenotype determination unit 230 may determine the immune phenotype of a first ROI 612, which has a density of immune cells in the cancer area equal to or greater than a first threshold density, as immune inflamed. In addition, the immune phenotype determination unit 230 may determine the immune phenotype of a second ROI 614, which has a density of immune cells in the cancer area less than the first threshold density and a density of immune cells in the cancer stroma equal to or greater than the second threshold density, as immune excluded. In addition, the immune phenotype determination unit 230 may determine the immune phenotype of a third ROI 616 having a density of immune cells in the cancer area less than the first threshold density and having a density of immune cells in the cancer stroma less than the second threshold density is immune desert. In this example, the first threshold density may be determined based on a distribution of the density of the immune cells in the cancer area in each of a plurality of ROIs in the plurality of pathology slide images. Likewise, the second threshold density may be determined based on a distribution of the density of the immune cells in the cancer stroma in each of the plurality of ROIs in the plurality of pathology slide images.
In another embodiment, the immune phenotype determination unit 230 may input the feature for each of the one or more ROIs to the artificial neural network immune phenotype classification model to determine the immune phenotype and/or immune phenotype score of each of the one or more ROIs. In this example, the artificial neural network immune phenotype classification model may correspond to a classifier that is trained to determine the immune phenotype of the reference ROI as one of immune inflamed, immune excluded, or immune desert upon input of the feature for the reference ROI. In addition, in this example, the feature for each of one or more ROIs may include a statistical feature for one or more target items in each of one or more ROIs (e.g., density, number, and the like, of specific target items in the ROI), a geometric feature for one or more target items (e.g., a feature including relative position information between specific target items, and the like), and/or an image feature, and the like corresponding to each of the one or more ROIs (e.g., a feature extracted from a plurality of pixels included in the ROIs, an image vector corresponding to the ROIs, and the like). Additionally or alternatively, the feature for each of the one or more ROIs may include a feature obtained by concatenating two or more features from among the statistical feature for one or more target items in each of the one or more ROIs, the geometric feature for the one or more target items, or the image feature corresponding to each of the one or more ROIs.
As illustrated, the immune phenotype determination unit 230 may receive one or more ROIs 612, 614, and 616, and generate the immune phenotype determination result 620. In this example, one or more ROIs 612, 614, and 616 may include the detection result for target item for the corresponding ROI. Accordingly, the immune phenotype determination unit 230 may generate the immune phenotype determination result 620 for the ROI that includes the detection result for target item. For example, the immune phenotype determination unit 230 may generate the immune phenotype determination result 620 including the immune phenotype of the corresponding ROI and/or the immune phenotype score of the corresponding ROI. The generated immune phenotype determination result 620 and/or one or more ROIs 612, 614, and 616 may be provided to the user terminal.
To this end, the user terminal may generate an image including a visual representation corresponding to one or more immune phenotype scores for one or more ROIs. For example, in a region corresponding to one or more ROIs, the user terminal may generate an image including a visual representation corresponding to one or more immune phenotype scores for the ROI. In this example, the one or more immune phenotype scores may include at least one of a score for immune inflamed, a score for immune excluded, or a score for immune desert. In addition, the visual representation may include color (e.g., color, brightness, saturation, and the like), text, image, mark, figure, and the like.
For example, with respect to the score for immune inflamed (that is, the immune inflamed score), the user terminal may generate an image including a color with higher saturation for a region corresponding to ROI having a higher immune inflamed score, and generate an image including a color with lower saturation for a region corresponding to ROI having a lower immune inflamed score. As another example, with respect to the score for the immune inflamed, the user terminal may generate an image including a first visual representation in a region corresponding to the ROI having the immune inflamed score corresponding to a first score section, including a second visual representation in a region corresponding to the ROI having the immune inflamed score corresponding to a second score section, and including a third visual representation in a region corresponding to the ROI having the immune inflamed score corresponding to a third score section. In this example, the first visual representation, the second visual representation, and the third visual representation may be different from each other.
In an embodiment, the user terminal may output one or more ROIs in the pathology slide image together with an image including a visual representation. For example, the user terminal may simultaneously display the image (e.g., the image including a visual representation) generated as described above and at least some regions of the pathology slide image (e.g., the region corresponding to the generated image) on a display device. In another embodiment, the user terminal may overlay the image including the visual representation on one or more ROIs in the pathology slide image. For example, the user terminal may overlap the image (e.g., the image including a visual representation) generated as described above on at least some regions of the pathology slide image (e.g., the region corresponding to the generated image), and display on the display device.
For example, as illustrated, the user terminal may generate an image 710 in which the region corresponding to the ROI having the immune inflamed score falling in the first score section is displayed in white, the region corresponding to the ROI having the immune inflamed score falling in the second score section is displayed in light gray, the region corresponding to the ROI having the immune inflamed score falling in the third score section is displayed in dark gray, and the regions other than the ROIs are displayed in black. Then, the user terminal may arrange, in parallel, the image 710 including a visual representation and a region 720 in the pathology slide image which corresponds to the generated image 710, and display them together on the user interface.
Additionally or alternatively, the user terminal may further display on the user interface the information associated with immune phenotype, such as text, numerical values, graphs, and the like for a “total tissue region” (e.g., a tissue region in a pathology slide image), a “cancer region” (e.g., a cancer area in a pathology slide image, an “analyzable region” (e.g., an ROI in pathology slide image), and an “immune phenotype proportion” (e.g., immune phenotype proportion). Additionally or alternatively, a region in the pathology slide images, which is displayed on the display device, may include the detection result for one or more target items for the corresponding region. That is, when outputting an image of at least some regions of the pathology slide image, the user terminal may output an image of at least some regions displaying the detection results for one or more target items through the display device.
To this end, the user terminal may generate an image including a visual representation corresponding to the immune phenotype of one or more ROIs. For example, the user terminal may generate an image including a visual representation corresponding to the immune phenotype of the corresponding ROI in a region corresponding to one or more ROIs. That is, an image may be generated, which includes the first visual representation in a region corresponding to the ROI having the immune phenotype of immune inflamed, includes the second visual representation in a region corresponding to the ROI having the immune phenotype of immune excluded, and includes the third visual representation in a region corresponding to the ROI having the immune phenotype of immune desert. In this case, the visual representation may include color (e.g., color, brightness, saturation, and the like), text, image, mark, figure, and the like, to distinguish each immune phenotype. For example, the first visual representation indicating immune inflamed may be red color, the second visual representation indicating immune excluded may be green color, and the third visual representation indicating immune desert may be blue color. Additionally or alternatively, the first visual representation indicating immune inflamed may be a circle mark, the second visual representation indicating immune excluded may be a triangle mark, and the third visual representation indicating immune desert may be an X mark.
In an embodiment, the user terminal may output one or more ROIs in the pathology slide image together with an image including the visual representation. For example, the user terminal may display the image (e.g., the image including a visual representation) generated as described above together with at least some regions of the pathology slide image (e.g., the region corresponding to the generated image) on the display device. In another embodiment, the user terminal may overlay the image including the visual representation on one or more ROIs in the pathology slide image. For example, the user terminal may overlap the image (e.g., the image including a visual representation) generated as described above transparently, translucently, or opaquely on at least some regions of the pathology slide image (e.g., the region corresponding to the generated image), and display on the display device.
For example, as illustrated, the user terminal may generate an image that includes a diagonal marker in a region corresponding to the ROI having the immune phenotype of immune inflamed, includes a vertical line marker in a region corresponding to the ROI having the immune phenotype of immune excluded, and includes a horizontal line marker in a region corresponding to the ROI having the immune phenotype of immune desert. Then, the user terminal may display on the user interface an image 810 including the image including a visual representation overlaid on a corresponding region of the pathology slide image. Additionally, the user terminal may also display on the user interface an image 820 of at least some regions of the pathology slide image (e.g., region corresponding to an image including a visual representation and/or the entire pathology slide image) separately (e.g., in a form of a minimap).
Additionally or alternatively, the user terminal may display on the user interface the information associated with immune phenotype, such as text, numerical values, graphs, and the like for “analysis summary”, “biomarker findings”, “score” (e.g., immune inflamed score, and the like), “cutoff” (e.g., reference value used in determining the immune phenotype, and the like), “total tissue region”, “cancer region”, “analyzable region”, “immune phenotype proportion”, “tumor infiltrating lymphocyte density” (e.g., immune cell density in the cancer area, immune cell density in the cancer stromal region, and the like). Additionally or alternatively, the region in the pathology slide images, which is displayed on the display device, may include the detection result for one or more target items for the corresponding region. That is, when outputting an image of at least some regions of the pathology slide image, the user terminal may output an image of at least some regions displaying the detection results for one or more target items through the display device.
According to an embodiment, the artificial neural network model 900 may represent a machine learning model that obtains a problem solving ability by repeatedly adjusting the weights of synapses by the nodes that are artificial neurons forming the network through synaptic combinations as in the biological neural networks, thus training to reduce errors between a target output corresponding to a specific input and a deduced output. For example, the artificial neural network model 900 may include any probability model, neural network model, and the like, that is used in artificial intelligence learning methods such as machine learning and deep learning.
According to an embodiment, the artificial neural network model 900 may include an artificial neural network model configured to detect one or more target items from a pathology slide image being inputted. Additionally or alternatively, the artificial neural network model 900 may include an artificial neural network model configured to determine one or more ROIs from an input pathology slide image.
The artificial neural network model 900 is implemented as a multilayer perceptron (MLP) formed of multiple nodes and connections between them. The artificial neural network model 900 according to an embodiment may be implemented using one of various artificial neural network model structures including the MLP. As shown in
The method of training the artificial neural network model 900 includes the supervised learning that trains to optimize for solving a problem with inputs of teacher signals (correct answer), and the unsupervised learning that does not require a teacher signal. In an embodiment, the information processing system may train the artificial neural network model 900 by supervised learning and/or unsupervised learning to detect one or more target items from a pathology slide image. For example, the information processing system may train the artificial neural network model 900 by supervised learning to detect one or more target items from the pathology slide image by using a reference pathology slide image and label information for one or more reference target items.
In another embodiment, the information processing system may train the artificial neural network model 900 by supervised learning and/or unsupervised learning to determine one or more ROIs from the pathology slide image. For example, the information processing system may train the artificial neural network model 900 by supervised learning to determine one or more ROIs from the pathology slide image using the reference pathology slide image and/or the detection result for one or more target items for the reference pathology slide image (e.g., reference pathology slide image including the detection result for target item), and the label information on the reference ROI. In this example, the ROI and/or the reference ROI may include at least some regions in the pathology slide image and/or the reference pathology slide image that satisfy the condition associated with one or more target items.
The artificial neural network model 900 trained as described above may be stored in a memory (not illustrated) of the information processing system, and in response to an input for the pathology slide image received from the communication module and/or the memory, may detect one or more target items in the pathology slide image. Additionally or alternatively, the artificial neural network model 900 may determine one or more ROIs from the pathology slide image, in response to the detection result for the one or more target items for the pathology slide image and/or an input to the pathology slide image.
According to an embodiment, the input variable of the artificial neural network model for detecting the target item may be one or more pathology slide images (e.g., H&E-stained pathology slide images, IHC-stained pathology slide images). For example, the input variable input to the input layer 920 of the artificial neural network model 900 may be the image vector 910 which may be one or more pathology slide images configured as one vector data element. The output variable output from the output layer 940 of the artificial neural network model 900 in response to the input of the image may be a vector 950 representing or characterizing one or more target items detected in the pathology slide image. That is, the output layer 940 of the artificial neural network model 900 may be configured to output a vector representing or characterizing one or more target items detected from the pathology slide image. In the present disclosure, the output variable of the artificial neural network model 900 is not limited to the types described above, and may include any information/data indicative of one or more target items detected from the pathology slide image. In addition, the output layer 940 of the artificial neural network model 900 may be configured to output a vector indicative of the reliability and/or accuracy of the output detection result for target item and the like.
In another embodiment, the input variable of the machine learning model for determining the ROI, that is, the input variable of the artificial neural network model 900 may be the detection result for one or more target items for the pathology slide image (e.g., detection data for the target item in the pathology slide image) and/or the pathology slide image. For example, the input variable input to the input layer 920 of the artificial neural network model 900 may be the image vector 910 which may be the detection result for the one or more target items for the pathology slide image and/or the pathology slide image configured as one vector data element. An output variable output from the output layer 940 of the artificial neural network model 900 in response to input for the detection result for one or more target items for the pathology slide image and/or the pathology slide image may be the vector 950 that indicates or characterizes one or more ROIs. In the present disclosure, the output variable of the artificial neural network model 900 is not limited to the types described above, and may include any information/data indicative of one or more ROIs.
As described above, the input layer 920 and the output layer 940 of the artificial neural network model 900 are respectively matched with a plurality of output variables corresponding to a plurality of input variables, and the synaptic values between nodes included in the input layer 920, the hidden layers 930_1 to 930_n, and the output layer 940 are adjusted, so that by training, a correct output corresponding to a specific input can be extracted. Through this training process, the features hidden in the input variables of the artificial neural network model 900 may be confirmed, and the synaptic values (or weights) between the nodes of the artificial neural network model 900 may be adjusted so as to reduce the errors between the output variable calculated based on the input variable and the target output. By using the artificial neural network model 900 trained in this way, in response to input of the pathology slide image, the detection result for target item may be output. Additionally or alternatively, by using the artificial neural network model 900, one or more ROIs may be output in response to input of the pathology slide image and/or the detection result for one or more target items for the pathology slide image (e.g., the pathology slide image including the detection result for target item).
The processors 1010 control the overall operation of components of the computing device 1000. The processors 1010 may be configured to include a central processing unit (CPU), a microprocessor unit (MPU), a micro controller unit (MCU), a graphic processing unit (GPU), or any type of processor well known in the technical field of the present disclosure. In addition, the processors 1010 may perform an arithmetic operation on at least one application or program for executing the method according to the embodiments of the present disclosure. The computing device 1000 may include one or more processors.
The memory 1020 may store various types of data, commands, and/or information. The memory 1020 may load one or more computer programs 1060 from the storage module 1050 in order to execute a method/operation according to various embodiments of the present disclosure. The memory 1020 may be implemented as a volatile memory such as RAM, but the technical scope of the present disclosure is not limited thereto.
The bus 1030 may provide a communication function between components of the computing device 1000. The bus 1030 may be implemented as various types of buses such as an address bus, a data bus, a control bus, or the like.
The communication interface 1040 may support wired/wireless Internet communication of the computing device 1000. In addition, the communication interface 1040 may support various other communication methods in addition to the Internet communication. To this end, the communication interface 1040 may be configured to include a communication module well known in the technical field of the present disclosure.
The storage module 1050 may non-temporarily store one or more computer programs 1060. The storage module 1050 may be configured to include a nonvolatile memory such as a read only memory (ROM), an erasable programmable ROM (EPROM), an electrically erasable programmable ROM (EEPROM), a flash memory, and the like, a hard disk, a detachable disk, or any type of computer-readable recording medium well known in the art to which the present disclosure pertains.
The computer program 1060 may include one or more instructions that, when loaded into the memory 1020, cause the processors 1010 to perform an operation/method in accordance with various embodiments of the present disclosure. That is, the processors 1010 may perform operations/methods according to various embodiments of the present disclosure by executing one or more instructions.
For example, the computer program 1060 may include one or more instructions for causing the following operations to be performed: obtaining information associated with immune phenotype for one or more ROIs in the pathology slide image; generating an image indicative of information associated with immune phenotype based on the information associated with immune phenotype for one or more ROIs; outputting the image indicative of the information associated with immune phenotype, and the like. In this case, a system for predicting a response to immune checkpoint inhibitor according to some embodiments of the present disclosure may be implemented through the computing device 1000.
The processor may generate an image indicating the detection result for one or more target items, and output the generated image indicating the detection result for one or more target items. In an embodiment, the processor may generate and output an image including visual representations capable of distinguishing respective target items, based on the detection result for one or more target items from the pathology slide image. For example, the processor may generate an image including a visual representation indicating each of the target items (e.g., cancer area, cancer stromal region, blood vessel, cancer cell, immune cell, positive/negative cells according to the expression amount of the biomarker, and the like) that may be considered in determining the immune phenotype. Additionally or alternatively, based on the detection result for one or more target items, the processor may generate an image including a segmentation map for the target item in units of regions, a contour for the target item having a specific structure (e.g., a blood vessel, and the like), a center point of the target item in units of cells, or a contour indicative of the shape of the target item in units of cells, and the like.
In an embodiment, the processor may output one or more ROIs in the pathology slide image and an image indicative of the detection result for one or more target items for one or more ROIs. For example, the processor may output the image generated as described above (e.g., the image including a visual representation) together with at least some regions of the pathology slide image (e.g., region corresponding to the generated image). As another example, the processor may overlay the image indicative of the detection result for one or more target items (e.g., the image including a visual representation) on one or more ROIs in the pathology slide image. For example, the user terminal may overlap the image (e.g., the image including a visual representation) generated as described above on at least some regions of the pathology slide image (e.g., the region corresponding to the generated image), and display it on a display device connected to the user terminal.
For example, as illustrated, the processor may display on the user interface an image 1110 displaying a first color indicating a background other than the target item, a second color indicating a cancer epithelium area, a third color indicating a cancer stromal region, a fourth color indicating an immune cell (e.g., a central point of an immune cell), and a fifth color indicating a cancer cell (e.g., a center point of a cancer cell) on the corresponding regions of the pathology slide image. In this example, the image 1110 may correspond to an image obtained by overlaying (or merging) the image indicative of the detection result for one or more target items on a corresponding region of the pathology slide image. Further, the first color, the second color, the third color, the fourth color and/or the fifth color may correspond to different colors that can be distinguished from each other.
Additionally or alternatively, as the information associated with immune phenotype, the user terminal may display on the user interface visual representations corresponding to respective target items (e.g., color information corresponding to respective target items, and the like), text, numerical values, markers, graphs, and the like for “analysis summary”. Additionally or alternatively, the user terminal may output a user interface through which the user can select a target item to be displayed with a visual representation in a corresponding region in the pathology slide image. In
For example, as illustrated, the processor may display on the user interface an image 1210 displaying a first color indicative of immune desert, a second color indicative of immune excluded, a third color indicative of immune inflamed, a fourth color indicative of cancer epithelium region, a fifth color indicative of a cancer stromal region, a sixth color indicative of an immune cell (e.g., a center point of immune cell), and a seventh color indicative of a cancer cell (e.g., a center point of a cancer cell) are displayed on the corresponding region in the pathology slide images. In this example, the image 1210 may correspond to an image in which an image (e.g., an immune phenotype map) indicative of the information associated with immune phenotype for one or more ROIs and an image indicative of the detection result for one or more target items are overlaid on (or merged with) the corresponding region in the pathology slide image. Further, the first color, the second color, the third color, the fourth color, the fifth color, the sixth color and/or the seventh color may correspond to different colors that can be distinguished from each other.
Additionally or alternatively, the processor may also display on the user interface an image 1220 of at least some regions of the pathology slide image (e.g., regions corresponding to the image including the visual representation and/or the entire pathology slide image) separately (e.g., in a form of a minimap). Additionally or alternatively, as the information associated with immune phenotype, the processor may display on the user interface the visual representations corresponding to the respective target items (e.g., information on color corresponding to respective target items, and the like), text, numerical values, graphs, and the like for “analysis summary”. Additionally or alternatively, a user interface may be output, through which the user may select the information associated with target item and/or immune phenotype to mark with a visual representation, in a corresponding region in the pathology slide image. In
The above description of the present disclosure is provided to enable those skilled in the art to make or use the present disclosure. Various modifications of the present disclosure will be readily apparent to those skilled in the art, and the general principles defined herein may be applied to various modifications without departing from the spirit or scope of the present disclosure. Thus, the present disclosure is not intended to be limited to the examples described herein but is intended to be accorded the broadest scope consistent with the principles and novel features disclosed herein.
Although example implementations may refer to utilizing aspects of the presently disclosed subject matter in the context of one or more standalone computer systems, the subject matter is not so limited, and they may be implemented in conjunction with any computing environment, such as a network or distributed computing environment. Furthermore, aspects of the presently disclosed subject matter may be implemented in or across a plurality of processing chips or devices, and storage may be similarly influenced across a plurality of devices. Such devices may include PCs, network servers, and handheld devices.
Although the present disclosure has been described in connection with some embodiments herein, it should be understood that various modifications and changes can be made without departing from the scope of the present disclosure, which can be understood by those skilled in the art to which the present disclosure pertains. In addition, such modifications and changes should be considered within the scope of the claims appended herein.
Claims
1. A method, performed by at least one computing device, for providing information associated with immune phenotype for pathology slide image, comprising:
- obtaining information associated with immune phenotype for one or more regions of interest (ROIs) in a pathology slide image;
- generating, based on the information associated with the immune phenotype for one or more ROIs, an image indicative of the information associated with the immune phenotype; and
- outputting the image indicative of the information associated with the immune phenotype.
2. The method according to claim 1, wherein the one or more ROIs are determined based on a detection result for one or more target items for the pathology slide image.
3. The method according to claim 2, wherein the one or more ROIs include at least some regions in the pathology slide image that satisfy a condition associated with the one or more target items.
4. The method according to claim 2, wherein the one or more ROIs are regions being output upon input of the detection result for the one or more target items for the pathology slide image or the pathology slide image to a region-of-interest (ROI) extraction model, and
- the ROI extraction model is trained to output a reference ROI upon input of the detection result for one or more target items for a reference pathology slide image or a reference pathology slide image.
5. The method according to claim 1, wherein the obtaining includes obtaining the immune phenotype of the one or more ROIs,
- the generating includes generating an image including a visual representation corresponding to the immune phenotype of the one or more ROIs, and
- the immune phenotype includes at least one of immune inflamed, immune excluded, or immune desert.
6. The method according to claim 5, wherein the outputting includes outputting the one or more ROIs in the pathology slide image together with the image including the visual representation.
7. The method according to claim 5, wherein the outputting includes overlaying the image including the visual representation on the one or more ROIs in the pathology slide image.
8. The method according to claim 1, wherein the obtaining includes obtaining one or more immune phenotype scores for the one or more ROIs,
- the generating includes generating an image including a visual representation corresponding to the one or more immune phenotype scores, and
- the one or more immune phenotype scores include at least one of a score for immune inflamed, a score for immune excluded, or a score for immune desert.
9. The method according to claim 8, wherein the outputting includes outputting the one or more ROIs in the pathology slide image together with the image including the visual representation.
10. The method according to claim 8, wherein the outputting includes overlaying the image including the visual representation on the one or more ROIs in the pathology slide image.
11. The method according to claim 1, wherein the obtaining includes obtaining a feature associated with one or more immune phenotypes for the one or more ROIs,
- the generating includes generating an image including a visual representation corresponding to the feature associated with the one or more immune phenotypes, and
- the feature associated with the one or more immune phenotypes includes at least one of a statistical value or a vector associated with the immune phenotype.
12. The method according to claim 11, wherein the outputting includes outputting the one or more ROIs in the pathology slide image together with the image including the visual representation.
13. The method according to claim 11, wherein the outputting includes overlaying the image including the visual representation on the one or more ROIs in the pathology slide image.
14. The method according to claim 1, further comprising:
- obtaining a detection result for one or more target items from the pathology slide image;
- generating an image indicative of the detection result for one or more target items; and
- outputting the image indicative of the detection result for one or more target items.
15. A non-transitory computer-readable recording medium storing a computer program for executing, on a computer, the method for providing the information associated with the immune phenotype for the pathology slide image according to claim 1.
16. A computing device comprising:
- a memory storing one or more instructions; and
- a processor configured to execute the stored one or more instructions to: obtain information associated with immune phenotype for one or more regions of interest (ROIs) in a pathology slide image; generate, based on the information associated with the immune phenotype for one or more ROIs, an image indicative of the information associated with the immune phenotype; and output the image indicative of the information associated with the immune phenotype.
17. The computing device according to claim 16, wherein the processor is further configured to:
- obtaining the immune phenotype of the one or more ROIs; and
- generate an image including a visual representation corresponding to the immune phenotype of the one or more ROIs, and
- the immune phenotype includes at least one of immune inflamed, immune excluded, or immune desert.
18. The computing device according to claim 17, wherein the processor is further configured to output the one or more ROIs in the pathology slide image together with the image including the visual representation.
19. The computing device according to claim 17, wherein the processor is further configured to overlay the image including the visual representation on the one or more ROIs in the pathology slide image.
20. The computing device according to claim 16, wherein the processor is further configured to:
- obtain one or more immune phenotype scores for the one or more ROIs; and
- generate an image including a visual representation corresponding to the one or more immune phenotype scores, and
- the one or more immune phenotype scores include at least one of a score for immune inflamed, a score for immune excluded, or a score for immune desert.
Type: Application
Filed: Oct 15, 2021
Publication Date: Feb 3, 2022
Applicant: LUNIT Inc. (Seoul)
Inventors: Donggeun Yoo (Seoul), Chanyoung Ock (Seoul), Kyunghyun Paeng (Seoul)
Application Number: 17/502,661