Skip to content

ai4colonoscopy/IntelliScope

Folders and files

NameName
Last commit message
Last commit date

Latest commit

ย 

History

35 Commits
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 

Repository files navigation


๐Ÿ”ฌ Paper (๐Ÿ”— arXiv, ๐Ÿค— Huggingface, ๐Ÿค– AIModels.fyi | ๐Ÿ“– ColonSurvey (๐Ÿ”— Online Sheet) | ๐Ÿฅ ColonINST (๐Ÿ”— Google Drive, ๐Ÿค— Huggingface) | ๐Ÿค– ColonGPT (๐Ÿ”— Google Drive, ๐Ÿค— Huggingface) | ๐Ÿ‡ Multimodal benchmark (๐Ÿ”— Google Drive, ๐Ÿ”— PaperWithCode)

Keyworks: Intelligent Colonoscopy, Multimodal Colonoscopy Dataset, Multimodal Language Model, Vision-language Understanding, Endoscopic Image Analysis, Healthcare AI, Abdomen.


Colonoscopy is currently one of the most sensitive screening methods for colorectal cancer (๐Ÿ”— Wikipedia). Have you ever wondered how to make colonoscopy smarter? Well, buckle up, let's enter the exciting world of intelligent colonoscopy!

  • Our vision. To explore the frontiers of intelligent colonoscopy techniques and their potential impact on multimodal medical applications.
  • Why use IntelliScope? It combines "Intelligent" and "colonoScope", where "Intelli" reflects the intelligent processing and decision-making capabilities of the system, and "Scope" refers to the colonoscope device used in medical endoscopy. Together, they imply a cutting-edge multimodal system designed to improve colonoscopy with advanced AI technologies.
  • Project members. Ge-Peng Ji (๐Ÿ‡ฆ๐Ÿ‡บ ANU), Jingyi Liu (๐Ÿ‡ฏ๐Ÿ‡ต Keio), Peng Xu (๐Ÿ‡จ๐Ÿ‡ณ THU), Nick Barnes (๐Ÿ‡ฆ๐Ÿ‡บ ANU), Fahad Shahbaz Khan (๐Ÿ‡ฆ๐Ÿ‡ช MBZUAI), Salman Khan (๐Ÿ‡ฆ๐Ÿ‡ช MBZUAI), Deng-Ping Fan (๐Ÿ‡จ๐Ÿ‡ณ NKU)
  • Let's join our IntelliScope community. We are building a discussion forum for the convenience of researchers to ๐Ÿ’ฌ ask any questions, ๐Ÿ’ฌ showcase/promote your work, ๐Ÿ’ฌ access data resources, and ๐Ÿ’ฌ share research ideas.
  • Quick view. Next, we present some features of our colonoscopy-specific AI chatbot, ColonGPT. This is a domain-pioneering multimodal language model that can help endoscopists perform various user-driven tasks through interactive dialogues.


Updates

  • [Nov/20/2024] We are working to a series of colonoscopy-specific multimodal chatbots with different lightweight LLM-brains. Watch our project and stay tuned!
  • [Oct/30/2024] We've set up an online benchmark on the paper-with-code website.
  • [Oct/16/2024] Open-source the whole codebase of project.

๐Ÿ”ฅ Research Highlights


Figure 1: Introductary diagram.

  • Survey on colonoscopic scene perception (CSP) โžก๏ธ "We assess the current landscape to sort out domain challenges and under-researched areas in the AI era."
    • ๐Ÿ“– ColonSurvey. We investigate the latest research progress in four colonoscopic scene perception tasks from both data-centric and model-centric perspectives. Our investigation summarises key features of 63 datasets and 137 representative deep techniques published since 2015. In addition, we highlight emerging trends and opportunities for future study. (๐Ÿ”— Hyperlink)
  • ๐Ÿ’ฅ Multimodal AI Initiatives โžก๏ธ "We advocate three initiatives to embrace the coming multimodal era in colonoscopy."
    • ๐Ÿฅ ColonINST. We introduce a pioneering instruction tuning dataset for multimodal colonoscopy research, aimed at instructing models to execute user-driven tasks interactively. This dataset comprises of 62 categories, 300K+ colonoscopic images, 128K+ medical captions (GPT-4V) generated), and 450K+ human-machine dialogues. (๐Ÿ”— Hyperlink)
    • ๐Ÿค– ColonGPT. We develop a domain-specific multimodal language model to assist endoscopists through interactive dialogues. To ensure reproducibility for average community users, we implement ColonGPT in a resource-friendly way, including three core designs: a 0.4B-parameter visual encoder ๐Ÿค— SigLIP-SO from Google, a 1.3B-parameter lightweight language model ๐Ÿค— Phi1.5 from Microsoft, and a multigranularity adapter for token reducing from 100% to only 33.74% but not compromise to performance. (๐Ÿ”— Hyperlink)
    • ๐Ÿ’ฏ Multimodal Benchmark. We contribute a multimodal benchmark, including six general-purpose models and two designed for medical purposes, across three colonoscopy tasks to enable fair and rapid comparisons going forward. (๐Ÿ”— Hyperlink)

๐Ÿ“– ColonSurvey

Our "ColonSurvey" project contributes various useful resources for the community. We investigate 63 colonoscopy datasets and 137 deep learning models focused on colonoscopic scene perception, all sourced from leading conferences or journals since 2015. This is a quick overview of our investigation; for a more detailed discussion, please refer to our paper in PDF format.


Figure 2: The investigation of colonoscopy datasets and models.

To better understand developments in this rapidly changing field and accelerate researchersโ€™ progress, we are building a ๐Ÿ“– paper reading list, which includes a number of AI-based scientific studies on colonoscopy imaging from the past 12 years. [UPDATE ON OCT-14-2024] In detail, our online list contains:

Make our community great again. If we miss your valuable work in google sheet, please add it and this project would be a nice platform to promote your work. Or anyone can inform us via email (๐Ÿ“ฎ gepengai.ji@gmail.com) or push a PR in github. We will work on your request as soon as possible. Thank you for your active feedback.

๐Ÿฅ ColonINST (A multimodal instruction tuning dataset)


Figure 3: Details of our multimodal instruction tuning dataset, ColonINST. (a) Three sequential steps to create the instruction tuning dataset for multimodal research. (b) Numbers of colonoscopy images designated for training, validation, and testing purposes. (c) Data taxonomy of three-level categories. (d) A word cloud of the category distribution by name size. (e) Caption generation pipeline using the VL prompting mode of GPT-4V. (f) Numbers of human-machine dialogues created for four downstream tasks.

Our data contains two parts: colonoscopy images and human-machine dialogues (available at ๐Ÿค— huggingface and ๐Ÿ”— google drive). However, due to privacy-preserving concerns, we can not directly share the origin medical images without its authorization. DO NOT WORRY! We prepare a data download list and an easy-to-use script to organise our ColonINST. The operation instructions are detailed in our document (๐Ÿ”— .docs/guideline-for-ColonINST.md)

๐Ÿค– ColonGPT (A colonoscopy-specific multimodal Language Model)


Figure 4: Details of our multimodal language model, ColonGPT.

Our ColonGPT is a standard multimodal language model, which has been released at ๐Ÿ”— google drive platform. It contains four basic components: a language tokenizer, an visual encoder (๐Ÿค— SigLIP-SO or ๐Ÿ”— google drive), a multimodal connector, and a language model (๐Ÿค— Phi1.5 or ๐Ÿ”— google drive).

โœ… Quick start

We show a code snippet to show you how to quickly try-on our ColonGPT model with HuggingFace transformers quickly. For convenience, we manually combined some configuration and code files and merged the weights. Please note that this is a quick code, we recommend you installing ColonGPT's source code to explore more.

  • Before running the snippet, you only need to install the following minimum dependencies.

    conda create -n quickstart python=3.10
    conda activate quickstart
    pip install torch transformers accelerate pillow
  • Then you can use python script/quick_start/quickstart.py to start.

    import torch
    import transformers
    from transformers import AutoModelForCausalLM, AutoTokenizer, StoppingCriteria
    from PIL import Image
    import warnings
    
    transformers.logging.set_verbosity_error()
    transformers.logging.disable_progress_bar()
    warnings.filterwarnings('ignore')
    
    device = 'cuda'  # or cpu
    torch.set_default_device(device)
    
    model_name = "ai4colonoscopy/ColonGPT-v1"
    
    model = AutoModelForCausalLM.from_pretrained(
        model_name,
        torch_dtype=torch.float16,  # or float32 for cpu
        device_map='auto',
        trust_remote_code=True
    )
    
    tokenizer = AutoTokenizer.from_pretrained(
        model_name,
        trust_remote_code=True
    )
    
    class KeywordsStoppingCriteria(StoppingCriteria):
        def __init__(self, keyword, tokenizer, input_ids):
            self.keyword_id = tokenizer(keyword).input_ids
            self.tokenizer = tokenizer
            self.start_len = input_ids.shape[1]
    
        def __call__(self, input_ids: torch.LongTensor, scores: torch.FloatTensor, **kwargs) -> bool:
            for keyword_id in self.keyword_id:
                if keyword_id in input_ids[0, -len(self.keyword_id):]:
                    return True
            return False
    
    prompt = "Describe what you see in the image." 
    text = f"USER: <image>\n{prompt} ASSISTANT:"
    text_chunks = [tokenizer(chunk).input_ids for chunk in text.split('<image>')]
    input_ids = torch.tensor(text_chunks[0] + [-200] + text_chunks[1], dtype=torch.long).unsqueeze(0).to(device)
    
    image = Image.open('cache/examples/example2.png')
    image_tensor = model.process_images([image], model.config).to(dtype=model.dtype, device=device)
    
    stop_str = "<|endoftext|>"
    stopping_criteria = KeywordsStoppingCriteria(stop_str, tokenizer, input_ids)
    
    output_ids = model.generate(
        input_ids,
        images=image_tensor,
        do_sample=False,
        temperature=0,
        max_new_tokens=512,
        use_cache=True,
        stopping_criteria=[stopping_criteria]
    )
    
    outputs = tokenizer.decode(output_ids[0, input_ids.shape[1]:]).replace("<|endoftext|>", "").strip()
    print(outputs)

โœ… Try full-version ColonGPT

The complete workflow scheme has been officially released, providing a streamlined and user-friendly process. This allows even average community users to easily develop, customize, and enhance their own models. To ensure a smooth experience, we have included comprehensive step-by-step instructions (๐Ÿ”— .docs/guideline-for-ColonGPT.md), which can be found in our detailed documentation. These resources are designed to guide users through every stage of the development process, making it accessible and efficient for both beginners and advanced practitioners.

  • ๐Ÿค– How to train? Our model is a resource-friendly multimodal language model that can be trained in under five hours using just four A100-40GB GPUs. This efficient model is also compatible with other Amperer GPUs, such as NVIDIA 4090 card, making it accessible to a wider range of hardware setups.
  • ๐Ÿ‘จโ€๐Ÿ”ฌ How to infer? We support three inference modes, including batch inference, CLI inference, and Gradio WebUI inference.

๐Ÿ’ฏ Multimodal benchmark


Figure 4: Multimodal benchmark.

We provide a comprehensive benchmark of eight latest multimodal competitors across three multimodal colonoscopy tasks, including MiniGPT-V2, LLaVA-v1, LLaVA-v1.5, Bunny-v1.0-3B, Mini-Gemini-2B, MobileVLM-1.7B, and LLaVA-Med-v1.0, and LLaVA-Med-v1.5. We provide ๐Ÿ”— the meta prediction files, ๐Ÿ”— the evaluation instructions. We believe these resources facilitate everyone to conveniently access their newly developed model, or rapidly conduct proof-of-concept development for follow-up research.

๐Ÿ™ Acknowledgement

We gratefully acknowledge the contributions of the following projects, which served as the foundation and inspiration for our work:

  • Bunny: A family of lightweight multimodal models
  • LAVIS: A Library for Language-Vision Intelligence
  • LLaVA: Large Language and Vision Assistant

๐Ÿ‘ Citation

Please use the following reference if you find this project useful for your research or applications:

@article{ji2024frontiers,
  author = {Ji, Ge-Peng and Liu, Jingyi and Xu, Peng and Barnes, Nick and Khan, Fahad Shahbaz and Khan, Salman and Fan, Deng-Ping},
  title = {Frontiers in Intelligent Colonoscopy},
  journal = {arXiv preprint arXiv:2410.17241},
  year = {2024}
}

๐Ÿšจ Ethical and Responsible Use

ColonGPT is designed to assist in medical colonoscopy by leveraging multimodal capabilities, but it comes with no guarantees regarding its predictive accuracy or reliability in clinical practice. Users should be aware that the datasets and pre-trained models used in ColonGPT may contain inherent biases, including socioeconomic factors, which can lead to misclassification or other undesirable behaviors, such as the generation of offensive or inappropriate content.

We urge users and developers to carefully review and validate the performance of pre-trained models, particularly those integrated through the ColonGPT framework, before considering practical applications in a clinical setting. It is crucial that any AI-driven tool used in healthcare undergoes rigorous testing to ensure patient safety and avoid unintended consequences. Our commitment to ethical AI use extends to ongoing efforts to investigate, address, and mitigate the risks of bias and inappropriate behavior in ColonGPT. Continuous improvement of this codebase is a priority to ensure that the system aligns with responsible and equitable healthcare standards.