AI_for_biomedical_students

Projects

Projects for students

Project ideas

Code to perform CFD and solve ARC tasks https://github.com/neelsoumya/CFD_LLM_Accelerate24

https://github.com/neelsoumya/hands-on-llms/blob/main/Notebooks/arc_solver.ipynb

Code to create a healthcare AI chatbot https://github.com/neelsoumya/LLM-Handon

Code to use science-fiction to re-envision AI using LLMs https://github.com/neelsoumya/science_fiction_LLM

https://github.com/neelsoumya/science_fiction_AI_LLM

Code for using science-fiction and LLMs to re-envision life as we do not know it https://github.com/neelsoumya/science_fiction_life_LLM

Code to create a deepresearch tool for stock market and portfolio analysis https://github.com/neelsoumya/stock_portfolio_LLM_agent

Code to create a deepresearch tool to analyzing national AI strategies of different countries https://github.com/neelsoumya/intro_to_LMMs/blob/main/deep_research_agent_ag2_nationalAI_strategy.ipynb

Code for open source LLM for probing superintelligence (model organism of misalignment). https://github.com/neelsoumya/intro_to_LMMs/blob/main/L2_NLP_transformers.ipynb

Coding agents using smolagents https://github.com/neelsoumya/intro_to_LMMs/tree/main/agents/smolagents/L3

https://github.com/neelsoumya/intro_to_LMMs/blob/main/agents/smolagents/L3/Lesson3.ipynb

Other LLM projects https://github.com/neelsoumya/llm_projects

User interfaces Streamlit

https://docs.science.ai.cam.ac.uk/large-language-models/streamlit/

Dashboards

Explainable AI

Modify the code below

https://www.medrxiv.org/content/10.1101/2023.10.06.23296591v1

https://github.com/neelsoumya/complex_stories_explanations

https://www.medrxiv.org/content/10.1101/2024.03.25.24304824v1

https://github.com/Sharday/Enhancing_patient_stratification_explainable_AI

https://github.com/JoshanParmar/TuneableCounterfactuals

Miscellaneous projects

Higher level projects

Application areas can be any field of medicine such as oncology, cardiovascular, infectious diseases, etc.

Responsible AI

  1. Can AI internalize human morality? Objective: Investigate whether AI systems can form and apply moral judgments that reflect human values. Skills & methods: Survey design, crowdsourced data collection, statistical analysis, and moral choice theory.

  2. Co-developing a taxonomy of AI risks Objective: Build a grounded classification of AI-related risks by collaborating with a variety of stakeholders. Skills & methods: Co-design approaches, survey creation, statistical methods, and ethics of AI.

  3. Participatory auditing of AI systems Objective: Create and assess community-driven auditing techniques for evaluating AI behaviour and harms. Skills & methods: Survey and experimental design, (web) development, and applied AI ethics.

  4. Visualization tools for responsible decision-making with AI Objective: Design visual interfaces that help stakeholders make informed, responsible choices using AI outputs. Skills & methods: HCI, web development, and user-centred research.

  5. AI risks, opportunities, and media analysis at scale Objective: Analyze how news media portray AI — its benefits and dangers — using large-scale text analysis. Skills & methods: Large-scale data analysis, natural language processing, and ethical reflection.

  6. AI governance from an organizational lens Objective: Study how institutions govern AI by applying network and organisational analysis techniques. Skills & methods: Survey research, large-scale data analysis, and network science.

Small language models (SLMs) for healthcare in Global South

For students are medics first and programmers second, the goal is to leverage their domain expertise for data curation and evaluation while using “low-code” tools for the actual training.

In the Global South, the primary barriers aren’t just lack of doctors; they are language barriers, medical jargon, and remote triage. Small Language Models (SLMs) in the 1B–7B parameter range are perfect here because they can run on consumer laptops or cheap cloud instances.

Here are four concrete, small-scale project ideas:


1. The “Jargon Buster” for Rural Patients

The Goal: Fine-tune a model (like Llama-3.2-3B or Phi-3.5-Mini) to translate complex clinical diagnoses into local cultural metaphors or simplified language (e.g., Swahili, Hindi, or basic English).

2. Low-Resource Triage for Community Health Workers

The Goal: A decision-support SLM that helps non-doctor health workers in remote areas decide if a patient needs immediate transport to a city hospital.

3. Multilingual “Voice-to-Prescription” Assistant

The Goal: A model that takes a messy, spoken-style summary of a consultation in a local dialect and extracts a structured “Prescription & Plan” table.

4. Chronic Disease “Check-in” Bot for SMS/WhatsApp

The Goal: A very “thin” model (1B parameters, like Gemma-2B) designed to handle follow-up questions for Diabetes or Hypertension via low-bandwidth text.


The “Medic-Friendly” Technical Toolkit

  1. Model Selection: Stick to Llama-3.2 (1B/3B) or Phi-3.5-Mini. They are incredibly smart for their tiny size.
  2. Fine-Tuning Tool: Use Unsloth (via Google Colab). It’s a “wrapper” that makes fine-tuning 2x faster and uses 70% less memory. It’s essentially a 10-line Python script where they just point to their dataset.
  3. Data Generation: Have the medics use a “Teacher Model” (like GPT-4o) to help them expand their 50 handwritten examples into 500 synthetic examples, which they then manually “audit” for medical accuracy.
  4. Evaluation: This is where they shine. Instead of math metrics, have them perform a “Blind Clinical Review”—comparing the SLM’s output against a human doctor and scoring it on a 1-5 scale for safety and empathy.

More ideas

Pactical, low-math, and runnable with Python + simple web UI or notebooks. Each idea shows a clear undergraduate scope and a natural master’s-level extension.

Quick notes on common stacks

Python, Jupyter/Colab, Flask or Streamlit for a small UI, simple ML libs (scikit-learn, Hugging Face Transformers for prebuilt models), and basic REST calls to an LLM API (if allowed by your institution) are sufficient. Emphasize reproducibility: notebooks + README + short demo video.


Project ideas (title → short plan, UG scope, MSc extension, tools & deliverables)

  1. Red-Team / Blue-Team Simulator for Prompt Safety
  1. Output Provenance & Watermark Checker (Text)
  1. Adaptive Content Moderation Pipeline
  1. “Safe Suggestor” — Provide Non-Harmful Alternatives
  1. Honeypot Interaction Logger to Study Malicious Prompts (Ethical, Simulated)
  1. Model Output Robustness Checker (Paraphrase & Noise)
  1. Low-Cost Sandboxing: Policy-Enforced LLM Proxy
  1. Synthetic Data Detector for Downstream Classifiers
  1. Interactive User Study: How People Perceive Model Safety Interventions
  1. Explainable Flagging: Why Was This Output Blocked?
  1. Lightweight Threat Score Card for Model Responses
  1. Policy Update Simulator: How Fast Can a System Adapt?

Grading rubric (suggested)


Ethical & safety guidance to give students (very important)


Low code/no code project ideas for medics (human centered design/HCI)

The following is inspired by an HCI course from the University of Washington

To design for students with little programming experience, the focus must shift from back-end engineering to interaction design, ethics, and social contextualization.

Using a framework of Low AI (WhatsUp), Co-creative AI (Note Assist), and AI-Intensive (Replika).

These are designed to be built using “no-code” or “low-code” tools (like Voiceflow, Landbot, or simple GPT-wrappers) so students can focus on Human-Centered Design (HCD).


1. “The Cultural Navigator” (Low AI / Resource-Based)

Social Context: First-generation immigrants or refugees navigating a new healthcare system (e.g., the NHS in the UK or Medicare in the US).

2. “Neuro-Transitioner” (Co-creative / Note Assist style)

Social Context: Neurodivergent students (ADHD/Autism) transitioning from high school to university life.

3. “The Grief Archivist” (AI-Intensive / Replika style)

Social Context: Individuals dealing with “ambiguous loss” (e.g., a family member with dementia) or long-term bereavement.

4. “Caregiver’s Handover” (Co-creative / Administrative)

Social Context: Unpaid family caregivers (e.g., a daughter looking after an elderly parent) who need to communicate with rotating professional nurses.

5. “Safe-Space Simulator” (Low AI / Roleplay)

Social Context: LGBTQ+ youth in restrictive environments who are practicing “coming out” to a healthcare provider or asking for gender-affirming care.


Pedagogical Tip: The “Social Context” Audit

Students to complete a “Context Canvas” before they touch any software:

  1. The Actor: Who is the user (e.g., an elderly man living alone)?
  2. The Becoming: Who does he want to be? (e.g., He doesn’t want to be a “patient”; he wants to be “independent”).
  3. The AI’s Role: Is the AI a Mirror (Reflecting his thoughts), a Scaffold (Supporting his weaknesses), or a Bridge (Connecting him to a human doctor)?
  4. The Friction: What is one thing the AI should not do to preserve the user’s dignity?

By framing the project this way, the students are assessed on their empathy and logic rather than their ability to write Python code.

Scientific agents for biomedical research

Foundation models for health

Building multi modal foundation models for health (electronic healthcare records data, omics data, etc.)

Multi-agent systems for automated science

Building multi-agent systems, generative agents and generative AI systems for automated science.

We will also look at applying these frameworks to other sectors (such as finance, economics, defence, environment, policy, etc.)

Look at:

https://github.com/AstroPilot-AI/Denario

https://github.com/CMBAgents/cmbagent

https://github.com/CMBAgents/cmbagent/blob/main/docs/notebooks/cmbagent_beta2_demo.ipynb

https://arxiv.org/pdf/2507.07257

https://www.aisi.gov.uk/work/long-form-tasks