Low-Compute Test-Time Adaptation Research Track

Overview

This page is for students who are interested in improving models at test time while keeping GPU usage as small as possible.

You do not need to start with large-scale training or expensive hardware. This track is designed for students who want a modern research topic with realistic experiments and clear questions.

The goal of this track is to explore questions such as:

If you are interested in these questions, this track may be a good place to start.

What to avoid at the beginning

Do not begin with:

When to contact me

If you read some of the papers on this page and feel interested, feel free to contact me.

Part I. A simple starting path

A good starting path is the following:

You do not need to do everything at once.

Part II. Core background for efficient test-time learning

These are the most important shared starting points for this page.

1. CLIP (ICML 2021)

Paper: Learning Transferable Visual Models From Natural Language Supervision
Code: https://github.com/openai/CLIP

Why read it: a strong starting point for modern vision-language transfer
Focus on: image-text alignment, zero-shot classification, prompt-based inference

2. Tent (ICLR 2021)

Paper: Fully Test-Time Adaptation by Entropy Minimization
Code: https://github.com/DequanWang/tent

Why read it: one of the simplest and most influential starting points for test-time adaptation
Focus on: entropy minimization, online adaptation, normalization-based updates

3. TDA (CVPR 2024)

Paper: Efficient Test-Time Adaptation of Vision-Language Models
Code: https://github.com/kdiAAA/TDA

Why read it: a clear starting point for efficient test-time adaptation in multimodal settings
Focus on: training-free adaptation, cache-based updates, low-cost improvement at inference time

Part III. Main track — Efficient Test-Time Adaptation

Typical question:

How can we improve a pretrained model on shifted data without expensive retraining?

Why this track is good

This track is suitable for students who want:

Possible directions

A. Vision-only test-time adaptation

Why study it: often the simplest place to begin
Good for students because: experiments are lighter and the core adaptation issue is easier to isolate

Recommended papers:

B. Language or multimodal test-time adaptation

Why study it: modern models increasingly rely on cross-modal or language-conditioned inference
Good for students because: this direction is timely and closely connected to current foundation model research

Recommended papers:

C. Vision-Language-Action adaptation (emerging direction)

Why study it: action-conditioned systems face deployment shift and low-latency constraints, so test-time improvement is especially valuable at deployment time.
Good for students because: this can become a distinctive topic if framed carefully around test-time robustness, online correction, and compute limits

Recommended papers:

Why these papers matter:

Possible direction:

What students should reproduce first in this track

Choose one:

Then compare it against:

Good starter experiments for this track

Part IV. A promising publication direction

A strong project in this track could be:

Realistic and efficient test-time adaptation under strict compute constraints

This direction is attractive because it combines:

The page should stay focused on one question:

How can we improve a deployed model at test time with minimal extra compute?

Part V. Good starter benchmarks

Avoid very large datasets at the beginning.

Part VI. Suggested first mini-project

A strong first project is:

This is a good starting point because it answers a concrete and modern question:

Can test-time adaptation remain useful when compute is limited and deployment is realistic?

Final note

It is better to have one clear track than to force separate tracks with the same papers.

So this page should stay centered on: