(a) The input is composed of a live cell microscopy sequence of length T and the corresponding sequence of label maps.
(b) Each cell instance in the sequence is represented by a feature vector which includes DML and spatio-temporal features.
(c) The entire microscopy sequence is encoded as a direct graph where the cell instances are represented by its nodes and their associations are represented by the graph edges. Each node and edge in the graph has its own embedded feature vector.
(d) These feature vectors are encoded and updated using Graph Neural Network (GNN). The GNN is composed of L message passing blocks which enable an update of edge and node features by their L-th order neighbors (i.e., cell instances which are up to L frames apart).
(e) The GNN’s edge feature output is the input for an edge classifier network which classifies the edges into active (solid lines) and non-active (dashed lines). During training, the predicted classification is compared to the GT classification for the loss computation. Since all the framework components are connected in an end-to-end manner the loss backpropogates throughout the entire network.
(f) At inference time, cell tracks are constructed by concatenating sequences of active edges that connect cells in consecutive frames.
@inproceedings{ben2022graph,
title={Graph Neural Network for Cell Tracking in Microscopy Videos},
author={Ben-Haim, Tal and Riklin-Raviv, Tammy},
booktitle={Proceedings of the European Conference on Computer Vision (ECCV)},
year={2022},
}