Token-consistent Dropout for Calibrated Vision Transformers
Nikola Popovic, Danda Pani Paudel, Thomas Probst, Luc Van Gool
-
SPS
IEEE Members: $11.00
Non-members: $15.00
We introduce token-consistent dropout in vision transformers, which improves network calibration without causing any severe drop in performance. We use linear layers with token-consistent stochastic parameters inside the multilayer perceptron blocks, without altering the architecture of the transformer. The stochastic parameters are sampled from the uniform distribution, both during training and inference. The applied linear operations preserve the topological structure, formed by the set of tokens passing through the shared multilayer perceptron. This operation encourages the learning of the recognition task to rely on the topological structures of the tokens, instead of their values, which in turn offers the desired behavior. We compare our method to established baselines and applicable state-of-the-art alternatives, to demonstrate its impact on the calibration of vision transformers.