Over-parameterisation


Created: 17 Nov 2022, 05:16 PM | Modified: =dateformat(this.file.mtime,"dd MMM yyyy, hh:mm a") Tags: knowledge, MTL


MTL_MS1_2022 Presentation

[From MTL Problem Analysis Report:]{.underline}

Regularisation methods need to be carefully designed to avoid over-parameterised network such as AdaShare. <https://arxiv.org/abs/1911.12423>

  • Why is over-parameterisation bad here?

[From MTL research direction analysis report:]{.underline}

The study aims to design a deep MTL method using over-parameterised convolutional layers to improve model expressivity and to learn task-specific variations. In a broader perspective, this study falls into both the MTL architecture design and model compression domains.

  • Why is over-parameterisation good here?
  • Why do you do over-parameterisation for training and not test? Why does that help?

What does over-parameterisation mean?