Please use this identifier to cite or link to this item: https://hdl.handle.net/2440/129158
Citations
Scopus Web of Science® Altmetric
?
?
Full metadata record
DC FieldValueLanguage
dc.contributor.authorTeney, D.-
dc.contributor.authorAbbasnejad, M.-
dc.contributor.authorVan Den Hengel, A.-
dc.contributor.editorVedaldi, A.-
dc.contributor.editorBischof, H.-
dc.contributor.editorBrox, T.-
dc.contributor.editorFrahm, J.-M.-
dc.date.issued2020-
dc.identifier.citationLecture Notes in Artificial Intelligence, 2020 / Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (ed./s), vol.12355, pp.580-599-
dc.identifier.isbn3030586065-
dc.identifier.isbn9783030586065-
dc.identifier.issn0302-9743-
dc.identifier.issn1611-3349-
dc.identifier.urihttp://hdl.handle.net/2440/129158-
dc.description.abstractOne of the primary challenges limiting the applicability of deep learning is its susceptibility to learning spurious correlations rather than the underlying mechanisms of the task of interest. The resulting failure to generalise cannot be addressed by simply using more data from the same distribution. We propose an auxiliary training objective that improves the generalization capabilities of neural networks by leveraging an overlooked supervisory signal found in existing datasets. We use pairs of minimally-different examples with different labels, a.k.a counterfactual or contrasting examples, which provide a signal indicative of the underlying causal structure of the task. We show that such pairs can be identified in a number of existing datasets in computer vision (visual question answering, multi-label image classification) and natural language processing (sentiment analysis, natural language inference). The new training objective orients the gradient of a model’s decision function with pairs of counterfactual examples. Models trained with this technique demonstrate improved performance on out-of-distribution test sets.-
dc.description.statementofresponsibilityDamien Teney, Ehsan Abbasnedjad, and Anton van den Hengel-
dc.language.isoen-
dc.publisherSpringer-
dc.relation.ispartofseriesLecture Notes in Computer Science; 12355-
dc.rights© Springer Nature Switzerland AG 2020-
dc.source.urihttps://link.springer.com/book/10.1007/978-3-030-58607-2-
dc.titleLearning what makes a difference from counterfactual examples and gradient supervision-
dc.typeConference paper-
dc.contributor.conferenceEuropean Conference on Computer Vision Workshops (ECCV) (23 Aug 2020 - 28 Aug 2020 : virtual online)-
dc.identifier.doi10.1007/978-3-030-58607-2_34-
dc.publisher.placeSwitzerland-
pubs.publication-statusPublished-
dc.identifier.orcidTeney, D. [0000-0003-2130-6650]-
dc.identifier.orcidVan Den Hengel, A. [0000-0003-3027-8364]-
Appears in Collections:Aurora harvest 4
Australian Institute for Machine Learning publications

Files in This Item:
There are no files associated with this item.


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.