Please use this identifier to cite or link to this item:
https://hdl.handle.net/10356/153544
Full metadata record
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Guo, Xu | en_US |
dc.contributor.author | Li, Boyang | en_US |
dc.contributor.author | Yu, Han | en_US |
dc.contributor.author | Miao, Chunyan | en_US |
dc.date.accessioned | 2021-12-12T07:12:13Z | - |
dc.date.available | 2021-12-12T07:12:13Z | - |
dc.date.issued | 2021 | - |
dc.identifier.citation | Guo, X., Li, B., Yu, H. & Miao, C. (2021). Latent-optimized adversarial neural transfer for sarcasm detection. Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, 5394-5407. | en_US |
dc.identifier.other | https://aclanthology.org/volumes/2021.naacl-main/ | - |
dc.identifier.uri | https://hdl.handle.net/10356/153544 | - |
dc.description.abstract | The existence of multiple datasets for sarcasm detection prompts us to apply transfer learning to exploit their commonality. The adversarial neural transfer (ANT) framework utilizes multiple loss terms that encourage the source-domain and the target-domain feature distributions to be similar while optimizing for domain-specific performance. However, these objectives may be in conflict, which can lead to optimization difficulties and sometimes diminished transfer. We propose a generalized latent optimization strategy that allows different losses to accommodate each other and improves training dynamics. The proposed method outperforms transfer learning and meta-learning baselines. In particular, we achieve 10.02% absolute performance gain over the previous state of the art on the iSarcasm dataset. | en_US |
dc.description.sponsorship | AI Singapore | en_US |
dc.description.sponsorship | Nanyang Technological University | en_US |
dc.description.sponsorship | National Research Foundation (NRF) | en_US |
dc.language.iso | en | en_US |
dc.relation | AISG2-RP-2020-019 | en_US |
dc.relation | NRF-NRFI05-2019-0002 | en_US |
dc.relation | NRF-NRFF13-2021-0006 | en_US |
dc.relation | NWJ2020-008 | en_US |
dc.relation | A20G8b0102 | en_US |
dc.relation | NSC-2019-011 | en_US |
dc.rights | © 2021 Association for Computational Linguistics. This is an open-access article distributed under the terms of the Creative Commons Attribution License. | en_US |
dc.subject | Engineering::Computer science and engineering | en_US |
dc.title | Latent-optimized adversarial neural transfer for sarcasm detection | en_US |
dc.type | Conference Paper | en |
dc.contributor.school | School of Computer Science and Engineering | en_US |
dc.contributor.conference | Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies | en_US |
dc.contributor.research | Joint NTU-UBC Research Centre of Excellence in Active Living for the Elderly (LILY) | en_US |
dc.description.version | Published version | en_US |
dc.identifier.spage | 5394 | en_US |
dc.identifier.epage | 5407 | en_US |
dc.subject.keywords | Transfer Learning | en_US |
dc.subject.keywords | Deep Learning Optimization | en_US |
dc.subject.keywords | Sarcasm Detection | en_US |
dc.citation.conferencelocation | Online | en_US |
dc.description.acknowledgement | This research is supported by the National Research Foundation, Singapore under its the AI Singapore Programme (AISG2-RP-2020-019), NRF Investigatorship (NRF-NRFI05-2019-0002), and NRF Fellowship (NRF-NRFF13-2021-0006); the Joint NTU-WeBank Research Centre on Fintech (NWJ-2020-008); the Nanyang Assistant/Associate Professorships (NAP); the RIE 2020 Advanced Manufacturing and Engineering Programmatic Fund (A20G8b0102), Singapore; NTU-SDU-CFAIR (NSC-2019-011). | en_US |
item.fulltext | With Fulltext | - |
item.grantfulltext | open | - |
Appears in Collections: | SCSE Conference Papers |
Files in This Item:
File | Description | Size | Format | |
---|---|---|---|---|
2021.naacl-main.425.pdf | 1.01 MB | Adobe PDF | View/Open |
Page view(s)
55
Updated on May 24, 2022
Download(s)
15
Updated on May 24, 2022
Google ScholarTM
Check
Items in DR-NTU are protected by copyright, with all rights reserved, unless otherwise indicated.