Knowledge-selective pretraining for attribute value extraction
Findings of the Association for Computational Linguistics: EMNLP 2023, 2023•aclanthology.org
Abstract Attribute Value Extraction (AVE) aims to retrieve the values of attributes from the
product profiles. The state-of-the-art methods tackle the AVE task through a question-
answering (QA) paradigm, where the value is predicted from the context (ie product profile)
given a query (ie attributes). Despite of the substantial advancements that have been made,
the performance of existing methods on rare attributes is still far from satisfaction, and they
cannot be easily extended to unseen attributes due to the poor generalization ability. In this …
product profiles. The state-of-the-art methods tackle the AVE task through a question-
answering (QA) paradigm, where the value is predicted from the context (ie product profile)
given a query (ie attributes). Despite of the substantial advancements that have been made,
the performance of existing methods on rare attributes is still far from satisfaction, and they
cannot be easily extended to unseen attributes due to the poor generalization ability. In this …
Abstract
Attribute Value Extraction (AVE) aims to retrieve the values of attributes from the product profiles. The state-of-the-art methods tackle the AVE task through a question-answering (QA) paradigm, where the value is predicted from the context (ie product profile) given a query (ie attributes). Despite of the substantial advancements that have been made, the performance of existing methods on rare attributes is still far from satisfaction, and they cannot be easily extended to unseen attributes due to the poor generalization ability. In this work, we propose to leverage pretraining and transfer learning to address the aforementioned weaknesses. We first collect the product information from various E-commerce stores and retrieve a large number of (profile, attribute, value) triples, which will be used as the pretraining corpus. To more effectively utilize the retrieved corpus, we further design a Knowledge-Selective Framework (KSelF) based on query expansion that can be closely combined with the pretraining corpus to boost the performance. Meanwhile, considering the public AE-pub dataset contains considerable noise, we construct and contribute a larger benchmark EC-AVE collected from E-commerce websites. We conduct evaluation on both of these datasets. The experimental results demonstrate that our proposed KSelF achieves new state-of-the-art performance without pretraining. When incorporated with the pretraining corpus, the performance of KSelF can be further improved, particularly on the attributes with limited training resources.
aclanthology.org
Showing the best result for this search. See all results