Homogeneous-listing-augmented Self-supervised Multimodal Product Title Refinement
Paper
| Paper/Presentation Title | Homogeneous-listing-augmented Self-supervised Multimodal Product Title Refinement |
|---|---|
| Presentation Type | Paper |
| Authors | Deng, Jiaqi, Shi, Kaize, Huo, Huan, Wang, Dingxian and Xu, Guandong |
| Journal or Proceedings Title | Proceedings of the 47th International ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR ’24) |
| Journal Citation | pp. 2870-2874 |
| Number of Pages | 5 |
| Year | 2024 |
| Publisher | Association for Computing Machinery (ACM) |
| Place of Publication | United States |
| ISBN | 9798400704314 |
| Digital Object Identifier (DOI) | https://doi.org/10.1145/3626772.3661347 |
| Web Address (URL) of Paper | https://dl.acm.org/doi/abs/10.1145/3626772.3661347 |
| Web Address (URL) of Conference Proceedings | https://dl.acm.org/doi/proceedings/10.1145/3626772 |
| Conference/Event | 47th International ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR ’24) |
| Event Details | 47th International ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR ’24) Parent ACM International Conference on Research and Development in Information Retrieval Delivery In person Event Date 14 to end of 18 Jul 2024 Event Location Washington DC, United States |
| Abstract | Product titles on e-commerce marketplaces often suffer from verbosity and inaccuracy, hindering effective communication of essential product details to customers. Refining titles to be more concise and informative is crucial for better user experience and product promotion. Recent solutions to product title refinement follow the standard text extractive and generative methods. Some also leverage multimodal information, e.g. using product images to supplement original titles with visual knowledge. However, these generative methods often produce additional terms not endorsed by sellers. Thus, it remains challenging to incorporate visual information missing from original titles into refined titles without excessively introducing novel terms. Additionally, most existing methods require human-labeled datasets, which are laborious to construct. In response to the two challenges, we present a self-supervised multimodal framework (HLATR) for title refinement that comprises two key modules: (1) a perturbated sample generator that constructs training data by systematically mining homogeneous listing information and (2) a title refinement network that effectively harnesses visual information to refine the original titles. To explicitly balance the extraction from original titles and the generation of supplementary novel terms, we adapt the copy mechanism that is guided by a focused refinement loss. Extensive experiments demonstrate that our proposed framework consistently outperforms others in generating refined titles that contain essential multimodal semantics with minimal deviation from the original ones. |
| Keywords | Product title refinement; Multimodal generative mod; Self-supervised learning |
| Contains Sensitive Content | Does not contain sensitive content |
| ANZSRC Field of Research 2020 | 4602. Artificial intelligence |
| Byline Affiliations | University of Technology Sydney |
| Etsy.com, United States |
https://research.usq.edu.au/item/100973/homogeneous-listing-augmented-self-supervised-multimodal-product-title-refinement
Download files
80
total views4
total downloads74
views this month2
downloads this month