Search Filter Ranking with Language-Aware Label Embeddings

Publikation: Beitrag in Buch/Bericht/Sammelwerk/KonferenzbandAufsatz in KonferenzbandForschungPeer-Review

Autoren

  • Jacek Golebiowski
  • Felice Antonio Merra
  • Ziawasch Abedjan
  • Felix Biessmann

Externe Organisationen

  • Amazon.com, Inc.
  • Politecnico di Bari
  • Berlin International University of Applied Sciences
Forschungs-netzwerk anzeigen

Details

OriginalspracheEnglisch
Titel des SammelwerksWWW 2022 - Companion Proceedings of the Web Conference 2022
Seiten121-125
Seitenumfang5
ISBN (elektronisch)9781450391306
PublikationsstatusVeröffentlicht - 25 Apr. 2022
Veranstaltung31st ACM Web Conference, WWW 2022 - Virtual, Online, Frankreich
Dauer: 25 Apr. 202229 Apr. 2022

Abstract

A search on the major eCommerce platforms returns up to thousands of relevant products making it impossible for an average customer to audit all the results. Browsing the list of relevant items can be simplified using search filters for specific requirements (e.g., shoes of the wrong size). The complete list of available filters is often overwhelming and hard to visualize. Thus, successful user interfaces desire to display only the ones relevant to customer queries. In this work, we frame the filter selection task as an extreme multi-label classification (XMLC) problem based on historical interaction with eCommerce sites. We learn from customers' clicks and purchases which subset of filters is most relevant to their queries treating the relevant/not-relevant signal as binary labels. A common problem in classification settings with a large number of classes is that some classes are underrepresented. These rare categories are difficult to predict. Building on previous work we show that classification performance for rare classes can be improved by accounting for the language structure of the class labels. Furthermore, our results demonstrate that including language structure in category names enables relatively simple deep learning models to achieve better predictive performance than transformer networks with much higher capacity.

ASJC Scopus Sachgebiete

Zitieren

Search Filter Ranking with Language-Aware Label Embeddings. / Golebiowski, Jacek; Merra, Felice Antonio; Abedjan, Ziawasch et al.
WWW 2022 - Companion Proceedings of the Web Conference 2022. 2022. S. 121-125.

Publikation: Beitrag in Buch/Bericht/Sammelwerk/KonferenzbandAufsatz in KonferenzbandForschungPeer-Review

Golebiowski, J, Merra, FA, Abedjan, Z & Biessmann, F 2022, Search Filter Ranking with Language-Aware Label Embeddings. in WWW 2022 - Companion Proceedings of the Web Conference 2022. S. 121-125, 31st ACM Web Conference, WWW 2022, Virtual, Online, Frankreich, 25 Apr. 2022. https://doi.org/10.1145/3487553.3524218
Golebiowski, J., Merra, F. A., Abedjan, Z., & Biessmann, F. (2022). Search Filter Ranking with Language-Aware Label Embeddings. In WWW 2022 - Companion Proceedings of the Web Conference 2022 (S. 121-125) https://doi.org/10.1145/3487553.3524218
Golebiowski J, Merra FA, Abedjan Z, Biessmann F. Search Filter Ranking with Language-Aware Label Embeddings. in WWW 2022 - Companion Proceedings of the Web Conference 2022. 2022. S. 121-125 doi: 10.1145/3487553.3524218
Golebiowski, Jacek ; Merra, Felice Antonio ; Abedjan, Ziawasch et al. / Search Filter Ranking with Language-Aware Label Embeddings. WWW 2022 - Companion Proceedings of the Web Conference 2022. 2022. S. 121-125
Download
@inproceedings{43b73db8ce364ea9be98ab0465c8d600,
title = "Search Filter Ranking with Language-Aware Label Embeddings",
abstract = "A search on the major eCommerce platforms returns up to thousands of relevant products making it impossible for an average customer to audit all the results. Browsing the list of relevant items can be simplified using search filters for specific requirements (e.g., shoes of the wrong size). The complete list of available filters is often overwhelming and hard to visualize. Thus, successful user interfaces desire to display only the ones relevant to customer queries. In this work, we frame the filter selection task as an extreme multi-label classification (XMLC) problem based on historical interaction with eCommerce sites. We learn from customers' clicks and purchases which subset of filters is most relevant to their queries treating the relevant/not-relevant signal as binary labels. A common problem in classification settings with a large number of classes is that some classes are underrepresented. These rare categories are difficult to predict. Building on previous work we show that classification performance for rare classes can be improved by accounting for the language structure of the class labels. Furthermore, our results demonstrate that including language structure in category names enables relatively simple deep learning models to achieve better predictive performance than transformer networks with much higher capacity.",
keywords = "Information Retrieval, Ranking, Search Filters",
author = "Jacek Golebiowski and Merra, {Felice Antonio} and Ziawasch Abedjan and Felix Biessmann",
year = "2022",
month = apr,
day = "25",
doi = "10.1145/3487553.3524218",
language = "English",
pages = "121--125",
booktitle = "WWW 2022 - Companion Proceedings of the Web Conference 2022",
note = "31st ACM Web Conference, WWW 2022 ; Conference date: 25-04-2022 Through 29-04-2022",

}

Download

TY - GEN

T1 - Search Filter Ranking with Language-Aware Label Embeddings

AU - Golebiowski, Jacek

AU - Merra, Felice Antonio

AU - Abedjan, Ziawasch

AU - Biessmann, Felix

PY - 2022/4/25

Y1 - 2022/4/25

N2 - A search on the major eCommerce platforms returns up to thousands of relevant products making it impossible for an average customer to audit all the results. Browsing the list of relevant items can be simplified using search filters for specific requirements (e.g., shoes of the wrong size). The complete list of available filters is often overwhelming and hard to visualize. Thus, successful user interfaces desire to display only the ones relevant to customer queries. In this work, we frame the filter selection task as an extreme multi-label classification (XMLC) problem based on historical interaction with eCommerce sites. We learn from customers' clicks and purchases which subset of filters is most relevant to their queries treating the relevant/not-relevant signal as binary labels. A common problem in classification settings with a large number of classes is that some classes are underrepresented. These rare categories are difficult to predict. Building on previous work we show that classification performance for rare classes can be improved by accounting for the language structure of the class labels. Furthermore, our results demonstrate that including language structure in category names enables relatively simple deep learning models to achieve better predictive performance than transformer networks with much higher capacity.

AB - A search on the major eCommerce platforms returns up to thousands of relevant products making it impossible for an average customer to audit all the results. Browsing the list of relevant items can be simplified using search filters for specific requirements (e.g., shoes of the wrong size). The complete list of available filters is often overwhelming and hard to visualize. Thus, successful user interfaces desire to display only the ones relevant to customer queries. In this work, we frame the filter selection task as an extreme multi-label classification (XMLC) problem based on historical interaction with eCommerce sites. We learn from customers' clicks and purchases which subset of filters is most relevant to their queries treating the relevant/not-relevant signal as binary labels. A common problem in classification settings with a large number of classes is that some classes are underrepresented. These rare categories are difficult to predict. Building on previous work we show that classification performance for rare classes can be improved by accounting for the language structure of the class labels. Furthermore, our results demonstrate that including language structure in category names enables relatively simple deep learning models to achieve better predictive performance than transformer networks with much higher capacity.

KW - Information Retrieval

KW - Ranking

KW - Search Filters

UR - http://www.scopus.com/inward/record.url?scp=85137457223&partnerID=8YFLogxK

U2 - 10.1145/3487553.3524218

DO - 10.1145/3487553.3524218

M3 - Conference contribution

AN - SCOPUS:85137457223

SP - 121

EP - 125

BT - WWW 2022 - Companion Proceedings of the Web Conference 2022

T2 - 31st ACM Web Conference, WWW 2022

Y2 - 25 April 2022 through 29 April 2022

ER -