Workshop on Large Language Models' Interpretability and Trustworthiness: (LLMIT)

Publikation: Beitrag in Buch/Bericht/Sammelwerk/KonferenzbandAufsatz in KonferenzbandForschungPeer-Review

Autoren

  • Tulika Saha
  • Sriparna Saha
  • Debasis Ganguly
  • Prasenjit Mitra

Organisationseinheiten

Externe Organisationen

  • The University of Liverpool
  • Indian Institute of Technology Patna (IITP)
  • University of Glasgow
  • Pennsylvania State University
Forschungs-netzwerk anzeigen

Details

OriginalspracheEnglisch
Titel des SammelwerksCIKM 2023
UntertitelProceedings of the 32nd ACM International Conference on Information and Knowledge Management
Herausgeber (Verlag)Association for Computing Machinery (ACM)
Seiten5290-5293
Seitenumfang4
ISBN (elektronisch)9798400701245
PublikationsstatusVeröffentlicht - 21 Okt. 2023
Veranstaltung32nd ACM International Conference on Information and Knowledge Management, CIKM 2023 - Birmingham, Großbritannien / Vereinigtes Königreich
Dauer: 21 Okt. 202325 Okt. 2023

Publikationsreihe

NameInternational Conference on Information and Knowledge Management, Proceedings

Abstract

Large language models (LLMs), when scaled from millions to billions of parameters, have been demonstrated to exhibit the so-called 'emergence' effect, in that they are not only able to produce semantically correct and coherent text, but are also able to adapt themselves surprisingly well with small changes in contexts supplied as inputs (commonly called prompts). Despite producing semantically coherent and potentially relevant text for a given context, LLMs are vulnerable to yield incorrect information. This misinformation generation, or the so-called hallucination problem of an LLM, gets worse when an adversary manipulates the prompts to their own advantage, e.g., generating false propaganda to disrupt communal harmony, generating false information to trap consumers with target consumables etc. Not only does the consumption of an LLM-generated hallucinated content by humans pose societal threats, such misinformation, when used as prompts, may lead to detrimental effects for in-context learning (also known as few-shot prompt learning). With reference to the above-mentioned problems of LLM usage, we argue that it is necessary to foster research on topics related to not only identifying misinformation from LLM-generated content, but also to mitigate the propagation effects of this generated misinformation on downstream predictive tasks thus leading to more robust and effective leveraging in-context learning.

Zitieren

Workshop on Large Language Models' Interpretability and Trustworthiness: (LLMIT). / Saha, Tulika; Saha, Sriparna; Ganguly, Debasis et al.
CIKM 2023 : Proceedings of the 32nd ACM International Conference on Information and Knowledge Management. Association for Computing Machinery (ACM), 2023. S. 5290-5293 (International Conference on Information and Knowledge Management, Proceedings).

Publikation: Beitrag in Buch/Bericht/Sammelwerk/KonferenzbandAufsatz in KonferenzbandForschungPeer-Review

Saha, T, Saha, S, Ganguly, D & Mitra, P 2023, Workshop on Large Language Models' Interpretability and Trustworthiness: (LLMIT). in CIKM 2023 : Proceedings of the 32nd ACM International Conference on Information and Knowledge Management. International Conference on Information and Knowledge Management, Proceedings, Association for Computing Machinery (ACM), S. 5290-5293, 32nd ACM International Conference on Information and Knowledge Management, CIKM 2023, Birmingham, Großbritannien / Vereinigtes Königreich, 21 Okt. 2023. https://doi.org/10.1145/3583780.3615311
Saha, T., Saha, S., Ganguly, D., & Mitra, P. (2023). Workshop on Large Language Models' Interpretability and Trustworthiness: (LLMIT). In CIKM 2023 : Proceedings of the 32nd ACM International Conference on Information and Knowledge Management (S. 5290-5293). (International Conference on Information and Knowledge Management, Proceedings). Association for Computing Machinery (ACM). https://doi.org/10.1145/3583780.3615311
Saha T, Saha S, Ganguly D, Mitra P. Workshop on Large Language Models' Interpretability and Trustworthiness: (LLMIT). in CIKM 2023 : Proceedings of the 32nd ACM International Conference on Information and Knowledge Management. Association for Computing Machinery (ACM). 2023. S. 5290-5293. (International Conference on Information and Knowledge Management, Proceedings). doi: 10.1145/3583780.3615311
Saha, Tulika ; Saha, Sriparna ; Ganguly, Debasis et al. / Workshop on Large Language Models' Interpretability and Trustworthiness : (LLMIT). CIKM 2023 : Proceedings of the 32nd ACM International Conference on Information and Knowledge Management. Association for Computing Machinery (ACM), 2023. S. 5290-5293 (International Conference on Information and Knowledge Management, Proceedings).
Download
@inproceedings{871f95a6657442f783d443d021a21dc5,
title = "Workshop on Large Language Models' Interpretability and Trustworthiness: (LLMIT)",
abstract = "Large language models (LLMs), when scaled from millions to billions of parameters, have been demonstrated to exhibit the so-called 'emergence' effect, in that they are not only able to produce semantically correct and coherent text, but are also able to adapt themselves surprisingly well with small changes in contexts supplied as inputs (commonly called prompts). Despite producing semantically coherent and potentially relevant text for a given context, LLMs are vulnerable to yield incorrect information. This misinformation generation, or the so-called hallucination problem of an LLM, gets worse when an adversary manipulates the prompts to their own advantage, e.g., generating false propaganda to disrupt communal harmony, generating false information to trap consumers with target consumables etc. Not only does the consumption of an LLM-generated hallucinated content by humans pose societal threats, such misinformation, when used as prompts, may lead to detrimental effects for in-context learning (also known as few-shot prompt learning). With reference to the above-mentioned problems of LLM usage, we argue that it is necessary to foster research on topics related to not only identifying misinformation from LLM-generated content, but also to mitigate the propagation effects of this generated misinformation on downstream predictive tasks thus leading to more robust and effective leveraging in-context learning.",
keywords = "Explainability, In-context Learning, Interpretability, Large Language Model, Trustworthiness",
author = "Tulika Saha and Sriparna Saha and Debasis Ganguly and Prasenjit Mitra",
note = "Funding information: (4) Prasenjit Mitra is a Professor at The Pennsylvania State Uni-versity and a visiting Professor at the L3S Center at the Leibniz University at Hannover, Germany. He obtained his Ph.D. from Stanford University in 2003 in Electrical Engineering and has been at Penn State since. His research interests are in artifi-cial intelligence, applied machine learning, natural language processing, etc. His research has been supported by the NSF CAREER award, the DoE, DoD, Microsoft Research, Raytheon, Lockheed Martin, Dow Chemicals, McDonnell Foundation, etc. His has published over 200 peer-reviewed papers at top con-ferences and journals, supervised or co-supervised 15-20 Ph.D. dissertations; his work has been widely cited (h-index 60) and over 12,500 citations. Along with his co-authors, he has won the test of time award at the IEEE VIS and a best paper award at ISCRAM, etc. He has been the co-chair of several workshops, including a workshop previously collocated with CIKM. They are listed below: • Program Chair, Big-O(Q){\textquoteright}15: Workshop on Big-Graphs Online Querying in VLDB{\textquoteright}15: the 41st International Conference on Very Large Databases (2015). • Program Chair, WIDM{\textquoteright}12: The 12th International Workshop on Web Information and Data Management in CIKM{\textquoteright}12: the 21st ACM International Conference on Information and Knowledge Management. (2012). • Program Co-Chair, WIDM{\textquoteright}09: The 11th International Work-shop on Web Information and Data Management in CIKM{\textquoteright}09: the 18th ACM International Conference on Information and Knowledge Management. (2009). • Program Co-Chair, SNAKDD{\textquoteright}09: The 2nd International Work-shop on Social Network Mining and Analysis in KDD{\textquoteright}08: the 14th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. (2008).; 32nd ACM International Conference on Information and Knowledge Management, CIKM 2023 ; Conference date: 21-10-2023 Through 25-10-2023",
year = "2023",
month = oct,
day = "21",
doi = "10.1145/3583780.3615311",
language = "English",
series = "International Conference on Information and Knowledge Management, Proceedings",
publisher = "Association for Computing Machinery (ACM)",
pages = "5290--5293",
booktitle = "CIKM 2023",
address = "United States",

}

Download

TY - GEN

T1 - Workshop on Large Language Models' Interpretability and Trustworthiness

T2 - 32nd ACM International Conference on Information and Knowledge Management, CIKM 2023

AU - Saha, Tulika

AU - Saha, Sriparna

AU - Ganguly, Debasis

AU - Mitra, Prasenjit

N1 - Funding information: (4) Prasenjit Mitra is a Professor at The Pennsylvania State Uni-versity and a visiting Professor at the L3S Center at the Leibniz University at Hannover, Germany. He obtained his Ph.D. from Stanford University in 2003 in Electrical Engineering and has been at Penn State since. His research interests are in artifi-cial intelligence, applied machine learning, natural language processing, etc. His research has been supported by the NSF CAREER award, the DoE, DoD, Microsoft Research, Raytheon, Lockheed Martin, Dow Chemicals, McDonnell Foundation, etc. His has published over 200 peer-reviewed papers at top con-ferences and journals, supervised or co-supervised 15-20 Ph.D. dissertations; his work has been widely cited (h-index 60) and over 12,500 citations. Along with his co-authors, he has won the test of time award at the IEEE VIS and a best paper award at ISCRAM, etc. He has been the co-chair of several workshops, including a workshop previously collocated with CIKM. They are listed below: • Program Chair, Big-O(Q)’15: Workshop on Big-Graphs Online Querying in VLDB’15: the 41st International Conference on Very Large Databases (2015). • Program Chair, WIDM’12: The 12th International Workshop on Web Information and Data Management in CIKM’12: the 21st ACM International Conference on Information and Knowledge Management. (2012). • Program Co-Chair, WIDM’09: The 11th International Work-shop on Web Information and Data Management in CIKM’09: the 18th ACM International Conference on Information and Knowledge Management. (2009). • Program Co-Chair, SNAKDD’09: The 2nd International Work-shop on Social Network Mining and Analysis in KDD’08: the 14th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. (2008).

PY - 2023/10/21

Y1 - 2023/10/21

N2 - Large language models (LLMs), when scaled from millions to billions of parameters, have been demonstrated to exhibit the so-called 'emergence' effect, in that they are not only able to produce semantically correct and coherent text, but are also able to adapt themselves surprisingly well with small changes in contexts supplied as inputs (commonly called prompts). Despite producing semantically coherent and potentially relevant text for a given context, LLMs are vulnerable to yield incorrect information. This misinformation generation, or the so-called hallucination problem of an LLM, gets worse when an adversary manipulates the prompts to their own advantage, e.g., generating false propaganda to disrupt communal harmony, generating false information to trap consumers with target consumables etc. Not only does the consumption of an LLM-generated hallucinated content by humans pose societal threats, such misinformation, when used as prompts, may lead to detrimental effects for in-context learning (also known as few-shot prompt learning). With reference to the above-mentioned problems of LLM usage, we argue that it is necessary to foster research on topics related to not only identifying misinformation from LLM-generated content, but also to mitigate the propagation effects of this generated misinformation on downstream predictive tasks thus leading to more robust and effective leveraging in-context learning.

AB - Large language models (LLMs), when scaled from millions to billions of parameters, have been demonstrated to exhibit the so-called 'emergence' effect, in that they are not only able to produce semantically correct and coherent text, but are also able to adapt themselves surprisingly well with small changes in contexts supplied as inputs (commonly called prompts). Despite producing semantically coherent and potentially relevant text for a given context, LLMs are vulnerable to yield incorrect information. This misinformation generation, or the so-called hallucination problem of an LLM, gets worse when an adversary manipulates the prompts to their own advantage, e.g., generating false propaganda to disrupt communal harmony, generating false information to trap consumers with target consumables etc. Not only does the consumption of an LLM-generated hallucinated content by humans pose societal threats, such misinformation, when used as prompts, may lead to detrimental effects for in-context learning (also known as few-shot prompt learning). With reference to the above-mentioned problems of LLM usage, we argue that it is necessary to foster research on topics related to not only identifying misinformation from LLM-generated content, but also to mitigate the propagation effects of this generated misinformation on downstream predictive tasks thus leading to more robust and effective leveraging in-context learning.

KW - Explainability

KW - In-context Learning

KW - Interpretability

KW - Large Language Model

KW - Trustworthiness

UR - http://www.scopus.com/inward/record.url?scp=85178112362&partnerID=8YFLogxK

U2 - 10.1145/3583780.3615311

DO - 10.1145/3583780.3615311

M3 - Conference contribution

AN - SCOPUS:85178112362

T3 - International Conference on Information and Knowledge Management, Proceedings

SP - 5290

EP - 5293

BT - CIKM 2023

PB - Association for Computing Machinery (ACM)

Y2 - 21 October 2023 through 25 October 2023

ER -