Logo des Repositoriums
 

Evaluating Dangerous Capabilities of Large Language Models: An Examination of Situational Awareness

dc.contributor.authorYadav, Dipendra
dc.contributor.editorStolzenburg, Frieder
dc.date.accessioned2023-09-20T04:20:43Z
dc.date.available2023-09-20T04:20:43Z
dc.date.issued2023
dc.description.abstractThe focal point of this research proposal pertains to a thorough examination of the inherent risks and potential challenges associated with the use of Large Language Models (LLMs). Emphasis has been laid on the facet of situational awareness, an attribute signifying a model’s understanding of its environment, its own state, and the implications of its actions. The proposed research aims to design a robust and reliable metric system and a methodology to gauge situational awareness, followed by an in-depth analysis of major LLMs using this developed metric. The intention is to pinpoint any latent hazards and suggest effective strategies to mitigate these issues, with the ultimate goal of promoting the responsible and secure advancement of artificial intelligence technologies.en
dc.identifier.doi10.18420/ki2023-dc-10
dc.identifier.urihttps://dl.gi.de/handle/20.500.12116/42398
dc.language.isoen
dc.pubPlaceBonn
dc.publisherGesellschaft für Informatik e.V.
dc.relation.ispartofDC@KI2023: Proceedings of Doctoral Consortium at KI 2023
dc.titleEvaluating Dangerous Capabilities of Large Language Models: An Examination of Situational Awarenessen
dc.typeText
gi.citation.endPage93
gi.citation.startPage88
gi.conference.date45195
gi.conference.locationBerlin
gi.conference.sessiontitleDoctoral Consortium at KI 2023
gi.document.qualitydigidoc

Dateien

Originalbündel
1 - 1 von 1
Vorschaubild nicht verfügbar
Name:
D1-paper.pdf
Größe:
122.9 KB
Format:
Adobe Portable Document Format