Description
Chair: Dr. Wolfgang Stille (hessian.AI)
Content/Abstract:
This session examines two critical aspects of European AI: sovereignty and safety. First, it highlights the European LLM landscape, featuring an open-source language model initiative and advancements in code language models. Second, it explores approaches to address safety and fairness concerns, as well as model safeguarding techniques.
Agenda:
- Welcome: 5 min
- 4 contributed talks (each 15 min + 5 min discussion)
In the burgeoning field of artificial intelligence, ensuring the safety and fairness of larg-scale models is paramount. This talk presents Aurora-M, a state-of-the-art multilingual model designed specifically for non-English, e.g. European, languages. Our focus lies in the safety tuning process of Aurora-M, highlighting innovative methodologies to mitigate biases and enhance the model's...
Although pre-trained language models (PLMs) on code get significantly better, code is largely treated as sequential.
By ignoring easily extractable structural rules -- through static analysis -- that programming languages and algorithmic concepts follow, significant potential for improvement is lost.
Some previous work used abstract syntax trees (ASTs) and their extended versions by...
We introduce LlavaGuard, a family of VLM-based safeguard models, offering a versatile framework for evaluating the safety compliance of visual content.
Specifically, we designed LlavaGuard for dataset annotation and generative model safeguarding.
To this end, we collected and annotated a high-quality visual dataset incorporating a broad safety taxonomy, which we use to tune VLMs on...
Large language models (LLMs) have emerged as transformative tools, revolutionizing various natural language processing tasks. Despite their remarkable potential, the LLM landscape is predominantly shaped by US tech companies, leaving Europe with limited access and influence. This talk will present Occiglot - an ongoing research collective for open-source language models for and by Europe. More...