Session

🇬🇧 Session 4. Large AI Models by and for Europe

19 Sept 2024, 09:45
Emmy-Noether-Saal

Emmy-Noether-Saal

Description


Chair: Dr. Wolfgang Stille (hessian.AI)

Content/Abstract:
This session examines two critical aspects of European AI: sovereignty and safety. First, it highlights the European LLM landscape, featuring an open-source language model initiative and advancements in code language models. Second, it explores approaches to address safety and fairness concerns, as well as model safeguarding techniques.

Agenda:
- Welcome: 5 min
- 4 contributed talks (each 15 min + 5 min discussion)

Presentation materials

There are no materials yet.

  1. Felix Friedrich (Hessian.AI, TU Darmstadt)
    19/09/2024, 09:45
    Session 4. Large AI Models by and for Europe

    In the burgeoning field of artificial intelligence, ensuring the safety and fairness of larg-scale models is paramount. This talk presents Aurora-M, a state-of-the-art multilingual model designed specifically for non-English, e.g. European, languages. Our focus lies in the safety tuning process of Aurora-M, highlighting innovative methodologies to mitigate biases and enhance the model's...

    Go to contribution page
  2. Mert Tiftikci (Hessian.AI, TU Darmstadt)
    19/09/2024, 10:05
    Session 4. Large AI Models by and for Europe

    Although pre-trained language models (PLMs) on code get significantly better, code is largely treated as sequential.
    By ignoring easily extractable structural rules -- through static analysis -- that programming languages and algorithmic concepts follow, significant potential for improvement is lost.
    Some previous work used abstract syntax trees (ASTs) and their extended versions by...

    Go to contribution page
  3. Lukas Helff (Hessian.AI, TU Darmstadt)
    19/09/2024, 10:25
    Session 4. Large AI Models by and for Europe

    We introduce LlavaGuard, a family of VLM-based safeguard models, offering a versatile framework for evaluating the safety compliance of visual content.
    Specifically, we designed LlavaGuard for dataset annotation and generative model safeguarding.
    To this end, we collected and annotated a high-quality visual dataset incorporating a broad safety taxonomy, which we use to tune VLMs on...

    Go to contribution page
  4. Malte Ostendorff (German Research Center for Artificial Intelligence (DFKI), Occiglot)
    19/09/2024, 10:45
    Session 4. Large AI Models by and for Europe

    Large language models (LLMs) have emerged as transformative tools, revolutionizing various natural language processing tasks. Despite their remarkable potential, the LLM landscape is predominantly shaped by US tech companies, leaving Europe with limited access and influence. This talk will present Occiglot - an ongoing research collective for open-source language models for and by Europe. More...

    Go to contribution page
  5. Ali Doosthosseini (GWDG)
    19/09/2024, 11:05
Building timetable...