エピソード

  • A Society of AI Agents
    2024/10/10

    In this podcast, the hosts discuss a research paper that explores how large language models (LLMs), like the ones used in chatbots, behave when placed in a simulated prison scenario. The researchers built a custom tool, zAImbardo, to simulate interactions between a guard and a prisoner, focusing on two key behaviors: persuasion, where the prisoner tries to convince the guard to allow extra privileges (like more yard time or an escape), and anti-social behavior, such as being toxic or violent. The study found that while some LLMs struggle to stay in character or hold meaningful conversations, others show distinct patterns of persuasion and anti-social actions. It also reveals that the personality of the guard (another LLM) can greatly influence whether the prisoner succeeds in persuading them or if harmful behaviors emerge, pointing to the potential dangers of LLMs in power-based interactions without human oversight.


    Original paper:

    Campedelli, G. M., Penzo, N., Stefan, M., Dessì, R., Guerini, M., Lepri, B., & Staiano, J. (2024). I want to break free! Anti-social behavior and persuasion ability of LLMs in multi-agent settings with social hierarchy. arXiv. https://arxiv.org/abs/2410.07109

    続きを読む 一部表示
    10 分
  • Making AI safer
    2024/10/09

    This podcast episode discusses a research paper focused on making large language models (LLMs) safer and better aligned with human values. The authors introduce a new technique called DATA ADVISOR, which helps LLMs create safer and more reliable data by following guiding principles. DATA ADVISOR works by continuously reviewing the data generated by the model, spotting gaps or issues, and suggesting improvements for the next round of data creation. The study shows that this method makes LLMs safer without reducing their overall effectiveness, and it performs better than other current approaches for generating safer data.


    Original paper:

    Wang, F., Mehrabi, N., Goyal, P., Gupta, R., Chang, K.-W., & Galstyan, A. (2024). Data Advisor: Dynamic data curation for safety alignment of large language models. arXiv. https://arxiv.org/abs/2410.05269

    続きを読む 一部表示
    10 分
  • Altering Body Shape with AI
    2024/10/07

    In this episode, we dive into the fascinating world of BodyShapeGPT, a breakthrough approach that turns simple text into realistic 3D human avatars. By using the power of LLaMA-3, a fine-tuned large language model, researchers have developed a system that can accurately shape a virtual body based on descriptive language. We'll explore how a unique dataset and custom algorithms make it all possible, and why this could revolutionize everything from gaming to virtual reality. Tune in to discover how technology is bridging the gap between words and virtual worlds!


    Original paper:

    Árbol, B. R., & Casas, D. (2024). BodyShapeGPT: SMPL Body Shape Manipulation with LLMs. https://arxiv.org/abs/2410.03556

    続きを読む 一部表示
    8 分
  • Making AI Understand Dialects
    2024/10/06

    In this episode, we're diving into the world of languages and computers! Have you ever wondered how people who speak in different ways or dialects can still communicate with the help of technology? Today, we’ll learn about a cool new method scientists are using to help computers understand Finnish dialects and turn them into standard Finnish! We’ll explore how this helps computers do a better job at things like reading and talking to us. Join us as we talk about how this amazing technology works and how it can even help people who speak other languages! Original paper

    Partanen, N., Hämäläinen, M., & Alnajjar, K. (2019). Dialect text normalization to normative standard Finnish. In Workshop on Noisy User-generated Text (pp. 141-146). The Association for Computational Linguistics. https://aclanthology.org/D19-5519/

    続きを読む 一部表示
    8 分
  • How to ensure responsible AI?
    2024/10/05

    Join us on an exciting journey into the world of Artificial Intelligence (AI)! In this episode, we explore how countries around the world are creating smart strategies to use AI in the best way possible. You'll learn about the EPIC framework, which focuses on four important things—Education, Partnership, Infrastructure, and Community—to make sure AI helps everyone. Whether you're curious about robots, computers, or just how technology works, this fun and easy-to-understand episode will show you how we can use AI to make the world a better place for all! Original paper:

    Tjondronegoro, D. W. (2024). Strategic AI Governance: Insights from Leading Nations. https://arxiv.org/abs/2410.01819

    続きを読む 一部表示
    9 分
  • Can AI Give Legal Counsel?
    2024/10/04

    In this episode, we dive into groundbreaking research on how large language models (LLMs) handle complex legal reasoning. We discuss the challenges LLMs face when distinguishing between similar legal charges and explore a new framework called MALR, which uses a multi-agent approach and non-parametric learning to enhance AI's understanding of legal concepts. Tune in to learn how this innovative approach improves AI's performance, even surpassing human capabilities in some legal reasoning tasks.


    Original paper:

    Yuan, W., Cao, J., Jiang, Z., Kang, Y., Lin, J., Song, K., tianqianjin lin, Yan, P., Sun, C., & Liu, X. (2024). Can Large Language Models Grasp Legal Theories? Enhance Legal Reasoning with Insights from Multi-Agent Collaboration. https://arxiv.org/abs/2410.02507


    続きを読む 一部表示
    9 分
  • What is One-Shot Style Adaptation?
    2024/10/02

    In this episode, we dive into a groundbreaking method called One-Shot Style Adaptation (OSSA), designed to tackle a common challenge in deep learning: performance drop-offs when models face different environments than they were trained for. Unlike traditional approaches that need large amounts of data, OSSA requires only a single image to adjust the model, making it highly efficient. From weather changes to synthetic-to-real scenarios, OSSA shows promise in real-world applications with limited data. Join us as we explore this innovative and practical solution for object detection! Original paper:

    Gerster, R., Caesar, H., Rapp, M., Wolpert, A., & Teutsch, M. (2024). OSSA: Unsupervised One-Shot Style Adaptation. https://arxiv.org/abs/2410.00900

    続きを読む 一部表示
    11 分