Tokyo AI Safety Conference 2025
TAIS 2025 is a one-day workshop on AI Safety, to be held on Saturday, April 12th, 2025 from 12:00 - 17:00 at Tokyo Midtown Tower.
For our 2025 conference, we hope to welcome back all the attendees who joined us in 2024, plus any other domestic and international researchers and professionals interested in discussing AI Safety.
We welcome attendees from all backgrounds, regardless of your prior research experience. This event is free, so we welcome you to come and join us!
Last year, we ran TAIS 2024: a small non-archival open academic conference structured as a lecture series. It ran over the course of 2 days from April 5th–6th 2024 at the International Conference Hall of the Plaza Heisei in Odaiba, Tokyo.
Recordings from the conference are hosted on our youtube channel.
Call for Papers
We are inviting submissions of short papers (maximum 8 pages) outlining new research, with a deadline of January 1, 2025. We welcome papers on any of the following topics, or anything else where the authors convincingly argue that it advances the field of AI safety in Japan.
Mechanistic Interpretability: the detailed study of particular neural networks. Interpretability researchers often take a circuits-style approach following the foundational work of Chris Olah, or use causal linear probes to understand directions in latent space. Work that investigates how and why structure emerges in neural networks during training can also fall under this category.
Developmental Interpretability: understanding the process by which neural networks learn. By applying the tools of Watanabe's singular learning theory to modern neural network architectures, developmental interpretability researchers aim to detect phase transitions during training, thereby detecting new capabilities as they emerge.
Agent Foundations: the name given to a number of parallel research efforts, each of which aims to designing AI that is provably safe. Problems in agent foundations include embedded agency and cognition, corrigibility and the shutdown problem, natural abstractions, and game theory. With sufficient rigour, agent foundations research hopes to build AGI that is safe by construction.
Scalable Oversight: a prosaic alignment approach that aims to make advanced AI systems amenable to human oversight. Often this involves training less powerful AIs to oversee more powerful AIs in a hierarchical way with a human at the root, as in Paul Christiano's Iterated Amplification. Other scalable oversight approaches aim to make even the largest AIs directly overseeable.
ALIFE: a broad theme grouping approaches that understand artificial life as it relates to natural life. Often, as in active inference or collective intelligence, this involves replicating natural systems in silico such that we can understand and improve on them. Other ALIFE approaches overlap with the agent foundations research agenda, as in the study of emergent agentic phenomena in cellular automata.
Artificial Consciousness and Posthuman Ethics: thinking seriously about whether machines can be conscious or worthy of moral patienthood, including determining whether current systems are conscious or the intentional construction of conscious artificial systems. Works that address the ethical treatment of machines by humans and humans by machines also fall into this category.
AI Governance: research into how frontier AI systems can be effectively deployed to benefit Japan and the rest of the world and what regulation is necessary or sufficient to ensure the safety of the public, from both regular and existential risks.
Reviewing and Submission Policy
Please submit your paper using the form on the submission page.
Submissions are non-archival. We are happy to receive submissions that are also undergoing peer review elsewhere at the time of submission, but we will not accept submissions that have already been previously published or accepted for publication at peer-reviewed conferences or journals. Submission is permitted for papers presented or to be presented at other non-archival venues.
Most accepted papers will be offered a poster presentation. In rare cases, we might ask the authors to lead a workshop (1-2 hours). If you would like to be considered for a workshop, please indicate as such on your submission form.
TAIS 2025 is not peer-reviewed. Our team of editors will review each submission and decide whether to accept a paper as a poster, invite the authors to lead a workshop, or reject it. Evaluation of submissions will be based on originality and novelty, technical strength, and relevance to the workshop topics. Notifications of acceptance will be sent to applicants by email.
Important Dates
Submissions open: October 11, 2024
Submission Deadline: January 1, 2025
Notification of Acceptance: February 1, 2025
Camera-ready Deadline: April 1, 2025
Workshop Date: Saturday, April 12, 2025
All deadlines are 11:59PM UTC-12:00 (“anywhere on Earth”).
We encourage you to consider submitting to TAIS 2025 even if you think your work might not be relevant; irrelevant submissions are not a substantial burden on our editorial team, and our editors like reading oblique and hard-to-categorize research.
We also welcome position pieces bringing clarification to complex topics (especially topics not prominent in the Japanese research community), discussing the need for research in overlooked areas, or pointing out structural issues that present obstacles to those who would otherwise do good research.
Schedule of Events & Workshops
Time Event
12:00 - 12:30 Welcome + Talk 1 + Sponsors
Workshops
12:30 - 14:15 Workshop 1
14:15 - 15:00 Networking Break
15:00 - 16:45 Workshop 2
16:45 - 17:00 Farewells, Breakaway for Evening Social (TBD)
Concurrent Poster Sessions
12:30 - 16:45 Poster Sessions
TAIS 2025 is currently courting industry sponsors.
-
Company Name and Logo Displayed on TAIS Website
Acknowledgement in Opening Ceremony
Distribution of your Company Flyer among participants
-
Company Name and Logo Displayed on TAIS Website
5-minute Company Presentation to Attendees
Optional presentation or demonstration (limited to 10 minutes) during a workshop
-
Company Name and Logo Displayed on TAIS Website
5-minute Company Presentation to Attendees
Lead a workshop on a topic relevant to the conference
Access to Attendee List
If you are interested in becoming a sponsor, please contact someone@aisafety.tokyo.
Interested in attending TAIS 2025?
Please sign up on luma.