<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Our Team on AI Safety Türkiye</title><link>https://aisafetyturkiye.org/en/team/</link><description>Recent content in Our Team on AI Safety Türkiye</description><generator>Hugo</generator><language>tr</language><copyright>Copyright (c) 2024-2026 AI Safety Türkiye</copyright><lastBuildDate>Thu, 03 Apr 2025 00:00:00 +0000</lastBuildDate><atom:link href="https://aisafetyturkiye.org/en/team/index.xml" rel="self" type="application/rss+xml"/><item><title>Bengüsu Özcan</title><link>https://aisafetyturkiye.org/en/team/bengusu-ozcan/</link><pubDate>Sat, 12 Aug 2023 00:00:00 +0000</pubDate><guid>https://aisafetyturkiye.org/en/team/bengusu-ozcan/</guid><description>&lt;p&gt;Bengüsu studied engineering and psychology at Sabancı University and holds a master&amp;rsquo;s degree in social data science from Columbia University. She is the co-founder and director of AI Safety Turkey. She conducts research on advanced AI governance and policy at the Center for Future Generations, a Brussels-based research institute. Her areas of interest include international coordination in advanced AI governance, AI safety standards, and semiconductor industry-based security policies.&lt;/p&gt;</description></item><item><title>Berke Çelik</title><link>https://aisafetyturkiye.org/en/team/berke-celik/</link><pubDate>Sat, 12 Aug 2023 00:00:00 +0000</pubDate><guid>https://aisafetyturkiye.org/en/team/berke-celik/</guid><description>&lt;p&gt;Berke graduated from the Philosophy Department at Boğaziçi University and is the co-founder and director of AI Safety Turkey. He serves as director of the development program at Global Policy Research Group. His other areas of interest include AI policy in developing countries and decision theory.&lt;/p&gt;</description></item><item><title>Sayhan Yalvaçer</title><link>https://aisafetyturkiye.org/en/team/sayhan-yalvacer/</link><pubDate>Sat, 12 Aug 2023 00:00:00 +0000</pubDate><guid>https://aisafetyturkiye.org/en/team/sayhan-yalvacer/</guid><description>&lt;p&gt;Sayhan, an alumnus of the Department of Philosophy at Boğaziçi University, counts the alignment problem, autoregressive transformers, the rationalist community (LW), and secular demographic trends among his principal areas of interest.&lt;/p&gt;</description></item><item><title>Alparslan Bayrak</title><link>https://aisafetyturkiye.org/en/team/alparslan-bayrak/</link><pubDate>Sat, 12 Aug 2023 00:00:00 +0000</pubDate><guid>https://aisafetyturkiye.org/en/team/alparslan-bayrak/</guid><description>&lt;p&gt;Alparslan is a senior philosophy student at Bilkent University. His main areas of interest are global priorities and suffering risks (s-risks) research. He focuses on conflict scenarios that may arise from the development and deployment of advanced AI systems, and the philosophical aspects of cooperative AI.&lt;/p&gt;</description></item><item><title>İrem Sena Karakoç</title><link>https://aisafetyturkiye.org/en/team/irem-sena-karakoc/</link><pubDate>Thu, 03 Apr 2025 00:00:00 +0000</pubDate><guid>https://aisafetyturkiye.org/en/team/irem-sena-karakoc/</guid><description>&lt;p&gt;İrem graduated from Bilkent University with a degree in Philosophy. Her primary research interests include AI-driven systemic risks, the economic impacts of transformative AI, and AI governance strategies in developing nations.&lt;/p&gt;</description></item></channel></rss>