AI Alignment people
people doing ai alignment research
Created by
@kabirkumar.bsky.social
@tom4everitt.bsky.social
AGI safety researcher at Google DeepMind, leading causalincentives.com Personal website: tomeveritt.se
@sleepinyourhat.bsky.social
AI safety at Anthropic, on leave from a faculty job at NYU. Views not employers'. I think you should join Giving What We Can. cims.nyu.edu/~sbowman
@peterbhase.bsky.social
AI safety researcher. PhD from UNC Chapel Hill (Google PhD Fellow). Previously: Anthropic, AI2, Google, Meta
@samuelalbanie.bsky.social
@turntrout.bsky.social
Research scientist at Google DeepMind. All opinions are my own. https://turntrout.com
@amuuueller.bsky.social
Postdoc at Northeastern and incoming Asst. Prof. at Boston U. Working on NLP, interpretability, causality. Previously: JHU, Meta, AWS
@sebfar.bsky.social
Senior Research Scientist at Google DeepMind. AGI Alignment researcher. Views my dog's.
@bshlgrs.bsky.social
@norabelrose.bsky.social
AI, philosophy, spirituality Head of interpretability research at EleutherAI, but posts are my own views, not Eleuther’s.
@karimabdel.bsky.social
Intern at CHAI, UC Berkeley | Ex-Research intern at the Krueger AI Safety Lab, University of Cambridge Interested in RL, AI Safety, Cooperative AI, TCS https://karim-abdel.github.io
@ankareuel.bsky.social
Computer Science PhD Student @ Stanford | Geopolitics & Technology Fellow @ Harvard Kennedy School/Belfer | Vice Chair EU AI Code of Practice | Views are my own
@kajsotala.bsky.social
This is a profile. There are many like it, but this one's mine. Blogs: https://kajsotala.fi , https://kajsotala.substack.com/ .
@gsarti.com
PhD Student at @gronlp.bsky.social 🐮, core dev @inseq.org. Interpretability ∩ HCI ∩ #NLProc. gsarti.com
@wordscompute.bsky.social
nlp/ml phding @ usc, interpretability & reasoning & pretraining & emergence 한american, she, iglee.me, likes ??= bookmarks
@florasalim.bsky.social
Professor, CSE, UNSW Sydney. #AI #ML #UbiComp #LLM #MFM #timeseries #ST #multimodal #sensors #continuallearning #trustworthyAI ❤️ #coffee Why am I here? Scouting for a new platform to discover and learn new papers (let’s see if it’s the one)
@stephaniebrandl.bsky.social
Assistant Professor in NLP (Fairness, Interpretability and lately interested in Political Science) at the University of Copenhagen ✨ Before: PostDoc in NLP at Uni of CPH, PhD student in ML at TU Berlin
@mimansaj.bsky.social
Robustness, Data & Annotations, Evaluation & Interpretability in LLMs http://mimansajaiswal.github.io/
@variint.bsky.social
Lost in translation | Interpretability of modular convnets applied to 👁️ and 🛰️🐝 | she/her 🦒💕 variint.github.io
@mdlhx.bsky.social
NLP assistant prof at KU Leuven, PI @lagom-nlp.bsky.social. I like syntax more than most people. Also multilingual NLP, interpretability, mountains and beer. (She/her)
@christophmolnar.bsky.social
Author of Interpretable Machine Learning and other books Newsletter: https://mindfulmodeler.substack.com/ Website: https://christophmolnar.com/
@stellaathena.bsky.social
I make sure that OpenAI et al. aren't the only people who are able to study large scale AI systems.
@romapatel.bsky.social
research scientist @deepmind. language & multi-agent rl & interpretability. phd @BrownUniversity '22 under ellie pavlick (she/her) https://roma-patel.github.io
@soniajoseph.bsky.social
AI researcher at Mila, visiting researcher at Meta Also on X: @soniajoseph_
@matijafranklin.bsky.social
Researching AI Alignment and Manipulation. Conduct CogSci experiments
@aashiqmuhamed.bsky.social
Machine Learning PhD at Carnegie Mellon @mldcmu Ex-Applied Scientist @amazon Search & @AWSAI @Stanford & @LTIatCMU MS and @iitroorkee President's Gold Medalist
@stevebyrnes.bsky.social
Researching Artificial General Intelligence Safety, via thinking about neuroscience and algorithms, at Astera Institute. https://sjbyrnes.com/agi.html
@dhadfieldmenell.bsky.social
Assistant Prof of AI & Decision-Making @MIT EECS I run the Algorithmic Alignment Group (https://algorithmicalignment.csail.mit.edu/) in CSAIL. I work on value (mis)alignment in AI systems. https://people.csail.mit.edu/dhm/
@stanislavfort.bsky.social
AI + security | Stanford PhD in AI & Cambridge physics | techno-optimism + alignment + progress + growth | 🇺🇸🇨🇿
@kartikchandra.bsky.social
I'm a PhD student at MIT CSAIL. More about me: https://cs.stanford.edu/~kach
@xuanalogue.bsky.social
PhD Student. MIT ProbComp / CoCoSci. Inverting Bayesian models of human reasoning and decision-making. Pronouns: 祂/伊
@kabirkumar.bsky.social
I run AI-Plans, an AI Safety lab focused on very precisely evaluating AI Alignment Plans. For several weeks I used a stone for a pillow. I once spent a quarter of my paycheck on cheese. Ping me! DM me! SurpassAI