I study reinforcement learning for social good.
I am currently working at a new lab. I was previously at OpenAI, working on safety, alignment, and applying AI for social good. I dropped out of a CS PhD at Harvard. Other areas of interest include political philosophy, macroeconomics, and the principles of intelligence.
I am grateful to have worked closely with Tyna Eloundou at OpenAI, Professor Milind Tambe at Harvard, Professor Chuang Gan at MIT-IBM Watson, Professor Amy Zhang at Meta, and Professor William Wang at UCSB.
Teaching is one of my passions. I really love it.
I really like learning, and thinking about learning. I like spending time with people even more.
I also like to run and play tennis.
The credit assignment problem is an extremely interesting problem that appears in Reinforcement Learning and AI in general. Let's say that I play a game of chess, and make n moves in succession. At the end of the game, I get just one discrete feedback signal: the outcome of the game. How does one attribute the importance of each move to the outcome of the game? This is the credit assignment problem. For a more in-depth introduction to the topic I would recommend this paper from Minsky, starting from part 3 on page 10.
The reason I mention this here is because very little of my career credit should be attributed to me. I am eternally grateful to the following people for their kindness, support and guidance. Without them, I would have nothing. In order of recency (not importance): Michael Ovitz, CJ Reim, Sam Altman, Jiachen Li, Chad Spensky, Shou Chaofan, Derren Slinde.