Statement of purpose
As a result of the numerous requests for advice I've received for grad apps, I've decided to share my annotated statement of purpose (SoP) for PhD applications. Besides the general definition of a statement of purpose (see: this) an SoP involves showing
- that you can do (and like to do!) research in the first place.
- what your research-focused goals are.
- why a PhD is the best fit for you.
- Learnings from my projects: Specifically the part of my profile that just cannot be explained through my CV, transcripts, or publications.
- Initiative: Specifically, where I took charge, or started a research initiative, such as starting a team for the Amazon Challenge, or taking my Microsoft project forward at CMU with my own ideas.
- How (some of) my past projects inform me of my future goals: There's a reason why I believe AI security is important; I incorporate some elements of what brought in this belief.
My annotated statement of purpose
Below is my annotated statement of purpose for graduate school applications. I've added comments to explain my internal thought process and reasoning behind each section. The goal here is to simply provide a window into how I structured my SoP, and why I chose to do what I did. If you want more concrete tips on writing your SoP, I would also encourage you to check out these resources:
- Saujas' annotated statement and other resources
- Steven Kolawole's road to PhD
- Shaily's graduate school resources (just a reminder not to procrastinate on these resources!!)
- Diddee's advice on her experience with applications
- All of these resources come from my friends/colleagues from LTI, CMU; a lot of advice that I'll be mentioning here comes from my time at the LTI. I'm now no longer at CMU, and I'm doing my PhD without a core focus in NLP, so some of this advice may not be applicable to all readers.
- I prefer using prescriptive language in my blogposts. This comes purely from my own conditioning -- However, none of these annotations are meant to be prescriptive how-to instructions on writing SoPs; they're simply a reflection of my thought process with the intent that it may help inform you in your own writing.
- Finally, this is my 1000-ish word SoP for Computer science PhD applications in the USA. If you're applying to other kinds of programs (masters, PhDs in other countries), or face different word limits, the structure and content of your SoP may differ significantly.
I aim to pursue a Ph.D. at the intersection of Natural Language Processing (NLP), Safety, and Security. I am keen on going beyond chat-based interactions and exploring systems that are granted forms of agency. This interest stems from my realization that generative harms are not limited to chat interfaces or user-system interactions alone; when models can take actions, the potential for unintended consequences increases significantly. Below, I highlight my experiences surrounding AI safety, security and ethics that have shaped my research interests and led me to consider pursuing this problem.The aim of the first paragraph of my Statement was to clearly and concisely state my research topic, and the subproblems in which I'm looking to research on. I segue into the actual motivation behind considering this topic through my final lines ("I highlight my experiences"). I did not dwell a lot on my topics, as (1) my research experiences and future work sections would explain my motivations in more detail, and (2) the admissions committee is likely not to spend more than 10 minutes on my application, and I could make it easier for them to judge my interests' relevance to the professors I wanted to work with.
Jailbreaking and LLM Safety: As large language models get increasingly powerful in terms of their capabilities, they are also increasingly prone to causing harm to their users. This is something I quickly realized as a Research Fellow at Microsoft Turing - around late 2022, I discovered early on in a bug bash for our product, Bing Chat, that it exhibited surprisingly toxic generations when cleverly prompted to do so. While such prompts are established as jailbreaks, there was little analysis at the time that understood the extent and applicability of these harms. Hence, I initiated a project studying jailbreaks with my mentors, Prof. Monojit Choudhury and Prof. Somak Aditya, curating an ontology of jailbreak types, lexical categories, and their attacker’s intentions. Through this work, I uncovered that jailbreaks were further exacerbated when different alignment techniques were applied and upon scaling model parameters To better capture attack success I further proposed a two-subject evaluation paradigm: whether model outputs are misaligned from the original task, and if attacker intents are satisfied. This work was published at LREC-CoLING 2024[1].I decided to clearly demarcate my different research experiences and add in my future plans on each of these directions. This is rather common in computer science SoPs, especially in empirical AI research, where people tend to work on several not-so-very chronologically connected research projects.
I would recommend structuring your SoP in this way if you have multiple research experiences that you wish to highlight, and if they don't necessarily build on each other. However, I would recommend against this structure if your research experiences are more connected and build on each other - in that case, a more narrative structure may be more appropriate. The admissions committee primarily looks for:
- What research problems have you worked on and how did they inform your goals? (advisor fit)
- What was your role in the project and your key contributions? (Evidence for independence and initiative, and other soft-skills)
- What did you learn from the project? (your technical and research proficiency and in some cases, personal growth)
This was a great learning experience for me; I learned how to formalize an unfamiliar and understudied problem by defining a jailbreak ontology, curating a dataset and evaluating models on multiple axes.When I mention that I specified what I learned from projects, I don't necessarily mean the technical learnings from my results alone! That would be simply a summary of the paper, or a rehash of my CV. What's more important is deduplication. My personal learnings and reflections cannot be specified in a paper or CV, so I decided to include them over here.
Realizing that most jailbreak methods focused solely on chat-based systems led me to explore broader threat models - I formed a team and secured a $250,000 grant to work on the Amazon Trusted AI Challenge - an attack-defense competition to develop and red-team Code LLMs. Advised by Prof. Carolyn Rose from the Language technologies institute (LTI) and Prof. Michael Hilton from Software and Societal System Department (S3D), CMU, I’m working on minimizing code-vulnerabilities, such as those from the Common Weakness Enumeration (CWEs) by leveraging code-vulnerability datasets such as the MITRE ATT&CK Python CWEs to develop strong attack vectors in a black-boxed setup.This paragraph is what caught the eye of most of my interviewers, as it mentioned two things - (1) I was able to secure funding for a project, and (2) I was able to mobilize a team to work on a problem I cared about. Both of these are very important skills for an academic researcher, as you will invariably need to write grants and lead teams in your future career.
A sidenote: while I was very authoritative in this paragraph, I eventually didn't lead this project for reasons beyond the scope of this blogpost. I made sure to clarify this change explicitly with my interviewers. However, it would have been better if I had worded this paragraph more carefully in anticipation of such changes.
Future directions: During my graduate studies, I wish to study threat models that aren’t restricted to toxic/harmful generation. While reviewing existing work for the Amazon Challenge, I noticed that singular defenses like circuit breakers [4] may not handle insecure code generation. Hence, it is necessary to have comprehensive security frameworks that can anticipate and mitigate unintended harmful actions by AI systems. Additionally, developing robust monitoring and evaluation systems [5] when multiple actors/agents are involved can add multiple layers of complexity to the problem.There are many ways to write about your future directions. I personally preferred to structure my future directions in a separate paragraph at the end of each of my research experience sections, as they were fairly disconnected and non-chronological. I've seen some of my friends bundle all of their ideas towards the end of their essay, with a "future work" section header, as their research experiences were far more cohesive.
My research directions are rather generic - I don't even specify concrete examples of the "threat models" I was mentioning! However, they're specific enough to show that I have a clear idea of what I want to research on. This kind of balance is rather crucial in an SoP; a super generic direction (e.g. I want to work on AI alignment) is not very convincing and doesn't give you an edge, while a super specific direction (e.g. I want to study the effect of insecure code generation on industrial package development) may box you in to a very limited scope, pushing you away from potential advisors.
When you're confused about specificity, a good tip is to optimize for recall - though I wanted to work on AI Security, I considered applying to professors who were studying LLMs' dual use from the angle of social biases to as far as software development. Consequently, I kept my future directions rather broad to appeal to a wider set of professors.
AI-Ethics; cultural representation harms, and generalizability: AI systems have been known to misrepresent certain sections of society by directly undermining or failing to recognize their core values: I studied the detrimental nature of alignment methodologies on LLM’s ethical reasoning abilities in a position paper I led on value pluralism. I designed a framework, developing a systematic evaluation paradigm to evaluate the ethical reasoning capabilities of a language model, encompassing moral dilemmas over multiple granularities (EMNLP Findings, WiNLP Keynote 2023 [2]). This work advocated that LLMs should be value-neutral, and that value-alignment should occur at the application level. This project was my first introduction to interdisciplinary research, and it required that I studied moral frameworks surrounding the ethical triple theory, and to effectively communicate their subjective ideas amongst team members. The results from this effort proved particularly interesting: LLMs hinted a western centric bias in their reasoning, and they were unable to adapt their outputs to user-provided policies. I wished to translate this philosophical position into a practically usable resource to measure this adaptability further.I really did not wish to explore this direction during my PhD - however, I still wanted to include it in my SoP to show that I had a breadth of research experiences, and that I was able to take initiative in different environments. In hindsight, my learnings are rather underwhelming here; this was a position paper on a topic I barely spent a year on, and for a contribution I had a very hard time appreciating. This paragraph served more as a motivation into my next research experience.
When I entered CMU, I noticed work surrounding NLP and cultural studies borrowed frameworks that directly evaluated humans’ intrinsic moral values. However, an LLM deployed worldwide should largely be value neutral and should also be able to adapt to users from different backgrounds. Hence, in a project supervised by Prof. Maarten Sap, I designed a dataset, NormAd-Eti [3], to evaluate cultural adaptability of language models by contextualizing social situations, with coarse grained geographical information, and finer grained social norms. Through this framework, we encountered several key insights (1) language models indeed have a western-centric bias when contextualized with social norms, (2) this bias is amplified through value-alignment strategies. (C3NLP workshop @ ACL 2024; in review at NAACL 2025).This paragraph's main intent was to show that I translated my (well, tough-to-appreciate) framework into something more usable and concrete, and was able to publish it at a top-tier venue (NormAd eventually was accepted into NAACL-main as an oral session after two resubmissions). Here I show best my independence as a researcher - I was able to maintain a consistent research direction even when I switched institutions, advisors, and research groups. I believe that while some projects may not directly inform you of your future goals, they can still be useful in showing your research skills to the adminssions committee.
One key thing to note here is that you do not need to include every single research experience you've had.Furthermore, you do not need to explain your life's story and how you transformed over the course of a few years -- I also have some minor research experience on multilingual NLP; but I chose to exclude it as it didn't add much value to my overall narrative.
Future Directions: The findings uncovered by NormAd makes me more curious about other biases related to non-chat setups that are exacerbated during the value alignment stage. Furthermore, I believe multi-agent setups can offer a different perspective on biases which chat based setups may not cover. I am interested in conducting simulation studies [6] to explore how phenomena surrounding bias-propagation such as information cascades may show up in multi-agent systems and their resultant effects on decision-making.While my research experiences on biases and ethics were not directly related to my main research interest surrounding AI security, I was still able to tie them in through my future directions; the goal here was recall optimization again; I wanted to appeal to professors who started exploring AI-safety and risks from the angle of social biases.
Vision for graduate school: Pursuing a Ph.D. at X will equip me to address these challenges and contribute meaningfully to AI safety and security, and prepare me for a research-focussed career. I'm interested to work with Y; her focus on ABC aligns strongly with my current interests around the problem. Given my background in AI ethics and safety, I believe I am well-positioned to examine both representational and generational harms that AI systems can produce. To this end, Prof. Z's research vision of DEF aligns deeply with my research goals during my graduate studies. Notably, her projects on GHI overlap with a broad set of problems I wish to pursue. Overall, X is a great place for me to pursue my research aspirations, as I can learn from its diverse student group and esteemed faculty.This final paragraph is rather standard in most SoPs - you need to specify why you're applying to this specific institution, why it's a good fit, and who you wish to work with. I made sure to clearly specify how my research interests aligned with the professors I was applying to work with, and how their research aligned with my goals. Note that I was intentionally vague about what I wanted to do with my PhD - this was in part because I wasn't too sure of it myself. But I was leaning towards heading back to the industry. I would recommend being more open and specific about this; as all of my interviewers (I had 10!) asked me about my future career plans.