Currently:
- ๐งช Researching compute-efficient model architecture at OpenAI's Training team
- ๐ฌ Understanding how vision-language models build syntactic representations
- โคด Scaling up neural satisfiability solvers
- ๐ Exploring how scaling laws scale with data complexity
- ๐พ Fine-tuning LLM agents to play games with online RL
- โฉ๏ธ Replacing backprop in autoregressive language models
Previously:
- ๐ Engineer #1 @ Reworkd working on multimodal code-generation for web data extraction
- ๐ Graduated from Carnegie Mellon '23 with Honors in Computer Science
- ๐ My thesis on vision-language semantics is cited by Google Brain, Meta AI, Stanford, etc.
- ๐ Published papers at ACL, ICLR, EMNLP, & EACL conferences and NeurIPS & ICCV workshops
- ๐งโ๐ป Exited a content research micro-SaaS with some cool clustering, fact checking, & generation features
- ๐ค Fine-tuned language models at Microsoft AI over summer '22
- ๐ ๏ธ Worked on information retrieval, question answering, & summarization at various startups '20-21
- ๐ง Developed brain-computer interfaces with NSF funding and placed 1st nationally at NeuroTechX '20
- ๐ Won 10+ hackathons including 1st @ Facebook '19, 2nd @ UCLA '19, 3rd @ MIT '20
Warning: has not learnt the Bitter Lesson. Prone to getting nerd-sniped by linguistically & cognitively motivated AI research directions.