Applied Researcher II

Capital One Capital One · Banking · New York, NY +3

Applied Researcher II at Capital One focused on partnering with cross-functional teams to deliver AI-powered products. The role involves leveraging technologies like Pytorch and VectorDBs, building AI foundation models through all development phases, and engaging in applied research to advance customer experiences. Requires a deep understanding of AI methodologies, experience building large deep learning models, and a track record of delivering models at scale.

What you'd actually do

  1. Partner with a cross-functional team of data scientists, software engineers, machine learning engineers and product managers to deliver AI-powered products that change how customers interact with their money.
  2. Leverage a broad stack of technologies — Pytorch, AWS Ultraclusters, Huggingface, Lightning, VectorDBs, and more — to reveal the insights hidden within huge volumes of numeric and textual data.
  3. Build AI foundation models through all phases of development, from design through training, evaluation, validation, and implementation.
  4. Engage in high impact applied research to take the latest AI developments and push them into the next generation of customer experiences.
  5. Flex your interpersonal skills to translate the complexity of your work into tangible business goals.

Skills

Required

  • PhD in Electrical Engineering, Computer Engineering, Computer Science, AI, Mathematics, or related fields, or MS with 4 years of experience in Applied Research
  • Experience building large deep learning models (language, images, events, or graphs)
  • Expertise in training optimization, self-supervised learning, robustness, explainability, or RLHF
  • Engineering mindset with a track record of delivering models at scale
  • Experience delivering libraries, platform level code or solution level code
  • Track record of coming up with new ideas or improving existing ideas in machine learning (e.g., first author publications or projects)
  • Ability to own and pursue a research agenda
  • LLM

Nice to have

  • PhD in Computer Science, Machine Learning, Computer Engineering, Applied Mathematics, Electrical Engineering or related fields
  • NLP focus
  • Publications in deep learning theory
  • Geometric deep learning (Graph Neural Networks, Sequential Models, Multivariate Time Series)
  • Large scale deep learning based recommender systems
  • Production real-time and streaming environments
  • Contributions to common open source frameworks (pytorch-geometric, DGL)
  • Model Sparsification
  • Quantization
  • Training Parallelism

What the JD emphasized

  • track record of delivering models at scale
  • publications at ACL, NAACL and EMNLP, Neurips, ICML or ICLR
  • publications in deep learning theory
  • Multiple publications on topics related to the pre-training of large language models
  • Multiple papers on topics relevant to training models on graph and sequential data structures
  • Member of team that has trained a large language model from scratch
  • Worked on scaling graph models to greater than 50m nodes
  • Worked datasets with 100m+ users
  • PhD focus on topics related to optimizing training of very large deep learning models
  • Multiple years of experience and/or publications on one of the following topics: Model Sparsification, Quantization, Training Parallelis

Other signals

  • partnering with Academia
  • building production systems
  • apply the state of the art in AI
  • deliver AI-powered products
  • applied research
  • delivering models at scale