+
Skip to content
View 7vik's full-sized avatar

Organizations

@bitsacm

Block or report 7vik

Block user

Prevent this user from interacting with your repositories and sending you notifications. Learn more about blocking users.

You must be logged in to block users.

Maximum 250 characters. Please don't include any personal information such as legal names or email addresses. Markdown supported. This note will be visible to only you.
Report abuse

Contact GitHub support about this user’s behavior. Learn more about reporting abuse.

Report abuse
7vik/README.md

Satvik Golechha

Note that this will soon get outdated. To know more about me and my research, a much better place is my website: 7vik.io.

Independent Researcher | AGI Safety • Interpretability • Reinforcement Learning
📍 UC Berkeley (CHAI), MATS Program
📧 zsatvik@gmail.com | 🌐 7vik.io | Scholar | LinkedIn | GitHub


On a quest to understand intelligence and ensure that advanced AGI is safe and beneficial.

🧠 About Me

I’m an independent AI safety researcher currently working with:

  • CHAI, UC Berkeley — Optimal exploration and long-horizon planning in RL.
  • Adrià Garriga-Alonso (FAR AI) — Studying deceptive behavior in frontier AI systems at the MATS Program.
  • Nandi Schoots (Oxford) — Hierarchical representations and modular training for interpretability.

Previously:

  • Microsoft Research — Worked with Neeraj Kayal on representation learning theory, and Amit Sharma and Amit Deshpande on ICL robustness in LLMs.
  • Wadhwani AI — Formulated AI problems in public health and trained robust and interpretable ML large-scale deployments in India.
  • Mentored a SPAR 2025 project on zero-knowledge auditing for undesired behaviors.

🔬 Research Interests

  • AI Alignment & Safety
  • Interpretability & Feature Geometry
  • Long-horizon RL & Planning
  • Representation Learning & Theory

📰 Selected Publications

=Equal contribution; full list at Google Scholar


🧪 Notable Projects

  • AmongUs – Agentic deception sandbox
  • nice-icl – ICL optimization tools
  • grokking – Measuring grokking dynamics
  • byoeb – Healthcare LLM deployment platform


✉️ Get in Touch

  • Email: zsatvik@gmail.com
  • Website: 7vik.io
  • LinkedIn: @7vik
  • Open to collaborations in interpretability, alignment, deception audits, and theoretical ML.

Pinned Loading

  1. mats_2024 mats_2024 Public

    All the code for MATS in Summer 2024.

    Jupyter Notebook

  2. microsoft/nice-icl microsoft/nice-icl Public

    NICE: Normalized Invariance to Choice of Example

    Python 5 2

  3. bitsacm/Slack-Stock-DAG bitsacm/Slack-Stock-DAG Public

    This repository holds a list of cool resources for Silica.

    97 42

  4. erplag-cc erplag-cc Public

    Compiler for the custom language 'ERPLAG' in C.

    C 6 3

  5. AmongUs AmongUs Public

    Make open-weight LLM agents play the game "Among Us", and study how the models learn and express lying and deception in the game.

    Jupyter Notebook 13 1

点击 这是indexloc提供的php浏览器服务,不要输入任何密码和下载