Skip to content
View pps121's full-sized avatar
🎯
Focusing
🎯
Focusing

Block or report pps121

Block user

Prevent this user from interacting with your repositories and sending you notifications. Learn more about blocking users.

You must be logged in to block users.

Maximum 250 characters. Please don’t include any personal information such as legal names or email addresses. Markdown is supported. This note will only be visible to you.
Report abuse

Contact GitHub support about this user’s behavior. Learn more about reporting abuse.

Report abuse
pps121/README.md

Email GitHub Website


🧭 Research Vision

Can the geometry of thought reveal how alignment works?

I develop differential-geometric frameworks for understanding how large language models encode, transform, and ultimately suppress beliefs β€” with a focus on mechanistic interpretability and AI alignment. The Torsional Belief Vector Field (TBVF), models transformer hidden states as discrete curves on a Riemannian belief manifold equipped with a Cartan torsion connection. This reveals, for the first time, where and how DPO/RLHF alignment geometrically reshapes model internals β€” creating what I call "brake layers": localized, geometrically distinct suppression mechanisms.

I am actively seeking fully-funded PhD positions at world-class research universities in mechanistic interpretability, geometric deep learning, and AI alignment.


πŸ”¬ Torsional Belief Vector Fields

Torsional Belief Vector Field treats each transformer layer's hidden state as a point on a high-dimensional Riemannian manifold with Fisher-Rao metric. The torsion tensor β€” antisymmetric component of cross-layer covariance β€” measures rotational mismatch between consecutive belief updates.

Layer 27 (Peak Brake Layer):
━━━━━━━━━━━━━━━━━━━━━━━━━━━
SFT torsion norm:  β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ 0.66
DPO torsion norm:  β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ       0.42
Cohen's d:         0.741 ***
p-value:           7.7 Γ— 10⁻¹³
━━━━━━━━━━━━━━━━━━━━━━━━━━━━
"Brake layers are geometry."

Pinned Loading

  1. 100-Days-Of-ML-Code 100-Days-Of-ML-Code Public

    Forked from Avik-Jain/100-Days-Of-ML-Code

    100 Days of ML Coding

    Jupyter Notebook 1 1

  2. ACL2022_KnowledgeNLP_Tutorial ACL2022_KnowledgeNLP_Tutorial Public

    Forked from zcgzcgzcg1/ACL2022_KnowledgeNLP_Tutorial

    Materials for ACL-2022 tutorial: Knowledge-Augmented Methods for Natural Language Processing

    1

  3. annotated_research_papers annotated_research_papers Public

    Forked from AakashKumarNain/annotated_research_papers

    This repo contains annotated research papers that I found really good and useful

    1

  4. awesome-nlp awesome-nlp Public

    Forked from abi-aryan/awesome-nlp

    πŸ“– A curated list of resources dedicated to Natural Language Processing (NLP)

    1

  5. awesome-notebooks awesome-notebooks Public

    Forked from jupyter-naas/awesome-notebooks

    Ready to use data science templates, organized by tools to jumpstart your projects and data products in minutes. 😎 published by the Naas community.

    Jupyter Notebook 1

  6. google-research google-research Public

    Forked from google-research/google-research

    Google Research

    Jupyter Notebook