Skip to content
View diksha-shrivastava13's full-sized avatar

Organizations

@Entangled-Causality

Block or report diksha-shrivastava13

Block user

Prevent this user from interacting with your repositories and sending you notifications. Learn more about blocking users.

You must be logged in to block users.

Maximum 250 characters. Please don’t include any personal information such as legal names or email addresses. Markdown is supported. This note will only be visible to you.
Report abuse

Contact GitHub support about this user’s behavior. Learn more about reporting abuse.

Report abuse

Diksha Shrivastava

Diksha is an AI Safety Researcher based in India, who has spent the last eight months working full-time on safety research at a capability-first lab and is now moving to independent work. Her research sits at the intersection of dynamic agency, multi-agent risks, developmental interpretability and scalable oversight — she’s building upon Causal Incentives research to study how co-evolving environments shape temporal goal structures in RL agents using regret-based Unsupervised Environment Design. She’s particularly interested in what it means for an agent to model its own training process, and what that implies for oversight. Alongside her research, she volunteers in reading groups and mentors people new to AI Safety. She’s always glad to talk about risks from open-endedness, agent epistemics, or alignment as an environment design problem.

diksha-shrivastava13.github.io

Note: I’m not active on social media — the best way to reach me is by email.

Pinned Loading

  1. goal-composition goal-composition Public

    Somewhere between shard theory, open-endedness and causal modeling of agency

    Python

  2. AI-Makerspace/AI-Makerspace AI-Makerspace/AI-Makerspace Public

    AI Makerspace: Blueprints for developing AI applications with cutting-edge technologies.

    JavaScript 74 38