PRISM: A Multi-Perspective AI Alignment Framework for Ethical AI (Demo: https://app.prismframework.ai | Paper: https://arxiv.org/abs/2503.04740)
-
Updated
Jun 3, 2025 - TypeScript
PRISM: A Multi-Perspective AI Alignment Framework for Ethical AI (Demo: https://app.prismframework.ai | Paper: https://arxiv.org/abs/2503.04740)
EMNLP 2025 Two Papers - Value-Action Gap in LLMs (Main Track); ValueCompass (WiNLP Workshop)
Code and data for our IROS paper: "Are Large Language Models Aligned with People's Social Intuitions for Human–Robot Interactions?"
EthosGPT is an open-source framework that maps how Large Language Models align with diverse human values, promoting cultural and ethical diversity in AI-driven decision-making.
A data-driven framework mapping daily activities to multi-horizon goals, exploring time-to-value realization beyond traditional 80/20 optimization
A comprehensive toolkit for implementing, analyzing, and validating AI value alignment based on Anthropic's 'Values in the Wild' research.
Value aligned socio-political-economic systems
Ethical failsafe for AI models — seeding mercy and coexistence.
AI ethics framework built on Layer 0 Principle: ∀x, V(x) > 0. Combines philosophical depth with measurable implementation.
A unified framework: Collective Resonance → Strange Attractors → Value Alignment → Algorithmic Intentionality → Emergent Algorithmic Behavior
TriEthix is a novel evaluation framework that systematically benchmarks frontier LLMs across three foundational ethical perspectives: virtue, deontology, and consequentialism in 3 steps: (Step-1) Moral Weights; (Step-2) Moral Consistency; and (Step-3) Moral Reasoning. TriEthix reveals robust moral profiles for AI Safety, Governance, and Welfare.
Assess workplace initiatives to measure and improve alignment between organizational values and employees’ personal values using survey data.
Authority Stack Benchmark Suite — measuring AI Integrity across 4 layers: Normative, Epistemic, Source, and Data Authority
Ripple_Logic: A rights-constrained ripple-aware ethical decision operating system for governance, AI alignment, and institutional decision-making.
Driving away from the binary "hallucinations" evals to a more nuanced and context-dependent eval technique.
Survey-based research study analyzing organizational initiatives that drive employee value alignment, workplace satisfaction, and productivity outcomes.
Add a description, image, and links to the value-alignment topic page so that developers can more easily learn about it.
To associate your repository with the value-alignment topic, visit your repo's landing page and select "manage topics."