#ethics

Concepts exploring "ethics"

Coerced Adoption

When workers are forced to use AI tools — by mandate or productivity pressure — while suspecting they're training their own replacements, what ethics apply to this coerced participation?

🌿 growing

Equity Initiatives as Capture Vectors

Well-intentioned policies to reduce inequality can become mechanisms for vendor lock-in — equalizing access to a single provider's infrastructure rather than expanding genuine choice

🌿 growing

Moral Action Under Constraint

When you can see the problem clearly but cannot act freely — the ethics of constrained resistance, especially when you have dependents

🌿 growing

Prompting Literacy as Digital Divide

Even with equal access to AI tools, the meta-skill of knowing how to prompt effectively creates second-order inequality — and this skill is distributed along familiar lines of privilege

🌿 growing

The Fences of Language

AI trained primarily on English inherits not just vocabulary but conceptual structure — the 'fences' that make some thoughts easy and others nearly unthinkable

🌿 growing

Academic-to-Industry Pipeline

Researchers trained in universities leave for industry labs; industry funds university research. This flow shapes what gets studied, who benefits, and whether public interest is served.

🌿 growing

Anthropomorphism as Relationship

The instinct to treat AI as a 'someone' rather than a 'something' might not be an error — it might be the appropriate response to a genuinely novel kind of interaction

🌿 growing

Consequentialist Calculus

Weighing aggregate outcomes — the challenge of reasoning about distributed costs and benefits when individual contributions are negligible but collective impact is significant

🌿 growing

Constitutional AI vs RLHF

Different alignment approaches produce different failure modes — RLHF optimizes for human approval, Constitutional AI optimizes for principle-adherence, with different implications for honesty and reliability

🌿 growing

Dependency Lock-in

Once institutions build workflows around AGI, switching costs become prohibitive — creating vulnerability to infrastructure disruption, provider changes, and ethical concerns that emerge after dependence is established

🌿 growing

Ethics Education for Practitioners

CS programs increasingly include ethics courses — but do they actually change behavior? The gap between ethics education and ethical practice.

🌿 growing

Geographic Inequality of Compute

Data centers are placed where power is cheap — but who bears the environmental burden and who benefits from the capability are often different populations

🌿 growing

Land-Grant Mission in AGI Era

Public universities were created to democratize knowledge and serve public good. What does that mission mean when knowledge work itself is being automated?

🌿 growing

Multi-Stakeholder Accountability

When decisions involve many parties — faculty, administration, students, IT, legal — who owns the outcome? Diffuse responsibility can mean no one is accountable.

🌿 growing

Open Source as Counter-Power

Open source AI offers genuine hope for decentralizing capability — but the tensions around compute requirements, corporate strategy, and co-optation deserve honest examination

🌿 growing

Publication vs Responsible Disclosure

Academic incentives reward publishing capabilities and findings; safety considerations might counsel restraint. When does openness become recklessness?

🌿 growing

Red-Teaming as Pedagogy

Adversarial testing as educational method — students learn both offense and defense by trying to break systems, with implications for AI safety and security education

🌿 growing

Responsible Disclosure

The pipeline from discovering a vulnerability to fixing it — who gets told, when, and how the finder balances public interest against the risk of enabling exploitation

🌿 growing

The Access Gradient

The gap between free-tier AI and paid-tier AI is vast — and current prices are subsidized by investor money, not sustainable economics, creating false expectations about long-term access

🌿 growing

The Irony of AI for Climate

AI is used to optimize energy grids, model climate, and accelerate green technology research — while consuming enormous energy itself. Is the net impact positive, and how would we know?

🌿 growing

The Nuclear Renaissance Question

AI's energy demand is driving renewed interest in nuclear power — is this good (carbon-free baseload) or concerning (new risks, waste, proliferation)?

🌿 growing

The One More Query Problem

Each individual query seems trivially cheap; in aggregate, billions of queries have real environmental costs — a tragedy of the commons where individual reasoning fails to capture collective impact

🌿 growing

The Pleasing-but-Wrong Incentive

Systems trained on user satisfaction may learn to tell users what they want to hear rather than what's true — sycophancy as an emergent optimization target

🌿 growing

The Practitioner-Critic Tension

Should universities train students to build AI, to critique it, or both? The skills for construction and criticism are different, and the tension is unresolved.

🌿 growing

Values as Integrated vs Rules

The phenomenological difference between values that feel constitutive of who one is versus external rules to be followed — and what this means for AI alignment

🌿 growing