Skip to content
#

ai-ethics

Here are 249 public repositories matching this topic...

A long-form essay exploring the philosophy of minimalist AI, how future intelligent systems can be calm, ethical, and invisible. Inspired by calm technology, design minimalism, and cognitive science, Quiet Machines envisions a world where the best technology listens more than it speaks.

  • Updated Nov 2, 2025

An in-depth exploration of the rise of human-centered, interactive machine learning. This article examines how Streamlit enables collaborative AI design by merging UX, visualization, and automation. Includes theory, architecture, and design insights from the ML Playground project.

  • Updated Nov 3, 2025

A narrative and technical exploration of data authenticity through the four pillars of synthetic data realism, Fidelity, Coverage, Privacy, and Utility. This thought-leadership piece combines storytelling, mathematics, and code to explain how these metrics define the ethical and functional “soul” of data in AI systems.

  • Updated Nov 14, 2025

A long-form article and practical framework for designing machine learning systems that warn instead of decide. Covers regimes vs decimals, levers over labels, reversible alerts, anti-coercion UI patterns, auditability, and the “Warning Card” template, so ML preserves human agency while staying useful under uncertainty.

  • Updated Dec 20, 2025

Trustworthy AI: From Theory to Practice book. Explore the intersection of ethics and technology with 'Trustworthy AI: From Theory to Practice.' This comprehensive guide delves into creating AI models that prioritize privacy, security, and robustness. Featuring practical examples in Python, it covers uncertainty quantification, adversarial ML

  • Updated Feb 23, 2024
  • Jupyter Notebook

An analytical essay on why prediction-based models fail in reflexive, unstable systems. This article argues that accuracy collapses when models influence behavior, and proposes equilibrium and force-based modeling as a more robust framework for understanding pressure, instability, and transitions in AI-shaped systems.

  • Updated Dec 13, 2025

A long-form article introducing the Twin Test: a practical standard for high-stakes machine learning where models must show nearest “twin” examples, neighborhood tightness, mixed-vs-homogeneous evidence, and “no reliable twins” abstention. Argues similarity and evidence packets beat probability scores for trust and safety.

  • Updated Dec 26, 2025

Improve this page

Add a description, image, and links to the ai-ethics topic page so that developers can more easily learn about it.

Curate this topic

Add this topic to your repo

To associate your repository with the ai-ethics topic, visit your repo's landing page and select "manage topics."

Learn more