Theory and critique to choreograph accountable intelligence.
Philosophy is the compass of Traceremove. Artur’s research blends critical theory, ethics, and futures thinking to design machine intelligence that honours human dignity, interrogates power, and invites collective stewardship.
Philosophical terrains
Each discipline grounds Traceremove’s research in rigorous theory, ensuring design choices honour cultural nuance, justice, and long-term planetary wellbeing.
Philosophy of technology
Investigates how sociotechnical systems sculpt agency, power, and everyday rituals across cultures.
AI ethics & responsibility
Designs normative frameworks that align machine decision-making with human dignity and collective flourishing.
Epistemology & intelligence
Explores how AI acquires, represents, and justifies knowledge within contested epistemic communities.
Mind, agency, and embodiment
Interrogates consciousness, intentionality, and embodiment to choreograph humane collaborations with AI.
Guiding principles
These philosophical commitments anchor every research sprint, partnership, and product decision at Traceremove.
Human dignity first
Design choices begin with the lived experiences of people most impacted by automation, centring justice and accessibility.
Transparency by default
Every system earns trust through legible documentation, participatory oversight, and reproducible evidence.
Futures thinking
Philosophical foresight maps long-term social consequences so intelligence evolves responsibly.
Selected works
Research across journals, conferences, and books translates philosophical inquiry into actionable playbooks for responsible intelligence.
The ethics of agentic AI systems
A comprehensive examination of moral accountability in autonomous systems and the obligations of designers.
Access paperTechnology, society, and human flourishing
Frameworks for aligning computation with human values through participatory governance and civic imagination.
Access paperEpistemic responsibility in AI development
Investigates duties of care for research teams curating, labelling, and deploying data-intensive systems.
Access paperDigital rights and human dignity
Philosophical foundations for rights-preserving infrastructures in a world of pervasive automation.
Access paperPhilosophical reflections
“Technology is never neutral; it is an argument about how we choose to live together. Our task is to script these arguments for collective thriving.”
“The question is not whether AI can think, but whether we can think responsibly about the agency we grant to machines.”
“Progress in intelligence requires philosophical courage – to interrogate what it means to remain human amid algorithmic futures.”