Could AI systems introspect on their own (potential) consciousness?
Is longtermism helpful?
A new argument for the simulation hypothesis
Thorstad's case against the singularity hypothesis
Can we understand AI as a rational agent?