—Jessica Hamzelou
This week, I’ve been engaged on a chunk about an AI-based device that might assist information end-of-life care. We’re speaking concerning the sorts of life-and-death selections that come up for very unwell folks.
Usually, the affected person isn’t capable of make these selections—as an alternative, the duty falls to a surrogate. It may be a particularly troublesome and distressing expertise.
A bunch of ethicists have an concept for an AI device that they consider might assist make issues simpler. The device can be educated on details about the individual, drawn from issues like emails, social media exercise, and looking historical past. And it might predict, from these elements, what the affected person would possibly select. The group describe the device, which has not but been constructed, as a “digital psychological twin.”
There are many questions that have to be answered earlier than we introduce something like this into hospitals or care settings. We don’t know the way correct it might be, or how we are able to guarantee it received’t be misused. However maybe the largest query is: Would anybody wish to use it? Learn the total story.
This story first appeared in The Checkup, our weekly publication providing you with the within observe on all issues well being and biotech. Join to obtain it in your inbox each Thursday.
In case you’re excited by AI and human mortality, why not try:
+ The messy morality of letting AI make life-and-death selections. Automation might help us make laborious decisions, however it will possibly’t do it alone. Learn the total story.
+ …however AI methods mirror the people who construct them, and they’re riddled with biases. So we must always rigorously query how a lot decision-making we actually wish to flip over to.
—Jessica Hamzelou
This week, I’ve been engaged on a chunk about an AI-based device that might assist information end-of-life care. We’re speaking concerning the sorts of life-and-death selections that come up for very unwell folks.
Usually, the affected person isn’t capable of make these selections—as an alternative, the duty falls to a surrogate. It may be a particularly troublesome and distressing expertise.
A bunch of ethicists have an concept for an AI device that they consider might assist make issues simpler. The device can be educated on details about the individual, drawn from issues like emails, social media exercise, and looking historical past. And it might predict, from these elements, what the affected person would possibly select. The group describe the device, which has not but been constructed, as a “digital psychological twin.”
There are many questions that have to be answered earlier than we introduce something like this into hospitals or care settings. We don’t know the way correct it might be, or how we are able to guarantee it received’t be misused. However maybe the largest query is: Would anybody wish to use it? Learn the total story.
This story first appeared in The Checkup, our weekly publication providing you with the within observe on all issues well being and biotech. Join to obtain it in your inbox each Thursday.
In case you’re excited by AI and human mortality, why not try:
+ The messy morality of letting AI make life-and-death selections. Automation might help us make laborious decisions, however it will possibly’t do it alone. Learn the total story.
+ …however AI methods mirror the people who construct them, and they’re riddled with biases. So we must always rigorously query how a lot decision-making we actually wish to flip over to.