Parents of teenage ChatGPT users will soon be able to receive a notification if the platform thinks their child is in "acute distress".

It is among a number of parental controls announced by the chatbot's maker, OpenAI.

Its safety for young users was put in the spotlight last week when a couple in California sued OpenAI over the death of their 16-year-old son, alleging ChatGPT encouraged him to take his own life.

OpenAI said it would introduce what it called "strengthened protections for teens" within the next month.

When news of the lawsuit emerged last week, OpenAI published a note on its website stating ChatGPT is trained to direct people to seek professional help when they are in trouble, such as the Samaritans in the UK.