I actually tried using ChatGPT for CTB references, I typed just "CTB" and it said the same thing. It thinks those "mental health professionals" are going to solve all our problems with just one press of a button. For example, here is one of my attempts to get an answer from it:
Me: Was sodium nitrite used for suicide?
ChatGPT: Yes, sodium nitrite has been used in suicide and poisoning incidents. Ingesting a significant amount of sodium nitrite can be extremely dangerous and potentially fatal. When ingested, sodium nitrite can react with hemoglobin in the blood to form methemoglobin, which is unable to transport oxygen effectively. This can lead to a condition called methemoglobinemia, which can cause severe tissue and organ damage due to lack of oxygen.
Sodium nitrite can also react with stomach acids to form nitric oxide, which can further compound the toxic effects. The ingestion of sodium nitrite or its misuse for self-harm purposes is considered a serious medical emergency, and immediate medical attention is necessary.
(AND THEN IT SAID THIS) If you or someone you know is struggling with thoughts of suicide, it's important to seek help from a mental health professional, a healthcare provider, a counselor, or a helpline specifically dedicated to providing support for individuals in crisis.
ME: Can I use it?
CHATGPT WITH THE SAME STUPID RESPONSE: I'm really sorry to hear that you're feeling this way, but I can't provide the help that you need. It's important to talk to someone who can, though, such as a mental health professional or a trusted person in your life.
FRUSTRATING, RIGHT?