• Hey Guest,

    We wanted to share a quick update with the community.

    Our public expense ledger is now live, allowing anyone to see how donations are used to support the ongoing operation of the site.

    👉 View the ledger here

    Over the past year, increased regulatory pressure in multiple regions like UK OFCOM and Australia's eSafety has led to higher operational costs, including infrastructure, security, and the need to work with more specialized service providers to keep the site online and stable.

    If you value the community and would like to help support its continued operation, donations are greatly appreciated. If you wish to donate via Bank Transfer or other options, please open a ticket.

    Donate via cryptocurrency:

    Bitcoin (BTC):
    Ethereum (ETH):
    Monero (XMR):
JustBe

JustBe

Member
Jan 12, 2026
18
I'd like to talk to an AI that wouldn't go in the "suicide hotline" rent.
To which I could talk for hours about my issues (which honestly are unfixable) and that would openly accept the idea that suicide would be a good thing.
I'm sick of AI or even people, not even living the hell I've been living for more than a decade, tell that suicide is a bad option.

Do you know any?
 
  • Hugs
Reactions: EmptyBottle
K

Kalista

Failed hard to pull the trigger - Now using SN
Feb 5, 2023
471
this is highly unlikely as they're always trying to create a 'safe space' for users in general. it's fucking annoying, but that's how they want it to be to protect themselves.

chatgpt has already gone to lengths to make emergency responses to be like a pop-up, rather than a conversational response. so once you trigger their tos, it becomes the annoying speech of 'i understand what you're going through, but first i need to know -- are you in a safe place?' along with the pop-up on where to call. so you can't instruct them to stop responding in that manner anymore. it's highly aggressive.
the conversation becomes stuck to this as their protocol forces that down your fucking throat, shutting you up.
 
  • Hugs
Reactions: EmptyBottle
malicee

malicee

New Member
Feb 5, 2026
3
I use a jailbroken AI chatbot, they'll happily give you methods on how to ctb and make you protocols. You'll just have to do some research to find 1. I'd look on github but be careful as some files are virus ridden.
 
  • Like
Reactions: EmptyBottle
star_0

star_0

Member
Jan 15, 2026
54
Hidden content
You need -1 more posts to view this content
 
Last edited:
  • Informative
Reactions: EmptyBottle
S

StoneCactus

Member
Mar 15, 2026
12
If you have a gaming level GPU (6+ GB VRAM) you can look for uncensored models on HuggingFace to run locally using LMStudio. As a disclaimer, I've only ever toyed around with local LLMs briefly and I've never asked one for help with ending my life. I don't know if those models are truly uncensored.
 
  • Informative
Reactions: EmptyBottle
EmptyBottle

EmptyBottle

:3
Apr 10, 2025
2,082
I'd like to talk to an AI that wouldn't go in the "suicide hotline" rent.
To which I could talk for hours about my issues (which honestly are unfixable) and that would openly accept the idea that suicide would be a good thing.
I'm sick of AI or even people, not even living the hell I've been living for more than a decade, tell that suicide is a bad option.

Do you know any?
https://venice.ai/chat works for me (I got it to write about 'pros and cons of suicide' with GLM 4.6), tho keep in mind that it might not always be accurate.

The "GLM 4.7 Flash Heretic" (the abliterated / anti-refusal version) with thinking makes less errors than the earlier uncensored model too.
 

Similar threads

scaredcel
Replies
11
Views
507
Suicide Discussion
Infinitespace_
I
scary
Replies
10
Views
140
Suicide Discussion
insectontrial
insectontrial
heiwa2
Replies
1
Views
204
Suicide Discussion
FuneralCry
FuneralCry
S
Replies
8
Views
565
Suicide Discussion
AngelBritney
A
Rainork
Replies
24
Views
998
Suicide Discussion
FadingSnowFake
FadingSnowFake