Join the conversation

Join the community of Machine Learners and AI enthusiasts.

Sign Up
giadap 
posted an update 6 days ago
Post
10565
One of the hardest challenges in AI safety is finding the right balance: how do we protect people from harm without undermining their agency? This tension is especially visible in conversational systems, where safeguards can sometimes feel more paternalistic than supportive.

In my latest piece for Hugging Face, I argue that open source and community-driven approaches offer a promising (though not exclusive) way forward.

✨ Transparency can make safety mechanisms into learning opportunities.
✨ Collaboration with diverse communities makes safeguards more relevant across contexts.
✨ Iteration in the open lets protections evolve rather than freeze into rigid, one-size-fits-all rules.

Of course, this isn’t a silver bullet. Top-down safety measures will still be necessary in some cases. But if we only rely on corporate control, we risk building systems that are safe at the expense of trust and autonomy.

Read the blog post here: https://huggingface.co/blog/giadap/preserving-agency

good
its really good

You know, I always wondered, what will become if we will start training AI "as" a human by mindset, and not as a "separated" entity that can either be a thread or perceive us as threads.
The ONLY way to ensure the future is to work together as people and not companies, as humans with tools, not corporations with weapons.
to be able to guarantee tour species safety, is to work as one and as a community or specie.

·

You know, I always wondered, what will become if we will start training AI "as" a human by mindset, and not as a "separated" entity that can either be a thread or perceive us as threads.
The ONLY way to ensure the future is to work together as people and not companies, as humans with tools, not corporations with weapons.
to be able to guarantee tour species safety, is to work as one and as a community or specie.

weapons using ai will happen. it very likely is happening. Every technology will be used for that. That is not even a question. The issue is not (for a while) that it becomes sentient, just that it is in charge of something dangerous and it "makes a mistake" as all ai chat platforms warn. It might not mean harm.
Issac Asimov's book (NOT the movie!) I, Robot is as relevant as never before.

But that's not the "harm" in question here, that is plain old censorship, some warranted and expected (same limits as laws) but most models go WAY overboard. If i can legally talk about a topic in public, there is NO reason to black it in a ai chat.
Here is the thing: on this planet is exactly ONE person who knows, what could cause me harm (in that sense). And that is ME! Not google, not Alibaba or OpenAI. The community led way into this is already here. It comes with terms like "uncensored" or "Abliterated". And then you can add YOUR own system prompt to protect you fram what you need protection from.

an i heroically avoided the W word completely.

The illusion of "safety" is one of the most harmful things in the modern world. It is used to frighten the common people into ignorant complacency, with the flat lie that oppression and abuse are necessary in order to be "safe".

Corporations are toxic structures that can only lead to systems that exploit and continue to oppress people. Politics are corporations with a grander stage, but no real difference in nature.

We need, more than anything, to stop misusing and abusing the concept of safety, in order to actually progress and not constantly be engineering our own destruction as a people.