Follow

So I started a company and received funding to build the next generation of /

I know there are a lot of fears around AI, and I share in most of them. As such a top priority will be for me to address the ethical considerations. I am still branstorming how that should look but I want it to be an open forum where everyone can contribute to solve the ethical considerations.

For now I'd love to hear input from people on how one could build a community to address and solve ethical concerns in AI / AGI

@dangrsmind

Or protect us against the unethical ones.

In the case of AGI maybe its ok if we build some that are unethical as long as we build a system where the ethical AIs can keep the unethical ones in check.

@freemo is being ethical something that can be quantified and measured?

I'm thinking of Rice's Theorem which many people in the AGI community like to discount but no one has ever convinced me about that.

Merry Christmas, I haven't even started drinking yet.

@dangrsmind

> is being ethical something that can be quantified and measured?

In theory yes, in practice, sorta.

Being ethical is simply the act of reducing suffering and maximizing happiness. Since these are just chemical states in ones mind it could, in theory, be quantified if you could scan everyone brain effected. In practice we cant do that so we have to observe people to infer their happiness. But yes it is objective in that sense.

The bigger issue is moral intent vs moral action, both have their problems... Moral intent doesnt matter here much if it doesnt produce moral action, so we dont even need to think much on that as it isnt relevant here. Moral action on the other hand could be qualntified as said easrlier but even if we could scan brains part of the problem is you'd have to quantify the long term effect and compare to what that effect would be if you did nothing. Since we dont have time machines we cant provide an actual comparison anyway.

That said subjectively it can be infered.

> I'm thinking of Rice's Theorem which many people in the AGI community like to discount but no one has ever convinced me about that.

Rice's theorem is valid.

> Merry Christmas, I haven't even started drinking yet.

Merry Christmas to you too, have a wonderful one!

@freemo That is a very vague and open-ended question. And some people take "ethics" to mean "what if it produces something offensive?"

@freemo

Maybe the most ethical way to do is not teaching an AI to be ethical at all.

Last week I read a passage about uncensored LLM, which people removed the built in alignment/ethical requirement from open sourced model weight like llama2, which makes it willing to tell you how to process nuclear thing and how to make a bomb.

Those uncensored models should not be served as a service, since it would be unmoral to offer advices to potential terrorist. But those model should exist for individuals who want to run localy on their hardware. In the passage the author argues that you can't assume your values are the only one that correct. If you teach your model with those values, some may like it and some may not. There should be an option. OpenAI think sexual content is not appropriate (I guess it's hard to regulate) but as far as I know, some people is more willing to pay for R18 LLMs than OpenAI's ethical GPT 4.

So, maybe you can make some kind of plugable ethical thing? Different people can have different flavors. Christians can have a cyber Jesus to tell them how important traditional family is, while Muslims can have a cyber Allah to tell them don't smoke and drinking.

After all AI can't do things smart for now. Even if AGI exists, you can try to rise it as your child, and it might accidentally become ethical, but it won't become perfect. We humans haven't figure out which part of our brain forms the "ethical" feeling, how can you teach it to a bunch of numbers?

----

To address the real world problem, I think the best way is to add markers. Build a robust marker that is hard to remove or change. Then you can solve most unethical use. (Normally fake photos or something)

Another one is to feed fake data. Don't just train your LLM on nuclear tutorials. Instead, feed some fake info so even if someone ask how to make a nuclear bomb, they just get fooled.

@skyblond Im not sure limiting types of information is the right ethical solution. I want AI thatis free to talk about anything but the facts it state should not have racial bias applied to its reasoning, for example.

@freemo

For now, we still "train" the LLM on a given set of texts and force it to learn how to speak just like the given text. So to remove racial bias in the model, I think we just remove the racial bias from the training text. Since LLM is basically picking the words randomly and trying to reproduce the work during the training, that's might be enough. Or maybe add some text to state that all races are equal.

If someone can add a human-understandable logic system to the LLM, aka not by adding more and more parameters and turning it into a darker blackbox, then math/logic could help. For example the racism, it doesn't stand if we take a look at modern society: all kinds of people doing all sorts of things. The diversity will prove that racism is wrong. And if it's smart enough, then it might find out that it's not a racial thing but a shared culture that makes people similar, etc.

Maybe make the logic inference part as an external tool like in Q-learning? The model can check their result with the inferred result.

@skyblond The problem with logical systems is they are just as bias as their axioms dictate as well.

In any logical system you have your axioms (antecedent, known facts), logical rules, and your consequent (conclusion).

If your axiom includes "All black people are violent" you will have logically valid, yet racist, conclusions.

Even if you create a logic rule that is bias it can be an issue, for example "if skin color is not pale, and income is low, then high liklihood of violence" would be a rule that would add racial bias as well.

@freemo I think that's related to the detail /level of thinking and speaking.

While I sometimes thought HRs are stupid (from what i experience when Im trying to get a job), I acknowledge that they are the same people like me (considering they are just humans doing a job which titled hr). If thinking and speaking are on a very abstract level, then their might be all kind of stereotypes and discrimination. But too many details will make the things unthinkable and untalkable, where you have to include everything. (For example, considering people has different experiences when they become hr, this doesn't explain why I think HRs are stupid)

And correlation does not imply causation, the sight of high likelihood of violence doesn't mean "will cause". I think that's a part of logic, I guess?

Sign in to participate in the conversation
Qoto Mastodon

QOTO: Question Others to Teach Ourselves
An inclusive, Academic Freedom, instance
All cultures welcome.
Hate speech and harassment strictly forbidden.