All of soth02's Comments + Replies

soth02
-1-1

There is a problem in that any group that is generating alpha would likely lose alpha/person if they allow random additional people into their group.

Think Renaissance Medallion fund.  It's been closed to outside investment since near its inception 30 years ago.  Prerequisites for the average person joining would be something like true-genius level Phd in a related STEM field.

An analogue which is closely related is poker players who use solvers to improve their game.  The starting stakes are a bit lower.  The solvers are like a few thous... (read more)

soth02
40

Coincidentally, that scene in The Big Short takes place on January 11 (2007) :D

2gjm
Yeah, could be.
soth02
10

https://www.lesswrong.com/posts/jnyTqPRHwcieAXgrA/finding-goals-in-the-world-model

Could it be possible to poison the world model an AGI is based on to cripple its power?

Use generated text/data to train world models based on faulty science like miasma, phlogiston, ether, etc.

Remove all references to the internet or connectivity based technology.

Create a new programming language that has zero real world adoption, and use that for all code based data in the training set.

soth02
10

There might be a way to elicit how aligned/unaligned the putative AGI is.

  1. Enter into a Prisoner's Dilemma type scenario with the putative AGI.
  2. Start off in the non-Nash equilibrium of cooperate/cooperate.
  3. The number of rounds is specified at random and isn't known to participants. (possible variant is declare false last rounds, and then continue playing for x rounds).
  4. Observe when/if the putative AGI defects in the 'last' round.
soth02
10

Does there have to be a reward?  This is using brute force to create the underlying world model.  It's just adjusting weights right?

1Evan R. Murphy
I think there has to be some kind of reward or loss function, in the current paradigm anyway. That's what gradient descent uses to know such weights to adjust on each update. Like what are you imagining is the input output channel of this AI? Maybe discussing this a bit would help us clarify.
soth02
20

Brute force alignment by adding billions of tokens of object level examples of love, kindness, etc to the dataset.  Have the majority of humanity contribute essays, comments, and (later) video.

1Evan R. Murphy
What would be the reward you're training the AI on with this dataset? If you're not careful you could inadvertently train a learned optimizer, e.g. a "hugging humans maximizer" to take a silly example. That may sound nice but could have torturous results, e.g. the AI forcing humans to hug, or replacing biological humans with server farms housing simulations of quadrillions of humans hugging.
soth02
30

I wonder what kind of signatures a civilization gives off when AGI is nascent.

soth02
10

Develop a training set for alignment via brute force.  We can't defer alignment to the ubernerds.  If enough ordinary people (millions? tens of millions?) contribute billions or trillions of tokens, maybe we can increase the chance of alignment.  It's almost like we need to offer prayers of kindness and love to the future AGI: writing alignment essays of kindness that are posted to reddit, or videos  extolling the virtue of love that are uploaded to youtube.

soth02
10

AI presents both staggering opportunity and chilling peril. Developing intelligent machines could help eradicate disease, poverty, and hunger within our lifetime. But uncontrolled AI could spell the end of the human race. As Stephen Hawking warned, "Success in creating AI would be the biggest event in human history. Unfortunately, it might also be the last, unless we learn how to avoid the risks."

soth02
10

AI safety is essential for the ethical development of artificial intelligence."

soth02
10

"AI safety is the best insurance policy against an uncertain future."

soth02
10

"AI safety is not a luxury, it's a necessity."

soth02
10

While it is true that AI has the potential to do a lot of good in the world, it is also true that it has the potential to do a lot of harm. That is why it is so important to ensure that AI safety is a top priority. As Google Brain co-founder Andrew Ng has said, "AI is the new electricity." Just as we have rules and regulations in place to ensure that electricity is used safely, we need to have rules and regulations in place to ensure that AI is used safely. Otherwise, we run the risk of causing great harm to ourselves and to the world around us.

soth02
10

I'm soliciting input from people with more LLM experience to tell me why this naive idea will fail.  I'm hoping it's not in the category of "not even wrong".  If there's a 2%+ shot this will succeed, i'll start coding.

From what I gather, the scrapers look for links on reddit to external text files.  I could also collate submissions, zip them and upload to github/IPFS.  Which ever format is easiest for inclusion into a Pile.