Philip Niewold
Philip Niewold has not written any posts yet.

Philip Niewold has not written any posts yet.

You make a number of interesting points.
Interpretabilty would certainly help a lot, but I am worried about our inability to recognize (or even agree) to leaving local optima that we believe are 'aligned'.
Like when we force a child to go to bed by his parents when it doesn't want to, because the parents know it is better for the child in the long run, but the child is still unable to extend his understanding to this wider dimension.
Ar some point, we might experience what we think of as misaligned behaviour when the A.I. is trying to push us out of a local optimum that we experience as 'good'.
Similar to how current A.I.'s... (read more)
While I acknowledge this is important, it is a truly hard problem, as it often involves looking not just at first-order consequences, but also at second-order consequences and so on. People are notoriously bad at predicting, let alone managing side-effects.
Besides, if you look at it more fundamentally, human natures and technological progress in a broad sense has many of these side effects, where you basically need to combat human nature itself to have people take into account the side-effects.
We are still struggling coming to terms and accurately classifying things like environmental pollution, global warming and such. Understanding the illegible problems -> explaining them to policy makers -> thinking of legible solution ->... (read more)
An LLM it a tool of communicative expression, but so it the written or spoken word, music etc. It is a medium throug which the intent travels. As a Dutchman, I have a preference of being direct and clear, but the impact of my words sometimes have the opposite effect, as my listeners do not have my context and can react emotionally to a worded message that is meant factually. If an LLM can help me translate such expression to a language that is better for my target audience to understand, then it is similar to translating into another language.
Still, the written word is no substitute for the full breadth of human... (read more)
Our exploration system is very useful, but it takes a lot of energy (and anxiety), because of the inherent cost of failure which genetics baked into our brain. Hence, doing something new everyday in a society as complex and everchanging as our own is very useful, but very hard with our outdated brain hardware and software.
Add to that the distractions that hijack our outdated brain mechanisms: we have gotten better and better and such hijacking, creating an additional barrier. Doing this is comparably difficult to keeping to a strict diet and exercise regime while mouthwatering delicacies and relaxed convenciences are offered to you at every turn.
You are trying to break patterns (habits), but it is extremely hard to create a habit/pattern of newness, for habits/patterns are fundamentally opposed to doing things in a novel way.
I think your claim the rudimentary abilities arrive before transformational ones cannot be applied to A.I. the same as human intelligence. While humans might have taken millennia to go from caveman painting to our current ability to produce artistic images, it is clear that A.I. became transformational very quickly in that particular field. You see the same transformational abilities in text writing, music and video too and software development is getting there.
Some of the more artistic of these abilities don't have a clear benchmark, but even with more fuzzy criteria for success, they already outcompute most humans.
Some of the building blocks of A.I. are fundamentally different from us, that is why the... (read more)
Your working paper, "Open Global Investment as a Governance Model for AGI." It provides a clear, pragmatic, and much-needed baseline for discussion by grounding a potential governance model in existing legal and economic structures. The argument that OGI is more incentive-compatible and achievable in the short term than more idealistic international proposals is a compelling one.
However, I wish to offer a critique based on the concern that the OGI model, by its very nature, may be fundamentally misaligned with the scale and type of challenge that AGI presents. My reservations can be grouped into three main points.
1. The Inherent Limitations of Shareholder Primacy in the Face of Existential Stakes
The core of the... (read 548 more words →)
I don't think people in general react well to societal existential risks, regardless how well or courageous the message is framed. These are abstract concerns. The fact that we are talking about AI (an abstract thing in itself) makes it even worse.
I'm also a very big opponent of arguing by authority (I really don't care how many nobel laureates are of the opinion of something, it is the content of their argument I care about, now how many authorities are saying it). That is simply that I cannot determine the motives of these authorities and hence their opinions, while I can't argue with logic and facts)
Usually it is better to make people... (read more)
Social messaging is fine balancing act: people like to offload responsibility and effort, especially if it doesn't come at the cost of status. And, to be honest, you don't know if your question would impose upon the other (in terms of cognitive load, social pressure or responsibility), so you it is smart to start your social bid low and see if the other wants to raise the price. Sometimes they work, creating a feedback loop similar to how superstitions evolve: if it is minimal effort and sometimes it is effective, better continue using it.
As a child, I despised a lot of these practices, to me it felt like people were lying all... (read more)
Of course it is perfectly rational to do so, but only from a wider context. From the context of the equilibrium it isn't. The rationality your example is found because you are able to adjudicate your lifetime and the game is given in 10 second intervals. Suppose you don't know how long you have to live, or, in fact, now that you only have 30 seconden more to live. What would you choose?
This information is not given by the game, even though it impacts the decision, since the given game does rely on real-world equivalency to give it weight and impact.
Excellent from-the-heart post. Predictability and stability is a great good, and if you have a large imagination and good intellect, you can become lost in your own projections easily. I know I do.
You have just realized that just working towards some future is not a viable path to living. This is a lesson most people take decades to discover. Perhaps you look happier because your mind was forced to live more and the here and now and less in the future, and living in the here and now is really.
It is hard to both grasp and let go. But that is really the only option we have.