LESSWRONG
LW

2713
Martin Vlach
7251151
Message
Dialogue
Subscribe

If you get an email from aisafetyresearch@gmail.com , that is most likely me. I also read it weekly, so you can pass a message into my mind that way.
Other ~personal contacts: https://linktr.ee/uhuge 

Posts

Sorted by New

Wikitag Contributions

Comments

Sorted by
Newest
1Martin Vlach's Shortform
3y
35
GPT-oss is an extremely stupid model
Martin Vlach6d10

I'd bet "re-based" model ala https://huggingface.co/jxm/gpt-oss-20b-base when instruction-tuned would do same as similarly sized Qwen models.

Reply
Project Vend: Can Claude run a small shop?
Martin Vlach2mo10

It's provided the current time together with other 20k sys-prompt tokens, so substantially more diluted influence on the behaviours..?

Reply
So You Think You've Awoken ChatGPT
Martin Vlach2mo10

Folks like this guy hit it on hyperspeed - 

https://www.facebook.com/reel/1130046385837121/?mibextid=rS40aB7S9Ucbxw6v

 

I still remember university teacher explaining how early TV transmission were very often including/displaying ghosts of dead people, especially dead relatives.

As the tech matures from art these phenomena or hallucinations evaporate.

Reply
Energy-Based Transformers are Scalable Learners and Thinkers
Martin Vlach2mo40

 you seem to report one OOM less than this picture in https://alexiglad.github.io/blog/2025/ebt/#:~:text=a%20log%20function).-,Figure%208,-%3A%20Scaling%20for

Reply
Open Thread - Summer 2025
Martin Vlach3mo10

Link to Induction section on https://www.lesswrong.com/lw/dhg/an_intuitive_explanation_of_solomonoff_induction/#induction seems broken on mobile Chrome, @habryka 

Reply
Open Thread - Summer 2025
Martin Vlach3mo20

I've heard that hypothesis in a review of that blog post of Anthropic, likely by 

AI Explained

maybe by

bycloud

.

 

They've called it "Chekov's gun".

Reply
Open Thread - Summer 2025
Martin Vlach3mo10

What's your view on sceptic claims about RL on transformer LMs like https://arxiv.org/abs/2504.13837v2 or one that CoT instruction yields better results than <thinking> training?

Reply
Open Source Search (Summary)
Martin Vlach3mo10

Not the content I expect labeled AIb Capabilities,

although I see how that'd be vindicated. 

 

By the way, if I write an article about LMs generating SVG, that's a plaintext and if I put an SVG illustration up, that's an image, not a plaintext?

Reply1
Martin Vlach's Shortform
Martin Vlach3mo10

Trivial, but do token-based LMs follow instructions like "only output tokens '1', '2', '3'" where they'd output 123 as one token without that instruction?

Reply1
Martin Vlach's Shortform
Martin Vlach4mo10

I'd update my take from a very pessimist/gloom one to an (additional) excited one: Those more intelligent models building a clear view of the person they/it interacts with is a sign of emerging empathy, which is a hopeful property for alignment/respect.

Reply
Load More
Zombies
a year ago
(+52/-50)
4Draft: A concise theory of agentic consciousness
3mo
2
0Thou shalt not command an alighned AI
4mo
4
8G.D. as Capitalist Evolution, and the claim for humanity's (temporary) upper hand
4mo
3
6Would it be useful to collect the contexts, where various LLMs think the same?
Q
2y
Q
1
1Martin Vlach's Shortform
3y
35