Frustrated by claims that "enlightenment" and similar meditative/introspective practices can't be explained and that you only understand if you experience them, Kaj set out to write his own detailed gears-level, non-mysterious, non-"woo" explanation of how meditation, etc., work in the same way you might explain the operation of an internal combustion engine.
/the-political-is-personal/
It seems like many people propose "generalization from their own example" as a model for the entire humanity. And it can be quite annoying when people around you agree on a model that doesn't fit you at all... and when you point it out, they dismiss it by saying that you are in a denial. Because they have examined their own minds deeply, and found out that it was true... yeah, possibly so, but that doesn't necessarily make it true about the others.
This isn’t primarily about how I write. It’s about how other people write, and what advice they give on how to write, and how I react to and relate to that advice.
I’ve been collecting those notes for a while. I figured I would share.
At some point in the future, I’ll talk more about my own process – my guess is that what I do very much wouldn’t work for most people, but would be excellent for some.
Aspies certainly seem to do this less!
You mean, like him as a blogger? Or as a person in real life?
The latter? Like, I subconsciously parse his blogging voice not unlike as if it were a person in my tribal surroundings, and I like/admire/relate to that virtual person, and I think this is what causes some extra persuasion
I mean yes it's embarrassing, but it's what I see in myself and what seems to be most consistent with what everyone else is doing, certainly more consistent than what they claim they're doing.
E.g. it seems rare for someone who activel...
Learn the official language of the place you are migrating to.
Yes, this sounds completely obvious to me.
Of course, learning languages takes time, and may be more difficult for older people. So I wouldn't expect fluent speech from the start, and maybe from the older generation even in a year or two -- just a gesture of trying. The important thing is that they do not isolate their kids and themselves from the local society behind the language barrier. Become bilingual.
Heck, if I had to emigrate somewhere, I would want my kids to speak the local language, bec...
Agreed. If I'm talking to someone who I expect to be able to recalibrate, I just explain that I think the standard norms are dumb, the norms I actually follow, and then give an honest and balanced assessment. If I'm talking to someone I don't really know, I generally give a positive but not very detailed reference or don't reply, depending on context.
I’m considering translating my work into English to share it with the LessWrong community, but I’d like to first ask if it aligns with the community's interests and could be valuable. Below is a summary of the work to help evaluate its relevance:
We explore the specific causal mechanisms linking humor recognition to learning outcomes, including the computational and neurological pathways involved.
This study began with a practical goal: to evaluate the use of humor as a pedagogical tool in Cardiopulmonary Resuscitation (CPR) courses through a randomized trial. However, the lack of clear criteria to define and operationalize "humor" in educational contexts led us to explore its conceptual foundations. Initially, we adopted Clarke's formula, which describes humor as "a pleasant...
Scheming AIs may have secrets that are salient to them, such as:
Extracting these secrets would help reduce AI risk, but how do you do that? One hope is that you can do fuzzing of LLMs,[1] e.g. by adding noise to LLM weights or activations.
While LLMs under fuzzing might produce many incorrect generations, sometimes-correct generations can still be very helpful if you or the LLM itself can tell if a given answer is correct. But it’s still unclear if this works at all: there are probably some intermediate activations that would result in an LLM telling you the secret, but...
By doing more search around promising vectors found with random search or MELBO, you could get more powerful vectors, and that could be useful for unlocking / fuzzing-adversarial-training. It's unclear if that would be more effective than just fine-tuning the model on the generation from the best random vectors, but it would be worth trying.
For interp, I don't know what interp metric you want to optimize. Vector norm is a really bad metric: effective MELBO vectors have a much smaller norm, but qualitatively I find their results are sometimes much more erra...
I didn’t want to write this.
Not for lack of courage—I’d meme-storm Putin’s Instagram if given half a chance. But why?
Okay, maybe a little lack of courage.
And yet, something can be extracted from all this madness, right?
Then comes someone named Gwern. He completely ignores my thesis and simply asks:
"Tell military firefighter stories."
My first instinct was to dismiss him as an oddball—until a friend told me I was dealing with a legend of rationality. I have to admit: I nearly shit myself. His comment got more likes than the post I’d spent years working on.
Someone with,...