Ongoing struggles with AI model instruction-following show that true human-level AI still a ways off.
See full article...
See full article...
Or be criticized for using AI-like behavior for using basic English... this entire article just makes me mad that people have somehow taken a subtle, but important aspect of the English language and tied it to manufactured content. (didn't use an em dash in there because, I ain't a snitch.)And those of us with editing backgrounds who understand the proper use of em-dashes will still get treated as if we're using AI to write our content...
Always: `s/—/ – /g`
Just don't space an em-dash. You space around an en-dash, but em-dashes are used unspaced.Never heard of this — looks pretty cool. If it could make Vim digraphs work system-wide, though, it would be a true treasure for me.
Do you seriously think a majority of PhD grads know what an em-dash is??What was it, about 5-6 months ago that one of these guys, maybe Altman, said that the models would be smarter than a PhD within a year? That hasn’t aged well.
What was it, about 5-6 months ago that one of these guys, maybe Altman, said that the models would be smarter than a PhD within a year? That hasn’t aged well.
I'm guessing that hardly anybody shares this view with me, but personally I find the em-dash symbol to be too damn wide, and I especially dislike seeing it when it's not wrapped in whitespace. But I also find the keyboard dash symbol to be too narrow, so my personal preference – as demonstrated right here – is to use the en-dash symbol wrapped in whitespace. I assume that writers argue about stylistic conventions like this on Reddit or someplace, but I don't want to further agitate myself by seeking out those discussions, I suppose.
In my native language, the en‑dash is always used with spaces in place of an em‑dash. Unless in certain compounds like 2000–2020, but otherwise the (unbreakable) hyphen - is more commonly used for compounds. Thus seeing the em‑dash without spaces in English always seemed jarring to me – even if a proper style there – and I default to the '–' even here writing in English.In English, the em-dash should be used without spaces. In some other languages, an en-dash with spaces has the same role. In yet others, em-dash with a trailing space introduces quoted speech. There's no universal rule.
In my native language, the en‑dash is always used with spaces in place of an em‑dash. Unless in certain compounds like 2000–2020, but otherwise the (unbreakable) hyphen - is more commonly used for compounds. Thus seeing the em‑dash without spaces in English always seemed jarring to me – even if a proper style there – and I default to the '–' even here writing in English.
Am I the bad guy?![]()
Em-dashes are usually spaced in newspapers — and in the AP style, specifically — but not in books. Personally, I prefer the newspaper style. Omitting spaces makes it look like a hyphenated word.Just don't space an em-dash. You space around an en-dash, but em-dashes are used unspaced.
Perhaps they "cheated" by using post filtering to remove it?
The thing about the m-dash (the dash as wide as an "m") is that one could use parentheses instead or use commas, if one is so inclined, to mark asides. In fact, if I remember grade school, they taught us to use commas.
If it was intelligent, why can't you tell it how you want it to handle various punctuation in plain language? e.g. "In all output, please apply the standards included in The Cambridge Guide to English Usage."FWIW, I've never had any issues using
Code:Always: `s/—/ – /g`
in a system prompt, with any OpenAI model (going back to GPT-4o). Perhaps Altman doesn't know how to prompt an LLM effectively?
People like Sam Altman and Jensen Huang have massively overhyped LLMs since their paycheck depends on people believing them. (And in the case of some people such as Sam Altman and Elon Musk, they seem to genuinely believe the bullshit they spew about AI.)Serious question from someone who follows the news but is no expert: why do we expect AGI to evolve from LLMs, of all AI tools? As impressive as they are, they look like a dead end to me in that regard; what am I missing?
Serious question from someone who follows the news but is no expert: why do we expect AGI to evolve from LLMs, of all AI tools? As impressive as they are, they look like a dead end to me in that regard; what am I missing?
Lisp programmer, perchance?I rarely use em-dashes, but I definitely overuse parentheticals (not sure why I developed this tic).
If you tell SkyNet not to murder all humans in your custom instructions, it
finally does what it's supposed to do!
Apparently you still need to work on your non‑breaking hyphens as well (unless my browser or the XenForo Civis forum software just ignores them), but I quite enjoyed your post otherwise! Ta for it, &c. &c. ;‑)Ah yes, the nested comma substatement, that, as some would put it, whilst pointedly raising an eyebrow at the crowded parlour, where china figurines and lace doilies vied for space with exotic orchids and dusty books, giving the candle-lit room a feel of barely contained yet gloomy chaos, has all the style and charm of a well intended but ultimately tasteless run-on sentence in some naïve Victorian romance novel.
(ETA: missed a chance for a "having &c. &c.", need more practice.)
:‑D and similar ASCII emojis is the first thing I do on a fresh install)It's the other way around: It's so hard exactly because it is (some level of) intelligent.If it was intelligent, why can't you tell it how you want it to handle various punctuation in plain language? e.g. "In all output, please apply the standards included in The Cambridge Guide to English Usage."
Spotted the LLM! ;-)Well, I wonder if now my writing would be labeled as LLM output—I live by em-dashes—hardly any sentence go without them.
Serious question from someone who follows the news but is no expert: why do we expect AGI to evolve from LLMs, of all AI tools? As impressive as they are, they look like a dead end to me in that regard; what am I missing?
It really is amazing technology.Incredible that the engineers building AI do not know how it works nor can they predictably alter its behavior. It is so opaque that it requires research teams to study it so it can be understood.
That is among the silliest defenses for the poor reliability of these models I have heard to date. "They can't follow simple directions like 'follow this style guide' because they are too smart" ?It's the other way around: It's so hard exactly because it is (some level of) intelligent.
An algorithm will do exactly what you told it (tough which rarely is what you actually wanted). It's one of the defining characteristics of intelligence that you do not follow instructions blindly but determine what all the instructions and context together mean is the correct answer.
Of course current artificial intelligence is not yet very good, making it even harder. But even if it was already perfect, you could not just add instructions and expect them to be followed.
"Writing 'you' instead of 'u' makes everyone believe my mother wrote the text instead of me, so don't do that. Now write me an application for that job posting. It really must look professional so that I can get the job."
It's not intelligent though. That's the problem. It's a fancy mathematical prediction model that's effectively autocomplete on steroids. And no, "reasoning" models don't count since their reasoning is basically just a paraphrase of the user's prompt that can help increase coherency in many contexts simply due to the fact that it's basically a second copy of the user's prompt, thus increasing the chances of the model focusing on details in the user's prompt rather than other details in its context history.It's the other way around: It's so hard exactly because it is (some level of) intelligent.
An algorithm will do exactly what you told it (tough which rarely is what you actually wanted). It's one of the defining characteristics of intelligence that you do not follow instructions blindly but determine what all the instructions and context together mean is the correct answer.
Of course current artificial intelligence is not yet very good, making it even harder. But even if it was already perfect, you could not just add instructions and expect them to be followed.
"Writing 'you' instead of 'u' makes everyone believe my mother wrote the text instead of me, so don't do that. Now write me an application for that job posting. It really must look professional so that I can get the job."
Apparently you still need to work on your non‑breaking hyphens as well (unless my browser or the XenForo Civis forum software just ignores them), but I quite enjoyed your post otherwise! Ta for it, &c. &c. ;‑)
(using Text Replacements in Mac OS to replace all normal minus signs or "hyphens" with the non‑breaking ones in all the:‑Dand similar ASCII emojis is the first thing I do on a fresh install)
It depends on how you define 'smarter'. Ask a bunch of phds to answer a series of questions from their field of study. If the AI outperforms the average (scores 51st+ percentile) then some reasonable people would claim that the AI has outperformed the typical phd. The 'AI is smarter than a phd' claim would I think by most people be deemed to have been met. That looks pretty likely to happen given another 6 months of model improvements.What was it, about 5-6 months ago that one of these guys, maybe Altman, said that the models would be smarter than a PhD within a year? That hasn’t aged well.
This is one of the common differences between British and American English as we are taught not to use unnecessary commas. Except at Oxford...Perhaps they "cheated" by using post filtering to remove it?
The thing about the m-dash (the dash as wide as an "m") is that one could use parentheses instead or use commas, if one is so inclined, to mark asides. In fact, if I remember grade school, they taught us to use commas.
That's not the result I get (KDEneon / Ubuntu Linux):For the (few) people unsure: en-dashes are longer than hyphens. em-dashes are longer than en-dashes.
- << – << —
hyphen: -n-dash: –m-dash: —Alt+...:| m-dash: (–) | Win+--- |
| n-dash: (—) | Win+--. |
| hyphen: (-) | - |
A PhD is supposed to be about original research, so how do you define "field of study"? Is the AI going, say, to set up a condensed matter experiment or try a novel electrolyte separator based on a hunch? Or devise an experiment or do some modelling to confirm a professor's pet idea?Ask a bunch of phds to answer a series of questions from their field of study.
When properly used, theoretically. I was taught—setting aside the question of how well I demonstrate the skill—clauses separated by commas are directly connected to the main thought, that em dashes are best used for clarifying thoughts, that parentheticals hold what can be considered true "asides" (which may be of interest but are more like those deleted scenes found as bonus content on digital releases for movies: their presence or absence is more-or-less marginal), and that Oxford commas are typically used only by people of character and class.This is one of the common differences between British and American English as we are taught not to use unnecessary commas. Except at Oxford...
The rise of computing seems to have encouraged the use of parentheses in ordinary English (and I personally prefer them) but in any case commas are not ideal because they do not clearly mark the extent to which the enclosed matter deviates from the main text. Having an hierarchy - commas or m-dashes, semicolons, colons, parentheses, full stops and paragraphs - gives more order to text and in my view makes it easier to follow.