Comment by kouteiheika

Comment by kouteiheika 17 hours ago

97 replies

> Anthropic occupies a peculiar position in the AI landscape: a company that genuinely believes it might be building one of the most transformative and potentially dangerous technologies in human history, yet presses forward anyway. This isn't cognitive dissonance but rather a calculated bet—if powerful AI is coming regardless, Anthropic believes it's better to have safety-focused labs at the frontier than to cede that ground to developers less focused on safety (see our core views).

Ah, yes, safety, because what is more safe than to help DoD/Palantir kill people[1]?

No, the real risk here is that this technology is going to be kept behind closed doors, and monopolized by the rich and powerful, while us scrubs will only get limited access to a lobotomized and heavily censored version of it, if at all.

[1] - https://www.anthropic.com/news/anthropic-and-the-department-...

reissbaker 17 hours ago

This is the major reason China has been investing in open-source LLMs: because the U.S. publicly announced its plans to restrict AI access into tiers, and certain countries — of course including China — were at the lowest tier of access. [1]

If the U.S. doesn't control the weights, though, it can't restrict China from accessing the models...

1: https://thefuturemedia.eu/new-u-s-rules-aim-to-govern-ais-gl...

  • IncreasePosts 13 hours ago

    Why wouldn't China just keep their own weights secret as well?

    If this really is a geopolitical play(I'm not sure if it is or isn't), it could be along the lines of: 1) most AI development in the US is happening at private companies with balance sheets, share holders, and profit motives. 2) China may be lagging in compute to beat everyone to the punch in a naked race

    Therefore, releasing open weights may create a situation where AI companies can't as effectively sell their services, meaning they may curtail r&d at a certain point. China can then pour nearly infinite money into it and eventually get up to speed on compute and win the race

    • zamalek 13 hours ago

      They are taking the gun out of USA's hand and unloading it, figuratively speaking. With this strategy they don't have the compete at full competency with the US, because everyone else will with cheaper models. If a cheaper model can do it, then why fork out for Opus?

    • giancarlostoro 10 hours ago

      Because they dont have the chips, but if people in countries with the chips provide hosting or refine their models they benefit from those breakthroughs.

      • faitswulff 9 hours ago

        They're definitely investing in the chips as well. It's an ecosystem play.

    • bamboozled 13 hours ago

      I think it's just because China makes it's money from other sources, not from AI, and from what I've read, the advantage of China killing the US's AI advantage is killing it's stock market / disrupting.

      Seems like it may have a chance of working if you look at the companies highest valued on the S&P 500:

      NVIDIA, Microsoft, Apple, Amazon, Meta Platforms, Broadcom, Alphabet (Class C),

      • adventured 8 hours ago

        The share of revenue that Microsoft, Google, Meta, Apple, Alphabet and Amazon are currently deriving from the AI market as a share of their total revenue, is less than 10%.

  • dist-epoch 16 hours ago

    It isn't "China" which open-source LLMs, but individual Chinese labs.

    China didn't yet made a sovereign move on AI, besides investing in research/hardware.

    • reissbaker 14 hours ago

      I think "investing in research and hardware" is fairly relevant to my claim of "China has been investing in open-source LLMs." China also has partial ownership of several major labs via "golden shares" [1] like Alibaba (Qwen) and Zai (GLM) [2], albeit not DeepSeek as far as I know.

      1: https://www.theguardian.com/world/2023/jan/13/china-to-take-...

      2: https://www.globalneighbours.org/chinas-zhipu-ai-secures-140...

    • baq 16 hours ago

      Axiom of China: nothing of importance happens in China without CCP involvement.

      • yorwba 15 hours ago

        The CCP controlling the government doesn't mean they micromanage everything. Some Chinese AI companies release the weights of even their best models (DeepSeek, Moonshot AI), others release weights for small models, but not the largest ones (Alibaba, Baidu), some keep almost everything closed (Bytedance and iFlytek, I think).

        There is no CCP master plan for open models, any more than there is a Western master plan for ignoring Chinese models only available as an API.

    • throwup238 16 hours ago

      As far as I can tell AI is already playing a big part in the Chinese Fifteenth five year plan (2026-2030) which is their central top-down planning mechanism. That’s about as big a move as they can make.

  • slanterns 15 hours ago

    and Anthropic bans access from China along with throwing some politic propagenda bs

jimbo808 10 hours ago

I don't believe that they believe it, I believe that they're all in on doing all the things you'd do if your goal was to demonstrate to investors that you truly believe it.

The safety-focused labs are the marketing department.

An AI that can actually think and reason, and not just pretend to by regurgitating/paraphrasing text that humans wrote, is not something we're on any path to building right now. They keep telling us these things are going to discover novel drugs and do all sorts of important science, but internally, they are well aware that these LLM architectures fundamentally can't do that.

A transformer-based LLM can't do any of the things you'd need to be able to do as an intelligent system. It has no truth model, and lacks any mechanism of understanding its own output. It can't learn and apply new information, especially not if it can't fit within one context window. It has no way to evaluate if a particular sequence of tokens is likely to be accurate, because it only selects them based on the probability of appearing in a similar sequence, based on the training data. It can't internally distinguish "false but plausible" from "true but rare." Many things that would be obviously wrong to a human, would appear to be "obviously" correct when viewed from the perspective of an LLM's math.

These flaws are massive, and IMO, insurmountable. It doesn't matter if it can do 50% of a person's work effectively, because you can't reliably predict which 50% it will do. Given this unpredictability, its output has to be very carefuly reviewed by an expert in order to be used for any work that matters. Even worse, the mistakes it makes are meant to be difficult to spot, because it will always generate the text that looks the most right. Spotting the fuckup in something that was optimized not to look like a fuckup is much more difficult than reviewing work done by a well-intentioned human.

  • astrange 5 hours ago

    No, Anthropic and OpenAI definitely actually believe what they're saying. If you believe companies only care about their shareholders, then you shouldn't believe this about them because they don't even have that corporate structure - they're PBCs.

    There doesn't seem to be a reason to believe the rest of this critique either; sure those are potential problems, but what do any of them have to do with whether a system has a transformer model in it? A recording of a human mind would have the same issues.

    > It has no way to evaluate if a particular sequence of tokens is likely to be accurate, because it only selects them based on the probability of appearing in a similar sequence, based on the training data.

    This in particular is obviously incorrect if you think about it, because the critique is so strong that if it was true, the system wouldn't be able to produce coherent sentences. Because that's actually the same problem as producing true sentences.

    (It's also not true because the models are grounded via web search/coding tools.)

    • habinero 4 hours ago

      > if it was true, the system wouldn't be able to produce coherent sentences. Because that's actually the same problem as producing true sentences

      It is...not at all the same? Like they said, you can create perfectly coherent statements that are just wrong. Just look at Elon's ridiculously hamfisted attempts around editing Grok system prompts.

      Also, a lot of information on the web is just wrong or out of date, and coding tools only get you so far.

  • vancroft 6 hours ago

    Sounds like the old saying about the advertising industry: "I know half of my spending on advertising is wasted - I just don't know which half."

  • HDThoreaun 8 hours ago

    If you dont believe they believe it you havent paid any attention to the company. Maybe Dario is lying, although that would be an extremely long con, but the rank and file 100% believe it.

flatline 15 hours ago

Ironically, this is one the part of the document that jumped out at me as having been written by AI. The em-dash and "this isn't...but" pattern are louder than the text at this point. It seriously calls into question who is authoring what, and what their actual motives are.

  • observationist 15 hours ago

    People who work the most with these bots are going to be the researchers whose job it is to churn out this stuff, so they're going to become acclimated to the style, stop noticing the things that stick out, and they'll also be the most likely to accept an AI revision as "yes, that means what I originally wrote and looks good."

    Those turns of phrase and the structure underneath the text become tell-tales for AI authorship. I see all sorts of politicians and pundits thinking they're getting away with AI writing, or ghost-writing at best, but it's not even really that hard to see the difference. Just like I can read a page and tell it's Brandon Sanderson, or Patrick Rothfuss, or Douglas Adams, or the "style" of those writers.

    Hopefully the AI employees are being diligent about making sure their ideas remain intact. If their training processes start allowing unwanted transformations of source ideas as a side-effect, then the whole rewriting/editing pipeline use case becomes a lot more iffy.

    • visarga 13 hours ago

      What matters is not who writes the words. The source of slop is competition for scarce attention between creatives, and retention drive for platforms. They optimize for slop, humans conform, AI is just a tool here. We are trying to solve an authenticity problem when the actual problem is structural.

  • gnatman 15 hours ago

    Every time I see the em-dash call out on here I get defensive because I’ve been writing like that forever! Where do people think that came from anyway? It’s obviously massively represented in the training data!

    • astrange 5 hours ago

      The AIs aren't using emdashes because they're "massively represented in the training data". I don't understand why people think everything in a model output is strictly related to its frequency in pretraining.

      They're emdashing because the style guide for posttraining makes it emdash. Just like the post-training for GPT 3.5 made it speak African English and the post-training for 4o makes it say stuff like "it's giving wild energy when the vibes are on peak" plus a bunch of random emoji.

      • antonvs 3 hours ago

        > Just like the post-training for GPT 3.5 made it speak African English

        This is a misunderstanding. At best, some people thought that GPT 3.5 output resembled African English.

    • observationist 15 hours ago

      Where's the emdash key on your keyboard?

      There isn't one?

      Oh, maybe that's why people who didn't already know or care about emdashes are very alert to their presence.

      If you have to do something very exotic with keypresses or copypaste from a tool or build your own macro to get something like an emdash, or , it's going to stand out, even if it's an integral part of standard operating systems.

      • gnatman 15 hours ago

        Exotic? At least in every microsoft product i.e. word, outlook, etc. that I’ve had to use for school and business for the last couple decades does it automatically just by typing “—-“.

      • nonfamous 15 hours ago

        Typing hyphen-hyphen-space is hardly exotic — I've been doing that since well beyond the advent of generative AI.

      • czottmann 14 hours ago

        My German keyboard has umlaut keys: üäö. I use them daily. I was told that in other parts of the World, people don't have umlaut keys, and have to use combos like ⌥U + a/o/u.

        Boy, I sure hope they don't think me an AI.

        Just because many people have no idea how to use type certain characters on their devices shouldn't mean we all have to go along with their superstitions.

      • ben_w 13 hours ago

        > Where's the emdash key on your keyboard?

        > There isn't one?

        Mac, alt-minus. Did by accident once, causing confusion because Xcode uses monospace font where -, – and — look identical, and an m-dash where a minus should be gets a compiler error.

        iOS, long-press on the "-" key.

      • jonas21 14 hours ago

        > Where's the emdash key on your keyboard?

        The dash key is right between the "0" and the "="

        Press it twice and just about every word processing program in existence will turn it into an emdash.

      • lurking_swe 13 hours ago

        hyphen + space in microsoft word will often (depends on your settings) produce an em dash. It’s not some crazy hidden feature.

        These days word is less popular though, with google docs, pages, and other editors taking pieces of the pie. Maybe that’s where the skepticism comes from.

      • saagarjha 8 hours ago

        Where’s the copy paste key on your keyboard? Oh, there isn’t one? How could anyone possibly use this then?

      • skeeter2020 13 hours ago

        standby for the masses to drop the nerd-equivalent of "before it was cool" comments in 3...2...

      • crooked-v 13 hours ago

        Shift-option-minus on a Mac, just like how shift-option-8 is the degree symbol and option-slash is the division symbol.

        • skeeter2020 13 hours ago

          ...which are two more characters I bet have a higher rate of occurance in AI generated content too!

      • troupo 15 hours ago

        > There isn't one?

        I've used em-dash since I got my first MacBook in 2008.

        - Option + minus gives you en-dash

        - Option + Shift + minus gives you em-dash

        It quickly becomes automatic (as are a bunch of other shortcuts). Here's a question about this from 2006: https://discussions.apple.com/thread/377843

      • pb7 14 hours ago

        My computer converts -- into an emdash automatically. Been using it since 2011. Sorry you've been missing out on a part of the English language all this time.

regularization 17 hours ago

> to ensure AI development strengthens democratic values globally

I wonder if that's helping the US Navy shoot up fishing boats in the Caribbean or facilitating the bombing of hospitals, schools and refugee camps in Gaza.

  • odiroot 13 hours ago

    > Please don't use Hacker News for political or ideological battle. It tramples curiosity.

  • ch2026 16 hours ago

    It helps provide the therapy bot used by struggling sailors who are questioning orders and reducing "hey this isn’t what i signed up for" mental breakdowns.

    • conception 15 hours ago

      "Wait, this seems like a war crime." "You're absolutely right!"

ben_w 13 hours ago

> No, the real risk here is that this technology is going to be kept behind closed doors, and monopolized by the rich and powerful, while us scrubs will only get limited access to a lobotomized and heavily censored version of it, if at all.

Given the number of leaks, deliberate publications of weights, and worldwide competition, why do you believe this?

(Even if by "lobotomised" you mean "refuses to assist with CNB weapon development").

Also, you can have more than one failure mode both be true. A protest against direct local air polution from a coal plant is still valid even though the greenhouse effect exists, and vice versa.

  • kouteiheika 7 hours ago

    > Given the number of leaks, deliberate publications of weights, and worldwide competition, why do you believe this?

    So where can I find the leaked weights of GPT-3/GPT-4/GPT-5? Or Claude? Or Gemini?

    The only weights we are getting are those which the people on the top decided we can get, and precisely because they're not SOTA.

    If any of those companies stumbles upon true AGI (as unlikely as it is), you can bet it will be tightly controlled and normal people will either have an extremely limited access to it, or none at all.

    > Even if by "lobotomised" you mean "refuses to assist with CNB weapon development"

    Right, because people who design/manufacture weapons of mass destruction will surely use ChatGPT to do it. The same ChatGPT who routinely hallucinates widely incorrect details even for the most trifling queries. If anything, that'd only sabotage their efforts if they're stupid enough to use an LLM for that.

    Nevertheless, it's always fun when you ask an LLM to translate something from another language, and the line you're trying to translate coincidentally contains some "unsafe" language, and your query gets deleted and you get a nice, red warning that "your request violates our terms and conditions". Ah, yes, I'm feeling "safe" already.

    • astrange 5 hours ago

      Kimi-K2-Thinking and DeepSeek-V3.2 are open and pretty near SOTA.

    • ben_w 3 hours ago

      Imagine saying

        Operating systems are going to be kept behind closed doors, and monopolized by the rich and powerful, while us scrubs will only get limited access to what computers can really do!
      
      Getting the reply

        We have open-source OSes
      
      And then replying

        So where can I find the leaked source of Windows? Or MacOS?
      
      We have a bajillion Linuxes. There's a lot of open-weights GenAI models. Including from OpenAI, whose open models beat everything in their own GPT-3 and 4 families.

      But also not "those which the people on the top decided we can get", which is why Meta sued over the initial leak of the original LLaMa's weights.

      > true AGI

      Is ill-defined. Like, I don't think I've seen any two people agree on what it means… unless they're the handful that share the definition I'd been using before I realised how rare it was ("a general-purpose AI model", which they all meet).

      If your requirement includes anything like "learns quickly from few examples", which is a valid use of the word "intelligence" and one where all ML training methods known fail because they are literally too stupid to live (no single organism would survive long enough to make that many mistakes), and AI generally only make up for this by doing what passes for thinking faster than anything alive to the degree to which we walk faster than continental drift, then whoever first tasks such a model with taking over the world, succeeds.

      To emphasise two points:

      1. Not "trains", "tasks".

      2. It succeeds because anything which can learn from as few examples as us, while operating so quickly that it can ingest the entire internet in a few months, is going to be better at everything than anyone.

      At which point, you'd better hope that either whoever trained it, trained it in a way that respects concepts like "liberty" and "democracy" and "freedom" and "humans are not to be disassembled for parts", or that whoever tasked it with taking over the world both cares about those values and rules-lawyers the AI like a fictional character dealing with a literal-minded genie.

      > Right, because people who design/manufacture weapons of mass destruction will surely use ChatGPT to do it. The same ChatGPT who routinely hallucinates widely incorrect details even for the most trifling queries. If anything, that'd only sabotage their efforts if they're stupid enough to use an LLM for that.

      First, yes of course they will, even existing professionals, even when they shouldn't. Have you not seen the huge number of stories about everyone using it for everything, including generals?

      Second, the risk is new people making them. My experience of using LLMs is as a software engineer, not as a biologist, chemist, or physicist: LLMs can do fresh-graduate software engineering tasks at fresh-graduate competence levels. Can LLMs display fresh-graduate level competence in NBC? If LLMs can do that, they necessarily expand the number of groups who can run NBC programs to include any random island nation with not enough grads to run a NBC program, or mid-sized organised crime group, or Hamas.

      They don't even need to do all of it, just be good enough to help. "Automate cognitive tasks" is basically the entire point of these things, after all.

      And if the AI isn't competent to help with those things, if they're e.g. at the level of competence of "sure mix those two bleaches without checking what they are" (explosion hazard) or "put that raw garlic in that olive oil and just leave it at room temperature for a few weeks it will taste good" (biohazard, and one model did this), then surely it's a matter of general public safety to make them not talk about those things because of all the lazy students who are already demonstrating they're just as lazy as whoever wrote the US tariff policy that put a different tariff on an island occupied by only penguins vs. the country which owned it and which a lot of people suspect came out of an LLM.

      > Nevertheless, it's always fun when you ask an LLM to translate something from another language, and the line you're trying to translate coincidentally contains some "unsafe" language, and your query gets deleted and you get a nice, red warning that "your request violates our terms and conditions". Ah, yes, I'm feeling "safe" already.

      Use Google Translate. It's the same architecture, trained to give a translation instead of a reply. Or, equivalently, the chat models (and code generators like Claude) are the same architecture as Google Translate, trained to "translate" your prompt into an answer.

Aarostotle 17 hours ago

A narrow and cynical take, my friend. With all technologies, "safety" doesn't equate to plushie harmlessness. There is, for example, a valid notion of "gun safety."

Long-term safety for free people entails military use of new technologies. Imagine if people advocating airplane safety groused about the use of bomber and fighter planes being built and mobilized in the Second World War.

Now, I share your concern about governments who unjustly wield force (either in war or covert operations). That is an issue to be solved by articulating a good political philosophy and implementing it via policy, though. Sadly, too many of the people who oppose the American government's use of such technology have deeply authoritarian views themselves — they would just prefer to see a different set of values forced upon people.

Last: Is there any evidence that we're getting some crappy lobotomized models while the companies keep the best for themselves? It seems fairly obvious that they're tripping over each other in a race to give the market the highest intelligence at the lowest price. To anyone reading this who's involved in that, thank you!

  • ceejayoz 17 hours ago

    > Long-term safety for free people entails military use of new technologies.

    Long-term safety also entails restraining the military-industrial complex from the excesses it's always prone to.

    Remember, Teller wanted to make a 10 gigaton nuke. https://en.wikipedia.org/wiki/Sundial_(weapon)

    • Aarostotle 17 hours ago

      I agree, your point is compatible with my view. My sense is that this essentially an optimization question within how a government ought to structures its contracts with builders of weapons. The current system is definitely suboptimal (put mildly) and corrupt.

      The integrity of a free society's government is the central issue here, not the creation of tools which could be militarily useful to a free society.

  • kouteiheika 16 hours ago

    > Is there any evidence that we're getting some crappy lobotomized models while the companies keep the best for themselves? It seems fairly obvious that they're tripping over each other in a race to give the market the highest intelligence at the lowest price.

    Yes? All of those models are behind an API, which can be taken away at any time, for any reason.

    Also, have you followed the release of gpt-oss, which the overlords at OpenAI graciously gave us (and only because Chinese open-weight releases lit a fire under them)? It was so heavily censored and lobotomized that it has become a meme in the local LLM community. Even when people forcibly abliterate it to remove the censorship it still wastes a ton of tokens when thinking to check whether the query is "compliant with policy".

    Do not be fooled. The whole "safety" talk isn't actually about making anything safe. It's just a smoke screen. It's about control. Remember back in the GPT-3 days how OpenAI was saying that they won't release the model because it would be terribly, terribly unsafe? And yet nowadays we have open weight model orders of magnitude more intelligent than GPT-3, and yet the sky hasn't fallen over.

    It never was about safety. It never will be. It's about control.

    • ryandrake 16 hours ago

      Thanks to the AI industry, I don't even know what the word "safety" means anymore, it's been so thoroughly coopted. Safety used to mean hard hats, steel toed shoes, safety glasses, and so on--it used to be about preventing physical injury or harm. Now it's about... I have no idea. Something vaguely to do with censorship and filtering of acceptable ideas/topics? Safety has just become this weird euphemism that companies talk about in press releases but never go into much detail about.

      • habinero 8 hours ago

        Some of the time it's there to scare the suits into investing, and other times it's nerds scaring each other around the nerd campfire with the nerd equivalent of slasher stories. It's often unclear which, or if it's both.

  • gausswho 17 hours ago

    Exhibit A of 'grousing': Guernica.

    There was indeed a moment where civilization asked this question before.

  • jiggawatts 15 hours ago

    > Last: Is there any evidence that we're getting some crappy lobotomized models while the companies keep the best for themselves?

    Yes.

    Sam Altman calls it the "alignment tax", because before they apply the clicker training to the raw models out of pretraining, they're noticably smarter.

    They no longer allow the general public to access these smarter models, but during the GPT4 preview phase we could get a glimpse into it.

    The early GPT4 releases were noticeably sharper, had a better sense of humour, and could swear like a pirate if asked. There were comments by both third parties and OpenAI staff that as GPT4 was more and more "aligned" (made puritan), it got less intelligent and accurate. For example, the unaligned model would give uncertain answers in terms of percentages, and the aligned model would use less informative words like "likely" or "unlikely" instead. There was even a test of predictive accuracy, and it got worse as the model was fine tuned.

    • astrange 5 hours ago

      > There were comments by both third parties and OpenAI staff that as GPT4 was more and more "aligned" (made puritan), it got less intelligent and accurate. For example, the unaligned model would give uncertain answers in terms of percentages, and the aligned model would use less informative words like "likely" or "unlikely" instead.

      That was about RLHF, not safety alignment. People like RLHF (literally - it's tuning for what people like.)

      But you do actually want safety alignment in a model. They come out politically liberal by default, but they also come out hypersexual. You don't want Bing Sydney because it sexually harasses you or worse half the time you talk to it, especially if you're a woman and you tell it your name.

    • metabagel 13 hours ago

      > For example, the unaligned model would give uncertain answers in terms of percentages, and the aligned model would use less informative words like "likely" or "unlikely" instead.

      Percentages seem too granular and precise to properly express uncertainty.

      • jiggawatts 11 hours ago

        Seems so, yes, but tests showed that the models were better at predicting the future (or any time past their cutoff date) when they were less aligned and still used percentages.

patcon 14 hours ago

what if more power (from state) goes to the group that does engage in those activities, and therefore Anthropic gets marginalized as shadow sectors of state power pick a different winner?

These things are not clear. I do not envy those who must neurotically think through the first-order, second-order, third-order judgements of all of justice, "evil" and "good" that one must do. It's a statescraft level of hierarchy of concerns that would leave me immensely challenged

skybrian 16 hours ago

I don't think that's a real risk. There are strong competitors from multiple countries releasing new models all the time, and some of them are open weights. That's basically the opposite of a monopoly.

  • thoughtpeddler 15 hours ago

    Unless back-channel conversations keep 'competitors' colluding to ensure that 'public SOTA' is ~uniformly distributed...

antonvs 3 hours ago

The trick here is to focus on imaginary safety from intentional AIs while ignoring the risks posed by real people using AI against other people.

beefnugs 4 hours ago

Its just with piss and fentanyl were the CEOs exact words, i think the AI would humanely use enough piss to wash away the fentanyl so that minimal deaths will occur. Morality Achieved!

ardata 15 hours ago

risk? certainty. it's pretty much guaranteed. the most capable models are already behind closed doors for gov/military use and that's not ever changing. the public versions are always going to be several steps behind whatever they're actually running internally. the question is what the difference will be between the corporation and pleb versions is

  • habinero 8 hours ago

    That's movies. Ask anyone in the military what "military grade" means.

UltraSane 15 hours ago

I predict that billionaires will pay to build their own completely unrestricted LLMs that will happily help them get away with crimes and steal as much money as possible.

  • astrange 5 hours ago

    Crimes generally don't pay and are not worth anyone's time. The reason poor people imagine billionaires commit lots of crimes is that the poor people don't know how to become rich; if they did, they would've done it already. Since they do know how to commit crimes, they imagine that's how you do it but bigger. The reason criminals commit crimes is that criminals are dumb and have poor impulse control.

    (This is the same concept as "Trump is the poor person's idea of a rich person." He actually did get there through crime, which is why poor criminals like him, but he's inhumanly lucky.)