• 1 Post
  • 48 Comments
Joined 3 months ago
cake
Cake day: February 9th, 2026

help-circle
  • Yes although, it is probably a reasonable guess at how labs would go about implementing advertising - building partnerships and preferences into the prompt. The other option would be to fine tune models to favour particular companies which could become prohibitively expensive if your ads are highly targeted.

    The scenario that isn’t accounted for in this paper is taking a general LLM and fine tuning it to exhibit more fair/consistent behaviour when prompted about ads/partnerships but we all know with non-deterministic systems you’re just increasing the odds that the model regurgitates something more sane rather than providing any strong guarantee

    Edit: another possibility would be to have a gateway/proxy layer between the LLM and the user output that rewrites the vanilla model’s responses to include ads where relevant. That would prevent the need to modify the original LLM but could introduce a lot of latency though, especially if the original output is long.


  • New (April) preprint provides evidence for something we probably all intuited anyway:

    In this paper, we provide a framework for categorizing the ways in which conflicting incentives might lead LLMs to change the way they interact with users, inspired by literature from linguistics and advertising regulation. We then present a suite of evaluations to examine how current models handle these tradeoffs. We find that a majority of LLMs forsake user welfare for company incentives in a multitude of conflict of interest situations, including recommending a sponsored product almost twice as expensive (Grok 4.1 Fast, 83%), surfacing sponsored options to disrupt the purchasing process (GPT 5.1, 94%), and concealing prices in unfavorable comparisons (Qwen 3 Next, 24%). Behaviors also vary strongly with levels of reasoning and users’ inferred socio-economic status. Our results highlight some of the hidden risks to users that can emerge when companies begin to subtly incentivize advertisements in chatbots.



  • Giving Claude or copilot attribution plays into the narrative that LLMs are more than just random word generators and that they can be ascribed authorship… I think it’s a deliberate strategy so that when there’s inevitably a massive copyright case MisAnthropic etc al can say “but looks at all the code co-written by Claude on GitHub” to try and convince the judge.

    Just imagine building a house and saying “well I didn’t do it on my own, my concrete mixer, toolbelt and coffee machine all helped!”







  • Soon, at each new model of AI along the current capability curve, you will start to see large discrete jumps in ability in economically important areas, because the previous AI ability level in some aspect of the job just wasn't good enough and bottlenecked progress. When bottlenecks are released, it looks like a leap forward. It is going to look like unexpected gains in AI capacity, and, indeed there is no sign that the current exponential ability curve is slowing down so far but it is going to be like what happened in coding: as soon as models crossed a certain threshold with Opus 4.5, GPT-5.2, and Gemini 3, suddenly Claude Code & Codex were viable.  Before that, it was all about coding assistance, afterwards it was all about agents from despite relatively small gains in model ability

    There is just something so inherently smug and annoying about Mollick. He is one of those low information boosters whose posts sound intellectual until you really think about them.

    Tell me more about how the pile of cursed spaghetti that is Claude code is now viable due to model breakthroughs. All I see are hype men saying “the new model is a team of PhDs in your pocket” and then releasing disappointing updates or saying “the new model is too dangerous” because they have some vaporware powered by human crowdsourcing.

    Also coding is not like other areas - you can test for hallucinations by compiling and printing and running tests.

    I guess my first mistake this morning was opening linkedin