> We assert that artificial intelligence is a natural evolution of human tools developed throughout history to facilitate the creation, organization, and dissemination of ideas, and argue that it is paramount that the development and application of AI remain fundamentally human-centered.
While this is a noble goal, it seems obvious that this isn't how it usually goes.
For instance, "free market" is often used as a dogma against companies that are actively harmful to society, as "globalization" might be.
An unstoppable force, so any form of opposition is "luddite behavior".
Another one is easier transport and remote communication, that generally broke down the social fabric.
Or social media wreaking havoc among teen's minds.
From there, it's easy to see why the technological system might be seen as an inherent evil.
In 1872's Erewhon, Butler already described the technological system as a force that human society could contain as soon as it tolerated it.
There are already many companies persecuting their employees for not using AI enough, even when the employee's response is that the quality of its output is not good enough for the work at hand, rather than any ideological reason.
I'm neither optimistic nor pessimistic about the changes that AI might bring, but hoping it to become "human-centered" seems almost as optimistic as hoping for "humane wars".
> "free market" is often used as a dogma against companies that are actively harmful to society
This is a predominantly America-specific piece of propaganda, and it's pretty recent.
Adam Smith's ideas are primarily arguments against mercantilism (e.g. things like using tariffs to wield self-interested state power), something he showed to be against the common good. The "invisible hand" concept is used to show how self-interested action can, under conditions of *competitive markets*, lead to unintentional alignment with the common good.
Obviously that's a significant departure from the way it's commonly used today, where Thiel's book has influenced so many entrepreneurs into believing Monopolies are Good.
But the history of this is very Cold War-influenced, where "free markets" were politically positioned as alternatives to the USSR's "planned economy", and slowly pushed to depart further and further from Adam Smith's original argument about moral philosophy.
The government is a rare example of an extremely strong monopoly and not just a duopoly or a company holding significant marketshare. And yet people never seem to criticise it on those grounds despite it suffering from all of the same problems that corporate monopolies are accused of.
It has been funny to watch the rise of "China is beating us" rhetoric against the steady backdrop of "mercantilism is obsolete/bad" dogma, because the elephant in the room is that China has been running a textbook mercantilist playbook.
> Thiel's book has influenced so many entrepreneurs into believing Monopolies are Good.
Haven't read his book, but the idea that monopolies are good isn't typically made in a vacuum, it's made relative to alternatives, most often "ham-fisted government intervention". It's easier to take down a badly behaving monopoly than to change government, so believing monopolies are better than the alternatives seems like a decent heuristic.
Economic behavior is inherently game theoretic - agents take various actions and get some positive/negative reward as a result. Whether an agent's reward is positive or negative and of what magnitude, depends on the strategies employed by all agents. If some agents adopt new strategies, the reward calculus for everyone involved can completely change [1].
Over the past few centuries, countless new economic structures and strategies have been discovered and practiced. The rewards for the same action today and in the past can be completely different due to this.
So to me, if someone claimed more than a few decades ago that certain economic strategies and structures are good or bad, its simply not worth listening to them, unless someone reconfirms that the old finding still holds with the latest range of strategies. In that case, the credit and citation goes to that new someone, not the ghosts of the past.
Not even sure if AI was ever "human-centered" . DARPA funded a large amount of AI reseearch from even the 1960s. The DART tool used in the Gulf War made back all of their previous investments. [1]
I just re-read my comment (too late for edits) and there are a number of typos (including missing "not") that significantly degraded the syntax, but the point kind of came across anyway.
I skimmed the paper a couple of times, hoping to find the promised (from the abstract)
> pathway to integrating AI into our most challenging and intellectually rigorous fields to the benefit of all humankind.
There's very little insight here though. It seems mostly a retread of conversations we've been having in the academic community for a few years now. In particular, I was hoping to see some discussion of how we might restructure our educational institutions around this technology, when the machines rob students of the opportunity to develop critical thinking skills. Right now our best idea seems to be a retreat to oral and written examinations; an idea which doesn't scale and which ignores the supposed benefits of human+AI reasoning. The alternative suggestion I've seen is to teach prompt engineering, which seems (a) hard for foundational subjects and (b) again, seems to outsource much of the thinking to the AI, instead of extending the reach of human thought.
> We assert that artificial intelligence is a natural evolution of human tools.
While nowhere in the paper this is actually asserted but the abstract, a whiggish narrative of a genuinely unprecedented technology --such that it can replace and supersede human "labour" altogether (one is reminded of The Evolution of Human Science by Ted Chiang)-- sounds naive at best, dangerous at worst.
Software Engineer job openings for instance is at 2 year high (still far lower than covid dislocations though), but arguably all Enterprise AI was built or deployed in the last two years. We should have seen a crash in the job openings if the AI job replacement claim was correct.
Is there a better illustration of the power of UX than the fact that a messaging chat interface was able to set free all of human knowledge from copyright, whereas a bittorrent client couldn't?
In fact, the paper has an error in the argument that AI might find Fermat's theorem to be incorrect due to definition of natural numbers including a zero, because paper's version of a theorem explicitly says that the number should be greater than two, and zero cannot be greater than two.
Surprisingly, this mistake proves the author's point that human can implicitly understand what was said, and that it still has value to it, even if it's incorrect.
> This is an unabridged version of a solicited article for a forthcoming Blackwell Companion to the Philosophy of Mathematics. […] took over a year to write – which means, at the current pace of development in the field, that some of it is already slightly out of date.
It's not "the age of AI", it's just a Slop Decade.
And the tools did not become "exponentially sophisticated", one thing it's logarithmic, another is that the improvements are questionable. But "pervasive" - yes, granted.
93 comments
> We assert that artificial intelligence is a natural evolution of human tools developed throughout history to facilitate the creation, organization, and dissemination of ideas, and argue that it is paramount that the development and application of AI remain fundamentally human-centered.
While this is a noble goal, it seems obvious that this isn't how it usually goes. For instance, "free market" is often used as a dogma against companies that are actively harmful to society, as "globalization" might be. An unstoppable force, so any form of opposition is "luddite behavior". Another one is easier transport and remote communication, that generally broke down the social fabric. Or social media wreaking havoc among teen's minds. From there, it's easy to see why the technological system might be seen as an inherent evil. In 1872's Erewhon, Butler already described the technological system as a force that human society could contain as soon as it tolerated it. There are already many companies persecuting their employees for not using AI enough, even when the employee's response is that the quality of its output is not good enough for the work at hand, rather than any ideological reason.
I'm neither optimistic nor pessimistic about the changes that AI might bring, but hoping it to become "human-centered" seems almost as optimistic as hoping for "humane wars".
> "free market" is often used as a dogma against companies that are actively harmful to society
This is a predominantly America-specific piece of propaganda, and it's pretty recent.
Adam Smith's ideas are primarily arguments against mercantilism (e.g. things like using tariffs to wield self-interested state power), something he showed to be against the common good. The "invisible hand" concept is used to show how self-interested action can, under conditions of *competitive markets*, lead to unintentional alignment with the common good.
Obviously that's a significant departure from the way it's commonly used today, where Thiel's book has influenced so many entrepreneurs into believing Monopolies are Good.
But the history of this is very Cold War-influenced, where "free markets" were politically positioned as alternatives to the USSR's "planned economy", and slowly pushed to depart further and further from Adam Smith's original argument about moral philosophy.
> arguments against mercantilism
It has been funny to watch the rise of "China is beating us" rhetoric against the steady backdrop of "mercantilism is obsolete/bad" dogma, because the elephant in the room is that China has been running a textbook mercantilist playbook.
> Thiel's book has influenced so many entrepreneurs into believing Monopolies are Good.
Haven't read his book, but the idea that monopolies are good isn't typically made in a vacuum, it's made relative to alternatives, most often "ham-fisted government intervention". It's easier to take down a badly behaving monopoly than to change government, so believing monopolies are better than the alternatives seems like a decent heuristic.
Over the past few centuries, countless new economic structures and strategies have been discovered and practiced. The rewards for the same action today and in the past can be completely different due to this.
So to me, if someone claimed more than a few decades ago that certain economic strategies and structures are good or bad, its simply not worth listening to them, unless someone reconfirms that the old finding still holds with the latest range of strategies. In that case, the credit and citation goes to that new someone, not the ghosts of the past.
[1] A good interactive demo https://ncase.me/trust/
[1] https://en.wikipedia.org/wiki/Dynamic_Analysis_and_Replannin...
> pathway to integrating AI into our most challenging and intellectually rigorous fields to the benefit of all humankind.
There's very little insight here though. It seems mostly a retread of conversations we've been having in the academic community for a few years now. In particular, I was hoping to see some discussion of how we might restructure our educational institutions around this technology, when the machines rob students of the opportunity to develop critical thinking skills. Right now our best idea seems to be a retreat to oral and written examinations; an idea which doesn't scale and which ignores the supposed benefits of human+AI reasoning. The alternative suggestion I've seen is to teach prompt engineering, which seems (a) hard for foundational subjects and (b) again, seems to outsource much of the thinking to the AI, instead of extending the reach of human thought.
> We assert that artificial intelligence is a natural evolution of human tools.
While nowhere in the paper this is actually asserted but the abstract, a whiggish narrative of a genuinely unprecedented technology --such that it can replace and supersede human "labour" altogether (one is reminded of The Evolution of Human Science by Ted Chiang)-- sounds naive at best, dangerous at worst.
> Today, unlike in the Luddites’ time, we are already seeing skilled workers replaced not with lower-wage human labor, but with AI.
To me this is the weakest claim of the article. This claim been thrown around endlessly without proof.
https://fred.stlouisfed.org/series/IHLIDXUSTPSOFTDEVE
Software Engineer job openings for instance is at 2 year high (still far lower than covid dislocations though), but arguably all Enterprise AI was built or deployed in the last two years. We should have seen a crash in the job openings if the AI job replacement claim was correct.
This is something I've spend some time thinking about (personally written article, not AI slop): https://www.signalbloom.ai/posts/why-task-proficiency-doesnt...
Surprisingly, this mistake proves the author's point that human can implicitly understand what was said, and that it still has value to it, even if it's incorrect.
In particular:
> This is an unabridged version of a solicited article for a forthcoming Blackwell Companion to the Philosophy of Mathematics. […] took over a year to write – which means, at the current pace of development in the field, that some of it is already slightly out of date.
----
Edit: The post at https://mathstodon.xyz/@tao/116319186983426174 mentions also an (entirely unrelated) popular-math presentation titled “What does it mean to think like a mathematician?” https://terrytao.wordpress.com/wp-content/uploads/2026/03/ta... which is interesting too (despite the ChatGPT-generated illustrations and repeating stuff Tao has said before, on his blog etc.)
https://www.youtube.com/watch?v=zJvuaRVc8Bg
And the tools did not become "exponentially sophisticated", one thing it's logarithmic, another is that the improvements are questionable. But "pervasive" - yes, granted.