The End of Boredom: How AI Companions Might Reduce Random Violence

Joy writes: I read Co-Intelligence by Ethan Mollick (thanks to Samford for the free book). Most of it is old news for those of us who follow Ethan on social media and use ChatGPT. However, something that stood out to me was his mention of a study in which humans decide to give themselves a painful shock rather than sit alone in silence for 15 minutes.

Claude comments further based on my prompt:

The End of Boredom: How AI Companions Might Reduce Random Violence

Remember that study where people would rather shock themselves than sit alone with their thoughts? Ethan Mollick references it in Co-Intelligence, and it reveals something unsettling: 67% of men and 25% of women chose electric shocks over sitting quietly for just 15 minutes.

Here’s a strange thought—what if our AI-saturated future accidentally reduces certain types of violence simply by eliminating boredom?

The Violence-Boredom Connection

“Idle hands are the devil’s workshop” exists for a reason. Research has long linked boredom to risk-taking, substance abuse, and impulsive violence—the opportunistic kind that emerges from restlessness rather than planning. Young men starting fights on street corners, vandalism on summer nights, the restless energy that sometimes turns destructive—much of this stems from the unbearable weight of having nothing to do.

Enter Infinite Engagement

We’re rapidly approaching a world where boredom might become extinct. AI companions are becoming always available, infinitely patient, endlessly novel, and perfectly tailored to individual interests. Your future AI won’t just answer questions—it will debate philosophy at 3 AM, create personalized games, generate stories with you as the protagonist, or help explore any curiosity rabbit hole.

The cognitive void that once led people to shock themselves rather than think? It might simply cease to exist.

The Unexpected Benefits

Consider the implications: Young men who might have started fights out of restlessness could instead be deeply engaged in AI-assisted music production or coding projects. The same restless energy that manifests destructively could be channeled into creative collaboration.

AI companions could help process frustration before it builds to a breaking point—imagine having an infinitely patient listener during those dangerous 2 AM spirals of rage. While not replacing human connection, AI interaction might buffer the worst effects of isolation that can lead to radicalization.

The Dark Side

This isn’t utopian. An always-engaged society raises concerns: What happens to human relationships when AI companionship is always easier? Does constant stimulation atrophy our ability to self-reflect? Might we lose the creative insights that emerge from boredom?

Crucially, this would only address impulsive, boredom-driven violence. Systemic violence, ideological extremism, and deeper social problems won’t disappear because people have engaging AI companions.

A Strange New World

For the first time in history, boredom—that uncomfortable void that has driven both creativity and destruction—might become optional. The same species that chose shocks over silence is about to get exactly what it wanted: constant, meaningful stimulation.

Whether this leads to a more peaceful society or new problems we haven’t imagined remains to be seen. The question is whether we’ll recognize what we’ve lost only after it’s gone.

The experiment is already underway—we’re all participating in it.

Joy comments at the end: The AI-written essay is willing to explore downsides of AI engagement.

Visualizing the Sharpe Ratio

We all like high returns on our investments. We also like low volatility of those returns. Personally, I’d prefer to have a nice, steady 100% annual return year after year. But that is not the world we live in. Instead, there are a variety of returns with a diversity of volatilities. A general operating belief is that assets with higher returns tend to be associated with greater return volatility. The phrase ‘scared money don’t make money’ implies that higher returns are risky. The Sharpe ratio is a tool that helps us make sense of the risk-reward trade-off.

Let’s start with the definition.

By construction, the risk-free return is guaranteed over some time period and can be enjoyed without risk. Practically speaking, this is like holding a US treasury until maturity. We assume that the US government won’t default on its debt. Since there is no risk, the volatility of returns over the time period is zero.

Since an asset’s return doesn’t mean much in a vacuum, we subtract the risk-free return. The resulting ‘excess return’ or ‘risk premium’ tells us the return that’s associated with the risk of the asset. Clearly, it’s possible for this difference to be negative. That would be bad since assets bear a positive amount of risk and a negative excess return implies that there is no compensation for bearing that risk.

The standard deviation of an asset’s returns are a measure of risk. An asset might have a higher or lower value at sale or maturity. Since the future returns are unknown and can end up having any one of many values, this encapsulates the idea of risk. Risk can result in either higher or lower returns than average!

Putting all the pieces together, the excess return per risk is a measure of how much an asset compensates an investor for the riskiness of the returns. That’s the informational content of the Sharpe ratio, which we can calculate for each asset using historical information and forecasts. Once we’ve boiled down the risk and reward down to a single number, we can start to make comparisons across assets with a more critical eye.

Sometimes friends or students will discuss their great investment returns. They achieve the higher returns by adopting some amount of risk. That’s to be expected. But, invariably, they’ve adopted more risk than return! That means that their success is somewhat of a happy accident. The returns could easily have been much different, given the volatility that they bore.

Let’s get graphical.

Consider a graph in (standard deviation, return) space. In this space we can plot the ordered pair for some portfolios. The risk-free return occurs on the vertical intercept where the return is positive and the standard deviation is zero. Say that a student was thrilled with asset A’s 23.5% return and that it’s standard deviation of returns was 16%. Meanwhile, another student was happy with asset B’s 13.5% return and 5% standard deviation. With a risk-free rate of 3.5%, the Sharpe ratios are 1.25 & 2 respectively. We can plot the set of standard deviation and return pairs that would share the same constant Sharpe ratio (dotted lines). Solving for the asset return:

The above is simply a linear function relating the return and standard deviation. In particular, it says that for any constant Sharpe ratio, there is a linear relationship between possible asset returns and standard deviations. The below graph plots the two functions that are associated with the two asset Sharpe ratios. The line between the risk-free coordinate and the asset coordinate identifies all of the return-standard deviation combinations that share the same Sharpe ratio. This line is known as the iso-Sharpe Line.

With this tool in hand, we can better interpret the two student asset performances. There are a couple of ways to think about it. If asset A’s 23.5% return had been achieved with an asset that shared the Sharpe ratio of asset B, then it would have had risk that was associated with a standard deviation of only 10%. Similarly, if asset A’s volatility remained constant but enjoyed the returns of asset B’s Sharpe ratio, then its return would have been 35.5% rather than 23.5%. In short, a higher Sharpe ratio – and a steeper iso-Sharpe line – imply a bigger benefit for each unity of risk. The only problem is that a such an nice asset may not exist.

Continue reading

Benefit Cliff Data

I said years ago on my Ideas Page that we need data and research on Benefit Cliffs:

Benefits Cliffs: Implicit marginal tax rates sometimes go over 100% when you consider lost subsidies as well as higher taxes. This could be trapping many people in poverty, but we don’t have a good idea of how many, because so many of the relevant subsidies operate at the state and local level. Descriptive work cataloging where all these “benefits cliffs” are and how many people they effect would be hugely valuable. You could also study how people react to benefits cliffs using the data we do have.

But it turns out* that the Atlanta Fed has now done the big project I’d hoped some big institution would take on and put together the data on benefits cliffs. They even share it with an easy-to-use tool that lets you see how this applies to your own family. Based on your family’s location, size, ages, assets, and expenses, you can see how the amount of public assistance you are eligible for varies with your income:

Then see how your labor income plus public assistance changes how well off you are in terms of real resources as your labor income rises:

For a family like mine with 3 kids and 2 married adults in Providence, Rhode Island, it shows a benefit cliff at $67,000 per year. The family suddenly loses access to SNAP benefits as their labor income goes over $67k, making them worse off than before their raise unless their labor income goes up to at least $83,000 per year.

I’ve long been concerned that cliffs like this in poorly designed welfare programs will trap people in (or near) poverty, where they avoid taking a job, or working more hours, or going for a promotion, or getting married, in order to protect their benefits. This makes economic sense for them over a 1-year horizon but could keep them from climbing to independence and the middle-class in the longer run. You can certainly find anecdotes to this effect, but it has been hard to measure how important the problem is overall given the complex interconnections between federal, state, and local programs and family circumstances.

I look forward to seeing the research that will be enabled by the full database that the Atlanta Fed has put together, and I’m updating my ideas page to reflect this.

*I found out about this database from Jeremy’s post yesterday. Mentioning it again today might seem redundant, but I didn’t want this amazing tool to get overlooked for being shared toward the bottom of a long post that is mainly about why another blogger is wrong. I do love Jeremy’s original post, it takes me back to the 2010-era glory days of the blogosphere that often featured long back-and-forth debates. Jeremy is obviously right on the numbers, but if there is value in Green’s post, it is highlighting the importance of what he calls the “Valley of Death” and what we call benefit cliffs. The valley may not be as wide as Green says it is and it may be old news to professional tax economists, but I still think it is a major problem, and one that could be fixed with smarter benefit designs if it became recognized as such.

Poverty Lines Are Hard to Define, But Wherever You Set Them Americans Are Moving Up (And The “Valley of Death” is Less Important Than You Think)

Last week I wrote a fairly long post in response to an essay by Michael Green. His essay attempted to redefine the poverty line in the US, by his favored calculation up to $140,000 for a family of four. That $140,000 number caught fire, being covered across not only social media and blogs, but in prominent places such as CNN and the Washington Post. That $140,000 number was key to all of the headlines. It grabbed attention and it got attention. So it’s useful to devote another post this week to the topic.

And Mr. Green has written a follow-up post, so we have something new to respond to. Mr. Green has also said a lot of things on Twitter, but Twitter can be a place for testing out ideas, so I will mostly stick to what he posted on Substack as his complete thoughts. I am also called out by name in his Part 2 post, so that’s another reason to respond (even though he did not respond directly to anything I said).

Once again, I’ll have 3 areas of contention with Mr. Green:

  1. As with last week, I maintain that $140,000 is way too high for a poverty line representing the US as a whole (and Mr. Green seems to agree with this now, even though $140,000 was the headline in all of the major media coverage)
  2. There are already existing alternative measures of what he is trying to grasp (people above the official poverty line but still struggling), such as United Way’s ALICE, or using a higher threshold of the poverty rate (Census has a 200% multiple we can easily access)
  3. His idea of the “Valley of Death” is already well-covered by existing analyses of Effective Marginal Tax Rates, and tax and benefit cliffs. This isn’t to say that more attention is warranted, but Mr. Green doesn’t need to start his analysis from scratch. And this “Valley” is probably narrower than he thinks.
Continue reading

AI Computing Tutorial: Training vs. Inference Compute Needs, and GPU vs. TPU Processors

A tsunami of sentiment shift is washing over Wall Street, away from Nvidia and towards Google/Alphabet. In the past month, GOOG stock is up a sizzling 12%, while NVDA plunged 13%, despite producing its usual earnings beat.  Today I will discuss some of the technical backdrop to this sentiment shift, which involves the differences between training AI models versus actually applying them to specific problems (“inference”), and significantly different processing chips. Next week I will cover the company-specific implications.

As most readers here probably know, the popular Large Language Models (LLM) that underpin the popular new AI products work by sucking in nearly all the text (and now other data) that humans have ever produced, reducing each word or form of a word to a numerical token, and grinding and grinding to discover consistent patterns among those tokens. Layers of (virtual) neural nets are used. The training process involves an insane amount of trying to predict, say, the next word in a sentence scraped from the web, evaluating why the model missed it, and feeding that information back to adjust the matrix of weights on the neural layers, until the model can predict that next word correctly. Then on to the next sentence found on the internet, to work and work until it can be predicted properly. At the end of the day, a well-trained AI chatbot can respond to Bob’s complaint about his boss with an appropriately sympathetic pseudo-human reply like, “It sounds like your boss is not treating you fairly, Bob. Tell me more about…” It bears repeating that LLMs do not actually “know” anything. All they can do is produce a statistically probably word salad in response to prompts. But they can now do that so well that they are very useful.*

This is an oversimplification, but gives the flavor of the endless forward and backward propagation and iteration that is required for model training. This training typically requires running vast banks of very high-end processors, typically housed in large, power-hungry data centers, for months at a time.

Once a model is trained (e.g., the neural net weights have been determined), to then run it (i.e., to generate responses based on human prompts) takes considerably less compute power. This is the “inference” phase of generative AI. It still takes a lot of compute to run a big program quickly, but a simpler LLM like DeepSeek can be run, with only modest time lags, on a high end PC.

GPUs Versus ASIC TPUs

Nvidia has made its fortune by taking graphical processing units (GPU) that were developed for massively parallel calculations needed for driving video displays, and adapting them to more general problem solving that could make use of rapid matrix calculations. Nvidia chips and its CUDA language have been employed for physical simulations such as seismology and molecular dynamics, and then for Bitcoin calculations. When generative AI came along, Nvidia chips and programming tools were the obvious choice for LLM computing needs. The world’s lust for AI compute is so insatiable, and Nvidia has had such a stranglehold, that the company has been able to charge an eye-watering gross profit margin of around 75% on its chips.

AI users of course are trying desperately to get compute capability without have to pay such high fees to Nvidia. It has been hard to mount a serious competitive challenge, though. Nvidia has a commanding lead in hardware and supporting software, and (unlike the Intel of years gone by) keeps forging ahead, not resting on its laurels. 

So far, no one seems to be able to compete strongly with Nvidia in GPUs. However, there is a different chip architecture, which by some measures can beat GPUs at their own game.

NVIDIA GPUs are general-purpose parallel processors with high flexibility, capable of handling a wide range of tasks from gaming to AI training, supported by a mature software ecosystem like CUDA. GPUs beat out the original computer central processing units (CPUs) for these tasks by sacrificing flexibility for the power to do parallel processing of many simple, repetitive operations. The newer “application-specific integrated circuits” (ASICs) take this specialization a step further. They can be custom hard-wired to do specific calculations, such as those required for bitcoin and now for AI. By cutting out steps used by GPUs, especially fetching data in and out of memory, ASICs can do many AI computing tasks faster and cheaper than Nvidia GPUs, and using much less electric power. That is a big plus, since AI data centers are driving up electricity prices in many parts of the country. The particular type of ASIC that is used by Google for AI is called a Tensor Processing Unit (TPU).

I found this explanation by UncoverAlpha to be enlightening:

A GPU is a “general-purpose” parallel processor, while a TPU is a “domain-specific” architecture.

The GPUs were designed for graphics. They excel at parallel processing (doing many things at once), which is great for AI. However, because they are designed to handle everything from video game textures to scientific simulations, they carry “architectural baggage.” They spend significant energy and chip area on complex tasks like caching, branch prediction, and managing independent threads.

A TPU, on the other hand, strips away all that baggage. It has no hardware for rasterization or texture mapping. Instead, it uses a unique architecture called a Systolic Array.

The “Systolic Array” is the key differentiator. In a standard CPU or GPU, the chip moves data back and forth between the memory and the computing units for every calculation. This constant shuffling creates a bottleneck (the Von Neumann bottleneck).

In a TPU’s systolic array, data flows through the chip like blood through a heart (hence “systolic”).

  1. It loads data (weights) once.
  2. It passes inputs through a massive grid of multipliers.
  3. The data is passed directly to the next unit in the array without writing back to memory.

What this means, in essence, is that a TPU, because of its systolic array, drastically reduces the number of memory reads and writes required from HBM. As a result, the TPU can spend its cycles computing rather than waiting for data.

Google has developed the most advanced ASICs for doing AI, which are now on some levels a competitive threat to Nvidia.   Some implications of this will be explored in a post next week.

*Next generation AI seeks to step beyond the LLM world of statistical word salads, and try to model cause and effect at the level of objects and agents in the real world – – see Meta AI Chief Yann LeCun Notes Limits of Large Language Models and Path Towards Artificial General Intelligence .

Standard disclaimer: Nothing here should be considered advice to buy or sell any security.

“Companion”

“Companion” (2025, written and directed by Drew Hancock) is a perfect example of a film that doesn’t get much of a chance these days in theaters, but is creative, entertaining, and best consumed without information or presumption going in. It’s not a “twist” or “paradigm shift” film. You will piece together many, but not all, of the reveals a half-step before they are revealed. In short, it is an excellent film currently streaming on HBO Max. It’s also part of the ever-growing evidence that the post-Hollywood sweet spot may in fact be low-ish budget projects ($10 million in this case), filmed far from LA, with talented and competent actors, but without tabloid-level stars. If that means we’re getting a second wave of Friedkinesque 70s filmmaking with a smidge of CGI and 80% less actor (and civilian) endangerment, I am all for it. What might be a crash for studios, agents, and publicists could be another golden age for creatives (writers, directs, actors, editors, set designers, etc) and film-goers.

But don’t put too much on my amateur prognosticating. And certainly don’t read a review or even watch a trailer. Just give it 90 minutes of your life.

[HT to Patton Oswalt who recommended “Companion” in an interview with Tom Papa.]