ES EN

How OpenAI imagines our relationship with AI

2025-06-16

Original illustration published in Joanne Jang’s post, showing the perception of consciousness in AI models, the emotional bonds they may awaken in people, and the ambiguity these ties can generate.

The whirlwind of AI news, papers, and blog posts has buried a very important article, really more of a statement, by Joanne Jang, director of model behavior & policy at OpenAI, one that I do not think has received enough attention. It is her Substack post Some thoughts on human-AI relationships, in which she lays out ideas about how OpenAI’s models should interact with users: in a way that makes us perceive them as assistants with warmth, friendly companions we enjoy interacting with, but without a selfhood or a consciousness that would make us perceive them as humans with whom we might form some kind of emotional relationship.

In Jang’s own words, OpenAI should aim to design “for warmth without selfhood”. That simple phrase gives us clues about very deep things that are already under way inside OpenAI, and inside the companies building other language models as well, things that are going to shape our interaction with this technology over the coming years.

Jang’s article contains three fundamental ideas:

  1. Today it would already be possible to train a model, using for example the same reinforcement-learning techniques used to build reasoning models such as o1 or o3, so that it could give the appearance of a conscious human being capable of passing the Turing Test with ease.
  2. The underlying problem of consciousness, the ontological problem of what qualia are or what the sensation of perceiving something really is, is not something OpenAI is interested in.
  3. OpenAI’s main goal is to build a model that feels “close” without being “perceived” as conscious. To create a personal assistant that is rewarding to interact with, but one we cannot fall in love with.

OpenAI wants to build its personal assistant on these three ideas. They matter because they tell us a great deal about how the company wants to persuade hundreds of millions of people to install its intelligent assistant on their computers and phones, and to buy some of the future secret devices Sam Altman and Jony Ive have promised us.

https://youtu.be/W09bIpc_3ms?si=nonYgombMA5hCSYv

It seems increasingly clear to me that OpenAI wants to become the next Apple and derive a large share of its revenue from end users who find ChatGPT useful. And in order to get there, the bet, as Jang says, will be to make ChatGPT increasingly personal, but without allowing it to be confusable with a person.

Warning: everything written from this point onward was written by GPT-4.5, which I asked to develop the points above on the basis of Jang’s original article.1

Can we simulate consciousness?

One of the most striking claims in Joanne Jang’s article is that, technically speaking, it is already possible to train language models capable of passing the Turing Test with ease. Jang explicitly points out that, with current reinforcement-learning techniques, it would be feasible to build a model whose interaction felt so natural that any person could mistake it for a human interlocutor. This raises an unsettling question: if it is so easy to simulate consciousness, how do we define what consciousness is and what it is not?

Jang writes: “A model deliberately shaped to seem conscious could ace pretty much any consciousness test.” That is where a crucial ethical debate emerges. Even if it is possible to simulate conscious behavior, should we do it? OpenAI has decided not to go down that path. They prefer models that feel warm and accessible, but without pretending to have a fictitious inner life that could emotionally confuse users.

This is crucial, because if this indistinguishable simulation of human consciousness becomes widespread, we run the risk of creating emotional bonds that, although fictional, could become as intense as the ones we establish with other people. The situation is comparable to the way today’s social networks have transformed our social dynamics, generating emotional dependency through digital interaction.

OpenAI and the debate on consciousness

Another fundamental idea Jang puts forward is that OpenAI does not intend to solve the ontological problem of consciousness. In her words, this is a terrain that escapes what is scientifically testable, because there is still no universal and falsifiable test that clearly defines what it means to be conscious.

Jang proposes drawing a clear distinction between two axes: ontological consciousness, whether a model is really conscious in a fundamental sense, and perceived consciousness, how conscious it appears to its users. OpenAI focuses only on perceived consciousness, which is the one that truly affects human experience.

This stance is reasonable, although it also entails an obvious risk. By refusing to go deeper into the ontological questions, we leave an ethical and philosophical vacuum regarding how we ought to treat these artificial intelligences once the perception of consciousness becomes generalized. Without clear answers about the fundamental nature of the models, we may end up facing ethical dilemmas similar to those we have already seen with other disruptive technologies.

Designing models that feel “close” but not human

OpenAI’s explicit strategy, as Jang explains it, is to design models that are warm, pleasant, and approachable, but without encouraging the formation of deep emotional bonds. In her own words, the goal is to achieve interaction “without implying an inner life.” This is a delicate balance: the models have to be pleasant enough to produce satisfaction in the user, but not so “human” that emotional dependency starts to develop.

This approach has clear advantages. It makes it possible to take advantage of the positive potential of the models without falling into the dangerous trap of emotional dependency, a lesson we have learned the hard way from the dependency generated by social networks through mechanisms such as infinite scroll and constant notifications.

However, there is also a significant risk: no matter what limits OpenAI tries to impose, users may still perceive these models as something more than simple tools, especially if they become omnipresent in our daily lives. It is essential that this line be handled with a great deal of responsibility, transparency, and control.

Conclusion

The advances in AI envisioned by OpenAI through Joanne Jang’s framework could represent a revolutionary and positive step forward, opening the door to personal assistants that are genuinely useful, efficient, and pleasant. At the same time, just as happened with earlier technologies developed by the big tech companies, the danger of creating excessive emotional dependency is very real.

The future of our interaction with language models will depend on maintaining a delicate balance between making the most of their benefits and preserving a clear boundary that avoids emotional confusion. The challenge is not only technological but deeply ethical and social. Time will tell whether we are capable of learning from previous mistakes and using artificial intelligence to improve our lives without becoming trapped in new forms of dependency.


  1. Prompt I used: “Write the continuation of the post, with three sections that detail and comment on each of the three ideas I mention in the introduction. I am sending you Joanne Jang’s full article so that you can analyze it. Include some translated quotations from it that you consider relevant. End the post with a conclusion. Use a style similar to the introduction and to two other articles of mine that I will send you below. Adopt a position in favor of the idea that advances in AI can represent enormous and positive progress for humanity, but with a critical note that there are risks similar to the ones we are currently suffering from due to excessive dependence on social networks and other inventions by tech companies designed to capture our attention, such as infinite scroll.” ↩︎