It happened. The switch flipped and I‘m now one of those people who believe that it‘s far more productive to design in code than to move boxes and text in some design software.
I spent the last decade not wanting to believe the people who praised designers who code, but I‘m convinced now.
It‘s been about a year since I worked through 100 Days of SwiftUI. I built four iOS apps and about 4-5 web projects using React since then. I‘m obviously still a coding-baby but it‘s already very clear to me that being able to code made me a better designer.
AR interfaces are going to take this up a notch.
Three years ago, when I had an epiphany and realized that AR/VR interfaces are going to be the future of computing, I wondered how current design software would ever be able to allow me to do a good job designing AR interfaces.
I came to the conclusion that it wouldn‘t. It couldn’t.
Continue reading “Being a designer who doesn‘t code might make you a bad designer in 2024 and forward”
There’s this concept of self-efficacy in psychology that really resonates with me. I see a lot of life through this lens. Here’s Wikipedia’s definition:
In psychology, self-efficacy is an individual’s belief in their capacity to act in the ways necessary to reach specific goals. […] A strong sense of self-efficacy promotes human accomplishment and personal well-being. A person with high self-efficacy views challenges as things that are supposed to be mastered rather than threats to avoid.
I believe that you can choose to be self-efficacious and things you do can make you feel self-efficacious. Most people fail to recognize when these moments occur, and even fewer make a conscious effort to intentionally create such moments for themselves.
Changing the physical world does this to humans. That’s one of the reasons so many people daydream about gardening and why pottery feels a bit like therapy. You create something that wasn’t there before. You moved something and it stayed in place. You’ve literally made a teeny-tiny dent in the universe.
You won’t be able to describe to a person who never experienced anything like it, how gardening makes you feel. Starting with nothing, spending hours of work, accepting failures and imperfections to then see a result of something you made, tickles the core of what we are. Sure, you can explain all the steps of the process and tell them you felt “good” doing so but there’s no way to describe the intensity of that feeling.
Turning a rotary dial to call someone, pressing buttons to control a SNES video game character and swiping and tapping on glass to send an email did this to us with ever increasing amounts of directness. Every evolution of digital technology helped us feel more self-efficacy.
For better or worse.
I think AR interfaces are the inevitable next step in computing because they make us feel more self-efficacious. You won’t be able to properly describe how moving digital windows in the physical space of real life made you feel. It’s counter-intuitive to even think that the way you interact with the window your bursting inbox makes a difference, yet it does.
Spatial computing can’t be described. It must be felt to be understood.
I can’t get over the fact that Apple introduced an AR headset yesterday. Everybody, including myself, expected a VR headset with AR features. However, Vision Pro is clearly AR first with VR as an optional mode you deliberately have to activate if you feel the need for it.
It even seems like they nailed the interaction paradigm. Based on my experience with current VR (lame) headsets I wasn’t very happy with how hand tracking worked. You have to hold your hands in front of you for the sensors to see them properly. Apple knowingly showed people in several situations lounging on their couches with their hands by their sides, only making tiny gestures to interact with what’s being looked at.
If it were possible, I would order a Vision Pro right now.