With the advent of the internet and the virtual world, scarcity is one aspect from the physical dimension many had hoped to leave behind. But alas, scarcity has meandered its way into the digital realm, and introduced another barrier into what could be an endless plane of possibility. It started out as Digital Rights Management, to ensure fair compensation for content, but quickly exploded into crypto currency, essentially the most sugar coated form of manufactured scarcity. So the question is, will scarcity put a cap the limitless possibility of the internet, or will it tear it open further?
Why do we wear masks? There is an inherent seduction that comes packaged with the mystique of a secret identity. We feel empowered by secrecy, it is like a drug, making us feel as if we can do anything, because no one is looking. The KKK used animosity as a weapon through hiding behind a mask and committing treacheries under the backdrop of the moon and acting like different men with the sun’s rise. Technology, like a mask, empowers us through secrecy, we can attack, hurl vicious insults and destroy people with a keyboard and mouse, without them ever seeing our face. Virtual secrecy is a danger often left unchecked, as in the past, fear of public humiliation or shame for attacking others, verbally or physically, has kept us in line. But now, technology enables the deepest levels of secrecy, and people use this to their advantage to wage virtual war with bombs and planes of derogatory remarks and insults. So the question is, will we become less sensitive, or more power high?
Fractal geometry, to put it simply, is the fourth dimension, it is supplementary to the first three or “euclidian” dimensions, length, width, and height, and it describes the space in between the first three dimensions. For example, a euclidian definition of a mountain would be a cone, as that can easily be translated from its length, width, and height, whereas a “fractallian” definition would describe it based on its components. This is done as pertaining to the three principles of fractals: self-similarity, recursiveness, and initiation, which state respectively, that each component of a fractal is similar to the whole fractal, and that the self-similarity is infinite in detail, going on forever. So how do I go from writing about iPads and the future of computing to this, I have to state that such a jump is not that far of one. You see, in my mind, fractals hold the key to the future of computing. A key element of fractals is their property of infinite detail in a limited amount of space, an element that could be translated into computing quite clearly. The use of standardized data utilizing minuscule variations to such data could allow for self similar binary (1s and 0s for the non tech savvy) that could be infinite in detail while still being finite in the amount of space it takes up. This would allow for infinite storage and memory, which would simultaneously expedite the convergence of memory and storage, and null another limitation to compute power. However this concept fringes on the translation of Mandelbrot’s formula (Z [the complete fractal, think the complete triangle from pascals triangle]= Z * C^2 [The little triangles multiplied by their amount]) to binary, which would be done in a similar way into its translation into color, where an integer for Z translates to a specified color as it inches towards infinity and a rational number with a decimal translates to black as it inches towards zero. Once this could be translated, a major barrier in computing could be broken, and major technological leaps could be made.
10 years ago on this day, the post pc era began. At least according to Steve Jobs it did. 10 years ago, Steve Jobs sauntered onto a San Francisco stage to announce possibly the most anticipated product in his company’s history: the iPad. To Jobs, the iPad was more than just a tablet, it was the final nail in the PC’s coffin and the first glimpse at the future of computing. But 10 years after its unveil, many are still hesitant to call the iPad a computer. This has nothing to do with the iPads functionality today, and everything to do with the iPads functionality 10 years ago. When the iPad first came out, it seemed to be focused on content consumption. Its reveal keynote highlighted its potential as an e reader, or its video streaming prowess, rather than paying attention to applications in writing, business environments or content creation, all fields that were dominated by the PC. While its obvious why the iPad was so centered around content consumption, with the tempting ease of advertising the device as a portable tv or a slate that lets you carry the Library of Alexandria in your backpack, these weren’t the features people were looking for out of a tablet, they wanted Word, Photoshop, GarageBand, and other apps previously exclusive to PCs and Macs. This association between the iPad and content consumption is still strong in the minds of many today, and is the key reason why so many people have such a difficult time seeing the iPad as a computer in the same way they see their laptop or desktop as one, where in reality the iPads of today are entirely different beasts than the first ten years ago. Modern iPads check all the boxes of a computer, they have physical keyboards, file managers, productivity and content creation applications, large screens, and other features key to a computer replacement, so yes, the iPad can be called a computer, and it should be, but that doesn’t mean that everyone will.
Apple is known for making its devices as simple as possible by taking away what they deem to be unnecessary additions. This policy goes all the way back to the Mac, which didn’t feature arrow keys to push users to use the mouse. More recently, it can be seen in the removal of the headphone jack on the iPhone and the narrow port selection on current MacBooks. While this practice has come under fire, it has allowed for the adoption of more futuristic technologies, such as the aforementioned mouse or wireless earbuds like Apples own AirPods, which wouldn’t have been nearly as popular as they are if you didn’t have to use a dongle to connect regular wired headphones. However, we might not be ready for the next step in this line of removing complications from our devices. According to multiple reports, the one with the most weight coming from CNBC, Apple could ditch the lightning port for wireless charging only power delivery on the next iPhone. Such a move would certainly be true to Apple’s history, but are we ready yet? Wireless charging has seen some widespread adoption, and the technology has some clear benefits over the conventional charging over wire, such as being easier to charge your device, easier to remove it from a charger, and more sleek design. It’s also important to note that, unlike the case of the removal of the headphone jack, the removal of the charging port would not be that big of a loss. With the massive batteries in iPhones these days that allow them to last all day without needing a charge, one of the major advantages of conventional charging is eliminated, that being the portability. You don’t really need a charging port when you have a wireless charger, which will presumably be included with the new iPhone in place of a conventional charging cable and brick, and when you don’t need to top off your phone, you wouldn’t need a charging brick and cable to bring with you. Sure, when going on road trips you might need a wired charger to charge your phone, but I foresee new cars coming with built in wireless charging pads and for the right now, many companies make portable wireless chargers. I think, just like in the case of the arrow keys, headphone jack, any countless other examples, taking something away would give be giving you an experience that is that much greater.
Last week the Consumer Electronics Show-Or CES for short-was held in Las Vegas. Tons of future tech products were shown off, from 5G phones, to 8K TVs, to self-driving cars, but one technology showed up the most: foldables. Dell, Huawei, and HP all showed off new foldable phones, laptops, and tablets. So if this is the future of consumer electronics, why hasn’t Apple made any foldable yet? Have they lost their edge over their competitors? Have they fallen behind the pack? The answer to both of these questions: No. The reason why Apple hasn’t made any folding iPhones, iPads, or MacBooks when every other tech company has products using this technology yet is not because every other tech company is smarter than them, but because they are smarter than every other tech company. Right now, folding phones suck. Apple knows this. They saw what happened with the Galaxy Fold, how the review units broke within days of journalists receiving them, everyone did. Apple knows that folding technology isn’t ready yet, and they know that, if they want to make a truly great folding product, they need it to be. Apple has done this before, when they were developing the original iMac, Steve Jobs refused to use a tray loading disk drive, even though cd-burning drives didn’t come in the slot loading form factor yet. Jobs knew it was better to have a better user experience, one where users didn’t have to press a button to insert and play disks, than give them the functionality of being able to burn cds. This core appreciation for the user experience to the extent of the loss of some functionality is the core foundation of Apple today, what sets it apart from Dell, HP, Lenovo, and the rest of the pack, and the policy that has made them the most popular-and most profitable-tech company in the world.
The year is 1984, IBM has a tight grasp over the personal computer market, and many think it will stay that way. But then, Apple, a fledgling company with a small but loyal group of users, introduces a new personal computer, one that seeks to dethrone Big Blue as the king of the PC, and its called the Mac. Small, well designed, and friendly looking, the original Mac didn’t look like anything else in its league, which was filled with ugly beige boxes and towers. While it wasn’t their first computer, the Mac put Apple on the map, and was there their main product line for a majority of their time as a company. The iMac, the iPod, even the iPhone, iPad, and Apple Watch, none of these would be possible without the first Mac. But now, 35 Years after Steve Jobs walked onstage and unveiled the computer that would change the world, the purpose and the future of the Mac is as unclear as ever. A product line that used to carry the weight of the company is now relegated to a side thought. Sure, Apple has paid more attention to the Mac when it comes to pro products, but their focus on consumer computers is set on the iPad. Furthermore, having two lines of personal computers is pretty confusing for customers to navigate through, making it more difficult for users to decide between an iPad, arguably the better personal computer and the more forward thinking product, and the MacBook, the faulty keyboard adorned, weaker product of the two. To me, it seems that the future of the Mac is that of a professional device, for programmers, filmmakers, and other creative professionals, while the iPad serves the purpose of the consumer computer. This would fix Apple’s confusion problem, while giving them a strong foot in both the consumer and professional markets, allowing them to reach the widest range of customers while retaining the amazing user experience that came with the original Mac, way back in 1984.
Perhaps the most outspoken criticism of Apple in this day and age targets the price of their products. First it was the iPhone, then the iPad Pro, most recently, the Mac Pro. The second Apple releases a new product, someone is already making a twitter post about its “unbearable” price tag. These calls for cheaper products often go unanswered, isolated in the echo chambers that are Twitter and Facebook groups. But according to recent rumors, Apple could finally be listening to calls for more affordable products. According to a rumor first first publicized by Bloomberg, Apple could be working on a cheaper iPhone, rumored to be called the iPhone 9, that would serve as a successor to the popular iPhone SE, which was a more affordable offering from a few years ago. The new model would reportedly go back to the older home button design, last seen on the iPhone 8, with lesser specs and a smaller screen to keep costs down. But while introducing a cheaper iPhone sounds like music to some people’s ears, there are several less clear problems that could arise with such a move. The first pertains to their product lineup. As I have previously stated, Apple’s current product lineup is in pretty rough shape when compared those of previous years. Each product line has way to many devices in it, each with names that mean almost nothing and say almost nothing about the product, I mean, on their website, they’re selling four different iPhones, four different iPads, and four different Mac desktops. I’m all for covering all your bases, but making as many products as you can is the not the way of going about it, for each new product that you add, you lose some focus, and you make buying products harder for consumers, while simultaneously hindering the user experience of products as you make them more specialized and less general. This brings me to my biggest concern about introducing a cheaper iPhone. A lesser user experience. Steve Jobs understood the importance of imputing, or giving a good first impression, to your customers. This understanding led to the amazing user experience that built the foundation for Apple to become the most valuable company in the world. When you introduce cheaper products, with a more limited set of features and worse user experience, you impute a lesser image to your customers. I understand the need to reach users through financial accessibility, and I can see the business benefits of this move, but I think the effect that it would have on the brand and brand image as a whole far outweighs the benefits of a little bit more marketshare and a little bit more money to talk about at your annual report.
Before iTunes, the only legal way to download music onto your computer was through subscription services. But these subscription services were not like the Spotifys and Pandoras of today. No, they truly sucked. They were clunky, hard to navigate, and lacked a lot of popular songs as they were often produced by the record labels. In people’s eyes however, the biggest problem with these subscription services was that they didn’t own the music they were paying for. Steve Jobs understood this, and made owning the music you purchased the foundation for iTunes, on of Apple’s most successful products. But in 2019, subscription services rule the market, and only a small percentage of people still buy their music. So what changed? Was it us? Was Steve Jobs wrong? No. What really made iTunes so much better than all the streaming services it was contemporary to was not that you owned the music sold on it, but the connection you felt with the music for owning it. Today, music streaming services replicate this connection through user profiles, which allow us to define ourselves through the music we listen to and the playlists we make. We have this connection with the music on these services not because of whether or not we own it, but because of the things we can do with it. We can make playlists and share them with the world, and that makes the music feel like it is truly ours. User profiles that suggest music based on our tastes contribute to this connection as well. We feel connected because we are, things that are suggested to us are based on what we listen to, and our user profiles, playlists, and recommended feeds feel like a true reflection of us, just as much as owning a record or a cd.