programming4us
programming4us
ENTERPRISE

The End Of Wintel (Part 1)

2/28/2013 5:56:14 PM

Wintel has ruled the PC world for almost a decade now, but seems to have lost its footing. Is it too late for it to scramble back to the top?

Microsoft (Windows) and Intel have been defining the computing experience of the world for well over a decade and the word “Wintel” is now synonymous with the PC the way Google is with the internet. There was a time when Wintel commanded over 90 percent of the global PC market, but times have changed and with the rise of the smartphone/phablet/tablet and the promise of portability and power, the true definition of personal computing has been blurred to include all these fancy new “computational devices.” In this new world, it is the AAA that rule, Android, Apple and ARM, with the all-powerful Wintel now being relegated to a more humble 20 percent market share (http://dgit.in/ULwHN9). With Microsoft branching out to ARM, and Intel unable to compete in the mobile computing space, is the era of Wintel finally over? What is ARM exactly and why is Intel, the largest chip manufacturer in the world, unable to compete? To understand this we need to understand the way these CPUs process data and why this is so important to us.

Wintel 3.0 for Mac

Wintel 3.0 for Mac

Intel chips are based on a CISC (Complex Instruction Set Computing) architecture that essentially allows the CPU to process multiple low-level instructions simultaneously, which also reduces the number of instructions required to complete a task; thus simplifying the programmer’s (and compiler’s) job.

Brainless monkeys

RISC (Reduced Instruction Set Computing), an architecture used by ARM, is the exact opposite of CISC where only one instruction is processed at a time, in sequence. This increases the compiler’s workload as each task would have to be split into multiple steps. This would be better explained with a simple analogy; if someone were to tell you to get them some water, you would just go to the water source, fill a glass with water and bring it back to that person. This is CISC where you are the processor and the “someone” is the OS or compiler. In RISC mode, you would need to be given specific instructions; turn 180 degrees, walk 5 meters, turn left 90 degrees, walk 7 meters, and so on. In this situation, you, as a CPU, don’t need to think too much and the instructions can be as easily followed by a person as well as a somewhat well-trained monkey. If the instructions are given at the right time, there wouldn’t even be a time lag between the two methods of operation and there is not much brain power involved (no offence to monkeys). The problem arises when, say, the glass is not there at the location you specified, the monkey would not know what to do next and may not even be able to tell you that. In computing terms what this means is that the program will just get stuck and/or will return an incorrect result. Current processors do have methods to prevent such a thing from happening. The main advantages of RISC can be summed up in the following points:

RISC (Reduced Instruction Set Computing), an architecture used by ARM, is the exact opposite of CISC where only one instruction is processed at a time, in sequence.

RISC (Reduced Instruction Set Computing), an architecture used by ARM, is the exact opposite of CISC where only one instruction is processed at a time, in sequence.

·         A simpler architecture, meaning fewer transistors and it’s hence, cheaper to produce.

·         Very high efficiency per clock (if the code is compiled correctly) and per watt, translating to lower operating temperatures.

·         Lower power consumption because of fewer “components” and higher efficiency.

The end of Geekdom

As can be seen from the advantages of RISC listed above, this architecture is ideal for mobile devices because it provides stellar battery life (compared to CISC processors) and negligible heat (again, only in comparison to CISC). Comparatively, CISC architecture such as Intel’s x86 (except Atom) is leagues ahead of any ARM processor in terms of sheer power. The problem here is that most people don’t need that much power. What does the average person’s computing usage involve? Nothing more than Facebook, a bit of YouTube, some Instagraming and maybe a bit of mild video editing. Most work is also something that just involves office suites and e-mails. You don’t need a quad-core i7 to do that. In fact, you don’t even need an i3. An ARM-based tablet or phone can do that much admirably while still giving you portability and stellar battery life. Nowadays building your own computer is almost considered to be the equivalent of building your own table or chair. Computing is such an integral part of our lives that it has just been taken for granted. You might think that this only applies to first world countries but you’d be wrong. More people have smartphones than PCs and internet access is just a SIM card away. Almost everyone knows how to operate a phone, send an email and browse the web. Where is Windows or Intel here? They just don’t matter anymore. Internet Explorer 6 was the most popular browser for so long, despite being so flawed and people still used it and not just because they didn’t know any better. They didn’t know any better but it was more the fact that the prospect of using a PC (and Windows) was such a daunting task that they just preferred to stick to something they knew rather than try something new which might create a whole pile of other unknown issues.

a quad-core i7

a quad-core i7

It is this situation that Apple managed to take advantage of when it launched the iPad. People were comfortable with it, they never really felt that they could damage the OS. They got a good device, a good browser, it could access the internet and it seemed way better than the Windows experience that they had been used to for so long. There was no more messing around with settings, no installing scary applications that might install a “virus”. Curated content and a curated experience became the norm and it was enough. The geeks did fret and fume about it, but Google lent a hand and Android filled the gap.

Microsoft (and Intel), for whatever reason, remained completely oblivious to this space and while they did ultimately catch on, they were far too late. The products they released just seemed slapped together and while they had the right idea, they were too late and tried too hard to adapt in too short a time period.

Other  
  •  Windows System Programming : Exception Handling - Exceptions and Their Handlers
  •  Building Applications with Visual Studio 2010 : Getting Smart about Smart Clients - Using the Properties Window
  •  Building Applications with Visual Studio 2010 : Getting Smart about Smart Clients - Manipulating Controls
  •  .NET Debugging : Introduction to the Tools - Reflector for .NET, PowerDbg, Managed Debugging Assistants
  •  .NET Debugging : Introduction to the Tools - CLR Profiler, Performance Counters
  •  Microsoft .NET : Design Principles and Patterns - From Principles to Patterns (part 2)
  •  Microsoft .NET : Design Principles and Patterns - From Principles to Patterns (part 1)
  •  All We Need To Know About Green Computing (Part 4)
  •  All We Need To Know About Green Computing (Part 3)
  •  All We Need To Know About Green Computing (Part 2)
  •  
    video
     
    Video tutorials
    - How To Install Windows 8

    - How To Install Windows Server 2012

    - How To Install Windows Server 2012 On VirtualBox

    - How To Disable Windows 8 Metro UI

    - How To Install Windows Store Apps From Windows 8 Classic Desktop

    - How To Disable Windows Update in Windows 8

    - How To Disable Windows 8 Metro UI

    - How To Add Widgets To Windows 8 Lock Screen

    - How to create your first Swimlane Diagram or Cross-Functional Flowchart Diagram by using Microsoft Visio 2010
    programming4us programming4us
    programming4us
     
     
    programming4us