The age of the Internet has broke down the barriers that once existed between Australia and the rest of the world. We’re keenly aware that there are vast numbers of products and services available overseas that we want to take advantage of but either can’t, because they don’t want to bring it to us, or won’t because it’s far too expensive. We’re a resourceful bunch though and whilst companies will try their darnedest to make us pay the dreaded Australia Tax we’ll find a way around it, legitimately or otherwise. Probably the most popular of services like this is Netflix which, even though it’s not available here, attracts some 200,000 subscribers here in Australia. That number could soon rocket skywards as Netflix has finally announced that they’ll be coming to our shores early next year.
Australia will be the 16th country to receive the Netflix service, 7 years after they originally launched in the USA. Whilst there’s been demand for them to come Australia for some time now the critical mass of semi-legitimate users, plus the maturity of the cloud infrastructure they will need to deliver it here (Netflix uses AWS), has finally reached a point where an actual presence is warranted. Details are scant on exactly what they’ll be offering in Australia but looking at the other 14 non-US countries to get Netflix we can get a pretty good idea of what to expect when they finally hit the go live button for the Australian service.
For starters the full catalogue of shows that the USA service has will likely not be available to Netflix Australia subscribers. Whilst original content, like House of Cards or Orange is the New Black, will be available the content deals inked by rights holders with other companies in Australia will unfortunately take precedent over Netflix. This doesn’t mean that this won’t change over time as it’s highly likely that rights holders will look to move onto Netflix as old contracts expire but it might put a damper on the initial uptake rate. Considering that there are numerous services to change your Netflix region to get the full catalogue though I’m sure the restriction won’t have too much of an effect.
The DVD service probably won’t be making it here either, although I don’t think anyone really cares about that anyway.
Probably the biggest issue that Netflix will face coming to Australia is the dismal state of the Internet infrastructure here. Whilst most of us have enough speed to support some level of streaming the numbers of us that can do anything above 720p is a much more limited market. As long time readers will know I have little faith in the MTM NBN to provide the speeds required to support services like Netflix so I don’t think this is a problem that will be going away any time soon. Well, unless everyone realises their mistake at the next election.
Overall this is good news for Australia as it has the potential to break the iron grip that many of the pay TV providers have on the content that Australians want. It might not be the service that many are lusting after for but over time I can see Netflix becoming the dominant content platform in Australia. Hopefully other content providers will follow suit not long after this and Australia will finally get all the services it’s been lusting after for far too long. Maybe then people will realise the benefits of a properly implemented FTTP NBN and I’ll finally be able to stop ranting about it.
Microsoft isn’t a company you’d associate with open source. Indeed if you wound back the clock 10 years or so you’d find a company that was outright hostile to the idea, often going to great lengths to ensure open source projects that competed with their offerings would never see the light of day. The Microsoft of today is vastly different, contributing to dozens of open sourced projects and working hard with partner organisations to develop their presence in the ecosystem. For the most part however this has usually been done with an integration view towards their proprietary products which isn’t exactly in-line with the open source ethos. That may be set to change however as Microsoft will be fully open sourcing its .NET framework, the building blocks of all Microsoft applications.
For the uninitiated Microsoft .NET is a development framework that’s been around since the Windows XP days that exposed a consistent set of capabilities which applications could make use of. Essentially this meant that developing a .NET application meant you could guarantee it would work on any computer running that framework, something which wasn’t entirely a given before its inception. It’s since then grown substantially in capability, allowing developers to create some very capable programs using nothing more than the functionality built directly into Windows. Indeed it was so successful in accomplishing its aims that there was already a project going to make it work on non-Windows platforms, dubbed Mono, and it is with them that Microsoft is seeking to release a full open source implementation of the .NET framework.
Whilst this still falls in line with Microsoft’s open source strategy of “things to get people onto the Microsoft platform” it does open up a lot of opportunities for software to be freed from the Microsoft platform. The .NET framework underpins a lot of applications that run on Windows, some that only run on Windows, and an implementation of that framework on another platform could quickly elevate them to cross platform status. Sure, the work to translate them would still likely be non-trivial, however it’ll be a damn sight easier with a full implementation available, possibly enough to tempt some companies to make the investment.
One particularly exciting application of an open sourced .NET framework is games which, traditionally, have an extremely high opportunity cost when porting between platforms. Whilst everything about games development on Windows isn’t strictly .NET there are a lot of .NET based frameworks out there that will be readily portable to new platforms once the open sourcing is complete. I’m not expecting miracles, of course, but it does mean that the future of cross-platform releases is looking a whole bunch brighter than it was just a week ago.
This is probably one of Microsoft’s longest bets in a while as it’s going to be years before the .NET framework sees any kind of solid adoption among the non-Windows crowd. However this does drastically increase the potential of C# and .NET to become the cross platform framework of favour with developers, especially considering the large .NET developer community that already exists today. It’s going to be an area that many of us will be watching with keen interest as it’s yet another signal that Microsoft isn’t the company it used to be, a likely never will be again in the future.
There’s two distinct schools of thought when it comes to the modular smartphone idea. The first is that it’s the way phones were meant to be made, giving users the ability to customize every aspect of their device and reducing e-waste at the same time. The other flips that idea on its head, stating that the idea is infeasible due to the limitations inherent in a modular platform and reliance on manufacturers to build components specifically for the platform. Since I tend towards the latter I thought that Project Ara, Google’s (nee Motorola’s) attempt at the idea, would likely never see the light of day but as it turns out the platform is very real and they even have a working prototype.
The essence of the idea hasn’t changed much since Motorola first talked about it at the end of last year, being a more restrained version of the Phonebloks idea. The layout is the same as the original design prototypes, giving you space on the back of the unit for about 7 modular units and space on the front for a large screen and a speaker attachment. However they also showed off a new, slim version which has space for fewer modules but is a much sleeker unit overall. Google also mentioned that they were working on a phablet design as well which was interesting considering that the current prototype unit was looking to be almost phablet sized. The whole unit, dubbed Spiral 1, was fully functional including module removal and swapping so the idea has definitely come a long way since it’s initial inception late last year.
There are a few things that stand out about the device in its current form, primarily the way in which some of the blocks don’t conform to the same dimensions as other ones. Most notably you can see this with the blood oxygen sensor they have sticking out of the top however you’ll also notice that the battery module is about twice the height of anything else. This highlights one of the bigger issues with modular design as much of the heft in modern phones is due to the increasingly large batteries they carry with them. The limited space of the modular blocks means that either the batteries have significantly reduced capacity or have to be bigger than the other modules, neither of which is a particularly desirable attribute.
In fact the more the I think about Project Ara the more I feel it’s oriented towards those looking to develop hardware for mobile platforms than it is for actual phone users. Being able to develop your specific functionality without having to worry about the rest of the platform frees up a significant amount of time which can then be spent on getting said functionality into other phones. In that regards Project Ara is amazing however that same flexibility is likely what will turn many consumers off such a device. Sure, having a phone tailored to your exact specifications has a certain allure, but I can’t help but feel that that market is vanishingly small.
It will be interesting to see how the Project Ara platform progresses as they have hinted that there’s a much better prototype floating around (called Spiral 2) which they’re looking to release to hardware developers in the near future. Whilst having a proof of concept is great there’s still a lot of questions around module development, available functionality and, above all, the usability of the system when its complete. It’s looking like a full consumer version likely isn’t due out until late next year or early 2016 so we’re going to have to wait a while to see what the fully fledged modular smartphone will look like.
For the longest time, far too long in my opinion, XP had been the beast that couldn’t be slayed. The numerous releases of Windows after it never seemed to make much more than a slight dent in its usage stats and it reigned as the most used operating system worldwide for an astonishing 10 years after its initial release. It finally lost its crown to Windows 7 back in October of 2011 but it still managed to hold on a market share that dwarfed many of its competitors. It’s decline was slow though, much slower than an operating system which was fast approaching end of life should have been. However last quarter saw it drop an amazing 6% in total usage, finally dropping it below the combined usage of Windows 8 and 8.1.
The reasons behind this drop are wide and varied but it finally appears that people are starting to take Microsoft’s warnings that their product is no longer supported seriously and are looking for upgrades. Surprisingly though the vast majority of people transitioning away from the aging operating system aren’t going for Windows 7, they’re going straight to Windows 8.1. This isn’t to say that 8.1 is eating away at 7’s market share however, it’s up about half a percent in the same time frame, and the upgrade path is likely due to the fact that Microsoft has ceased selling OEM copies of Windows 7. Most of those new licenses do come with downgrade rights however though I’m sure few people actually use them.
If XP’s current downward trend continues along this path then it’s likely to hit the low single digit usage figures sometime around the middle of next year. On the surface this would appear to be a good thing for Microsoft as it means that the majority of their user base will be on a far more modern platform. However at the same time the decline might just be a little too swift for people to consider upgrading to Windows 10 which isn’t expected to be RTM until late next year. Considering the take up performance of Windows 8 and 8.1 this could be something of a concern for Microsoft although there is another potential avenue: Windows 7 users.
The last time Microsoft has a disastrous release like Windows 8 the next version of Windows to take the majority of the market share was 7, a decade after the original had released. Whilst it’s easy to argue that this time will be different (like everyone does) a repeat performance of that nature would see Windows 7 being the dominant platform all the way up until 2019. Certainly this is something that Microsoft wants to avoid so it will be interesting to see how fast Windows 10 gets picked up and which segments of Microsoft’s business it will cannibalize. Should it be primarily Windows 7 based then I’d say everything would be rosy for them, however if it’s all Windows 8/8.1 then we could be seeing history repeat itself.
Microsoft is on the cusp of either reinventing itself with Windows 10 or being doomed to forever repeat the cycle which consumers have forced them into. To Microsoft’s credit they have been trying their best to break out of this mould however it’s hard to argue with the demands of the consumer and there’s only so much they can do before they lose their customer’s faith completely. The next year will be very telling for how the Microsoft of the future will look and how much of history will repeat itself.
If you’re a long time PC gamer chances are that you’ve considered getting yourself a gaming laptop at one point or another. The main attraction from such a device is portability, especially back in the heydays of LANs where steel cases and giant CRTs were a right pain to lug around. However they always came at a cost, both financially and opportunity as once you bought yourself a gaming laptop you were locked into those specs until you bought yourself another one. Alienware, a longtime manufacturer of gaming laptops, has cottoned onto this issue and has developed what they’re calling the Graphics Amplifier in order to bring desktop level grunt and upgradeability to their line of laptops.
On the surface it looks like a giant external hard drive but inside it are all the components required to run any PCIe graphics card. It contains a small circuit board with a PCIe x16 slot, a 450W power supply and a host of other connections because why not. There’s no fans or anything else to speak of however so you’re going to want to get a card with a blower style fan system on it, something which you’ll only see on reference cards these days. This then connects back to an Alienware laptop through a proprietary connection (unfortunately) which then allows the graphics card to act as if it’s installed in the system. The enclosure retails for about $300 without the graphics card included in it which means you’re up for about $600+ if you’re going to buy one for it. That’s certainly not out of reach for those who are already investing $1800+ in the requisite laptop but it’s certainly enough to make you reconsider the laptop purchase in the first place.
You see whilst this external case does appear to work as advertised (judging by the various articles that have popped up with it) it essentially removes the most attractive thing about having a gaming capable laptop: the portability. Sure this is probably more portable than a mini tower and a monitor but at the same time this case is likely to weigh more than the laptop itself and won’t fit into your laptop carry bag. The argument could be made that you wouldn’t need to take this with you, this is only for home use or something, but even then I’d argue you’d likely be better off with a gaming desktop and some slim, far more portable laptop to take with you (both of which could be had for the combined cost of this and the laptop).
Honestly though the days have long since passed when it was necessary to upgrade your hardware on a near yearly basis in order to be able to play the latest games. My current rig is well over 3 years old now and is still quite capable of playing all current releases, even if I have to dial back a setting or two on occasion. With that in mind you’d be better off spending the extra cash that you’d sink into this device plus the graphics card into the actual laptop itself which would likely net you the same overall performance. Then, when the laptop finally starts to show its age, you’ll likely be in the market for a replacement anyway.
I’m sure there’ll be a few people out there who’ll find some value in a device like this but honestly I just can’t see it. Sure it’s a cool piece of technology, a complete product where there’s only been DIY solutions in the past, but it’s uses are extremely limited and not likely to appeal to those who it’ll be marketed too. Indeed it feels much like Razer’s modular PC project, a cool idea that just simply won’t have a market to sell its product to. It’ll be interesting to see if this catches on though but since Alienware are the first (and only) company to be doing this I don’t have a high hopes.
It was just over 2 years ago that Felix Baumgartner leapt from the Red Bull Stratos capsule from a height of 39KMs above the Earth’s atmosphere, breaking a record that had stood for over 50 years. The amount of effort that went into creating that project left many, including myself, thinking that Baumgartner’s record would stand for a pretty long time as few have the resources and desire to do something of that nature. However as it turns out one of Google’s Senior Vice Presidents, Alan Eustace, had been working on breaking that record in secret for the past 3 years and on Friday last week he descended to Earth from a height of 135,890 feet (41.4KM), shattering Baumgartner’s record by an incredible 7,000 feet.
The 2 jumps could not be more different, both technically and generally. For starters the Red Bull Stratos project was primarily a marketing exercise for Red Bull, the science that happened on the side was just a benefit for the rest of us. Eustace’s project on the other hand was done primarily in secret, with him eschewing any help from Google in order to avoid it becoming a marketing event. Indeed I don’t think anyone bar those working on the project knew that this was coming and the fact that they managed to achieve what Stratos did with a fraction of the funding speaks volumes to the team Eustace created to achieve this.
Looking at the above picture, which shows Eustace dangling from a tenuous tether as he ascends upwards, it’s plain to see that their approach was radically different to Stratos. Instead of building a capsule to transport Eustace, like Stratos and Kittinger’s project both did, they instead went for a direct tether to his pressure suit. This meant he spent the long journey skywards dangling face down which, whilst being nightmare material for some, would’ve given him an unparalleled view of the Earth disappearing from him. It also means that the load the balloon needed to carry was greatly reduced by comparison which likely allowed him to ascend much quicker.
Indeed the whole set up is incredibly bare bones with Eustace’s suit lacking many of the ancillary systems that Baumgartner’s had. One that amazed me was the lack of any kind of cooling system, something which meant that any heat he generated would stick around for an uncomfortably long period of time. To get around this he essentially remained motionless for the entire ascent, responding to ground control by moving one of this legs which they could monitor on camera. They did include a specially developed kind of parachute though, called Saber, which ensured that he didn’t suffer from the same control issues that Baumgartner did during his descent.
It’s simply astounding how Eustace and his team managed to achieve this, given their short time frame and comparatively limited budget. I’m also wildly impressed that they managed to keep this whole thing a secret for that period of time too as it would’ve been very easy for them to overshadow the Stratos project, especially given some of the issues they encountered. Whilst we might not all be doing high altitude jumps any time soon the technology behind this could find its way into safety systems in the coming generation of private space flight vehicles, something they will all need in no short order.
Windows has always had a troubled relationship with security. As the most popular desktop operating system it’s frequently the target of all sorts of weird and wonderful attacks which, to Microsoft’s credit, they’ve done their best to combat. However it’s hard to forget the numerous missteps along the way like the abhorrent User Access Control system which, in its default state, did little to improve security and just added another level of frustration for users. However if the features coming from the technical preview of Windows 10 are anything to go by Microsoft might finally be making big boy steps towards improving security on their flagship OS.
Whilst there’s numerous third party solutions to 2 factor authentication on Windows, like smartcards or tokens, the OS itself has never had that capability natively. This means that for the vast majority of Windows users this heightened security mode has been unavailable. Windows 10 brings with it the Next Generation Credentials service which allows users (both consumer and corporate) the ability to enrol a device to function as a second factor for authentication. The larger mechanics of how this work are still being worked out however the application has a PIN which would prevent unauthorized access to the code, ensuring that losing your device doesn’t mean someone automatically gains access to your Windows login. Considering this kind of technology has been freely available for years (hell my World of Warcraft characters have had it for years) it’s good to see it finally making its way into Windows as native functionality.
There’s also extensive customization abilities available thanks to Microsoft adopting the FIDO Alliance standard rather than developing their own proprietary solution. In addition to the traditional code-generation 2 factor auth you can also use your smartphone as a sort of smartcard with it being automatically recognised when brought next to a bluetooth enabled PC. This opens up the possibility for your phone to be a second factor for a whole range of services and products that currently make use of Microsoft technology, like Active Directory integrated applications. Whilst some might lament that possibility the fact that it’s based on open standards means that such functionality won’t be limited to the Microsoft family of products.
Microsoft has also announced a whole suite of better security features, many of which have been third party products for the better part of a decade. Encryption is now available for the open and save dialogs natively within the Windows APIs, allowing developers to easily integrate encryption functionality into their applications. This comes hand in hand with controls around which applications can access said encrypted data, ensuring that data handling measures can’t be circumvented by using non-standard applications. Device lock down is also now natively supported, eliminating the need for other device access control software like Lumension (which, if you’ve worked with, will likely be thankful for).
It might not be the sexiest thing to be happening in Windows 10 but it’s by far one of the more important. As the defacto platform for many people increases in Windows security are very much welcome and hopefully this will lead to a much more secure computing world for us all. These measures aren’t a silver bullet by any stretch of the imagination but they’ll go a long way to making Windows far more secure than it has been in the past.
For the last 6 months I’ve been on the lookout for the next phone that will replace my Xperia Z. Don’t get me wrong, it’s still quite a capable phone, however not a year has gone by in the past decade that there hasn’t been one phone that triggered my geeky lust, forcing me to part ways with several hundred dollars. However the improvements made since I acquired my last handset have just been evolutionary steps forward, none of which have been compelling enough to make me get my wallet out. I had hoped that the Nexus 6 would be the solution to my woes and, whilst it’s not exactly the technological marvel I was hoping for, Google might just be fortunate enough to get my money this time around.
The Nexus 6 jumps on the huge screen bandwagon bringing us an (almost) 6″ display boasting a 2560 x 1440 resolution on an AMOLED panel. The specs under the hood are pretty impressive with it sporting a quad core 2.7 GHz SOC with 3GB RAM and a 3220mAh battery. The rest of it is a rather standard affair including things such as the standard array of sensors that everyone has come to expect, a decent camera (that can do usable 4K video) and a choice between 32GB and 64GB worth of storage. If you were upgrading every 2 years or so the Nexus 6 would be an impressive step up however compared to what’s been available in the market for a while now it’s not much more than a giant screen.
You can’t help but compare this phone to the recently released iPhone 6+ which also sports a giant screen and similar specifications. In terms of who comes out ahead it’s not exactly clear as they both seem to win out in various categories (the Nexus 6 has the better screen, the iPhone 6+ is lighter) but then again the main driver of which one of these you’d go for would be more heavily driven by which ecosystem you’d already bought into. I’d be interested to see how these devices compare side by side however as there’s only so much you can tell by looking at spec sheets.
As someone who’s grown accustom to his 5″ screen I was hoping there’d be a diminutive sister of the Nexus 6, much like the iPhone 6. You can still get the Nexus 5, which now sports Android L, however the specs are the same as they ever were which means there’s far less incentive for people like me to upgrade. Talking to friends who’ve made the switch to giant phones like this (and seeing my wife, with her tiny hands, deftly use her Galaxy Note) it seems like the upgrade wouldn’t be too much of a stretch. Had there been a smaller screen I would probably be a little bit more excited about acquiring one as I don’t really have a use case for a much bigger screen than what I have now. That could change once I get some time with the device, though.
So whilst I might not be frothing at the mouth to get Google’s latest handset they might just end up getting my money anyway as there just enough new features for me to justify upgrading my near 2 year old handset. There’s no mistaking that the Nexus 6 is the iPhone 6+ for those on the Android ecosystem and I’m sure there will be many a water cooler conversation over which one of them is the better overall device. For me though the main draw is the stock Android interface with updates that are unimpeded by manufacturers and carriers, something which has been the bane of my Android existence for far too long. Indeed that’s probably the only compelling reason I can see to upgrade to the Nexus 6 at the moment, which is likely enough for some.
The Tesla Model S as we know it today is quite an impressive car. Whilst it’s not exactly within the everyman’s price range yet (getting one landed in Australia likely won’t see much change from $100K) it’s gone a long way to making a high performing electric vehicle available to the masses, especially considering Tesla stance on their patents. Before that electric cars were more of a niche product for the ultra environmentally conscious, combining tiny engines with small frames that would have just enough power to get you to work and back. Now they’re far more easily compared to high end luxury cars and with the new things that Elon announced last week electric cars are heading into a class all of their own.
Elon teased last week that he was going to unveil the D soon (and seemingly forgot how much of a dirty mind the entire Internet has) and “something else”. The D was for their new drive train system that incorporates 2 motors, making the Tesla Model S one of the few fully electric all wheel drive cars. The something else turned out to be the debut of their autopilot system, a sort of cut down version of the Google self-driving car. Whilst the D version of the Model S won’t be available for another couple months (although you can order one today) all Model S cars built within the last couple weeks shipped with the autopilot hardware. Suffice to say both these announcements are pretty exciting although the latter probably more so.
The dual motors is an interesting upgrade for the Model S as it’s a pretty common feature among higher end luxury cars, something which it has been lacking. Of particular note is how the dual motor upgrade affects the various aspects of the car, like slashing 0.8 seconds off the 0-100 time (3.2 seconds) and increasing range by about 3.5%, all whilst granting the benefits that all wheel drive provides. Typically you’d be taking a decent hit to range and efficiency due to the increased weight and power requirements but the Model S has managed to come out on top in all respects. Should those figures hold up in real world testing then it’ll speak volumes to the engineering team that Tesla has managed to cultivate.
However the most interesting part for me was the debut of Tesla’s autopilot system. Elon Musk had always been of the mind that a self driving car didn’t need to be an all encompassing thing, instead they should aim to do the majority of tasks first before looking to take the next leap into full automation. Tesla’s autopilot system is the embodiment of that philosophy, taking some of the technology that’s currently available (emergency braking, lane keeping, collision avoidance) and combining it into one seamless package. It won’t get you from point A to point B without human intervention but it’ll happy take over on the highway, park itself in the garage and even meet you at a certain location. It might not be as comprehensive at what Google is looking to create but it’s available today and does almost everything you’d need it to.
I really shouldn’t be surprised that a Musk created company is managing to innovate so quickly in an industry that has long been one of the slowest movers but honestly these two announcements blew me away. The dual motors might not exactly be a revolutionary piece of technology but the way Telsa has done it speaks volumes to the calibre of people that they have working there. The introduction of autopilot in just over a year since they first talked about it really is quite amazing and whilst it might not be the all encompassing system that Google is seeking it will likely be the standard for many years to come. I can’t wait to see what Tesla has in store for us next as they don’t seem to have any intentions of stopping their brisk innovating pace any time soon.
The time has long since past when a computer manufacturer could get by on shipping tin. The margins on computer equipment are so low that, most of the time, the equipment they sell is just a loss leader for another part of the business. Nowadays the vast majority of most large computer company’s revenue comes from their services division, usually under the guise of providing the customer a holistic solution rather than just another piece of tin. Thus for many companies the past couple decades have seen them transform from pure hardware businesses into more services focused companies, with several attempting more radical transformations in order to stay relevant. HP has become the most recent company to do this, announcing that they will be splitting the company in half.
HP will now divest itself into 2 different companies. The first will be Hewlett Packard Enterprise comprising of their server market, services branch and software group. The second will be purely consumer focused, comprising of their personal computer business and their printing branch. If you were going to split a PC business this is pretty much how you’d do it as whilst these functions are somewhat complimentary to each other (especially if you want to be the “end to end” supplier for all things computing) there’s just as many times when they’re at odds. HP’s overarching strategy with this split is to have two companies that can be more agile and innovative in their respective markets and, hopefully, see better margins because of it.
When I first heard the rumours swirling about this potential split the first question that popped into my head was “Where is the services business going?”. As I alluded to before the services business is the money maker for pretty much every large PC manufacturer these days and in this case the enterprise part of HP has come away with it. The numbers only give a slight lead to the new enterprise business in terms of revenue and profit however with the hardware business has been on a slow decline for the past few years which, if I’m honest, paints a bleak picture for HP Inc. going forward. There’s nothing to stop them from developing a services capability (indeed parts of the consumer business already have that) however in its current form I’d put my money on HP Inc. being the one who’s worse off out of this deal.
That could change however if HP’s rhetoric has some merit to it. HP, as it stands today, is an amalgamation of dozens of large companies that it acquired over the years and whilst they all had a similar core theme of being in the IT business there really wasn’t a driving overarching goal for them to adhere to. The split gives them an opportunity to define that more clearly for each of the respective companies, allowing them to more clearly define their mission within each of their designated market segments. Whether that will translate into the innovation and agility that they’re seeking is something we’ll have to see as this is yet another unprecedented change from a large IT conglomerate.
As someone who’s been involved in the IT industry for the better part of 2 decades now the amount of change that’s happened in the last couple years has been, honestly, staggering. We’ve seen IBM sell off some of its core manufacturing capability (the one no one got fired for buying), Dell buy back all its stock to become a private company again and now HP, the last of the 3 PC giants, divest itself into 2 companies. It will likely take years before all the effects of these changes are really felt but suffice to say that the PC industry of the future will look radically different to that of the past.
FULL DISCLOSURE: The writer is a current employee of Dell. All opinions expressed in this article are of the writer’s own and are not representative of Dell.