ITExpo West — Achieving HD Voice On Smartphones

I will be moderating a panel discussion at ITExpo West on Tuesday 5th October at 11:30 am in room 306B: “Achieving HD Voice On Smartphones.”

Here’s the session description:

The communications market has been evolving to fixed high definition voice services for some time now, and nearly every desktop phone manufacturer is including support for G.722 and other codecs now. Why? Because HD voice makes the entire communications experience a much better one than we are used to.

But what does it mean for the wireless industry? When will wireless communications become part of the HD revolution? How will handset vendors, network equipment providers, and service providers have to adapt their current technologies in order to deliver wireless HD voice? How will HD impact service delivery? What are the business models around mobile HD voice?

This session will answer these questions and more, discussing both the technology and business aspects of bringing HD into the mobile space.

The panelists are:

This is a deeply experienced panel; each of the panelists is a world-class expert in his field. We can expect a highly informative session, so come armed with your toughest questions.

Intel Infineon: history repeats itself

Unlike its perplexing McAfee move, Intel had to acquire Infineon. Intel must make the Atom succeed. The smartphone market is growing fast, and the media tablet market is in the starting blocks. Chips in these devices are increasingly systems-on-chip, combining multiple functions. To sell application processors in phones, you must have a baseband story. Infineon’s RF expertise is a further benefit.

As Linley Gwennap said when he predicted the acquisition a month ago, the fit is natural. Intel needs 3G and LTE basebands, Infineon has no application processor.

Linley also pointed out Intel’s abysmal track record for acquisitions.

Intel has been through this movie before, for the same strategic reasons. It acquired DSP Communications in 1999 for $1.6 Bn. The idea there was to enter the cellphone chip market with DSP’s baseband plus the XScale ARM processor that Intel got from DEC. It was great in theory, and XScale got solid design wins in the early smart-phones, but Intel neglected XScale, letting its performance lead over other ARM implementations dwindle, and its only significant baseband customer was RIM.

In 2005, Paul Otellini became CEO; at that time AMD was beginning to make worrying inroads into Intel’s market share. Otellini regrouped – he focused in on Intel’s core business, which he saw as being “Intel Architecture” chips. But XScale runs an instruction set architecture that competes with IA, namely ARM. So rather than continuing to invest in its competition, Intel instead dumped off its flagging cellphone chip business (baseband and XScale) to Marvell for $0.6 Bn, and set out to create an IA chip that could compete with ARM in size, power consumption and price. Hence Atom.

But does instruction set architecture matter that much any more? Intel’s pitch on Atom-based netbooks was that you could have “the whole Internet” on them, including the parts that run only on IA chips. But now there are no such parts. Everything relevant on the Internet works fine on ARM-based systems like Android phones. iPhones are doing great even without Adobe Flash.

So from Intel’s point of view, this decade-later redo of its entry into the cellphone chip business is different. It is doing it right, with a coherent corporate strategy. But from the point of view of the customers (the phone OEMs and carriers) it may not look so different. They will judge Intel’s offerings on price, performance, power efficiency, wireless quality and how easy Intel makes it to design-in the new chips. The same criteria as last time.

Rethink Wireless has some interesting insights on this topic…

iPhone 4 gets competition

When the iPhone came out it redefined what a smartphone is. The others scrambled to catch up, and now with Android they pretty much have. The iPhone 4 is not in a different league from its competitors the way the original iPhone was. So I have been trying to decide between the iPhone 4 and the EVO for a while. I didn’t look at the Droid X or the Samsung Galaxy S, either of which may be better in some ways than the EVO.

Each hardware and software has stronger and weaker points. The Apple wins on the subtle user interface ingredients that add up to delight. It is a more polished user experience. Lots of little things. For example I was looking at the clock applications. The Apple stopwatch has a lap feature and the Android doesn’t. I use the timer a lot; the Android timer copied the Apple look and feel almost exactly, but a little worse. It added a seconds display, which is good, but the spin-wheel to set the timer doesn’t wrap. To get from 59 seconds to 0 seconds you have to spin the display all the way back through. The whole idea of a clock is that it wraps, so this indicates that the Android clock programmer didn’t really understand time. Plus when the timer is actually running, the Android cutely just animates the time-set display, while the Apple timer clears the screen and shows a count-down. This is debatable, but I think the Apple way is better. The countdown display is less cluttered, more readable, and more clearly in a “timer running” state. The Android clock has a wonderful “desk clock” mode, which the iPhone lacks, I was delighted with the idea, especially the night mode which dims the screen and lets you use it as a bedside clock. Unfortunately when I came to actually use it the hardware let the software down. Even in night mode the screen is uncomfortably bright, so I had to turn the phone face down on the bedside table.

The EVO wins on screen size. Its 4.3 inch screen is way better than the iPhone’s 3.5 inch screen. The “retina” definition on the iPhone may look like a better specification but the difference in image quality is indistinguishable to my eye, and the greater size of the EVO screen is a compelling advantage.

The iPhone has far more apps, but there are some good ones on the Android that are missing on the iPhone, for example the amazing Wi-Fi Analyzer. On the other hand, this is also an example of the immaturity of the Android platform, since there is a bug in Android’s Wi-Fi support that makes the Wi-Fi Analyzer report out-of-date results. Other nice Android features are the voice search feature and the universal “back” button. Of course you can get the same voice search with the iPhone Google app, but the iPhone lacks a universal “back” button.

The GPS on the EVO blows away the GPS on the iPhone for accuracy and responsiveness. I experimented with the Google Maps app on each phone, walking up and down my street. Apple changed the GPS chip in this rev of the iPhone, going from an Infineon/GlobalLocate to a Broadcom/GlobalLocate. The EVO’s GPS is built-in to the Qualcomm transceiver chip. The superior performance may be a side effect of assistance from the CDMA radio network.

Incidentally, the GPS test revealed that the screens are equally horrible under bright sunshine.

The iPhone is smaller and thinner, though the smallness is partly a function of the smaller screen size.

The EVO has better WAN speed, thanks to the Clearwire WiMax network, but my data-heavy usage is mainly over Wi-Fi in my home, so that’s not a huge concern for me.

Battery life is an issue. I haven’t done proper tests, but I have noticed that the EVO seems to need charging more often than the iPhone.

Shutter lag is a major concern for me. On almost all digital cameras and phones I end up taking many photos of my shoes as I put the camera back in my pocket after pressing the shutter button and assuming the photo got taken at that time rather than half a second later. I just can’t get into the habit of standing still and waiting for a while after pressing the shutter button. The iPhone and the EVO are about even on this score, both sometimes taking an inordinately long time to respond to the shutter – presumably auto-focusing. The pictures taken with the iPhone and the EVO look very different; the iPhone camera has a wider angle, but the picture quality of each is adequate for snapshots. On balance the iPhone photos appeal to my eye more than the EVO ones.

For me the antenna issue is significant. After dropping several calls I stuck some black electrical tape over the corner of the phone which seems to have somewhat fixed it. Coverage inside my home in the middle of Dallas is horrible for both AT&T and Sprint.

The iPhone’s FM radio chip isn’t enabled, so I was pleased when I saw FM radio as a built-in app on the EVO, but disappointed when I fired it up and discovered that it needed a headset to be plugged in to act as an antenna. Modern FM chips should work with internal antennas. In any case, the killer app for FM radio is on the transmit side, so you can play music from your phone through your car stereo. Neither phone supports that yet.

So on the plus side, the EVO’s compelling advantage is the screen size. On the negative side, it is bulkier, the battery life is less, the software experience isn’t quite so polished.

The bottom line is that the iPhone is no longer in a class of its own. The Android iClones are respectable alternatives.

It was a tough decision, but I ended up sticking with the iPhone.

Samsung GT-S8500 is first with 11n, BT 3.0 certifications

Engadget reports that the Samsung GT-S8500 is the first phone to support Bluetooth 3.0. A look at the Wi-Fi Alliance website reveals that it was also the first feature phone to gain 802.11n certification.

The certificate is dated December 28th 2009, the same date that the first smartphone was certified for 802.11n – the LG Veri/VS750. The VS750 Wi-Fi appears to be more advanced than the Samsung, since it is certified for short guard interval and WMM Power Save.

While these are the first phones to gain Wi-Fi certification for 802.11n, they may not be the first to market.

First 802.11n handset spotted in the wild – what took so long?

The fall 2009 crop of ultimate smartphones looks more penultimate to me, with its lack of 11n. But a handset with 802.11n has come in under the wire for 2009. Not officially, but actually. Slashgear reports a hack that kicks the Wi-Fi chip in the HTC HD2 phone into 11n mode. And the first ultimate smartphone of 2010, the HTC Google Nexus One is also rumored to support 802.11n.

These are the drops before the deluge. Questions to chip suppliers have elicited mild surprise that there are still no Wi-Fi Alliance certifications for handsets with 802.11n. All the flagship chips from all the handset Wi-Fi chipmakers are 802.11n. Broadcom is already shipping volumes of its BCM4329 11n combo chip to Apple for the iTouch (and I would guess the new Apple tablet), though the 3GS still sports the older BCM4325.

Some fear that 802.11n is a relative power hog, and will flatten your battery. For example, a GSMArena report on the HD2 hack says:

There are several good reasons why Wi-Fi 802.11n hasn’t made its way into mobile phones hardware just yet. Increased power consumption is just not worth it if the speed will be limited by other factors such as under-powered CPU or slow-memory…

But is it true that 802.11n increases power consumption at a system level? In some cases it may be: the Slashgear report linked above says: “some users have reported significant increases in battery consumption when the higher-speed wireless is switched on.”

This reality appears to contradict the opinion of one of the most knowledgeable engineers in the Wi-Fi industry, Bill McFarland, CTO at Atheros, who says:

The important metric here is the energy-per-bit transferred, which is the average power consumption divided by the average data rate. This energy can be measured in nanojoules (nJ) per bit transferred, and is the metric to determine how long a battery will last while doing tasks such as VoIP, video transmissions, or file transfers.

For example, Table 1 shows that for 802.11g the data rate is 22 Mbps and the corresponding receive power-consumption average is around 140 mW. While actively receiving, the energy consumed in receiving each bit is about 6.4 nJ. On the transmit side, the energy is about 20.4 nJ per bit.

Looking at these same cases for 802.11n, the data rate has gone up by almost a factor of 10, while power consumption has gone up by only a factor of 5, or in the transmit case, not even a factor of 3.

Thus, the energy efficiency in terms of nJ per bit is greater for 802.11n.

Here is his table that illustrates that point:
Effect of Data Rate on Power Consumption

Source: Wireless Net DesignLine 06/03/2008

The discrepancy between this theoretical superiority of 802.11n’s power efficiency, and the complaints from the field may be explained several ways. For example, the power efficiency may actually be better and the reports wrong. Or there may be some error in the particular implementation of 802.11n in the HD2 – a problem that led HTC to disable it for the initial shipments.

Either way, 2010 will be the year for 802.11n in handsets. I expect all dual-mode handset announcements in the latter part of the year to have 802.11n.

As to why it took so long, I don’t think it did, really. The chips only started shipping this year, and there is a manufacturing lag between chip and phone. I suppose a phone could have started shipping around the same time as the latest iTouch, which was September. But 3 months is not an egregious lag.

Dual-mode technology maturing

The Rethink Wireless newsletter is always worth reading. An article in today’s edition says that according to ABI dual mode handset shipments are on track to double from 2008 to 2010, and more than double from 2009-2011 (144 million units to 300 million units).

Rethink’s Matt Lewis cites improved performance and usability as driving forces, plus a change in the attitudes of carriers towards hot-spots. Wireless network operators now often have captive Wi-Fi networks and can use them to offload their cellular networks.

The upshot is a prediction of 300 million dual mode handsets to ship in 2011: 100% of the smartphone market plus high end feature phones.

The attach rate of Wi-Fi will continue to grow. By 2011 the effects of Bluetooth 3.0 will be kicking in, pushing Wi-Fi attachment towards 100% in camera phones and music phones in ensuing years.

Bluetooth 3.0 arrives

The Bluetooth 3.0 specification has finally been ratified.

The main new feature is the Alternate MAC/PHY (AMP), that lets Bluetooth piggyback on Wi-Fi for high speed data transfers. The way it works is that applications write to the traditional Bluetooth Profile APIs, and connections are negotiated using the traditional Bluetooth radio. But then for high-speed data transfers the system switches to a direct peer-to-peer Wi-Fi session. This enables things like bulk syncing of photos from your phone to your PC, or streaming uncompressed CD stereo audio to wireless loudspeakers.

I wrote about Bluetooth AMP before, wondering why it retained a dependency on Bluetooth radio. The answer is that in idle, listening mode waiting for activity, Bluetooth is more power efficient than Wi-Fi, while Wi-Fi is more power efficient for bulk data transfers. This makes Bluetooth’s other next big thing, LE (formerly Wibree), an interesting complement to AMP: for power efficiency Bluetooth devices will reside in two modes, very low power idle mode (LE), and Wi-Fi mode when transferring data.

The Bluetooth 3.0 specification talks about 802.11 rather than 11g or 11n, since 802.11n is not yet ratified, but some of the companies involved will be supporting draft 802.11n anyway.

From an industry point of view there are several interesting aspects to this announcement, among them:

  • Atheros’ ascendence. Atheros, a leader in Wi-Fi, only recently got into the Bluetooth market, and currently only plays in the PC Bluetooth market. It dabbled in headset Bluetooth and got out, and has not yet announced Bluetooth for handsets. So Atheros is a minor player in Bluetooth, eclipsed by CSR and Broadcom, and several others. But Kevin Hayes of Atheros was the technical editor for the 802.11 Protocol Adaptation Layer of the Bluetooth 3.0 specification, and Atheros supplied the video and the demo of AMP at the 3.0 announcement event.
  • Potential movement of Wi-Fi into feature phones. Handset makers slice the market into four main segments: ultra low cost phones, basic phones, feature phones and smart phones. Wi-Fi is now pretty much ubiquitous in new smartphones, but effectively absent in all other types of cell phone. But feature phones have music and cameras which generate exactly the data that Bluetooth 3.0 was designed to sync with PCs, so Bluetooth 3.0 provides a motivation to handset manufacturers to add Wi-Fi to their feature phones. This will vastly boost the Wi-Fi attach rate in 2010 and beyond.
  • Another nail in the coffin of UWB (Ultra Wide-Band). In its original conception, AMP was to use WiMedia’s flavor of UWB. Later Wi-Fi was added to the mix, and now UWB is absent from the spec. UWB has so far failed to meet its performance expectations, and rather than fix it the WiMedia Alliance threw in the towel in March 2009. I suppose it is possible that the few companies still toiling away on fixing UWB will eventually overcome its performance woes, and that it will get adopted into the Bluetooth specification.

AT&T to deploy Voice over Wi-Fi on iPhones

Don’t get too excited by Apple’s announcement of a Voice over IP service on the iPhone 3.0. It strains credulity that AT&T would open up the iPhone to work on third party VoIP networks, so presumably the iPhone’s VoIP service will be locked down to AT&T.

AT&T has a large network of Wi-Fi hotspots where iPhone users can get free Wi-Fi service. The iPhone VoIP announcement indicates that AT&T may be rolling out voice over Wi-Fi service for the iPhone. It will probably be SIP, rather than UMA, the technology that T-Mobile uses for this type of service. It is likely to be based on some flavor of IMS, especially since AT&T has recently been rumored to be spinning up its IMS efforts for its U-verse service, which happens to include VoIP. AT&T is talking about a June launch.

An advantage of the SIP flavor of Voice over Wi-Fi is that unlike UMA it can theoretically negotiate any codec, allowing HD Voice conversations between subscribers when they are both on Wi-Fi; wouldn’t that be great? The reference to the “Voice over IP service” in the announcement is too cryptic to determine what’s involved. It may not even include seamless roaming of a call between the cellular and Wi-Fi networks (VCC).

AT&T has several Wi-Fi smartphones in addition to the iPhone. They are mostly based on Windows Mobile, so they can probably be enabled for this service with a software download. The same goes for Blackberries. Actually, RIM may be ahead of the game, since it already has FMC products in the field with T-Mobile, albeit on UMA rather than SIP, while Windows Mobile phones are generally ill-suited to VoIP.

Sharing Wi-Fi 2 – Atheros turns a cellphone into an access point

There are several smartphone applications that allow a cell phone to act as a wireless WAN router and Wi-Fi access point, creating a wireless LAN with Internet access. For the (jailbroken) iPhone there’s PDAnet, for Windows Mobile there’s WM Wi-Fi Router and for Symbian there’s Walking HotSpot and JoikuSpot. Now Atheros has proposed to bake this functionality into their low power Wi-Fi chipset.

An idea that is as patent jargon goes “obvious to one skilled in the art,” can sometimes have obvious handicaps to one experienced in the industry. While exposing a broadband wireless data connection through a smartphone’s Wi-Fi radio is massively useful to consumers, it is unlikely to appeal to network service providers, who would prefer you to buy a wireless data card (and an additional service subscription) for your laptop rather than to simply use the wireless data connection that you are already paying for on your phone.

It will be interesting to see where this goes. I will be stunned if Atheros’ implementation appears on any phone subsidized by (or even distributed by) a wireless carrier, until they can figure out a way to charge extra for it. As Tim Wu says in his Wireless Carterfone paper (the Wireless Carterfone concept was promoted by Skype, and rejected by the FCC last April):

carriers are in a position to exercise strong control over the design of mobile equipment. They have used that power to force equipment developers to omit or cripple many consumer-friendly features.

The billing issue may not be that intractable. Closely related models already exist. You can get routers from Cisco and other vendors that have a slot for a wireless WAN card, and the service providers have subscription plans for them. More similarly, this could be viewed as a kind of “tethering” But tethering only lets one PC at a time access the wireless WAN connection – unless that PC happens to support My Wi-Fi.

Update: Marvell has announced a similar capability for its 88W8688 chip.

A not so perfect Storm

The Verizon Storm may be heading for failure in more than one way. A raft of reviewers, led by David Pogue of the New York Times are trashing its usability. This means that even with the marketing might of Verizon behind it it may not fulfill its goal of being a bulwark against the iPhone in the enterprise.

But the Storm was an experiment in another way by Verizon. The other three major American mobile network operators have capitulated to Wi-Fi in smartphones. Against the new conventional wisdom, Verizon decided to launch a new flagship smartphone without Wi-Fi. The Storm looks like a trial balloon to see whether Wi-Fi is optional in modern smartphones. If the Storm is a success, it will demonstrate that it is possible to have credible business smartphones without Wi-Fi. But if it turns out to be a flop because of other factors, it will not be a proof point for Wi-Fi either way.

But Wi-Fi is a closed issue by now for all the network operators, perhaps even including Verizon. Phones have lead times of the order of a year or so, and controversies active back then may now be resolved. Verizon covered its bets by launching three other smartphones around the same time as the Storm, all with Wi-Fi (HTC Touch Pro, Samsung Omnia, Samsung Saga).

Before its launch, AT&T hoped that the iPhone would stimulate use of the cellular data network. It succeeded in this, so far beyond AT&T’s hopes that it revealed a potential problem with the concept of 3G (and 4G) data. The network slows to a crawl if enough subscribers use data intensively in small areas like airports and conferences. Mobile network operators used to fear that if phones had Wi-Fi subscribers would use it instead of the cellular data network, causing a revenue leak. AT&T solved that problem with the iPhone by making a subscription to the data service obligatory. T-Mobile followed suit with the Google phone. So no revenue leak. With the data subscription in hand, Wi-Fi is a good thing for the network operators because it offloads the 3G network. In residences and businesses all the data that goes through Wi-Fi is a reduction in the potential load on the network. In other words, a savings in infrastructure investment, which translates to profit. This may be some of the thinking behind AT&T’s recent acquisition of Wayport. The bandwidth acquired with Wayport offloads the AT&T network relatively cheaply. AT&T’s enthusiasm for Wi-Fi is such that it is selling some new Wi-Fi phones without requiring a data subscription.

The enterprise market is one that mobile network operators have long neglected. It is small relative to the consumer market, and harder to fit into a one-size-fits-all model. Even so, in these times of scraping for revenue in every corner, and with the steady rise of the Blackberry, the network operators are taking a serious look at the enterprise market.

The device manufacturers are way ahead of the network operators on this issue: the iPhone now comes with a lot of enterprise readiness Kool-Aid; Windows Mobile makes manageability representations, as does Nokia with its Eseries handsets. RIM, the current king of the enterprise smartphone vendors also pitches its IT-friendliness.

Wi-Fi in smartphones has benefits and drawbacks for enterprises. One benefit is that you have another smart device on the corporate LAN to enhance productivity. A drawback is that you have another smart device on the corporate LAN ripe for viruses and other security breaches. But that issue is mitigated to some extent if smartphones don’t have Wi-Fi. So it’s arguable that the Storm may be more enterprise-friendly as a result of its lack of Wi-Fi. Again, if the Storm becomes a hit in enterprises that argument will turn out to hold water. If the Storm is a flop for other reasons, we still won’t know, and it will have failed as a trial balloon for Wi-Fi-less enterprise smartphones.