FCC to issue Net Neutrality rules

In a speech to the Brookings Institution today, FCC Chairman Julius Genachowski announced that the FCC is initiating a public process to formulate net neutrality rules for broadband network operators based on six principles:

  1. Open access to Content
  2. Open access to Applications
  3. Open access to Services
  4. Freedom for users to attach devices to the network
  5. Non-discrimination for content and applications
  6. Transparency of network management practices

The first four of these principles were initially articulated by former FCC Chairman Michael Powell in 2004 as the “Four Freedoms.” Numbers 5 and 6 are new. The forthcoming rules will apply these six principles to all broadband access technologies, including wireless.

Genachowski made the case that Internet openness is essential and that it is threatened. He acknowledged that network providers need to manage their networks, and said that they can control spam and help to maintain intellectual property integrity without compromising these principles.

The threats to Internet openness come from reduced competition among ISPs and conflicts of interest within the ISPs, because they are also trying to be content providers.

Genachowski rightly sees these threats as serious:

This is not about protecting the internet against imaginary dangers. We’re seeing the breaks and cracks emerge, and they threaten to change the Internet’s fundamental architecture of openness. This would shrink opportunities for innovators, content creators and small businesses around the country, and limit the full and free expression the internet promises. This is about preserving and maintaining something profoundly successful and ensuring that it’s not distorted or undermined.

These rules will be very tough to enforce. The fundamental structure of the business works against them. A more effective approach may be to break up the ISPs into multiple independent companies, for example: Internet access operations, wide area network operations, and service/content/application operations. The neutrality problem is in the access networks – the WANs and the services are healthier. With only the telcos (DSL and fiber) and the MSOs (cable) there is not enough competition for a free market to develop. This is why Intel pushed so hard for WiMAX as a third mode of broadband access, though it hasn’t panned out that way. It is also why municipal dark fiber makes sense, following the model of roads, water and sewers.

VoIP on the cellular data channel

In a recent letter to the FCC, AT&T said that it had no objection to VoIP applications on the iPhone that communicate over the Wi-Fi connection. It furthermore said:

Consistent with this approach, we plan to take a fresh look at possibly authorizing VoIP capabilities on the iPhone for use on AT&T’s 3G network.

So why would anybody want to do VoIP on the cellular data channel, when there is a cellular voice channel already? Wouldn’t voice on the data channel cost more? And since the voice channel is optimized for voice and the data channel isn’t, wouldn’t voice on the data channel sound even worse than cellular voice already does?

Let’s look at the “why bother?” question first. There are actually at least four reasons you might want to do voice on the cellular data channel:

  1. To save money. If your voice plan has some expensive types of call (for example international calls) you may want to use VoIP on the data channel for toll by-pass. The alternative to this is to use the voice channel to call a local access number for an international toll by-pass service (like RebTel.)
  2. To get better sound quality: the cellular voice codecs are very low bandwidth and sound horrible. You can choose which codec to run over the data network and even go wideband. At IT Expo West a couple of weeks ago David Frankel of ZipDX demoed a wideband voice call on his laptop going through a Sprint Wireless Data Card. The audio quality was excellent.
  3. To get additional service features: companies like DiVitas offer roaming between the cellular and Wi-Fi networks that makes your cell phone act as an extension behind your corporate PBX. All these solutions currently use the cellular voice channel when out of Wi-Fi range, but if they were to go to the data channel they could offer wideband codecs and other differentiating features.
  4. For cases where there is no voice channel. In the example of David Frankel’s demo, the wireless data card doesn’t offer a voice channel, so VoIP on the data channel is the only option for a voice connection.

Moving on to the issue of cost, an iPhone unlimited data plan is $30 per month. “Unlimited” is AT&T’s euphemism for “limited to 5GB per month,” but translated to voice that’s a lot of minutes: even with IP packet overhead the bit-rate of compressed HD voice is going to be around 50K bits per second, which works out to about 13,000 minutes in 5GB. So using it for voice is unlikely to increase your bill. On the other hand, many voice plans are already effectively unlimited, what with rollover minutes, friend and family minutes, night and weekend minutes and whatnot, and you can’t get a phone without a voice plan. So for normal (non-international) use voice on the data channel is not going to reduce your bill, but it is unlikely to increase it, either.

Finally we come to the issue of whether voice sounds better on the voice channel or the data channel. The answer is, it depends on several factors, primarily the codec and the network QoS. With VoIP you can radically improve the sound quality of a call by using a wideband codec, but do impairments on the data channel nullify this benefit?

Technically, the answer is yes. The cellular data channel is not engineered for low latency. Variable delays are introduced by network routing decisions and by router queuing decisions. Latencies in the hundreds of milliseconds are not unusual. This will change with the advent of LTE, where the latencies will be of the order of 10 milliseconds. The available bandwidth is also highly variable, in contrast to the fixed bandwidth allocation of the voice channel. It can sometimes drop below what is needed for voice with even an aggressive variable rate codec.

In practice VoIP on the cellular data channel can sometimes sound much better than regular cellular voice. I mentioned above David Frankel’s demo at IT Expo West. I performed a similar experiment this morning with Michael Graves, with similarly good results. I was on a Polycom desk phone, Michael used Eyebeam on a laptop, and the codec was G.722. The latency on this call was appreciable – I estimated it at around 1 second round trip. There was also some packet loss – not bad for me, but it caused a sub-par experience for Michael. Earlier this week at Jeff Pulver’s HD Connect conference in New York, researchers from Qualcomm demoed a handset running on the Verizon network using EVRC-WB, transcoding to G.722 on Polycom and Gigaset phones in their lab in San Diego. The sound quality was excellent, but the latency was very high – I estimated it at around two seconds round trip.

The ITU addresses latency (delay) in Recommendation G.114. Delay is a problem because normal conversation depends on turn taking. Most people insert pauses of up to about 400 ms as they talk. If nobody else speaks during a pause, they continue. This means that if the one-way delay on a phone conversation is greater than 200 ms, the talker doesn’t hear an interruption within the 400 ms break, and starts talking again, causing frustrating collisions.
The ITU E-Model for call quality identifies a threshold at about 170 ms one-way at which latency becomes a problem. The E-Model also tells us that increasing latency amplifies other impairments – notably echo, which can be severe at low latencies without being a problem, but at high latencies even relatively quiet echo can severely disrupt a talker.

Some people may be able to handle long latencies better than others. Michael observed that he can get used to high latency echo after a few minutes of conversation.

Nokia no longer the only VoWi-Fi friendly phone maker

Until now, Nokia has been top of the heap in the category of VoIP-friendliness. When I spoke with Richard Watson, CTO of DiVitas, last year in the course of my test drive of the DiVitas system, he pointed out that dual-mode phones are not normally VoIP-friendly. At that time the only phone he recommended was the Nokia E71. There are several reasons for this, primarily the treatment of the voice path and the ease of integration of the VoIP software with the built-in phone software user interfaces. Since then, DiVitas has been working closely with Samsung, and now Richard says several Samsung phones are well suited to Voice over Wi-Fi. Let’s hope this shakes the other phone OEMs loose and gets them working on improving Voice over Wi-Fi performance.

Apple’s App-roval process

I wrote earlier about AT&T’s responses to FCC’s questions concerning the iPhone App Store and Google Voice.

Now Apple has posted its responses to the same questions, which are basically the same as AT&T’s. Among the differences are that Apple’s responses contain some hard numbers on its controversial App Store approval process:

  • 80% of applications are approved as originally submitted.
  • 95% of applications are approved within 14 days of submission.
  • 65,000 applications have been approved.
  • 200,000 submissions and re-submissions have been made.
  • 8,500 submissions are coming in each week.
  • Each submission is reviewed by two reviewers.
  • There are 40 reviewers.

These numbers don’t really add up. So what Apple probably means is that 95% of the applications that have been approved were approved within 14 days of their final submission. Even so, each reviewer must look at an average of 425 submissions per week (8,500*2/40), which is 10 per hour per reviewer – an average of 12 minutes of reviewer time per submission, which doesn’t seem to justify the terms “comprehensive” and “rigorous” used in Apple’s description of the process:

Apple developed a comprehensive review process that looks at every iPhone application that is submitted to Apple. Applications and marketing text are submitted through a web interface. Submitted applications undergo a rigorous review process that tests for vulnerabilities such as software bugs, instability on the iPhone platform, and the use of unauthorized protocols. Applications are also reviewed to try to prevent privacy issues, safeguard children from exposure to inappropriate content, and avoid applications that degrade the core experience of the iPhone. There are more than 40 full-time trained reviewers, and at least two different reviewers study each application so that the review process is applied uniformly. Apple also established an App Store executive review board that determines procedures and sets policy for the review process, as well as reviews applications that are escalated to the board because they raise new or complex issues. The review board meets weekly and is comprised of senior management with responsibilities for the App Store. 95% of applications are approved within 14 days of being submitted.

Of course much of this might be automated, which would explain both the superhuman productivity of the reviewers and the alleged mindlessness of the decision-making.

Dual-mode technology maturing

The Rethink Wireless newsletter is always worth reading. An article in today’s edition says that according to ABI dual mode handset shipments are on track to double from 2008 to 2010, and more than double from 2009-2011 (144 million units to 300 million units).

Rethink’s Matt Lewis cites improved performance and usability as driving forces, plus a change in the attitudes of carriers towards hot-spots. Wireless network operators now often have captive Wi-Fi networks and can use them to offload their cellular networks.

The upshot is a prediction of 300 million dual mode handsets to ship in 2011: 100% of the smartphone market plus high end feature phones.

The attach rate of Wi-Fi will continue to grow. By 2011 the effects of Bluetooth 3.0 will be kicking in, pushing Wi-Fi attachment towards 100% in camera phones and music phones in ensuing years.

AT&T, Apple and VoIP on the iPhone

The phone OEMs are customer-driven, and I mean that in a bad way. They view service providers rather than consumers as their customers, and therefore have historically tended to be relatively uninterested in ease of use or performance, concentrating on packing in long checklists of features, many of which went unused by baffled consumers. Nokia seemed to have factions that were more user-oriented, but it took the chutzpah of Steve Jobs to really change the game.

A recent FCC inquiry has provoked a fascinating letter from AT&T on the background of the iPhone and AT&T’s relationship with Apple, including Voice over IP on the iPhone. On the topic of VoIP, the letter says that AT&T bound Apple to not create a VoIP capability for the iPhone, but Apple did not commit to prevent third parties from doing so. AT&T says that it never had any objection to iPhone VoIP applications that run over Wi-Fi, and that it is currently reconsidering its opposition to VoIP applications that run over the 3G data connection. Since the argument that AT&T presents in the letter in favor of restrictions on VoIP is weak, such a reconsideration seems in order.

The argument goes as follows: the explosion of the mobile Internet led by the iPhone was catalyzed by cheap iPhones. iPhones are cheap because of massive subsidies. The subsidies are paid for by the voice services. Therefore, AT&T is justified in protecting its voice service revenues because the subsidies they allow had such a great result: the flourishing of the mobile Internet. The reason this argument is weak is that voice service revenues are not the only way to recoup subsidies. AT&T has discovered that it can charge for the mobile Internet directly, and recoup its subsidies that way. It will not sell a subsidized iPhone without an unlimited data plan, and it increased the price of that mandatory plan by 50% last year. Even with this price increase iPhone sales continued to burgeon. In other words, AT&T may be able to recoup lost voice revenues by charging more for its data services.

This is exactly what the “dumb pipes” crowd has been advocating for over a decade now: connectivity providers should charge a realistic price for connectivity, and not try to subsidize it with unrealistic charges for other services.

HD Communications Project

As part of the preparation for the fall HD Communications Summit, Jeff Pulver has put up a video clip promoting HD Voice for phone calls. It goes over the familiar arguments:

  • Sound quality on phone calls hasn’t improved since 1937. Since most calls are now made on cell phones, it has actually deteriorated considerably.
  • The move to VoIP has made it technically feasible to make phone calls with CD quality sound or better, yet instead VoIP calls are usually engineered to sound worse than circuit-switched calls (except in the case of Skype.)
  • Improved sound quality on phone calls yields undisputed productivity benefits, particularly when the calls involve multiple accents.
  • Voice has become a commodity service, with minimal margins for service providers, yet HD Voice offers an opportunity for differentiation and potentially improved margins.

The HD Communications Summit is part of the HD Connect Project. The HD Connect Project aims to provide a coordination point for the various companies that have an interest in propagating HD Voice. These companies include equipment and component manufacturers, software developers and service providers.

Among the initiatives of the HD Connect Project is a logo program, like the Wi-Fi Alliance logo program. The logo requirements are currently technically lax, providing an indicator of good intentions rather than certain interoperability. Here’s a draft of the new logo:

HD Connect Draft Logo

Another ingredient of the HD Connect project is the HDConnectNow.org website, billed as “the news and information place for The HD Connect Project.”

It is great that Jeff is stepping up to push HD Voice like this. With the major exception of Skype almost no phone calls are made with wideband codecs (HD Voice). Over the past few years the foundation has been laid for this to change. Several good wideband codecs are now available royalty free, and all the major business phone manufacturers sell mostly (or solely) wideband-capable phones. Residential phones aren’t there yet, but this will change rapidly: the latest DECT standards are wideband, Gigasets are already wideband-capable, and Uniden is enthusiastic about wideband, too. As the installed base of wideband-capable phones grows, wideband calling can begin to happen.

Since most dialing is still done with old-style (E.164) phone numbers, wideband calls will become common within companies before there is much uptake between companies. That will come as VoIP trunking displaces circuit-switched, and as ENUM databases are deployed and used.

Tsera, LLC v. Apple Inc. et al. Patent troll?

Here’s a little fairy-tale about what might have happened: Chuang Li took an idea to his boss at Actiontec who determined that it wasn’t of interest to the company and told him the idea was all his if he wanted to pursue it. Chuang ultimately refined the idea into the user interface that would become ubiquitous on MP3 players. The Patent Office rejected his application on a technicality. Chuang labored for years patiently jumping through hoops for the patent examiner while educating him on the validity of his claims. When the patent was finally issued, Chuang took it to Apple and requested a reasonable compensation for his idea. Unable to reach agreement with Apple after five years of effort, Chuang found a reputable firm of New York lawyers who were willing to take the case on a contingency basis.

Now here are some facts: the slashdotscape is alive with outrage about another patent lawsuit, this time filed by a company called Tsera LLC against Apple and 18 other companies over a touchpad interface to personal media player type devices (iPods).

Tsera was formed a couple of weeks ago on July 10th, and has filed no ownership or officer information with the Texas Secretary of State. Its registered agent is National Registered Agents, Inc. of New Jersey. Five days after the company was formed, Chuang Li, the inventor of US patent number 6,639,584 assigned that patent to Tsera. That same day Tsera filed suit in the notorious U.S. District Court for the Eastern District of Texas. Tsera’s attorneys are Kaye Scholer LLP of New York (specifically James S. Blank, Patricia A. Carson, Leora Ben-Ami, Oliver C. Bennett and Tsung-Lin Fu.) Tsera’s local counsel are Jack Wesley Hill and Otis W. Carroll of Ireland, Carroll and Kelley, P.C. of Tyler Texas.

The complaint reveals that Tsera has no parent corporation, and that no public company owns 10% or more of its stock.

The original application for patent 6,639,584 was made exactly ten years ago, on July 29, 1999. At that time (and still) the inventor, Chuang Li was apparently working for Actiontec Electronics, Inc. of Sunnyvale, CA. Chuang Li did not assign the patent to Actiontec, although other patents he applied for around that time were assigned to Actiontec. Companies like Actiontec normally require their workers to assign all the intellectual property they generate, especially when it is relevant to their business. Actiontec makes an MP3 player called the PocketRave.

Chuang Li’s application was rejected on October 9, 2001, just two weeks before Apple launched the first iPod (which did not have a touch-sensitive interface.) Three months later, on 28th January 2002, Chuang Li submitted an amended application, which was again rejected, in November 2002. In May of 2003 Chuang Li appealed the rejection and submitted another amendment. The appeal was ultimately successful, and the patent was issued on the 9th of October, 2003.

Looking at this chronology, you can see that the patent application was amended after the July 2002 launch of the touch-wheel iPod. Numerous rejections and resubmissions are common in the patent process, but they can be symptomatic of a “submarine patent,” where an inventor (famously Jerome H. Lemelson) files a vague patent and tweaks it over the course of several years to make it apply to some successful product that has appeared in the interim. The most egregious type of patent trolling is when the patent at issue is meritless, but the troll demands a settlement that the defending company determines is cheaper to pay than to go to court over.

The Tsera patent doesn’t cite as prior art Xerox’s US Patent 5596656, filed in 1995 and issued in January 1997. The basic idea of the Xerox patent is to replace a keyboard by forming strokes on a touch-pad, while the basic idea of the Tsera patent is to replace buttons and knobs on a portable electronic device by forming strokes on a touch-pad. The Xerox patent has a system of “unistrokes” on a touch-sensitive surface that can be performed “eyes free” and in which unistroke symbols can “correlate with user invokeable control functions.” The Tsera patent has a “device controlled by a user tracing a command pattern on the touch-sensitive surface with a finger,” “without requiring the user to view the portable electronic device,” with each of the “patterns corresponding to a predefined function of the portable electronic device.”

These descriptions actually apply better to the iPhone and the iPod Touch than they do to the canonical iPod touch-wheel, where the annular touch sensitive surface doesn’t really accommodate free-form strokes.

The Xerox patent was the subject of extensively reported litigation running from 1997 to 2006.

The Xerox patent, which is prior art to the Tsera patent, includes free-form touchpad strokes used as control functions. I am not a lawyer, but it seems to me that application of this same idea to portable devices would have to be non-obvious or the term “control functions” would have to be narrowly defined in order for the Tsera patent to be valid.

Is Tsera acting as a patent troll? You be the judge.

GIPS webinar on Wideband Voice

GIPS yesterday sponsored a webinar on HD voice in the teleconferencing and videoconferencing industries. It was introduced by Elliot Gold of Telespan Publishing. Elliot made the point that HD video will be running at 90% of video equipment sales by the end of this year, and that HD audio is at least equally important to the user experience and should be the next technology transformation in the equipment and services markets.

Dovid Coplon of GIPS gave a more technical presentation, starting with a very accessible explanation of the benefits of HD audio.

Dovid made some interesting observations. He named some users of GIPS wideband technology including Google Talk and Yahoo Messenger. He cited a May 2009 poll that showed 10% of respondents used HD audio all the time or whenever they could. To me this is a surprisingly high number, possibly reflecting a biased poll sample, or perhaps a large number of respondents that didn’t understand the question.

Dovid compared the benefits of the GIPS wideband codecs, iSAC and iPCM-WB. iSAC is a lower-bit-rate, higher complexity codec than IPCM-WB. Dovid pointed out that with IP packet overhead being so high, decreasing the bit-rate of a codec is not as useful as one might think. The implication was that the lower complexity of IPCM-WB outweighed its bandwidth disadvantage in mobile applications. He also included a codec quality comparison chart based on a MUSHRA scale. The chart predictably showed iSAC way better than any of the others, and anomalously showed Speex at 24 kbps as inferior to Speex at 16 kbps.

Dovid also echoed all the codec engineers I have talked with in emphasizing that the codec is a small piece of the audio quality puzzle, and that network and physical acoustic impairments can have a greater effect on user experience.

You can download the slides here.