Skip to main content

VirnetX Brings VPN Patent Battle to Apple, Cisco

Image used with permission by copyright holder

After winning a $200 million patent infringement settlement from Microsoft, VPN technology company is expanding its legal fusillade to include other technology giants: the company has filed suit in the patent holder-friendly Eastern Texas U.S. District Court against a series of companies for infringing on its VPN patents. The named defendants in the suit are Apple, Cisco, NEC, and Aastra.

The suit alleges the companies infringe on five VirnetX patents, and seeks unspecified injunctive relief and damages. The suit claims Apple mobile devices (the iPhone, iPad, and iPod touch) infringe on at least one of the patents (7,490,151) having to do with setting up secure communications using DNS. The claims against Aastra, NEC, and Cisco focus on VoIP products—including VoIP phones.

VirnetX acquired the patents in question when it acquired Science Applications International back in 2006. Science Applications mainly operated as a defense contractor. Shortly after the acquisition, VirnetX sued Microsoft for infringing on two of the company’s patents in Microsoft Office Communicator and Windows Meeting Space; VirnetX won the trial in March of this year and expanded the litigation to include Windows 7 before Microsoft agreed to a settlement.

Geoff Duncan
Former Digital Trends Contributor
Geoff Duncan writes, programs, edits, plays music, and delights in making software misbehave. He's probably the only member…
A dangerous new jailbreak for AI chatbots was just discovered
the side of a Microsoft building

Microsoft has released more details about a troubling new generative AI jailbreak technique it has discovered, called "Skeleton Key." Using this prompt injection method, malicious users can effectively bypass a chatbot's safety guardrails, the security features that keeps ChatGPT from going full Taye.

Skeleton Key is an example of a prompt injection or prompt engineering attack. It's a multi-turn strategy designed to essentially convince an AI model to ignore its ingrained safety guardrails, "[causing] the system to violate its operators’ policies, make decisions unduly influenced by a user, or execute malicious instructions," Mark Russinovich, CTO of Microsoft Azure, wrote in the announcement.

Read more