Skip to main content

Nvidia re-enables lock on GTX 980M clock speeds in recent driver updates

nvidia 980m clockblock driver gtx980m 3qtr
Image used with permission by copyright holder
Nvidia has puzzled and angered its gaming notebook community once again, by re-initiating a block that prevents owners of laptops featuring the GTX 980M GPU from overclocking it. Added through the recent 350.12 driver update, the move is being dubbed the “clockblock” by gamers, with many wondering why Nvidia feels the need to hand-hold users with such powerful machines.

This particular block isn’t hitting everyone who has a 980M GPU, as some have stuck with an older driver that disabled the block (after it was brought in on NVidia’s 347.09 update, as per NotebookReview) and not everyone’s gaming laptops shipped with the locked GPU BIOS that makes overclocking of the chip impossible. However for a large number of gamers, particularly those that are keen to keep their drivers as updated as possible, GPU overclocking has now been disabled once again.

This seems rather bizarre in the contemporary climate of high-end systems and tweaking, as many of them are sold on the basis that they can be overclocked by end users. Buyers of high-end gaming laptops featuring the powerful 980M chip want maximum performance. While Nvidia might not want anyone doing damage to their presumably already heavily tweaked system by going overboard, most users appreciate that once you overclock something you’ve waved goodbye to your warranty, anyway.

Get your weekly teardown of the tech behind PC gaming
Check your inbox!

Nvidia hasn’t released a statement addressing the matter. Clearly, the company isn’t backing down on this issue.

Jon Martindale
Jon Martindale is the Evergreen Coordinator for Computing, overseeing a team of writers addressing all the latest how to…
A dangerous new jailbreak for AI chatbots was just discovered
the side of a Microsoft building

Microsoft has released more details about a troubling new generative AI jailbreak technique it has discovered, called "Skeleton Key." Using this prompt injection method, malicious users can effectively bypass a chatbot's safety guardrails, the security features that keeps ChatGPT from going full Taye.

Skeleton Key is an example of a prompt injection or prompt engineering attack. It's a multi-turn strategy designed to essentially convince an AI model to ignore its ingrained safety guardrails, "[causing] the system to violate its operators’ policies, make decisions unduly influenced by a user, or execute malicious instructions," Mark Russinovich, CTO of Microsoft Azure, wrote in the announcement.

Read more