Skip to main content

New QLC flash memory from Toshiba could push SSD capacity as high as 128TB

toshiba ups its storage game with reveal of 4 bit mlc memory qlc roadmap
Marc Sauter / Golem.de
The solid state drive news continues to come fast and furious from this year’s Flash Memory Summit, where the world’s leading flash memory manufacturers are unveiling the latest technology. One tech innovation that’s not gaining much attention, but that could greatly increase data densities, is Toshiba’s Quadruple Level Cell.

Many of today’s solid state drives use a type of flash called Triple Level Cell (TLC) or, interchangeably, 3-bit Multi-level Cell (3-bit MLC). This kind of flash can store 3 bits of data per cell. As the name implies, Toshiba’s technology boosts that to 4 bits per cell.

That permits every memory cell in a QLC drive to store more data, resulting in higher data densities. Toshiba believes that QLC and other technologies will drastically increase SSD drive size in the near future, resulting in capacities of up to 128TB by 2018.

The announcement of QLC seems to have come as a bit of a surprise, largely because it’s extremely complex. In order to make this possible, Toshiba had to find a way to handle 16 narrow voltage ranges that represent data stored by the drive, and that doesn’t leave much room for variation or error.

While QLC may enable extremely high data densities, it does appear to have notable disadvantages. Drive endurance is lower than other forms of flash memory, as the sensitive cells wear out more quickly, and performance is not expected to be on par with current SSDs. As a result, Toshiba is positioning the technology as a solution for archival storage – big, affordable drives that aren’t often used.

Whatever their use, the drives won’t be coming to stores anytime soon. Toshiba’s current roadmap suggests they’ll come out sometime late in 2016. As you might imagine, given the timeframe, pricing is also unavailable.

Matthew S. Smith
Matthew S. Smith is the former Lead Editor, Reviews at Digital Trends. He previously guided the Products Team, which dives…
A dangerous new jailbreak for AI chatbots was just discovered
the side of a Microsoft building

Microsoft has released more details about a troubling new generative AI jailbreak technique it has discovered, called "Skeleton Key." Using this prompt injection method, malicious users can effectively bypass a chatbot's safety guardrails, the security features that keeps ChatGPT from going full Taye.

Skeleton Key is an example of a prompt injection or prompt engineering attack. It's a multi-turn strategy designed to essentially convince an AI model to ignore its ingrained safety guardrails, "[causing] the system to violate its operators’ policies, make decisions unduly influenced by a user, or execute malicious instructions," Mark Russinovich, CTO of Microsoft Azure, wrote in the announcement.

Read more