New directions in data storage solutions

Jun 07, 2011 By Tiffany Fox

(PhysOrg.com) -- One day in the not-too-distant future, the entire contents of the Library of Congress might be stored on a device the size of a postage stamp.

It seems far-fetched, but computer engineers at the Center for Magnetic Recording Research (CMRR) at the UC San Diego Jacobs School of Engineering, are tucked away in a suite of windowless labs filled with whirring , working hard to harness emerging nanotechnologies to build and flash drives that can store data at unprecedented levels. And they’ll be doing so under new leadership, beginning next month.

Eric Fullerton, who is a professor of Electrical and Computer Engineering (ECE) and NanoEngineering (NE) at UC San Diego, is expected to replace outgoing CMRR director and ECE professor Paul Siegel, who served the center for 11 years and helped establish its reputation as one of the world’s leading research institutions for magnetic recording.

The very underpinnings of our digital society rely on research in this somewhat arcane field, which focuses on increasing the capacity of data in a data-saturated world.

“Magnetic storage media are widely used in information systems at all scales – from supercomputers to desktop workstations to smartphones,” explains Siegel. “It’s a low-cost form of non-volatile memory, meaning that it can retain information even when the computer is not powered on.”

As one of his final duties as director, Siegel led the CMRR in its twice-yearly research review. The invitation-only review took place earlier this month and provided an opportunity for faculty and students from the Center to meet with CMRR’s industry partners, including Hitachi, LSI, Marvell Semiconductor, NHK, Seagate Technologies, Toshiba, and Western Digital, as well as representatives of the Information Storage Industry Consortium and the recently established Advanced Storage Technology Consortium.

Says Siegel: “The idea is to give a series of 15-20-minute presentations that describe the work we do at CMRR. It’s a great way of not only exposing our research to industry reps, but exposing students to industry.”

Two students -- ECE Ph.D candidates Eitan Yaakobi and Uwe Boettcher -- were awarded the Sheldon Schultz Prize for Excellence in Graduate Student Research at the research review for their work in error-correction coding for flash memories and dual-stage actuators for magnetic storage devices. The prize, which was established in 2003 in conjunction with the Center’s 20th anniversary celebration, is named in honor of former CMRR Director, Sheldon Schultz, and is presented to CMRR graduate students who have distinguished themselves through the creativity of their research and the impact of their publications.

“We have a great track record of sending our students directly into industry jobs, and the industry as a whole has been very successful,” notes Siegel. “I think it’s fair to say that the people who have come out of this center and other centers like it have made a significant difference in the evolution of these technologies.”

It’s an evolution that has taken place with remarkable speed, with the density of hard drives increasing by a factor of about 5,000 in only the past 15 years -- from 100 megabits (100 million bits) per square inch to about 500 gigabits (500 billion bits) per square inch. Meanwhile, the cost per bit has dropped by a factor of 5,000, from $5 per megabyte to less than one-tenth of a cent per megabyte.

Siegel puts it another way: “IBM’s first disk drive, built in 1956, was the size of two refrigerators, cost more than $50,000 and held a total of 5 megabytes,” which is only enough storage capacity to handle about 30 seconds of broadcast-quality video. “Now you can buy a three-terabyte hard drive for $130 on Amazon.com” and have enough capacity to store the equivalent of an entire academic research library.

The growth of the data storage industry has been so explosive, in fact, that in many years, it has exceeded the trend in computing hardware known as Moore’s Law. Named for Intel Corp. co-founder Gordon Moore, the law predicts that the number of transistors that can be placed inexpensively on an integrated circuit will double approximately every two years.

By contrast, areal density of data storage has doubled on average every 18 months, and between 1997 and 2002, it doubled every year.

This cramming of more and more bits onto ever-shrinking hard drives has prompted the coining of a new law, known as Kryder’s law. The law, which is named for Seagate’s chief technology officer Mark Kryder, predicts that the growth of areal density will not follow a perfect logarithmic vertical scale, as is demonstrated by Moore’s Law, but will instead see several vertical jumps in growth as game-changing innovations are introduced.

Explains Siegel: “Growth jumped to 60 percent per year in the early ’80s with the introduction of several advances in magnetic recording, and when the Giant Magnetoresistive (GMR) Head was introduced, it jumped to 100 percent.”

CMRR’s incoming director, for one, was directly responsible for one of the field’s most important innovations. In 2001, while a researcher with IBM/Hitachi, Eric Fullerton helped invent a new type of magnetic medium called antiferromagnetically-coupled (AFC) media. AFC is a three-atom-thick layer of the element ruthenium sandwiched between two magnetic layers. This development alone made it possible for hard disk drives to store 100 billion bits (gigabits) of data per square inch of disk area by counteracting the superparamagnetic effect, which results when magnetic regions on a hard disk become too small and cannot retain data.

Fullerton’s advance made it possible for the upward arc of data storage capacity to progress, even as the devices themselves continued to become smaller. Siegel notes that this emphasis on ‘scaling’ has been another important advance in the field.

“The disks in hard drives used to be two feet across, and now they’re 2.5 inches across,” he points out. “We’re always striving to make things better, thinner and smaller. But because of the physics involved, we can’t scale things the way we used to.”

So what’s the theoretical physical limit for the density of data storage? It all depends on where and how the data are being stored. Now that the milestone of 100 Gigabits per square inch in commercial disk drive technology has been achieved and the most advanced hard drives can store up to 600 Gigabits per square inch, researchers at CMRR and elsewhere are striving toward -- and are only a factor of two away from -- the ambitious goal of 1 Terabit (1 trillion bits), which is thought to be the physical limit for data storage using current technologies.

This limit exists primarily because the magnetized areas on hard disks are becoming so small that it is difficult for them to remain stable as the read/write head flies across the surface of the disk. Flash drives use electrically charged transistors rather than moving parts to record information, making them a bit more stable (and better for mobile technologies, such as smartphones, says Fullerton) but also more expensive.

“Making predictions about the death of hard drives is a very old and prosperous field,” jokes Fullerton. “Being a factor of two away from what’s perceived as the limit is very daunting. What happens after we reach that threshold is that density will increase much more slowly, capacity will change much more slowly. To get around that, we have to do something new.”

Computer engineers at CMRR and elsewhere are now focusing their efforts on two distinct approaches for increasing capacity at a rate consistent with Kryder’s Law. One approach, called Bit Patterned Media, creates an array of tiny magnetic ‘islands’ -- each of which stores a single bit -- on the surface of hard disks. The challenge with this approach lies in the precise fabrication of the islands, which must now be so small, Fullerton says, that “engineers have to pattern the entire surface of the disk at a resolution of 10 nanometers [1/8000 the width of a human hair], plus do it quickly and cheaply.”

“Currently, the cost to directly make the pattern on one disk is about $2 million,” he notes. “Somehow we have to bring that cost down from $2 million to $1.”
Another approach, Heat Assisted (HAMR) maximizes density by harnessing a laser to write data onto a disk. But this is no small feat -- for the data to be written, the laser must heat a point 50 nanometers in size to 200 degrees Celsius in one nanosecond, and then cool it down in one nanosecond.

“That gap between nanoseconds is when the data are written” explains Fullerton.

He says that it’s not clear that either approach will work, but overcoming these initial hurdles in recording data at the nanoscale is a necessary step if computer engineers are to achieve the final frontier: Storing data near the atomic or molecular level. According to IBM, such a storage capability would enable nearly 30,000 feature length movies or the entire contents of YouTube – millions of videos estimated to be more than 1,000 trillion bits of data – to fit in a device the size of an iPod.

Although engineers are still a factor of 10,000 away from building data storage drives at the atomic level, the question now becomes: Do people actually need such gargantuan storage capability? Is it really necessary to carry around 30,000 feature-length films in one’s back pocket?

“I can remember when people would ask, ‘Why would you ever need a 1 gigabit hard drive?” recalls Fullerton. “But it’s true that there might come a point where the level of current storage capacity is ‘good enough.’ After all, nobody goes around saying, ‘Why aren’t airplanes any faster than they were 20 years ago?’ We could all travel around in planes that fly faster than the speed of sound, but that would be incredibly expensive, and not necessarily any more efficient.

“For all of this to go forward, either people have to be more efficient at storing data, or they have to store less of it.”

As their most recent research review demonstrated, at CMRR are actively working to overcome barriers related to cost, efficiency and stabilization as data storage moves into its next incarnation. Several presentations, for example, addressed new methods for not only determining the height of the read/write head, but also ways for controlling it.

Knowing the height of the head -- which Siegel describes as a ‘glorified electromagnet’ -- is crucial to hard drive functionality, says Siegel. “One of the worst things that could happen is if the head crashes. It could wipe out the disk.”

Presentations also focused on recording channels, the mechanics of disk and tape drives, high-speed high-resolution instrumentation, and simulations of the physics involved in the data-recording process.

There were also technical sessions featuring CMRR’s research on other non-volatile storage technologies, such as magnetic random access memory (MRAM), phase-change memory (PCM), and flash memory.  “These solid-state, non-mechanical memory technologies are increasingly important in applications ranging from consumer electronics to large-scale enterprise storage systems,” said Siegel.  “They require a complete rethinking about the way future data-intensive computers can most effectively harness the potential of these new storage technologies.”

CMRR is teaming up with other research groups on campus to do just that. Among them is the San Diego Supercomputer Center, which is leading a $15M NSF-funded effort to build a data-crunching computer, nicknamed “Gordon,” that relies on massive amounts of high-speed solid-state memory, in essence a huge flash drive.  There is also the Non-Volatile Systems Laboratory in the Computer Science and Engineering Department, where researchers are exploring performance characteristics of new solid-state storage devices and using their findings to help future computers handle data faster and in larger quantities.  Together with these centers and the California Institute for Telecommunications and Information Technology (Calit2), CMRR co-organized and hosted the recent UCSD Non-Volatile Memories Workshop (NVMW 2011), which attracted over 225 participants from academia, industry, and government research labs.

“With the technical expertise, innovative thinking, and collaborative spirit at CMRR and these other campus research centers, UC San Diego is arguably the premiere university world-wide in all aspects of advanced data storage technology and systems,” says Siegel. “This is a very exciting direction for the campus, and an exciting time for the storage industry as a whole.”

Explore further: Researchers propose network-based evaluation tool to assess relief operations feasibility

add to favorites email to friend print save as pdf

Related Stories

New device may revolutionize computer memory

Jan 20, 2011

(PhysOrg.com) -- Researchers from North Carolina State University have developed a new device that represents a significant advance for computer memory, making large-scale "server farms" more energy efficient and allowing ...

World record data density for ferroelectric recording

Aug 17, 2010

Scientists at Tohoku University in Japan have recorded data at a density of 4 trillion bits per square inch, which is a world record for the experimental "ferroelectric" data storage method. As described the journal Applied Ph ...

The kids are alright

May 26, 2011

Children should be seen and not heard... who says? A Philosophy academic at The University of Nottingham is challenging the adage by teaching primary school children to argue properly.

Recommended for you

Large streams of data warn cars, banks and oil drillers

17 hours ago

Better warning systems that alert motorists to a collision, make banks aware of the risk of losses on bad customers, and tell oil companies about potential problems with new drilling. This is the aim of AMIDST, the EU project ...

User comments : 2

Adjust slider to filter visible comments by rank

Display comments: newest first

dirk_bruere
not rated yet Jun 07, 2011
Mechanical memory systems are inelegant. The biggest holdup to more powerful computing is the bandwidth and access time of HDDs.
Beard
not rated yet Jun 08, 2011
It's true that most people would not make full use of a 1000TB HD today. But it won't happen today, it'll happen 12 years from now. As a reminder; 12 years ago was 1999, the huge advancements since then will continue onwards and I'm sure you'll need all 1000TB and more when the time comes.

More news stories

Simplicity is key to co-operative robots

A way of making hundreds—or even thousands—of tiny robots cluster to carry out tasks without using any memory or processing power has been developed by engineers at the University of Sheffield, UK.

Microsoft CEO is driving data-culture mindset

(Phys.org) —Microsoft's future strategy: is all about leveraging data, from different sources, coming together using one cohesive Microsoft architecture. Microsoft CEO Satya Nadella on Tuesday, both in ...

Floating nuclear plants could ride out tsunamis

When an earthquake and tsunami struck the Fukushima Daiichi nuclear plant complex in 2011, neither the quake nor the inundation caused the ensuing contamination. Rather, it was the aftereffects—specifically, ...

Patent talk: Google sharpens contact lens vision

(Phys.org) —A report from Patent Bolt brings us one step closer to what Google may have in mind in developing smart contact lenses. According to the discussion Google is interested in the concept of contact ...

Quantenna promises 10-gigabit Wi-Fi by next year

(Phys.org) —Quantenna Communications has announced that it has plans for releasing a chipset that will be capable of delivering 10Gbps WiFi to/from routers, bridges and computers by sometime next year. ...