Tag Archives: archives

More memory, less space and a walk down the cryptocurrency road

Libraries, archives, records management, oral history, etc. there are many institutions and names for how we manage collective and personal memory. You might call it a peculiarly human obsession stretching back into antiquity. For example, there’s the Library of Alexandria (Wikipedia entry) founded in the third, or possibly 2nd, century BCE (before the common era) and reputed to store all the knowledge in the world. It was destroyed although accounts differ as to when and how but its loss remains a potent reminder of memory’s fragility.

These days, the technology community is terribly concerned with storing ever more bits of data on materials that are reaching their limits for storage.I have news of a possible solution,  an interview of sorts with the researchers working on this new technology, and some very recent research into policies for cryptocurrency mining and development. That bit about cryptocurrency makes more sense when you read what the response to one of the interview questions.

Memory

It seems University of Alberta researchers may have found a way to increase memory exponentially, from a July 23, 2018 news item on ScienceDaily,

The most dense solid-state memory ever created could soon exceed the capabilities of current computer storage devices by 1,000 times, thanks to a new technique scientists at the University of Alberta have perfected.

“Essentially, you can take all 45 million songs on iTunes and store them on the surface of one quarter,” said Roshan Achal, PhD student in Department of Physics and lead author on the new research. “Five years ago, this wasn’t even something we thought possible.”

A July 23, 2018 University of Alberta news release (also on EurekAlert) by Jennifer-Anne Pascoe, which originated the news item, provides more information,

Previous discoveries were stable only at cryogenic conditions, meaning this new finding puts society light years closer to meeting the need for more storage for the current and continued deluge of data. One of the most exciting features of this memory is that it’s road-ready for real-world temperatures, as it can withstand normal use and transportation beyond the lab.

“What is often overlooked in the nanofabrication business is actual transportation to an end user, that simply was not possible until now given temperature restrictions,” continued Achal. “Our memory is stable well above room temperature and precise down to the atom.”

Achal explained that immediate applications will be data archival. Next steps will be increasing readout and writing speeds, meaning even more flexible applications.

More memory, less space

Achal works with University of Alberta physics professor Robert Wolkow, a pioneer in the field of atomic-scale physics. Wolkow perfected the art of the science behind nanotip technology, which, thanks to Wolkow and his team’s continued work, has now reached a tipping point, meaning scaling up atomic-scale manufacturing for commercialization.

“With this last piece of the puzzle now in-hand, atom-scale fabrication will become a commercial reality in the very near future,” said Wolkow. Wolkow’s Spin-off [sic] company, Quantum Silicon Inc., is hard at work on commercializing atom-scale fabrication for use in all areas of the technology sector.

To demonstrate the new discovery, Achal, Wolkow, and their fellow scientists not only fabricated the world’s smallest maple leaf, they also encoded the entire alphabet at a density of 138 terabytes, roughly equivalent to writing 350,000 letters across a grain of rice. For a playful twist, Achal also encoded music as an atom-sized song, the first 24 notes of which will make any video-game player of the 80s and 90s nostalgic for yesteryear but excited for the future of technology and society.

As noted in the news release, there is an atom-sized song, which is available in this video,

As for the nano-sized maple leaf, I highlighted that bit of whimsy in a June 30, 2017 posting.

Here’s a link to and a citation for the paper,

Lithography for robust and editable atomic-scale silicon devices and memories by Roshan Achal, Mohammad Rashidi, Jeremiah Croshaw, David Churchill, Marco Taucer, Taleana Huff, Martin Cloutier, Jason Pitters, & Robert A. Wolkow. Nature Communicationsvolume 9, Article number: 2778 (2018) DOI: https://doi.org/10.1038/s41467-018-05171-y Published 23 July 2018

This paper is open access.

For interested parties, you can find Quantum Silicon (QSI) here. My Edmonton geography is all but nonexistent, still, it seems to me the company address on Saskatchewan Drive is a University of Alberta address. It’s also the address for the National Research Council of Canada. Perhaps this is a university/government spin-off company?

The ‘interview’

I sent some questions to the researchers at the University of Alberta who very kindly provided me with the following answers. Roshan Achal passed on one of the questions to his colleague Taleana Huff for her response. Both Achal and Huff are associated with QSI.

Unfortunately I could not find any pictures of all three researchers (Achal, Huff, and Wolkow) together.

Roshan Achal (left) used nanotechnology perfected by his PhD supervisor, Robert Wolkow (right) to create atomic-scale computer memory that could exceed the capacity of today’s solid-state storage drives by 1,000 times. (Photo: Faculty of Science)

(1) SHRINKING THE MANUFACTURING PROCESS TO THE ATOMIC SCALE HAS
ATTRACTED A LOT OF ATTENTION OVER THE YEARS STARTING WITH SCIENCE
FICTION OR RICHARD FEYNMAN OR K. ERIC DREXLER, ETC. IN ANY EVENT, THE
ORIGINS ARE CONTESTED SO I WON’T PUT YOU ON THE SPOT BY ASKING WHO
STARTED IT ALL INSTEAD ASKING HOW DID YOU GET STARTED?

I got started in this field about 6 years ago, when I undertook a MSc
with Dr. Wolkow here at the University of Alberta. Before that point, I
had only ever heard of a scanning tunneling microscope from what was
taught in my classes. I was aware of the famous IBM logo made up from
just a handful of atoms using this machine, but I didn’t know what
else could be done. Here, Dr. Wolkow introduced me to his line of
research, and I saw the immense potential for growth in this area and
decided to pursue it further. I had the chance to interact with and
learn from nanofabrication experts and gain the skills necessary to
begin playing around with my own techniques and ideas during my PhD.

(2) AS I UNDERSTAND IT, THESE ARE THE PIECES YOU’VE BEEN
WORKING ON: (1) THE TUNGSTEN MICROSCOPE TIP, WHICH MAKE[s] (2) THE SMALLEST
QUANTUM DOTS (SINGLE ATOMS OF SILICON), (3) THE AUTOMATION OF THE
QUANTUM DOT PRODUCTION PROCESS, AND (4) THE “MOST DENSE SOLID-STATE
MEMORY EVER CREATED.” WHAT’S MISSING FROM THE LIST AND IS THAT WHAT
YOU’RE WORKING ON NOW?

One of the things missing from the list, that we are currently working
on, is the ability to easily communicate (electrically) from the
macroscale (our world) to the nanoscale, without the use of a scanning
tunneling microscope. With this, we would be able to then construct
devices using the other pieces we’ve developed up to this point, and
then integrate them with more conventional electronics. This would bring
us yet another step closer to the realization of atomic-scale
electronics.

(3) PERHAPS YOU COULD CLARIFY SOMETHING FOR ME. USUALLY WHEN SOLID STATE
MEMORY IS MENTIONED, THERE’S GREAT CONCERN ABOUT MOORE’S LAW. IS
THIS WORK GOING TO CREATE A NEW LAW? AND, WHAT IF ANYTHING DOES
;YOUR MEMORY DEVICE HAVE TO DO WITH QUANTUM COMPUTING?

That is an interesting question. With the density we’ve achieved,
there are not too many surfaces where atomic sites are more closely
spaced to allow for another factor of two improvement. In that sense, it
would be difficult to improve memory densities further using these
techniques alone. In order to continue Moore’s law, new techniques, or
storage methods would have to be developed to move beyond atomic-scale
storage.

The memory design itself does not have anything to do with quantum
computing, however, the lithographic techniques developed through our
work, may enable the development of certain quantum-dot-based quantum
computing schemes.

(4) THIS MAY BE A LITTLE OUT OF LEFT FIELD (OR FURTHER OUT THAN THE
OTHERS), COULD;YOUR MEMORY DEVICE HAVE AN IMPACT ON THE
DEVELOPMENT OF CRYPTOCURRENCY AND BLOCKCHAIN? IF SO, WHAT MIGHT THAT
IMPACT BE?

I am not very familiar with these topics, however, co-author Taleana
Huff has provided some thoughts:

Taleana Huff (downloaded from https://ca.linkedin.com/in/taleana-huff]

“The memory, as we’ve designed it, might not have too much of an
impact in and of itself. Cryptocurrencies fall into two categories.
Proof of Work and Proof of Stake. Proof of Work relies on raw
computational power to solve a difficult math problem. If you solve it,
you get rewarded with a small amount of that coin. The problem is that
it can take a lot of power and energy for your computer to crunch
through that problem. Faster access to memory alone could perhaps
streamline small parts of this slightly, but it would be very slight.
Proof of Stake is already quite power efficient and wouldn’t really
have a drastic advantage from better faster computers.

Now, atomic-scale circuitry built using these new lithographic
techniques that we’ve developed, which could perform computations at
significantly lower energy costs, would be huge for Proof of Work coins.
One of the things holding bitcoin back, for example, is that mining it
is now consuming power on the order of the annual energy consumption
required by small countries. A more efficient way to mine while still
taking the same amount of time to solve the problem would make bitcoin
much more attractive as a currency.”

Thank you to Roshan Achal and Taleana Huff for helping me to further explore the implications of their work with Dr. Wolkow.

Comments

As usual, after receiving the replies I have more questions but these people have other things to do so I’ll content myself with noting that there is something extraordinary in the fact that we can imagine a near future where atomic scale manufacturing is possible and where as Achal says, ” … storage methods would have to be developed to move beyond atomic-scale [emphasis mine] storage”. In decades past it was the stuff of science fiction or of theorists who didn’t have the tools to turn the idea into a reality. With Wolkow’s, Achal’s, Hauff’s, and their colleagues’ work, atomic scale manufacturing is attainable in the foreseeable future.

Hopefully we’ll be wiser than we have been in the past in how we deploy these new manufacturing techniques. Of course, before we need the wisdom, scientists, as  Achal notes,  need to find a new way to communicate between the macroscale and the nanoscale.

As for Huff’s comments about cryptocurrencies and cyptocurrency and blockchain technology, I stumbled across this very recent research, from a July 31, 2018 Elsevier press release (also on EurekAlert),

A study [behind a paywall] published in Energy Research & Social Science warns that failure to lower the energy use by Bitcoin and similar Blockchain designs may prevent nations from reaching their climate change mitigation obligations under the Paris Agreement.

The study, authored by Jon Truby, PhD, Assistant Professor, Director of the Centre for Law & Development, College of Law, Qatar University, Doha, Qatar, evaluates the financial and legal options available to lawmakers to moderate blockchain-related energy consumption and foster a sustainable and innovative technology sector. Based on this rigorous review and analysis of the technologies, ownership models, and jurisdictional case law and practices, the article recommends an approach that imposes new taxes, charges, or restrictions to reduce demand by users, miners, and miner manufacturers who employ polluting technologies, and offers incentives that encourage developers to create less energy-intensive/carbon-neutral Blockchain.

“Digital currency mining is the first major industry developed from Blockchain, because its transactions alone consume more electricity than entire nations,” said Dr. Truby. “It needs to be directed towards sustainability if it is to realize its potential advantages.

“Many developers have taken no account of the environmental impact of their designs, so we must encourage them to adopt consensus protocols that do not result in high emissions. Taking no action means we are subsidizing high energy-consuming technology and causing future Blockchain developers to follow the same harmful path. We need to de-socialize the environmental costs involved while continuing to encourage progress of this important technology to unlock its potential economic, environmental, and social benefits,” explained Dr. Truby.

As a digital ledger that is accessible to, and trusted by all participants, Blockchain technology decentralizes and transforms the exchange of assets through peer-to-peer verification and payments. Blockchain technology has been advocated as being capable of delivering environmental and social benefits under the UN’s Sustainable Development Goals. However, Bitcoin’s system has been built in a way that is reminiscent of physical mining of natural resources – costs and efforts rise as the system reaches the ultimate resource limit and the mining of new resources requires increasing hardware resources, which consume huge amounts of electricity.

Putting this into perspective, Dr. Truby said, “the processes involved in a single Bitcoin transaction could provide electricity to a British home for a month – with the environmental costs socialized for private benefit.

“Bitcoin is here to stay, and so, future models must be designed without reliance on energy consumption so disproportionate on their economic or social benefits.”

The study evaluates various Blockchain technologies by their carbon footprints and recommends how to tax or restrict Blockchain types at different phases of production and use to discourage polluting versions and encourage cleaner alternatives. It also analyzes the legal measures that can be introduced to encourage technology innovators to develop low-emissions Blockchain designs. The specific recommendations include imposing levies to prevent path-dependent inertia from constraining innovation:

  • Registration fees collected by brokers from digital coin buyers.
  • “Bitcoin Sin Tax” surcharge on digital currency ownership.
  • Green taxes and restrictions on machinery purchases/imports (e.g. Bitcoin mining machines).
  • Smart contract transaction charges.

According to Dr. Truby, these findings may lead to new taxes, charges or restrictions, but could also lead to financial rewards for innovators developing carbon-neutral Blockchain.

The press release doesn’t fully reflect Dr. Truby’s thoughtfulness or the incentives he has suggested. it’s not all surcharges, taxes, and fees constitute encouragement.  Here’s a sample from the conclusion,

The possibilities of Blockchain are endless and incentivisation can help solve various climate change issues, such as through the development of digital currencies to fund climate finance programmes. This type of public-private finance initiative is envisioned in the Paris Agreement, and fiscal tools can incentivize innovators to design financially rewarding Blockchain technology that also achieves environmental goals. Bitcoin, for example, has various utilitarian intentions in its White Paper, which may or may not turn out to be as envisioned, but it would not have been such a success without investors seeking remarkable returns. Embracing such technology, and promoting a shift in behaviour with such fiscal tools, can turn the industry itself towards achieving innovative solutions for environmental goals.

I realize Wolkow, et. al, are not focused on cryptocurrency and blockchain technology per se but as Huff notes in her reply, “… new lithographic techniques that we’ve developed, which could perform computations at significantly lower energy costs, would be huge for Proof of Work coins.”

Whether or not there are implications for cryptocurrencies, energy needs, climate change, etc., it’s the kind of innovative work being done by scientists at the University of Alberta which may have implications in fields far beyond the researchers’ original intentions such as more efficient computation and data storage.

ETA Aug. 6, 2018: Dexter Johnson weighed in with an August 3, 2018 posting on his Nanoclast blog (on the IEEE [Institute of Electrical and Electronics Engineers] website),

Researchers at the University of Alberta in Canada have developed a new approach to rewritable data storage technology by using a scanning tunneling microscope (STM) to remove and replace hydrogen atoms from the surface of a silicon wafer. If this approach realizes its potential, it could lead to a data storage technology capable of storing 1,000 times more data than today’s hard drives, up to 138 terabytes per square inch.

As a bit of background, Gerd Binnig and Heinrich Rohrer developed the first STM in 1986 for which they later received the Nobel Prize in physics. In the over 30 years since an STM first imaged an atom by exploiting a phenomenon known as tunneling—which causes electrons to jump from the surface atoms of a material to the tip of an ultrasharp electrode suspended a few angstroms above—the technology has become the backbone of so-called nanotechnology.

In addition to imaging the world on the atomic scale for the last thirty years, STMs have been experimented with as a potential data storage device. Last year, we reported on how IBM (where Binnig and Rohrer first developed the STM) used an STM in combination with an iron atom to serve as an electron-spin resonance sensor to read the magnetic pole of holmium atoms. The north and south poles of the holmium atoms served as the 0 and 1 of digital logic.

The Canadian researchers have taken a somewhat different approach to making an STM into a data storage device by automating a known technique that uses the ultrasharp tip of the STM to apply a voltage pulse above an atom to remove individual hydrogen atoms from the surface of a silicon wafer. Once the atom has been removed, there is a vacancy on the surface. These vacancies can be patterned on the surface to create devices and memories.

If you have the time, I recommend reading Dexter’s posting as he provides clear explanations, additional insight into the work, and more historical detail.

World heritage music stored in DNA

It seems a Swiss team from the École Polytechnique de Lausanne (EPFL) have collaborated with American companies Twist Bioscience and Microsoft, as well as, the University of Washington (state) to preserve two iconic jazz pieces on DNA (deoxyribonucleic acid) according to a Sept. 29, 2017 news item on phys.org,,

Thanks to an innovative technology for encoding data in DNA strands, two items of world heritage – songs recorded at the Montreux Jazz Festival [held in Switzerland] and digitized by EPFL – have been safeguarded for eternity. This marks the first time that cultural artifacts granted UNESCO heritage status have been saved in such a manner, ensuring they are preserved for thousands of years. The method was developed by US company Twist Bioscience and is being unveiled today in a demonstrator created at the EPFL+ECAL Lab.

“Tutu” by Miles Davis and “Smoke on the Water” by Deep Purple have already made their mark on music history. Now they have entered the annals of science, for eternity. Recordings of these two legendary songs were digitized by the Ecole Polytechnique Fédérale de Lausanne (EPFL) as part of the Montreux Jazz Digital Project, and they are the first to be stored in the form of a DNA sequence that can be subsequently decoded and listened to without any reduction in quality.

A Sept. 29, 2017 EPFL press release by Emmanuel Barraud, which originated the news item, provides more details,

This feat was achieved by US company Twist Bioscience working in association with Microsoft Research and the University of Washington. The pioneering technology is actually based on a mechanism that has been at work on Earth for billions of years: storing information in the form of DNA strands. This fundamental process is what has allowed all living species, plants and animals alike, to live on from generation to generation.

The entire world wide web in a shoe box

All electronic data storage involves encoding data in binary format – a series of zeros and ones – and then recording it on a physical medium. DNA works in a similar way, but is composed of long strands of series of four nucleotides (A, T, C and G) that make up a “code.” While the basic principle may be the same, the two methods differ greatly in terms of efficiency: if all the information currently on the internet was stored in the form of DNA, it would fit in a shoe box!

Recent advances in biotechnology now make it possible for humans to do what Mother Nature has always done. Today’s scientists can create artificial DNA strands, “record” any kind of genetic code on them and then analyze them using a sequencer to reconstruct the original data. What’s more, DNA is extraordinarily stable, as evidenced by prehistoric fragments that have been preserved in amber. Artificial strands created by scientists and carefully encapsulated should likewise last for millennia.

To help demonstrate the feasibility of this new method, EPFL’s Metamedia Center provided recordings of two famous songs played at the Montreux Jazz Festival: “Tutu” by Miles Davis, and “Smoke on the Water” by Deep Purple. Twist Bioscience and its research partners encoded the recordings, transformed them into DNA strands and then sequenced and decoded them and played them again – without any reduction in quality.

The amount of artificial DNA strands needed to record the two songs is invisible to the naked eye, and the amount needed to record all 50 years of the Festival’s archives, which have been included in UNESCO’s [United Nations Educational, Scientific and Cultural Organization] Memory of the World Register, would be equal in size to a grain of sand. “Our partnership with EPFL in digitizing our archives aims not only at their positive exploration, but also at their preservation for the next generations,” says Thierry Amsallem, president of the Claude Nobs Foundation. “By taking part in this pioneering experiment which writes the songs into DNA strands, we can be certain that they will be saved on a medium that will never become obsolete!”

A new concept of time

At EPFL’s first-ever ArtTech forum, attendees got to hear the two songs played after being stored in DNA, using a demonstrator developed at the EPFL+ECAL Lab. The system shows that being able to store data for thousands of years is a revolutionary breakthrough that can completely change our relationship with data, memory and time. “For us, it means looking into radically new ways of interacting with cultural heritage that can potentially cut across civilizations,” says Nicolas Henchoz, head of the EPFL+ECAL Lab.

Quincy Jones, a longstanding Festival supporter, is particularly enthusiastic about this technological breakthrough: “With advancements in nanotechnology, I believe we can expect to see people living prolonged lives, and with that, we can also expect to see more developments in the enhancement of how we live. For me, life is all about learning where you came from in order to get where you want to go, but in order to do so, you need access to history! And with the unreliability of how archives are often stored, I sometimes worry that our future generations will be left without such access… So, it absolutely makes my soul smile to know that EPFL, Twist Bioscience and their partners are coming together to preserve the beauty and history of the Montreux Jazz Festival for our future generations, on DNA! I’ve been a part of this festival for decades and it truly is a magnificent representation of what happens when different cultures unite for the sake of music. Absolute magic. And I’m proud to know that the memory of this special place will never be lost.

A Sept. 29, 2017 Twist Bioscience news release is repetitive in some ways but interesting nonetheless,

Twist Bioscience, a company accelerating science and innovation through rapid, high-quality DNA synthesis, today announced that, working with Microsoft and University of Washington researchers, they have successfully stored archival-quality audio recordings of two important music performances from the archives of the world-renowned Montreux Jazz Festival.
These selections are encoded and stored in nature’s preferred storage medium, DNA, for the first time. These tiny specks of DNA will preserve a part of UNESCO’s Memory of the World Archive, where valuable cultural heritage collections are recorded. This is the first time DNA has been used as a long-term archival-quality storage medium.
Quincy Jones, world-renowned Entertainment Executive, Music Composer and Arranger, Musician and Music Producer said, “With advancements in nanotechnology, I believe we can expect to see people living prolonged lives, and with that, we can also expect to see more developments in the enhancement of how we live. For me, life is all about learning where you came from in order to get where you want to go, but in order to do so, you need access to history! And with the unreliability of how archives are often stored, I sometimes worry that our future generations will be left without such access…So, it absolutely makes my soul smile to know that EPFL, Twist Bioscience and others are coming together to preserve the beauty and history of the Montreux Jazz Festival for our future generations, on DNA!…I’ve been a part of this festival for decades and it truly is a magnificent representation of what happens when different cultures unite for the sake of music. Absolute magic. And I’m proud to know that the memory of this special place will never be lost.”
“Our partnership with EPFL in digitizing our archives aims not only at their positive exploration, but also at their preservation for the next generations,” says Thierry Amsallem, president of the Claude Nobs Foundation. “By taking part in this pioneering experiment which writes the songs into DNA strands, we can be certain that they will be saved on a medium that will never become obsolete!”
The Montreux Jazz Digital Project is a collaboration between the Claude Nobs Foundation, curator of the Montreux Jazz Festival audio-visual collection and the École Polytechnique Fédérale de Lausanne (EPFL) to digitize, enrich, store, show, and preserve this notable legacy created by Claude Nobs, the Festival’s founder.
In this proof-of-principle project, two quintessential music performances from the Montreux Jazz Festival – Smoke on the Water, performed by Deep Purple and Tutu, performed by Miles Davis – have been encoded onto DNA and read back with 100 percent accuracy. After being decoded, the songs were played on September 29th [2017] at the ArtTech Forum (see below) in Lausanne, Switzerland. Smoke on the Water was selected as a tribute to Claude Nobs, the Montreux Jazz Festival’s founder. The song memorializes a fire and Funky Claude’s rescue efforts at the Casino Barrière de Montreux during a Frank Zappa concert promoted by Claude Nobs. Miles Davis’ Tutu was selected for the role he played in music history and the Montreux Jazz Festival’s success. Miles Davis died in 1991.
“We archived two magical musical pieces on DNA of this historic collection, equating to 140MB of stored data in DNA,” said Karin Strauss, Ph.D., a Senior Researcher at Microsoft, and one of the project’s leaders.  “The amount of DNA used to store these songs is much smaller than one grain of sand. Amazingly, storing the entire six petabyte Montreux Jazz Festival’s collection would result in DNA smaller than one grain of rice.”
Luis Ceze, Ph.D., a professor in the Paul G. Allen School of Computer Science & Engineering at the University of Washington, said, “DNA, nature’s preferred information storage medium, is an ideal fit for digital archives because of its durability, density and eternal relevance. Storing items from the Montreux Jazz Festival is a perfect way to show how fast DNA digital data storage is becoming real.”
Nature’s Preferred Storage Medium
Nature selected DNA as its hard drive billions of years ago to encode all the genetic instructions necessary for life. These instructions include all the information necessary for survival. DNA molecules encode information with sequences of discrete units. In computers, these discrete units are the 0s and 1s of “binary code,” whereas in DNA molecules, the units are the four distinct nucleotide bases: adenine (A), cytosine (C), guanine (G) and thymine (T).
“DNA is a remarkably efficient molecule that can remain stable for millennia,” said Bill Peck, Ph.D., chief technology officer of Twist Bioscience.  “This is a very exciting project: we are now in an age where we can use the remarkable efficiencies of nature to archive master copies of our cultural heritage in DNA.   As we develop the economies of this process new performances can be added any time.  Unlike current storage technologies, nature’s media will not change and will remain readable through time. There will be no new technology to replace DNA, nature has already optimized the format.”
DNA: Far More Efficient Than a Computer 
Each cell within the human body contains approximately three billion base pairs of DNA. With 75 trillion cells in the human body, this equates to the storage of 150 zettabytes (1021) of information within each body. By comparison, the largest data centers can be hundreds of thousands to even millions of square feet to hold a comparable amount of stored data.
The Elegance of DNA as a Storage Medium
Like music, which can be widely varied with a finite number of notes, DNA encodes individuality with only four different letters in varied combinations. When using DNA as a storage medium, there are several advantages in addition to the universality of the format and incredible storage density. DNA can be stable for thousands of years when stored in a cool dry place and is easy to copy using polymerase chain reaction to create back-up copies of archived material. In addition, because of PCR, small data sets can be targeted and recovered quickly from a large dataset without needing to read the entire file.
How to Store Digital Data in DNA
To encode the music performances into archival storage copies in DNA, Twist Bioscience worked with Microsoft and University of Washington researchers to complete four steps: Coding, synthesis/storage, retrieval and decoding. First, the digital files were converted from the binary code using 0s and 1s into sequences of A, C, T and G. For purposes of the example, 00 represents A, 10 represents C, 01 represents G and 11 represents T. Twist Bioscience then synthesizes the DNA in short segments in the sequence order provided. The short DNA segments each contain about 12 bytes of data as well as a sequence number to indicate their place within the overall sequence. This is the process of storage. And finally, to ensure that the file is stored accurately, the sequence is read back to ensure 100 percent accuracy, and then decoded from A, C, T or G into a two-digit binary representation.
Importantly, to encapsulate and preserve encoded DNA, the collaborators are working with Professor Dr. Robert Grass of ETH Zurich. Grass has developed an innovative technology inspired by preservation of DNA within prehistoric fossils.  With this technology, digital data encoded in DNA remains preserved for millennia.
About UNESCO’s Memory of the World Register
UNESCO established the Memory of the World Register in 1992 in response to a growing awareness of the perilous state of preservation of, and access to, documentary heritage in various parts of the world.  Through its National Commissions, UNESCO prepared a list of endangered library and archive holdings and a world list of national cinematic heritage.
A range of pilot projects employing contemporary technology to reproduce original documentary heritage on other media began. These included, for example, a CD-ROM of the 13th Century Radzivill Chronicle, tracing the origins of the peoples of Europe, and Memoria de Iberoamerica, a joint newspaper microfilming project involving seven Latin American countries. These projects enhanced access to this documentary heritage and contributed to its preservation.
“We are incredibly proud to be a part of this momentous event, with the first archived songs placed into the UNESCO Memory of the World Register,” said Emily Leproust, Ph.D., CEO of Twist Bioscience.
About ArtTech
The ArtTech Foundation, created by renowned scientists and dignitaries from Crans-Montana, Switzerland, wishes to stimulate reflection and support pioneering and innovative projects beyond the known boundaries of culture and science.
Benefitting from the establishment of a favorable environment for the creation of technology companies, the Foundation aims to position itself as key promoter of ideas and innovative endeavors within a landscape of “Culture and Science” that is still being shaped.
Several initiatives, including our annual global platform launched in the spring of 2017, are helping to create a community that brings together researchers, celebrities in the world of culture and the arts, as well as investors and entrepreneurs from Switzerland and across the globe.
 
About EPFL
EPFL, one of the two Swiss Federal Institutes of Technology, based in Lausanne, is Europe’s most cosmopolitan technical university with students, professors and staff from over 120 nations. A dynamic environment, open to Switzerland and the world, EPFL is centered on its three missions: teaching, research and technology transfer. EPFL works together with an extensive network of partners including other universities and institutes of technology, developing and emerging countries, secondary schools and colleges, industry and economy, political circles and the general public, to bring about real impact for society.
About Twist Bioscience
At Twist Bioscience, our expertise is accelerating science and innovation by leveraging the power of scale. We have developed a proprietary semiconductor-based synthetic DNA manufacturing process featuring a high throughput silicon platform capable of producing synthetic biology tools, including genes, oligonucleotide pools and variant libraries. By synthesizing DNA on silicon instead of on traditional 96-well plastic plates, our platform overcomes the current inefficiencies of synthetic DNA production, and enables cost-effective, rapid, high-quality and high throughput synthetic gene production, which in turn, expedites the design, build and test cycle to enable personalized medicines, pharmaceuticals, sustainable chemical production, improved agriculture production, diagnostics and biodetection. We are also developing new technologies to address large scale data storage. For more information, please visit www.twistbioscience.com. Twist Bioscience is on Twitter. Sign up to follow our Twitter feed @TwistBioscience at https://twitter.com/TwistBioscience.

If you hadn’t read the EPFL press release first, it might have taken a minute to figure out why EPFL is being mentioned in the Twist Bioscience news release. Presumably someone was rushing to make a deadline. Ah well, I’ve seen and written worse.

I haven’t been able to find any video or audio recordings of the DNA-preserved performances but there is an informational video (originally published July 7, 2016) from Microsoft and the University of Washington describing the DNA-based technology,

I also found this description of listening to the DNA-preserved music in an Oct. 6, 2017 blog posting for the Canadian Broadcasting Corporation’s (CBC) Day 6 radio programme,

To listen to them, one must first suspend the DNA holding the songs in a solution. Next, one can use a DNA sequencer to read the letters of the bases forming the molecules. Then, algorithms can determine the digital code those letters form. From that code, comes the music.

It’s complicated but Ceze says his team performed this process without error.

You can find out more about UNESCO’s Memory of the World and its register here , more about the EPFL+ECAL Lab here, and more about Twist Bioscience here.

Shocker! Science data being lost at a disturbing rate

A Dec. 19, 2013 University of British Columbia (Vancouver, Canada) news release (on EurekAlert) provides a shock for anyone unfamiliar with the problems of accessing ‘old’ data,

Eighty per cent of scientific data are lost within two decades, according to a new study that tracks the accessibility of data over time.

The culprits? Old e-mail addresses and obsolete storage devices.

“Publicly funded science generates an extraordinary amount of data each year,” says Tim Vines, a visiting scholar at the University of British Columbia. “Much of these data are unique to a time and place, and is thus irreplaceable, and many other datasets are expensive to regenerate.

“The current system of leaving data with authors means that almost all of it is lost over time, unavailable for validation of the original results or to use for entirely new purposes.”

For the analysis, published today in Current Biology, Vines and colleagues attempted to collect original research data from a random set of 516 studies published between 1991 and 2011. They found that while all datasets were available two years after publication, the odds of obtaining the underlying data dropped by 17 per cent per year after that.

“I don’t think anybody expects to easily obtain data from a 50-year-old paper, but to find that almost all the datasets are gone at 20 years was a bit of a surprise.”

Vines is calling on scientific journals to require authors to upload data onto public archives as a condition for publication, adding that papers with readily accessible data are more valuable for society and thus should get priority for publication.

“Losing data is a waste of research funds and it limits how we can do science,” says Vines. “Concerted action is needed to ensure it is saved for future research.”

Unfortunately, there’s nothing about the research methodology in the news release. It would be nice to know how the researchers approached the topic and whether or not they focused on biological sciences and are generalizing those results to all of the sciences,including the social sciences. It is likely more or less true of all the sciences as there is a major issue with being able to access data over time. Whether or not the researcher can provide access to the data set, which is a problem in itself, there’s also the issue of obsolete hardware, software, and formats, problems that haunt the arts, the sciences, and the humanities, as well as, business and government. One of my more recent postings about the issue of archiving data is this March 8, 2012 posting and there’s this March 9, 2010 posting (I believe it was my first on the topic). I also mentioned the current Council of Canadian Academies assessment Memory Institutions and the Digital Revolution in a June 5, 2013 posting.

Here’s a link to and a citation for the UBC study,

The Availability of Research Data Declines Rapidly with Article Age by Timothy H. Vines, Arianne Y.K. Albert, Rose L. Andrew, Florence Débarre, Dan G. Bock, Michelle T. Franklin, Kimberly J. Gilbert, Jean-Sébastien Moore, Sébastien Renaut, Diana J. Rennison. Current Biology, 19 December 2013 DOI: 10.1016/j.cub.2013.11.014
Copyright © 2014 Elsevier Ltd All rights reserved.

This paper is behind a paywall.

Memories, science, archiving, and authenticity

This is going to be one of my more freewheeling excursions into archiving and memory. I’ll be starting with  a movement afoot in the US government to give citizens open access to science research moving onto a network dedicated to archiving nanoscience- and nanotechnology-oriented information, examining the notion of authenticity in regard to the Tiananmen Square incident on June 4, 1989, and finishing with the Council of Canadian Academies’ Expert Panel on Memory Institutions and the Digital Revolution.

In his June 4, 2013 posting on the Pasco Phronesis blog, David Bruggeman features information and an overview of  the US Office of Science and Technology Policy’s efforts to introduce open access to science research for citizens (Note: Links have been removed),

Back in February, the Office of Science and Technology Policy (OSTP) issued a memorandum to federal science agencies on public access for research results.  Federal agencies with over $100 million in research funding have until August 22 to submit their access plans to OSTP.  This access includes research publications, metadata on those publications, and underlying research data (in a digital format).

A collection of academic publishers, including the Association of American Publishers and the organization formerly known as the American Association for the Advancement of Science (publisher of Science), has offered a proposal for a publishing industry repository for pubic access to federally funded research that they publish.

David provides a somewhat caustic perspective on the publishers’ proposal while Jocelyn Kaiser’s June 4, 2013 article for ScienceInsider details the proposal in more detail (Note: Links have been removed),

Organized in part by the Association of American Publishers (AAP), which represents many commercial and nonprofit journals, the group calls its project the Clearinghouse for the Open Research of the United States (CHORUS). In a fact sheet that AAP gave to reporters, the publishers describe CHORUS as a “framework” that would “provide a full solution for agencies to comply with the OSTP memo.”

As a starting point, the publishers have begun to index papers by the federal grant numbers that supported the work. That index, called FundRef, debuted in beta form last week. You can search by agency and get a list of papers linked to the journal’s own websites through digital object identifiers (DOIs), widely used ID codes for individual papers. The pilot project involved just a few agencies and publishers, but many more will soon join FundRef, says Fred Dylla, executive director of the American Institute of Physics. (AAAS, which publishes ScienceInsider, is among them and has also signed on to CHORUS.)

The next step is to make the full-text papers freely available after agencies decide on embargo dates, Dylla says. (The OSTP memo suggests 12 months but says that this may need to be adjusted for some fields and journals.) Eventually, the full CHORUS project will also allow searches of the full-text articles. “We will make the corpus available for anybody’s search tool,” says Dylla, who adds that search agreements will be similar to those that publishers already have with Google Scholar and Microsoft Academic Search.

I couldn’t find any mention in Kaiser’s article as to how long the materials would be available. Is this supposed to be an archive, as well as, a repository? Regardless, I found the beta project, FundRef, a little confusing. The link from the ScienceInsider article takes you to this May 28, 2013 news release,

FundRef, the funder identification service from CrossRef [crossref.org], is now available for publishers to contribute funding data and for retrieval of that information. FundRef is the result of collaboration between funding agencies and publishers that correlates grants and other funding with the scholarly output of that support.

Publishers participating in FundRef add funding data to the bibliographic metadata they already provide to CrossRef for reference linking. FundRef data includes the name of the funder and a grant or award number. Manuscript tracking systems can incorporate a taxonomy of 4000 global funder names, which includes alternate names, aliases, and abbreviations enabling authors to choose from a standard list of funding names. Then the tagged funding data will travel through publishers’ production systems to be stored at CrossRef.

I was hoping that clicking on the FundRef button would take me to a database that I could test or tour. At this point, I wouldn’t have described the project as being at the beta stage (from a user’s perspective) as they are still building it and gathering data. However, there is lots of information on the FundRef webpage including an Additional Resources section featuring a webinar,

Attend an Introduction to FundRef Webinar – Thursday, June 6, 2013 at 11:00 am EDT

You do need to sign up for the webinar. Happily, it is open to international participants, as well as, US participants.

Getting back to my question on whether or not this effort is also an archive of sorts, there is a project closer to home (nanotechnologywise, anyway) that touches on these issues from an unexpected perspective, from the Nanoscience and Emerging Technologies in Society (NETS); sharing research and learning tools About webpage,

The Nanoscience and Emerging Technologies in Society: Sharing Research and Learning Tools (NETS) is an IMLS-funded [Institute of Museum and Library Services] project to investigate the development of a disciplinary repository for the Ethical, Legal and Social Implications (ELSI) of nanoscience and emerging technologies research. NETS partners will explore future integration of digital services for researchers studying ethical, legal, and social implications associated with the development of nanotechnology and other emerging technologies.

NETS will investigate digital resources to advance the collection, dissemination, and preservation of this body of research,  addressing the challenge of marshaling resources, academic collaborators, appropriately skilled data managers, and digital repository services for large-scale, multi-institutional and disciplinary research projects. The central activity of this project involves a spring 2013 workshop that will gather key researchers in the field and digital librarians together to plan the development of a disciplinary repository of data, curricula, and methodological tools.

Societal dimensions research investigating the impacts of new and emerging technologies in nanoscience is among the largest research programs of its kind in the United States, with an explicit mission to communicate outcomes and insights to the public. By 2015, scholars across the country affiliated with this program will have spent ten years collecting qualitative and quantitative data and developing analytic and methodological tools for examining the human dimensions of nanotechnology. The sharing of data and research tools in this field will foster a new kind of social science inquiry and ensure that the outcomes of research reach public audiences through multiple pathways.

NETS will be holding a stakeholders workshop June 27 – 28, 2013 (invite only), from the workshop description webpage,

What is the value of creating a dedicated Nano ELSI repository?
The benefits of having these data in a shared infrastructure are: the centralization of research and ease of discovery; uniformity of access; standardization of metadata and the description of projects; and facilitation of compliance with funder requirements for data management going forward. Additional benefits of this project will be the expansion of data curation capabilities for data repositories into the nanotechnology domain, and research into the development of disciplinary repositories, for which very little literature exists.

What would a dedicated Nano ELSI repository contain?
Potential materials that need to be curated are both qualitative and quantitative in nature, including:

  • survey instruments, data, and analyses
  • interview transcriptions and analyses
  • images or multimedia
  • reports
  • research papers, books, and their supplemental data
  • curricular materials

What will the Stakeholder Workshop accomplish?
The Stakeholder Workshop aims to bring together the key researchers and digital librarians to draft a detailed project plan for the implementation of a dedicated Nano ELSI repository. The Workshop will be used as a venue to discuss questions such as:

  • How can a repository extend research in this area?
  • What is the best way to collect all the research in this area?
  • What tools would users envision using with this resource?
  • Who should maintain and staff a repository like this?
  • How much would a repository like this cost?
  • How long will it take to implement?

What is expected of Workshop participants?
The workshop will bring together key researchers and digital librarians to discuss the requirements for a dedicated Nano ELSI repository. To inform that discussion, some participants will be requested to present on their current or past research projects and collaborations. In addition, workshop participants will be enlisted to contribute to the draft of the final project report and make recommendations for the implementation plan.

While my proposal did not get accepted (full disclosure), I do look forward to hearing more about the repository although I notice there’s no mention made of archiving the materials.

The importance of repositories and archives was brought home to me when I came across a June 4, 2013 article by Glyn Moody for Techdirt about the Tiananmen Square incident and subtle and unsubtle ways of censoring access to information,

Today is June 4th, a day pretty much like any other day in most parts of the world. But in China, June 4th has a unique significance because of the events that took place in Tiananmen Square on that day in 1989.

Moody recounts some of the ways in which people have attempted to commemorate the day online while evading the authorities’ censorship efforts. Do check out the article for the inside scoop on why ‘Big Yellow Duck’ is a censored term. One of the more subtle censorship efforts provides some chills (from the Moody article),

… according to this article in the Wall Street Journal, it looks like the Chinese authorities are trying out a new tactic for handling this dangerous topic:

On Friday, a China Real Time search for “Tiananmen Incident” did not return the customary message from Sina informing the user that search results could not be displayed due to “relevant laws, regulations and policies.” Instead the search returned results about a separate Tiananmen incident that occurred on Tomb Sweeping Day in 1976, when Beijing residents flooded the area to protest after they were prevented from mourning the recently deceased Premiere [sic] Zhou Enlai.

This business of eliminating and substituting a traumatic and disturbing historical event with something less contentious reminded me both of the saying ‘history is written by the victors’ and of Luciana Duranti and her talk titled, Trust and Authenticity in the Digital Environment: An Increasingly Cloudy Issue, which took place in Vancouver (Canada) last year (mentioned in my May 18, 2012 posting).

Duranti raised many, many issues that most of us don’t consider when we blithely store information in the ‘cloud’ or create blogs that turn out to be repositories of a sort (and then don’t know what to do with them; ça c’est moi). She also previewed a Sept. 26 – 28, 2013 conference to be hosted in Vancouver by UNESCO [United Nations Educational, Scientific, and Cultural Organization), “Memory of the World in the Digital Age: Digitization and Preservation.” (UNESCO’s Memory of the World programme hosts a number of these themed conferences and workshops.)

The Sept. 2013 UNESCO ‘memory of the world’ conference in Vancouver seems rather timely in retrospect. The Council of Canadian Academies (CCA) announced that Dr. Doug Owram would be chairing their Memory Institutions and the Digital Revolution assessment (mentioned in my Feb. 22, 2013 posting; scroll down 80% of the way) and, after checking recently, I noticed that the Expert Panel has been assembled and it includes Duranti. Here’s the assessment description from the CCA’s ‘memory institutions’ webpage,

Library and Archives Canada has asked the Council of Canadian Academies to assess how memory institutions, which include archives, libraries, museums, and other cultural institutions, can embrace the opportunities and challenges of the changing ways in which Canadians are communicating and working in the digital age.
Background

Over the past three decades, Canadians have seen a dramatic transformation in both personal and professional forms of communication due to new technologies. Where the early personal computer and word-processing systems were largely used and understood as extensions of the typewriter, advances in technology since the 1980s have enabled people to adopt different approaches to communicating and documenting their lives, culture, and work. Increased computing power, inexpensive electronic storage, and the widespread adoption of broadband computer networks have thrust methods of communication far ahead of our ability to grasp the implications of these advances.

These trends present both significant challenges and opportunities for traditional memory institutions as they work towards ensuring that valuable information is safeguarded and maintained for the long term and for the benefit of future generations. It requires that they keep track of new types of records that may be of future cultural significance, and of any changes in how decisions are being documented. As part of this assessment, the Council’s expert panel will examine the evidence as it relates to emerging trends, international best practices in archiving, and strengths and weaknesses in how Canada’s memory institutions are responding to these opportunities and challenges. Once complete, this assessment will provide an in-depth and balanced report that will support Library and Archives Canada and other memory institutions as they consider how best to manage and preserve the mass quantity of communications records generated as a result of new and emerging technologies.

The Council’s assessment is running concurrently with the Royal Society of Canada’s expert panel assessment on Libraries and Archives in 21st century Canada. Though similar in subject matter, these assessments have a different focus and follow a different process. The Council’s assessment is concerned foremost with opportunities and challenges for memory institutions as they adapt to a rapidly changing digital environment. In navigating these issues, the Council will draw on a highly qualified and multidisciplinary expert panel to undertake a rigorous assessment of the evidence and of significant international trends in policy and technology now underway. The final report will provide Canadians, policy-makers, and decision-makers with the evidence and information needed to consider policy directions. In contrast, the RSC panel focuses on the status and future of libraries and archives, and will draw upon a public engagement process.

Question

How might memory institutions embrace the opportunities and challenges posed by the changing ways in which Canadians are communicating and working in the digital age?

Sub-questions

With the use of new communication technologies, what types of records are being created and how are decisions being documented?
How is information being safeguarded for usefulness in the immediate to mid-term across technologies considering the major changes that are occurring?
How are memory institutions addressing issues posed by new technologies regarding their traditional roles in assigning value, respecting rights, and assuring authenticity and reliability?
How can memory institutions remain relevant as a trusted source of continuing information by taking advantage of the collaborative opportunities presented by new social media?

From the Expert Panel webpage (go there for all the links), here’s a complete listing of the experts,

Expert Panel on Memory Institutions and the Digital Revolution

Dr. Doug Owram, FRSC, Chair
Professor and Former Deputy Vice-Chancellor and Principal, University of British Columbia Okanagan Campus (Kelowna, BC)

Sebastian Chan     Director of Digital and Emerging Media, Smithsonian Cooper-Hewitt National Design Museum (New York, NY)

C. Colleen Cook     Trenholme Dean of Libraries, McGill University (Montréal, QC)

Luciana Duranti   Chair and Professor of Archival Studies, the School of Library, Archival and Information Studies at the University of British Columbia (Vancouver, BC)

Lesley Ellen Harris     Copyright Lawyer; Consultant, Author, and Educator; Owner, Copyrightlaws.com (Washington, D.C.)

Kate Hennessy     Assistant Professor, Simon Fraser University, School of Interactive Arts and Technology (Surrey, BC)

Kevin Kee     Associate Vice-President Research (Social Sciences and Humanities) and Canada Research Chair in Digital Humanities, Brock University (St. Catharines, ON)

Slavko Manojlovich     Associate University Librarian (Information Technology), Memorial University of Newfoundland (St. John’s, NL)

David Nostbakken     President/CEO of Nostbakken and Nostbakken, Inc. (N + N); Instructor of Strategic Communication and Social Entrepreneurship at the School of Journalism and Communication, Carleton University (Ottawa, ON)

George Oates     Art Director, Stamen Design (San Francisco, CA)

Seamus Ross     Dean and Professor, iSchool, University of Toronto (Toronto, ON)

Bill Waiser, SOM, FRSC     Professor of History and A.S. Morton Distinguished Research Chair, University of Saskatchewan (Saskatoon, SK)

Barry Wellman, FRSC     S.D. Clark Professor, Department of Sociology, University of Toronto (Toronto, ON)

I notice they have a lawyer whose specialty is copyright, Lesley Ellen Harris. I did check out her website, copyrightlaws.com and could not find anything that hinted at any strong opinions on the topic. She seems to feel that copyright is a good thing but how far she’d like to take this is a mystery to me based on the blog postings I viewed.

I’ve also noticed that this panel has 13 people, four of whom are women which equals a little more (June 5, 2013, 1:35 pm PDT, I substituted the word ‘less’ for the word ‘more’; my apologies for the arithmetic error) than 25% representation. That’s a surprising percentage given how heavily weighted the fields of library and archival studies are weighted towards women.

I have meandered somewhat but my key points are this:

  • How we are going to keep information available? It’s all very well to have repository but how long will the data be kept in the repository and where does it go afterwards?
  • There’s a bias certainly with the NETS workshop and, likely, the CCA Expert Panel on Memory Institutions and the Digital Revolution toward institutions as the source for information that’s worth keeping for however long or short a time that should be. What about individual efforts? e.g. Don’t Leave Canada Behind ; FrogHeart; Techdirt; The Last Word on Nothing, and many other blogs?
  • The online redirection of Tiananmen Square incident queries is chilling but I’ve often wondered what happen if someone wanted to remove ‘objectionable material’ from an e-book, e.g. To Kill a Mockingbird. A new reader wouldn’t notice the loss if the material has been excised in a subtle or professional  fashion.

As for how this has an impact on science, it’s been claimed that Isaac Newton attempted to excise Robert Hooke from history (my Jan. 19, 2012 posting). Whether it’s true or not, there is remarkably little about Robert Hooke despite his accomplishments and his languishment is a reminder that we must always take care that we retain our memories.

ETA June 6, 2013: David Bruggeman added some more information links about CHORUS in his June 5, 2013 post (On The Novelty Of Corporate-Government Partnership In STEM Education),

Before I dive into today’s post, a brief word about CHORUS. Thanks to commenter Joe Kraus for pointing me to this Inside Higher Ed post, which includes a link to the fact sheet CHORUS organizers distributed to reporters. While there are additional details, there are still not many details to sink one’s teeth in. And I remain surprised at the relative lack of attention the announcement has received. On a related note, nobody who’s been following open access should be surprised by Michael Eisen’s reaction to CHORUS.

I encourage you to check out David’s post as he provides some information about a new STEM (science, technology, engineering, mathematics) collaboration between the US National Science Foundation and companies such as GE and Intel.

NanoRosetta; a Kickstarter archiving project for the human genome

The NanoRosetta Kickstarter project needs to raise $100,000 by June 2, 2013 if the organizers are to shrink the human genome to the nanoscale and archive it by printing  it on five nickel discs which will be good for 10,000 years. From the NanoRosetta campaign page,

NanoRosetta is seeking to bring the archival industry into the modern age by using nanotechnology to print analog information onto nickel discs. With a life-span of 10,000 years, no other technology can match the durability and longevity of these discs, and because of the microscopic size of the images printed on the discs, we are able to print high volumes of data that were once thought to be unprintable.

To showcase this paradigm-shifting technology, we are seeking to print multiple sets of all 3.2 BILLION characters of the Human genome on five nickel discs about the size of CDs. Previously, this task would have required a room of books to archive the information as analog data.

Storing this information digitally may be effective in the short term, but for the purposes of long-term archiving, the computer, the operating system and the software would also need to be archived equally well. This is the Achilles heel of digital archiving, and the reason why an analog system is the only way to properly archive important data.

Such passion for archiving warms my heart and I love this notion which reminds me of certain types of science fiction novels (from NanoRosetta campaign page),

No matter how well something is stored, it is always susceptible to a single point of failure.

This is where you come in.

With 80 custodians of the Human genome, with at least one custodian located on each continent, we can avoid the problem of a single point of failure and give you a unique piece to hang on your wall.

After the campaign started in early April, the team partnered with the Moon Arts Project with a plan to send the discs into outer space,

This Kickstarter has just gone from a 10,000 year archiving project to a 1 Billion year archiving project! Thanks to the vision of Lowry Burgess, the former dean of the College of Fine Arts at Carnegie Mellon University and head of the mission’s Moon Arts Project, the university’s engineers and artists have made room for the Human genome discs on their lunar lander.

“One of these days, one of these days…” 

Launching in 2015.

More information on the university’s mission, their status, and the Moon Arts Project can be found at: http://www.cmu.edu/google-lunar-x/
http://moonarts.org/

As for the original Rosetta Stone which provides the inspiration for this project, here’s a little information from its Wikipedia essay (Note: Links have been removed),

Originally displayed within a temple, the stone was probably moved during the early Christian or medieval period and eventually used as building material in the construction of Fort Julien near the town of Rashid (Rosetta) in the Nile Delta. It was rediscovered there in 1799 by a soldier, Pierre-François Bouchard, of the French expedition to Egypt. As the first Ancient Egyptian bilingual text recovered in modern times, the Rosetta Stone aroused widespread public interest with its potential to decipher this hitherto untranslated ancient language. Lithographic copies and plaster casts began circulating among European museums and scholars. Meanwhile, British troops defeated the French in Egypt in 1801, and the original stone came into British possession under the Capitulation of Alexandria. Transported to London, it has been on public display at the British Museum since 1802. It is the most-visited object in the British Museum.

The NanoRosetta team has created a campaign video,

For those who like to know something about the people behind a project, this team doesn’t provide much information (from the campaign page),

Bruce Ha, John Bishop, and Jakub Svec make up the NanoRosetta team bringing technical expertise and the archiving industry together.

Websites

The team (also known as Norsam Technologies) notes on its campaign page that it has worked with the Long Now Foundation on that organization’s, The Rosetta Project.

I wish them the best of luck with NanoRosetta.

There are other pieces about archives on the blog but this Mar. 8, 2012 posting, Digital disasters, probably provides the best justifications for this NanoRosetta project.

Digital disasters

What would happen if we had a digital disaster? Try to imagine a situation where all or most of our information has been destroyed on all global networks. It may seem unlikely but it’s not entirely impossible as Luciana Duranti, then a professor at the University of British Columbia School of Library, Archival, and Information Sciences, suggested to reporter Mike Roberts in a 2006 interview. She cited a few examples of what we had already lost, (excerpted from my March 9, 2010 posting)

… she commented about the memories we had already lost. From the article,

Alas, she says, every day something else is irretrievably lost.

The research records of the U.S. Marines for the past 25 years? Gone.

East German land-survey records vital to the reunification of Germany? Toast.

A piece of digital interactive music recorded by Canadian composer Keith Hamel just eight years ago?

“Inaccessible, over, finito,” says Duranti, educated in her native Italy and a UBC prof since 1987.

Duranti, director of InterPARES (International Research on Permanent Authentic Records in Electronic Systems), an international cyber-preservation project comprising 20 countries and 60 global archivists, says original documentation is a thing of the past.

Glyn Moody’s March 5, 2012 posting on Techdirt notes a recent attempt to address the possible loss of ‘memory’ along with other issues specific to the digitization of information (I have removed links),

But there’s a problem: as more people turn to digital books as their preferred way of consuming text, libraries are starting to throw out their physical copies. Some, because nobody reads them much these days; some, because they take up too much space, and cost too much to keep; some, even on the grounds that Google has already scanned the book, and so the physical copy isn’t needed. Whatever the underlying reason, the natural assumption that we can always go back to traditional libraries to digitize or re-scan works is looking increasingly dubious.

Fortunately, Brewster Kahle, the man behind the Alexa Web traffic and ranking company (named after the Library of Alexandria, and sold to Amazon), and the Internet Archive — itself a kind of digital Library of Alexandria — has spotted the danger, and is now creating yet another ambitious library, this time of physical books …

For some reason this all reminded me of a Canticle for Leibowitz, a book I read many years ago and remember chiefly as a warning that information can be lost. There’s more about the book here. As for Kahle’s plan, I wish him the best of luck.

Trip down memory lane courtesy of the Royal Society

It’s a long trip down memory lane, courtesy of the Royal Society, all the way back to 1665 when they first started published their Philosophical Transactions. In her Oct. 26, 2011 posting in Punctuated Equilibrium on the Guardian science blogs site, GrrlScientist writes,

Beginning today, the historical archives of the peer-reviewed journal, Philosophical Transactions of the Royal Society, are permanently free to online access from anywhere in the world, according to an announcement by The Royal Society.

The Royal Society, established in 1660, began publishing the Philosophical Transactions of the Royal Society — world’s first scientific journal — in March 1665. In 1886, it was divided into two journals, Philosophical Transactions A (mathematics, physics and engineering) and Philosophical Transactions B (biological sciences), both of which are published to this day. Its historical archives are defined as all scientific papers published 70 years or longer ago. These historical archives include more than 60,000 scientific papers.

I took a peek at the 1865-1866 issue and it is quite the experience to see what was being published. Here’s an excerpt from the Table of Contents for the 1st issue (Note: I have removed links to the documents),

Epistle Dedicatory

Phil. Trans. 1665 1: doi:10.1098/rstl.1665.0001

  • ·  The Introduction

Phil. Trans. 1665 1:1-2; doi:10.1098/rstl.1665.0002

  • ·  An Accompt of the Improvement of Optick Glasses

Phil. Trans. 1665 1:2-3; doi:10.1098/rstl.1665.0003

  • ·  A Spot in One of the Belts of Jupiter

Phil. Trans. 1665 1:3; doi:10.1098/rstl.1665.0005

  • ·  The Motion of the Late Comet Praedicted

Phil. Trans. 1665 1:3-8; doi:10.1098/rstl.1665.0004

  • ·  An Experimental History of Cold

Phil. Trans. 1665 1:8-9; doi:10.1098/rstl.1665.0006

An Account of a Very Odd Monstrous Calf

Phil. Trans. 1665 1:10; doi:10.1098/rstl.1665.0007

  • ·  Of a Peculiar Lead-Ore of Germany, and the Use Thereof

Phil. Trans. 1665 1:10-11; doi:10.1098/rstl.1665.0008

I did take a look at one of the articles and found it easy to read, other than the spelling. Here’s a little more about the Philosophical Transactions from the Royal Society publishing website,

In 1662, the newly formed ‘Royal Society of London for Improving Natural Knowledge’ was granted a charter to publish by King Charles II and on 6 March 1665, the first issue of Philosophical Transactions was published under the visionary editorship of Henry Oldenburg, who was also the Secretary of the Society. … In 1886, the breadth and scope of scientific discovery had increased to such an extent that it became necessary to divide the journal into two, Philosophical Transactions A and B, covering the physical sciences and the life sciences respectively.

This initiative is part of a larger commitment to open access publishing (more from GrrlScientist’s Oct. 26, 2011 posting),

Opening its historical archive is part of the Royal Society’s ongoing commitment to open access in scientific publishing. It coincides with The Royal Society’s 5th annual Open Access Week, and also comes soon after the launch of its first ever fully open access journal, Open Biology. All of the Royal Society’s journals provide free access to selected papers, hot-off-the-presses.

There are more details about when and which journals give full open access in GrrlScientist’s post.

French want more nanotech public debates; British science oral history project

After last month’s post about disturbances (causing at least one cancellation) taking place during a series of nanotechnology public debates in France, it was a surprise to find that at least one French group wants to continue the ‘discussion’. This last series of  events has been completed with a report due in April 2010. According to a news item on Chemical Watch, France Nature Environnement (FNE) is urging more public debates. From Chemical Watch,

The French public debate on nanotechnologies that began in September ended this week. An official summary of the 17 debates will be published at the end of April, but environmental organisation France Nature Environnement (FNE) says in its conclusions that further discussion is needed to decide where the technology is useful for human advancement and where its use is unacceptable.

You can look at the FNE news item here but it is in French and the site doesn’t seem hospitable to Firefox,  so do try another browser.

Meanwhile, the Brits are embarking on an oral history of British science. From the news item on BBC News,

The British Library has begun a project to create a vast, online oral history and archive of British science.

The three-year project will see 200 British scientists interviewed and their recollections recorded for the audio library.

“We have long been painfully aware that there’s a marked absence of significant recordings of scientists,” said Dr Rob Perks, curator of oral history at the British Library.

For instance, said Dr Perks, in the current sound archives there are only two recordings of Ernest Rutherford, none of computer pioneer Alan Turing, hovercraft inventor Christopher Cockerell or AV Hill, a physiologist and Nobel laureate.

A study carried out prior to the project being started found that in the last ten years, 30 leading British scientists including 9 Nobel winners have died leaving little or no archive of their work.

I’m glad to hear that this oral history is being preserved although I do wonder about the recording formats. One of the problems with archiving materials is maintaining to access them afterwards.

Coincidentally, one of the local Vancouver papers (The Georgia Straight) has an article by Rhiannon Coppin (in the Feb. 25 – March 4, 2010 issue) about the City of Vancouver archives and their attempts at digital archiving. From the article,

Every day, Vancouver’s city archivist and director of records and archives runs a rescue operation on our past. Les Mobbs might send out film reels from the ’30s for repair, or he could receive a donation of early-20th-century photographic negatives that need to be catalogued, scanned, and put into cold storage.

Lately, Mobbs has been putting equal consideration into how to preserve our future. More and more of the city’s legal and cultural record is being created in a digital format; in other words, it’s “born digital”, he told the Georgia Straight.

The pitfall in digital archiving is that we’re poor caretakers of electronic file formats. In 50 or 100 years, we’ll know we’ve won the preservation game if we can open and read a computer document created today. But even in 2010, we’re missing out on 20-year-old WordStar files stuck on five-and-a-quarter-inch floppy disks. Ironically, it may be safer to keep a paper copy of a document than to store the original computer file.

“We’ve been dealing with paper for 2,000 years,” Mobbs said. “We have a lot of experience with what paper is, what it looks like, and how it’s preserved.”

While acid decay, mould, brittleness, and water damage are formidable but vanquishable foes, machine decay, format obsolescence, and file integrity degradation are virtually unconquerable. The short lifetime of many licensed software formats and the quick deaths of so much hardware (remember LaserDisc?) have posed a particular challenge for archivists like Mobbs.

“How do we preserve material that is, for all intents and purposes, essentially transitory?” he asked.

While this discussion might seem irrelevant on a mostly science-oriented blog, the ‘memristor’ story highlights why information about the past is so important. In 2008, R. Stanley Williams (HP Labs) and his colleagues published two papers, the first proving the existence of a fourth member, a memristor, of electrical engineering’s ‘holy trinity’ of the resistor, capacitor, and inductor and the second paper where they established engineering control over the memristor. Williams  and his team both solved a problem they were experiencing in the lab and made engineering history, in part  by reviewing engineering theories dating back at least 30 years. You can read my post about it here.

Imagine if those theories had been locked into formats that were no longer accessible. That’s one of the major reasons for preserving the past, it can yield important information.

In the interest of full disclosure, I once worked for the City of Vancouver archives.