Tag Archives: Luciana Duranti

Internet Archive backup in Canada?

It’s a good idea whether or not the backup site is in Canada and regardless of who is president of the United States, i.e., having a backup for the world’s digital memory. The Internet Archives has announced that it is raising funds to allow for the creation of a backup site. Here’s more from a Dec. 1, 2016 news item on phys.org,

The Internet Archive, which keeps historical records of Web pages, is creating a new backup center in Canada, citing concerns about surveillance following the US presidential election of Donald Trump.

“On November 9 in America, we woke up to a new administration promising radical change. It was a firm reminder that institutions like ours, built for the long term, need to design for change,” said a blog post from Brewster Kahle, founder and digital librarian at the organization.

“For us, it means keeping our cultural materials safe, private and perpetually accessible. It means preparing for a Web that may face greater restrictions.”

While Trump has announced no new digital policies, his campaign comments have raised concerns his administration would be more active on government surveillance and less sensitive to civil liberties.

Glyn Moody in a Nov. 30, 2016 posting on Techdirt eloquently describes the Internet Archive’s role (Note: Links have been removed),

The Internet Archive is probably the most important site that most people have never heard of, much less used. It is an amazing thing: not just a huge collection of freely-available digitized materials, but a backup copy of much of today’s Web, available through something known as the Wayback Machine. It gets its name from the fact that it lets visitors view snapshots of vast numbers of Web pages as they have changed over the last two decades since the Internet Archive was founded — some 279 billion pages currently. That feature makes it an indispensable — and generally unique — record of pages and information that have since disappeared, sometimes because somebody powerful found them inconvenient.

Even more eloquently, Brewster Kahle explains the initiative in his Nov. 29, 2016 posting on one of the Internet Archive blogs,

The history of libraries is one of loss.  The Library of Alexandria is best known for its disappearance.

Libraries like ours are susceptible to different fault lines:

Earthquakes,

Legal regimes,

Institutional failure.

So this year, we have set a new goal: to create a copy of Internet Archive’s digital collections in another country. We are building the Internet Archive of Canada because, to quote our friends at LOCKSS, “lots of copies keep stuff safe.” This project will cost millions. So this is the one time of the year I will ask you: please make a tax-deductible donation to help make sure the Internet Archive lasts forever. (FAQ on this effort).

Throughout history, libraries have fought against terrible violations of privacy—where people have been rounded up simply for what they read.  At the Internet Archive, we are fighting to protect our readers’ privacy in the digital world.

We can do this because we are independent, thanks to broad support from many of you. The Internet Archive is a non-profit library built on trust. Our mission: to give everyone access to all knowledge, forever. For free. The Internet Archive has only 150 staff but runs one of the top-250 websites in the world. Reader privacy is very important to us, so we don’t accept ads that track your behavior.  We don’t even collect your IP address. But we still need to pay for the increasing costs of servers, staff and rent.

You may not know this, but your support for the Internet Archive makes more than 3 million e-books available for free to millions of Open Library patrons around the world.

Your support has fueled the work of journalists who used our Political TV Ad Archive in their fact-checking of candidates’ claims.

It keeps the Wayback Machine going, saving 300 million Web pages each week, so no one will ever be able to change the past just because there is no digital record of it. The Web needs a memory, the ability to look back.

My two most relevant past posts on the topic of archives and memories are this May 18, 2012 piece about Luciana Duranti’s talk about authenticity and trust regarding digital documents and this March 8, 2012 posting about digital memory, which also features a mention of Brewster Kahle and the Internet Archives.

Does digitizing material mean it’s safe? A tale of Canada’s Fisheries and Oceans scientific libraries

As has been noted elsewhere the federal government of Canada has shut down a number of Fisheries and Oceans Canada libraries in a cost-saving exercise. The government is hoping to save some $440,000 in the 2014-15 fiscal year by digitizing, consolidating, and discarding the libraries and their holdings.

One would imagine that this is being done in a measured, thoughtful fashion but one would be wrong.

Andrew Nikiforuk in a December 23, 2013 article for The Tyee wrote one of the first articles about the closure of the fisheries libraries,

Scientists say the closure of some of the world’s finest fishery, ocean and environmental libraries by the Harper government has been so chaotic that irreplaceable collections of intellectual capital built by Canadian taxpayers for future generations has been lost forever.

Glyn Moody in a Jan. 7, 2014 post on Techdirt noted this,

What’s strange is that even though the rationale for this mass destruction is apparently in order to reduce costs, opportunities to sell off more valuable items have been ignored. A scientist is quoted as follows:

“Hundreds of bound journals, technical reports and texts still on the shelves, presumably meant for the garbage or shredding. I saw one famous monograph on zooplankton, which would probably fetch a pretty penny at a used science bookstore… anybody could go in and help themselves, with no record kept of who got what.”

Gloria Galloway in a Jan. 7, 2014 article for the Globe and Mail adds more details about what has been lost,

Peter Wells, an adjunct professor and senior research fellow at the International Ocean Institute at Dalhousie University in Halifax, said it is not surprising few members of the public used the libraries. But “the public benefits by the researchers and the different research labs being able to access the information,” he said.

Scientists say it is true that most modern research is done online.

But much of the material in the DFO libraries was not available digitally, Dr. Wells said, adding that some of it had great historical value. And some was data from decades ago that researchers use to determine how lakes and rivers have changed.

“I see this situation as a national tragedy, done under the pretext of cost savings, which, when examined closely, will prove to be a false motive,” Dr. Wells said. “A modern democratic society should value its information resources, not reduce, or worse, trash them.”

Dr. Ayles [Burton Ayles, a former DFO regional director and the former director of science for the Freshwater Institute in Winnipeg] said the Freshwater Institute had reports from the 1880s and some that were available nowhere else. “There was a whole core people who used that library on a regular basis,” he said.

Dr. Ayles pointed to a collection of three-ringed binders, occupying seven metres of shelf space, that contained the data collected during a study in the 1960s and 1970s of the proposed Mackenzie Valley pipeline. For a similar study in the early years of this century, he said, “scientists could go back to that information and say, ‘What was the baseline 30 years ago? What was there then and what is there now?’ ”

When asked how much of the discarded information has been digitized, the government did not provide an answer, but said the process continues.

Today, Margo McDiarmid’s Jan. 30, 2014 article for the Canadian Broadcasting Corporation (CBC) news online further explores digitization of the holdings,

Fisheries and Oceans is closing seven of its 11 libraries by 2015. It’s hoping to save more than $443,000 in 2014-15 by consolidating its collections into four remaining libraries.

Shea [Fisheries and Oceans Minister Gail Shea] told CBC News in a statement Jan. 6 that all copyrighted material has been digitized and the rest of the collection will be soon. The government says that putting material online is a more efficient way of handling it.

But documents from her office show there’s no way of really knowing that is happening.

“The Department of Fisheries and Oceans’ systems do not enable us to determine the number of items digitized by location and collection,” says the response by the minister’s office to MacAulay’s inquiry. [emphasis mine]

The documents also that show the department had to figure out what to do with 242,207 books and research documents from the libraries being closed. It kept 158,140 items and offered the remaining 84,067 to libraries outside the federal government.

Shea’s office told CBC that the books were also “offered to the general public and recycled in a ‘green fashion’ if there were no takers.”

The fate of thousands of books appears to be “unknown,” although the documents’ numbers show 160 items from the Maurice Lamontagne Library in Mont Jolie, Que., were “discarded.”  A Radio-Canada story in June about the library showed piles of volumes in dumpsters.

And the numbers prove a lot more material was tossed out. The bill to discard material from four of the seven libraries totals $22,816.76

Leaving aside the issue of whether or not rare books were given away or put in dumpsters, It’s not confidence-building when the government minister can’t offer information about which books have been digitized and where they might located online.

Interestingly,  Fisheries and Oceans is not the only department/ministry shutting down libraries (from McDiarmid’s CBC article),

Fisheries and Oceans is just one of the 14 federal departments, including Health Canada and Environment Canada, that have been shutting physical libraries and digitizing or consolidating the material into closed central book vaults.

I was unaware of the problems with Health Canada’s libraries but Laura Payton’s and Max Paris’ Jan. 20, 2014 article for CBC news online certainly raised my eyebrows,

Health Canada scientists are so concerned about losing access to their research library that they’re finding workarounds, with one squirrelling away journals and books in his basement for colleagues to consult, says a report obtained by CBC News.

The draft report from a consultant hired by the department warned it not to close its library, but the report was rejected as flawed and the advice went unheeded.

Before the main library closed, the inter-library loan functions were outsourced to a private company called Infotrieve, the consultant wrote in a report ordered by the department. The library’s physical collection was moved to the National Science Library on the Ottawa campus of the National Research Council last year.

“Staff requests have dropped 90 per cent over in-house service levels prior to the outsource. This statistic has been heralded as a cost savings by senior HC [Health Canada] management,” the report said.

“However, HC scientists have repeatedly said during the interview process that the decrease is because the information has become inaccessible — either it cannot arrive in due time, or it is unaffordable due to the fee structure in place.”

….

The report noted the workarounds scientists used to overcome their access problems.

Mueller [Dr. Rudi Mueller, who left the department in 2012] used his contacts in industry for scientific literature. He also went to university libraries where he had a faculty connection.

The report said Health Canada scientists sometimes use the library cards of university students in co-operative programs at the department.

Unsanctioned libraries have been created by science staff.

“One group moved its 250 feet of published materials to an employee’s basement. When you need a book, you email ‘Fred,’ and ‘Fred’ brings the book in with him the next day,” the consultant wrote in his report.

“I think it’s part of being a scientist. You find a way around the problems,” Mueller told CBC News.

Unsanctioned, underground libraries aside, the assumption that digitizing documents and books ensures access is false.  Glyn Moody in a Nov. 12, 2013 article for Techdirt gives a chastening example of how vulnerable our digital memories are,

The Internet Archive is the world’s online memory, holding the only copies of many historic (and not-so-historic) Web pages that have long disappeared from the Web itself.

Bad news:

This morning at about 3:30 a.m. a fire started at the Internet Archive’s San Francisco scanning center.

Good news:

no one was hurt and no data was lost. Our main building was not affected except for damage to one electrical run. This power issue caused us to lose power to some servers for a while.

Bad news:

Some physical materials were in the scanning center because they were being digitized, but most were in a separate locked room or in our physical archive and were not lost. Of those materials we did unfortunately lose, about half had already been digitized. We are working with our library partners now to assess.

That loss is unfortunate, but imagine if the fire had been in the main server room holding the Internet Archive’s 2 petabytes of data. Wisely, the project has placed copies at other locations …

That’s good to know, but it seems rather foolish for the world to depend on the Internet Archive always being able to keep all its copies up to date, especially as the quantity of data that it stores continues to rise. This digital library is so important in historical and cultural terms: surely it’s time to start mirroring the Internet Archive around the world in many locations, with direct and sustained support from multiple governments.

In addition to the issue of vulnerability, there’s also the issue of authenticity, from my June 5, 2013 posting about science, archives and memories,

… Luciana Duranti [Professor and Chair, MAS {Master of Archival Studies}Program at the University of British Columbia and Director, InterPARES] and her talk titled, Trust and Authenticity in the Digital Environment: An Increasingly Cloudy Issue, which took place in Vancouver (Canada) last year (mentioned in my May 18, 2012 posting).

Duranti raised many, many issues that most of us don’t consider when we blithely store information in the ‘cloud’ or create blogs that turn out to be repositories of a sort (and then don’t know what to do with them; ça c’est moi). She also previewed a Sept. 26 – 28, 2013 conference to be hosted in Vancouver by UNESCO (United Nations Educational, Scientific, and Cultural Organization), “Memory of the World in the Digital Age: Digitization and Preservation.” (UNESCO’s Memory of the World programme hosts a number of these themed conferences and workshops.)

The Sept. 2013 UNESCO ‘memory of the world’ conference in Vancouver seems rather timely in retrospect. The Council of Canadian Academies (CCA) announced that Dr. Doug Owram would be chairing their Memory Institutions and the Digital Revolution assessment (mentioned in my Feb. 22, 2013 posting; scroll down 80% of the way) and, after checking recently, I noticed that the Expert Panel has been assembled and it includes Duranti. Here’s the assessment description from the CCA’s ‘memory institutions’ webpage,

Library and Archives Canada has asked the Council of Canadian Academies to assess how memory institutions, which include archives, libraries, museums, and other cultural institutions, can embrace the opportunities and challenges of the changing ways in which Canadians are communicating and working in the digital age.

Background

Over the past three decades, Canadians have seen a dramatic transformation in both personal and professional forms of communication due to new technologies. Where the early personal computer and word-processing systems were largely used and understood as extensions of the typewriter, advances in technology since the 1980s have enabled people to adopt different approaches to communicating and documenting their lives, culture, and work. Increased computing power, inexpensive electronic storage, and the widespread adoption of broadband computer networks have thrust methods of communication far ahead of our ability to grasp the implications of these advances.

These trends present both significant challenges and opportunities for traditional memory institutions as they work towards ensuring that valuable information is safeguarded and maintained for the long term and for the benefit of future generations. It requires that they keep track of new types of records that may be of future cultural significance, and of any changes in how decisions are being documented. As part of this assessment, the Council’s expert panel will examine the evidence as it relates to emerging trends, international best practices in archiving, and strengths and weaknesses in how Canada’s memory institutions are responding to these opportunities and challenges. Once complete, this assessment will provide an in-depth and balanced report that will support Library and Archives Canada and other memory institutions as they consider how best to manage and preserve the mass quantity of communications records generated as a result of new and emerging technologies.

The Council’s assessment is running concurrently with the Royal Society of Canada’s expert panel assessment on Libraries and Archives in 21st century Canada. Though similar in subject matter, these assessments have a different focus and follow a different process. The Council’s assessment is concerned foremost with opportunities and challenges for memory institutions as they adapt to a rapidly changing digital environment. In navigating these issues, the Council will draw on a highly qualified and multidisciplinary expert panel to undertake a rigorous assessment of the evidence and of significant international trends in policy and technology now underway. The final report will provide Canadians, policy-makers, and decision-makers with the evidence and information needed to consider policy directions. In contrast, the RSC panel focuses on the status and future of libraries and archives, and will draw upon a public engagement process.

So, the government is shutting down libraries in order to save money and they’re praying (?) that the materials have been digitized and adequate care has been taken to ensure that they will not be lost in some disaster or other. Meanwhile the Council of Canadian Academies is conducting an assessment of memory institutions in the digital age. The approach seems to backwards.

On a more amusing note, Rick Mercer parodies at lease one way scientists are finding to circumvent the cost-cutting exercise in an excerpt (approximately 1 min.)  from his Jan. 29, 2014 Rick Mercer Report telecast (thanks Roz),

Mercer’s comment about sports and Canada’s Prime Minister, Stephen Harper’s preferences is a reference to Harper’s expressed desire to write a book about hockey and possibly a veiled reference to Harper’s successful move to prorogue parliament during the 2010 Winter Olympic games in Vancouver in what many observers suggested was a strategy allowing Harper to attend the games at his leisure.

Whether or not you agree with the decision to shutdown some libraries, the implementation seems to have been a remarkably sloppy affair.

Memories, science, archiving, and authenticity

This is going to be one of my more freewheeling excursions into archiving and memory. I’ll be starting with  a movement afoot in the US government to give citizens open access to science research moving onto a network dedicated to archiving nanoscience- and nanotechnology-oriented information, examining the notion of authenticity in regard to the Tiananmen Square incident on June 4, 1989, and finishing with the Council of Canadian Academies’ Expert Panel on Memory Institutions and the Digital Revolution.

In his June 4, 2013 posting on the Pasco Phronesis blog, David Bruggeman features information and an overview of  the US Office of Science and Technology Policy’s efforts to introduce open access to science research for citizens (Note: Links have been removed),

Back in February, the Office of Science and Technology Policy (OSTP) issued a memorandum to federal science agencies on public access for research results.  Federal agencies with over $100 million in research funding have until August 22 to submit their access plans to OSTP.  This access includes research publications, metadata on those publications, and underlying research data (in a digital format).

A collection of academic publishers, including the Association of American Publishers and the organization formerly known as the American Association for the Advancement of Science (publisher of Science), has offered a proposal for a publishing industry repository for pubic access to federally funded research that they publish.

David provides a somewhat caustic perspective on the publishers’ proposal while Jocelyn Kaiser’s June 4, 2013 article for ScienceInsider details the proposal in more detail (Note: Links have been removed),

Organized in part by the Association of American Publishers (AAP), which represents many commercial and nonprofit journals, the group calls its project the Clearinghouse for the Open Research of the United States (CHORUS). In a fact sheet that AAP gave to reporters, the publishers describe CHORUS as a “framework” that would “provide a full solution for agencies to comply with the OSTP memo.”

As a starting point, the publishers have begun to index papers by the federal grant numbers that supported the work. That index, called FundRef, debuted in beta form last week. You can search by agency and get a list of papers linked to the journal’s own websites through digital object identifiers (DOIs), widely used ID codes for individual papers. The pilot project involved just a few agencies and publishers, but many more will soon join FundRef, says Fred Dylla, executive director of the American Institute of Physics. (AAAS, which publishes ScienceInsider, is among them and has also signed on to CHORUS.)

The next step is to make the full-text papers freely available after agencies decide on embargo dates, Dylla says. (The OSTP memo suggests 12 months but says that this may need to be adjusted for some fields and journals.) Eventually, the full CHORUS project will also allow searches of the full-text articles. “We will make the corpus available for anybody’s search tool,” says Dylla, who adds that search agreements will be similar to those that publishers already have with Google Scholar and Microsoft Academic Search.

I couldn’t find any mention in Kaiser’s article as to how long the materials would be available. Is this supposed to be an archive, as well as, a repository? Regardless, I found the beta project, FundRef, a little confusing. The link from the ScienceInsider article takes you to this May 28, 2013 news release,

FundRef, the funder identification service from CrossRef [crossref.org], is now available for publishers to contribute funding data and for retrieval of that information. FundRef is the result of collaboration between funding agencies and publishers that correlates grants and other funding with the scholarly output of that support.

Publishers participating in FundRef add funding data to the bibliographic metadata they already provide to CrossRef for reference linking. FundRef data includes the name of the funder and a grant or award number. Manuscript tracking systems can incorporate a taxonomy of 4000 global funder names, which includes alternate names, aliases, and abbreviations enabling authors to choose from a standard list of funding names. Then the tagged funding data will travel through publishers’ production systems to be stored at CrossRef.

I was hoping that clicking on the FundRef button would take me to a database that I could test or tour. At this point, I wouldn’t have described the project as being at the beta stage (from a user’s perspective) as they are still building it and gathering data. However, there is lots of information on the FundRef webpage including an Additional Resources section featuring a webinar,

Attend an Introduction to FundRef Webinar – Thursday, June 6, 2013 at 11:00 am EDT

You do need to sign up for the webinar. Happily, it is open to international participants, as well as, US participants.

Getting back to my question on whether or not this effort is also an archive of sorts, there is a project closer to home (nanotechnologywise, anyway) that touches on these issues from an unexpected perspective, from the Nanoscience and Emerging Technologies in Society (NETS); sharing research and learning tools About webpage,

The Nanoscience and Emerging Technologies in Society: Sharing Research and Learning Tools (NETS) is an IMLS-funded [Institute of Museum and Library Services] project to investigate the development of a disciplinary repository for the Ethical, Legal and Social Implications (ELSI) of nanoscience and emerging technologies research. NETS partners will explore future integration of digital services for researchers studying ethical, legal, and social implications associated with the development of nanotechnology and other emerging technologies.

NETS will investigate digital resources to advance the collection, dissemination, and preservation of this body of research,  addressing the challenge of marshaling resources, academic collaborators, appropriately skilled data managers, and digital repository services for large-scale, multi-institutional and disciplinary research projects. The central activity of this project involves a spring 2013 workshop that will gather key researchers in the field and digital librarians together to plan the development of a disciplinary repository of data, curricula, and methodological tools.

Societal dimensions research investigating the impacts of new and emerging technologies in nanoscience is among the largest research programs of its kind in the United States, with an explicit mission to communicate outcomes and insights to the public. By 2015, scholars across the country affiliated with this program will have spent ten years collecting qualitative and quantitative data and developing analytic and methodological tools for examining the human dimensions of nanotechnology. The sharing of data and research tools in this field will foster a new kind of social science inquiry and ensure that the outcomes of research reach public audiences through multiple pathways.

NETS will be holding a stakeholders workshop June 27 – 28, 2013 (invite only), from the workshop description webpage,

What is the value of creating a dedicated Nano ELSI repository?
The benefits of having these data in a shared infrastructure are: the centralization of research and ease of discovery; uniformity of access; standardization of metadata and the description of projects; and facilitation of compliance with funder requirements for data management going forward. Additional benefits of this project will be the expansion of data curation capabilities for data repositories into the nanotechnology domain, and research into the development of disciplinary repositories, for which very little literature exists.

What would a dedicated Nano ELSI repository contain?
Potential materials that need to be curated are both qualitative and quantitative in nature, including:

  • survey instruments, data, and analyses
  • interview transcriptions and analyses
  • images or multimedia
  • reports
  • research papers, books, and their supplemental data
  • curricular materials

What will the Stakeholder Workshop accomplish?
The Stakeholder Workshop aims to bring together the key researchers and digital librarians to draft a detailed project plan for the implementation of a dedicated Nano ELSI repository. The Workshop will be used as a venue to discuss questions such as:

  • How can a repository extend research in this area?
  • What is the best way to collect all the research in this area?
  • What tools would users envision using with this resource?
  • Who should maintain and staff a repository like this?
  • How much would a repository like this cost?
  • How long will it take to implement?

What is expected of Workshop participants?
The workshop will bring together key researchers and digital librarians to discuss the requirements for a dedicated Nano ELSI repository. To inform that discussion, some participants will be requested to present on their current or past research projects and collaborations. In addition, workshop participants will be enlisted to contribute to the draft of the final project report and make recommendations for the implementation plan.

While my proposal did not get accepted (full disclosure), I do look forward to hearing more about the repository although I notice there’s no mention made of archiving the materials.

The importance of repositories and archives was brought home to me when I came across a June 4, 2013 article by Glyn Moody for Techdirt about the Tiananmen Square incident and subtle and unsubtle ways of censoring access to information,

Today is June 4th, a day pretty much like any other day in most parts of the world. But in China, June 4th has a unique significance because of the events that took place in Tiananmen Square on that day in 1989.

Moody recounts some of the ways in which people have attempted to commemorate the day online while evading the authorities’ censorship efforts. Do check out the article for the inside scoop on why ‘Big Yellow Duck’ is a censored term. One of the more subtle censorship efforts provides some chills (from the Moody article),

… according to this article in the Wall Street Journal, it looks like the Chinese authorities are trying out a new tactic for handling this dangerous topic:

On Friday, a China Real Time search for “Tiananmen Incident” did not return the customary message from Sina informing the user that search results could not be displayed due to “relevant laws, regulations and policies.” Instead the search returned results about a separate Tiananmen incident that occurred on Tomb Sweeping Day in 1976, when Beijing residents flooded the area to protest after they were prevented from mourning the recently deceased Premiere [sic] Zhou Enlai.

This business of eliminating and substituting a traumatic and disturbing historical event with something less contentious reminded me both of the saying ‘history is written by the victors’ and of Luciana Duranti and her talk titled, Trust and Authenticity in the Digital Environment: An Increasingly Cloudy Issue, which took place in Vancouver (Canada) last year (mentioned in my May 18, 2012 posting).

Duranti raised many, many issues that most of us don’t consider when we blithely store information in the ‘cloud’ or create blogs that turn out to be repositories of a sort (and then don’t know what to do with them; ça c’est moi). She also previewed a Sept. 26 – 28, 2013 conference to be hosted in Vancouver by UNESCO [United Nations Educational, Scientific, and Cultural Organization), “Memory of the World in the Digital Age: Digitization and Preservation.” (UNESCO’s Memory of the World programme hosts a number of these themed conferences and workshops.)

The Sept. 2013 UNESCO ‘memory of the world’ conference in Vancouver seems rather timely in retrospect. The Council of Canadian Academies (CCA) announced that Dr. Doug Owram would be chairing their Memory Institutions and the Digital Revolution assessment (mentioned in my Feb. 22, 2013 posting; scroll down 80% of the way) and, after checking recently, I noticed that the Expert Panel has been assembled and it includes Duranti. Here’s the assessment description from the CCA’s ‘memory institutions’ webpage,

Library and Archives Canada has asked the Council of Canadian Academies to assess how memory institutions, which include archives, libraries, museums, and other cultural institutions, can embrace the opportunities and challenges of the changing ways in which Canadians are communicating and working in the digital age.
Background

Over the past three decades, Canadians have seen a dramatic transformation in both personal and professional forms of communication due to new technologies. Where the early personal computer and word-processing systems were largely used and understood as extensions of the typewriter, advances in technology since the 1980s have enabled people to adopt different approaches to communicating and documenting their lives, culture, and work. Increased computing power, inexpensive electronic storage, and the widespread adoption of broadband computer networks have thrust methods of communication far ahead of our ability to grasp the implications of these advances.

These trends present both significant challenges and opportunities for traditional memory institutions as they work towards ensuring that valuable information is safeguarded and maintained for the long term and for the benefit of future generations. It requires that they keep track of new types of records that may be of future cultural significance, and of any changes in how decisions are being documented. As part of this assessment, the Council’s expert panel will examine the evidence as it relates to emerging trends, international best practices in archiving, and strengths and weaknesses in how Canada’s memory institutions are responding to these opportunities and challenges. Once complete, this assessment will provide an in-depth and balanced report that will support Library and Archives Canada and other memory institutions as they consider how best to manage and preserve the mass quantity of communications records generated as a result of new and emerging technologies.

The Council’s assessment is running concurrently with the Royal Society of Canada’s expert panel assessment on Libraries and Archives in 21st century Canada. Though similar in subject matter, these assessments have a different focus and follow a different process. The Council’s assessment is concerned foremost with opportunities and challenges for memory institutions as they adapt to a rapidly changing digital environment. In navigating these issues, the Council will draw on a highly qualified and multidisciplinary expert panel to undertake a rigorous assessment of the evidence and of significant international trends in policy and technology now underway. The final report will provide Canadians, policy-makers, and decision-makers with the evidence and information needed to consider policy directions. In contrast, the RSC panel focuses on the status and future of libraries and archives, and will draw upon a public engagement process.

Question

How might memory institutions embrace the opportunities and challenges posed by the changing ways in which Canadians are communicating and working in the digital age?

Sub-questions

With the use of new communication technologies, what types of records are being created and how are decisions being documented?
How is information being safeguarded for usefulness in the immediate to mid-term across technologies considering the major changes that are occurring?
How are memory institutions addressing issues posed by new technologies regarding their traditional roles in assigning value, respecting rights, and assuring authenticity and reliability?
How can memory institutions remain relevant as a trusted source of continuing information by taking advantage of the collaborative opportunities presented by new social media?

From the Expert Panel webpage (go there for all the links), here’s a complete listing of the experts,

Expert Panel on Memory Institutions and the Digital Revolution

Dr. Doug Owram, FRSC, Chair
Professor and Former Deputy Vice-Chancellor and Principal, University of British Columbia Okanagan Campus (Kelowna, BC)

Sebastian Chan     Director of Digital and Emerging Media, Smithsonian Cooper-Hewitt National Design Museum (New York, NY)

C. Colleen Cook     Trenholme Dean of Libraries, McGill University (Montréal, QC)

Luciana Duranti   Chair and Professor of Archival Studies, the School of Library, Archival and Information Studies at the University of British Columbia (Vancouver, BC)

Lesley Ellen Harris     Copyright Lawyer; Consultant, Author, and Educator; Owner, Copyrightlaws.com (Washington, D.C.)

Kate Hennessy     Assistant Professor, Simon Fraser University, School of Interactive Arts and Technology (Surrey, BC)

Kevin Kee     Associate Vice-President Research (Social Sciences and Humanities) and Canada Research Chair in Digital Humanities, Brock University (St. Catharines, ON)

Slavko Manojlovich     Associate University Librarian (Information Technology), Memorial University of Newfoundland (St. John’s, NL)

David Nostbakken     President/CEO of Nostbakken and Nostbakken, Inc. (N + N); Instructor of Strategic Communication and Social Entrepreneurship at the School of Journalism and Communication, Carleton University (Ottawa, ON)

George Oates     Art Director, Stamen Design (San Francisco, CA)

Seamus Ross     Dean and Professor, iSchool, University of Toronto (Toronto, ON)

Bill Waiser, SOM, FRSC     Professor of History and A.S. Morton Distinguished Research Chair, University of Saskatchewan (Saskatoon, SK)

Barry Wellman, FRSC     S.D. Clark Professor, Department of Sociology, University of Toronto (Toronto, ON)

I notice they have a lawyer whose specialty is copyright, Lesley Ellen Harris. I did check out her website, copyrightlaws.com and could not find anything that hinted at any strong opinions on the topic. She seems to feel that copyright is a good thing but how far she’d like to take this is a mystery to me based on the blog postings I viewed.

I’ve also noticed that this panel has 13 people, four of whom are women which equals a little more (June 5, 2013, 1:35 pm PDT, I substituted the word ‘less’ for the word ‘more’; my apologies for the arithmetic error) than 25% representation. That’s a surprising percentage given how heavily weighted the fields of library and archival studies are weighted towards women.

I have meandered somewhat but my key points are this:

  • How we are going to keep information available? It’s all very well to have repository but how long will the data be kept in the repository and where does it go afterwards?
  • There’s a bias certainly with the NETS workshop and, likely, the CCA Expert Panel on Memory Institutions and the Digital Revolution toward institutions as the source for information that’s worth keeping for however long or short a time that should be. What about individual efforts? e.g. Don’t Leave Canada Behind ; FrogHeart; Techdirt; The Last Word on Nothing, and many other blogs?
  • The online redirection of Tiananmen Square incident queries is chilling but I’ve often wondered what happen if someone wanted to remove ‘objectionable material’ from an e-book, e.g. To Kill a Mockingbird. A new reader wouldn’t notice the loss if the material has been excised in a subtle or professional  fashion.

As for how this has an impact on science, it’s been claimed that Isaac Newton attempted to excise Robert Hooke from history (my Jan. 19, 2012 posting). Whether it’s true or not, there is remarkably little about Robert Hooke despite his accomplishments and his languishment is a reminder that we must always take care that we retain our memories.

ETA June 6, 2013: David Bruggeman added some more information links about CHORUS in his June 5, 2013 post (On The Novelty Of Corporate-Government Partnership In STEM Education),

Before I dive into today’s post, a brief word about CHORUS. Thanks to commenter Joe Kraus for pointing me to this Inside Higher Ed post, which includes a link to the fact sheet CHORUS organizers distributed to reporters. While there are additional details, there are still not many details to sink one’s teeth in. And I remain surprised at the relative lack of attention the announcement has received. On a related note, nobody who’s been following open access should be surprised by Michael Eisen’s reaction to CHORUS.

I encourage you to check out David’s post as he provides some information about a new STEM (science, technology, engineering, mathematics) collaboration between the US National Science Foundation and companies such as GE and Intel.

Memory of the world

The fact that UNESCO will be holding its International Conference: “Memory of the World in the Digital Age: Digitization and Preservation” in Vancouver (Canada), Sept. 26 – 28, 2012 was one of the many snippets of information that Luciana Duranti, Chair and Professor at the School of Library, Archival and Information Studies at the University of British Columbia, passed on during her talk on Thursday, May 17, 2012 in Vancouver.

Organized by ARPICO (Society of Italian Researchers and Professionals in Western Canada), Duranti’s talk Trust and Authenticity in the Digital Environment: An Increasingly Cloudy Issue, first delved into definitions of trust, authenticity and cloud computing before focusing on the issues presented by storing our data on the  ‘cloud’. As Duranti noted, this is a return, of sorts, to the 60s and its mainframe environment.  However, unlike the 60s our data is not stored on one server, it may be split amongst many servers in many countries making our data quite vulnerable. For example, different laws in different countries means you can lose data if the legal situation changes as it did in the US recently.  According to Duranti (as best as I can recall), one of Megaupload’s servers has been shut down in the state of Virginia because of a problem with data from one business. Unfortunately, all of the data held there was also destroyed.

On investigating this further, I found a more general discussion of the situation with Megaupload on Techdirt (May 1, 2012 posting by Mike Masnick) which highlights law professor Eric Goldman’s extraordinary indictment of the government’s action in his April 20, 2012 posting, excerpt of 2nd point,

2) Taking Megaupload offline. Megaupload’s website is analogous to a printing press that constantly published new content. Under our Constitution, the government can’t simply shut down a printing press, but that’s basically what our government did when it turned Megaupload off and seized all of the assets. Not surprisingly, shutting down a printing press suppresses countless legitimate content publications by legitimate users of Megaupload. Surprisingly (shockingly, even), the government apparently doesn’t care about this “collateral,” entirely foreseeable and deeply unconstitutional effect. The government’s further insistence that all user data, even legitimate data, should be destroyed is even more shocking. Destroying the evidence not only screws over the legitimate users, but it may make it impossible for Megaupload to mount a proper defense. It’s depressing our government isn’t above such cheap tricks in its zeal to win.

As Masnick notes on Techdirt,

The more we hear and see about the government’s case against Megaupload, it really appears that the government was relying almost entirely on the fact that Megaupload looked bad. It’s hard to deny that there were plenty of things that Kim (in particular) [CEO Kim Dotcom] did that makes him appear pretty obnoxious. But being a crass showoff doesn’t automatically make you a criminal.

The Jan. 19, 2012 article by Nate Anderson for Ars Technica seems more sympathetic to the government’s position, initially,

The US government dropped a nuclear bomb on “cyberlocker” site Megaupload today, seizing its domain names, grabbing $50 million in assets, and getting New Zealand police to arrest four of the site’s key employees, including enigmatic founder Kim Dotcom. In a 72-page indictment unsealed in a Virginia federal court, prosecutors charged that the site earned more than $175 million since its founding in 2005, most of it based on copyright infringement.

As for the site’s employees, they were paid lavishly and they spent lavishly. Even the graphic designer, 35-year-old Slovakian resident Julius Bencko, made more than $1 million in 2010 alone.

The indictment goes after six individuals, who between them owned 14 Mercedes-Benz automobiles with license plates such as “POLICE,” “MAFIA,” “V,” “STONED,” “CEO,” “HACKER,” GOOD,” “EVIL,” and—perhaps presciently—”GUILTY.” The group also had a 2010 Maserati, a 2008 Rolls-Royce, and a 1989 Lamborghini. They had not one but three Samsung 83″ TVs, and two Sharp 108″ TVs. Someone owned a “Predator statue.” …

Yet the indictment seems odd in some ways. When Viacom made many of the same charges against YouTube, it didn’t go to the government and try to get Eric Schmidt or Chad Hurley arrested.

Anderson mentions that Megaupload had 525 servers in Virginia state and many more around the world. (I’m not sure why Duranti stated that one server had been shut down in Virginia but perhaps she was using it as an example to demonstrate what happens when just one server is shut down.) Regardless of whether it’s one server or 525 , I’m with Eric Goldman when he points out that destroying legitimate data is shocking.

Duranti’s talk was illuminating and I look forward to hearing more about these issues when the UNESCO conference takes place here in Vancouver next September. From the conference news webpage,

Digital information has economic value as a cultural product and as a source of knowledge. It plays a major role in national sustainable development as, increasingly, personal, governmental and commercial information is created in digital form only. But digitized national assets also constitute an immense wealth of the countries concerned and of society at large. The disappearance of this heritage will engender economic and cultural impoverishment and hamper the advancement of knowledge.

Ensuring digital continuity of content can only be overcome if a range of legal, technological, social, financial, political and other obstacles are addressed. The Vancouver Conference therefore seeks to achieve:

  • the launch of specific initiatives related to digital preservation and to the fostering of access to documentary heritage through digitization;
  • the revision of the UNESCO Charter on the Preservation of Digital Heritage;
  • the identification of the legal frameworks and solutions to facilitate long-term digital preservation;
  • the agreement on the promotion and/or development of exchange standards;
  • the definition of the respective roles and responsibilities of all stakeholders and elaboration of a cooperation model.

I have mentioned Duranti and issues relating to digitization and archives before (March 8, 2012 posting: Digital disaster).

Digital disasters

What would happen if we had a digital disaster? Try to imagine a situation where all or most of our information has been destroyed on all global networks. It may seem unlikely but it’s not entirely impossible as Luciana Duranti, then a professor at the University of British Columbia School of Library, Archival, and Information Sciences, suggested to reporter Mike Roberts in a 2006 interview. She cited a few examples of what we had already lost, (excerpted from my March 9, 2010 posting)

… she commented about the memories we had already lost. From the article,

Alas, she says, every day something else is irretrievably lost.

The research records of the U.S. Marines for the past 25 years? Gone.

East German land-survey records vital to the reunification of Germany? Toast.

A piece of digital interactive music recorded by Canadian composer Keith Hamel just eight years ago?

“Inaccessible, over, finito,” says Duranti, educated in her native Italy and a UBC prof since 1987.

Duranti, director of InterPARES (International Research on Permanent Authentic Records in Electronic Systems), an international cyber-preservation project comprising 20 countries and 60 global archivists, says original documentation is a thing of the past.

Glyn Moody’s March 5, 2012 posting on Techdirt notes a recent attempt to address the possible loss of ‘memory’ along with other issues specific to the digitization of information (I have removed links),

But there’s a problem: as more people turn to digital books as their preferred way of consuming text, libraries are starting to throw out their physical copies. Some, because nobody reads them much these days; some, because they take up too much space, and cost too much to keep; some, even on the grounds that Google has already scanned the book, and so the physical copy isn’t needed. Whatever the underlying reason, the natural assumption that we can always go back to traditional libraries to digitize or re-scan works is looking increasingly dubious.

Fortunately, Brewster Kahle, the man behind the Alexa Web traffic and ranking company (named after the Library of Alexandria, and sold to Amazon), and the Internet Archive — itself a kind of digital Library of Alexandria — has spotted the danger, and is now creating yet another ambitious library, this time of physical books …

For some reason this all reminded me of a Canticle for Leibowitz, a book I read many years ago and remember chiefly as a warning that information can be lost. There’s more about the book here. As for Kahle’s plan, I wish him the best of luck.

Dem bones at McGill; innovation from the Canadian business community?; the archiving frontier; linking and copyright

I have a number of bits today amongst them, Canadian nanotechnology, Canadian business innovation, digital archiving, and copyrights and linking.

A Quebec biotech company, Enobia Pharma is working with Dr. Marc McKee on treatments for genetic bone diseases. From the news item on Nanowerk,

The field is known as biomineralization and it involves cutting-edge, nanotech investigation into the proteins, enzymes and other molecules that control the coupling of mineral ions (calcium and phosphate) to form nano-crystals within the bone structure. The treatment, enzyme replacement therapy to treat hypophosphatasia, is currently undergoing clinical testing in several countries including Canada. Hypophosphatasia is a rare and severe disorder resulting in poor bone mineralization. In infants, symptoms include respiratory insufficiency, failure to thrive and rickets.

This research in biomineralization (coupling of mineral ions to form nano-crystals) could lead to better treatments for other conditions such as cardiovascular diseases, arthritis, and kidney stones.

McKee’s research is being funded in part by the Canadian Institutes of Health Research  From the Nanowerk news item,

McKee’s research program is a concrete example of how university researchers are working with private sector partners as an integral part of Canada’s innovative knowledge economy, and the positive outcomes their collaborations can offer.

I don’t think that businesses partnering with academic institutions in research collaborations is precisely what they mean when they talk about business innovation (research and development). From a March 2, 2010 article about innovation by Preston Manning in the Globe & Mail,

Government competition policy and support for science, technology, and innovation (STI) can complement business leadership on the innovation front, but it is not a substitute for such leadership. Action to increase innovation in the economy is first and foremost a business responsibility.

Manning goes on to describe what he’s done on this matter and asks for suggestions on how to encourage Canadian business to be more innovative. (Thanks to Pasco Phronesis for pointing me to Manning’s article.) I guess the problem is that what we’ve been doing has worked well enough and so there’s no great incentive to change.

I’ve been on an archiving kick lately and so here’s some more. The British Library recently (Feb.25.10) announced public access to their UK Web Archive, a project where they have been saving online materials. From the news release,

British Library Chief Executive, Dame Lynne Brindley said:

“Since 2004 the British Library has led the UK Web Archive in its mission to archive a record of the major cultural and social issues being discussed online. Throughout the project the Library has worked directly with copyright holders to capture and preserve over 6,000 carefully selected websites, helping to avoid the creation of a ‘digital black hole’ in the nation’s memory.

“Limited by the existing legal position, at the current rate it will be feasible to collect just 1% of all free UK websites by 2011. We hope the current DCMS consultation will enact the 2003 Legal Deposit Libraries Act and extend the provision of legal deposit through regulationto cover freely available UK websites, providingregular snapshots ofthe free UK web domain for the benefit of future research.”

Mike Masnick at Techdirt notes (here) that the British Library has to get permission (the legal position Dame Brindley refers to) to archive these materials and this would seem to be an instance where ‘fair use’ should be made to apply.

On the subject of losing data, I read an article by Mike Roberts for the Vancouver Province, January 22, 2006, p. B5 (digital copy here) that posed this question, What if the world lost its memory? It was essentially an interview with Luciana Duranti (chair of the Master of Archival Studies programme and professor at the School of Library, Archival and Information Studies at the University of British Columbia, Canada) where she commented about the memories we had already lost. From the article,

Alas, she says, every day something else is irretrievably lost.

The research records of the U.S. Marines for the past 25 years? Gone.

East German land-survey records vital to the reunification of Germany? Toast.

A piece of digital interactive music recorded by Canadian composer Keith Hamel just eight years ago?

“Inaccessible, over, finito,” says Duranti, educated in her native Italy and a UBC prof since 1987.

Duranti, director of InterPARES (International Research on Permanent Authentic Records in Electronic Systems), an international cyber-preservation project comprising 20 countries and 60 global archivists, says original documentation is a thing of the past.

I was shocked by how much ‘important’ information had been lost and I assume still is. (Getting back to the UK Web Archives, if they can only save 1% of the UK’s online material then a lot has got to be missing.)

For anyone curious about InterPARES, I got my link for the Roberts article from this page on the InterPARES 1 website.

Back to Techdirt and Mike Masnick who has educated me as to a practice I had noted but not realized is ‘the way things are done amongst journalists’. If you spend enough time on the web, you’ll notice stories that make their way to newspapers without any acknowledgment of  their web or writerly origins and I’m not talking about news releases which are designed for immediate placement in the media or rewritten/reworked before placement. From the post on Techdirt,

We recently wrote about how the NY Post was caught taking a blogger’s story and rewriting it for itself — noting the hypocrisy of a News Corp. newspaper copying from someone else, after Rupert Murdoch and his top execs have been going around decrying various news aggregators (and Google especially) for “stealing” from News Corp. newspapers. It’s even more ridiculous when you think about it — because the “stealing” that Rupert is upset about is Google linking to the original story — a step that his NY Post writer couldn’t even be bothered to do.

Of course, as a few people pointed out in the comments, this sort of “re-reporting” is quite common in the traditional news business. You see it all the time in newspapers, magazines and broadcast TV. They take a story that was found somewhere else and just “re-report” it, so that they have their own version of it.

That’s right, it’s ‘re-reporting’ without attributions or links. Masnick’s post (he’s bringing in Felix Salmon’s comments) attributes this to a ‘print’ mentality where reporters are accustomed to claiming first place and see acknowledgments and links as failure while ‘digital natives’ acknowledge and link regularly since they view these as signs of respect. I’m not going to disagree but I would like to point out that citing sources is pretty standard for academics or anyone trained in that field. I imagine most reporters have one university or college degree, surely they learned the importance of citing one’s sources. So does training as a journalist erode that understanding?

And, getting back to this morning’s archival subtheme, at the end of Clark Hoyt’s (blogger for NY Times) commentary about the plagiarism he had this to say,

Finally, The Times owes readers a full accounting. I asked [Philip] Corbett [standards editor] for the examples of Kouwe’s plagiarism and suggested that editors’ notes be appended to those articles on the Web site and in The Times’s electronic archives. Corbett would not provide the examples and said the paper was not inclined to flag them, partly because there were some clear-cut cases and others that were less clear. “Where do you draw the line?” he asked.

I’d draw it at those he regards as clear. To do otherwise is to leave a corrupted record within the archives of The Times. It is not the way to close the case.

One last thing, Heather Haley is one of the guests appearing tonight in Rock Against Prisons.

Tuesday, March 9, 2010

7:00pm – 11:55pm

Little Mountain Gallery

195 east 26th Ave [Vancouver, Canada]

More details from my previous announcement about this event here.