People who have been involved in the challenges of e-discovery for a while remember when email arrived on the scene nearly two decades ago. It changed the way people collaborate and left companies with mounds of digital information that was costly and time-consuming to sort through when litigation struck.
The arrival of social media is in many ways a repeat of those challenges. As was true of email, social media comes with new metadata and formats. But because of the similarities, there is an opportunity to avoid the mistakes made with email. One thing is clear: Companies that dive into social media without the right policies and solutions to govern usage will encounter information governance and e-discovery nightmares down the road.
With email, companies could plead ignorance about the e-discovery issues that arose. The digital revolution was new and case law and civil procedure rules were still in flux. With email as a precedent, however, companies cannot hide behind ignorance in the case of social media. Instead, they can get ahead of social media by putting in place governance policies, processes and tools to ensure that the email history lesson informs these new methods of collaboration.
Social media have seen widespread adoption. In order to avoid the mistakes made in the email generation, companies must figure out ways to best collect and preserve social-media content in the event it is needed for e-discovery. Today, this practice is extremely immature.
Recently, the eDJ Group conducted a survey on "The Cloud and eDiscovery" that looked at the experiences e-discovery professionals have had with collecting and preserving information from cloud-based sources such as Amazon, Rackspace and social-media publishers. At most, only 15% of the respondents indicated that they have had to collect from a popular social-media service. But that figure will surely rise.
Technological methods for collection and preservation
When it comes to the collection and preservation of social-media content, companies have several choices of technological methods, each with distinct pros and cons.
A Web crawler is a computer program that periodically browses the Web (in this case, a social-media URL). The crawler creates a copy of the page to be stored for processing into a preservation repository. Companies can set up Web crawlers to capture content from social-media sites at various intervals. Most of these systems store social-media content as static Web pages. However, Web crawling does not necessarily create a forensic capture of a Web page in its full context and therefore may not be sufficient in certain types of cases.
Companies can set up programs that will essentially take a screenshot, or screen scrape, of a Web page and then store that image as a record of the page at that point in time. In most cases, the image will be converted to a PDF (or similar) file so that it can be indexed and searched within a preservation repository. A screenshot, though, is not a full capture of the information in a Web page. It lacks metadata and other context that may be important depending on the matter.
Sign up for CIO Asia eNewsletters.