World Library  
Flag as Inappropriate
Email this Article

Web archive

Article Id: WHEBN0005089540
Reproduction Date:

Title: Web archive  
Author: World Heritage Encyclopedia
Language: English
Subject: Interstella 5555: The 5tory of the 5ecret 5tar 5ystem, Discovery (Daft Punk album), Jaos
Collection:
Publisher: World Heritage Encyclopedia
Publication
Date:
 

Web archive

"Web archive" redirects here. For web.archive.org, see Wayback Machine. For the Safari file format, see webarchive. For the Sun Microsystems file format, see WAR (Sun file format).

Web archiving is the process of collecting portions of the World Wide Web to ensure the information is preserved in an archive for future researchers, historians, and the public. Web archivists typically employ web crawlers for automated capture due to the massive size and amount of information on the Web. The largest web archiving organization based on a bulk crawling approach is the Internet Archive which strives to maintain an archive of the entire Web. National libraries, national archives and various consortia of organizations are also involved in archiving culturally important Web content. Commercial web archiving software and services are also available to organizations who need to archive their own web content for corporate heritage, regulatory, or legal purposes.

Collecting the web

Web archivists generally archive various types of web content including HTML web pages, style sheets, JavaScript, images, and video. They also archive metadata about the collected resources such as access time, MIME type, and content length. This metadata is useful in establishing authenticity and provenance of the archived collection.

Methods of collection

Remote harvesting

The most common web archiving technique uses web crawlers to automate the process of collecting web pages. Web crawlers typically access web pages in the same manner that users with a browser see the Web, and therefore provide a comparatively simple method of remote harvesting web content. Examples of web crawlers used for web archiving include:

On-demand

There are numerous services that may be used to archive web resources "on-demand", using web crawling techniques.

Free services open for public use
  • Archive.is, a free service which saves page and all its images. It can save Web 2.0 pages. Snapshots can be searched by URL wildcards.
  • ArchiveBay, allows users to submit a URL to be archived in the form of a screenshot. Users can browse all saved screenshots of a URL.
  • WebCite. Snapshots can be searched by exact URL.
  • .
  • WebCite, a free service specifically for scholarly authors, journal editors and publishers to permanently archive and retrieve cited Internet references.[4] Snapshots can be searched by exact URL.
Enterprise and subscription services
  • eDiscovery aimed to corporate (Global 500 market), legal and government industries.
  • Archive-It, a subscription service which allows institutions to build, manage and search their own web archive.
  • IM Website
  • Compliance WatchDog by SiteQuest Technologies, a subscription service that archives websites and allows users to browse the site as it appeared in the past. It also monitors sites for changes and alerts compliance personnel if a change is detected.
  • electronic discovery applications and services.
  • Iterasi, Provides enterprise web archiving for compliance, litigation protection, e-discovery and brand heritage. For enterprise companies, financial organizations, government agencies and more.
  • Nextpoint, offers an automated cloud-based, SaaS for marketing, compliance and litigation related needs including electronic discovery
  • CFTC regulations.
  • Federal Rules of Evidence and records management laws. Archives can be used as legal evidence.
  • Reed Archives, offers litigation protection, regulatory compliance & eDiscovery in the corporate, legal and government industries.
  • SiteReplay, a subscription service. Captures screen-shots of pages, transactions and user journeys using "actual browsers" for regulatory compliance. Screen-shots can be viewed online.
  • Smarsh Web Archiving is designed to capture, preserve and re-create the Web experience as it existed at any moment in time for e-discovery and regulatory compliance obligations. (Smarsh acquired Perpetually in May 2012)
  • California Digital Library.
  • webEchoFS, offers a subscription service that was created exclusively to meet the needs of Financial Services companies subject advertising regulations associated with FINRA and the Investment Advisors Act.
  • Cloud Testing technology.

Database archiving

Database archiving refers to methods for archiving the underlying content of database-driven websites. It typically requires the extraction of the XML schema, and the content exported into an XML document. Xinq then allows that content to be delivered online. Although the original layout and behavior of the website cannot be preserved exactly, Xinq does allow the basic querying and retrieval functionality to be replicated.

Transactional archiving

Transactional archiving is an event-driven approach, which collects the actual transactions which take place between a web server and a web browser. It is primarily used as a means of preserving evidence of the content which was actually viewed on a particular website, on a given date. This may be particularly important for organizations which need to comply with legal or regulatory requirements for disclosing and retaining information.

A transactional archiving system typically operates by intercepting every HTTP request to, and response from, the web server, filtering each response to eliminate duplicate content, and permanently storing the responses as bitstreams. A transactional archiving system requires the installation of software on the web server, and cannot therefore be used to collect content from a remote website.

Difficulties and limitations

Crawlers

Web archives which rely on web crawling as their primary means of collecting the Web are influenced by the difficulties of web crawling:

  • The robots exclusion protocol may request crawlers not access portions of a website. Some web archivists may ignore the request and crawl those portions anyway.
  • Large portions of a web site may be hidden in the deep Web. For example, the results page behind a web form lies in the deep Web because most crawlers cannot follow a link to the results page.
  • Crawler traps (e.g., calendars) may cause a crawler to download an infinite number of pages, so crawlers are usually configured to limit the number of dynamic pages they crawl.

However, it is important to note that a native format web archive, i.e., a fully browsable web archive, with working links, media, etc., is only really possible using crawler technology.

The Web is so large that crawling a significant portion of it takes a large amount of technical resources. The Web is changing so fast that portions of a website may change before a crawler has even finished crawling it.

General limitations

  • Some web servers are configured to return different pages to web archiver requests than they would in response to regular browser requests. This is typically done to fool search engines into directing more user traffic to a website, and is often done to avoid accountability, or to provide enhanced content only to those browsers that can display it.

Not only must web archivists deal with the technical challenges of web archiving, they must also contend with intellectual property laws. Peter Lyman[5] states that "although the Web is popularly regarded as a public domain resource, it is copyrighted; thus, archivists have no legal right to copy the Web". However national libraries in many countries do have a legal right to copy portions of the web under an extension of a legal deposit.

Some private non-profit web archives that are made publicly accessible like WebCite, the Internet Archive or Internet memory allow content owners to hide or remove archived content that they do not want the public to have access to. Other web archives are only accessible from certain locations or have regulated usage. WebCite cites a recent lawsuit against Google's caching, which Google won.[6]

Aspects of web curation

Web curation, like any digital curation, entails:

  • Certification of the trustworthiness and integrity of the collection content
  • Collecting verifiable Web assets
  • Providing Web asset search and retrieval
  • Semantic and ontological continuity and comparability of the collection content

Thus, besides the discussion on methods of collecting the Web, those of providing access, certification, and organizing must be included. There are a set of popular tools that addresses these curation steps:

A suite of tools for Web Curation by International Internet Preservation Consortium:

  • Heritrix - official website - collecting Web asset
  • NutchWAX - search Web archive collections
  • Wayback (Open source Wayback Machine) - search and navigate Web archive collections using NutchWax
  • Web Curator Tool - Selection and Management of Web Collection[7]

Other open source tools for manipulating web archives:

  • WARC Tools - for creating, reading, parsing and manipulating, web archives programmatically
  • Search Tools - for indexing and searching full-text and metadata within web archives

See also

References

Bibliography

External links

  • International Internet Preservation Consortium (IIPC) - International consortium whose mission is to acquire, preserve, and make accessible knowledge and information from the Internet for future generations
  • International Web Archiving Workshop (IWAW) - Annual workshop that focuses on web archiving
  • The Library of Congress, Digital Collections and Programs
  • National Library of Australia, Preserving Access to Digital Information (PADI)
  • Library of Congress - Web Archiving
  • Internet Memory
  • ArchivetheNet
  • Web archiving bibliography - Lengthy list of web-archiving resources
  • Web archiving discussion list - Used for discussing the technical, legal, and organizational aspects of web archiving
  • WebArchivist - Researchers that work with scholars, librarians, and archivists interested in preserving and analyzing Web resources
  • Julien Masanès, Bibliothèque Nationale de France - Towards continuous web archiving
  • Comparison of web archiving services
  • SWAT - Snappy Web Archiving Tool. A proof-of-concept software that archives web pages by harvesting all files and taking screenshots of each page. All META data is saved in XML (METS, PREMIS, MODS and ADDML).
  • The UK Government Web Archive at The National Archives - Archive of UK central government websites
  • The UK Web Archive provided by The British Library - Archive of selected websites of UK cultural, social and historical significance - archived with permission from content owners
This article was sourced from Creative Commons Attribution-ShareAlike License; additional terms may apply. World Heritage Encyclopedia content is assembled from numerous content providers, Open Access Publishing, and in compliance with The Fair Access to Science and Technology Research Act (FASTR), Wikimedia Foundation, Inc., Public Library of Science, The Encyclopedia of Life, Open Book Publishers (OBP), PubMed, U.S. National Library of Medicine, National Center for Biotechnology Information, U.S. National Library of Medicine, National Institutes of Health (NIH), U.S. Department of Health & Human Services, and USA.gov, which sources content from all federal, state, local, tribal, and territorial government publication portals (.gov, .mil, .edu). Funding for USA.gov and content contributors is made possible from the U.S. Congress, E-Government Act of 2002.
 
Crowd sourced content that is contributed to World Heritage Encyclopedia is peer reviewed and edited by our editorial staff to ensure quality scholarly research articles.
 
By using this site, you agree to the Terms of Use and Privacy Policy. World Heritage Encyclopedia™ is a registered trademark of the World Public Library Association, a non-profit organization.
 


Copyright © World Library Foundation. All rights reserved. eBooks from Project Gutenberg are sponsored by the World Library Foundation,
a 501c(4) Member's Support Non-Profit Organization, and is NOT affiliated with any governmental agency or department.