April 2023

S M T W T F S
      1
2345678
9101112131415
161718192021 22
23242526272829
30      

Page Summary

Style Credit

Expand Cut Tags

No cut tags
Monday, June 28th, 2010 08:38 am
A lot of time, we link to a website or an ongoing discussion rather than copying and pasting info over onto to Fanlore.  But once a website or a link is dead, that data is lost and your Fanlore entry may lack context or key info.

Your best shot is to head over to the Internet Archive (Wayback Machine) and see if the website has been archived. But since the Wayback Machine crawls and archives randomly, you won't know if your citation can be resurrected until it is too late.

Enter: WebCite. A service designed for scholars to create a static snapshot of a website so that you can cite  it (and the page contents) for longer periods.  It is user driven - you have to submit the website link before the website goes down (when you're creating your Fanlore entry). It comes with a few caveats: it won't create a snapshot of pages that have the 'no robots' code. It won't grab locked content and if you're grabbing a page from an adult Livejournal community, all you may see is the 'Adults only' warning.  And it is intended to be used in addition to the direct link to the website, not in place.

I've tested WebCite on the Professionals Fandom Timeline which pulls the bulk of its content from a few key LJ threads. We have already laboriously copied the data over to Fanlore (with permission), but it seemed like a good test candidate.

I also used WebCite to create links to a Stargate Award website that is not currently in the WayBack Machine.

If you have used this service before, or know anything more about it, please drop a note. I think it will be particularly useful for blogs and forum  posts which are prone to vanish quickly. It comes with an easy to use Bookmarklet that will allow you to cite a webpage with one click.

edited: I had a brief discussion with someone about WebCite in which they expressed discomfort with the use of this tool (and about whether aspects of the Fanlore project in general could be seen as a breach of fannish community mores/trust).   So I'll toss out this narrower question: How does using Webcite differ from our using the WayBack Machine/Internet Archive or Google as our citation sources? Both Webcite and the Wayback Machine are using the same caching process and both store the website snapshot on their servers.  What I like about WebCite is that it is much more limited - it cites only the one page and does not scrape and archive the entire website (like the WayBack Machine). This offers us a better level of control over what we're citing to, makes certain we give proper credit to the source of info and grabs the smallest portion of material. In other words, it seems (to me) to be a better form of 'fair use'. 

Thoughts? Input? Other ways of looking at the 'what to link, what to quote, what to cite' question? Is any use of any tool that caches a website (ex. Google, WayBack Machine, LJ Seek etc) something to avoid? I realize there may not be a single or uniform opinion, but like Fanlore, I think that plural POVs are good.

edited  to add:   I have to keep in mind that Fandom - and Fanlore - is not operating in isolation. Scholars, other Wikis, libraries, and historians are running into the same questions and looking at and evaluating the same tools.  In fact Wired had a recent article about the US and UK digital archives and their reliance on the Oakland Archive Policy of 2001.   More here.

And...
a recent Library Science article discussing yet another 'caching' service: Memento Web

And...
: links to legal articles on digital preservation and caching  below.

Tags:
Tuesday, June 29th, 2010 04:27 pm (UTC)
I'm somewhat bothered by the fact that the wiki on WebCite states this
Rather than relying on a web crawler which archives pages in a "random" fashion, WebCite users who want to cite web pages in a scholarly article can initiate the archiving process.
whereas the reasoning of Field v. Google expressly found
But when a
user requests a Web page contained in the Google cache by clicking on a “Cached” link, it is the
user, not Google, who creates and downloads a copy of the cached Web page. Google is passive
in this process. Google’s computers respond automatically to the user’s request. Without the
user’s request, the copy would not be created and sent to the user, and the alleged infringement at
issue in this case would not occur. The automated, non-volitional conduct by Google in response
to a user’s request does not constitute direct infringement under the Copyright Act.


The two operations are almost complete opposites of each other. Furthermore, the Google result also depends on an implied licence agreement which depended on Field knowing that Google was going to crawl over his material and cache it. As I uncerstand it, this isn't the case with WebCite; people make a choice whether or not to cache it and that choice is made on an item by item basis. Finally, the Court held that Google was entitled to the safe harbor provisions of the DCMA; again, I'm not sure if these would be available to WebCite.

I'm not arguing that WebCite's activities might not qualify for being fair use, but I don't think Field v. Google goes nearly so far to justify them as they seem to think it does.
Friday, July 2nd, 2010 07:19 am (UTC)
wow, that's the weirdest view of the word "cache" ever. I am certain that Google stores the full text of indexed documents both in the inverted (searchable) index and in the original order, in order to do snippets as well as view as html. I suppose it could be argued that it's a potential document until rendered, but then all web pages pretty much fit that criterion.

Luckily, Google has enough money to buy fleets of lawyers, and in this they seem to be on the side of public access. Unlike the Google Libraries digitization project, where they want to charge for access...
Friday, July 2nd, 2010 07:39 am (UTC)
I think the reasoning in Field was flaky in the extreme; basically the judge seems to have seen a clear public policy reason to allow Google caching and did whatever he needed to reach that result. Which is why I suspect WebCite are putting themselves out on a limb by relying on it for a different business model.
Friday, July 2nd, 2010 01:02 pm (UTC)
I'm not a lawyer, but I think WebCite is illegal under German Copyright law, and the Wayback Machine as well.
Friday, July 2nd, 2010 05:33 pm (UTC)
The ruling says that publishing a website (and optimizing it for search engines) without a robots.txt implies consenting to being indexed in the Google Image Search. You can take back your consent, and then a search engine isn't allowed to show your work anymore. There's a passage in the ruling where they point out that images that are deleted from the indexed website disappear from the Google Search as soon as possible, too.

But WebCite (and the Archive) preserve those sites for when they get deleted, and I think that deleting a website probably counts as withdrawing consent.

Disclaimer: What I've found wrt WebCite's legality in Germany predates the BGH ruling (and is all in German. :() Generally most rulings wrt copyright used to stress the necessity of opt-in procedures instead of this opt-out, so we'll see how it goes.
Edited 2010-07-02 05:33 pm (UTC)