-
Notifications
You must be signed in to change notification settings - Fork 24
Xkcd #21
Comments
@fazo96 how are you going to manage the comics that are dynamic or contain multiple sequential images? Or the map ones that have a larger version available on click? I would loooooove to have this. But we should make sure that randall is okay with it first, i'm not sure if there are any sort of copyrights involved here. (too bad he doesnt use github, or we could just ping him) |
absolutely. thanks for saying this. looks like everything is released CC-BY-NC
i'm sure has an account. just have to find it \o/
(Alternatively, mirror the RSS feed?) |
The title and the alt text could be stored in the png metadata. You can use ImageMagic: see this. |
Looks like license is not an issue as long as we provide credit to randall and include a copy of the license. Also:
Uhm, I just found this in the About page of xkcd.com:
Getting the data will be a lot easier this way (no html parsing involved) EDIT: I wrote a node script that downloads and organizes data from xkcd.com and it worked! I created a partial copy of xkcd.com to see if you like the setup (so that we can create a full copy later). I included Randall's about and license pages and my script in the folder 👍 You can check it out here: QmSeYATNaa2fSR3eMqRD8uXwujVLT2JU9wQvSjCd1Rf8pZ I'm thinking about writing a simple |
👍
I think #7 is also quite relevant here. |
I completed the archive (now every image file and more is available via ipfs), it just a needs a viewer and probably better folder structure. Here you go: QmPVP4sDre9rtYahGvcjv3Fqet3oQyqrH5xS33d4YBVFme |
@fazo96 👏 |
@fazo96 👍 Can we zero pad the numbers on the next pass? :) |
@cryptix yeah I figured it was necessary :) if you'd like a try, the script I used to generate the directory tree is included in the directory. It's named If I have time I'll implement it |
I have scraped the entirety of xkcd.com and
Switch explanation:
Archiving notes:
Archive links (newest to oldest): (My IPNS entry might be more up to date) |
Looks like the currently referenced version on the website isn't fully available. |
@fazo96 do you have the original archive that's currently linked to on the archives.ipfs.io site? https://ipfs.io/ipfs/QmPVP4sDre9rtYahGvcjv3Fqet3oQyqrH5xS33d4YBVFme It doesn't currently seem to be fully available, but if you still have it I can pin it to my ipfs node. I'd try to reproduce the archive using the script in the archive, but I could only guess what the exact text was in the about and license files. |
FWIW I just generated a new version of fazo96's archive that's linked to from the site and pinned it to my ipfs node, so the comics that I couldn't access through the gateway before (in the archive linked to from archives.ipfs.io) now seem to be accessible. The |
Awesome, gonna pull that onto one of our storage nodes too. @leerspace wanna make a PR to update the site? |
Cool thanks, I just updated https://archives.ipfs.io |
@leerspace sorry for replying late, looks like I lost my copy of the original archive. Thanks for updating it! 👍 |
Hello, I've updated the archive using the |
Awesome, thanks @chosenken -- also pinned it on nihal.i.ipfs.io |
Updated again to 1864, but this time attached it to an ipns: QmTaW8vRj4SkM6JhqVhAsibQE9PdJb5PQ2FMwPPc6gBi2h. I might work on a script that pulls new comics down and updates the ipns when it changes. |
I'd like to update this one again, but to facilitate programmatic access, I'd like to change the structure slightly to something more like:
where the comic files are contained within a 'folder' defined by the number rather than number and name. Any issues with this? I can host on our server, but I'd also be happy to submit a PR to update the archives. |
@carsonfarmer that'd be rad. I've no objection to simplifying the folder structure. |
I plan to feature this data set on the start page of the new IPLD Explorer page in the ipfs-webui. |
@carsonfarmer could we get some zero padding on those indexes?
|
Ah sorry, was on vacation. Yes I'll update the indexes and post here when ready. |
I've written a new program in go that creates an archive such as the following, /ipfs/QmdAChzF2JQCx9icrmYHZhFdRSv9TpRjq5q1v5b3ANpxRf. It also includes a csv with an index of post titles, published date and post number. I have submitted a pr, ipfs/awesome-ipfs#193 |
If I were someone who wanted to start pinning content like XKCD their own node(s) to help network redundancy, which hash would I use? There are many different hashes presented in this thread and it's not clear to me which one is the most relevant or up-to-date. This seems to be the most up-to-date: Qmb8wsGZNXt5VXZh1pEmYynjB6Euqpq3HYyeAdw2vScTkQ This one seems to be an exact replica of the first but it has a different hash, perhaps due to the lack of .html on "about" and "license": QmPVP4sDre9rtYahGvcjv3Fqet3oQyqrH5xS33d4YBVFme https://archives.ipfs.io/ seems to favor the first one, TkQ. And this seems to be the same content as everything above but in a different structure: https://ipfs.io/ipfs/QmdAChzF2JQCx9icrmYHZhFdRSv9TpRjq5q1v5b3ANpxRf So, which hash do I pin? |
I think it would be a good idea to zero pad the numbers: right know it's impossible to browse. |
I added padding to the downloader: /ipfs/QmX4pR3KKdivwY9Pn5mHNYi5FRhtTqapMfFmW4SYesstxU/xkcd-downloader.js |
I've just created an XKCD archive at |
@Stebalien should we update the index to |
I take it your comment is moot now that this has all moved to awesome.ipfs.io? |
Is it possible to have two links? Ideally, we'd link to an immutable version as well. |
@Stebalien we could also add an immutable version, but that would be a snapshot somewhere in the past. But yes, we could add it as a description perhaps. |
Yeah, I know. It's just that archives that rely on DNS sketch me out a bit. |
Please see ipfs/awesome-ipfs#261. |
I plan to archive all the comics in http://xkcd.com/
I think i'll use (comicnumber)-(comictitle).png for the image and figure out how to save the alt-text in the png metadata
Please post if you want to keep a copy of the archive or you manage to create it before I do :)
The text was updated successfully, but these errors were encountered: