--- Log opened Sat Feb 18 00:00:00 2012 --- Day changed Sat Feb 18 2012 00:00 < rdb> those are pretty much the three reasons people come to amsterdam 00:00 < yashgaroth> I'm meeting family there anyway, but surely there's museums and stuff 00:00 < yashgaroth> though I was there a couple years ago and saw all the museums, so hookers & drugs it is 00:03 < rdb> you used to be able to buy psilocybin mushrooms as a tourist, but they banned that in 2008, you can only buy a growkit now. psilocybin truffles are still legal though 00:03 < Stee|> rdb: How far is anywhere in the netherlands from anywhere else in the netherlands? 00:03 * rdb hates governments. 00:03 < yashgaroth> 10 minutes 00:03 < rdb> Stee|, how do you mean that? 00:03 < Stee|> like, how far would it take you to get to amsterdam 00:04 < rdb> I live relatively close, but it probably would take me still an hour or two 00:04 < Stee|> how long, rather 00:04 < Stee|> ah 00:04 < Stee|> clearly you should go get drunk with yashgaroth 00:04 < yashgaroth> where you at rdb, the hague? 00:04 < rdb> I don't drink alcohol. 00:05 < rdb> I don't think that ethanol brings me any effects that I find useful, so... plus its destructive, unlike many other drugs 00:05 < rdb> yashgaroth, near gouda 00:06 < yashgaroth> ah, I hope to visit the cheese market 00:06 < rdb> <3 gouda cheese 00:06 < yashgaroth> awww yeeee 00:23 < Stee|> welp 00:23 < Stee|> I'mm going to go lay down I think 00:23 < Stee|> *I'm 00:25 < yashgaroth> same here, g'night 00:25 -!- yashgaroth [~f@cpe-24-94-5-223.san.res.rr.com] has quit [Quit: Leaving] 00:54 -!- Mokbortolan_ [~Nate@c-71-59-241-82.hsd1.or.comcast.net] has joined ##hplusroadmap 00:55 -!- Mokbortolan_1 [~Nate@c-71-59-241-82.hsd1.or.comcast.net] has quit [Quit: Leaving.] 01:03 -!- ianmathwiz7 [~chatzilla@x-134-84-100-61.reshalls.umn.edu] has quit [Ping timeout: 245 seconds] 02:23 -!- d3nd3 [~dende@cpc10-croy17-2-0-cust245.croy.cable.virginmedia.com] has joined ##hplusroadmap 02:38 -!- klafka [~textual@ip-64-139-28-14.sjc.megapath.net] has joined ##hplusroadmap 02:55 -!- marainein [~marainein@114-198-65-190.dyn.iinet.net.au] has joined ##hplusroadmap 03:06 -!- chris_99 [~chris_99@unaffiliated/chris-99/x-3062929] has joined ##hplusroadmap 03:48 < rdb> the more I learn about it, the more I want to get a magnetic implant 04:19 -!- yottabit [~heath@unaffiliated/ybit] has joined ##hplusroadmap 04:32 -!- yottabit [~heath@unaffiliated/ybit] has quit [Quit: Konversation terminated!] 04:50 < chris_99> http://www.genome.gov/images/content/cost_per_genome.jpg 04:51 -!- ThomasEgi [~thomas@pppdyn-6e.stud-ko.rz-online.net] has joined ##hplusroadmap 04:51 -!- ThomasEgi [~thomas@pppdyn-6e.stud-ko.rz-online.net] has quit [Changing host] 04:51 -!- ThomasEgi [~thomas@panda3d/ThomasEgi] has joined ##hplusroadmap 05:06 < archels> kanzure: crazy prices for this old mechanical stuff, http://www.sciquip.com/browses/detailed_item_view.asp?productID=26051&Mfg=MICROMANIPULATOR&Mdl=550 05:14 < archels> http://www.ebay.de/itm/Motorisierter-dreiachsiger-Mikromanipulator-Marzhauser-Steuergerat-/290669989423?pt=Laborger%C3%A4te_instrumente&hash=item43ad480e2f 05:16 < chris_99> what's that archels? 05:21 < archels> two XYZ micromanipulators 05:41 -!- jmil [~jmil@c-68-81-252-40.hsd1.pa.comcast.net] has quit [Read error: Connection reset by peer] 06:12 -!- augur [~augur@208.58.5.87] has quit [Remote host closed the connection] 06:45 -!- jmil [~jmil@c-68-81-252-40.hsd1.pa.comcast.net] has joined ##hplusroadmap 07:15 -!- ParahSailin__ [~parahsail@adsl-69-151-205-240.dsl.hstntx.swbell.net] has joined ##hplusroadmap 07:17 -!- ParahSailin [~parahsail@unaffiliated/parahsailin] has quit [Ping timeout: 248 seconds] 07:25 -!- JayDugger [~duggerj@pool-173-74-78-36.dllstx.fios.verizon.net] has quit [Quit: Leaving.] 08:33 -!- anelma [~elmom@hoas-fe3ddd00-25.dhcp.inet.fi] has quit [Remote host closed the connection] 08:33 -!- augur [~augur@129.2.129.35] has joined ##hplusroadmap 08:37 < ParahSailin__> morning 08:42 < rdb> morning 08:42 < rdb> evening actually 08:43 -!- jmil [~jmil@c-68-81-252-40.hsd1.pa.comcast.net] has quit [Quit: jmil] 08:59 -!- ianmathwiz7 [~chatzilla@x-134-84-100-61.reshalls.umn.edu] has joined ##hplusroadmap 09:01 -!- d3nd3 [~dende@cpc10-croy17-2-0-cust245.croy.cable.virginmedia.com] has quit [Remote host closed the connection] 09:17 -!- jmil [~jmil@SEASNet-148-05.seas.upenn.edu] has joined ##hplusroadmap 09:19 -!- elmom [~elmom@hoas-fe3ddd00-25.dhcp.inet.fi] has joined ##hplusroadmap 09:37 -!- pasky_ [pasky@nikam.ms.mff.cuni.cz] has joined ##hplusroadmap 09:37 -!- jrayhawk_ [~jrayhawk@nursie.omgwallhack.org] has joined ##hplusroadmap 09:37 -!- Mokbortolan_ [~Nate@c-71-59-241-82.hsd1.or.comcast.net] has quit [Ping timeout: 240 seconds] 09:37 -!- pasky [pasky@nikam.ms.mff.cuni.cz] has quit [Ping timeout: 240 seconds] 09:37 -!- jrayhawk [~jrayhawk@nursie.omgwallhack.org] has quit [Ping timeout: 240 seconds] 09:37 -!- gedankenstuecke [~bastian@phylomemetic-tree.de] has quit [Ping timeout: 240 seconds] 09:37 -!- epitron [~epitron@unaffiliated/epitron] has quit [Ping timeout: 240 seconds] 09:37 -!- epitron [~epitron@bito.ponzo.net] has joined ##hplusroadmap 09:37 -!- gedankenstuecke [~bastian@phylomemetic-tree.de] has joined ##hplusroadmap 09:37 -!- epitron [~epitron@bito.ponzo.net] has quit [Changing host] 09:37 -!- epitron [~epitron@unaffiliated/epitron] has joined ##hplusroadmap 09:38 -!- Mokbortolan_ [~Nate@c-71-59-241-82.hsd1.or.comcast.net] has joined ##hplusroadmap 09:43 < chris_99> anyone heard of transcranial direct current stimulation 09:49 < archels> Some dude in ##neuroscience just asked about it. ;) 09:50 < mag1strate> isnt it mostly used for pschological disorders? 09:51 < chris_99> yeah i know, thats why i was asking archels 09:52 < chris_99> and yes it does seem so, although it might be similar to TMS 09:52 < chris_99> in some ways 09:54 < mag1strate> hmmm I had no clue 09:54 < mag1strate> it would be interesting to look into something like this 09:56 < chris_99> yeah and it looks easy to experiment with 09:56 < chris_99> as its just low current DC 09:56 < chris_99> low voltage too 09:57 < chris_99> although i don't really fancy attaching electrodes with electricity to my head 09:58 < mag1strate> thats usually a smart move lol 09:58 < mag1strate> the problem is I dont see any practical applications for it 09:59 < mag1strate> most is used to treat disease 09:59 < mag1strate> unless we can control where the current will be flowing to 09:59 < chris_99> could could make a DIY TMS device 10:00 < mag1strate> you can make one 10:00 < chris_99> oops bad spelling there 10:00 < chris_99> has anyone done that? 10:00 < mag1strate> but it would be nice to experiement where the electrodes would go to make it benificial 10:12 < kanzure> yes lots of others in here have heard about tdcs 10:12 < kanzure> and i think one or two built a tdcs setup 10:13 < kanzure> although i think collecctively this channel has more experience with magnetic stimulation 10:13 -!- strages_home [~strages@adsl-98-67-175-14.shv.bellsouth.net] has joined ##hplusroadmap 10:15 < chris_99> has anyone built a magnetic stimulation device? 10:18 < kanzure> superkuh worked on something 10:18 < kanzure> at the moment i'm more interested in ultrasound stimulation 10:18 < kanzure> http://diyhpl.us/~bryan/papers2/neuro/ultrasound/ 10:19 < mag1strate> I enjoy that I can't give myself and want to share with someone special. 10:19 < mag1strate> I enjoy that I can't give myself and want to share with someone special. 10:19 < mag1strate> lol 10:19 < mag1strate> my middle click button is a paste and enter button 10:20 < kanzure> "it is thought that the nonthermal actions of US are understood in terms of cavitation - for example, radiation force, acoustic streaming, shock waves, and strain neuromodulation, where US produces fluid-mechanical effects on the cellular environments of neurons to modulate their resting membrane potentials." 10:20 < kanzure> "The direct activation of ion channels by US may also represent a mechanism of action, since many of the voltage-gated sodium, potassium, and calcium channels influencing neuronal excitability possess mechanically sensitive gating kinetics (Morris and Juranka, 2007)." 10:20 < chris_99> oh i've not heard of ultrasound stimulation 10:21 < mag1strate> do you know of any positive effects of this kanzure? 10:21 < chris_99> i'm hopefully in the process of ordering some ultrasound transducers from china 10:21 < kanzure> mag1strate: 2mm targetting of regions in the brain 10:21 < mag1strate> I can see maybe positive effects on the cellular environemtn level 10:21 < kanzure> it's neural stimulation 10:21 < mag1strate> hmmm 10:21 < kanzure> soo if you have a 2mm chunk you want to stimulate somewhere.. it's pretty useful 10:21 < mag1strate> I've never really seen US used on the brian 10:21 < kanzure> rTMS has more like 1cm resolution 10:21 < kanzure> mag1strate: check those papers.. 10:22 < kanzure> one of the studies was to remove an inoperable brain tumor 10:22 < kanzure> by melting the tumor. 10:25 < mag1strate> was it successful? 10:25 < kanzure> yes 10:26 < mag1strate> noice 10:26 < mag1strate> if it was for melting the tumor, would it effect actual brain tissue? 10:26 < kanzure> really the ideal setup would be one where you can apply a certain amount of energy to any location within the brain 10:26 < kanzure> mag1strate: the brain tumor study was just a high-power version 10:27 < kanzure> here's a low power version: 10:27 < kanzure> http://www.youtube.com/watch?v=RGEP6iWLsvQ 10:28 < mag1strate> interesting 10:28 < mag1strate> it seemed the low power version seemed almost like an impulsive shock to the brain area 10:29 < kanzure> yes.. it's a mechanical compression wave that goes into the skull 10:29 < kanzure> when you have 10 or 50 transducers the compression waves add up 10:30 < kanzure> so when they geometrically intersect the power delivery increases 10:30 < kanzure> erm.. the total mW/mm^2 increases. you get the idea. 10:31 < kanzure> i just had the most fascinating time traveling dream 10:32 < kanzure> apparently it's a dream of mine to one day own giraffes and t-rex's and force them to fight against each other 10:32 < mag1strate> lolol 10:32 < mag1strate> That would be really cool actually 10:33 < mag1strate> I have always wanted sharks with lazer beams on their heads :/ 10:33 < kanzure> a shark tank doesn't cost that much 10:38 < kanzure> http://omicsomics.blogspot.com/2012/02/oxford-nanopore-doesnt-disappoint.html 10:40 < chris_99> is their presentation online? 10:40 < mag1strate> kanzure: lol 10:41 < kanzure> however you might have to go wrestle your own shark off the coast 10:48 < mag1strate> Thats the easy part 10:48 < mag1strate> the hardest part is the lazer beams 10:49 < ParahSailin__> as insty would say "faster, please" 11:02 < kanzure> i prefer this edit of doc brown: http://www.youtube.com/watch?v=KJRh-37H4fA 11:06 < chris_99> haha 11:12 < chris_99> isn't it rather dangerous someone could get the DNA for the pneumonic plague off the net? 11:14 < rkos> i think synthesizing companies ban sequences of too dangerous things 11:14 < chris_99> thats scary as hell to me though 11:14 < kanzure> chris_99: doesn't matter, they already have it 11:15 < chris_99> who already have it? 11:15 < kanzure> obscurity is not security 11:15 < chris_99> true i agree with that normally, but in this case 11:15 < kanzure> the best defense against plagues is a biological solution 11:15 < kanzure> we have an immune system for a reason 11:16 < chris_99> is there a vaccine for the plague? 11:16 < kanzure> if there isn't, sounds like an important thing to make, no? 11:17 < chris_99> it does yeah, but i'm really suprised the DNA is available 11:18 < Stee|> no hangover, this is good 11:18 < rkos> is it available? 11:18 < chris_99> yes 11:19 < rkos> but dont dna foundries refuse to synthesize sequences of viruses etc? 11:20 < kanzure> jrayhawk_: do you have any interest in doing a mirror of ftp://ftp.ncbi.nlm.nih.gov/genomes/ 11:20 < jrayhawk_> Could do. 11:21 -!- jrayhawk_ is now known as jrayhawk 11:21 < kanzure> to my knowledge, there are no mirrors 11:21 < jrayhawk> Huh. 11:21 < kanzure> which is bad. 11:22 < kanzure> oh 11:22 < kanzure> http://biomirror.aarnet.edu.au/biomirror/ncbigenomes/ 11:22 < kanzure> http://mirrors.vbi.vt.edu/mirrors/ftp.ncbi.nih.gov/genomes/ 11:23 < kanzure> well, there should be a non-institutional mirror somewhere 11:23 < kanzure> http://ftp.cbi.pku.edu.cn/pub/database/genomes/ 11:23 < jrayhawk> What's wrong with institutional mirrors 11:23 < kanzure> and some of these mirrors look a bit stale (that last one was from 2010?) 11:23 < kanzure> i don't trust universities to always keep them up 11:23 < kanzure> if feds come knocking, etc. 11:23 < jrayhawk> Ah, I see. 11:24 < kanzure> also, apparently i don't trust these universities to keep their mirrors current o__o 11:24 < kanzure> nice set of backups: http://ftp.cbi.pku.edu.cn/pub/database/ 11:26 < kanzure> weird they're using some perl module i think, but it doesn't appear on cpan 11:26 < kanzure> http://ftp.cbi.pku.edu.cn/pub/biomirror/software/biomirror/BioMirror.pm 11:26 < kanzure> http://ftp.cbi.pku.edu.cn/pub/biomirror/software/biomirror/BioMirror/ 11:26 -!- delinquentme [~asdfasdf@c-67-171-66-113.hsd1.pa.comcast.net] has joined ##hplusroadmap 11:49 < Urchin> mirrors of what? 12:03 -!- _sol_ [Sol@c-174-57-58-11.hsd1.pa.comcast.net] has quit [Ping timeout: 260 seconds] 12:08 -!- _sol_ [Sol@c-174-57-58-11.hsd1.pa.comcast.net] has joined ##hplusroadmap 12:27 < kanzure> Urchin: gnomes 12:27 < kanzure> *genomes 12:29 < jrayhawk> the world has enough gnome mirrors 12:33 -!- Technicus [~Technicus@108.198.137.39] has joined ##hplusroadmap 12:40 -!- Technicus [~Technicus@108.198.137.39] has quit [] 12:47 -!- Jaakko96 [~Jaakko@host86-131-178-213.range86-131.btcentralplus.com] has joined ##hplusroadmap 12:50 < kanzure> hi jrayhawk 12:50 < kanzure> erm.. Jaakko96 12:59 < ParahSailin__> kanzure, i saw that mutual banking essay on your pdf directory -- this is similar http://praxeology.net/FDT-VS.htm 13:01 < kanzure> alright 13:20 -!- Mokbortolan_1 [~Nate@c-71-59-241-82.hsd1.or.comcast.net] has joined ##hplusroadmap 13:22 -!- Mokbortolan_ [~Nate@c-71-59-241-82.hsd1.or.comcast.net] has quit [Ping timeout: 265 seconds] 13:35 -!- yashgaroth [~f@cpe-24-94-5-223.san.res.rr.com] has joined ##hplusroadmap 13:35 -!- archels [~foo@sascha.esrac.ele.tue.nl] has quit [Ping timeout: 245 seconds] 13:37 < kanzure> hi yashgaroth 13:37 < yashgaroth> hello 13:41 < kanzure> why would onLoadFinished() be called 3 times for saks, but not 3 times for heybryan? http://pastebin.com/4csM0qSC 13:41 < kanzure> ^for anyone who wants to help out with some javascript 13:47 -!- archels [~foo@sascha.esrac.ele.tue.nl] has joined ##hplusroadmap 13:51 -!- ParahSailin__ [~parahsail@adsl-69-151-205-240.dsl.hstntx.swbell.net] has quit [Ping timeout: 260 seconds] 14:07 -!- ParahSailin__ [~parahsail@adsl-69-151-205-240.dsl.hstntx.swbell.net] has joined ##hplusroadmap 14:13 < kanzure> aha.. http://code.google.com/p/phantomjs/issues/detail?id=122 14:14 -!- _sol_ [Sol@c-174-57-58-11.hsd1.pa.comcast.net] has quit [Ping timeout: 240 seconds] 14:21 -!- ianmathwiz7 [~chatzilla@x-134-84-100-61.reshalls.umn.edu] has quit [Ping timeout: 260 seconds] 14:28 -!- Jaakko96 [~Jaakko@host86-131-178-213.range86-131.btcentralplus.com] has quit [Quit: Nettalk6 - www.ntalk.de] 14:36 -!- strages_home [~strages@adsl-98-67-175-14.shv.bellsouth.net] has quit [Ping timeout: 245 seconds] 14:38 -!- strages_home [~strages@adsl-98-67-175-14.shv.bellsouth.net] has joined ##hplusroadmap 14:38 -!- jmil [~jmil@SEASNet-148-05.seas.upenn.edu] has quit [Read error: Operation timed out] 14:40 -!- d3nd3 [~dende@cpc10-croy17-2-0-cust245.croy.cable.virginmedia.com] has joined ##hplusroadmap 14:42 -!- d3nd3 [~dende@cpc10-croy17-2-0-cust245.croy.cable.virginmedia.com] has quit [Client Quit] 14:59 -!- ianmathwiz7 [~chatzilla@x-134-84-100-61.reshalls.umn.edu] has joined ##hplusroadmap 14:59 < kanzure> hi ianmathwiz7 15:01 < ianmathwiz7> hey 15:06 < ThomasEgi> hoho ianmathwiz7 , long time no chat^ 15:06 < ianmathwiz7> yeah 15:06 < ianmathwiz7> I've been on the ##biohack channel from time to time 15:07 < ianmathwiz7> but I haven't been spending too much time on IRC, lately 15:13 -!- jmil [~jmil@c-68-81-252-40.hsd1.pa.comcast.net] has joined ##hplusroadmap 15:18 -!- ianmathwiz7 [~chatzilla@x-134-84-100-61.reshalls.umn.edu] has quit [Quit: ChatZilla 0.9.88 [SeaMonkey 2.7.1/20120208224119]] 15:22 < kanzure> win 13 15:23 -!- augur [~augur@129.2.129.35] has quit [Remote host closed the connection] 15:23 < bkero> lose 12 15:29 -!- chris_99 [~chris_99@unaffiliated/chris-99/x-3062929] has quit [Quit: Leaving] 16:03 < kanzure> http://www.reddit.com/r/science/comments/pv7tn/the_high_price_of_knowledge_should_journals/ 16:03 < kanzure> http://thepiratebay.se/torrent/6554331/Papers_from_Philosophical_Transactions_of_the_Royal_Society__fro 16:05 < ParahSailin__> https://docs.google.com/open?id=0B3qaT-ZL6aeKMWFhNmIwOGYtNWM2Yi00ZTU0LTkxZjMtZGYzNjUwNWJhZTBm 16:05 < ParahSailin__> like bitcoin but different 16:05 < kanzure> what is this? 16:05 < kanzure> http://bib.tiera.ru/static 16:05 < kanzure> seems to be 24 MB 16:06 < kanzure> hmm this is a very poorly organized collection 16:07 < kanzure> 16:08 < kanzure> Cavaleiro A. - Nanostructured Coatingspdfen 16:08 < kanzure> how the fuck am i supposed to fix this? 16:10 < superkuh> I don't know. But at least it's easy to search through and extract URLs. Neat. 16:10 < superkuh> Tangentially, http://erewhon.superkuh.com/library/ - always on (fast) daily mirror of my library for when I turn off the main server. 16:11 < kanzure> superkuh: we need some way of helping people who keep these mirrors to have better indexes 16:11 < kanzure> for instance, at the moment i can't confirm that their copy of journal xyz is complete or not 16:11 < kanzure> and they probably don't know which papers belong to which journals either 16:12 < kanzure> bibtex could probably work for this? with some tools to search through directories and get file hashes 16:13 < delinquentme> http://imgur.com/r/atheism/VEKo5 17:29 -!- marainein [~marainein@114-198-65-190.dyn.iinet.net.au] has quit [Ping timeout: 260 seconds] 17:43 -!- ThomasEgi [~thomas@panda3d/ThomasEgi] has quit [Remote host closed the connection] 18:14 < uniqanomaly> http://news.slashdot.org/story/12/02/18/2130245/universities-agree-to-email-monitoring-for-copyright-agency :> 18:31 < kanzure> eww slashdot 18:31 < kanzure> how is that not dead yet 18:56 < kanzure> Agrawal S. Protocols for Oligonucleotide Conjugates[c] Synthesis and Analytical Techniquespdfen25521154034 18:57 < kanzure> http://bib.tiera.ru//DVD-022/Arnold_F.H.,_Georgiou_G._Directed_Evolution_Library_Creation._Methods_and_Protocols_(2003)(en)(232s).pdf 18:58 < kanzure> http://bib.tiera.ru//DVD-022/Cutler_P._Protein_Purification_Protocols_(2003)(2nd)(en)(496s).pdf 18:58 < kanzure> http://bib.tiera.ru//DVD-035/Dooman_S._Protein_Purification_Protocols_(1996)(en)(416s).pdf 18:59 < kanzure> http://bib.tiera.ru//DVD-022/English_L.B._(ed.)_Combinatorial_Library_Methods_and_Protocols_(2002)(en)(383s).pdf 18:59 < kanzure> http://bib.tiera.ru//DVD-022/Federico_M._Lentivirus_Gene_Engineering_Protocols_(2003)(en)(328s).pdf 18:59 < kanzure> http://bib.tiera.ru//DVD-022/Fedoroff_S.,_Richardson_A._Protocols_for_Neural_Cell_Culture_(2001)(3rd)(en)(384s).rar 19:00 < kanzure> http://bib.tiera.ru//DVD-022/Findeis_M.A._Nonviral_Vectors_for_Gene_Therapy._Methods_and_Protocols_(2001)(en)(416s).pdf 19:01 < kanzure> http://bib.tiera.ru//dvd57/Foster G. D. (Ed), Taylor S. (Ed) - Plant Virology Protocols, Vol. 81(1998)(571).rar 19:01 < kanzure> http://bib.tiera.ru//DVD-028/Gartland_K.M.A.,_Davey_M.R._(eds.)_Agrobacterium_Protocols_(1995)(en)(432s).pdf 19:01 < kanzure> probably useless http://bib.tiera.ru//DVD-022/Graham_C.A._(ed.),_Hill_A._(ed.)_DNA_Sequencing_Protocols_(2001)(2-nd)(en)(244s).rar 19:01 < kanzure> http://bib.tiera.ru//DVD-022/Gray_J.,_Desselberger_U._Rotaviruses._Methods_and_Protocols_(2000)(en)(272s).pdf 19:02 < kanzure> http://bib.tiera.ru//DVD-028/Harwood_A.J._Basic_DNA_and_RNA_Protocols_(1996)(en)(528s).pdf 19:02 < kanzure> http://bib.tiera.ru//DVD-022/Hofker_M.H.,_van_Deursen_J._Transgenic_Mouse_Methods_and_Protocols_(2002)(en)(392s).pdf 19:03 < kanzure> http://bib.tiera.ru//DVD-022/Howe_P.H._Transforming_Growth_Factor-Beta_Protocols_(2000)(en)(176s).pdf 19:03 < kanzure> http://bib.tiera.ru//DVD-030/Irvine_G.B.,_Williams_C.H._Neuropeptide_Protocols_(1996)(en)(381s).pdf 19:03 < kanzure> http://bib.tiera.ru//DVD-035/Jones_G.E._Human_Cell_Culture_Protocols_(1996)(en)(560s).pdf 19:03 < kanzure> http://bib.tiera.ru//DVD-031/Jones_H._Plant_Gene_Transfer_and_Expression_Protocols_(1995)(en)(462s).pdf 19:04 < kanzure> http://bib.tiera.ru//DVD-034/Kendall_D.A.,_Hill_S.J._Signal_Transduction_Protocols_(1995)(en)(316s).pdf 19:04 < kanzure> http://bib.tiera.ru//DVD-022/Kmiec_E.B._(ed.)_Gene_Targeting_Protocols_(1999)(en)(450s).pdf 19:04 < kanzure> http://bib.tiera.ru//DVD-022/Kola_I.,_Tymms_M.J._Gene_Knockout_Protocols_(2001)(en)(448s).pdf 19:04 < kanzure> http://bib.tiera.ru//DVD-029/LaRossa_R.A._Bioluminescence_Methods_and_Protocols_(1998)(en)(320s).pdf 19:05 < kanzure> http://bib.tiera.ru//DVD-022/Lieberman_B.A._Steroid_Receptor_Methods._Protocols_and_Assays_(2001)(en)(400s).pdf 19:05 < fenn> books books books books books books books books books 19:05 < kanzure> http://bib.tiera.ru//DVD-022/Lo_B._(ed.)_Antibody_Engineering,_Methods_and_Protocols_(2003)(en)(550s).pdf 19:05 < kanzure> booooooks 19:05 < kanzure> fenn: help me figure out a way to fix this 19:05 < fenn> what's the problem? 19:05 < kanzure> lots of people scrape and download lots of this content (which is great) 19:05 < kanzure> their organization skills suck 19:06 < kanzure> and they can't show that they have a complete collection for a given journal, edition, volume, or whatever 19:06 < fenn> hmm 19:06 < kanzure> ideally there's some folder organization scheme, or metadata file format that we should all be using 19:06 < kanzure> and writing tools to use. 19:06 < kanzure> and then some web service that collects this information and maps it all together. 19:06 < fenn> i assume the metadata is publically available 19:06 < kanzure> it's definitely available at the original soruce 19:06 < kanzure> *source 19:06 < fenn> then it's a "simple matter of coding" to match the files to the metadata 19:07 < kanzure> but most people just take a pdf and give it a title and they feel they are done 19:07 < kanzure> i know that's what i did (because i'm an idiot) 19:07 < fenn> unfortunately you will probably have to rewrite parsing code for each publisher 19:07 < kanzure> well, how about bibtex 19:07 < fenn> yeah it's the curse of filesystems 19:08 < fenn> what, have a bibtex file for every pdf? i think the problem is people only download a pdf so the (computer parseable) metadata gets lost 19:08 < kanzure> well let's only consider scraping scenarios 19:08 < fenn> you could build up a hash table of pdf to metadata 19:08 < kanzure> where the programmer can afford to take extra metadata 19:08 < fenn> sha256(foo.pdf) => metadata 19:08 < kanzure> except the pdfs are modified each time you download them 19:08 < fenn> o rly 19:09 < fenn> what doesn't change then 19:09 < kanzure> yeah like AdobePDFStamperPro (watermarking) 19:09 < kanzure> sometimes there's metadata in the file itself 19:09 < fenn> is there a title field or something at least that stays the same? 19:09 < kanzure> sometimes, but for the vast majority of content it's just a scanned image 19:09 < fenn> how does mendeley do it? 19:09 < kanzure> proprietary ocr 19:09 < kanzure> as a fallback. 19:10 < fenn> how hard is OCR then? 19:10 < fenn> i mean, you just need to match the text to something in a list of titles 19:10 < kanzure> tesseract was pretty awful. i don't know, i think the pain of ocr probably goes down for sufficiently large collections 19:10 < kanzure> anyway, let's assume this isn't a problem 19:10 < kanzure> let's assume that programmers will scrape metadata too 19:10 < fenn> "programmers"? 19:10 < kanzure> i'm thinking of a particular scenario where people are scraping content 19:11 < kanzure> and contributing it to the master collection 19:11 < kanzure> and then being able to say "Ok, the internet has pages 1-400 of journal xyz" based on collected records on the server 19:11 < kanzure> "THE INTERNET" well.. "this public service" 19:12 < kanzure> then if you are feeling like you want to contribute, maybe you'd write a compatible scraper to gather, dump and upload data for "x, y and z" that the server says is missing 19:13 < fenn> how to keep it from falling apart again when the site gets shut down? 19:13 < fenn> presumably people would have backups or partial backups 19:13 < kanzure> and the software is separate anyway 19:13 < fenn> but if the pdf hash changes.. 19:14 < kanzure> *shrug* that just hurts verification 19:14 < kanzure> web of trust bullshit can maybe offset that 19:14 < kanzure> "hey look, all the contributions from anonymous user with this public key are all awful" 19:14 < fenn> oh i didnt even think of that 19:15 < fenn> something like git annex would be preferred 19:15 < fenn> so you as the library maintainer would only accept patches that look legit 19:15 < fenn> i've never seen any working web of trust software 19:15 < kanzure> sure. and again different computers and friendlies will contribute 'chunks' of scraped content / metadata that gets distributed from some primary server 19:15 < fenn> (doesnt mean it doesnt exist) 19:16 < kanzure> well, i do think contributions would need to be reviewed somehow 19:16 < kanzure> a basic method might be "show us your scraper" 19:16 < fenn> wouldnt looking at the content make more sense? 19:17 < kanzure> yes but how am i going to manually look at 7 million articles a month 19:17 < kanzure> there might be some computational way to determine if a paper looks like what it says it is 19:17 < fenn> you do a random sampling 19:17 < kanzure> well ok. 19:17 < fenn> yes, document clustering 19:17 < kanzure> so far i've met nobody that has indicated they would want to spam a service like this 19:17 < fenn> there really arent very many people working for journals 19:18 < kanzure> hm? 19:18 < fenn> many many more pissed off students who don't have access and want to fix this broken situation 19:18 < kanzure> right.. and "Here's 20,000 files with some bad names" does not help as much as it could 19:18 < fenn> i think you're just going to have to ignore filenames 19:18 < kanzure> fine by me 19:19 < fenn> start with the nature archive 19:19 < kanzure> i don't want to bother with parsing everyone's horrible dump 19:19 < fenn> get metadata, figure out how to computationally extract matching metadata from the articles themselves 19:19 < kanzure> what's wrong with me forcing them to go get metadata 19:19 < kanzure> yeah, i can't easily reverse from the nature archive "what the original url was so i can go grab metadata" 19:19 < fenn> well, how do you attach the metadata to the files in the first place? 19:20 < kanzure> bibtex, some simple yaml format, let's make something up 19:20 < kanzure> then that will be the definition, and the service will grow around that definition. 19:20 * fenn reads about bibtex 19:20 < kanzure> it's latex except a citation subset 19:20 < kanzure> i don't know. everything supports bibtex. 19:21 < fenn> "It is possible to use BibTeX outside of a LaTeX-Environment, namely MS Word using the tool Bibshare. " okay but how is this supposed to help for the pdf scenario 19:21 < kanzure> example: http://diyhpl.us/~bryan/sciencedirect/microelectronics.journal.txt 19:21 < fenn> cat you just cat foo.bibtex >> foo.pdf 19:21 < kanzure> well i've always considered a .pdf.tar format that would include that, but whatever 19:22 < kanzure> yes there's metadata portions in the pdf format, and probably some way to attach files.. but i don't care that much 19:22 < kanzure> so my example link 19:22 < kanzure> i had a bibtex file for each volume of each issue of this journal 19:22 < kanzure> and cat'd it together. so think of this as a single issue of a journal with a looong list of articles 19:23 < fenn> .pdf.tar sucks because your pdf indexing software won't read it 19:23 < kanzure> proprietary pdf indexing software is not the solution anyway 19:24 < kanzure> it's a part of the problem. proprietary ocr? thanks mendeley, that doesn't actually help 19:24 < fenn> tracker-search does pdf indexing 19:24 < fenn> time tracker-search transcranial 19:24 < fenn> real 0m0.080s 19:24 < fenn> unfortunately it returns urls instead of paths 19:25 < kanzure> how is this better than just keeping the original metadata from the webpage clean 19:25 < fenn> eh? 19:25 < fenn> it works with just the pdf files 19:25 < kanzure> i don't trust pdf files to have this information, and i don't trust ocr that much 19:25 < fenn> essentially you have to write a scraper either way 19:25 < kanzure> sure 19:25 < fenn> scrape metadata from the webpage, or scrape it from the pdf 19:26 < kanzure> well you have to get the webpage anyway to get to the pdf 19:26 < fenn> i'd rather scrape from the pdf because we already have those 19:26 < kanzure> but you can't do that reliably :/ 19:26 < fenn> you're guaranteed to have the pdf, but the webpage could disappear, change, or never exist in the first place 19:26 < kanzure> lots of pdfs are just images 19:26 < kanzure> well you only get the webpage once, you see 19:26 < kanzure> after you extrat all the data it might as well disappear (who cares) 19:26 < fenn> okay maybe you should just do that first 19:27 < kanzure> ? 19:27 < fenn> get all the journal metadata 19:27 < fenn> there's "only" 80,000 journals 19:27 < kanzure> right, i'm presently doing that for most of elsevier (although it seems they hide some of their metadata to me unless i'm on a university network) 19:27 < fenn> wow really? what do they hide? 19:27 < kanzure> like issues 19:27 < kanzure> one of the journals i was looking at went back to 2003 19:28 < kanzure> but on another computer, i saw that it went back to 1990something 19:28 < kanzure> same site. 19:28 < fenn> is this data considered "copyright"? i mean it's basically a library card catalog, so there shouldn't be any problem with hosting journal metadata out in the open 19:28 < kanzure> exactly 19:28 < kanzure> i'm sure they will complain anyway 19:29 < fenn> but the mere existence would help others doing the same sort of thing 19:29 < fenn> i mean there's no reason everyone should have to scrape metadata from the web 19:29 < kanzure> and then we can coordinate multiple scrapers simultaneously or at least help people to not duplicate work 19:29 -!- _sol_ [Sol@c-174-57-58-11.hsd1.pa.comcast.net] has joined ##hplusroadmap 19:34 < kanzure> so, i guess it's just a matter of bibtex+pdf? 19:34 < fenn> bibtex is just a text format 19:34 < kanzure> sure 19:34 < fenn> how does that fix anything? 19:34 < kanzure> well ultimately what we need from a scraper is the pdf plus metadata 19:35 < kanzure> i guess not. that doesn't fix it. 19:35 < fenn> you want to know if a collection is "complete", no? 19:35 < kanzure> i want each collection to have an index and know what it has 19:35 < kanzure> and then to describe each item. 19:35 < fenn> okay 19:35 < fenn> the index could just be bibtex 19:36 < fenn> one big file with bibtex entries for all pdf files i the collection 19:36 < kanzure> i guess so. does that cover everything? 19:36 < fenn> alternatively, you could have one bibtex file for each pdf file 19:36 < fenn> they amount to the same thing i guess 19:37 < fenn> neither one "sticks" to the pdf file though 19:37 < fenn> it would be nice to somehow append bibtex to the pdf 19:37 < fenn> i dont know enough about modifying pdf's to know if this is easy 19:38 < fenn> pdf ends with %%EOF so you could just literally append bibtex and it shouldnt hurt anything 19:38 < kanzure> pdftk html_tidy.pdf attach_files command_ref.html to_page 24 output html_tidy_book.pdf 19:38 < fenn> metadata doesn't need to be visible in the pdf viewer 19:39 < kanzure> you mean EOF or a literal that says '%%EOF' 19:39 < fenn> literal 19:39 < fenn> tail foo.pdf 19:40 < fenn> bbl maybe 19:41 < kanzure> well that's weird. 19:42 < kanzure> so doesn't citeseer do citation tracking 19:42 < kanzure> or citeulike 19:42 < kanzure> and i guess mendeley has an ok collection by now. 19:42 < kanzure> "Scientific Literature Digital Library incorporating autonomous citation indexing, awareness and tracking, citation context, related document retrieval, similar" 19:42 < kanzure> http://citeseer.ist.psu.edu/index 19:45 < fenn> i think this is not solving the same problem 19:45 < kanzure> no not quite 19:45 < fenn> there's the forwards metadata problem (start with a catalog, link to the articles) 19:45 < fenn> and the backwards metadata problem (start with the articles, link to the catalog) 19:50 < fenn> wow citeseer search sucks balls 19:51 < fenn> "here are some articles that contain keywords that you typed in, in no particular order" 20:02 < kanzure> http://diyhpl.us/~bryan/papers2/bib.tiera.ru/protocols.txt 20:04 < fenn> document clustering is probably more useful than journal indexes anyway 20:05 < fenn> i'm wondering how many of my books are essentially duplicates 20:06 < fenn> what's bib.tiera.ru? 20:06 < kanzure> no clue 20:06 < kanzure> just found it today 20:06 < kanzure> looks a bit more hand-curated than libgenesis 20:08 < kanzure> If you have any stuff to upload (or to donate), write us
20:14 < kanzure> http://f2.tiera.ru//TEXTBOOKS3/ELSEVIER-Referex/1-Chemical Petrochemical and Process Collection/CD3/RICE, R. G. (1994). Applied Mathematics and Modelin 20:14 < kanzure> g for Chemical Engineers/03771_08.pdf 20:14 < kanzure> hmm ELSEVIER-Referex? 20:15 < fenn> wonder what's up with their css and general lack of content http://fennetic.net/irc/facebook_huh.png 20:17 < fenn> hmm maybe i broke it with adblock 20:18 < kanzure> i'm looking at tiera.ru's index 20:18 < kanzure> and i'm not sure, but i think some of these were mine 20:19 < fenn> hehehe 20:19 < kanzure> /other/other3/Chemistry/Chemical engineering/Technology and processing of polymers/ 20:19 < kanzure> nobody makes awful paths like i do! 20:19 < fenn> spaces! 20:19 < kanzure> :/ 20:20 < fenn> hey why not use freenet 20:21 < fenn> for distributing papers 20:21 < kanzure> i don't think distribution is a problem 20:21 < kanzure> and realistically for maximum impact you need http 20:25 < fenn> how big do you think "all" of the journal archives would be if properly OCR'ed? 20:26 < kanzure> "ScienceDirect publishes 250,000 articles a year in 2,000 journals." 20:26 < kanzure> sciencedirect indexes >5000 journals though, so i don't know what's up there 20:26 < fenn> for the sake of analysis, limit "all" to everything currently offered in digital format online 20:27 < fenn> sciencedirect is a brand of elsevier, right? 20:27 < kanzure> yes 20:27 < kanzure> i think i've seen estimates of 7 to 8 million articles per year at the moment 20:27 < fenn> "ScienceDirect is Elsevier's platform for online electronic access to its journals" 20:27 < kanzure> yeah 20:28 < fenn> so maybe they only have metadata for the other 3000 20:28 < fenn> 8 million per year sounds way too high 20:28 < kanzure> they include some things they have "Intellectual Property Rights to index" or something lame 20:28 < kanzure> http://www.quora.com/How-many-academic-papers-are-published-each-year 20:29 < kanzure> "1.486 million peer-reviewed papers published within 2010" 20:29 < kanzure> "They estimate that 1.346 million articles were published in 23,750 journals within 200" 20:29 < kanzure> ok that was originally: 20:29 < kanzure> "They estimate that 1.346 million articles were published in 23.750 journals within 200" 20:29 < kanzure> why would 23.750 make sense 20:29 < kanzure> either say 1,346 million and 23,750 or 1.346 million and 23750.00 dfkadkfadja 20:30 < kanzure> ny times published this graph once that showed the rate of increase in publications for all countries for at least 40 years 20:31 < Stee|> 23.750 is used in some european countries 20:32 < fenn> it's confusing because they're inconsistently using the period for either decimal place or thousands separator 20:33 < fenn> ok so something like 2million per year 20:33 < fenn> this means, at minimum your scraper has to be able to handle 5000+ papers a day 20:33 < kanzure> http://duncan.hull.name/2010/07/15/fifty-million/ 20:34 < fenn> conversely, 15 seconds maximum processing time per paper (assuming no parallelization) 20:35 < kanzure> "One paper per minute is based on 679,858 papers per year in 2009 / 365 days / 24 hours / 60 minutes = 1.29 papers per minute." 20:35 < kanzure> hrm actually this might be an ok method to help approximate the collection completeness 20:36 < kanzure> "768,341 papers have been written so far in 2012. We have 20,000." 20:37 < fenn> maybe we should just focus on getting papers that aren't crap 20:37 < kanzure> well that too. 20:37 < kanzure> it would be nice to start with some commonly-read journals 20:38 < fenn> maybe take your collection and use that as a set of seed points for a web of science crawler 20:38 < kanzure> ? 20:39 < fenn> presumably anything not cited by or citing the papers cited by the papers in your collection (ugh) is not worth reading 20:39 < kanzure> i find strange outliers all the time 20:39 < kanzure> i guess i haven't looked at the citation network at all. 20:40 < fenn> for all it's been abused by academia, citation network is a pretty cool thing 20:40 -!- nuba [~nuba@pauleira.com] has quit [Ping timeout: 260 seconds] 20:41 < fenn> s/for/despite/ 20:41 < kanzure> i've never been sure how researchers remember which paper they read something in 20:41 -!- nuba [~nuba@pauleira.com] has joined ##hplusroadmap 20:41 < kanzure> i can remember big papers where important things happen 20:41 < kanzure> but for the small results, that seems much harder 20:41 < fenn> they don't; it's all made up 20:41 < fenn> you just have to make up a bibliography when writing your paper so you go look around for the originals 20:41 < kanzure> i guess if you go looking for something to cite, it's easier 20:42 -!- jmil [~jmil@c-68-81-252-40.hsd1.pa.comcast.net] has quit [Quit: jmil] 20:42 < kanzure> orr they probably just read a review paper and pick out some crap 20:42 < fenn> also a trick i've seen is people just look at what papers they've downloaded 20:42 < fenn> most academics don't have a very extensive library 20:43 -!- jmil [~jmil@c-68-81-252-40.hsd1.pa.comcast.net] has joined ##hplusroadmap 20:43 < fenn> that's the library's job after all 20:43 < kanzure> the libraries are busy paying too much for all these journals 20:45 < fenn> why is there no "complete" journal metadata index? 20:45 -!- jmil [~jmil@c-68-81-252-40.hsd1.pa.comcast.net] has quit [Client Quit] 20:46 < fenn> i mean why does this fall on the shoulders of a couple disgruntled hackers 20:46 < kanzure> http://diyhpl.us/~bryan/irc/sciencedirect_journals.json 20:46 < fenn> there should be armies of librarians tackling this 20:46 -!- Stee| [~Steel@cpe-67-246-36-165.nycap.res.rr.com] has quit [Ping timeout: 276 seconds] 20:47 < kanzure> it's because "others" got to science first, before the internet got it. 20:47 < kanzure> *got to it. 20:47 < fenn> but it's really not a hard problem to solve 20:47 < kanzure> nope 20:47 < fenn> "scrape all the indices, merge, repeat" 20:48 < kanzure> except they do 20:48 < kanzure> and then they pay for it for some reason 20:48 < kanzure> and call it proprietary 20:48 < kanzure> isi... 20:48 < kanzure> :/ 20:48 < fenn> isi is supposed to be the complete metadata index? 20:48 < kanzure> hrmm i'm not sure. 20:49 -!- Stee| [~Steel@cpe-67-246-36-165.nycap.res.rr.com] has joined ##hplusroadmap 20:49 < kanzure> ". This intelligent research platform provides access to the world's leading citation databases, including powerful cited reference searching, the Analyze Tool, and over 100 years of comprehensive backfile and citation data." 20:49 < fenn> so there are these papers studying h-index etc., what is their data set? 20:49 < kanzure> sometimes isi access i think 20:49 < kanzure> http://wokinfo.com/about/whatitis/ 20:49 < kanzure> it's a hodgepodge of commercial databases i think 20:50 < kanzure> "Calculate an accurate h-index by ensuring the full extent of an author’s past research is taken into account." 20:50 < fenn> ok this is the citation network, which while cool, is a superset of what i'm asking for 20:50 < kanzure> oh right 20:50 < fenn> i just want a list of all articles published 20:53 < kanzure> worldcat? i think that stops at individual "titles" (of volumes/books) 20:53 < Stee|> kanz, any idea how to get access to internal zaibatsu journals? 20:53 < fenn> i should be able to google "list of all journal articles published" and find a webpage that lets me download some file with all the metadata 20:53 < fenn> Stee|: walk into the office wearing a jumpsuit, say you're here to fix the router 20:53 < Stee|> hhaa 20:54 < fenn> maybe a bit harder if you're not japanese 20:54 < fenn> worldcat is exactly the right model 20:54 < fenn> but for some mysterious reason no similar thing exists for journals 20:54 < kanzure> worldcat is the supplier of things like ezproxy and interlibrary loans 20:55 < fenn> "With authorization from OCLC, you can download a subset of the WorldCat database for harvesting by your search engine or other enterprise Web application." 20:55 < kanzure> yep.. commercial 20:55 < fenn> http://www.worldcat.org/partnership/harvestset/worldcat_sample_data.xml 20:55 < kanzure> 'enterprise' 20:55 < fenn> does that mean they sell it or what? 20:56 < kanzure> they might be doing license agreements if you end up selling/distributing it, but free for 'research' 20:56 < kanzure> but i don't know for sure 20:56 < kanzure> what about library of congress? i don't think they track individual articles 20:56 < fenn> "Each party shall be entirely responsible for meeting its own costs incurred with respect to the matters described in this Agreement and neither shall be obligated to make any payment to the other under the terms of this Agreement." 20:57 < kanzure> uhuh 20:57 < fenn> so apparently it's free, if you appear "legit" 20:57 < fenn> "Under no circumstances shall Institution/Company sell, license, publish, display, distribute or otherwise transfer to any third party WorldCat Metadata or holdings information or any copy thereof, in whole or in part, except as expressly permitted" 20:57 < kanzure> hrm it looks like they do track some articles? 20:57 < kanzure> http://www.worldcat.org/search?q=george+whitesides&qt=owc_search#x0%253Aartchap-%2Cx0%253Aartchap%2Bx4%253Adigitalformat 20:57 < kanzure> wow what 20:58 < kanzure> well what is expressly permitted? 20:58 < fenn> it just means you don't have permission to redistribute 20:58 < kanzure> haha how useless 20:58 < kanzure> god our libraries suck 20:59 < fenn> yeah somehow the librarians missed the "information wants to be free" bandwagon 21:02 < kanzure> wasn't google supposed to fix this? 21:02 < kanzure> or was it only supposed to index the shitty information 21:03 < fenn> google "grew up" and got all responsible n shit 21:03 < fenn> so now they can't do anything 21:04 < fenn> "OCLC does claim copyright rights in WorldCat as a compilation. In accordance with US copyright law, those 21:04 < fenn> rights are based on OCLC's substantial intellectual contribution to WorldCat as a whole, including OCLC’s selection, 21:04 < fenn> arrangement, and coordination of the material in WorldCat" 21:04 < kanzure> riight. 21:04 < fenn> fwiw, google is awful about exporting data, despite their "data liberation front" 21:04 < kanzure> well then anyone can claim copyright on collection 21:04 < kanzure> science liberation front sounds like an awesome name 21:04 < fenn> heh 21:05 < fenn> you need to wear a ski mask and have stacks of liberated hard drives behind you 21:05 < kanzure> not a problem 21:06 < kanzure> it can also be like a drug bust: http://addictionrecoveryhope.com/wp-content/uploads/2009/09/Drug-Bust.jpg 21:06 < fenn> if anyone cares, here's some discussion and self-justification by oclc on why they don't allow redistribution http://www.oclc.org/worldcat/recorduse/policy/forum/forum.pdf 21:07 < kanzure> i think oclc was setup by a bunch of university librarians and that's why it has traction 21:07 < kanzure> i'm not really sure why they all agreed to this awful mess 21:07 < kanzure> but it's particularly nice of them to all be using ezproxy (it makes it easier once someone finds an exploit) 21:08 < kanzure> "the practical need to sustain the economic viability and value of WorldCat over the long term" 21:08 < fenn> what about ezproxy? 21:08 < kanzure> all universities use it 21:09 < kanzure> they all run a local instance of it 21:09 < kanzure> so if i was to find a backdoor, i'd have keys to the entire kingdom 21:09 < fenn> is that a mirroring software? 21:09 < kanzure> no 21:09 < kanzure> the best way to explain it is to show you 21:09 < fenn> how do you get the database if it's all just web pages? 21:09 < fenn> you dont need ezproxy to search worldcat 21:09 < fenn> does oclc do something else? 21:10 < kanzure> oclc does a lot of things 21:10 < kanzure> it's your usual clusterfuck of databases and library integrations 21:10 < fenn> here's the thing, there's a unique ISBN for every book 21:10 < kanzure> here's an example of ezproxy 21:11 < kanzure> http://webserver.macu.edu:2048/ 21:11 < kanzure> username: 3952 21:11 -!- Stee| is now known as Steel_ 21:11 < kanzure> password: 3952 21:11 < fenn> where's the data that goes along with that ISBN submitted to when the author publishes the book? 21:11 < kanzure> library of congress somewhere 21:11 < kanzure> http://www.loc.gov/rr/ 21:12 < fenn> so what good does worldcat do then? why can't we just download the ISBN's from LOC? 21:12 < kanzure> because worldcat is also indexing papers 21:12 < kanzure> also, worldcat.org is not the primary purpose of worldcat 21:13 < kanzure> "Some WorldCat libraries make their specialized reference databases available on their Web sites, but only to library members." 21:13 < kanzure> "worldcat libraries" 21:13 < fenn> oh great 21:14 < fenn> ISBN is run by a for-profit company 21:14 < fenn> http://en.wikipedia.org/wiki/R._R._Bowker 21:14 < fenn> now owned by ... (drumroll) 21:14 < fenn> Elsevier! 21:14 < kanzure> here's what oclc is: http://www.oclc.org/us/en/services/a-to-z.htm 21:15 < fenn> those fuckers 21:15 < kanzure> xisbn is oclc apparently 21:15 < kanzure> hah 21:15 < kanzure> look at how they list worldcat 21:15 < kanzure> "Global network of library content and services that lets your institution be more connected, open and productive" 21:15 < kanzure> it's about convincing libraries to sign up with them 21:16 < kanzure> even my high school had some weird worldcat integration (it was pretty broken) 21:17 < kanzure> did you try that ezproxy login? 21:17 < fenn> this is interesting http://isbndb.com/ 21:18 < fenn> yes, looks poorly configured 21:18 < fenn> reminds me of internet circa 1995 21:18 < kanzure> i'm pretty sure libraries were networked together pre-web 21:18 < kanzure> maybe that's why proprietary solutions are so dominant 21:19 < fenn> i dont really know what i'm looking at here 21:20 < fenn> it's a list of services they've purchased access to? 21:20 < kanzure> yes 21:20 < kanzure> but if you click, you have access 21:20 < kanzure> since you're logged in 21:20 < kanzure> normally these services authenticate you by ip address 21:20 < kanzure> ezproxy is inside the college's network 21:21 -!- yashgaroth [~f@cpe-24-94-5-223.san.res.rr.com] has quit [Ping timeout: 260 seconds] 21:21 < fenn> i assume at a more science-oriented university the list would be more useful? 21:21 < kanzure> heh yes this is a bad example 21:22 < fenn> i just realized what "WorldBook" is 21:22 < kanzure> microsoft? 21:22 < fenn> it's the old second-rate paper encyclopedia, for people who couldn't afford britannica 21:23 < kanzure> i thought that was ms encarta 21:23 < fenn> oo oxford english dictionary 21:24 < fenn> so ezproxy just forwards stuff from the uni's IP to the external internet 21:24 < kanzure> yes 21:24 < kanzure> sometimes these services have usernames/passwords instead of ip authentication 21:25 < kanzure> and ezproxy handles that configuration/setup too (apparently) 21:26 < kanzure> in the uk i think they have some federal system for paper access called athens? i don't know much about it 21:26 < kanzure> "Athens is an access management system which controls access to many of the Library's electronic information sources. When you login to an Athens protected resource it checks to see if you are a member of an institution that has paid to use that resource, and if your username and password are correct it lets you through." 21:27 < fenn> hmm want to scrap OED? they only have 275k entries http://www.oed.com.webserver.macu.edu:2048/viewdictionaryentry/Entry/274870 21:27 < kanzure> http://en.wikipedia.org/wiki/Athens_access_and_identity_management 21:27 < kanzure> i'm sure someone has a pdf of oed 21:27 < kanzure> i guess a pdf is less useful 21:28 < kanzure> "The Athens service is a trust federation where Identity Providers, Service Providers and Athens operate under common rules and licenses. Trust is enforced by the use of public-key cryptography and other security mechanisms." 21:28 < kanzure> "Athens is used extensively within UK Higher and Further Education institutions, the UK National Health Service, and in more than 90 countries worldwide. It has been adopted by over 2,000 organisations, and over 300 online resources since it was first launched in 1996. Over 4.5 million accounts are now registered with Athens." 21:28 < kanzure> "Conceived in 1996 at the University of Bath," we should get adrian bowyer to fix that 21:28 < fenn> this is the last entry added http://www.oed.com.webserver.macu.edu:2048/viewdictionaryentry/Entry/277114 21:28 < kanzure> it's by id 21:28 < kanzure> ? 21:29 < fenn> yep just fiddle with the number 21:29 < fenn> "once you're past the perimeter, there's no security! have a nice day" 21:30 < fenn> i do just fine with the 1907 dictionary, but someone might find it useful 21:31 < fenn> well i guess i should go socialize 21:31 * fenn sighs 21:41 -!- SDr [~SDr@unaffiliated/sdr] has joined ##hplusroadmap 21:47 -!- yashgaroth [~f@cpe-24-94-5-223.san.res.rr.com] has joined ##hplusroadmap 21:49 < kanzure> hi yashgaroth 21:49 < yashgaroth> hey hey 21:50 < _sol_> DO you think DIY orgo chem on a legit side for open source experimentation is possible or that it'd break the bank to make some DIY fume hood and accessory safety in the lab to make sure ya don't have vapors floating wiht a spark setting things soff? 21:50 < _sol_> er off 21:51 < _sol_> I was thinking about that th eother day reading some DIY sites, and wondering what legit open source means for DIY chemistry since ppl are trying to spring up microbio labs and such now 21:53 < yashgaroth> depends where you're trying to install it, and how legal you want to be about EPA regulations 21:53 < _sol_> Of coures, if ya own a glass beaker these days some countries enforcement agencies put books and a single glass beaker together with all the regs and assume the worst... 21:56 < kanzure> what does any of that have to do with breaking the bank 21:57 < _sol_> I guess I'm wondering if ya could try to start a DIY fume hood project among other things or is it all regulated as to what is needed for chem lab? 21:58 < _sol_> I mean the cost for breaking the bank... 21:58 < _sol_> the cost may be to much to make something safe for small DIY stuff 21:58 < _sol_> and try to be pretty safe although not sure and in whose eyes 21:58 < yashgaroth> how interesting are the chemicals you plan to use 21:58 < _sol_> don't know yet 21:59 < _sol_> I'm just wondering if there are projects out there already... 21:59 < _sol_> but I think some solvents in basic orgo chem experiments are still pretty volatile with sparks if I recall 22:00 < _sol_> if ya are doing a distillation process to seperate a heavier weight molecule from lighter weight via heating and using a water cooling over the glass to cool the vapors... 22:00 < _sol_> I'm just looking at how-tos 22:01 < kanzure> yes you can make a fume hood if you want to? 22:01 < _sol_> but if ya don't insulate the fan right? couldn't an electrical spark set stuff off? 22:01 < _sol_> I guess I'm just overthinking... 22:03 < _sol_> I have a chemist friend so I could ask him, but I'm wondering how big chemistry is in the DIY... 22:04 < _sol_> DIY community... etc which I tink this room sorta is , but its more open electronics and software maybe right now 22:04 < _sol_> biohacking somewhat I guess 22:04 < kanzure> you're welcome to bring your chemistry friends in here 22:04 < _sol_> I'll see if he is around later.. 22:04 < kanzure> drazak_: did you ever finish your distillation setup? 22:05 -!- augur [~augur@208.58.5.87] has joined ##hplusroadmap 22:10 < fenn> a fume hood is dead simple: a box where you do your work, a trapezoidal reducing flange, a fan, and a chimney 22:10 < fenn> unless you're venting extremely toxic fumes (in which case i question your methodology) dilution with lots of air will render it harmless 22:12 < fenn> as for explosion prevention, make sure you don't exceed the minimum concentration needed to explode stuffs 22:12 < fenn> for propane this is as little as 5% 22:13 < fenn> on the other hand, 1 mol of gas is only 22 liters so if you're going to evaporate a mol of whatever you need to add at least 1 m^3 of air to it to render it explosion-proof 22:14 < fenn> a bigger concern is crap building up on the chimney which really should be done with regular close-up visual inspection 22:14 -!- SDr [~SDr@unaffiliated/sdr] has quit [] 22:14 < kanzure> oh nice cysteine has Oligonucleotide Synthesis - Methods and Applications [Methods in Molec Bio 288] - P. Herdewijn (Humana, 2005) WW.pdf 22:15 < fenn> cysteine has papers on it already? 22:15 < fenn> or is that a book 22:15 < kanzure> it's a book 22:15 < kanzure> check /torrents/text/protocols/ 22:15 < kanzure> and /torrents/text/books/textbooks/Biology_And_Medicine/ 22:16 < kanzure> text/books/textbooks/.. damn the world sucks 22:16 < fenn> hey, four hour work week, was looking for that 22:20 < fenn> gah 96MB 22:23 < kanzure> fenn: so i've been using phantomjs a lot lately 22:23 < kanzure> and i keep looking at http://www.gnu.org/software/pythonwebkit/ 22:23 < kanzure> which is very ranty.. but accessing the dom from python seems much better than from javascript 22:24 < kanzure> if you'll notice, it's a giant rant by luke kenneth casson leighton 22:24 < kanzure> who you might remember from openscad 22:27 < kanzure> luke keeps saying that pythonwebkit is pyjamas 22:29 < fenn> tldr what? 22:30 < kanzure> web scraping with webkit bindings 22:30 < fenn> don't you need to run js to access links that get created at runtime? 22:30 < kanzure> yes that's what webkit does 22:30 < fenn> i thought that was the whole point of phantomjs 22:30 -!- Mokbortolan_1 [~Nate@c-71-59-241-82.hsd1.or.comcast.net] has quit [Read error: Connection reset by peer] 22:30 < kanzure> correct. pythonwebkit seems to be phantomjs except piloted by python instead of js 22:30 < kanzure> except not marketed like phantomjs 22:30 < fenn> ok 22:31 < fenn> well, that's nice 22:31 < kanzure> theoretically this should be more pleasant 22:31 < fenn> i had intended to learn js anyway 22:32 < fenn> lots of stuff can be scraped without js though 22:32 < fenn> just looking at page structure with yer eyeballs 22:32 < kanzure> i'm tired of beautifulsoup, lxml, mechanize and nokogiri 22:32 < kanzure> sometimes the html is poorly formatted and sometimes there's data written in the js headers 22:32 < fenn> i suppose it depends on how structured the data is you're trying to scrape in the first place 22:33 < kanzure> and then these parsers break and crap.. if any parser isn't gonnab reak, it's going to be a web browser 22:33 < fenn> really the html parser doesn't work? i thought the whole point of BS was that it didnt break on bad html 22:33 < kanzure> no i thought that was lxml 22:33 -!- Jaakko96 [~Jaakko@host86-131-178-213.range86-131.btcentralplus.com] has joined ##hplusroadmap 22:34 < fenn> The BeautifulSoup class turns arbitrarily bad HTML into a tree-like nested 22:34 < fenn> etc 22:34 < kanzure> hmmm okay 22:35 -!- Mokbortolan_ [~Nate@c-71-59-241-82.hsd1.or.comcast.net] has joined ##hplusroadmap 22:35 < kanzure> fwiw beautifulsoup is what i use anyway 22:36 < fenn> ug flattr keeps 10% for itself, why do people put up with this 22:37 < kanzure> because it has a slim chance of being better than paypal 22:38 < fenn> bah 22:38 < fenn> wepay is actually better than paypal 22:38 < fenn> for now at least 22:39 < fenn> problem is paypal tries to be all "don't worry, we'll refund you someone else's money if there's any problems" 22:40 < kanzure> git clone http://git.savannah.gnu.org/cgit/pythonwebkit.git 22:40 * kanzure gulps 22:45 < drazak_> kanzure: at home? nah, never ended up buying anything 22:45 < drazak_> kanzure: too expensive 22:47 < kanzure> too expensive! bah 22:53 < fenn> distillation is easy 22:54 < fenn> i'm surprised you can't just buy a still from walmart 22:54 < fenn> such things exist but only for water 22:58 < kanzure> wow what in the name of holy fuck 22:58 < kanzure> http://www.scholartime.com/index.php/journal-hosting 22:59 < kanzure> hosted ojs instances (openjournalsystem) 22:59 < kanzure> $600/year? 22:59 < kanzure> that should be more like $10 or $20/year 22:59 < kanzure> i guess that's $50/year.. but still 23:00 < kanzure> oops $50/month 23:01 < fenn> it's overpriced hosting but it's somewhat specialized knowledge 23:01 < kanzure> there's more money to be made in vertical integration with the research 23:01 < fenn> eh? 23:01 < kanzure> like "yo dawg, we noticed you're doing a bill of materials for reagents in your project.. we can hook you up aww yeah" 23:02 < fenn> "yo dawg i herd u liek bill o materials so i put a bill o materials in yo bill o materials yo" 23:02 < kanzure> well each research paper is the result of some $200k grant 23:03 < kanzure> reagents maybe costing some % of that.. which i guess is what um, that lab management webapp thing was trying to tap 23:03 < Mokbortolan_> Yo dawg, I heard you liked BoMs, so I put a bomb in your BoM so you can bomb while you BoM. 23:04 < fenn> i guess the idea there is they've already forked over cash so they don't feel so bad forking over more cash to the same entitty? 23:04 < kanzure> also! there's all sorts of weird stateful information during research that could be served by a platform 23:04 < kanzure> instead of keeping random spreadsheets on random computers about which petri dish is currently in which state 23:05 < fenn> dude this is just a web host with some custom software installed 23:05 < kanzure> yeah i know, there's no reason for this to cost so much 23:05 < fenn> they're not doing anything in the research phase at all 23:05 < fenn> it's merely for preserving the results for posterity 23:05 < kanzure> well they seem convinced they are a part of research 23:05 < kanzure> not this site in particular though 23:09 < kanzure> ok whatever. all these fees are stupid. 23:09 < Steel_> kanzure: worth starting a business for it? 23:09 < Steel_> undercut 'em? 23:09 < kanzure> no. there's only 20000 journals or something 23:10 < fenn> ah but think how many MORE there could be! 23:10 < kanzure> 20000 * $5/year = haha.. yeah 23:10 < kanzure> journals are a dumb structure anyway 23:10 < fenn> a journal for every lab! 23:10 < kanzure> isn't it just supposed to be an aggregator 23:10 < fenn> a journal for every paper! 23:10 < kanzure> yes! 23:10 < kanzure> wait. do you mean tags? 23:10 < Steel_> kanzure: If you have thoughts on a better one, I'd certainly love to see them written up somewhere so I can incorporate those ideas 23:10 < Steel_> tags are on my list 23:10 < fenn> how about scientists publish their own fucking papers 23:11 < kanzure> i don't trust people to maintain active web servers 23:11 < fenn> neither do i 23:11 < fenn> but at least someone would be able to aggregate them 23:11 < fenn> unlike now where we're all "oh noes the paywall is falling" 23:11 < kanzure> did those jerks update the arxiv torrents or is there still an anti-get-all-our-data thing going on there? 23:11 < kanzure> err right now if the paywalls fall there's nothing saved from inside. 23:12 < fenn> okay why is pythonwebkit.git > 1GB? 23:12 < kanzure> i have no fucking clue 23:12 < kanzure> it's still cloning 23:13 < kanzure> i have a clone on gnusha in /home/bryan/local/pythonwebkit/ if you want it. 23:13 < kanzure> i'm currently checking out python_codegen (the branch) 23:13 < fenn> can you make a copy of this without the bloat? 23:13 < kanzure> um i don't know what the issue is yet 23:13 < kanzure> i think it's a complete copy of webkit 23:13 < kanzure> i'll check what the working directory size is. 23:13 < fenn> still, shouldn't be that big 23:14 < kanzure> 2.9 GB? 23:14 -!- lkcl [~lkcl@host86-131-171-208.range86-131.btcentralplus.com] has joined ##hplusroadmap 23:14 < fenn> seems to be mostly ./LayoutTests and ./WebCore 23:15 < lkcl> morning folks 23:15 < kanzure> can we delete those 23:15 < fenn> yes 23:15 < fenn> actually wait 23:15 < fenn> dunno about webcore 23:15 < lkcl> i'm told that there are people trying to compile pythonwebkit around here 23:15 -!- uniqanomaly_ [~ua@dynamic-78-8-80-128.ssp.dialog.net.pl] has joined ##hplusroadmap 23:15 < fenn> anyway it's in .git as well 23:15 < lkcl> i'm the lead developer 23:15 < fenn> hi lkcl 23:15 < fenn> why is your repo 2.9GB? 23:15 < lkcl> so... what do you need? 23:16 < lkcl> because that's what the size of webkit git is 23:16 < kanzure> git-annex? 23:16 < lkcl> git://git.webkit.org/WebKit.git 23:16 < fenn> i'm interested in pythonwebkit primarily for scraping 23:16 < lkcl> ahhh 23:16 < lkcl> headless? 23:16 < kanzure> yeah i might have misunderstood you 23:16 < kanzure> headless would be great 23:17 < kanzure> maybe pythonwebkit isn't actually required? 23:17 < lkcl> ok, there's a couple of ways to do that 23:17 < lkcl> actually three. 23:17 < lkcl> what level of HTML compatibility do you need? 23:17 < kanzure> what are my options 23:17 < lkcl> just some notes from another conversation i'm going to cut/paste ok? 23:17 < lkcl> http://git.savannah.gnu.org/cgit/pythonwebkit.git/tree/pywebkitgtk/gtk/webkitgtkmodule.c?h=python_codegen 23:17 < lkcl> comment out line 82. 23:18 < lkcl> that will give you "headless" mode in pythonwebkit. ok, _should_ do :) 23:18 < fenn> heh 23:18 < lkcl> 1) KDE's KHTMLPart (HTML DOM TR2 compatible) 23:18 < lkcl> 2) pythonwebkit gtk mode, hacked to remove that line 82 23:19 < lkcl> 3) pythonwebkit "DirectFB" mode, hacked to remove the equivalent line - you want the python_codegen-directfb-2011-10-18 branch 23:19 -!- uniqanomaly [~ua@dynamic-78-8-80-186.ssp.dialog.net.pl] has quit [Ping timeout: 272 seconds] 23:19 < lkcl> 4) python-hulahop with xulrunner 9.0 - source code is here: http://lkcl.net/hulahop 23:20 < lkcl> then get the tutorial i wrote, here: 23:20 < lkcl> http://pyxpcomext.mozdev.org/no_wrap/tutorials/hulahop/xpcom-hulahop.html 23:20 < kanzure> what i'd like is the same level of html compatibility as phantomjs, which seems to just be headless vanilla webkit 23:20 < lkcl> and hack that to simply remove the pygtk2 equivalent of the window stuff 23:21 < lkcl> well it depends on whether you want Firefox headless HTML5 compatibility or Safari/Webkit/Android headless HTML5 compatibility 23:21 < lkcl> python-hulahop will get you Firefox 23:21 < lkcl> pythonwebkit will get you Webkit/Android/Safari 23:21 < lkcl> and KHTMLPart will get you.... mmmm.... compatibility with the internet circa 1998 :) 23:22 < fenn> i think we are misunderstanding something 23:22 < lkcl> you have to compile KDE with c++ runtime type checking enabled 23:22 < fenn> the idea is to run javascript and get text data out of the page, not to render anything 23:22 < lkcl> well, then you'll need to create a "port" of webkit which does no rendering. 23:22 < lkcl> 1sec... 23:23 < lkcl> let me look up what phantomjs is.... 23:23 < lkcl> oooh hoo hoo! 23:23 < lkcl> verrry coool. 23:23 < kanzure> heh. except i want this in fucking python 23:23 < lkcl> so they created a port that... oh shit, you want _what_??? :) 23:24 < lkcl> ooo hoo hoo, you're gonna have a lot of fun then. 23:24 < lkcl> ok, you have a couple of options 23:25 < lkcl> 1) work out the patches that i did to add python bindings and re-apply them to phantomjs 23:25 < lkcl> the pythonwebkit stuff *is* entirely with the exception of about .... 100 lines of code *entirely* screen-independent 23:26 < lkcl> 2) work out the phantomjs patches and reapply *those* to pythonwebkit 23:26 < kanzure> i don't think phantomjs patches webkit 23:26 < lkcl> 3) freak out at option 1 and 2, and give up and just run pythonwebkit *without* .... it doesn't?? 23:26 < kanzure> it just includes some stuff? 23:26 < lkcl> 1sec.... 23:27 < kanzure> https://github.com/ariya/phantomjs/blob/master/src/webpage.cpp 23:27 < lkcl> is it based on Webkit2? 23:27 < kanzure> don't know 23:27 < kanzure> it looks like it's not qtwebkit 2.2 23:28 < lkcl> what der f**??? 23:28 < kanzure> yeah it looks like it's webkit1 23:28 < kanzure> http://groups.google.com/group/phantomjs/browse_thread/thread/e8ffec54c440b0a1 23:28 < kanzure> "However, the fact that WebKit1 API is considered "obsolete" means that at some point, we will not be able to use the latest and greatest WebKit features anymore." 23:28 < lkcl> where's the makefile showing the #includes 23:28 < kanzure> they use qmake 23:29 < lkcl> ok. 23:29 < lkcl> right. 23:29 < kanzure> to be fair i don't know how qmake works :) 23:29 < lkcl> if they're using QtWebKit then all they are doing is exactly as i described above... except not calling the Qt version of "show window" 23:29 < lkcl> so they *are* still "rendering".... just not rendering *on-screen*. 23:30 < kanzure> correct 23:30 < lkcl> or, more specifically, the code _to_ render is there, but it's just not called. 23:30 < kanzure> so they didn't patch webkit? 23:30 < lkcl> the above commenting-out that i described is *exactly* the same trick. line 82 removes the gtk "show all windows" 23:30 < lkcl> that's correct - they didn't. 23:30 < lkcl> all they're doing is firing up a qtwebkit instance and then not showing it on-screen. 23:31 < lkcl> the same trick is pulled in one of the webkitgtk test applications, i forget its name. 23:32 < kanzure> in your python bindings is the dom-touching-python (in a WebPage i think it's called) sandboxed from the other code? 23:32 < lkcl> anyway - in that tutorial: 23:32 < lkcl> http://pyxpcomext.mozdev.org/no_wrap/tutorials/hulahop/xpcom-hulahop.html 23:32 < lkcl> just remove the "gtk.show" 23:32 < lkcl> and you'll achieve exactly the same thing 23:32 < lkcl> i have no idea what you mean by "sandboxed". 23:33 < kanzure> in phantomjs you create a page object and can call page.evaluate(anonymous js function) 23:33 < kanzure> but the contents of the function can't access anything outside of the page's 'context' 23:34 < kanzure> page.open('http://www.google.com/', function(status) { console.log(document.location); }); 23:34 < lkcl> yeah - ok, i didn't add javascript evaluation functions because webkitgtk doesn't have a means to convert the return results into meaningful information 23:34 < kanzure> sure 23:34 < kanzure> but you did seem to have python examples of accessing the dom 23:34 < lkcl> the webkitqt team did translation of results into qt object types 23:34 < lkcl> yes 23:34 < lkcl> it's done *entirely* through python. 23:34 < kanzure> right 23:34 < lkcl> there is absolutely *no* javascript involved, *whatsoever*. 23:34 < fenn> but js in the page can change the DOM in important ways 23:34 < lkcl> yes it can. 23:35 < kanzure> erm, my point is, the javascript is "sandboxed" in phantomjs.. like you're not ever touching the DOM from your main js 23:35 < lkcl> and python can change the DOM in exactly the same "important" ways... in a declarative fashion [from quotes outside quotes] 23:35 < kanzure> is it the same way in pythonwebkit? 23:35 < lkcl> no, because you cannot activate the running of any javascript *at all* from webkitgtk, period. 23:36 < lkcl> ok that's not quite true, but.... 23:36 < kanzure> i'm not talking about javascript :P hrmm 23:36 < kanzure> let's look at http://pyxpcomext.mozdev.org/no_wrap/tutorials/hulahop/xpcom-hulahop.html 23:36 < kanzure> under _loaded 23:36 < lkcl> yep sure 23:36 < kanzure> is that code normal python? 23:36 < kanzure> can it access globals or whatever 23:36 < lkcl> yes it is entirely normal python. 23:36 < lkcl> yes it can 23:36 < kanzure> ok. in phantomjs the answer is no ;) 23:37 < lkcl> because it's pure python. 23:37 < lkcl> right - ok, i see what you mean 23:37 < kanzure> alright cool 23:37 < kanzure> that's great 23:37 -!- Jaakko96 [~Jaakko@host86-131-178-213.range86-131.btcentralplus.com] has quit [Quit: Nettalk6 - www.ntalk.de] 23:37 < lkcl> it's a one-way street. 23:37 < kanzure> what? in pythonwebkit it is? 23:38 < lkcl> you can do tricks such as add a script node to the DOM however :) 23:39 < kanzure> phantomjs is a little weird because you can't actually access the DOM except from inside the page's javascript context 23:40 < kanzure> so you can only hope-and-pray by passing giant hashes/json back and forth between page.evaluate() calls 23:40 < kanzure> whatever.. 23:40 < lkcl> yes. it's a bitch. 23:41 < lkcl> someone actually did a port of pyjamas-desktop using a similar trick, to webkitqt4 23:41 < lkcl> it got a looong way before being declared a complete failure 23:41 < lkcl> execution of javascript code-snippets for *everything*. 23:41 < lkcl> truly truly dreadful :) 23:41 < kanzure> so! can you convince me to use pythonwebkit-gtk over hulahop? 23:41 < lkcl> nope - that's up to you. 23:41 < kanzure> bah 23:42 < kanzure> some code evangelist you are. 23:42 < lkcl> it depends on what you want / need 23:42 < kanzure> hrm 23:42 < lkcl> it makes no odds to me :) 23:42 < kanzure> well, depending on gecko seems a little weird 23:42 < lkcl> pyjamas-desktop works on both... *and* on MSHTML under w32! 23:42 < lkcl> so if you were a windows fiend you'd even be able to do the same trick there! 23:42 < lkcl> the code you'd be looking for is pyjd/mshtml.py 23:43 < lkcl> and, once again, you just don't show the w32 GUI window 23:43 * lkcl shrugs 23:43 < kanzure> hrmm i'm going to try out hulahop then. 23:43 < kanzure> since it probably doesn't have a 2.9 GB git repo 23:43 < lkcl> bottom line is: you can actually do *all three* major browser engines if you really wanted to 23:43 < lkcl> ha ha 23:43 < lkcl> you got debian? 23:43 < kanzure> yes 23:44 < lkcl> ok, don't use debian/unstable, use debian/testing 23:44 < lkcl> and grab xulrunner-9.0-dev 23:44 < kanzure> i think i'm on wheezy :/ 23:44 < kanzure> i'll take a look 23:44 < lkcl> do "apt-get build-dep python-hulahop" 23:44 < lkcl> etc. etc. 23:44 < lkcl> but then grab the source code from here: 23:44 < lkcl> http://lkcl.net/hulahop 23:44 < lkcl> don't for god's sake use xulrunner 10 23:45 < lkcl> https://bugzilla.mozilla.org/show_bug.cgi?id=728500 23:45 < kanzure> wait you also wrote hulahop? 23:45 < lkcl> https://bugzilla.mozilla.org/show_bug.cgi?id=728645 23:45 < lkcl> http://bugs.debian.org/cgi-bin/bugreport.cgi?bug=660178 23:45 < lkcl> hell no. 23:45 < kanzure> ok just a mirror :3 23:45 < lkcl> no it was a quick-hacked fix to get it to work 23:46 < lkcl> the olpc-sugar team gave up on hulahop 6 months ago... oh dearie me are they in for a shock 23:46 < kanzure> these are not the words that indicate to me that any of this is stable 23:46 < kanzure> heh 23:46 < kanzure> okay let me see if i can get a backport of xulrunner-9.0-dev 23:47 < lkcl> i recommend you just add debian/testing and don't worry about it. use apt-pin priorities 23:47 < kanzure> jrayhawk: how do i use that 23:47 < lkcl> or just use "apt-get -t testing install xulrunner-9.0-dev" etc. etc. 23:47 < lkcl> which will be rather long-winded 23:48 < kanzure> E: Unable to locate package xulrunner-9.0-dev 23:48 < lkcl> backporting of xulrunner will take several hours 23:48 < lkcl> 1sec... 23:48 < kanzure> well i guess i should update my sources 23:48 < lkcl> ii xulrunner-9.0 9.0.1-1 XUL + XPCOM application runner 23:48 < lkcl> ftp://ftp.uk.debian.org/debian/pool/main/p/pyxpcom/ 23:49 < lkcl> ftp://ftp.uk.debian.org/debian/pool/main/i/iceweasel/ 23:49 < kanzure> yeah there's still no xulrunner-9.0-dev package being found? 23:49 < lkcl> deb http://ftp.uk.debian.org/debian/ testing main contrib non-free 23:49 < lkcl> deb-src http://ftp.uk.debian.org/debian/ testing main contrib non-free 23:49 < lkcl> ftp://ftp.uk.debian.org/debian/pool/main/i/iceweasel/xulrunner-dev_9.0.1-1_amd64.deb 23:49 < kanzure> ok maybe it's in testing contrib 23:49 < lkcl> ftp://ftp.uk.debian.org/debian/pool/main/i/iceweasel/xulrunner-9.0_9.0.1-1_amd64.deb 23:50 < lkcl> ftp://ftp.uk.debian.org/debian/pool/main/i/iceweasel/libmozjs-dev_9.0.1-1_amd64.deb 23:50 < kanzure> i only see xulrunner-9.0-dbg and xulrunner-9.0 23:50 < lkcl> lkcl@teenymac:~/src/python-tv/WebKit$ apt-cache search xulrunner-9.0 23:50 < lkcl> xulrunner-9.0 - XUL + XPCOM application runner 23:51 < lkcl> hmmm.... i also have this: deb http://ftp.uk.debian.org/debian/ experimental non-free 23:52 < lkcl> but that's experimental *non-free*. hmm... 23:52 < lkcl> yep - don't know. up to you to sort out :) 23:52 < kanzure> pyxpcom is supposed to be in testing? 23:52 < lkcl> pyxpcom has been around for a while, but it was formerly part of xulrunner's source package 23:52 < fenn> maybe i'm a bit slow, but http://packages.debian.org/wheezy/xulrunner-dev 23:52 < lkcl> it's now independent 23:53 < rdb> Does anyone know how much the magnetic stirrers in ordinary lab heaters affect magnetic implants? 23:53 < lkcl> no, you _definitely_ don't want the older version 23:53 < lkcl> ok there's another way: if you can get debian/lenny such that you end up with xulrunner 1.9.1 23:53 < lkcl> or debian/blahblah with xulrunner 1.9.1 23:53 < fenn> rdb: quite a lot i'd imagine 23:53 < lkcl> you will *not* need to do any kind of source code compiling. 23:53 < rdb> It sounds quite painful. 23:54 < fenn> yep 23:54 < fenn> after a while tissue grows around the magnet and immobilizes it a bit 23:54 < rdb> Does it hurt when the magnet gets a tug and flips around? 23:54 < lkcl> the last "stable" version of python-hulahop which installs out-of-the-box was 18+ months ago, and it used xulrunner-1.9 23:54 < kanzure> lkcl: xulrunner 1.9.1 sounds a bit old? 23:54 < Steel_> rdb: the magnet shouldn't 23:55 < kanzure> hrm 23:55 < Steel_> one of the things I'm planning on running later this year hopefully are some FE simulations of magnet implants in flesh 23:55 < lkcl> kanzure: it does the job. i'm still using firefox 3.5 and that uses xulrunner 1.9.2 23:56 < lkcl> i get absolutely no problems with it, other than f****g stupid google advertising f****g chrome at me nyah nyah youuu're usiiing an ooold version of firefox that weeeee can't be bothered to suppport 23:56 < kanzure> the python-hulahop package is also grabbing xulrunner-9.0 23:56 < lkcl> whine, whine 23:56 < lkcl> yep there you go. 23:56 < kanzure> is that bad 23:56 < lkcl> no you need that - that's the runtime. 23:56 < kanzure> h it's python-xpcom 23:56 < kanzure> *ah it's 23:56 < lkcl> but if you want to recompile for yourself you _will_ need xulrunner-9.0-dev 23:57 < lkcl> obviously. 23:57 < lkcl> that you can get with "apt-get build-dep python-hulahop". 23:57 < kanzure> what do i need besides xulrunner, python-hulahop, python-xpcom? 23:57 < lkcl> once you've done that, then grab the source from the url i posted 23:57 < lkcl> nothing else. 23:57 < lkcl> oh... the source code from that tutorial, obviously. 23:58 < lkcl> then you do dpkg-buildpackage -rfakeroot -nc 23:58 < lkcl> cd into the hulahop directory obviously 23:58 < kanzure> you mean this? http://lkcl.net/hulahop/sugar-hulahop-0.8.1.success.tgz 23:58 < lkcl> first 23:58 < lkcl> yep that's it 23:58 < lkcl> so cd sugar-hulahop-0.8.1 23:58 < lkcl> then do 23:58 < lkcl> dpkg-buildpackage -rfakeroot -nc 23:59 < lkcl> then install the resultant .deb which will be in the directory *below* 23:59 < lkcl> and you're done. 23:59 < kanzure> dpkg-checkbuilddeps: Unmet build dependencies: cdbs (>= 0.4.90~) python-all-dev dh-buildinfo xulrunner-dev (>= 1.9~rc2) python-gtk2-dev 23:59 < lkcl> you're on your way 23:59 < kanzure> oh xulrunner-dev exists 23:59 < kanzure> i see 23:59 < lkcl> i _did_ say do "apt-get build-dep python-hulahop" :) --- Log closed Sun Feb 19 00:00:12 2012