Who am I online & what have I done? Contact information; sites I use; computers and software tools; things I’ve worked on; psychological profiles
This page is about me; for information about Gwern.net, see About This Website.
A transition from an author’s book to his conversation, is too often like an entrance into a large city, after a distant prospect. Remotely, we see nothing but spires of temples and turrets of palaces, and imagine it the residence of splendour, grandeur and magnificence; but when we have passed the gates, we find it perplexed with narrow passages, disgraced with despicable cottages, embarrassed with obstructions, and clouded with smoke.
Samuel Johnson; The Rambler, No. 14 (1750-05-05)1
Behind a remarkable scholar one often finds a mediocre man, and behind a mediocre artist, often, a remarkable man.
The reader lives faster than life, the writer lives slower.
James Richardson, “Even More Aphorisms and Ten-Second Essays from Vectors 3.0”
I am a freelance American writer & researcher. (To make ends meet, I have a Patreon, benefit from Bitcoin appreciation thanks to some old coins, and live frugally.) I have worked for, published in, or consulted for: Wired (2015), MIRI / SIAI2 (2012–2013), CFAR (2012), GiveWell (2017), the FBI (2016), Cool Tools (2013), Quantimodo (2013), New World Encyclopedia (2006), Bitcoin Weekly (2011), Mobify (2013–2014), Bellroy (2013–2014), Dominic Frisby (2014), and private clients (2009-); everything on Gwern.net should be considered my own viewpoint or writing unless otherwise specified by a representative or publication. I am currently not accepting new commissions.
‘I don’t speak’, Bijaz said. ‘I operate a machine called language. It creaks and groans, but is mine own.’
Social news, discussion:
Book reviews (defunct: Goodreads)
Anime reviews (MyAnimeList ratings: anime (XML) / manga (XML))
I have a theme song (MP3), courtesy of David Gerard5
I have no connection to the French singer or with
gwern.com, any locations in Wales, the gwern on MySpace, or either account on
Pivory.com (which are connected to an attempted extortion of me).
I have been active on the English Wikipedia and related projects since January 2004. Cumulatively6, I have over 90,000 edits and have written or worked on hundreds of articles; during my time as an English administrator, I performed thousands of administrative actions; I am an admin on the Haskell wiki, handling routine spam & vandalism:
I also ran a custom Google search tool at “Wikipedia Reliable Sources for anime & manga”; this is a custom Google search with >4542 websites on its black and whitelists. (The source/lists are publicly available.) It returns much more useful7 results for topics in popular culture, and as the name suggests, anime & manga in particular.
I’m sometimes asked about my tech “stack”, in the vein of “Uses This” or The Paris Review’s Writer At Work. I use FLOSS software with a text/CLI emphasis on a custom workstation designed for deep learning & reinforcement learning work, and an ergonomic home office with portrait-orientation monitor, Aeron chair, & trackball.
I run Ubuntu Linux with a tiling window manager & CLI-centric habits. (I prefer Debian but the support of NVIDIA drivers has been better with Ubuntu, so as long as I need GPU acceleration, I will be using Ubuntu). I began using tiling window managers with ratpoison and helped drive the initial development of StumpWM and then xmonad (my config), which I still use in conjunction with MATE, a fork of the last good GNOME desktop environment version before the crazy GNOME 3 ruined everything.
I spend most of my time in Emacs editing Markdown (my config), Firefox (extensions: Evernote plugin, HTTPS Everywhere, NoScript, uBlock origin, LastPass, RECAP), or urxvt/Bash/screen. Most of my programming of R/Haskell/Python is done in a REPL+Emacs. (Friends don’t let friends use heroin or org-mode—are you ever really going to make back the time it takes to learn & customize org-mode?)
Miscellaneously: I use Mnemosyne for spaced repetition, Liferea for RSS, Evernote/NixNote for clippings/notes, rTorrent for downloads, mpv/Clementine for media playing, irssi for IRC, arbtt for time-tracking, ledger for finances & Google Calendar for scheduling/reminders, Redshift for screen tinting at night to help bedtimes, and duplicity for backups.
As of June 2020, I use a workstation PC (which I built myself), a large Dell monitor mounted in portrait mode for reading8, a 200-foot Ethernet cable (which required I dig a trench to the next house), a Logitech thumb trackball, a generic keyboard (to be replaced by a Kinesis Advantage keyboard once I figure out the keymapping), and Bose noise-canceling earphones. In July 2020, I switched to the split ergonomic mechanical keyboard Ergodox EZ ($325). The workstation is plugged into a 900W UPS for protection against the not-infrequent lightning storms here, and a 6TB external drive for daily incremental backups, supplemented by Backblaze B2 (~$4/month) & miscellaneous external drives. While traveling, I use my ThinkPad P70 laptop, which replaced an Acer Aspire V17 (which died in a most unfortunate way), which replaced a Dell Studio 17, which replaced a PC I built ~2008.
I designed the workstation to be useful for deep learning, reinforcement learning, and Bayesian statistics, which made it much more expensive than I would’ve liked, settling on a Threadripper+dual-GPU design (while not forgetting that IO is often a bottleneck), but unfortunately those are fairly contradictory requirements (DRL wants RAM+CPU while DL wants just GPU), and the result wound up being expensive. (I went overboard on RAM in part because I was frustrated how I kept hitting RAM limits while testing out various dynamic programming algorithms for the Kelly coin-flip game, and because that much RAM means that entire datasets can be cached or worked with in-memory in R/Python, saving the considerable complexity of out-of-core algorithms or optimizations.)
The workstation is a liquid-cooled AMD Threadripper CPU build on a Gigabyte X399 Designare EX motherboard, 2×1080ti NVIDIA GPUs, 110GB RAM (nominally 128GB but final stick is unusable due to apparent BIOS issues), a 1TB NVMe drive for OS/home, and an 8TB internal HDD for bulk storage, all in a (unnecessary but too fun to not have) tempered-glass case. The process of putting it together was difficult—motherboards/CPUs/GPUs have gotten more complex since I last built a PC back in 2008—and the first motherboard stubbornly refused to boot, and after I RMA’d it to Newegg (at a cost of $36), the second one initially worked but then died overnight.9 After tinkering & procrastinating for months, I gave up on the Asus motherboard, checked what Puget Systems was using for their Threadripper builds (ThinkMate was still not offering any), and copied their choice of Gigabyte X399 Designare EX motherboards, reasoning that if they were shipping hundreds of such systems, it must be relatively reliable; that motherboard, plus much more forcefully inserting the Threadripper CPU, finally worked, and I was able to switch everything over in June 2018. While the final result was as powerful and useful as I hoped (especially for working with Danbooru2018, where the 16-cores+2-GPUs allows me to create many different specialized datasets & experiment with many different GAN architectures) the experience of building it has soured me on building my own PCs in the future: I clearly no longer know enough about PC hardware to do a good job, and the more expensive the components, the less I enjoy the risk or fact of bricking them. In the future I will probably either rely more on cloud solutions or bite the bullet & buy prebuilt systems. The workstation parts list (PCPartPicker.com sketch):
Case: Phanteks Enthoo Pro M Series PH-ES515PTG_BK Brushed Black Tempered Glass Window ATX Mid Tower
motherboard: GIGABYTE X399 Designare EX
GPU: 2× Asus ROG GeForce GTX 1080ti GPU
RAM: 4×: G.SKILL Ripjaws V Series 32GB (2 × 16GB) 288-Pin DDR4
Total (December 2017): ~$3800. (I ordered most of the parts in December 2017, which turned out to be a high water mark for both GPUs and RAM, due to the cryptocurrency bubble, unreleased next generation of GPUs, and the RAM price-fixing cartel; between that and the usual price decreases, an equivalent system would’ve cost at least 15% less by November 2018.)
For scanning books, I use a 12-inch guillotine paper cutter to debind books evenly (a big upgrade from using X-Acto knives with Fiskars curved blades), an Epson sheet-fed scanner with
imagescan for scanning &
gscan2pdf for post-processing.
My desk is an old desk made out of plywood & plumbing hardware by my great-grandfather for my aunt; I repurposed it when I realized it was the perfect size and height. In July 2020, because I failed to find any good standing desks I could buy used locally to test it out, I gave up and bought a 48x30 curved bamboo Jarvis standing desk ($609). I experimented with a treadmill desk but found it distracting, chronically unpleasant, and distressing to my cat. I put the desk in front of my bay window so I could enjoy the view and rest my eyes, while watching what happens on the river. The bay window unfortunately often has direct sunlight through it, so I added reflective sheeting, which greatly reduces the heat during the summer (at the cost of making it gloomier in winter, of course, but that is why I have bright LED bulbs). The chair is a used Aeron chair I bought off Craigslist for $225 in November 2016 (a bargain, although I doubt I would pay the list price); I replaced the mesh when it tore through in April 2021 for $60, which is much cheaper than the usual recommendation of replacing the entire seat-pan ($200). The sisal cat tree (Petco) provides an excellent perch for my cat, and I have added a pet flap with a cat window sill so he can more easily come & go, with acrylic sheeting to reduce air flow. (He turns out to greatly dislike soft surfaces, so half of the cat window sill was useless! I had to replace the foam padding & cover with a sheet of plywood I cut to fit.) The box fan by my feet (Walmart, $19) & the workstation both rest on rubber-cork anti-vibration pads. To reduce RSI, I keep a grip exerciser around to use during idle moments like watching videos. For making tea, I boil water in a simple adjustable electric tea kettle which I’ve made ‘programmable’ by drilling a hole into the clear plastic & inserting a meat thermometer (which combination is far cheaper than electronic kettles and more trustworthy); I then steep the tea in a Finum filter inside a big Colonial Williamsburg ceramic fox mug.
- Darcs: users/devel
- Haskell libraries
- Google “Power Searching” (2012); review: negative (I am probably too good at searching to benefit much from it; I have written a tutorial on more advanced Internet search skills which goes beyond “Power Searching”’s basics.)
- CMU “Probability & Statistics” (2012); review: positive
- Code School “Try R” (2012); review: mixed (short, basic introduction; cool interface, but my R knowledge was already past it)
- MRU Development Economics (2012–2013); review: mixed
- Coursera “Data Analysis” (2013); review: positive
- SICP (2009-)
- Coursera “Drugs and the Brain” (2012); review: negative
- Coursera “Probabilistic Graphical Models” (2013); as hard as its reputation (I couldn’t handle the material and learning Octave) but I hope to try again in the future
This section covers some of the most important things possible to know about me: my personality and mental description. No doubt some readers expected a carefully airbrushed & potted biography describing where & when I was raised, what my familial & tribal affiliations are, or what famous institutions I am affiliated with; even though this information is almost entirely useless—what can one predict about me if one knows that I was born in Illinois and raised on Long Island, but (maybe) my accent and a general liberalism? The irony—that people want most the information they will learn from least—will not be lost on those familiar with signaling. In contrast, standardized & validated psychometric instruments like the NEO-PI-R or RAPM really do have predictive validity for many life outcomes.
(Much of this data comes from YourMorals.org. I plan to retake the surveys, if possible, every decade; it will be interesting to see what changes.)
To describe my personality briefly: I am introverted, calm, neither particularly industrious nor lazy, contrary, and pathologically curious. I have made a copy of my 2011–2014 responses to the YourMorals.org corpus; discussed in more detail below. My scores on the “Big 5 Personality Inventory” (1/2/3):
- Openness to Experience10: high (short) or 87/87th percentile (long)
- Conscientiousness11: medium or 64/69th
- Extraversion12: low or 6/7th percentile
- Agreeableness13: medium-low or 3/3rd percentile
- Neuroticism14: medium-low or 16/13th percentile
For those who enjoy playing the game of ‘ad hominem via lay psychiatric diagnosis’, may I suggest not accusing me of Asperger syndrome—which is so overdone—but something more novel & scary-sounding like schizoid personality disorder?
The relevant results
- Email: email@example.com; I do not use Skype.
1Gb89tyJq3P5K5M3GcpFvPrMsw33cik9wX(canonical address; used for
- PGP key (mirror; fingerprint:
-----BEGIN PGP PUBLIC KEY BLOCK----- mQGNBGEs5RgBDADVRdbC0D0LfQwpQFvTNJLZX6s3Ew/Tk5LW8VgK32Z1zhQr6YdB nm0/U/mlm32SsoqEY+u5O/zvfF9GNA7sbatZwTPFlO5+Inq/FNqMq+dht52TSqiz I6Oya1ryuI7142nga90nQhXlHLkwkIZODdb6QYIcgIOrc0yI3bEDmFw5bVHXAMH5 J6rjADi27ZALZxLoU4L+u4vgtHOVBOWsuTOaXR4z0zTJ/Wi3Emi1EqZXu8Z4dTfq YbVLqpL//z3B8ZZLc+1dY8lzacPXWulFhGN9yHeMz/vXgXdOSZfHMQYytqTL3Z3k prv11QLSF3UaE4Fncm7FZkX2YjVP3vGghinZ3Uvv71bJdU3FbwM7KK6wbby2cBo0 9xaZlyYrv8e8MxqTa5LYJw+j84DOE6bQq2KyyJewjc0Q0vz+t43rNr5uwxD1DJuZ 90jb3ylvDPu6qMX/AVlgODXTmV+8GIKiJP0UytpDi549ZatjTJLTMHHySxytFONZ jlcu0+88MBpMcSEAEQEAAbQfR3dlcm4gQnJhbndlbiA8Z3dlcm5AZ3dlcm4ubmV0 PokB1AQTAQoAPhYhBAMp4TEp4I8Z7bpyUGeKxRbdaojPBQJhLOUYAhsDBQkDwmcA BQsJCAcCBhUKCQgLAgQWAgMBAh4BAheAAAoJEGeKxRbdaojP4tIMAIMkP60FS3Nm HZbwk0t48zJDGLzpqK0UWnz+BxLRxJPOgZY7+0P53nfcpSFiwiG+cFB+06eaK5Hg ttcGJd6NcLnAyHdHDqUOSF3Yf/8vIXIzggC9jt+mc2ug27GGHAT5asY4CNX8N5Ho VH1cnCvrTMDINRrir8BX1puPnsfRIshG8cVaKW/hXpDHRvpq3v2UbMULHUBTLPBE PDDH/O2VUG2hj267TIcM/dvaBwfC74R47Phkiis+nwcFr7OsYs3mIkCsZ4RNxYns xuNkv1aNSFI6agmvBSfsoLg9Ch2Viw6CEfLpeAwzFH/DIljzHWJJ3JqCYYEEEGkI eb7cAzfdCoR2iISSGR4GH1XNujaFSTXnuvX/vUsxdvmaxnwm2VkbNDLEwCAv8gxA O/eA6JJGZl1jbQq/oUouYWqGmLjMBVVjR2r3Y9KNJMsLOT6xOXKwDWAsdNS6N1uv d4B7ZyE16mPfpH0SB8uNpT155FSlvRGo9rWvAWIPZh3X9Ko3/aX8u7kBjQRhLOUY AQwAycf9rk7GFRm7d7EE38GnUzwrkEWug4uczdRhknXN5rP4/a2aBOaUEX8WE7Lu vyyiCWyI2qca3koVCzOt8DR3zUGhAAIZruHJmC8zWMSM9Bu4oLxmK1EvZ8zKpWKT 1vsfIwKd3NsjEFTWvC19BI1S5NRWRiK97Kx6IPGcchSRk8SV1dbLpMNFM/u1oUuZ lS0DGydjvl+mcDHRL4Zk2Xkj3XcmJ8kQIAT0814UtuOSHvbcEkVU7fDs64MgE6Z1 6J9nUZYv6nN870N7QHBWmoKhipGEeNNimHZ4OdZuJdsSS19iHKYu9LVM68YMtBC2 FypgLs7BOmaiqtfE/NpIOeJGkxW8S7UQtzsZwzfLzD81k0fXZectsMhS1Cf4aLMR vcWkucLHGW4QcvuUDOHCjtNPSX5MTHngI6qZOHQfVDV2bsc4Gge7TRtj8ghkhkDh g3ElnFExJQkzwQcd2lTy3vXx4Ne1bxtcqNJIYiGZ+HCYaBvD3y2QgtckfxxBt1D5 MRu/ABEBAAGJAbwEGAEKACYWIQQDKeExKeCPGe26clBnisUW3WqIzwUCYSzlGAIb DAUJA8JnAAAKCRBnisUW3WqIz7axC/9SDOtdsW/H2PezxDAEHasQvMJsS6/CxPL+ MvEEMx7GTQBUk2Ktz538lJhxUirw9UfixcxuIuaBwta/pDbeqqpHFg7gJbU+pvDQ ROqgv84gsvUcAYfOIFFKfKmyE1NO+dSGmyDvCaUr8Z5xKWWwAg5MEHNDi0zJWV/F lCRCb4Q6o7lkzHWljzzAAN+UHs2lq0089WMor/D7CXD5yQGUjcXGbLvoxJl5mDtQ 0TwRdGlF7RQDOSEvNro0eooed80im4bXW+ahhuZsylocVPIlgRoOjCkSNKQTnbd/ ysTzPhCdi59UYIvOuogRxGGBr/uDBcq53amZeVIsxHyT6v3njw0oGC8+2F1fPuSp 5A234E2+lNBmO1TZaM9SH5G0V0mDaV+35JS+gVWQ0R7qRix33+yXHcSCdpoIQTRj wnuJtQ6JQ4ToVdZsGyfnn6N6fkTvYFx3bYl2EV5d30cb5oQpcd4QiFDXzZg2yjCE p5IV9yhbHXXHKYyLpLwJN8KpBO4oM0s= =1NJ8 -----END PGP PUBLIC KEY BLOCK-----
#haskell, I was asked why I have no large programs to my credit; I replied, “My problem is that most programs I use already exist.”
I am not a bad Haskell programmer (although I am no guru like Simon Peyton-Jones, Apfelmus, or Don Stewart), but given how long I’ve been using Haskell, my contributions probably look pretty slim. This isn’t because I don’t like Haskell—I do, I find functional programming natural: defining transformation after transformation until the result is what I need. And of the functional languages, Haskell seems the best combination of power beyond basic arithmetic or list processing, one of the best ecosystems, and good basic language. (Which is not to say it’s perfect: there are some sharp edges in the basic math which irritate me when I’m messing around in the REPL.)
This is partly because of my style of contribution. I’ve always preferred to work on existing applications and libraries than to go write my own. I’ve always preferred to take someone else’s work and bring it up to snuff than write a clean implementation of my own. I’ve always preferred prodding the author or maintainer to do the right thing than to drop a large batch of patches onto them. Likewise, I view it as better to use Haskell standards like Cabal or Darcs than to use something like Autotools even if the latter lets us manage just a little more automation. I view it as better to upload to Hackage than to use any fancy site like Github or Sourceforge.
It’s better to do yeoman’s work taking two similar modules in two applications and split them out to a library than to write even the fanciest purely functional finger tree using monoids. Better to commit changes that reduce user configs by a line than to demonstrate once again the elegance of monads. Better by far to file a bug than wank around in
#haskell golfing expressions.
It is much better to find some people who have tried in the past to solve a problem and bring them together to solve it, than to solve it yourself—even if it means being a footnote (or less) in the announcement. What’s important is that it got done, and people will be using it. Not the credit. It is a high accomplishment indeed to factor out a bit of functionality into a library and make every possible user actually use it. Would that more Haskellers had this mindset! Indeed, would that more people in general had this mindset; as it is, people have bad habits of repeatedly failing when they think they have special information, are highly overconfident even in objective areas with quick feedback, and badly overestimate how many good ideas they can come up with15—indeed, most good ideas are Not Invented Here. One should be able to draw upon the wisdom of others.
This is an ethos I learned working with the inclusionists of Wikipedia. No code is so bad that it contains no good; the most valuable code is that used by other code; credit is less important than work; a steady stream of small trivial improvements is better than occasional massive edits.
A leader is best when people barely know that he exists, not so good when people obey and acclaim him, worst when they despise him. Fail to honor people, They fail to honor you. But of a good leader, who talks little, when his work is done, his aims fulfilled, they will all say, ‘We did this ourselves.’16
This is not an ethos calculated to impress. Filing bug reports, helping newbies, commenting on articles and code, cabalizing & uploading code—these are things hard to evaluate or take credit for. They are useful, useful indeed (shepheb or, eg. myself, never boast in
#xmonad of having helped 5 newbies today, but over the months and years, this friendliness and ready aid is of greater value than any module in all of XMonadContrib.) but they will never impress an interviewer or earn a fellowship. Is that too bad? Did I waste all my time?
I don’t think so. I value my contributions, and the Haskell community is better for it. It may have made my life a little more difficult—all that time spent on Haskell matters is time I did not devote to classes or jobs or what-have-you—but ultimately they did help somebody. One could do worse things with one’s time than that.
I mostly contribute to projects in Haskell, my favorite language; I have contributed to non-Haskell projects such as StumpWM, Mnemosyne, GNU Emacs17 etc. but not in major ways, so I do not list them here. After starting this website, I wound down my regular coding activities in favor of my writings; when I code, now it tends to be tools documented or hosted on this website (eg. Archiving URLs, Resorter) or integrated into writeups (eg. Generating Anime Faces with StyleGAN). For that code, you can browse by language tag: C/CSS/Haskell/JS/Python/R/Scheme/shell.
Below is a more detailed list of my old Haskell contributions, most of which is now of only historical interest.
- wrote tutorial on configuring the time-tracker & defining rules: “Effective Use of arbtt”
- documented dependencies, similar software, configuration syntax mode, CLI flag corrections
- Switched from FastPackedStrings to ByteStrings
- Low-level C optimization
- Initiated Cabalization (my work initially appeared as darcs-cabalized and then was merged into HEAD and
- Refactoring of shell tests
- Initiated switch from MoinMoin wiki to Gitit
- Identified performance issue & instigated addition of
--max-countoption for Filestore
regular XMonadContrib patch reviews
Improved Emacs keybindings
Initiated ‘Unicodify’ or ‘Pretty Lambdas’ feature for Haskell syntax highlighting
Added movement-related functions for improved incremental search
Comment support to cabal-mode
Adapted to use Mueval
Refactored out code in multiple packages:
Implemented run-in-any-directory functionality (previously Lambdabot could only run in the repository directory)
Maintained it (with Cale Gibbard)
- Wrote Darcs backend (which was moved to the filestore package and became Data.FileStore.Darcs)
- Wrote RSS support
- Wrote Interwiki plugin
- Wrote Date plugin
- Wrote WebArchiver & WebArchiverBot plugins (see later archiver standalone tool/library)
- Wrote Unicode plugin
- Wrote HCAR entry
- Misc. bug reports & suggestions
- Added PDF export functionality
- Integrated JQuery-based floating footnotes
archiver: Wrote and maintain it (see release ANN)
Mueval: Wrote it
wp-archivebot: Wrote it (see release ANN)
Change-monger: Wrote it
- Helped replace
Data.List.sortwith YHC’s more efficient implementation (see Ticket #2143)
Unix: fixed a possible runtime crash in
- Improved basic functionality
- Implemented an XMonad-style reload system to allow actual customization
- Maintained it
- Updated for GHC 6.8 & 6.1021
- Replaced the non-Free level data and graphics with Free ones
- Improved examples, docs
- Added UTF8 support
- Made use ghc-paths library
- Enabled QuickCheck support
- Added GHC-options support
Hlint: added GHCi integration
- Cleaned up their third-party modules
- Fixed up various Cabal issues
- Helped maintain it
Tagsoup: replaced old custom HTTP download code with standard library functions
Hashell: Updated for 6.8’s GHC API; Cleanup; Cabalized
As part of my effort to help shift the Haskell community to the use of centralized packaging repositories pioneered by CPAN, which is a fundamental requirement for any modern language, I made a systematic effort to get all extant Haskell code into Cabal format & uploaded to Hackage—whether the original authors wanted it or not. (For all the ruffled feathers and continued infelicities of Haskell packaging, a decade later, no Haskeller would go back to the pre-Cabal/Hackage Autotools days.) I cabalized and/or uploaded (according to the 2013-05-10 Hackage upload log):
- greencard & greencard-lib
Haskell In Space
This is a literary way of saying I am not as interesting as my writings, and in some respect, it should not matter who I am or what I have done because argument screens off authority.↩︎
When I say “research assistant”, I mean it in the older sense of someone who does detail work for another person’s original research—so I spent a lot of time reading up on specific areas and making notes about stuff my boss needs, and only occasionally do independent work. Not all my work can be made public, but some of it is. A partial list in rough chronological order:
“Intelligence Explosion: Evidence and Import”, 2012?
One of my first projects was working on this paper—digging up citations for various claims and giving general feedback; most of my changes or comments are private when we switched to working through Google Docs, but some of my initial comments were public.
finding predictions / estimates for whole-genome sequencing costs
“The Personality of (great / creative) Scientists: Open and Conscientious”
Summarizing some of the existing research; besides being intrinsically interesting, this is also relevant to computer intelligence—if achievement is just a reflection of IQ, then any computer intelligence will improve on humans only to the extent it has greater raw intelligence, but if achievement is limited past 130s IQ, then the picture becomes more complex. (Computer intelligence may be far more more powerful than predicted for a given intelligence level because there’s no apparent reason you couldn’t easily make its ‘personality’ open-minded—it not really being clear that truly random exploration and computing power don’t yield creativity in some fields already—and a computer can be tireless and self-disciplined, which takes care of Conscientiousness. On the other hand, Openness and Conscientiousness may correspond to subtle high-level dynamics which can’t be easily designed into an AI and can’t be produced by throwing computing power at the problem, in which case the AI will underperform whatever measure of intelligence we ultimately use, like AIXI-style IQ tests.)
Some quick notes on remote monitoring software MIRI could use to monitor freelancers such as myself. (It’s a good thing, really—trust but verify.)
Some info on the big Beijing genomics project, and estimates about personality and IQ, and past Chinese abuse of genetic information
A detailed examination of one cognitive bias asking whether it has been misinterpreted as a problem for individuals.
“Against Utilitarianism: Sobel’s attack on judging lives’ goodness”
Philosophy essay summarizing and attempting to analyze a recent class of objections to utilitarianism (that one cannot judge utilities of lives, and hence the utilitarian ethical project is impossible).
Neuroscience additions to Wikipedia; I had a hard time finding references for a MIRI paper, and donated my results to Wikipedia so at least no one else has to suffer so much:
- Exactly how big are chimpanzee & human brains, anyway?
- Primate and human brain size correlates with IQ (No, really. The correlation isn’t terribly big, though.)
- Rat IQ (mixed evidence, thankfully)
Not really enough data to do anything but point out problems for reliable assessment. My subjective impression is that polyamory is complex enough that it may not be a good idea even for polyamorists.
“Cashing Out Cognitive Biases as Behavior”
Spin off from “Are Sunk Costs Fallacies?”—2 studies examining consequences in personal life from vulnerability to cognitive biases in questionnaires
“Experiment: a good researcher is hard to find”
A simple email experiment on making applicants feel that their admission test is not a waste of time: they seem to do better work.
Geopolitical intelligence failures—notes from some reading in the history of intelligence failures
The following is a list of my submissions to LW I regard as substantive or particularly good, excluding content which can be found on Gwern.net, in chronological order with interesting ones highlighted:
“A good volunteer is hard to find”
“On the unpopularity of cryonics: life sucks, but at least then you die”
JET paper reviews:
“Inverse p-zombies: the other direction in the Hard Problem of Consciousness”
“The Personality of (great / creative) Scientists: Open and Conscientious”
“To Learn Critical Thinking, Study Critical Thinking” -(general philosophy studies are not as efficient in improving critical thinking as critical thinking courses)
“Against NHST” -(the statistical controversies over the p-value paradigm)
Of course, I don’t agree with every MIRI or LW position. The intellectual homogeneity has been much over-estimated by outsiders who have not bothered to look at the annual surveys, I think. Here are some major points for me:
MWI: I think that LWers who were persuaded by Eliezer’s MWI writings are wrong to do so, as they are unfamiliar with even the rudiments of any alternatives interpretations and cannot judge in the matter; how many LWers have ever seriously looked at all the competing theories, or could even name many alternatives? (“Collapse, MWI, uh…”), much less could discuss why they dislike pilot waves or whatever. Lacking any real understanding, they ought to simply adopt the expert consensus, where MWI seems to have a plurality or bare majority of adherents (with the weak confidence that implies).
Heuristics and cognitive biases: I am not much convinced that knowledge of heuristics & biases help in ordinary life. Feedback & learning are powerful tools in eliminating error, calibrating predictions, and justify committing what may look like the sunk cost fallacy; and feedback is what one gets in ordinary life.
Per Moravec’s paradox, where our knowledge of heuristics & biases will pay off most is in what Hanson would call “Far” scenarios: evolutionary novel situations with few precedents and only costly or non-existent feedback. (For example, the question of whether artificial intelligence will be developed by 2040: it will only happen or not once, there are few comparable events, the consequences may be dramatic, and our ordinary lives offer no useful insights.) As it happens, this describes much of futurism & forecasting but we cannot justify our futurism by claiming its techniques are incredibly valuable in ordinary life!
Cryonics girl: The donations appall me, for reasons I lay out at length there—they are a complete abandonment of core ideas like utilitarianism & optimal philanthropy.
Alicorn’s “Living Luminously” paradigm struck me as dubious, not backed by even token research, and likely idiosyncratic to her; I thought her Luminosity e-novel was merely OK despite the endless discussions on LW (rivaling those for Methods of Rationality itself) and that her followup, Radiance, was just terrible. Nevertheless, her novel career seems to continue.
There is a moderately funny story about how Gerard came to write it, based on my musical incompetence.↩︎
That is, summing up the (surviving) edits of my various accounts over the years: User:Gwern, User:Marudubshinki, & User:Rhwawn↩︎
Compare the CSE results with the Google Results for the anime Wings of Honnêamise. Which is more useful for an editor? For more details, see my release announcement.↩︎
A trick I discovered when visiting FHI in 2015—I had used widescreen laptops for so long I had forgotten how nice portrait-orientation was for reading.↩︎
My best guess is that my problem initially was that I seriously underestimated how much pressure it takes to insert a Threadripper CPU into its socket—it required a truly terrifying amount of force and I only got it right after triple-checking online tutorials & videos & discussions—and that was why the first motherboard never worked at all, and the second one was killed by static electricity or a short.↩︎
See also “Actively Open-Minded Thinking Scale”, “Clarity Scale”, “Engagement with Beauty”, & “a measure of what types of stories you enjoy”.↩︎
See also “Zimbardo Time Perspective Inventory”. Brent W. Roberts criticizes these two inventories when used to measure Conscientiousness.↩︎
See also “Relational Mobility scale”, “Empathizing and Systemizing scales” & “Rational vs Experiential Inventory”.↩︎
See also “Self-Report Psychopathy Scale”.↩︎
See also “Experience in Purchasing Behavior Scale” & “Kentucky Inventory of Mindfulness Skills”.↩︎
For further reading on overconfidence, see all LW articles so tagged. I once read in a book of a study in which subjects were asked to generate ideas for, IIRC, putting out a fire, and to stop only when they were convinced they had thought up all good ones, and usually stopping when they had thought up only a third; but I have been unable to refind it and would appreciate knowing details if this description rings any bells for a reader.↩︎
For example, my clean-up and extension of the
browse-urlmodule was completely rewritten by RMS; so I can hardly take credit there.↩︎
Henceforth, this implies I have a commit-bit (or equivalent) for that project.↩︎
Henceforth, ‘cleanup’ should be taken as referring to extensive miscellaneous changes which include (in no particular order):
-Wallor hlint warnings
replacing OPTION pragmas with LANGUAGE pragmas
tracking down licensing information
switching from Haskell98 imports to the standard hierarchical module imports
import Data.Char; nontrivial in some cases where Haskell98 modules were dispersed over multiple base modules
reorganizing the file tree
improving the Cabalization
whitespace formatting, and so on.
Henceforth, this typically implies that I uploaded it to Hackage as well↩︎
Henceforth, this implies that I made whatever changes necessary to get it compiling on GHC 6.8.x and 6.10.x↩︎