I’m trying out Adsense

You’ll see a banner between in between the title and post(or another location – this is just an experiment) for some items on the page and in the feed. I won’t see them, because Firefox is blocking them, but they looked relatively unobtrusive and highly targeted in IE. If you have any comments, please leave them here.

After a few months, I’ll re-consider whether it has been worth it.

Why isn’t there more research on cognitive enhancement?

One issue is that cognitive enhancement as a research area suffers from the same problems as life extension, unfortunately, in that it kinda attracts a different sort of person. Most researchers feel it’s more noble to fix something that’s broken instead of trying to improve something that’s currently working, however inefficiently. I generally agree with that sentiment, but certainly there is some worth to improving things, too. I’m not suggesting that more brainpower should be spent on inventing “smart drugs” than on curing cancer, but, as many have noted, this isn’t a zero sum game. There’s just so little that’s been done in the past decade on “nootropics” that there must be just bushels of relatively low-hanging fruit, ripe for the picking, and if gains could be made relatively quickly, they’d have a large payoff because they’d be applicable to everyone, instead of just the people with disorder or disease X. There would also be an additive effect of increasing the rate at which progress could be made in all the other fields.

This kind of work will face the same class issues as other types of medical research like genomic profiling, with people worrying that the discoveries will only be available to those who can afford it, thus widening the have-have not gap still further, but that’s not really a good argument for not doing the research, because it has to be available before it can become affordable, doesn’t it?

It’s a difficult thing to argue for, trying to fix something that ain’t broke, but I’m convinced given the utter lack of progress in this area over the past decade, that just one competent researcher in this area could make significant advances relatively quickly. Am I wrong? Has there been significant advancement in the last decade? Am I being too hard on sports performance researchers?

Checking Usenet, they’re still talking about the same stuff they were 10 years ago:
ginkgo
choline
huperzine
DMAE/centrophenoxine
vinpocetine
piracetam
hydergine

Little of this is available OTC in the US. The only thing I’m aware of with US OTC availability that’s not essentially repackaged caffeine is this stuff.1 Has anyone tried these? Let me know in the comments.

1. I personally like it; I’m not connected to the company in any way.

Helicos Biosciences is reporting the sequencing of M13 in Science.

Helicos Biosciences is reporting the re-sequencing of the M13 viral genome.
What’s really cool about this is that it’s sequencing-by-synthesis, requiring no amplification before sequence reading, so there’s no biasing of the populations of DNA fragments. Here’s a short overview of how their sequencing by synthesis works:

So it’s still a shotgun-style technique, and the problem has been getting long enough read lengths to assemble the pieces into a whole sequence. Think of it like a jigsaw puzzle. The smaller the pieces are, the more you have that look identical, and the harder it is to figure out where to put them. Given 3 x 109 bases in the human genome, you’d need a read length of about 17 base pairs to be certain of ending up with all different pieces. They report average read lengths of 23 bp, with each individual sequence chunk represented about 150 times, and every part of the genome represented. This allows them to keep the error rate low enough that they could sequence different strains of the virus, and reliably pick up the genomic differences between the two. The paper doesn’t say how long this takes, but their marketing material says the process takes only one day.

The technique as practiced by Helicos has one major drawback, however, that will limit how low it can take the cost/base. It requires very expensive optics, since you’re essentially doing FRET on an array of targets. The list price for their instrument is $1.3 Million, with a consumables price of $18,000/sample, placing it out of reach of many institutions.

Illumina, another company with a sequencing-by-synthesis application, gets around the expensive optics issue by doing a clever solid-phase amplification of their target strands, essentially growing little colonies of identical strands in situ. A sales rep told me they’re selling around $500K, but didn’t have details on the consumables cost.

Long-term, I’d expect the Polonator to be the most widely adopted platform, due to their aggressive pursuit of royalty-free technology and low instrument cost. There’s a good thread to follow here, if you’re interested.

23andme and Navigenics take note: They’re not offering a direct to consumer service, yet, but there’s no way a SNP scan can compete with a full sequence, once the cost comes down.

via HotCites, more on this story at in-sequence.

Science Debate 2008

I’ve been a big supporter of the idea of having all the candidates get together for a roundtable on science-related issues. I’m subscribed to the mailing list for Science Debate 2008, which has been trying to organize such a debate. The organizer recently pointed out that the candidates are attending the Compassion Forum, which is a debate on “moral issues”. I don’t know if there’s more going on behind the scenes regarding why the candidates would chose to attend one conference and not the other, but it just seems to me like the Democratic candidates, at least, would want to seize the advantage provided by the recent weakening of the religious right, and it also seems like attending a forum on moral issues might not be the best way to do that. Maybe they’re trying to catch some swing votes, but they certainly shouldn’t give out the impression that moral issues remain the most important issues for this election, because that’s exactly what led to their loss last two times. You know Obama and Clinton would both kill McCain in a science debate.

Here’s the email:

I am sorry to send two emails in such short succession, but I thought you should know that after declining our invitation to debate science in Pennsylvania, Barack Obama and Hillary Clinton yesterday agreed to attend “The Compassion Forum,” a forum of “wide-ranging and probing discussions of policies related to moral issues.” CNN will serve as the exclusive broadcaster of the “presidential-candidate forum on faith, values and other current issues” at Messiah College near Harrisburg, Pa., April 13 at 8 p.m. You can read more here.

Perhaps among the moral issues discussed should be whether they have a moral obligation to more fully engage on science issues, since the future viability of the planet may hang in the balance, for starters. Is there a larger moral imperative? How about the future economic health of the United States and the prosperity of its families? Science & engineering have driven half our economic growth since WWII, yet but 2010 if trends hold 90% of all scientists and engineers will live in Asia. Then there are the moral questions surrounding the health of our families with stem cell research, genomics, health insurance policy, and medical research. There’s biodiversity loss and the health of the oceans and the morality of balancing destruction of species against human needs and expenses, there’s population and development and clean energy research, there’s food supply and GMO crops and educating children to compete in the new global economy and securing competitive jobs. Science issues are moral issues.

I would encourage you to write letters to the editor, emails to the campaigns, and blog postings pointing this out. And if you can, support our ongoing effort to turn this country around.

Shawn Lawrence Otto

ScienceDebate2008.com

Distributed data and distributed analysis

Two smart people I read, who probably don’t know each other, and work in disparate fields, both have a post today about using a distributed collection of data and analysts to answer problems that we’re just now being able to address.

Here’s Jon Udell talking about distributed communities of climate scientists studying CO2 fluctuations.

Here’s Steve Hsu talking about the quantitative finance firm Horton Point, and how they’re assembling teams of academics from diverse fields such as psychology and data mining, to come up with dynamic models of how markets work.

The problems with current genomic association studies.

There’s a nice roundup of the problems faced by current chip-based techniques to find disease-gene associations. This is the technique practiced by 23andme.

Current chip-based technologies for genome-wide analysis, while having some success in identifying the lowest-hanging genetic fruit for many common diseases, seem to have already started to run up against barriers that are unlikely to be overcome by simply increasing sample sizes. These technologies should really be regarded as little more than a place-holder for whole-genome sequencing, which should become affordable enough to use for large-scale association studies within 3-5 years.

Nothing beats a large, diverse library of full sequences. There will be many a position available for the enterprising bioinformaticians needed to compile and interpret this data store.