By Ctein
Okay, now there's one thing you have to keep in mind. These folks are trying to do something fabulously difficult: write a computer program that can develop a sense of aesthetic judgment. That's so hard that humans can't even do it right or consistently. Read any sample of back columns on TOP and see how often there's unanimity on the aesthetic merit of a particular photograph, let alone any kind of cogent or coherent explanation of why it has merit. We humans don't agree and we're largely inarticulate on the subject. We certainly can't expect much from a machine.
Still, it's hard not to be amused by its efforts, the same way we're amused by inept computer language translation. That's not unrighteous, so long as we stifle the snarkiness. And not take it too seriously.
The program's called ACQUINE, short for "Aesthetic Quality Inference Engine."
Go here and you can enter URLs for (or upload) your photos. How could I resist? I picked a sampling from my website of my favorites (and my audience's, I must emphasize).
For the sake of calibration, there's a reference photo of the Macbeth chart from my website. ACQUINE rated it a 10 on a scale of 100, 50 being "a typical average-quality professional photograph" according to the programmers. Okay, so far so good.
Next I handed it my photograph of Castle Eileen Donan, and that garnered an amazing 98.5! "Wow, that computer has excellent taste!" I thought objectively.
It's a 98.5! Do I rock, or what!?
So I fed it a bunch more stuff, and while one picture garnered a mere 34 (obviously a computer error), the average score for the rest was over 80. Insert artist smugness here.
It bothered me to be using URLs for photos directly from my website. As I present them, they're all encumbered with dark gray "frames," grayscales, and title information. That has to be detracting (aesthetically) from the photograph. I figured the smart and sensible thing to do would be to strip off that extraneous matter and present ACQUINE with the photographs in a clean form.
So I did—and my scores plummeted! A few stayed respectable; the Castle photo only declined to 86.5. But two thirds of the photos dropped into the teens and the one below came back with an appalling 6.2.
In other words, ACQUINE thought it was lousy even in comparison to a Macbeth chart!
6.2...6.2??? I am being dissed by silicon.
I guess it's like they say: It's all in the presentation.
For those who like to wallow in data here are the URL's to the photos I tested, along with the scores in their presentation form and the scores after I stripped off the "frames."
"Scotland 1," with frame: 82.3; "bare": 15.9
"Scotland 2," with frame: 67.1; "bare": 6.2
"Scotland 14," with frame: 56.5; "bare": 14
"Scotland 15," with frame: 98.5; "bare": 86.5
"Scotland 16," with frame: 71.1; "bare": 15.2
"Scotland 17," with frame: 34; "bare": 12.3
"VAB Bay," with frame: 87.8; "bare": 58.1
"Rose," with frame: 91.6; "bare": 15.1
"Rushes," with frame: 94.1; "bare": 71.5
Featured Comment by Chuck Kimmerle: "That software's hard to please. After my third photo evaluation, it suggested I sell my camera and take up plumbing."
Featured Comment by Adam McAnaney: "For the record, the above picture of The Robot rated a 64.7."
"...and the one below came back with an appalling 6.2"
Don't feel too bad. This (which was a favorite of mine from last weekend's hike in the Wild Basin of Rocky Mountain National Park) came back as a 1.3.
A Macbeth chart is almost 8 times more aesthetically pleasing? Yikes!
Posted by: Chris May | Friday, 22 May 2009 at 09:31 AM
I played around with this yesterday and found that borders are key to high scores. Apparently nothing says aesthetic quality like the design sense of a motivational poster.
Posted by: The Wallbanger | Friday, 22 May 2009 at 09:39 AM
Inline with what I imagine will be many people's results, mine were a smattering to and fro.
However, I noticed when entering a certain black and white photo, my score was 89.3. A color version of the same photo went down to 39.5. Harrumph.
Posted by: Dustin | Friday, 22 May 2009 at 09:39 AM
Great tip, but have you read the bold fine print on the ACQUINE site; "A rule of thumb is that if the aesthetic quality of a photo is obvious to most people, it may not be worthwhile to seek Acquine's opinion on it because Acquine may choose to assign funny scores in such cases. Please be serious if you would like Acquine to help you"
So it would seem we're all not being serious enough ;-)
Now I'm off to rate my PICs...
Posted by: Christian Kurmann | Friday, 22 May 2009 at 09:43 AM
The results for Chris May's (very nice) photo proves that Black and White is aesthetically doomed.
bd
Posted by: bobdales | Friday, 22 May 2009 at 09:46 AM
"However, I noticed when entering a certain black and white photo, my score was 89.3. A color version of the same photo went down to 39.5."
Maybe there's something to this after all....
Mike
Posted by: Mike Johnston | Friday, 22 May 2009 at 09:59 AM
Now you know why people frame their prints.
Posted by: Lambert | Friday, 22 May 2009 at 10:00 AM
..and you seriously believe that's genuine software - some cackling performance artist hasn't hired a team of monkeys, or wired up a jellyfish or something?
Posted by: richardplondon | Friday, 22 May 2009 at 10:02 AM
I don't know.... I looked at the "top rated of all time" I don't trust it. Seems to be just like Flickr, a bunch of HDR and cat photos. Sometimes even Infrared HDR cat photos.
Posted by: Chad Thompson | Friday, 22 May 2009 at 10:27 AM
My scores for 5 pictures varied from 3.7 for this, to 97.7 for this. Which just proves I am a photographer takes can produce the near perfect picture and absolute rubbish, no news to me.
Posted by: Ruud van Ruitenbeek | Friday, 22 May 2009 at 10:46 AM
Hey, at least it can be bothered to look at your images and pass an objective (by its standards) comment. That makes it more valuable than 95% of online commenters.
Posted by: Ade | Friday, 22 May 2009 at 10:48 AM
It looks to me as though it really likes pictures that have a wide tonal range. Look at their top pictures of all time, and you'll quickly notice that they all seem to have large areas of full black. I suspect that adding a generous very dark area is why the frame boosts the score of your photographs.
I also noticed that somebody had loaded a bunch of Ansel Adams photos. "Moon and Half Dome" apparently rates a 99.5
Posted by: Roger Moore | Friday, 22 May 2009 at 10:49 AM
I believe it was Degas that said "The frame is the artist's reward" and apparently the software agrees.
Posted by: Jim Meeks | Friday, 22 May 2009 at 11:03 AM
I love richardplondon's take on this with reference to the "team of monkeys or wired up jelleyfish". I can visualize the setting of a low rent apartment with a bunch of chimps in front of a Mac classic monitor looking at images. I'm sure a framed jungle scene would do well and zoo photos, well.........or the jellyfish tank by the monitor all aglow with a nice sea scape and the light goes dim with a shot of the desert. So the program is called ACQUINE? You sure that's not EQUINE................Ass?
Posted by: John Sartin | Friday, 22 May 2009 at 11:18 AM
I'd just like to introduce the notion that maybe we should be striving for LOWER scores, on the assumption that anything considered laudable by a computer program is by definition a cliche.
Mike
Posted by: Mike Johnston | Friday, 22 May 2009 at 11:22 AM
You're all missing what's really going here. You see, what they're saying is that artistic merit is to a large extent arbitrary, or we'd have more agreement on standards and value. So, a software-analyzed photo is as arbitrary as any other, no better, no worse. If they can corner the market on computerized esthetic analysis, or at least be the first on the block, well, it could be the next big thing. Imagine doing away with all those expensive critics and gallery curators, art history professors, web forum contributors, etc., so that the whole world comes to your software for validation.
That's just the beginning. The next step is to develop image-taking software that matches that analyzer's value system, and then presto, you have a robot camera that will generate the world's best pictures and all photo buyers will have a central web location to buy those world's best pictures, the analyzer says so. You can fire all those people on staff who research and buy your images for you.
It's all over but the crying.
Posted by: Robert Roaldi | Friday, 22 May 2009 at 12:12 PM
I guess I'll think of this next time I read about an outrage perpetrated by a performance artist with NEA funding. My GOD, can you believe the way these people are throwing around taxpayer dollars from the NSF?!? It's APPALLING!
It reminds me of Painting by Numbers: Komar and Melamid's Scientific Guide to Art
Posted by: Andy Burday | Friday, 22 May 2009 at 12:13 PM
> ... maybe we should be striving
> for LOWER scores...
A subjectively good photo that gets a low rating by a cliche-based software?
That implies it works!
Alessandro
Posted by: Alessandro | Friday, 22 May 2009 at 12:13 PM
On a lark, I fed it the cover photo from the DOMAI website. Result: 7.3.
I will refrain from making any comment on ACQUINE's assessment relative to the Macbeth chart.
Posted by: Imnotme | Friday, 22 May 2009 at 12:17 PM
"Infrared HDR Cat photo" made me laugh. Of course then I googled it. Actually, the result _was_ pretty interesting... http://www.flickr.com/photos/platypusscooter/3023788754/
Posted by: Pen Waggener | Friday, 22 May 2009 at 12:19 PM
I went one step further and submitted two simple single-color backgrounds, a white and a black. Would you believe it? White scores higher than black. The light wins over darkness. It's very Tolkien-like.
The score? white: 41.7% - black: 24.6%.
I think the software needs serious tweaking.
Posted by: Vince | Friday, 22 May 2009 at 12:29 PM
Old news Ctein - editors have been Alpha testing this since 1826!
Posted by: Chris Gibbs | Friday, 22 May 2009 at 12:47 PM
Mike it's supposedly tougher than the human audience because they, again supposedly, strive against grade inflation.
But it has some silly biases for a site that's supposedly (:-)) about colour photography. I uploaded a colour photo with a thin white border. 64. The same photo with a thin black border, 77. And then this one, 98.4. And I'll admit this is a pretty cliched photo, aint' it?
Posted by: erlik | Friday, 22 May 2009 at 01:23 PM
I think that my personal sense of aesthetics must differ from that of the people who designed the program.
I'm also a bit sceptical about the consistency of the algorithm, given that after a while I started to see if I could game it by selecting pictures that were compositional matches for the high scoring ones. Nope.
Still, it provoked me into discovering that Zenfolio provides URLs for multiple sizes of my images, which is useful information!
Posted by: Rana | Friday, 22 May 2009 at 02:14 PM
Richard Plondon asks if "some cackling performance artist hasn't hired a team of monkeys, or wired up a jellyfish or something".
I suggest that some computer scientist just hired a performance artist to sit at a terminal and evaluate photographs.
;-)
Posted by: Mark Roberts | Friday, 22 May 2009 at 02:32 PM
The comments about it being a "team of monkeys, or a wired up jellyfish or something" have missed something important: a team of monkeys, at least, would probably do far better than this, which clearly just has no correlation at all with anything interesting. Monkeys are no fools, and many of them have remarkably good taste.
I'm not so sure about jellyfish, but I wouldn't be surprised if they do better too.
Posted by: Tim Bradshaw | Friday, 22 May 2009 at 02:35 PM
HCB was in fact a bad photographer, this one got a 2,7/100
I always thought he was overrated, now I'm sure...
The sadest thing is that some people are working hard on this, it must be really tough to program.
Posted by: Vincent Gautier | Friday, 22 May 2009 at 02:54 PM
According to this thing I am a better photographer than HCB
I knew that dude was over rated.
Posted by: charlie d | Friday, 22 May 2009 at 03:22 PM
I had my own ups and down with this program on Wednesday. Here are my results and what happened afterwards:
http://enticingthelight.wordpress.com/2009/05/21/why-i-am-quitting-photography/
Posted by: Miserere | Friday, 22 May 2009 at 04:33 PM
Could someone point me to the infrared HDR cat photo group on Flickr? I can't find it and it seems as if I could find some kindred spirts there.
Posted by: Tom Duffy | Friday, 22 May 2009 at 04:44 PM
"HCB was in fact a bad photographer"
Actually, it's not that bad. I enlarged the photo to 600x406 (they say it's better if the photos are at least 600px per side), added 25px of black border on each side and voila: 68.3.
Posted by: erlik | Friday, 22 May 2009 at 05:07 PM
Ctein has apparently been using the wrong color chart. The Kodak Q-60 gets a 57.3. Obviously making it 5.73 times better than the MacBeth.
Lee
Posted by: Lee | Friday, 22 May 2009 at 06:11 PM
For heaven´s sake!!!
You made me lose my whole afternoon!!!
This is the very best pastime ever!!!
And the sense of humour it shares, and the laughs it causes!
God, I´ve been stuck to my screen the whole afteroon, awestruck!!!
Can´t wait to show them to me lads!
Posted by: Iñaki | Friday, 22 May 2009 at 06:17 PM
I think we should all pour ourselves a cup/glass of our favorite brew (beer/ coffee/ whatever), flop down in puffy chair, and breath a sigh of relief. Our lives are already dominated by overly "thinking" machines. Yes, yes, I admit it is nice to have improved technology, and in some cases it can save lives, etc. But if and when someone creates a machine that can mimic or replicate the more complex facets of the human psyche, the jump to a HAL is but a short one.
Actually this reminds me of logistician who claimed he could analyze Zen texts to decode the experience of spiritual enlightenment. Well, the paper certainly evoked a response: one page in and the mind grows numb.
As Mike intimated, I'd prefer to leave the clichéd operations to the computers, and the creative innovation and interpretation of values to us fallible humans.
Alex
Posted by: Alex Vesey | Friday, 22 May 2009 at 07:11 PM
In a somewhat similar vein, check out what kind of funny paintings a poll will produce here: http://www.diacenter.org/km/homepage.html
Hint: LOTS of blue
Posted by: David Ray Carson | Friday, 22 May 2009 at 09:42 PM
Interesting that using a different size picture from Flickr got different results on the one I tried. Most of the B/W pictures got into the 90s and the highest rated one was 98.2 and not one I would think could get in the door. The color versions rated much lower...But heck..it was a good diversion this evening.
Mike T.
Posted by: Buckstop39 | Friday, 22 May 2009 at 09:43 PM
Here's a single line of shell script which delivers as-good or better results than this program.
wget -q -O- $IMAGEURL|cksum |awk '{ print $1 % 100}'
Posted by: Matthew Miller | Friday, 22 May 2009 at 10:32 PM
Give me a break. I tested it this morning.
It seemed to be all over the place with some of my B&W stuff---some good, some not. Sometimes it even rated the same photo in color higher than the B&W version and visa versa. Unbelievable!
More importantly, it could not consistently determine the superiority of photos taken with more modern, higher resolution dSLRs and some taken with an old Olympus ($110) P&S. Some of the photos from the P&S scored much higher than photos from a D40, D70, or even a D300. I suppose when I scan and upload a few from film, it will misjudge those too.
And what's wrong with infrared cat photos?
Posted by: David | Friday, 22 May 2009 at 10:34 PM
You all realize, don't you, that this will be the hot digicam feature of 2011? "Art Recognition" mode will only take the shot when the "aesthetic quality" rating of the scene is 90 or above, as indicated in real time by a bar gauge on the LCD. Who knows what will result, but it will be worth it for the contortions it evokes from earnest snappers everywhere. Barring too many mishaps from shooters backing into traffic or tripping over cacti, the next generation of the feature may let one specify styles: "Eggleston Recognition mode", "Vermeer Recognition mode", etc. A couple generations on and the camera will talk you through it. Mark my words.
Posted by: robert e | Saturday, 23 May 2009 at 02:06 AM
I tried half a dozen images that have won picture of the day competitions, and the scores ranged from 3.7 to 78.6. Amazing!
Ken from the UK
Posted by: Ken Croft | Saturday, 23 May 2009 at 04:28 AM
My night shot of a 1977 BMW R100RS motorcycle (http://inlinethumb34.webshots.com/43809/2661199120028267445S600x600Q85.jpg score=99.8) outscored an Ansel Adams photo who's link I submitted. I feel better now although I am pretty sure that Adams will continue to out sell me. Considering that I've never sold anything that should not be difficult. Interestingly, some of my personal favorite photos of my own scored poorly while I see some obvious composition flaws in the BMW photo.
Posted by: Doug | Saturday, 23 May 2009 at 06:00 AM
"HCB was in fact a bad photographer, this one got a 2,7/100 "
Looks like the deleteme pool was right after all.
Posted by: James B. | Saturday, 23 May 2009 at 11:57 AM
I *LIKE* that cat photo!
pax / Ctein
Posted by: ctein | Saturday, 23 May 2009 at 02:00 PM
After the national and international nightmares of the last decade, playing with this joke of a program for 10 minutes has almost completely restored my faith in humanity.
Posted by: Paul De Zan | Saturday, 23 May 2009 at 02:11 PM
Has anyone checked to see that the site isn't just a random generator?
Mike
Posted by: Mike Johnston | Saturday, 23 May 2009 at 04:28 PM
I don't think it's random, since it gives different copies of the same picture with different names from different URLs, roughly the same scores (I'm willing to concede a point or two to account for jpg compression differences - it seems to prefer higher-quality compression).
So if it's not random, and it's not a performance art piece, not chimps or jellyfish, if in fact someone is making a serious attempt to predict aesthetic appeal solely through computational methods ... well I just find that a really depressing waste of time.
And I'll be even more depressed if they eventually succeed in actually predicting the popularity of an image, though good luck finding an accurate way of quantifying that in the real world.
Meanwhile, plugging one's work into this tool feels a little like submitting one's prose to Microsoft Word for grammatical approval: as I said, just really depressing.
Posted by: Damon Schreiber | Saturday, 23 May 2009 at 07:41 PM
Was pleased to get a 91.5 on one of my photos, until I noticed that a totally black image scored higher. Also noticed that when I clicked on the "more from this user" icon under my photo, the site displayed a dozen photos I'd never seen before.
Posted by: Matt Mawson | Saturday, 23 May 2009 at 08:30 PM
That's not as funny, as it seems. Almost identical program is judging if I (or my colleagues) can take a mortage loan for apartment.
Posted by: Void | Monday, 25 May 2009 at 07:58 AM
Reminds me of a comment made by Sir Humphry Davy, English scientist and inventor of the Davy safety lamp.
He was asked his opinion after viewing an art exhibition in Paris and commented: "The finest collection of frames I ever saw".
Posted by: Steve Smith | Monday, 25 May 2009 at 05:30 PM
This program is clearly rubbish. Whilst this isn't the best picture in the world, ACQUINE rated it at 4.7 - poor girl...
Posted by: Mark T | Wednesday, 27 May 2009 at 03:26 AM
Dear Mark,
I don't assign numerical ratings, not being a computer, but the composition on that photo does seriously suck.
In general, I'm a bit surprised at how resistant people are to the very idea that a computer might be able to perform some of this sort of evaluation.
Do people really think that aesthetics is so obviously all contextual and spiritual that it has NO substantial psychophysical component? More to the point, are they so convinced this is true that they think the experiment is inherently unworthy of the effort?
pax / Ctein
Posted by: ctein | Wednesday, 27 May 2009 at 01:52 PM
This photo got 0.8. A new low record?
http://foto.no/linkeddata/bildekritikk/images/438000-438999/438450_1280x1024.jpg
Posted by: Ole Kristensen | Thursday, 28 May 2009 at 05:42 AM
Ctein - thanks for the critical analysis of my photo. In my defence, I was crushed in a crowd with the floats going past, simultaneously making sure my kids didn't run out into the road. Composition wasn't really an option... :-)
I think this sort of project could succeed, but only with a lot of training. I suspect something could be done with user feedback (think hotornot.com) and a neural net, perhaps working on a blurred or otherwise reduced version of the image to pick out the composition and colour balance (evaluated separately?) rather than the fine details. My Nikon camera allegedly has a 30,000 image database for the matrix metering - this sounds like a similar project to that, but with aesthetic evaluation as an output.
Posted by: Mark T | Thursday, 28 May 2009 at 09:01 AM
Dear Mark T,
I'm glad you took the critical analysis in the spirit in which it was made [ VBG ].
I also intentionally fed the program a couple of photographs of mine that I thought were pretty badly composed to see what it would do, and it largely agreed.
I think the project is actually being run the way you suggest. It's been around for a while, and if they were only interested in doing internal analyses, they wouldn't open it up to the entire world to submit photos. It's not about beta-testing a product, like some people think; it's about testing a model to destruction so you can learn why it fails.
In one sense, it is working pretty much like I would expect. It's not producing a lot of false positives (highly-rated photographs with sucky compositions), which means the most basic "aesthetic engine" has a fair understanding of good rules of design and composition. What causes the problems are the exceptional cases, the ones that produce false negatives, like my third figure and the photograph Ole just mentioned.
That's where the research gets interesting. Do those two photos fail the computer analysis because their appeal is contextual (that is, it's about what we're looking at, instead of how it's arranged) or is it that the aesthetic model isn't sophisticated enough yet?
I have my suspicions, but I honestly don't know the answer. Nor does anyone else reading this column. Which is what makes running experiments like this an inherently interesting and valuable endeavor.
Unless, that is, one is the sort who thinks the aesthetics should remain some ineffable, undefinable, human construct. That doesn't describe me, but it does describe some people.
~ pax \ Ctein
[ Please excuse any word-salad. MacSpeech in training! ]
======================================
-- Ctein's Online Gallery http://ctein.com
-- Digital Restorations http://photo-repair.com
======================================
Posted by: ctein | Thursday, 28 May 2009 at 04:31 PM
Since I have worked in that field for twenty years (before retiring to become a full-time photographer), I can state without doubt that the researchers who authored this software have excellent credentials, and have presented the research in extremely credible peer-reviewed venues.
It is possible that individual results appear erratic, however the whole approach includes sound ideas deserving of more examination than the dismissive and derogatory judgements used in some of the comments.
Posted by: Quang-Tuan Luong | Thursday, 04 June 2009 at 09:23 PM
Well human rating systems are not so great either. Or are they?
You guys have to take a look at this:
http://www.flickr.com/photos/andrerabelo/70458366/
Posted by: beuler | Monday, 15 June 2009 at 07:23 AM