{"id":1701,"date":"2022-12-25T23:25:45","date_gmt":"2022-12-26T04:25:45","guid":{"rendered":"https:\/\/www.fecundity.com\/nfw\/?p=1701"},"modified":"2022-12-25T23:25:45","modified_gmt":"2022-12-26T04:25:45","slug":"ai-cant-let-me-see-karl-marx","status":"publish","type":"post","link":"https:\/\/www.fecundity.com\/nfw\/2022\/12\/25\/ai-cant-let-me-see-karl-marx\/","title":{"rendered":"AI can&#8217;t let me see Karl Marx"},"content":{"rendered":"\n<p>Back in September, I wrote <a href=\"https:\/\/www.fecundity.com\/nfw\/2022\/09\/03\/can-ai-let-me-see-karl-marx\/\" data-type=\"post\" data-id=\"1606\">a post about generative AI and photographic transparency<\/a>. The gist of it was this: Kendall Walton famously argued that I actually see Karl Marx when I look at a photograph of him, in a way I don&#8217;t when I look at a painting. The painting is mediated by the beliefs of the painter in a way that the photograph is not mediated by the photographer&#8217;s beliefs. So, I asked, what about an AI-generated image of Marx?<\/p>\n\n\n\n<p>As I said in a footnote to that post, I wasn&#8217;t very happy with my answer to the question. As it happens, my Philosophy of Art class got interested in photographic transparency all on their own. So I made a mid-semester adjustment, added it to the syllabus, reread the Walton essay, and taught it to students in October. It turns out there was a part of the essay that I had forgotten when I wrote my post in September, and Walton gives us the resources for a better answer to the puzzle of AI-generated images.<\/p>\n\n\n\n<!--more-->\n\n\n\n<p>Near the end of the essay, Walton considers a machine which detects incoming light and prints out a text description of the scene.<span id='easy-footnote-1-1701' class='easy-footnote-margin-adjust'><\/span><span class='easy-footnote'><a href='https:\/\/www.fecundity.com\/nfw\/2022\/12\/25\/ai-cant-let-me-see-karl-marx\/#easy-footnote-bottom-1-1701' title=' \u201cTransparent Pictures: On the Nature of Photographic Realism\u201d, &lt;em&gt;Critical Inquiry&lt;\/em&gt;, Dec 1984, 11(2): 246-277.'><sup>1<\/sup><\/a><\/span> He thinks it is obvious that reading the output of such a device would not count as <em>seeing<\/em> the scene. Because the text descriptions would not be mediated by someone else&#8217;s beliefs (the way a painting is) there must be some further feature required for transparency\u2014 a feature that photographs have but which machine-generated text descriptions lack.<\/p>\n\n\n\n<p>The feature (Walton suggests) is that the kinds of mistakes we are inclined to make with photographs are like those that we make with ordinary direct seeing, but that text descriptions lend themselves to entirely different mistakes. He writes:<\/p>\n\n\n\n<blockquote class=\"wp-block-quote is-layout-flow wp-block-quote-is-layout-flow\">\n<p>A <em>house<\/em> is easily confused with a <em>horse<\/em> or a <em>hearse<\/em>, when our information comes from a verbal description, as is a <em>cat<\/em> with a <em>cot<\/em>, a <em>madam<\/em> with a <em>madman<\/em>, <em>intellectuality<\/em> with <em>ineffectuality<\/em>, and so on. When we confront things directly or via pictures, houses are more apt to be confused with barns or woodsheds, cats with puppies, and so forth.<span id='easy-footnote-2-1701' class='easy-footnote-margin-adjust'><\/span><span class='easy-footnote'><a href='https:\/\/www.fecundity.com\/nfw\/2022\/12\/25\/ai-cant-let-me-see-karl-marx\/#easy-footnote-bottom-2-1701' title=' p. 270.'><sup>2<\/sup><\/a><\/span><\/p>\n<\/blockquote>\n\n\n\n<p>Now, a generative AI is roughly the reverse of the machine that Walton imagined. Given a text prompt, it generates a digital image. The fact that it has text at one end makes it prone to the sorts of errors that Walton highlights. It trips up on homophones, depicting fruit when prompted for an orange color scheme. And things go wrong when the prompt is not spelled correctly. To consider a specific generative AI: Midjourney reliably generates pictures that look like Karl Marx when prompted for <em>Karl Marx<\/em>, but a prompt for <em>Karl Marks<\/em> yields a street scene, a jar of ointment, a bird perched on a door knob, and this guy (below).<\/p>\n\n\n<div class=\"wp-block-image\">\n<figure class=\"aligncenter size-full\"><a href=\"https:\/\/i0.wp.com\/www.fecundity.com\/nfw\/wp-content\/uploads\/2022\/12\/20221224karlmarks.jpg?ssl=1\"><img data-recalc-dims=\"1\" loading=\"lazy\" decoding=\"async\" width=\"512\" height=\"512\" src=\"https:\/\/i0.wp.com\/www.fecundity.com\/nfw\/wp-content\/uploads\/2022\/12\/20221224karlmarks.jpg?resize=512%2C512&#038;ssl=1\" alt=\"A scruffy-looking man in glasses.\" class=\"wp-image-1704\" srcset=\"https:\/\/i0.wp.com\/www.fecundity.com\/nfw\/wp-content\/uploads\/2022\/12\/20221224karlmarks.jpg?w=512&amp;ssl=1 512w, https:\/\/i0.wp.com\/www.fecundity.com\/nfw\/wp-content\/uploads\/2022\/12\/20221224karlmarks.jpg?resize=300%2C300&amp;ssl=1 300w, https:\/\/i0.wp.com\/www.fecundity.com\/nfw\/wp-content\/uploads\/2022\/12\/20221224karlmarks.jpg?resize=150%2C150&amp;ssl=1 150w\" sizes=\"auto, (max-width: 512px) 85vw, 512px\" \/><\/a><figcaption class=\"wp-element-caption\">An image generated by Midjourney in response to the prompt &#8220;karl marks&#8221;<\/figcaption><\/figure>\n<\/div>","protected":false},"excerpt":{"rendered":"<p>Back in September, I wrote a post about generative AI and photographic transparency. The gist of it was this: Kendall Walton famously argued that I actually see Karl Marx when I look at a photograph of him, in a way I don&#8217;t when I look at a painting. The painting is mediated by the beliefs &hellip; <a href=\"https:\/\/www.fecundity.com\/nfw\/2022\/12\/25\/ai-cant-let-me-see-karl-marx\/\" class=\"more-link\">Continue reading<span class=\"screen-reader-text\"> &#8220;AI can&#8217;t let me see Karl Marx&#8221;<\/span><\/a><\/p>\n","protected":false},"author":1,"featured_media":0,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"jetpack_post_was_ever_published":false,"_jetpack_newsletter_access":"","_jetpack_dont_email_post_to_subs":false,"_jetpack_newsletter_tier_id":0,"_jetpack_memberships_contains_paywalled_content":false,"_jetpack_memberships_contains_paid_content":false,"footnotes":"","jetpack_publicize_message":"","jetpack_publicize_feature_enabled":true,"jetpack_social_post_already_shared":false,"jetpack_social_options":{"image_generator_settings":{"template":"highway","default_image_id":0,"font":"","enabled":false},"version":2}},"categories":[9],"tags":[73,30,13],"class_list":["post-1701","post","type-post","status-publish","format-standard","hentry","category-philosophy","tag-ai","tag-art","tag-teaching"],"jetpack_publicize_connections":[],"jetpack_featured_media_url":"","jetpack_sharing_enabled":true,"jetpack_shortlink":"https:\/\/wp.me\/p7PjAo-rr","jetpack_likes_enabled":false,"_links":{"self":[{"href":"https:\/\/www.fecundity.com\/nfw\/wp-json\/wp\/v2\/posts\/1701","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.fecundity.com\/nfw\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.fecundity.com\/nfw\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.fecundity.com\/nfw\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/www.fecundity.com\/nfw\/wp-json\/wp\/v2\/comments?post=1701"}],"version-history":[{"count":3,"href":"https:\/\/www.fecundity.com\/nfw\/wp-json\/wp\/v2\/posts\/1701\/revisions"}],"predecessor-version":[{"id":1705,"href":"https:\/\/www.fecundity.com\/nfw\/wp-json\/wp\/v2\/posts\/1701\/revisions\/1705"}],"wp:attachment":[{"href":"https:\/\/www.fecundity.com\/nfw\/wp-json\/wp\/v2\/media?parent=1701"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.fecundity.com\/nfw\/wp-json\/wp\/v2\/categories?post=1701"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.fecundity.com\/nfw\/wp-json\/wp\/v2\/tags?post=1701"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}