{"id":8131,"date":"2024-10-22T22:03:57","date_gmt":"2024-10-22T22:03:57","guid":{"rendered":"https:\/\/www.tun.com\/home\/?p=8131"},"modified":"2024-10-22T22:03:58","modified_gmt":"2024-10-22T22:03:58","slug":"increasing-ai-transparency-and-diversity-boosts-user-trust-penn-state-study-finds","status":"publish","type":"post","link":"https:\/\/www.tun.com\/home\/increasing-ai-transparency-and-diversity-boosts-user-trust-penn-state-study-finds\/","title":{"rendered":"Increasing AI Transparency and Diversity Boosts User Trust, Penn State Study Finds"},"content":{"rendered":"\n<div class=\"wp-block-group\"><div class=\"wp-block-group__inner-container is-layout-constrained wp-block-group-is-layout-constrained\">\n<div class=\"wp-block-uagb-blockquote uagb-block-e7eb3fc3 uagb-blockquote__skin-border uagb-blockquote__stack-img-none\"><blockquote class=\"uagb-blockquote\"><div class=\"uagb-blockquote__content\">Research from Penn State demonstrates that transparency about the racial diversity in AI training data can improve user trust and perceived fairness of AI systems, fostering more ethical use.<\/div><footer><div class=\"uagb-blockquote__author-wrap uagb-blockquote__author-at-left\"><\/div><\/footer><\/blockquote><\/div>\n\n\n\n<div class=\"wp-block-group is-content-justification-space-between is-nowrap is-layout-flex wp-container-core-group-is-layout-0dfbf163 wp-block-group-is-layout-flex\"><div style=\"font-size:16px;\" class=\"has-text-align-left wp-block-post-author\"><div class=\"wp-block-post-author__content\"><p class=\"wp-block-post-author__name\">The University Network<\/p><\/div><\/div>\n\n\n<div class=\"wp-block-uagb-social-share uagb-social-share__outer-wrap uagb-social-share__layout-horizontal uagb-block-ee584a31\">\n<div class=\"wp-block-uagb-social-share-child uagb-ss-repeater uagb-ss__wrapper uagb-block-ec619ce7\"><span class=\"uagb-ss__link\" data-href=\"https:\/\/www.facebook.com\/sharer.php?u=\" tabindex=\"0\" role=\"button\" aria-label=\"facebook\"><span class=\"uagb-ss__source-wrap\"><span class=\"uagb-ss__source-icon\"><svg xmlns=\"https:\/\/www.w3.org\/2000\/svg\" viewBox=\"0 0 512 512\"><path d=\"M504 256C504 119 393 8 256 8S8 119 8 256c0 123.8 90.69 226.4 209.3 245V327.7h-63V256h63v-54.64c0-62.15 37-96.48 93.67-96.48 27.14 0 55.52 4.84 55.52 4.84v61h-31.28c-30.8 0-40.41 19.12-40.41 38.73V256h68.78l-11 71.69h-57.78V501C413.3 482.4 504 379.8 504 256z\"><\/path><\/svg><\/span><\/span><\/span><\/div>\n\n\n\n<div class=\"wp-block-uagb-social-share-child uagb-ss-repeater uagb-ss__wrapper uagb-block-32d99934\"><span class=\"uagb-ss__link\" data-href=\"https:\/\/twitter.com\/share?url=\" tabindex=\"0\" role=\"button\" aria-label=\"twitter\"><span class=\"uagb-ss__source-wrap\"><span class=\"uagb-ss__source-icon\"><svg xmlns=\"https:\/\/www.w3.org\/2000\/svg\" viewBox=\"0 0 512 512\"><path d=\"M389.2 48h70.6L305.6 224.2 487 464H345L233.7 318.6 106.5 464H35.8L200.7 275.5 26.8 48H172.4L272.9 180.9 389.2 48zM364.4 421.8h39.1L151.1 88h-42L364.4 421.8z\"><\/path><\/svg><\/span><\/span><\/span><\/div>\n\n\n\n<div class=\"wp-block-uagb-social-share-child uagb-ss-repeater uagb-ss__wrapper uagb-block-1d136f14\"><span class=\"uagb-ss__link\" data-href=\"https:\/\/www.linkedin.com\/shareArticle?url=\" tabindex=\"0\" role=\"button\" aria-label=\"linkedin\"><span class=\"uagb-ss__source-wrap\"><span class=\"uagb-ss__source-icon\"><svg xmlns=\"https:\/\/www.w3.org\/2000\/svg\" viewBox=\"0 0 448 512\"><path d=\"M416 32H31.9C14.3 32 0 46.5 0 64.3v383.4C0 465.5 14.3 480 31.9 480H416c17.6 0 32-14.5 32-32.3V64.3c0-17.8-14.4-32.3-32-32.3zM135.4 416H69V202.2h66.5V416zm-33.2-243c-21.3 0-38.5-17.3-38.5-38.5S80.9 96 102.2 96c21.2 0 38.5 17.3 38.5 38.5 0 21.3-17.2 38.5-38.5 38.5zm282.1 243h-66.4V312c0-24.8-.5-56.7-34.5-56.7-34.6 0-39.9 27-39.9 54.9V416h-66.4V202.2h63.7v29.2h.9c8.9-16.8 30.6-34.5 62.9-34.5 67.2 0 79.7 44.3 79.7 101.9V416z\"><\/path><\/svg><\/span><\/span><\/span><\/div>\n<\/div>\n<\/div>\n<\/div><\/div>\n\n\n\n<p>A new study from Penn State has found that increasing transparency about the diversity of training data used in artificial intelligence (AI) systems can significantly enhance user trust and perceived fairness. The research suggests that disclosing the racial composition of AI training data and the backgrounds of the data labelers helps users make more informed decisions about whether and how to utilize these technologies.<\/p>\n\n\n\n<p>AI systems are ubiquitous in modern technology, from home assistants to advanced search engines and large language models like ChatGPT. While these systems may appear omniscient, their outputs are only as robust as the data they are trained on. Despite the critical role of training data, users often remain unaware of potential biases it may contain. <\/p>\n\n\n\n<p>The research, conducted by S. Shyam Sundar, director of the Center for Socially Responsible Artificial Intelligence at Penn State, and Cheng &#8220;Chris&#8221; Chen, a former doctoral student at Penn State who is now an assistant professor of communication design at Elon University, explores the impact of revealing this information.<\/p>\n\n\n\n<h2 class=\"wp-block-heading\">Understanding Algorithmic Bias<\/h2>\n\n\n\n<p>\u201cUsers may not realize that they could be perpetuating biased human decision-making by using certain AI systems,\u201d Sundar said in a <a href=\"https:\/\/www.psu.edu\/news\/research\/story\/showing-ai-users-diversity-training-data-boosts-perceived-fairness-and-trust\" title=\"\">news release<\/a>. <\/p>\n\n\n\n<p>Chen, the lead author, emphasized that this bias becomes evident only after a user has completed a task, making it difficult to determine trustworthiness before use. <\/p>\n\n\n\n<p>\u201cThis bias presents itself after the user has completed their task, meaning the harm has already been inflicted, so users don\u2019t have enough information to decide if they trust the AI before they use it,\u201d Chen said in the news release.<\/p>\n\n\n\n<p><\/p>\n\n\n\n<p>To test their hypothesis, the researchers created two experimental scenarios: one showcasing a diverse dataset and the other a non-diverse dataset. <\/p>\n\n\n\n<p>In the diverse condition, the participants observed an equal racial distribution of training data and labelers, while the non-diverse condition displayed a significant majority from a single racial group. <\/p>\n\n\n\n<p>The participants analyzed the performance of an AI-powered tool named HireMe, which assessed job candidates based on facial expressions and tone during automated interviews.<\/p>\n\n\n\n<h2 class=\"wp-block-heading\">Findings and Implications<\/h2>\n\n\n\n<p>The study found that exposing users to diverse training data increased their trust in AI systems. <\/p>\n\n\n\n<p>\u201cWe found that showing racial diversity in training data and labelers\u2019 backgrounds increased users\u2019 trust in the AI,\u201d Chen added. <\/p>\n\n\n\n<p>Furthermore, providing feedback opportunities enhanced participants&#8217; sense of agency and future intention to use the AI system, he explained.<\/p>\n\n\n\n<p><\/p>\n\n\n\n<p>Interestingly, the study revealed that too much feedback might reduce usability, particularly for users who perceive the system as already fair and accurate. The presence of multiple diversity cues works independently, but both data diversity and labeler diversity cues are effective in shaping users\u2019 perceptions of an AI system\u2019s fairness.<\/p>\n\n\n\n<p>Sundar highlighted the importance of representation in training data to avoid misinterpretation of emotions across different racial groups. <\/p>\n\n\n\n<p>\u201cIf AI is just learning expressions labeled mostly by people of one race, the system may misinterpret emotions of other races,\u201d he added.<\/p>\n\n\n\n<h2 class=\"wp-block-heading\">Promoting Transparent and Ethical AI<\/h2>\n\n\n\n<p>The researchers underscored that for AI systems to garner user trust, the origins of their training data should be transparent. <\/p>\n\n\n\n<p>&#8220;Making this information accessible promotes transparency and accountability of AI systems,\u201d Sundar added. \u201cEven if users don\u2019t access this information, its availability signals ethical practice and fosters fairness and trust in these systems.\u201d<\/p>\n\n\n\n<p>As the role of AI continues to expand in various sectors, ensuring that these systems are fair, transparent and trustworthy is crucial. This study by Penn State contributes a significant step toward achieving that goal.<span class=\"Apple-converted-space\">\u00a0<\/span><\/p>\n\n\n\n<p>The full findings of this research can be found in the journal <a href=\"https:\/\/www.tandfonline.com\/doi\/full\/10.1080\/07370024.2024.2392494\" title=\"\">Human-Computer Interaction<\/a>.<\/p>\n","protected":false},"excerpt":{"rendered":"<p>A new study from Penn State has found that increasing transparency about the diversity of training data used in artificial intelligence (AI) systems can significantly enhance user trust and perceived fairness. The research suggests that disclosing the racial composition of AI training data and the backgrounds of the data labelers helps users make more informed [&hellip;]<\/p>\n","protected":false},"author":3,"featured_media":0,"comment_status":"open","ping_status":"open","sticky":false,"template":"single-no-separators","format":"standard","meta":{"_acf_changed":false,"_uag_custom_page_level_css":"","_monsterinsights_skip_tracking":false,"_monsterinsights_sitenote_active":false,"_monsterinsights_sitenote_note":"","_monsterinsights_sitenote_category":0,"footnotes":""},"categories":[8],"tags":[],"class_list":["post-8131","post","type-post","status-publish","format-standard","hentry","category-ai"],"acf":[],"aioseo_notices":[],"uagb_featured_image_src":{"full":false,"thumbnail":false,"medium":false,"medium_large":false,"large":false,"1536x1536":false,"2048x2048":false},"uagb_author_info":{"display_name":"The University Network","author_link":"https:\/\/www.tun.com\/home\/author\/funky_junkie\/"},"uagb_comment_info":0,"uagb_excerpt":"A new study from Penn State has found that increasing transparency about the diversity of training data used in artificial intelligence (AI) systems can significantly enhance user trust and perceived fairness. The research suggests that disclosing the racial composition of AI training data and the backgrounds of the data labelers helps users make more informed&hellip;","_links":{"self":[{"href":"https:\/\/www.tun.com\/home\/wp-json\/wp\/v2\/posts\/8131","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.tun.com\/home\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.tun.com\/home\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.tun.com\/home\/wp-json\/wp\/v2\/users\/3"}],"replies":[{"embeddable":true,"href":"https:\/\/www.tun.com\/home\/wp-json\/wp\/v2\/comments?post=8131"}],"version-history":[{"count":12,"href":"https:\/\/www.tun.com\/home\/wp-json\/wp\/v2\/posts\/8131\/revisions"}],"predecessor-version":[{"id":8194,"href":"https:\/\/www.tun.com\/home\/wp-json\/wp\/v2\/posts\/8131\/revisions\/8194"}],"wp:attachment":[{"href":"https:\/\/www.tun.com\/home\/wp-json\/wp\/v2\/media?parent=8131"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.tun.com\/home\/wp-json\/wp\/v2\/categories?post=8131"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.tun.com\/home\/wp-json\/wp\/v2\/tags?post=8131"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}