Page 272 - Touhpad Ai
P. 272

AI Bias in Real Life
              AI bias in real life means that the decisions made by AI systems are not always fair or accurate because they are
              influenced by biases. Just like people, AI systems can make unfair decisions if they are trained on biased data or
              programmed with biased rules.


                                     Tay was an Artificial Intelligence Chatbot that was originally launched
                                   by Microsoft Corporation on Twitter on March 23, 2016. When the bot
                                  started posting inflammatory and offensive tweets through its Twitter
                  BRAINY        account, it caused a subsequent controversy, prompting Microsoft to shut
                   FACT         down the service just 16 hours after its launch. According to Microsoft, this

                                 was caused by trolls that “attacked” the service, as the bot’s responses were based on its
                                                         interaction with people on Twitter.



                                Healthcare

                                AI model utilised for medical diagnosis can be biased if there is an over-representation of one
                                demographic group in the training data compared to another. For example, an AI system for
                                melanoma detection (a skin disease) trained primarily on fair skinned individuals may not perform
                                well for darker skinned ones, thus leading to misdiagnosis and delayed treatment.


              Education
              Drafting essays or grading standardised tests with AI can discriminate against test takers based
              on culture, dialect, or the way of writing. Failure to recognise cultural expressions and languages
              may unfairly disadvantage students from specific regions.

                                Finance

                                Biased training data or attributes may inadvertently lead some AI algorithms used in credit scoring
                                to discriminate against certain demographics. For instance, if historical loan approving data has
                                prejudice lending practices patterns, then AIs trained on such models might initiate them hence
                                making it difficult for lower-level communities to have equal access to loans.


              Criminal Justice
              If AI tools, trained on a system that had biases during the sentencing of inmates before their trial
              or parole process, then subsequently it could also be predictive of bias. For instance, because past
              cases disproportionately target people from certain races or socio-economic classes, this might
              result into categorising many people from these groups as high risks thereby leading to unfairness
              in sentencing procedures.



                 AI TASK                                                                      21 st
                                                                                             Century   #Experiential Learning
               Visit the website PortraitAI art generator (AI Gahaku) at https://ai-art.tokyo/en/   Skills
               or scan the QR code. Users upload their selfies here, and the Artificial Intelligence uses its
               understanding of Baroque and Renaissance portraits to draw their portrait in the style of a master.
               If a person is white, the result is very good. The drawback is that the most popular paintings of this
               era were of European whites. So, the database consists of mainly white people and an algorithm that
               relies on the same database tends to make you look ‘fairer’ while drawing your image.




                 270    Touchpad Artificial Intelligence - XI
   267   268   269   270   271   272   273   274   275   276   277