Written By: Aaron Schneider, Associate Solutions Engineer – Couchbase
Turn on your browser. You will see a vast number of Artificial Intelligence (AI) and AI-generated content—from writing, art and even music. Keying in inputs and watching your text morph into an image, sound or intelligent response can feel rather magical. To the keen observer, however, you will begin to see flaws in the replies. The images and text sometimes do not make sense. They may contain factual inaccuracies or be awkwardly phrased. Despite these limitations, AI has been highly praised on social media.
At the same time, a wave of scepticism has also surfaced. Much of it is on copyright and biased perspectives critics have uncovered. This has left the legal, social and ethical consequences of AI tools unaddressed. With such bias and errors occurring, how can users trust AI-based outcomes, especially when AI provides high levels of utility?
How Can AI Be More Trustworthy?
When the ChatGPT story broke, there were reactions to how it would first affect search engines. If users could obtain answers to queries from AI, how would companies such as Google, Bing and Yahoo respond? The potential risk of bias and factual errors could cost businesses revenue and reputation damage no matter how much the casual user trusts the AI outcomes.
A researcher from the University of California found out that ChatGPT would rank the value of a human being by gender and skin colour. These types of responses are risky, discriminatory and inflammatory. It also puts a dent in the reputation of businesses genuinely harnessing AI for good. This is not the AI’s fault to be fair as it lies solely on the inputs the technology receives. But it would behoove developers to nip bias and errors early on.
Under the Hood
At the end of the day, AI is based on Machine Learning (ML) models—with well-researched models and techniques used in creating predictive systems. Artificial intelligence systems such as ChatGPT and DALL-E need to ingest a vast volume of inputs from the internet before they could relationally match text to responses. Additionally, other tools and language models may be used to help the AI predict the words used in each response.
As such, it would be unsurprising if those AI systems produce offensive, racist or even sexist remarks. That is because the dataset used could contain potentially millions of highly illicit content. Unfiltered profanities in song lyrics or disrespectful language used on social media may have been included at some point. In such case, that would train the AI to say such phrases. This situation is not any different from teaching a toddler to speak with swear words.
It comes as no surprise that tech companies seem hesitant to release their complex AI models. That would be akin to handing over an unpredictable Pandora’s Box to users. For the business, it would be considered a highly risky manoeuvre as it has no control over the AI outputs. Any unreliable and obscene responses will be absolutely damaging to their reputation.
“To the keen observer, however, you will begin to see flaws in the replies. The images and text sometimes do not make sense. They may contain factual inaccuracies or be awkwardly phrased. Despite these limitations, AI has been highly praised on social media.”
Still, this conundrum is not something relatively new. AI bias occurs when AI models showcase the biases of their human authors through the datasets. What you give is what you get back in return. The challenge then is how businesses can reduce such biases in their product offerings and avoid the risk of putting perilous AI models into production. A successful intervention of atrocious AI models could herald a new age of information. This will, in turn, benefit technologies in various verticals from marketing tools to search engines and automation.
Built-In Limitations?
A quick study of ChatGPT tells us that OpenAI was well aware of its AI bias. And that is why it instinctively added limitations to its AI. This is as straightforward as it gets and prevents ChatGPT from making inappropriate responses. A list of prohibited keywords, phrases and guides could be put in place to prevent AI models from going erratic.
In most scenarios, this method has stopped ChatGPT from going overboard. This has not stopped AI enthusiasts from drawing out this bias through the DAN jailbreak. This only shows that underneath, AI models do suffer from bias from harmful content in its dataset. With time, OpenAI will add more guardrails and block these biases from escaping. This strategy, while successful, clearly is not foolproof, as gatekeeping content and processes is delimiting.
Rethinking an Approach for Reputation
A longer and more sustainable strategy would be rethinking what datasets are being ingested. When biases are removed before the AI could learn it, you can effectively negate the possibility of a biased AI. An attempt to ingest datasets the size of the internet will prove to be prohibitively expensive and sophisticated. Additionally, humans may have their own biases. What is offensive to one person may be harmless to another. This makes it harder to identify content in a consensual way. However, it works very well for smaller use cases.
A Preemptive Method: Removing Bias Automatically
For technology to remain a force of good, the user must be empowered to build unbiased AI models. By using a real-time event processing solution, researchers can utilise user-defined business logic to automatically remove undesired information from an AI dataset. This may require a cloud solution with a memory-first architecture that delivers unparalleled performance and makes SQL++ queries fast and highly efficient. Al projects will also benefit from the flexibility and potential of cloud NoSQL databases.
By empowering businesses and users with the ability to develop, manage and deploy data-driven business logic in a seamless environment, AI functions can be allowed to grow untethered to bias. By aggregating data from various sources, businesses can extract the diverse insights necessary to drive their business, drive AI-based Customer 360 automation or even form intelligent inventories and practice real-time logistics.
Archive
- October 2024(44)
- September 2024(94)
- August 2024(100)
- July 2024(99)
- June 2024(126)
- May 2024(155)
- April 2024(123)
- March 2024(112)
- February 2024(109)
- January 2024(95)
- December 2023(56)
- November 2023(86)
- October 2023(97)
- September 2023(89)
- August 2023(101)
- July 2023(104)
- June 2023(113)
- May 2023(103)
- April 2023(93)
- March 2023(129)
- February 2023(77)
- January 2023(91)
- December 2022(90)
- November 2022(125)
- October 2022(117)
- September 2022(137)
- August 2022(119)
- July 2022(99)
- June 2022(128)
- May 2022(112)
- April 2022(108)
- March 2022(121)
- February 2022(93)
- January 2022(110)
- December 2021(92)
- November 2021(107)
- October 2021(101)
- September 2021(81)
- August 2021(74)
- July 2021(78)
- June 2021(92)
- May 2021(67)
- April 2021(79)
- March 2021(79)
- February 2021(58)
- January 2021(55)
- December 2020(56)
- November 2020(59)
- October 2020(78)
- September 2020(72)
- August 2020(64)
- July 2020(71)
- June 2020(74)
- May 2020(50)
- April 2020(71)
- March 2020(71)
- February 2020(58)
- January 2020(62)
- December 2019(57)
- November 2019(64)
- October 2019(25)
- September 2019(24)
- August 2019(14)
- July 2019(23)
- June 2019(54)
- May 2019(82)
- April 2019(76)
- March 2019(71)
- February 2019(67)
- January 2019(75)
- December 2018(44)
- November 2018(47)
- October 2018(74)
- September 2018(54)
- August 2018(61)
- July 2018(72)
- June 2018(62)
- May 2018(62)
- April 2018(73)
- March 2018(76)
- February 2018(8)
- January 2018(7)
- December 2017(6)
- November 2017(8)
- October 2017(3)
- September 2017(4)
- August 2017(4)
- July 2017(2)
- June 2017(5)
- May 2017(6)
- April 2017(11)
- March 2017(8)
- February 2017(16)
- January 2017(10)
- December 2016(12)
- November 2016(20)
- October 2016(7)
- September 2016(102)
- August 2016(168)
- July 2016(141)
- June 2016(149)
- May 2016(117)
- April 2016(59)
- March 2016(85)
- February 2016(153)
- December 2015(150)