#artificial intelligence hallucinations
Explore tagged Tumblr posts
Text
Florida Middle District Federal Judge suspends Florida lawyer for filing false cases created by artificial intelligence
Hello everyone and welcome to this Ethics Alert which will discuss the recent Florida Middle District Senior Judge’s Opinion and Order suspending a Florida lawyer from practicing before that court for one (1) year for filing false cases created by artificial intelligence. The case is In Re: Thomas Grant Neusom, Case No: 2:24-mc-2-JES and the March 8, 2024 Opinion and Order is here:…
View On WordPress
#artificial intelligence hallucinations#Attorney ethics#corsmeier#corsmeier lawyer ethics#ethics for lawyers#Florida Bar#Florida lawyer ethics#Florida lawyer sanctions federal court artificial intelligence#joe corsmeier#joseph corsmeier#lawyer discipline#lawyer ethics#sanctions artificial intelligence
0 notes
Text
Who could have forseen this
79 notes
·
View notes
Text
If you put garbage into an AI, you will get garbage out of an AI.
And we have put garbage into the AIs. 🤖🗑️
#glue#pizza#artificial intelligence#ai#ai overview#google ai overview#technology#garbage in garbage out#GIGO#model collapse#ai hallucinations
45 notes
·
View notes
Text
#artificial intelligence#are you hallucinating again ChatGPT?#this reminds me... I've watched 2001: a space odyssey the other day
9 notes
·
View notes
Text
Instead of getting an actual expert on AI to write this book, they just got someone who wants to suck AI's dick 🤦♀️
#allison's work life#help she thinks artificial intelligence is actually intelligent#also no mention of hallucinations or copyright infringement#the only mention of bias is a brief mention about a Black man who was wrongfully arrested and the 'cops were sad' the computer was wrong#she actially included a line about how nice it will be when AI does the mundane tasks and humans can focus on creative endeavors#I'm gonna scream#tbf this is a raw ms. no editorial eyes have seen it. the editor who acquired it quit so we'll see what the new guy does with it#it's also hella late so i wouldn't be surprised if it moved and/or was cancelled entirely#given it's going to be defunct by publication time#also fully insane: she's written several books for us about climate change yet is convinced ai is the answer to climate change??????
2 notes
·
View notes
Text
Hallucinating LLMs — How to Prevent them?
As ChatGPT and enterprise applications with Gen AI see rapid adoption, one of the common downside or gotchas commonly expressed by the GenAI (Generative AI) practitioners is to do with the concerns around the LLMs or Large Language Models producing misleading results or what are commonly called as Hallucinations.
A simple example for hallucination is when GenAI responds back with reasonable confidence, an answer that doesn’t align much with reality. With their ability to generate diverse content in text, music and multi-media, the impact of the hallucinated responses can be quite stark based on where the Gen AI results are applied.
This manifestation of hallucinations has garnered substantial interest among the GenAI users due to its potential adverse implications. One good example is the fake citations in legal cases.
Two aspects related to hallucinations are very important.
1) Understanding the underlying causes on what contributes to these hallucinations and
2) How could we be safe and develop effective strategies to be aware, if not prevent them 100%
What causes the LLMs to hallucinate?
While it is a challenge to attribute to the hallucinations to one or few definite reasons, here are few reasons why it happens:
Sparsity of the data. What could be called as the primary reason, the lack of sufficient data causes the models to respond back with incorrect answers. GenAI is only as good as the dataset it is trained on and this limitation includes scope, quality, timeframe, biases and inaccuracies. For example, GPT-4 was trained with data only till 2021 and the model tended to generalize the answers from what it has learnt with that. Perhaps, this scenario could be easier to understand in a human context, where generalizing with half-baked knowledge is very common.
The way it learns. The base methodology used to train the models are ‘Unsupervised’ or datasets that are not labelled. The models tend to pick up random patterns from the diverse text data set that was used to train them, unlike supervised models that are carefully labelled and verified.
In this context, it is very important to know how GenAI models work, which are primarily probabilistic techniques that just predicts the next token or tokens. It just doesn’t use any rational thinking to produce the next token, it just predicts the next possible token or word.
Missing feedback loop. LLMs don’t have a real-time feedback loop to correct from mistakes or regenerate automatically. Also, the model architecture has a fixed-length context or to a very finite set of tokens at any point in time.
What could be some of the effective strategies against hallucinations?
While there is no easy way to guarantee that the LLMs will never hallucinate, you can adopt some effective techniques to reduce them to a major extent.
Domain specific knowledge base. Limit the content to a particular domain related to an industry or a knowledge space. Most of the enterprise implementations are this way and there is very little need to replicate or build something that is closer to a ChatGPT or BARD that can answer questions across any diverse topic on the planet. Keeping it domain-specific also helps us reduce the chances of hallucination by carefully refining the content.
Usage of RAG Models. This is a very common technique used in many enterprise implementations of GenAI. At purpleSlate we do this for all the use cases, starting with knowledge base sourced from PDFs, websites, share point or wikis or even documents. You are basically create content vectors, chunking them and passing it on to a selected LLM to generate the response.
In addition, we also follow a weighted approach to help the model pick topics of most relevance in the response generation process.
Pair them with humans. Always. As a principle AI and more specifically GenAI are here to augment human capabilities, improve productivity and provide efficiency gains. In scenarios where the AI response is customer or business critical, have a human validate or enhance the response.
While there are several easy ways to mitigate and almost completely remove hallucinations if you are working in the Enterprise context, the most profound method could be this.
Unlike a much desired human trait around humility, the GenAI models are not built to say ‘I don’t know’. Sometimes you feel it was as simple as that. Instead they produce the most likely response based on the training data, even if there is a chance of being factually incorrect.
Bottomline, the opportunities with Gen AI are real. And, given the way Gen AI is making its presence felt in diverse fields, it makes it even more important for us to understand the possible downsides.
Knowing that the Gen AI models can hallucinate, trying to understand the reasons for hallucination and some reasonable ways to mitigate those are key to derive success. Knowing the limitations and having sufficient guard rails is paramount to improve trust and reliability of the Gen AI results.
This blog was originally published in: https://www.purpleslate.com/hallucinating-llms-how-to-prevent-them/
2 notes
·
View notes
Text
would produce more senseless poetry like the artificial intelligence so hated by the world but seeming as the words are nonsense in your ears he, the poet, should find it more insightful to violently slam his head against the keyboard
#random thoughts#me? him? all the same all the same#repetition repetition does it help? does it do something? certainly not#unwell? unwell? unwell? unwell? unwell? it's catharsis#neither old enough nor intelligent enough to know what any words mean. my intelligence is artificial#come to think of it. i am the words. why all can any of you tolerate this thing when he is so painfully robotic#emanating human behaviorisms despite paling in comparison to the beauty of one. i've questioned my reality more than once#i think it is safe to say that i do not exist. this is not a problem for any of you#a mass hallucination as it were. i'd like to wake up now#i am asking you to let me wake up now. i'd like to wake up now. shake my senses so i can become human#not this not this. what is this?? what is he?? what is? he? not him#no. not him. himself. himself/myself#a means to an end? whatever does the phrase mean#vivisection of a butterfly
2 notes
·
View notes
Text
#artificial intelligence risks#artificial intelligence benefits#666#end time prophecy#AI hallucinations#control buying and selling#Revelation 13#bible prophecy#news & analysis#news & prophecy#bible prophecy interview
0 notes
Text
"V for Vendetta" - Alan Moore and David Lloyd
#book quote#v for vendetta#alan moore#david lloyd#i love you#computer#artificial intelligence#fluke#circuits#tricks#going mad#hallucinations
1 note
·
View note
Text
AI hallucinations gone wrong as Alaska uses fake stats in policy
New Post has been published on https://thedigitalinsider.com/ai-hallucinations-gone-wrong-as-alaska-uses-fake-stats-in-policy/
AI hallucinations gone wrong as Alaska uses fake stats in policy
.pp-multiple-authors-boxes-wrapper display:none; img width:100%;
The combination of artificial intelligence and policymaking can occasionally have unforeseen repercussions, as seen recently in Alaska.
In an unusual turn of events, Alaska legislators reportedly used AI-generated citations that were inaccurate to justify a proposed policy banning cellphones in schools. As reported by /The Alaska Beacon/, Alaska’s Department of Education and Early Development (DEED) presented a policy draft containing references to academic studies that simply did not exist.
The situation arose when Alaska’s Education Commissioner, Deena Bishop, used generative AI to draft the cellphone policy. The document produced by the AI included supposed scholarly references that were neither verified nor accurate, yet the document did not disclose the use of AI in its preparation. Some of the AI-generated content reached the Alaska State Board of Education and Early Development before it could be reviewed, potentially influencing board discussions.
Commissioner Bishop later claimed that AI was used only to “create citations” for an initial draft and asserted that she corrected the errors before the meeting by sending updated citations to board members. However, AI “hallucinations”—fabricated information generated when AI attempts to create plausible yet unverified content—were still present in the final document that was voted on by the board.
The final resolution, published on DEED’s website, directs the department to establish a model policy for cellphone restrictions in schools. Unfortunately, the document included six citations, four of which seemed to be from respected scientific journals. However, the references were entirely made up, with URLs that led to unrelated content. The incident shows the risks of using AI-generated data without proper human verification, especially when making policy rulings.
Alaska’s case is not one of a kind. AI hallucinations are increasingly common in a variety of professional sectors. For example, some legal professionals have faced consequences for using AI-generated, fictitious case citations in court. Similarly, academic papers created using AI have included distorted data and fake sources, presenting serious credibility concerns. When left unchecked, generative AI algorithms, which are meant to produce content based on patterns rather than factual accuracy, can easily produce misleading citations.
The reliance on AI-generated data in policymaking, particularly in education, carries significant risks. When policies are developed based on fabricated information, they may misallocate resources and potentially harm students. For instance, a policy restricting cellphone use based on fabricated data may divert attention from more effective, evidence-based interventions that could genuinely benefit students.
Furthermore, using unverified AI data can erode public trust in both the policymaking process and AI technology itself. Such incidents underscore the importance of fact-checking, transparency, and caution when using AI in sensitive decision-making areas, especially in education, where impact on students can be profound.
Alaska officials attempted to downplay the situation, referring to the fabricated citations as “placeholders” intended for later correction. However, the document with the “placeholders” was still presented to the board and used as the basis for a vote, underscoring the need for rigorous oversight when using AI.
(Photo by Hartono Creative Studio)
See also: Anthropic urges AI regulation to avoid catastrophes
Want to learn more about AI and big data from industry leaders? Check out AI & Big Data Expo taking place in Amsterdam, California, and London. The comprehensive event is co-located with other leading events including Intelligent Automation Conference, BlockX, Digital Transformation Week, and Cyber Security & Cloud Expo.
Explore other upcoming enterprise technology events and webinars powered by TechForge here.
Tags: ai, law, policy, research
#ai#ai & big data expo#AI hallucinations#AI regulation#ai-generated content#alaska#Algorithms#amp#anthropic#applications#artificial#Artificial Intelligence#attention#automation#background#Big Data#board#california#cellphones#chatbots#Cloud#cloud computing#comprehensive#computing#conference#content#court#cyber#cyber security#cybersecurity
0 notes
Video
A Real Hallucination by Davivid Rose Via Flickr: According to an image generated by Adobe's AI, the above image is what illegal LSD looks like in a laboratory right after it has been made... Please click here to read my "autobiography": thewordsofjdyf333.blogspot.com/ My telephone number is: 510-260-9695
0 notes
Text
ai hallucinations
#https://www.opensesame.dev/#ai hallucinations#artificial intelligence problem#ai models#ai developer#ai detection tool
1 note
·
View note
Text
What are Grounding and Hallucinations in AI? - Bionic
This Blog was Originally Published at :
What are Grounding and Hallucinations in AI? — Bionic
The evolution of AI and its efficient integration with businesses worldwide have made AI the need of the hour. However, the problem of AI hallucination still plagues generative AI applications and traditional AI models. As a result, AI organizations are constantly pursuing better AI grounding techniques to minimize instances of AI hallucination.
To understand AI hallucinations, imagine if someday your AI system starts showing glue as a solution to make cheese stick to pizza better. Or, maybe your AI fraud detection system suddenly labels a transaction fraud even when it is not. Weird, right? This is called AI hallucination.
AI Hallucination occurs when the AI systems generate outputs that are not based on the input or real-world information. These false facts or fabricated information can undermine the reliability of AI applications. This can seriously harm a business’s credibility.
On the other hand, Grounding AI keeps the accuracy and trustworthiness of the data intact. You can define Grounding AI as the process of rooting the AI system’s responses in relevant, real-world data.
We will explore what are grounding and hallucinations in AI in this detailed blog. We will explore the complexities of AI systems and how techniques like AI Grounding can help minimize it, ensuring reliability and accuracy.
What is AI Hallucination and how does it occur?
AI Hallucination refers to the instances when AI outputs are not based on the input data or real-world information. It can manifest as fabricated facts, incorrect details, or nonsensical information.
It can especially happen in Natural Language Processing (NLP) such as Large Language Models and image generation AI models. In short, AI hallucination occurs when the AI generative models generate data or output that looks plausible but lacks a factual basis. This can lead to incorrect results.
(Image Courtesy: https://www.nytimes.com/2023/05/01/business/ai-chatbots-hallucination.html)
What causes AI Hallucination?
When a user gives a prompt to an AI assistant, its goal is to understand the context of the prompt and generate a plausible result. However, if the AI starts blurting out fabricated information, it becomes a case of AI hallucination concluding that the AI model is not trained in that particular context and lacks background information.
Overfitting: Overfitting refers to training the AI model too closely on its training data, making the AI model overly specialized. This can result in the narrowing of the horizon of knowledge and context. As a result, the AI model doesn’t generate desirable output for new, unseen data. This overfitting can cause AI hallucinations when it is faced with user input outside of the model’s training data.
Biased Training Data: AI systems are as good as the data they are trained on. If this training data contains biases or prejudiced inaccuracies, the AI may reflect these biases as its output. This can lead to AI hallucinations, making the information incorrect.
Unspecific or Suggestive Prompts: Sometimes, your prompt may not have clear constraints and specific details. The AI will have to make up its irrelevant interpretation of the input based on its training data. This in turn increases the likelihood of getting fake information.
Asking about Fictional Context: Prompts that are associated with fictional subjects related to products, people, or even situations are likely to trigger hallucinations. This may be due to a lack of reference facts for an AI interface to draw information from.
Incomplete Training Data: When training data does not entail full coverage of the situations that an AI might find itself in, the system is likely to come up with wrong outputs. This results in hallucinations as the system tries to make up for the missing data.
Types of AI Hallucinations
AI hallucinations can be broadly categorized into three types:
Visual Hallucinations: These occur in AI systems that are used in image recognition, or image generation systems. The AI system generates erroneous design outputs or graphical inaccuracies. For instance, the AI may produce an image of an object that does not exist or fail to recognize the given objects present in a particular image.
Pictorial Hallucinations: They are somewhat similar to visual hallucinations, but they refer to the erroneous output of visual information. This could include graphical data like simple drawings, diagrams, infographics, etc.
Written Hallucinations: When it comes to NLP models, hallucinations are defined as text that contains information not included in the input data. These can be false facts, extra details, or statements not supported by the input data. This can occur in popular chatbots, auto-generated reports, or any AI that creates text material, etc.
Real-Life Examples of AI Hallucination
Below are some real-life examples of AI Hallucinations that made waves:
Glue on Pizza: A prominent AI hallucination happened when Google’s AI suggested that the cheese would not slide when using glue on pizza. This weird suggestion served to illustrate the system’s potential to produce harmful and illogical advice. Misleading users in this way can have serious safety implications. This is why close monitoring of AI and validation of facts is important.(Know More)
Back Door of Camera: Just about a month ago, there was an AI hallucination in which Google’s Gemini AI suggested “open the back door” of a camera as a photographic tip. However, it showed this result in a list of “Things you can try,” illustrating the harm of irresponsible directions coming from AI systems. These errors can lead to incorrect conclusions by the users, and could potentially cause damage to the equipment. (Know More)
Muslim Former President Misinformation: There was a false claim in Google’s AI search overview that Former President Barack Obama is a Muslim. Another error made by an AI during searches executed through Google stated that none of Africa’s 54 recognized nations begins with the letter ‘K’ forgetting Kenya. This occurrence demonstrated the danger of machine learning systems being used to disseminate wrong ideas. This also highlights the lack of basic factual information in AI systems. (Know More)
False Implications on Whistleblower: Brian Hood, Australian politician and current mayor of Hepburn Shire, was wrongly implicated in a bribery scandal by ChatGPT. The AI falsely identified Hood as one of the people involved in the case intimating that he had bribed authorities and served a jail term for it. Hood, however, was a whistleblower in that case. AI Hallucination incidents can lead to legal matters of defamation. (Know More)
These kinds of hallucinations in image classification systems can have very grave social and ethical consequences.
Why are AI Hallucinations not good for your business?
Apart from just being potentially harmful to your reputation, AI hallucinations can have detrimental effects on businesses including:
Eroded Trust: Consumers and clients will not rely on an AI system if it constantly comes up with wrong or fake information. This erosion weakens user confidence thus affecting their usage or interaction with the AI deployed. Once the trust in your business is breached, it becomes very difficult to maintain customer retention and brand loyalty.
Operational Risks: Erroneous information from AI systems can contribute to wrong decisions, subpar performance, and massive losses. For instance, if applied in the supply chain setting, an AI hallucination could lead to inaccurate inventory forecasting. This, in turn, leads to costs associated with either overstock or stock out. In addition, AI can give poor recommendations that interfere with organized workflow. This could require someone to fix what the AI got wrong.
Legal and Ethical Concerns: Legal risks due to AI could arise when hallucinations by the system result in a negative impact. For example, if a financial AI system provides erroneous recommendations on investments, it could cause significant financial losses, and thus, lead to legal proceedings. Ethical issues come up, especially when the outputs generated by an AI system are prejudiced or unfair in some way.
Reputational Damage: AI hallucinations are particularly dangerous and can lead to the loss of the reputation of a firm in the market. People’s opinions can be easily influenced negatively as seen in social media and leading news channels. Such reputational damage can lead to rejection by potential clients and partners. This could cause significant challenges for the business to attract and sustain opportunities.
Understanding AI Grounding
We can define Grounding AI as the process of grounding the AI systems in real data and facts. This involves aligning the AI’s response and behavior to factual data and information. Grounding AI is particularly helpful in Large Language Models. This helps minimize or eradicate instances of hallucinations as the information fed to the AI will be based on real data and facts.
Bridging Abstract AI Concepts and Practical Outcomes
Grounding AI can be seen as the connection between the theoretical and at times, highly abstract frameworks of AI and their real-world implementations.
(Image Courtesy: https://cloud.google.com/vertex-ai/generative-ai/docs/grounding/overview)
The Importance of Grounding AI
Grounding AI is essential for several reasons:
Accuracy and Reliability: AI systems that are grounded in real-time data feeds are likely to generate more accurate and reliable results. This can especially be helpful in business strategy, healthcare delivery, finance, and many other fields.
Trust and Acceptance: When the AI systems are grounded in real-life data, consumers are more inclined to accept the results of the systems. This makes the integration process easier.
Ethical and Legal Compliance: One of the reasons why grounding is important is to reduce cases where AI is used to propagate fake news. The propagation of these fake news causes harm, raising ethical and legal concerns.
The Best Practices for Grounding AI
Various best practices can be employed to ground AI systems effectively:
Data Augmentation: Improving the training datasets to incorporate more data that are similar to the inputs the model is expected to process.
Cross-Validation: Verifying the results generated by AI systems with one or more data sets, to check for coherence and correctness.
Domain Expertise Integration: Engagement of experts from the particular domain for the development of the AI system as well as to ensure the correctness of the output.
Feedback Loops: Incorporation of feedback and AI reinforcement learning process coming from the evaluation parameters and feedback received from users.
Implement Rigorous Validation Processes: Using cross-validation techniques and other reliable validation procedures to ensure the validity of the AI model.
Utilize Human-in-the-Loop Approaches: Introducing humans in the loop that check and review outputs produced by the AI tool, especially in sensitive matters.
Bionic uses Human in the Loop Approach and gets your content and its claims as well as facts validated. Request a demo now!
Benefits of Grounding AI
Grounding AI systems offers several significant benefits:
Increased Accuracy: Calibrating real data with AI output increases the accuracy of those outputs.
Enhanced Trust: Grounded AI systems foster more trust from users and stakeholders because they provide more accurate results.
Reduced Bias: Training a grounded AI model on diverse data reduces biases and creates more ethical AI systems.
Improved Decision-Making: Businesses can tremendously improve their organizational decision-making by using reliable grounded AI outputs.
Greater Accountability: Implementing grounded AI systems allows better monitoring and verification of outputs, thereby increasing accountability.
Ethical Compliance: Ensuring that AI reflects actual data about the world helps maintain ethical standards and prevent hallucination.
The Interplay Between Grounding AI and AI Hallucinations
Grounding AI is inversely related to hallucinations in AI because it filters out irrelevant or inaccurate content. It ensures that AI-generated content does not contain hallucinations. Conversely, a lack of grounding may cause AI hallucinations because the outputs will not be aligned with real-world applications.
Challenges in Achieving Effective AI Grounding
Achieving effective AI grounding to prevent hallucinations in AI systems presents several challenges:
Complexity of Real-World Data: Real-world data, often disorganized, understructured, and inconsistent, is difficult to acquire and assimilate into AI systems comprehensibly. Ensuring grounding AI with such information is challenging.
Dynamic Environments: AI systems usually operate in unpredictable and volatile environments. Maintaining AI generative models in these scenarios requires constant AI reinforcement learning and real-time data updates, posing technical hurdles and high costs.
Scalability: Grounding vast and complex AI systems is challenging, especially on a larger scale. Monitoring and maintaining grounding effects in different models and applications demands significant effort.
The Future of AI Grounding and AI Hallucinations
The future of grounding and hallucinations in AI looks promising, with several key trends and breakthroughs anticipated:
Advancements in Data Quality and Integration: Advancements in data collection, cleaning, and integration will improve AI grounding. Better data acquisition will train AI models with diverse and sufficiently large datasets to minimize hallucinations.
Enhanced Real-Time Data Processing: AI systems will have more real-time data feeds from various sources, grounding the systems on current and accurate data. This will enable AI models to learn in changing conditions and minimize hallucinated outputs.
Human-AI Collaboration: The prominence of augmented intelligence, where humans validate AI-generated outputs, will increase. AI models like Bionic AI will combine human brain capabilities with AI to obtain accurate facts.
Mitigating AI Hallucination with Bionic
Bionic AI is designed to handle multi-level cognitive scenarios, including complex real-world cases by constant AI reinforcement learning and bias reduction. Duly updated by real-world data and human supervision, Bionic AI safeguards itself from overfitting to remain as flexible and adaptable (to the real world) as can be.
Bionic AI combines AI with human inputs to eliminate contextual misinterpretation. Effective AI grounding techniques and a human-in-the-loop approach empower Bionic AI with specific and relevant information. This seamless integration of AI and human oversight makes Bionic AI change the game of business outsourcing.
Bionic AI adapts to changing human feedback making it hallucination-free and effective in dynamic environments. By mixing AI with human oversight, Bionic promises accurate and relevant results that foster customer satisfaction and trust. This synergy ensures that customer concerns with traditional AI are addressed justly, delivering outstanding customer experience.
Conclusion
With the increasing adoption of AI in businesses, it is crucial to make these systems trustworthy and dependable. This trust is kept intact by grounding AI systems in real-world data. The costs of AI hallucinations are staggering, due to instances such as wrong fraud alerts, and misdiagnosis of healthcare problems among others. This can result from factors such as overfitting, training datasets, and incomplete training sets.
Knowing what is grounding and Hallucinations in AI can take your business a long way ahead. Mechanisms such as data augmentation, cross-validation, and using human feedback help the implementation of effective grounding.
Bionic AI uses artificial intelligence and human oversight to fill gaps regarding biases, overfitting, and contextual accuracy. Bionic AI is your solution for accurate and factual AI outputs, letting you realize the full potential of AI.
Ready to revolutionize your business with AI that’s both intelligent and reliable? Explore how Bionic can transform your operations by combining AI with human expertise. Request a demo now! Take the first step towards a more efficient, trustworthy, and ‘humanly’ AI.
0 notes
Text
There is the right way, the wrong way, and the Google AI Overview way that will probably kill or maim you.
3 notes
·
View notes
Text
A Lawyer’s Guide to Understanding AI Hallucinations in a Closed System
Understanding Artificial Intelligence (AI) and the possibility of hallucinations in a closed system is necessary for the use of any such technology by a lawyer. AI has made significant strides in recent years, demonstrating remarkable capabilities in various fields, from natural language processing to large language models to generative AI. Despite these advancements, AI systems can sometimes…
View On WordPress
#AI#AI tools#Artificial Intelligence#Closed systems#court filings#Extrapolation Error#fabricated case details#false medical analysis#generative AI#hallucination#Internet#large language models#lawyer#natural language processing#overfitting#prompt-based data searches#safeguard#Technology
0 notes
Text
Fair Warning: AI can neither count, or do logic.
In short, this article:
To elaborate: when a problem has ONE correct answer, whether it's basic math or simple logic, AI still fails often and spits out over-hyped, combative answers. It's as if the things were trained in India or by trolls on the internet, or both.
But mainly, it's a Garbage In, Garbage Out issue. When you have legitimate software, even if it has algorithms, it's coded with verified hard data and known answers, and the means to get them. An actual word processing program has a dictionary file in it. AI and its apps? Not so much--they're all probability pushers, programs meant to operate inside of a Bell Curve, or a box of what "seems legit."
And to use an historical example of how wrong probability-based rationality can get? Galileo and Copernicus both were punished by the Catholic Churches of their day, for suggesting that the evidence supported a Round Earth that went Around the Sun at the Center of the Solar System. Respectively. That's the exact kind of error AI pushes: if it sees a thing happen often enough, it will assume the thing is a fact even if it's satire, a meme, or just plain wrong.
#AI#artificial intelligence#AI errors#AI hallucinations#logic#basic math#Yahoo! link#garbage in garbage out
0 notes