Vantage 3.0
Introducing a hybrid approach to using Document AI and GenAI
Supercharge AI automation with the power of reliable, accurate OCR
Increase straight-through document processing with data-driven insights
Integrate reliable Document AI in your automation workflows with just a few lines of code
PROCESS UNDERSTANDING
PROCESS OPTIMIZATION
Purpose-built AI for limitless automation.
Kick-start your automation with pre-trained AI extraction models.
Meet our contributors, explore assets, and more.
BY INDUSTRY
BY BUSINESS PROCESS
BY TECHNOLOGY
Build
Integrate advanced text recognition capabilities into your applications and workflows via API.
AI-ready document data for context grounded GenAI output with RAG.
Explore purpose-built AI for Intelligent Automation.
Grow
Connect with peers and experienced OCR, IDP, and AI professionals.
A distinguished title awarded to developers who demonstrate exceptional expertise in ABBYY AI.
Explore
Insights
Implementation
March 05, 2020
The potential for AI technologies to yield unintended and unfavorable consequences has sparked concern among policymakers and civil liberties groups. Regulations and guidelines surrounding the development and deployment of AI are becoming increasingly common. In order to create AI systems that are transparent, unbiased, and fair, there are a number of ethical considerations that must be evaluated.
Firstly, AI ethics requires a holistic approach. AI ethics must take into consideration not only the technology itself but also the data that is training and shaping the technology. AI is reflective of the data that powers it. If the underlying data is biased, then the resulting technology systems will likely also reflect that bias – which is especially problematic as AI becomes applied to an increasingly diverse set of use cases, including healthcare, criminal justice, and education.
The E.U. is leading the way in terms of developing standardized and enforceable guidelines for AI ethics. An expert panel commissioned by the E.U. published a set of guidelines that is designed for trustworthy AI. Furthermore, Europe’s recent GDPR also contains an article that affirms an individual’s right not to be subject to a decision solely based on automation. Researches have also noted that this article within the GDPR may pave the way for legal regulation that would require technology companies to reveal source code and algorithms.
Additionally, within the E.U., the U.K. House of Lords Select Committee on Artificial Intelligence has suggested an AI code the covers the following five principles:
While enforceable ethical guidelines surrounding AI is still nascent, the call for regulation and legal frameworks for ensuring fair and trustworthy AI is becoming louder – and even more necessary as AI becomes applied to systems that impact people’s lives in meaningful ways.
The full article, “The Ethical and Legal Challenges of AI,” can be read on the Association for Intelligent Information Management (AIIM) website. The article is part one of a three-part series, “Ethical Use of Data for Training Machine Learning Technology”.