Technology
Ethics In The Time of Artificial Intelligence (and their sentient attorneys)
By Matt Heisie, co-founder and head of product marketing of Ferret, an AI platform that provides critical intelligence to empower companies and individuals.
A Google engineer stated earlier this year that the company’s AI bot known as LaMDA not only had achieved sentience but perhaps in the most obvious sign of modern human intelligence, it (the bot) also had hired its own attorney. Perhaps it’s time we start thinking seriously about AI and ethics.
Actually, that’s pretty much what I’ve been doing for the past two years with my executive and engineering teams at Ferret – a new app that provides relationship intelligence about your contacts in real-time – as we tackle and, yes, at times struggle with question of ethics and AI. The burning question: How much do we reveal of someone’s past?
Let me provide a concrete example of our concerns: We recently researched an individual whose reputation was sterling … except for a misdemeanor arrest for marijuana possession when he was 18 (technically, an adult). The amount of marijuana that he had at the time is no longer illegal in the state where the “crime” took place and where they were arrested and continue to live. Include that arrest in their profile – yes or no?
AI ethics involves more than high-concept philosophies and futuristic dystopian fiction. It’s about the systems we interface with every day—and how the decisions made might impact our lives. From your home to your office, AI already imbues virtually every aspect of your life through facial recognition software, ad-blocking technology, smart home devices, online retail algorithms, search engines, relationship tracking, streaming entertainment and video game development. Yes, it’s here now and it’s everywhere.
At the heart of the question of AI ethics is an elegantly simple directive by UNESCO in its first ever global recommendations on AI ethics : “We need a human-centered AI. AI must be for the greater interest of the people, not the other way around.”
Potential Problems with AI
AI is the field of thinking machines. It’s often thought of as teaching machines to think like humans—ideally, though, AI can deliver more rational, relevant, extensive, and accurate information than the average person.
AI brings us relevant content and faster-than-ever information aggregation—but at what cost?
The innate objectivity of machines should protect us from human fallacies like cognitive bias and mismanagement, right? Well, that’s not exactly possible when you consider the fact that all AI is built by humans. If we’re not careful, individual biases can be baked into the algorithms created to operate the AI engines.
The biggest threat to ethical AI technology is a lack of foresight and caution. The only way to combat this is by examining, updating, and re-examining the product from every angle, ensuring it operates as intended for every user.
Ethical Concerns in AI
Our morals, principles, and values are at the heart of ethical AI development. In the same way that morals and ethics are developed by individuals, societies, and cultures, so too they are adapted to regulate AI. According to the Alan Turing Institute, “AI ethics is a set of values, principles, and techniques that employ widely accepted standards of right and wrong to guide moral conduct in the development and use of AI technologies.”
As the field of artificial intelligence ethics develops, there are a few central issues pushing their way to the forefront, including:
- Privacy and surveillance – Since the advent of the internet and cellular communication, personal privacy has been an ongoing global issue. Much of digital data remains unregulated and large data companies unaccountable.
- Bias and discrimination – While AI has the capacity to transcend the individual biases of humans, it also has the potential to follow the same biases as its creators. Because AI systems learn from human behaviors, they may replicate dangerous and unethical discriminatory practices. Even more concerning, the perceived impartiality of AI can make these biases seem credible or factual.
How to Develop Ethical AI
While we can analyze the potential risks and dangers associated with unethical systems, the factors necessary to build an ethical system are not as immediately apparent.
A truly ethical AI system will consider issues such as:
- Transparency – Provide users and stakeholders with clear and easily accessible insight into where public data comes from, how it’s analyzed, and how the AI makes its decisions.
- Impartiality – Eliminating bias and discriminatory practices are key when determining how to develop an ethical AI. An AI must become more impartial than a human. That means being very careful about what goes into the system, what trains the system, because biases present on the data in will be replicated and amplified on the analysis that comes out the other end.
- Design – Context is key. Think of the marijuana conviction discussed in the opening: it may not be ethical to hide a true article about someone’s past conviction, but it is ethical to add context about the following legalization, about biases in charging or convictions for youth offenders, or to pair the conviction with an article about a conviction being vacated. Design can reduce the bias that puts the inflammatory accusation on page 1, and the charged being thrown out in a footnote on page 32.
- Accountability – Allow third-party inspections and answer questions about how conclusions are reached.
- Security – AI systems must be secured from data breaches, hacks, and potential risks and threats both internally and externally.
- Regulated – AI systems should follow and abide by all known laws and regulations that have been established.
I suspect that the Google engineer who thought he witnessed the birth of AI sentience was merely looking at contextual pattern recognition – a decades old technology, albeit increasingly getting better in understanding its environment (physical and virtual) and reflecting the sentiments of its users. There’s no need at the moment to plan for robot apocalypse; AI can’t kill you. But make no mistake, it most certainly can kill your reputation. Ultimately, the answer to whether AI can be ethical is for us to decide.
-
Top Stories2 days ago
After VW plant victory, UAW sets its sights on Mercedes in Alabama
-
Investing2 days ago
Forex Market Trends to Watch Out For in 2024
-
Top Stories2 days ago
Hedge fund borrowing hits five-year peak, Goldman Sachs says
-
Business2 days ago
Mike Bahun and Fundraising University Make a Lasting Impact on Sports Programs Nationwide