paint-brush
4 Things Humans NEED in an AI-Led Worldby@dariasup
1,060 reads
1,060 reads

4 Things Humans NEED in an AI-Led World

by Daria LeshchenkoMarch 3rd, 2023
Read on Terminal Reader
Read this story w/o Javascript
tldt arrow

Too Long; Didn't Read

The data we use to train AI today is the biggest source of bias. People often tend to insert their own bias while training insurance systems, credit scoring algorithms, or education tools. Nipping this issue in the bud will ensure the AI is trained properly and can be impartial. Cybersecurity and AI will start being even more in sync with each other as with great amounts of data comes great responsibility.

People Mentioned

Mention Thumbnail
featured image - 4 Things Humans NEED in an AI-Led World
Daria Leshchenko HackerNoon profile picture

Just half a decade ago, people couldn’t imagine that AI would become almost a mundane part of our lives. Back then, it used to be something Isaac Asimov or H. G. Wells wrote about. Today, we use it to unlock our phones, communicate with chatbots, search for the best cheesecake recipe on Google (yes, AI helps with that as well), and so on.


Now, rather than denying the importance of AI, we better prepare for an AI-led world and dive deeper into the ways we can work with it.


So, in 2023, together with AI, humans will work:

With more ethics

The data we use to train AI today is the biggest source of bias. Among others, the most popular examples of bias in AI are:



As the world moves forward, those working with AI — meaning almost all of us — need to take the ethical matter into our own hands. We have to make sure only unbiased data is used for AI training.


The first step here would be choosing the right team. In 2023, finding open-minded professionals with diverse backgrounds, meaning the team we need to properly train AI, is not that hard. After all, people often tend to insert their own bias while training insurance systems, credit scoring algorithms, or education tools. Nipping this issue in the bud will ensure the AI is trained properly and can be impartial.


The next step is defining a target audience. Understanding who will use the AI and for what purpose will help tailor it for a specific audience and for specific tasks. Conducting test runs, assembling focus groups, and improving AI’s performance will make it as ethical and, as a result, effective, as possible.


But, no matter what steps an AI team takes to make their solution as unbiased as possible, it is important to remember about the security and privacy of data accumulated for the training. And as long as we are talking about security…

With more security

Cybersecurity and AI will start being even more in sync with each other as with great amounts of data comes great responsibility… to keep them secure.


In 2022, the number of cyberattacks grew 38%. Because of this, the industry started growing and is now predicted to reach $133 billion by 2030. This makes 2023 a perfect year to start learning more about the cooperation between AI and cybersecurity and everything it can bring with it.


By capitalizing on ML algorithms, cybersecurity specialists can conduct a better loss analysis, and make the process early enough to minimize the damage that can be done. The recent rise of LLMs will inevitably lead to growth of new AI-powered solutions’, such as targeted attack analytics tools like Symantec’s Dragonfly 2.0 and Sopho’s Intercept X.


As global businesses are planning to increase their cybersecurity spending this year, some of that simply needs to be invested into the development of fraud, malicious patterns, and compliance violation detection solutions.

With more regulations

AI is a relatively new payer in all markets. As it starts gaining more power and seeping into our everyday lives, we need to start controlling and regulating it more and more.


One of the best examples of it is the 2023 EU AI Act, which “aims to provide a legal framework to ensure the safety and fundamental rights of people and businesses in relation to AI systems placed on the EU market. At the same time, the rules seek to increase AI uptake and facilitate investments and innovation.” The USA is still catching up on the plans to regulate AI, but this clearly sets a precedent for the future.


My big concern here would be that more regulations would restrict the free and rapid development of AI when it is just starting to show us its true potential. But, the truth is — with the amount of data that goes into training AI and with the importance of tasks it is performing today, we cannot afford to have it unregulated any longer. If 2023 is to become the year for AI to become more controlled, then we need to start preparing for it today. And that is a conversation for a different day.

With a higher efficiency

Right now, AI’s efficiency is the most apparent in the world of business. For example, a US financial firm Trintech has utilized bots to perform routine accounting services. As a result, they were able to service three times as many customers as they used to.


Those who have already been working with AI already know how much good it can do for us. From automating routine tasks, to writing texts and ‘painting’ pictures (yes, generative AI, I’m looking at you) to helping us gather complex and extensive data — there are almost countless possibilities for AI to keep on revolutionizing our lives in 2023 and beyond.


What does the future hold for AI? Hard to tell. But 2023 is a perfect time to keep our eyes on AI and make up new ways it can help us become more productive.