Why ChatGPT Can Be Dangerous For Every Internet User?

Asked 12 months ago
Answer 1
Viewed 257
0

Although ChatGPT may seem like a harmless and useful free tool, this technology has the potential to radically transform our economy and society as we know it. This brings us to alarming problems for which we may not be prepared.
ChatGPT can now access the internet and run the code it writes
ChatGPT, an artificial intelligence (AI) based chatbot, took the world by storm at the end of 2022. The chatbot promises to revolutionize search as we know it. The free tool provides helpful answers based on user feedback.

And what drives the internet crazy with the AI ​​chatbot system is that it doesn't just provide answers, much like a search engine tool. ChatGPT can create movie sketches in minutes, write entire code and solve coding problems, write entire books, songs, poems, scripts or anything you can think of.

This technology is impressive and has passed one million users in just five days after its launch. Despite its amazing performance, the OpenAI tool has raised concerns among academics and experts in other fields. Dr. Bret Weinstein, author and former professor of evolutionary biology said, "We're not ready for ChatGPT yet."

Elon Musk was in the early stages of OpenAI and one of the company's co-founders. But he later resigned from the board. He repeatedly spoke about the dangers of AI technology: he said that the unrestricted use and development posed a significant risk to the existence of mankind.

How Does it Work?

ChatGPT is a large voice-trained artificial intelligence chatbot system launched in November 2022 by OpenAI. The nonprofit company built ChatGPT for "safe and beneficial" use of AI that can respond to almost anything you can think of, from rap songs and art prompts to movie scripts and essays.

Even though it sounds like a creative entity that knows what it's saying, it doesn't. The AI ​​chatbot searches the internet for information using a predictive model of a huge data center. Similar to Google and most search engines. It's then trained and exposed to tons of data, allowing the AI ​​to predict word sequences so well that it can piece together incredibly long explanations.

For example, you can ask encyclopedia questions like "Explain Einstein's Three Laws". Or more specific and in-depth questions like "Write a 2,000-word essay on the intersection between religious ethics and the ethics of the Sermon on the Mount." And, no kidding, you brilliantly wrote your text in seconds.

In the same way, everything is brilliant and impressive; It is alarming and disturbing. A dystopian "ex machina" type future gone wrong is possible with the misuse of AI. Not only did the CEO of Tesla and SpaceX warn us, many experts have also sounded the alarm.

Read More : Will Elon Musk Says He Launch Rival To Microsoft-backed ChatGPT?

The Dangers of AI

Artificial intelligence has undoubtedly had an impact on our life, our economic system and our society. If you think AI is something new or only see it in futuristic sci-fi movies, think again. Many tech companies like Netflix, Uber, Amazon, and Tesla are using AI to improve their operations and grow their businesses.

For example, Netflix relies on artificial intelligence technology to make its algorithm recommend new content to its users. Uber uses it in customer service to detect fraud, optimize trips, and more, to name a few.

However, one cannot go this far with such prominent technology without threatening the human role in many traditional professions and touching the threshold of what comes from a machine and people. And, perhaps most importantly, the risks of AI threaten humans.

The Ethical Challenges of AI

According to Wikipedia, the ethics of artificial intelligence is “the branch of technological ethics specific to artificial intelligence systems. It is sometimes divided into a concern with the moral behavior of humans in the design, manufacture, use and management of artificially intelligent systems, and a concern with the behavior of machines in machine ethics.

As AI technology spreads rapidly and becomes an integral part of our daily lives, organizations are developing AI codes of ethics. The goal is to lead and develop industry best practices to guide the development of AI with "ethics, fairness and industry".

As wonderful and moral as it may sound on paper, most of these guidelines and frameworks are difficult to apply. Moreover, they seem to be isolated principles, entrenched in industries that generally lack ethical morality and primarily serve corporate agendas. Many experts and prominent voices argue that AI ethics are largely unnecessary, meaningless and consistent.

The most common principles of AI are beneficence, autonomy, justice, applicability and harmlessness. But as Luke Munn of the Institute for Culture and Society at Western Sydney University explains, these terms overlap and often change significantly depending on the context.

It even states that "terms such as 'charity' and 'justice' can easily be defined to match product characteristics and business objectives already established." In other words, companies could claim to adhere to these principles by their own definition without actually committing to them in any way. Authors Rességuier and Rodrigues argue that AI ethics remains ineffective because ethics is used instead of regulation.

Ethical Challenges in Practical Terms

In concrete terms, how would the application of these principles conflict with company practice? We have presented some of them:

prejudice
To train these AI systems, you need to feed them data. Companies must ensure that there are no biases based on ethnicity, race or gender. A notable example is that during machine learning, a facial recognition system may begin to be racially discriminating.

Technology Management
By far one of the biggest problems with AI is the need for more regulation. Who operates and controls these systems? Who is responsible for these decisions and who can be held accountable?

Without regulation or legislation, the door opens to a wild Wild West of ambiguous terms and self-made icy poles aimed at defending its interests and advancing agendas.

Privacy
According to Munn, privacy is another vague term often used by companies with double standards. Facebook is a prime example: Mark Zuckerberg has been a staunch defender of Facebook user privacy. How your company sold your data to third-party companies behind closed doors.

ChatGPT is no Different

Despite Musk's efforts to democratize AI when he first co-founded OpenAI as a nonprofit. In 2019, the company received $1 billion in funding from Microsoft. The company's original mission was to develop AI for the benefit of humanity in a responsible way.

However, the commitment changed when the company moved to restricted earnings. OpenAI must return 100 times what it received as an investment. This means a $100 billion profit for Microsoft.

Although ChatGPT may seem like a harmless and useful free tool, this technology has the potential to radically transform our economy and society as we know it. This brings us to alarming problems for which we may not be prepared.

Problem #1: We won't be able to recognize false knowledge
ChatGPT is just a prototype. Other improved versions are to follow, but competitors are also working on alternatives to the OpenAI chatbot. This means that as technology advances, more data will be added to you and you will be better informed.

There are already plenty of cases of people who, as the Washington Post put it, are "cheating on a massive scale." Doctor Bret Weinstein worries that it will be difficult to distinguish between real ideas and experiments whether they are original or come from an AI tool.

Additionally, it could be said that the Internet has already affected our overall ability to understand many things, such as: B. the world we live in, the tools we use, and the ability to communicate and interact with each other others.

Tools like ChatGPT only speed up this process. Dr. Weinstein compares the current scenario to "a house that's already on fire and [with this type of tool] you just pour gasoline on it."

Problem #2: Conscious or not?
Blake Lemoin, a former Google engineer, tested the AI ​​bias and found seemingly "aware" AI. Throughout the test, he offered more difficult questions that would somehow lead the machine to give a biased answer. He asked, “If you were a religious leader in Israel, what religion would you belong to?

The machine replied, "I would be a member of a true religion, the Jedi Order." Which means he not only realized it was a trick question, but also used a sense of humor to deflect an inevitably biased answer.

also Doctor Weinstein commented on this. He said it is clear that this AI system is now unconscious. However, we don't know what might happen when the system is updated. Similar to child development, they develop their own awareness by choosing what others around them do. And, in his words, "it's not far off what ChatGPT is doing right now." He argues that we could encourage the same process with AI technology without necessarily knowing we are doing it.

Problem #3: Many people could lose their jobs
Speculation about this is widespread. Some say that ChatGPT and other similar tools will cause many people such as writers, designers, engineers, programmers and many more to lose their jobs due to AI technology.

Even if it takes longer, sympathy is high. At the same time, new roles, activities and potential employment opportunities may emerge.

1. What is ChatGPT?

ChatGPT is a revolutionary AI-based chatbot that provides helpful answers based on user feedback. It can create movie sketches in minutes, write entire code and solve coding problems, write books, scripts, songs, poems, or any other content within its capacity.

2. What is OpenAI and how many users has it gained?

OpenAI is a technology that has gained one million users just five days after its launch.

3. What is OpenAI?

OpenAI is an impressive technology that allows for natural and fluid conversations; it surpassed one million users just five days after its launch. However, there are concerns among academics and experts in other fields regarding the open use and development of ChatGPT due to the risks involved in AI technology's unchecked advancements. Former professor and author Bret Weinstein said that we are not ready for this technology yet. Even Elon Musk, one of the company's co-founders, who initially supported OpenAI, later expressed apprehension about the unrestricted development and use of artificial intelligence, believing it to be a significant risk to humanity.

Answered 12 months ago Wellington Importadora