国产三级大片在线观看-国产三级电影-国产三级电影经典在线看-国产三级电影久久久-国产三级电影免费-国产三级电影免费观看

Set as Homepage - Add to Favorites

【?? ??? ??】How are we regulating ChatGPT and other AI tools?

Source:Feature Flash Editor:hotspot Time:2025-07-03 05:19:27

ChatGPT is ?? ??? ??only two months old, but we've spent the time since it debuted debating how powerful it reallyis — and how we should regulate it. 

The artificial intelligence chatbot is being used by a significant number of people to help them with research; message people on dating apps; write code; brainstorm ideas for work, and more

Just because it can be helpful doesn't mean it can't also be harmful: Students can use it to write essaysfor them, and bad actors can use it to create malware. Even without malicious intent from users, it can generate misleading information, reflect biases, generate offensive content, store sensitive information, and — some people fear — degrade everyone's critical thinking skills due to over-reliance. Then there's the ever-present (if a bit unfounded) fear that RoBoTs ArE tAkInG oVeR.


You May Also Like

And ChatGPT can do all of that without much — if any — oversight from the U.S. government.

It's not that ChatGPT, or AI chatbots in general, are inherently bad, Nathan E. Sanders, a data scientist affiliated with the Berkman Klein Center at Harvard University, told Mashable. "In the democracy space, there are a lot of great, supportive applications for them that would help our society," Sanders said. It isn't that AI or ChatGPT shouldn't be used, but that we need to ensure it's being used responsibly. "Ideally, we want to be protecting vulnerable communities. We want to be protecting the interests of minority groups in that process so that the richest, most powerful interests are not the ones who dominate." 

Regulating something like ChatGPT is important because this kind of AI can show indifference toward individual personal rights like privacy, and bolster systematic biases with regard to race, gender, ethnicity, age, and others. We also don't know, yet, where risk and liability may reside when using the tool.

"We can harness and regulate AI to create a more utopian society or risk having an unchecked, unregulated AI push us toward a more dystopian future," Democratic California Rep. Ted Lieu wrote in a New York Timesop-ed last week. He also introduced a resolution to Congress written entirely by ChatGPT that directs the House of Representatives to support regulating AI. He used the prompt: "You are Congressman Ted Lieu. Write a comprehensive congressional resolution generally expressing support for Congress to focus on AI."

All of this adds up to a pretty unclear future for regulations on AI chatbots like ChatGPT. Some places are already placing regulations on the tool. Massachusetts State Sen. Barry Finegold penned a bill that would require companies that use AI chatbots, like ChatGPT, to conduct risk assessments and implement security measures to disclose to the government how their algorithms work. The bill would also require these tools to put a watermark on their work in order to prevent plagiarism. 

"This is such a powerful tool that there have to be regulations," Finegold told Axios.

Mashable Light Speed Want more out-of-this world tech, space and science stories? Sign up for Mashable's weekly Light Speed newsletter. By clicking Sign Me Up, you confirm you are 16+ and agree to our Terms of Use and Privacy Policy. Thanks for signing up!
SEE ALSO: Everything you need to know about ChatGPT

There are already some regulations on AIin general. The White House released an "AI Bill of Rights" that basically shows how protections that are already law — like civil rights, civil liberties, and privacy — affect AI. The EEOC is taking on AI-based hiring tools for the potential that they could discriminate against protected classes. Illinois requiresthat employers who rely on AI during the hiring process allow the government to check if the tool has a racial bias. Many states, including Vermont, Alabama, and Illinois, have commissions that work to ensure that AI is being used ethically. Colorado passed a billthat prohibits insurers from using AI that collects data that unfairly discriminates based on protected classes. And, of course, the EU is already ahead of the U.S. with regulations on AI: It passed the Artificial Intelligence Regulation Actlast December. None of these regulations are specific to ChatGPT or other AI chatbots.

While there are some state-wide regulations on AI, there isn't anything specific to chatbots like ChatGPT, neither state-wide nor nationally. The National Institute of Standards and Technology, part of the Department of Commerce, released an AI frameworkthat's supposed to give companies guidance on using, designing or deploying AI systems, but it's just that: a voluntary framework. There is no punishment for not sticking to it. Looking forward, the Federal Trade Commission appears to be creating new rulesfor companies that develop and deploy AI systems.

"Will the federal government somehow issue regulations or pass laws to oversee this stuff? I think that is highly, highly, highly unlikely," Dan Schwartz, an intellectual property partner with Nixon Peabody, told Mashable. "It is not likely you will see any federal regulation happening soon." In 2023, Schwartz predicts that the government will be looking into regulating the ownership of what ChatGPT produces. If you ask the tool to create code for you, for instance, do you own that code, or does OpenAI?

That second type of regulation — in the academia space — is likely to be private regulation. Noam Chompsky likens ChatGPT's contributions to educationas "high tech plagiarism," and when you plagiarize in school, you risk getting kicked out. That is how private regulation might work here, too.

We may run into a pretty big problem while attempting to regulate ChatGPT on the national level: AI systems can combat the very legislative regulatory system that would put them in check.

Sanders, the data scientist, explained in a piece for the New York Timesthat artificial intelligence like ChatGPT is "replacing humans in the democratic processes — not through voting, but through lobbying." That's because ChatGPT could automatically write comments and submit it in regulatory processes; write letters to submit to local newspapers and comment on news articles and post millions of social media posts every day.


Related Stories
  • Everything you need to know about ChatGPT
  • 3 things to know before talking to ChatGPT about your mental health
  • 5 of the best free online AI and ChatGPT courses available this week
  • Learn how to use AI and ChatGPT with the best free online courses on Udemy
  • 6 scary things ChatGPT has been used for already

Sanders explains to Mashable a concept called "the Red Queen's Race" in which someone — originally Lewis Caroll's Alice — exerts extreme effort only to make no forward progress. If you give an AI defensive and offensive capabilities, according to Sanders, you might get locked in a back and forth similar to a Red Queen's Race, and it could escalate out of control.

Sanders told Mashable the U.S. could potentially run into a problem: AI lobbyists trying to control the very legislation that is attempting to govern them. "It seems to me that's likely to be a losing battle for the human legislators," he said.

"My observation would be that the serious legislation that's been successfully passed for regulating machine learning in general has been painfully slow and insufficient to keep track with the progress in the field," Sanders said. "And I think it's easy to imagine that continuing into the future."

We have to be careful with how we regulate this, Sanders says, because you don't want to stifle innovation. So, you could, say, put in more roadblocks for people to submit comments to their legislators, like more captchas. But that could risk making it too difficult for regular people to engage in a democratic system.

"What I think is the most useful response is to try and encourage more democratic participation, try and encourage more humans to participate in the legislative process," Sanders said. "As AI presents challenges for scale and ubiquity, getting more people involved in the process and creating structures that allow legislators to hear from and be more responsive to real people, is a valid solution for combating that kind of threat of scale."

ChatGPT is in its infancy, and there are already plenty of ethical issues to take into account with its use. Yet, it isn't impossible to imagine a future in which sophisticated AI chatbots make our lives easier and our work better without risking the spread of misinformation and the downfall of the democratic system. It just might take a while for our government to put any meaningful regulation into action. After all, we've seen this play out before.

Topics Artificial Intelligence ChatGPT OpenAI

0.1485s , 9867.2578125 kb

Copyright © 2025 Powered by 【?? ??? ??】How are we regulating ChatGPT and other AI tools?,Feature Flash  

Sitemap

Top 主站蜘蛛池模板: 亚洲日韩精品无码久久 | 中文字幕乱码久久午夜 | 无码av中文一区二区三区桃花岛熟女电影国产狠狠免费视频 | h漫无码动漫av动漫在线播放 | 把女人弄特爽黄a大片 | 国产美女高潮流的白浆久久 | 亚洲sm另类一区二区三区 | 国产做爰又粗又大又深人物 | 国产精品99久久免费黑人人妻 | 国产精品原创视频一区二区 | 成人精品视频一区二区三区尤物 | 91精品福利 | 69无人区码一二三四区别 | 精品日韩免费播放器在线观看 | 国产丝袜免费视频网址 | 亚洲色婷婷开心综合久久一区 | 国内精品久久久久影院一蜜桃 | 纯肉巨黄H爆粗口男男分卷阅读 | 精产国品一区二区三产区 | 丰满少妇一区二区 | 亚洲精品国产一区二区在线 | 久久精品中文字幕一区 | 国产熟女一区 | 日本黄色免费网站 | av人妻精品一区二区三区 | 色综合小说久久综合图片 | 果冻传媒一二三区 | 精品综合久久久久久97超人 | 亚洲国产高清视频在线观看 | 熟女视频一区二区 | 欧美视频色偷偷 | 一本道波多野结衣一区二区 | 少妇丰满大乳被男人揉捏视频 | 国产精品免费网站 | 国产人妻人伦精品一区二区 | 麻豆国产在线 | 伊人中文字幕在线 | 欧美色综合天天久久综合精品 | 精品无码久久中文字幕 | 人妻av中文无码专区 | 久久精品国产免费中文 |