Latest News and Comment from Education

Wednesday, July 12, 2023

AI AND CHATGPT IN THE CLASSROOM: DANGER DANGER WILL ROBINSON

AI AND CHATGPT IN THE CLASSROOM

DANGER DANGER WILL ROBINSON

As we enter a new era of education, it's important to consider the potential dangers of AI in the classroom. Sure, it sounds cool to have a robot teacher who never gets tired or loses patience, but what happens when that robot becomes self-aware and decides to take over the world? Okay, maybe that's a bit extreme, but there are still some serious concerns to address.

First and foremost, we need to make sure our students are AI literate. And no, that doesn't mean they need to be able to program a robot from scratch. It means they need to understand the basics of how AI works and how it affects their lives. We don't want our students thinking that Siri is some kind of magical genie who lives inside their phone and grants wishes.

We also need to promote AI ethics and values. Just because a machine can do something doesn't mean it should. We need to teach our students to reflect on the ethical and social implications of AI, such as privacy, fairness, accountability, and human dignity. We don't want our students developing a sense of moral relativism where anything goes as long as it's done by a computer.

Of course, protecting student data and privacy is also a top priority. We don't want our students' personal information falling into the wrong hands. And let's face it, sometimes even the right hands can be a little sketchy. We need to make sure the AI systems we use in the classroom are compliant with data protection laws and regulations, and that they respect the rights and interests of students.

But perhaps most importantly, we need to monitor and report any potential risks or harms caused by AI systems. We don't want our students falling victim to cyberattacks, discrimination, bias, or misinformation. And if something does go wrong, we need to seek guidance and support from experts or authorities. Because let's be real, none of us are equipped to handle a rogue AI on our own.

So what can we do to prepare ourselves for this brave new world of AI in education? Luckily, the Big Education Ape blog has provided us with some great resources for teachers. The articles "What Every Teacher Needs to Know About ChatGPT Before the New School Year" is a must-read for anyone looking to incorporate AI into their curriculum. And their guide "CHATGPT: A Guide for Students and Teachers Using ChatGPT in the Classroom" is a great resource for anyone looking to get started with AI chatbots.

But let's not forget about the bigger picture here. AI safety and alignment are important topics for the future of artificial intelligence. We need to make sure we're building safe AI systems that don't harm humans or the environment. And we need to ensure that these systems are developed and deployed in a way that aligns with human values and goals.

Of course, there are some challenges that come with this. How do we instill complex values in AI systems that may not share our moral intuitions or preferences? How do we prevent emergent AI behaviors that may be harmful or undesirable? And how do we balance the trade-offs between efficiency, fairness, privacy, and accountability in AI systems?

Thankfully, there are many researchers and organizations working on these problems. And they're using some pretty cool methods like direct normativity, indirect normativity, cooperative inverse reinforcement learning, iterated amplification, and more. I have no idea what any of those mean, but they sound impressive.

So what are some of the dangers of AI? Well, there's the invasion of personal data, the risk of cyberattack, discrimination and bias, opacity and lack of transparency, accountability of AI-driven decisions, and replacement of jobs. But don't worry, there are plenty of groups working on AI safety and alignment. There are strategy groups, governance groups, technical groups, and movement building groups.

And there are some pretty awesome organizations doing this work too. The Future of Humanity Institute? The Center for Human-Compatible AI? The Partnership on AI? These sound like superhero teams! And let's not forget about DeepMind or the Center for AI Safety. These guys are like the Avengers of AI.

So while there are certainly some concerns when it comes to AI in education, there's also a lot of potential for good. We just need to make sure we're doing it right. And if all else fails, we can always call in Iron Man to save the day.


THE DEFT DIVE


Big Education Ape: WHAT EVERY TEACHER NEEDS TO KNOW ABOUT CHATGPT BEFORE THE NEW SCHOOL YEAR https://bigeducationape.blogspot.com/2023/06/what-every-teacher-needs-to-know-about.html 

Big Education Ape: CHATGPT: A GUIDE FOR STUDENTS AND TEACHER USING CHATGPT IN THE CLASSROOM https://bigeducationape.blogspot.com/2023/06/chatgpt-guide-for-students-and-teacher.html 

Educators can help keep AI safe for students by:

  • - **Teaching AI literacy and skills**: Educators can help students understand the basics of AI, how it works, and how it affects their lives. They can also help students develop critical thinking and problem-solving skills that are essential for using and creating AI systems responsibly¹².
  • - **Promoting AI ethics and values**: Educators can help students reflect on the ethical and social implications of AI, such as privacy, fairness, accountability, and human dignity. They can also help students develop a sense of civic responsibility and moral agency when interacting with or developing AI systems¹².
  • - **Protecting student data and privacy**: Educators can help students protect their personal data and privacy when using AI-enabled devices or platforms. They can also ensure that the AI systems they use in the classroom are compliant with data protection laws and regulations, and that they respect the rights and interests of students¹³.
  • - **Monitoring and reporting AI risks and harms**: Educators can help students identify and report any potential risks or harms caused by AI systems, such as cyberattacks, discrimination, bias, or misinformation. They can also seek guidance and support from experts or authorities when dealing with AI-related issues or incidents¹³.

Bing, 7/12/2023

(1) Our children are growing up with AI: what you need to know. https://www.weforum.org/agenda/2022/01/artificial-intelligence-children-technology/.

(2) Stanford AI Safety. http://aisafety.stanford.edu/.

(3) AI Detection In Schools: Key To Student Safety. https://elearningindustry.com/ai-detection-in-schools-key-to-student-safety.

(4) Berkeley AI Safety Student Initiative. https://berkeleyaisafety.com/.

What is 'AI alignment'? Silicon Valley's favourite way to think about AI safety misses the real issues https://theconversation.com/what-is-ai-alignment-silicon-valleys-favourite-way-to-think-about-ai-safety-misses-the-real-issues-209330?utm_source=twitter&utm_medium=bylinetwitterbutton via @ConversationEDU 

AI Alignment Can’t be “Solved” as OpenAI Says - The Hindu https://www.thehindu.com/sci-tech/technology/ai-alignment-cant-be-solved-as-openai-says/article67063877.ece 

AI safety and AI alignment are important topics for the future of artificial intelligence. AI safety is the study of building safe AI systems that do not harm humans or the environment. AI alignment is a subfield of AI safety that focuses on ensuring that AI systems are developed and deployed in a way that aligns with human values and goals. 

Some of the challenges in AI safety and alignment include:

  • - How to instill complex values in AI systems that may not share our moral intuitions or preferences
  • - How to develop honest and transparent AI systems that can be audited and interpreted by humans
  • - How to prevent emergent AI behaviors that may be harmful or undesirable, such as power-seeking, manipulation, or deception
  • - How to achieve consistency between the scientific model, the application solution, and the stakeholder needs
  • - How to deal with uncertainty, ambiguity, and ethical dilemmas in AI decision-making
  • - How to balance the trade-offs between efficiency, fairness, privacy, and accountability in AI systems

There are many researchers and organizations working on these problems, such as OpenAI, MIT CISR, the Partnership on AI, and the United Nations. They use various technical and philosophical approaches, such as direct normativity, indirect normativity, cooperative inverse reinforcement learning, iterated amplification, and more²³⁴⁵. 

AI safety and alignment are not only academic issues, but also practical ones that affect our society and economy. They require collaboration and coordination among multiple stakeholders, such as developers, users, regulators, policymakers, and the public. By addressing these issues, we can ensure that AI systems are beneficial for humanity and the planet.🌎

Bing, 7/12/2023

(1) What is 'AI alignment'? Silicon Valley's favourite way to think about .... https://theconversation.com/what-is-ai-alignment-silicon-valleys-favourite-way-to-think-about-ai-safety-misses-the-real-issues-209330.

(2) AI Alignment: A New Management Paradigm | MIT CISR. https://cisr.mit.edu/publication/2020_1101_AI-Alignment_WixomSomehGregory.

(3) AI alignment - Wikipedia. https://en.wikipedia.org/wiki/AI_alignment.

(4) AI safety - Wikipedia. https://en.wikipedia.org/wiki/AI_safety.

(5) From Risk To Reward: The Role Of AI Alignment In Shaping A Positive Future. https://www.linkedin.com/pulse/from-risk-reward-role-ai-alignment-shaping-positive.

Some of the dangers of AI are:

  • - **Invasion of personal data**: AI systems can collect and analyze vast amounts of personal data without the explicit consent of individuals¹.
  • - **Risk of cyberattack**: AI systems, like any other technology, can be vulnerable to cyberattacks that could compromise their security or functionality¹.
  • - **Discrimination and bias**: AI systems can reflect or amplify the biases and prejudices of their creators or data sources, leading to unfair or harmful outcomes for some groups or individuals¹².
  • - **Opacity and lack of transparency**: AI systems can be difficult to understand or explain, especially when they use complex algorithms or deep learning methods. This can make it hard to hold them accountable or to correct their errors¹².
  • - **Accountability of AI-driven decisions**: AI systems can make decisions that affect human lives or rights, but it can be unclear who is responsible or liable for the outcomes or errors. This can create legal and ethical challenges¹².
  • - **Replacement of jobs**: AI-powered automation can displace workers and reduce employment opportunities in various sectors. This can have social and economic impacts¹².

Some of the groups working on AI safety and alignment are:

  • - **Strategy**: These groups focus on the long-term vision and goals of AI development, as well as the potential scenarios and risks of superintelligent AI. They try to anticipate and plan for the possible impacts of AI on humanity and the world⁷⁸.
  • - **Governance**: These groups focus on the regulation and oversight of AI systems, as well as the norms and standards that guide their use. They try to ensure that AI is developed and deployed in a safe, ethical, and beneficial way⁷⁸.
  • - **Technical**: These groups focus on the design and implementation of AI systems, as well as the methods and tools that enable their safety and alignment. They try to solve the technical challenges and problems that arise in building and testing AI systems⁷⁸.
  • - **Movement Building**: These groups focus on the education and engagement of the public and stakeholders on AI issues, as well as the advocacy and outreach for AI safety and alignment. They try to raise awareness and support for the importance and urgency of AI safety and alignment⁷⁸.

Some examples of organizations that belong to these groups are:

  • - Strategy: Future of Humanity Institute, Center for Human-Compatible AI, Center for the Study of Existential Risk
  • - Governance: Partnership on AI, Center for a New American Security, OpenAI
  • - Technical: DeepMind, Center for AI Safety, Alignment Research Center
  • - Movement Building: Future of Life Institute, 80,000 Hours, Effective Altruism

Bing, 7/12/2023

(1) Dangers of Artificial Intelligence: 6 Risks and Concerns of AI - Flowclass. https://flowclass.io/marketing/dangers-of-artificial-intelligence/.

(2) Here's Why AI May Be Extremely Dangerous--Whether It's Conscious or Not .... https://www.scientificamerican.com/article/heres-why-ai-may-be-extremely-dangerous-whether-its-conscious-or-not/.

(3) The AI Safety community has four main work groups, Strategy ... - LessWrong. https://www.lesswrong.com/posts/zCYChCmnxsowBsMri/the-ai-safety-community-has-four-main-work-groups-strategy.

(4) AI bias and AI safety teams are divided on artificial intelligence - Vox. https://www.vox.com/future-perfect/2022/8/10/23298108/ai-dangers-ethics-alignment-present-future-risk.

(5) What is AI, is it dangerous and what jobs are at risk? - BBC. https://www.bbc.com/news/technology-65855333.

(6) SQ10. What are the most pressing dangers of AI?. https://ai100.stanford.edu/2021-report/standing-questions-and-responses/sq10-what-are-most-pressing-dangers-ai.

(7) New report assesses progress and risks of artificial intelligence. https://www.brown.edu/news/2021-09-16/ai100.

(8) The 15 Biggest Risks Of Artificial Intelligence - Forbes. https://www.forbes.com/sites/bernardmarr/2023/06/02/the-15-biggest-risks-of-artificial-intelligence/.

(9) Center for AI Safety (CAIS). https://www.safe.ai/.

(10) Part 1: The AI Safety community has four main work groups, Strategy .... https://forum.effectivealtruism.org/posts/5iQoR8mhEpvRT43jv/part-1-the-ai-safety-community-has-four-main-work-groups.