ReportWire

Tag: information literacy

  • Lessons from DENSI: Weaving digital citizenship into edtech innovation

    [ad_1]

    Key points:

    What happens when over 100 passionate educators converge in Chicago to celebrate two decades of educational innovation? A few weeks ago, I had the thrilling opportunity to immerse myself in the 20th anniversary of the Discovery Educator Network (the DEN), a week-long journey that reignited my passion for transforming classrooms.

    From sunrise to past sunset, my days at Loyola University were a whirlwind of learning, laughter, and relentless exploration. Living the dorm life, forging new connections, and rekindling old friendships, we collectively dove deep into the future of learning, creating experiences that went far beyond the typical professional development.

    As an inaugural DEN member, the professional learning community supported by Discovery Education, I was incredibly excited to return 20 years after its founding to guide a small group of educators through the bountiful innovations of the DEN Summer Institute (DENSI). Think scavenger hunts, enlightening workshops, and collaborative creations–every moment was packed with cutting-edge ideas and practical strategies for weaving technology seamlessly into our teaching, ensuring our students are truly future-ready.

    During my time at DENSI, I learned a lot of new tips and tricks that I will pass on to the educators I collaborate with. From AI’s potential to the various new ways to work together online, participants in this unique event learned a number of ways to weave digital citizenship into edtech innovation. I’ve narrowed them down to five core concepts; each a powerful step toward building future-ready classrooms and fostering truly responsible digital citizens.

    Use of artificial intelligence

    Technology integration: When modeling responsible AI use, key technology tools could include generative platforms like Gemini, NotebookLM, Magic School AI, and Brisk, acting as ‘thought partners’ for brainstorming, summarizing, and drafting. Integration also covers AI grammar/spell-checkers, data visualization tools, and feedback tools for refining writing, presenting information, and self-assessment, enhancing digital content interaction and production.

    Learning & application: Teaching students to ethically use AI is key. This involves modeling critical evaluation of AI content for bias and inaccuracies. For instance, providing students with an AI summary of a historical event to fact-check with credible sources. Students learn to apply AI as a thought partner, boosting creativity and collaboration, not replacing their own thinking. Fact-checking and integrating their unique voices are essential. An English class could use AI to brainstorm plot ideas, but students develop characters and write the narrative. Application includes using AI for writing refinement and data exploration, fostering understanding of AI’s academic capabilities and limitations.

    Connection to digital citizenship: This example predominantly connects to digital citizenship. Teaching responsible AI use promotes intellectual honesty and information literacy. Students can grasp ethical considerations like plagiarism and proper attribution. The “red, yellow, green” stoplight method provides a framework for AI use, teaching students when to use AI as a collaborator, editor, or thought partner–or not at all.This approach cultivates critical thinking and empowers students to navigate the digital landscape with integrity, preparing them as responsible digital citizens understanding AI’s implications.

    Digital communication

    Technology integration: Creating digital communication norms should focus on clarity with visuals like infographics, screenshots, and video clips. Canva is a key tool for a visual “Digital Communication Agreement” defining online interaction expectations. Include student voice by the integration and use of pictures and graphics to illustrate behaviors and potentially collaborative presentation / polling tools for student involvement in norm-setting.

    Learning & application: Establishing clear online interaction norms is the focus of digital communication. Applying clear principles teaches the importance of visuals and setting communication goals. Creating a visual “Digital Communication Agreement” with Canva is a practical application where students define respectful online language and netiquette. An elementary class might design a virtual classroom rules poster, showing chat emojis and explaining “think before you post.” Using screenshots and “SMART goals” for online discussions reinforces learning, teaching constructive feedback and respectful debate. In a middle school science discussion board, the teacher could model a respectful response like “I understand your point, but I’m wondering if…” This helps students apply effective digital communication principles.

    Connection to digital citizenship: This example fosters respectful communication, empathy, and understanding of online social norms. By creating and adhering to a “Digital Communication Agreement,” students develop responsibility for online interactions. Emphasizing respectful language and netiquette cultivates empathy and awareness of their words’ impact. This prepares them as considerate digital citizens, contributing positively to inclusive online communities.

    Content curation

    Technology integration: For understanding digital footprints, one primary tool is Google Drive when used as a digital folder to curate students’ content. The “Tech Toolbox” concept implies interaction with various digital platforms where online presence exists. Use of many tools to curate content allows students to leave traces on a range of technologies forming their collective digital footprint.

    Learning & application: This centers on educating students about their online presence’s permanence and nature. Teaching them to curate digital content in a structured way, like using a Google Drive folder, is key. A student could create a “Digital Portfolio” in Google Drive with online projects, proud social media posts, and reflections on their public identity. By collecting and reviewing online artifacts, students visualize their current “digital footprint.” The classroom “listening tour” encourages critical self-reflection, prompting students to think about why they share online and how to be intentional about their online identity. This might involve students reviewing anonymized social media profiles, discussing the impression given to future employers.

    Connection to digital citizenship: This example cultivates awareness of online permanence, privacy, responsible self-presentation, and reputation management. Understanding lasting digital traces empowers students to make informed decisions. The reflection process encourages the consideration of their footprint’s impact, fostering ownership and accountability for online behavior. This helps them become mindful, capable digital citizens.

    Promoting media literacy

    Technology integration: One way to promote media literacy is by using “Paperslides” for engaging content creation, leveraging cameras and simple video recording. This concept gained popularity at the beginning of the DEN through Dr. Lodge McCammon. Dr. Lodge’s popular 1-Take Paperslide Video strategy is to “hit record, present your material, then hit stop, and your product is done” style of video creation is something that anyone can start using tomorrow. Integration uses real-life examples (likely digital media) to share a variety of topics for any audience. Additionally, to apply “Pay Full Attention” in a digital context implies online viewing platforms and communication tools for modeling digital eye contact and verbal cues.

    Learning & application: Integrating critical media consumption with engaging content creation is the focus. Students learn to leverage “Paperslides” or another video creation method to explain topics or present research, moving beyond passive consumption. For a history project, students could create “Paperslides” explaining World War II causes, sourcing information and depicting events. Learning involves using real-life examples to discern credible online sources, understanding misinformation and bias. A lesson might show a satirical news article, guiding students to verify sources and claims through their storyboard portion. Applying “Pay Full Attention” teaches active, critical viewing, minimizing distractions. During a class viewing of an educational video, students could pause to discuss presenter credentials or unsupported claims, mimicking active listening. This fosters practical media literacy in creating and consuming digital content.

    Connection to digital citizenship: This example enhances media literacy, critical online information evaluation, and understanding persuasive techniques. Learning to create and critically consume content makes students informed, responsible digital participants. They identify and question sources, essential for navigating a digital information-saturated world. This empowers them as discerning digital citizens, contributing thoughtfully to online content.

    Collaborative problem-solving

    Technology integration: For practicing digital empathy and support, key tools are collaborative online documents like Google Docs and Google Slides. Integration extends to online discussion forums (Google Classroom, Flip) for empathetic dialogue, and project management tools (Trello, Asana) for transparent organization. 

    Learning & application: This focuses on developing effective collaborative skills and empathetic communication in digital spaces. Students learn to work together on shared documents, applying a “Co-Teacher or Model Lessons” approach where they “co-teach” each other new tools or concepts. In a group science experiment, students might use a shared Google Doc to plan methodology, with one “co-teaching” data table insertion from Google Sheets. They practice constructive feedback and model active listening in digital settings, using chat for clarification or emojis for feelings. The “red, yellow, green” policy provides a clear framework for online group work, teaching when to seek help, proceed cautiously, or move forward confidently. For a research project, “red” means needing a group huddle, “yellow” is proceeding with caution, and “green” is ready for review.

    Connection to digital citizenship: This example is central to digital citizenship, developing empathy, respectful collaboration, and responsible problem-solving in digital environments. Structured online group work teaches how to navigate disagreements and offers supportive feedback. Emphasis on active listening and empathetic responses helps internalize civility, preparing students as considerate digital citizens contributing positively to online communities.

    These examples offer a powerful roadmap for cultivating essential digital citizenship skills and preparing all learners to be future-ready. The collective impact of thoughtfully utilizing these or similar approaches , or even grab and go resources from programs such as Discovery Education’s Digital Citizenship Initiative, can provide the foundation for a strong academic and empathetic school year, empowering educators and students alike to navigate the digital world with confidence, integrity, and a deep understanding of their role as responsible digital citizens.

    In addition, this event reminded me of the power of professional learning communities.  Every educator needs and deserves a supportive community that will share ideas, push their thinking, and support their professional development. One of my long-standing communities is the Discovery Educator Network (which is currently accepting applications for membership). 

    Latest posts by eSchool Media Contributors (see all)

    [ad_2]

    Stephen Wakefield

    Source link

  • Critical thinking in the digital age of AI: Information literacy is key

    Critical thinking in the digital age of AI: Information literacy is key

    [ad_1]

    Key points:

    From New York to Texas, the pro-Palestinian protests sweeping U.S. colleges have become a flashpoint for viral disinformation, from falsely attributed “Jewish genocide” chants to debunked claims of Hamas presence. With the tenor of allegations reaching a fever pitch, Columbia University students have even launched their own fact-checking Twitter account. As this highly-charged moment collides with a hyper-partisan landscape, it offers a stark reminder of how disinformation thrives at the intersection of fierce emotions and polarized politics, threatening to drown out nuance, facts, and good-faith dialogue when they are needed most. All of this points to the urgency of tackling disinformation through information literacy.

    Disinformation has long played a role in global events. Technological change and increasingly global communications have made the deliberate spread of inaccurate information faster and more impactful. With the birth of AI, disinformation has entered a new era, rendering it critical to teach students how to question sources, spot fakes and be discerning consumers of news, social media, and information.  

    AI has dramatically complicated the information landscape by rapidly generating and amplifying deceptive narratives, deepfakes, and AI-generated visuals, drawing concern from global leaders as a major emerging challenge. The World Economic Forum’s latest Global Risks Report, which surveyed experts from academia, business, government, the international community, and civil society, named misinformation and disinformation from AI as the top global risk over the next two years–ahead of climate change and war.

    The stakes are high, especially as the U.S. approaches a critical election year–one that will undoubtedly be subject to disinformation, a force that voters will remember as having played a critical role in the 2016 and 2020 elections.

    As an academic who has studied how digital technology is used by governments and non-state actors for the purposes of repression and information control, these issues are especially concerning. There is an urgent need to promote greater critical thinking among young people, to give them the tools to detect what information is authentic and what has been manipulated. Information literacy, specifically across digital platforms, should be a mandatory part of every K-12 curriculum, to combat the rise of disinformation and develop more discerning students ready to take on an AI-driven future.

    How and where disinformation can take place

    Disinformation can show up anywhere, but it thrives on stories that appeal to emotions. Election issues and partisan politics are a prime example. During the pandemic, COVID-19 disinformation narratives, spanning the bizarre claims that the disease is spread by 5G and other conspiracies, spread faster than the virus itself–thanks to digital technology. Anti-vaccine groups essentially tricked Facebook’s algorithms into allowing posts that spread disinformation by using a carrot emoji in place of the word “vaccine.” Looking at climate change–another highly polarized and partisan issue–a probe into a subset of social media accounts revealed hundreds of AI-generated and stolen pictures used in greenwashing campaigns.

    Praying on the emotions that emerged after the deadly October 7th attacks and the ensuing attacks on Gaza, deepfakes powered by AI have spread at an unprecedented pace. Soon after October 7th, a fake story emerged that Qatar had threatened to cut off the world’s natural gas supply if Israel didn’t stop its bombing in Gaza, garnering millions of views before it was ultimately debunked. More recently, the United Nations Relief and Works Agency (UNRWA) has been a target of disinformation, thanks to a network of fake accounts and websites that have collaborated to spread accusations about the agency’s ties with Hamas. 

    Not only is disinformation incredibly damaging to the delivery of accurate, verifiable information, it has eroded the public’s trust in some of our most reliable institutions. Only 32 percent of Americans say they trust the mass media, a figure that is tied with record-low levels in 2016.

    Engaging with disinformation and AI as teachable moments

    Disinformation can be rectified through fact checking, but in many cases, a false story has already done its damage before it is corrected. Another strategy is ‘prebunking,’ a technique gaining momentum that helps to build preemptive resilience to misinformation.

    We can combat the spread of disinformation by encouraging and teaching more critical thinking, especially about AI, algorithms, and deception, and the value of greater subject matter knowledge.  

    Whether you are a teacher in K-12 schools, a university instructor, or simply an individual who actively engages in online platforms, there are many steps that can be taken to ensure a greater understanding and literacy around disinformation and AI. This will in turn instill greater trust in the institutions and organizations that disseminate the information we are seeking.

    Context-based case studies, such as videos of celebrities and influencers, can serve as important teaching moments. In my classes, I’ve challenged students to discern what is a deepfake or AI-generated image through exercises such as reverse image searches. This teaches them to detect clues such as fuzzy details, inconsistent lighting, out-of-sync audio and visuals, and the credibility of the image source. We spend time analyzing and discussing the spread, origins, and nature of social media manipulation, which equips students with important data literacy skills.

    Bringing the study of disinformation to the classroom

    What we know about the world ultimately informs how we approach disinformation and deception. Today’s students need a cross-disciplinary approach that starts early, so the foundations of critical thinking and information literacy are instilled at a young age and stick with them as they grow and mature.

    In Finland, media literacy constitutes a core component part of the national curriculum, starting in preschool. They start with understanding the basic elements of media, and build from there to understand more complex elements, such as identifying sources. It is not a single subject–rather, it is taught across different disciplines, including Finnish language and literature, math, and art to grow a well-rounded set of analytical skills. In a survey published by the Open Society Institute in Bulgaria, Finland has ranked No. 1 of 41 European countries on resilience against misinformation for the fifth time in a row. Finland’s population also has a higher level of trust in news and other institutions, with 76 percent of Finns considering print and digital newspapers to be reliable, according to a survey conducted by market research company IRO research.

    There is no denying the impact of disinformation and the stronghold it is having on political processes around the world. We will doubtless see the use of disinformation throughout 2024 U.S. presidential election battle, but a concerted effort on developing greater critical thinking can help alleviate the impact. By becoming more knowledgeable about what disinformation is, as well as different countries, cultures, and subjects, we can better navigate the array of disinformation scenarios in the digital world and foster a questioning mindset.

    Latest posts by eSchool Media Contributors (see all)

    [ad_2]

    Dr. Marc Owen Jones, Northwestern University in Qatar

    Source link