Report: Schools Are Being Duped by Marketers on Personalized Learning

Personalized learning often receives high praise for tailoring to the individual needs of each learner. However, researchers at the National Education Policy Center (NEPC) have found that personalization is often a threat to students’ privacy and a clever way for third-party products and services to collect personal student data for corporate gain.

The NEPC last week released its 19th annual report on school commercialism trends, in which researchers from the University of Colorado Boulder assert that personalized learning is turning student data into a marketable product and extending the reach of commercializing activities into students’ everyday lives at schools.

“Although marketers’ school-focused efforts are often billed as ‘innovative’ and ‘out-of-the-box,’ many of them are little more than repackaged marketing strategies that over the years have been seen again and again,” the report noted. Many K–12 educators have echoed similar sentiments, saying there is nothing personal or new about personalized learning. Some educators even argue that personalized learning further standardizes and privatizes education, while creating more hurdles for teachers in the classroom.

“Commercialization in schools is far from new,” the NEPC report noted, since they are a particularly desirable marketing environment. “Schoolchildren represent a captive audience required to spend hours in surroundings that, from a marketing perspective, is relatively uncluttered. Moreover, schools confer legitimacy on any product or worldview promoted there.”

Yet, schools are taking the bait and marketers are capitalizing, largely by employing the following strategies found in the report:

  • Appropriation of space on school property;
  • Sponsored programs and activities;
  • Exclusive agreements;
  • Sponsorship of supplementary educational materials;
  • Incentive programs;
  • Fundraising; and
  • Digital marketing.   

These tactics are luring schools interested in personalized learning and creating a “surveillance economy” that normalizes relentless monitoring for marketing purposes, according to the NEPC report.

“The dominant beliefs currently associated with technology and economic development are leading schools and districts to change their policies, pay huge sums of money to private vendors, and create systems for divulging vast amounts of children’s personal information to education technology companies,” the NEPC report stated. “Education applications, particularly those that attempt to ‘personalize’ student learning, are powered by proprietary algorithms that may harm children as they implement theories of learning without policymakers or teachers being able to examine how they work or how student data are being used.”

Furthermore, the researchers found that education policymaking is falling short of its duty to protect students’ personal data and have outlined several recommendations for policymakers. Many of the recommendations seem standard: Legislatures should barr schools from collecting student personal data unless safeguards are used. They should also hold school, districts and ed tech companies accountable for violations of student privacy. Notably, the NEPC suggests requiring algorithms used in classroom software “to be openly available for examination by educators and researchers.”

The full report, “Asleep at the Switch: Schoolhouse Commercialism, Student Privacy, and the Failure of Policymaking,” can be found on the NEPC site.

About the Author

Sri Ravipati is Web producer for THE Journal and Campus Technology. She can be reached at [email protected].

Featured

  • An elementary school teacher and young students interact with floating holographic screens displaying colorful charts and playful data visualizations in a minimalist classroom setting

    New AI Collaborative to Explore Use of Artificial Intelligence to Improve Teaching and Learning

    Education-focused nonprofits Leading Educators and The Learning Accelerator have partnered to launch the School Teams AI Collaborative, a yearlong pilot initiative that will convene school teams, educators, and thought leaders to explore ways that artificial intelligence can enhance instruction.

  • landscape photo with an AI rubber stamp on top

    California AI Watermarking Bill Supported by OpenAI

    OpenAI, creator of ChatGPT, is backing a California bill that would require tech companies to label AI-generated content in the form of a digital "watermark." The proposed legislation, known as the "California Digital Content Provenance Standards" (AB 3211), aims to ensure transparency in digital media by identifying content created through artificial intelligence. This requirement would apply to a broad range of AI-generated material, from harmless memes to deepfakes that could be used to spread misinformation about political candidates.

  • closeup of laptop and smartphone calendars

    2024 Tech Tactics in Education Conference Agenda Announced

    Registration is free for this fully virtual Sept. 25 event, focused on "Building the Future-Ready Institution" in K-12 and higher education.

  • cloud icon connected to a data network with an alert symbol (a triangle with an exclamation mark) overlaying the cloud

    U.S. Department of Commerce Proposes Reporting Requirements for AI, Cloud Providers

    The United States Department of Commerce is proposing a new reporting requirement for AI developers and cloud providers. This proposed rule from the department's Bureau of Industry and Security (BIS) aims to enhance national security by establishing reporting requirements for the development of advanced AI models and computing clusters.