Education
Anthropic’s AI Usage Study: Coding Still Dominates

editorially independent. We may make money when you click on links
to our partners.
Learn More
Anthropic has released its third Economic Index, a wide-ranging study of Claude AI usage across countries, US states, and enterprises. The September 2025 report shows coding still accounts for the largest share of activity worldwide, yet education and science are becoming more prominent.
The study also found that automation is increasingly common, with more users trusting Claude to complete entire tasks. Anthropic said the trend offers an early look at how AI is reshaping work and may determine which workers and regions gain the most from the technology.
From debugging code to classroom teaching
Software development remains Claude’s most common use case, making up more than a third of activity globally. But the kind of coding has shifted. New code creation doubled over the past eight months, while debugging declined, suggesting users rely on Claude for more advanced outputs in a single attempt.
Claude AI usage for education and science purposes also expanded. The share of educational tasks rose from about 9% at the start of 2025 to more than 12% by August. Scientific tasks climbed from 6% to over 7%. In contrast, business and financial operations fell from 6% to 3%, and management dropped from 5% to 3%.
Automation vs augmentation in AI tasks
Anthropic tracked not only what tasks Claude handled but also how people worked with the model. Conversations where users gave Claude a job and let it run with minimal input jumped from 27% in late 2024 to 39% in August 2025. For the first time, automation outweighed collaboration overall, making the balance of automation vs augmentation in AI tasks a central trend.

The AI company said two forces may explain the shift: improvements in Claude’s ability to deliver accurate results on the first try, and rising user confidence. That combination has made it more common for people to delegate complex tasks to the AI tool fully, rather than iterating step by step.
Global and enterprise patterns
The US accounted for 21.6% of global usage, far ahead of India, Brazil, Japan, and South Korea. Adjusted for population, smaller high-income countries led per capita. Israel, Singapore, Australia, New Zealand, and South Korea all used Claude far more than expected based on their working-age population.
Within the US, Washington, DC showed the highest per-capita usage, followed by Utah, California, New York, and Virginia. Anthropic said local economies shaped how people used the model, with IT-related requests most common in California and tourism tasks overrepresented in Hawaii.
Enterprise adoption leaned more heavily on automation than consumer use. About 44% of API traffic involved coding, compared with 36% on Claude.ai. Administrative tasks were also frequent, while education and creative work were far less common. Nearly eight in 10 business interactions were automated, showing companies are embedding Claude directly into workflows.
Anthropic said it will continue updating its Economic Index and releasing data to help researchers and policymakers prepare for AI’s economic effects. The findings point to an uneven but accelerating shift: coding still leads, but education, science, and automation are taking a growing share of AI’s early role in the economy.
Read about Anthropic’s recent decision to give Claude power to end harmful conversations and protect “model welfare.”
Education
Generative AI reshapes global education: Opportunities and risks

Generative artificial intelligence, or GenAI, is rapidly transforming the landscape of global education, with researchers stressing both the scale of opportunities and the depth of challenges. A new international study published in SAGE Open presents the first large-scale evidence of how generative AI has reshaped research, teaching, and learning in the 21st century.
The paper, titled “Are Generative AI Technologies Transforming Education for the 21st Century? Research Trends, Challenges, and Benefits,” provides an in-depth look at how education systems and academic research are responding to the rise of tools such as ChatGPT, GPT-4, and other generative models.
How has research on generative AI in education expanded?
The study finds that research output in this field was negligible before 2020, but momentum built rapidly from 2021, culminating in a surge after 2023. By 2024, scholarly production exceeded 1,800 publications in a single year, reflecting how generative AI technologies have become central to debates on the future of education. The rapid growth highlights not only academic interest but also the urgency of addressing new challenges created by generative AI in schools and universities.
The analysis identifies leading academic outlets driving the conversation. Journals such as Education and Information Technologies, Computers and Education: Artificial Intelligence, and Education Sciences emerged as the most influential platforms publishing research in this area. Highly cited studies concentrated on key issues such as the impact of ChatGPT on assessment systems, the risks to academic integrity, and the potential for generative AI to personalize learning.
The authors point out that this surge of academic attention reflects the disruptive influence of generative AI. Once viewed as experimental, these tools are now embedded in both teaching practices and student learning habits. The scale and pace of publications demonstrate that the academic community is mobilizing to study not only the potential benefits but also the far-reaching risks of generative AI adoption.
Who leads global research and what are the main themes?
The study provides a global mapping of research trends. The United States emerged as the clear leader in terms of the number of publications, followed by Australia and the United Kingdom. Interestingly, Singapore distinguished itself with one of the highest citation impacts despite producing fewer studies, suggesting that its contributions carry disproportionate influence. This global spread demonstrates that generative AI is not confined to a single region but is shaping education debates worldwide.
The bibliometric analysis also revealed three dominant research clusters. The first cluster focuses on the integration of generative AI into higher education, especially its impact on student learning outcomes and academic integrity. The second cluster addresses the technical and ethical implications of large language models in educational settings, reflecting concerns about bias, transparency, and governance. The third cluster centers on the role of generative AI in medical and nursing education, where AI-driven tools are used for simulations, skill-building, and clinical decision support.
These themes illustrate the dual role of generative AI as both an enabler and a disruptor. It enables educators to introduce innovative teaching methods and support personalized learning experiences, while at the same time disrupting traditional academic norms by challenging established practices of assessment and integrity. The geographic distribution of research suggests that while developed nations are leading in volume, emerging economies are beginning to participate actively in shaping this dialogue.
What benefits and challenges does generative AI bring to education?
The study stresses that the benefits of generative AI are significant. AI-powered tools can deliver personalized learning pathways, encourage creativity, and increase efficiency in administrative and academic tasks. In higher education, they have the potential to transform curricula, automate grading, and provide real-time feedback to students. In medical education, they are being deployed for realistic simulations and competency-based training.
At the same time, the challenges are substantial. Generative AI raises concerns about plagiarism, reduced critical thinking, and dependency on machine-generated outputs. Ethical issues such as data privacy, misinformation, and intellectual property rights compound these risks. The research warns that without strong governance frameworks, generative AI could undermine the credibility of academic institutions and weaken trust in education systems.
The authors argue that educational policymakers and institutions must act quickly to balance benefits with risks. The development of clear guidelines for responsible AI use, investment in AI literacy programs, and strategies to strengthen academic integrity are identified as urgent needs. Without such measures, the study cautions that the promises of generative AI could be overshadowed by its pitfalls.
Education
Upping Your Game – The Future of Compliance Education: Leveraging AI for Targeted Training | Thomas Fox – Compliance Evangelist

This podcast series, sponsored by Ethico and co-hosted with Ethico co-CEO Nick Gallo, hopes to meet Hui Chen’s challenge. They will discuss how See more +
This podcast series, sponsored by Ethico and co-hosted with Ethico co-CEO Nick Gallo, hopes to meet Hui Chen’s challenge. They will discuss how compliance professionals can ‘Up Their Game’ by utilizing currently existing Generative AI (GenAI) tools to improve their compliance programs significantly. As compliance professionals, it is critical to recognize that this moment is not merely about incremental improvements but about elevating our profession to an entirely new level of effectiveness, efficiency, and organizational value.
In this episode of Upping Your Game, Tom and Nick discuss the transformative potential of AI in compliance education. They explore how traditional compliance training methods with PowerPoints and annual tests are being surpassed by more dynamic, targeted approaches. Emphasizing the importance of delivering relevant messages to the right audience at the right time, they discuss how AI can integrate various training strategies and data-driven insights to enhance compliance. By tailoring content for specific roles and using crowd-sourced intelligence and accurate input data, compliance education can become more effective and engaging. The conversation highlights the importance of reducing organizational volatility by employing more targeted, frequent, and engaging training methods.
Key highlights:
• Evolution of Compliance Training
• The Role of AI in Compliance Education
• Targeted Training and Its Benefits
• Crowdsourcing and Input for Effective Training
• Reducing Volatility Through Compliance Education
Resources:
• Upping Your Game-How Compliance and Risk Management Move to 2030 and Beyond on Amazon.com: https://a.co/d/6kKYANX
• Nick Gallo on LinkedIn: https://www.linkedin.com/in/ngallo/
• Ethico: http://ethico.com/ See less –
Education
Here’s How Teachers Really Feel About the Rise of AI in K-12 Education

Artificial intelligence has been rapidly changing the K-12 education landscape—from providing opportunities for personalized learning to assisting with nonteaching tasks. As the new school year kicks into gear, how do educators expect to see AI incorporated into their school and classes?
EdWeek reporting shows that not many schools have policies in place to guide AI use just yet, although it’s a top-of-mind issue for educators and policymakers. According to a recent survey of state education technology officials, 40% of respondents said their states had released AI guidance, and supporting professional development on the issue was a top priority.
Additionally, educators remain divided on whether AI should be used in the classroom at all. There is concern among some educators about how AI may affect students’ critical thinking skills, as well as their ability to experiment and learn.
Still, most educators feel as if the use of AI in education is inevitable. In an informal Education Week LinkedIn poll with 700 votes, 87% of respondents said AI will affect the classroom, and 7% said it would not.
Educators shared more details about their opinions on AI’s impact in the comment section. Here are some of their responses, edited lightly for clarity.
Some educators fear AI will have a negative impact
It will be banned within three years once test scores plummet. It can’t be controlled, and 80% will use it to cheat.
It has negatively impacted the general public for sure. I hope it doesn’t enter the schools next.
Totally against it as an educator. If anything, I’m pulling BACK on technology in my classroom this year because it’s becoming more of a hindrance.
Other educators are embracing AI
I’m counting on it [having an impact]. And designing PD about exactly that.
Justin M.
Students are already using it, some to help them study and others to outright cheat. I’ve used it to help with planning and organizing ideas and even for inspiration sometimes. It’s here now and we can’t pretend that it isn’t, but we can show students how to use it in a productive way.
Emily H.
AI will shape classrooms in ways we can’t ignore. The opportunity? Freeing teachers from busywork so they can focus on relationships, creativity, and student growth. The challenge? Keeping humanity, critical thinking, and cultural awareness at the center.
Tricia T.
-
Business3 weeks ago
The Guardian view on Trump and the Fed: independence is no substitute for accountability | Editorial
-
Tools & Platforms1 month ago
Building Trust in Military AI Starts with Opening the Black Box – War on the Rocks
-
Ethics & Policy2 months ago
SDAIA Supports Saudi Arabia’s Leadership in Shaping Global AI Ethics, Policy, and Research – وكالة الأنباء السعودية
-
Events & Conferences4 months ago
Journey to 1000 models: Scaling Instagram’s recommendation system
-
Jobs & Careers3 months ago
Mumbai-based Perplexity Alternative Has 60k+ Users Without Funding
-
Podcasts & Talks2 months ago
Happy 4th of July! 🎆 Made with Veo 3 in Gemini
-
Education3 months ago
VEX Robotics launches AI-powered classroom robotics system
-
Education2 months ago
Macron says UK and France have duty to tackle illegal migration ‘with humanity, solidarity and firmness’ – UK politics live | Politics
-
Podcasts & Talks2 months ago
OpenAI 🤝 @teamganassi
-
Funding & Business3 months ago
Kayak and Expedia race to build AI travel agents that turn social posts into itineraries