AI Insights
Exploring AI literacy, attitudes toward AI, and intentions to use AI in clinical contexts among healthcare students in Korea: a cross-sectional study | BMC Medical Education

As AI literacy increasingly becomes a fundamental competency for future healthcare professionals, it is imperative that healthcare students are adequately prepared to navigate technological advancements in their future workplaces. To support this goal, assessing the current level of AI literacy among healthcare students is a necessary first step in designing targeted and effective educational interventions. In line with this objective, the present study explored healthcare students’ AI literacy and examined its associations with their attitudes toward AI, their intention to use AI in clinical contexts, and individual characteristics such as prior AI training and interest in AI.
To address this aim, SNAIL was first validated through confirmatory factor analysis in the Korean context. During this process, seven items were excluded, resulting in a final version consisting of 24 items (SNAIL-KR), which was used for the analysis. The results revealed that students demonstrated slightly below-average levels of AI literacy. These findings are consistent with those of Laupichler et al. [6], who reported a comparable level of AI literacy among German medical students, with a total mean score of 3.76 (3.79 in the present study). In their study, critical appraisal also showed the highest mean score (M = 4.89, M = 4.64 in this study) and technical understanding the lowest (M = 2.63, M = 3.19 in this study), demonstrating the same pattern observed in the present study. The relatively higher scores in critical appraisal may reflect increased exposure to media coverage on issues such as data privacy, ethics, and AI-related risks. However, given that the highest subscale score was only marginally above the midpoint of 4, the findings collectively underscore the need for more comprehensive AI literacy education. This interpretation is further supported by the fact that 85.17% of participants reported receiving less than 30 h of AI-related training. Moreover, systematic reviews by Kimiafar et al. [30] and Mousavi Baigi et al. [2] similarly reported that although healthcare professionals and students generally expressed motivation to adopt AI, they had received insufficient training and tended to possess limited knowledge and skills in using AI. These findings reinforce the present study’s results, highlighting the urgent need to strengthen AI literacy education to better prepare future healthcare professionals.
In addition, participants overall exhibited positive attitudes toward AI. However, the mean score for negative attitudes (reverse scored) was also slightly above the neutral midpoint. This suggests a balance between favorable and cautious perspectives—students recognized the potential benefits of AI while also acknowledging its current limitations. This result is consistent with the findings of Seo and Ahn [23], who examined the attitudes of 230 Korean nursing students using the same scale. On a 5-point Likert scale, the mean for positive attitudes was 3.69, while that for negative attitudes was 3.07, reflecting a pattern similar to that observed in the present study. Similarly, Gordon et al.’s [1] scoping review on attitudes toward the application of AI in clinical medicine reported the coexistence of support for and concerns about AI among healthcare learners. Such concerns may influence learners to adopt a cautious stance toward AI applications. Collectively, these findings highlight the importance of directly addressing both the opportunities and concerns related to AI in future education and training programs.
Furthermore, the findings indicate that students with higher levels of AI literacy are more likely to hold positive attitudes toward AI—and vice versa. Among the subcategories, practical application showed the strongest correlation with positive attitudes, suggesting that students who view AI favorably are more inclined to engage with it in daily life. Critical appraisal was positively associated with positive attitudes and negatively associated with negative attitudes, with a notably stronger correlation for positive attitudes. This implies that the more students understand both the benefits and potential concerns of AI, the more likely they are to adopt a positive perspective. In addition, a positive association was observed between positive attitudes toward AI and prior AI training. Collectively, these results highlight the importance of structured AI literacy education that addresses both the advantages and potential risks associated with AI.
Regarding the intention to use AI in clinical contexts, participants reported scores slightly above the neutral midpoint. This intention was significantly associated with both AI literacy and positive attitudes toward AI. These findings demonstrate that AI literacy is a critical factor influencing students’ intention to use AI in clinical practice, while also confirming previous research showing a positive association between positive attitudes toward AI and the intention to use it [1, 13].
Additionally, both AI literacy and positive attitudes toward AI were significantly correlated with students’ interest in AI and their prior AI training, with stronger correlations observed for interest in AI than for prior training. Interest in AI was also significantly associated with the intention to use AI, whereas prior AI training did not show a significant correlation with this variable. Furthermore, among the subcategories of AI literacy, practical application exhibited a notably stronger correlation with students’ interest in AI than with their prior AI training. This suggests that students with higher interest in AI are more likely to engage with AI technologies in their daily lives, thereby reinforcing their AI literacy and fostering positive attitude AI. The influential role of interest in AI on AI literacy aligns with findings by Laupichler et al. [6], who reported that students’ interest in AI was more strongly association with AI literacy than previous training experience. Moreover, there was a notable difference in AI literacy between students with hardly any training and those with less than 30 h training or more than 30 h of training, with mean differences of 0.68 and 1.1 point, respectively. These findings highlight the potential impact of AI education and suggest that programs providing more than 30 h of training may be necessary to meaningfully enhance students’ AI literacy. While the lack of significant association between prior AI training and intention to use AI may be attributable to the small sample size, further research is needed to clarify this relationship. Overall, the findings suggest that fostering interest in AI may be particularly effective in promoting AI literacy and that educational interventions should aim to stimulate engagement and curiosity about AI especially through programs offering more than 30 h of training.
No significant differences were observed by gender, year of study, or major for any of the measured variables in this study. These findings differ from those reported by Laupichler et al. [6], who found that male students scored higher in AI literacy and that academic semester was significantly associated with AI literacy levels. Similarly, Hashish and Alnajjar [5], in a study assessing nursing students’ digital health literacy—defined as perceived competence in using digital health tools—found that senior students demonstrated higher literacy levels. In addition, Kwak et al. [13] reported that senior students exhibited significantly higher positive attitudes toward AI, and Sumengen et al. [20] found that male nursing students showed more positive attitudes toward AI, as both measured by the GAAIS.
This discrepancies between these previous findings and the current results cannot be fully explained in this study. However, one possible explanation for the lack of significant gender difference is that participation in this survey was voluntary, and response rate is low. Thus, students who chose to participate—regardless of gender—may have had a greater interest in AI than the general student population. In addition, the absence of significant differences by year of study may reflect the limited integration of AI-related content into the curriculum. As previously noted, AI education in Korea is typically offered only as elective courses during pre-medical phase, which may contribute to a lack of variation in AI literacy and attitudes across academic years. Alternatively, these nonsignificant results may be attributed to the relatively small sample size. Regarding the lack of significant differences by major, further investigation is needed to explore whether and how AI literacy varies across academic disciplines within healthcare education. Future studies with larger and more diverse samples across multiple disciplines in healthcare education are needed to confirm and extend these findings.
The findings of this study highlight the urgent need to systematically integrate AI literacy into healthcare curricula. Current training opportunities appear insufficient, with most students reporting minimal exposure to AI education. Given the observed increase in AI literacy among students with more than 30 h of AI training, this study highlights the value of structured programs of sufficient length. Training that exceeds 30 h may be especially effective in enhancing both AI literacy and positive attitudes toward AI. In addition, integrating AI content into core curricula—rather than offering it solely as electives—can promote more equitable and comprehensive learning experiences. Educators should also leverage students’ existing interest in AI as a foundation for enhancing AI literacy. Instructional programs should be deliberately designed to spark and sustain students’ interest in AI, for example, by incorporating hands-on activities using AI tools, and interdisciplinary projects that apply AI to solve real-world healthcare problems. Furthermore, given that students may approach AI with a cautious stance toward AI applications, AI literacy programs should explicitly address their concerns. Promoting open dialogue, encouraging critical reflection, and providing exposure to practical applications can help learners develop a balanced and informed perspective on the role of AI in clinical practice.
Despite its contributions, this study has several limitations. First, the sample size was relatively small. The response rate of online survey was very low likely due to an ongoing two-year collective leave of absence among Korean medical students, triggered by political issues. Additionally, the sample was drawn from a single institution in one geographic region, which limits the generalizability of the findings. However, although the participants may not be fully representative of all Korean healthcare students, medical students in Korea tend to form a relatively homogeneous group due to the highly standardized and competitive admission process. Moreover, given the curricular similarities across Korean medical schools, significant variation in AI literacy between institutions may be limited. Future studies with larger and more diverse samples from multiple institutions are essential to validate and expand upon these results. Nonetheless, the present study offers preliminary insights that may serve as a valuable baseline for further investigation into AI literacy and attitudes among healthcare students. In addition, multinational studies could provide useful comparative perspectives on how AI literacy varies across educational and cultural contexts. Longitudinal or intervention-based research examining the development of AI literacy over time may also inform the design of future curricula.
Second, the study relied on self-reported data, which may be subject to response bias and subjective interpretation. To enhance the validity of future findings, objective measures—such as knowledge-based assessments or evaluations of practical AI-related skills—should be incorporated alongside self-report instruments. Finally, the adapted SNAIL-KR scale demonstrated both feasibility and internal consistency. The results obtained using SNAIL-KR were comparable to those reported by Laupichler et al. [6], despite the exclusion of seven items from the original version. Nevertheless, further validation with larger and more diverse samples is needed to confirm its reliability and applicability.
AI Insights
Artificial Intelligence Becomes Mandatory Discipline in All Kazakh Universities

ASTANA — Artificial intelligence has become a mandatory discipline across all universities in Kazakhstan. A total of 93 universities have already integrated AI into their academic programs, while 20 institutions have launched 25 new educational tracks, said Deputy Minister of Science and Higher Education Gulzat Kobenova.
Photo credit: gov.kz
According to her, this initiative will not only strengthen the education system but also bring Kazakhstan closer to leading countries in AI adoption, such as China, Finland and the United States, reported Khabar TV channel on Aug. 28.
Starting from the new academic year, the Aisana project will be included in university curricula, enabling every student to gain hands-on experience with AI technologies. So far, 390,000 students have completed a specialized AI course, with 3,000 of them earning official certificates.
Earlier this month, President Kassym-Jomart Tokayev called for urgent action to make AI a driver of national development.
One of the largest AI projects in Kazakhstan is the creation of the Alem.ai center, which is planned to host an educational center for school children and a startup campus. The center is expected to boost exports of AI solutions to $5 billion by 2029.
AI Insights
Application and Comparative Study of Generative Artificial Intelligence for Epidemic Prediction of Coronavirus Disease
AI Insights
Brave new world of AI casts shadow over Labor Day – Sentinel and Enterprise

As artificial intelligence reshapes the workforce, we mark Labor Day Weekend in the AI age with both the traditional recognition of workers’ achievements and a new reflection on the future of labor.
AI’s creating a new economic dynamic that profoundly affects the workplace. On this particular holiday weekend, the impact of these changes should be a primary concern for workers and policymakers.
Leading into Labor Day, Massachusetts business and technology leaders hashed out strategies to strengthen the state’s competitive edge in artificial intelligence and tackle emerging skills gaps.
More than half of the individual skills considered core parts of the top 15 job types “could face moderate-to-high disruption from AI,” according to a report from the Massachusetts High Technology Council and the Boston Consulting Group, highlighted Wednesday during a virtual briefing.
The most immediate impacts could occur in structured jobs with repetitive routines, including office clerks and accountants, “where 40-50% of core skills are at risk of high-to-complete disruption as GenAI automates tasks like scheduling, recordkeeping, invoicing and compliance checks,” the report states.
The report also discusses how AI will redefine other traditional jobs, including health-care professionals who can use AI to flag clinical risks, educators who can use analytics to personalize instruction, and financial analysts using generative AI to detect fraud.
Ahead of the start of the school year, the Department of Elementary and Secondary Education last week released guidance on using AI responsibly in schools, including in ways that are safe and ethical.
Given this transformative shift, employers have begun recruiting workers for “bilingual roles,” which involves AI fluency in areas like AI models, AI-based automation and AI-assisted learning tools.
But the report shows Massachusetts has a “retention problem,” with only 40% of AI-trained graduates staying here, compared to 80% in New York and California, said Anna Senko, project leader at the Boston Consulting Group.
“We also know that peer states are really investing quite aggressively in AI infrastructure and workforce pipelines as well, which will increase that competition for talent,” Senko said. “So we’re at a crossroads here in Massachusetts, and it isn’t about adopting AI for AI’s sake, but really making sure that AI strengthens Massachusetts’ competitive advantage, especially in some of our top industries.”
The report recommends the commonwealth invest in the talent pipeline from K-12 classrooms to those in mid-career learning new skills; bolster shared AI infrastructure and strategic partnerships; compete for federal AI dollars; and deploy grants, tax incentives and other tools to businesses.
The 2024 economic development law allocated $100 million for the Massachusetts AI Hub. Since launching, the hub has trained teachers in using AI, taught high school students how to use Python programming language, awarded grants in sectors like health care and manufacturing, and organized workforce development programs, said Massachusetts AI Hub Director Sabrina Mansur.
In New York, Empire AI is supported by more than $500 million in public and private funding. Connecticut and New Jersey are also launching AI hubs with smaller investments, according to the report.
“We need to move at the speed of AI in business because we’re not competing sufficiently with peer states, even though we have, I think, the best mix of resources of any commonwealth or any state,” said Chris Anderson, Massachusetts High Technology Council’s president. “And therefore, we need to help push government partners to join us, to be responsive in the right areas at the right time where it makes sense.”
The report from MHTC and BCG indicates that AI will trigger a variety of job disruptions. While clerks and accountants may see the biggest impact, even cashiers, retail salespeople, registered nurses, waiters, fast-food workers, and janitors and cleaners will feel the impact.
“If you think about buildings becoming smarter, predictive maintenance schedules, space use tracking and even autonomous cleaning pods, it’s easy to see how the responsibilities and ways of working a janitor role would fundamentally shift,” BCG consultant Trula Rael said.
We’d urge the state’s brightest minds to also concentrate on AI’s generational impact on displacing countless highly educated college grads just entering the workforce.
According to a CBS MoneyWatch report, artificial intelligence is already replacing entry-level workers whose function can be performed by generative AI tools like ChatGPT.
Early-career employees in fields most exposed to AI have experienced a 13% drop in employment since 2022, compared to more experienced workers in the same fields and people in sectors relatively unaffected by the fast-emerging technology, according to a recent working paper from Stanford economists Erik Brynjolfsson, Bharat Chandar and Ruyu Chen.
The study adds to the growing body of research that confirms the spread of generative AI in the workplace has especially disrupted the job market for younger workers, the report’s authors said.
The research highlights two fields in particular where AI already appears to be supplanting a significant number of young workers: software engineering and customer service. From late 2022 to July 2025, entry-level employment in those areas declined by roughly 20%, according to the report, while employment for older workers in the same jobs grew.
As a society, we must confront this seismic workforce shift by threading the needle that accepts AI’s continually increasing role in our daily workplace lives while not leaving our next generation of skilled labor behind.
-
Tools & Platforms3 weeks ago
Building Trust in Military AI Starts with Opening the Black Box – War on the Rocks
-
Ethics & Policy1 month ago
SDAIA Supports Saudi Arabia’s Leadership in Shaping Global AI Ethics, Policy, and Research – وكالة الأنباء السعودية
-
Events & Conferences3 months ago
Journey to 1000 models: Scaling Instagram’s recommendation system
-
Business2 days ago
The Guardian view on Trump and the Fed: independence is no substitute for accountability | Editorial
-
Jobs & Careers2 months ago
Mumbai-based Perplexity Alternative Has 60k+ Users Without Funding
-
Funding & Business2 months ago
Kayak and Expedia race to build AI travel agents that turn social posts into itineraries
-
Education2 months ago
VEX Robotics launches AI-powered classroom robotics system
-
Podcasts & Talks2 months ago
Happy 4th of July! 🎆 Made with Veo 3 in Gemini
-
Podcasts & Talks2 months ago
OpenAI 🤝 @teamganassi
-
Mergers & Acquisitions2 months ago
Donald Trump suggests US government review subsidies to Elon Musk’s companies