Connect with us

AI Insights

Trump hosts US tech leaders at White House dinner – minus Elon Musk | Elon Musk

Published

on


As Donald Trump hosted leaders from the biggest US tech companies at a lavish White House state dining room dinner on Thursday night, there was one notable absence. Elon Musk, once inseparable from Trump and a constant, contentious presence in the White House, was not in attendance.

The dinner, which included Meta’s Mark Zuckerberg, Microsoft’s Bill Gates, Apple’s Tim Cook and OpenAI’s Sam Altman, was exactly the type of event where Musk would have sat at Trump’s right hand only a few months ago. Instead, the Tesla CEO stated on his social media platform X that he had been invited but could not make it. He said he planned to send a representative and spent the day on X posting a familiar stream of attacks on immigration and trans people.

The White House did not respond for comment on why Musk would not be at the dinner.

The event, which was to have been held on the newly paved-over Rose Garden, until a forecast of thunderstorms forced the event indoors, began with televised words of praise for the president from several of the assembled tech leaders, and a brief series of questions from reporters.

Musk’s absence, even if voluntary, is a stark turnaround from when Trump repeatedly joked following the election that “Elon won’t go home, I can’t get rid of him”. The vacant seat highlights a divide that has emerged between the two men since their very public falling out earlier this year, one that has seen Musk’s influence over the government wane despite spending hundreds of millions of dollars to re-elect Trump during the 2024 election.

Musk’s omission from the list of attendees also echoes one of the seminal moments of his political evolution, another White House event. In 2022, then president Joe Biden failed to invite the Tesla CEO to a summit on electric vehicles over concerns it would draw backlash from autoworkers’ unions. Musk, who had not yet publicly aligned himself with the Republican party, lashed out at the White House for the snub and declared that he would not vote for Biden. The move proved enormously costly for Democrats.

The incident clearly stuck with Musk, who like Trump has shown a tendency to harbor long-term grudges. Even on the day of Trump’s dinner, he reserved his ire for Biden rather than the current president, retweeting a clip of himself from 2023 addressing Biden’s snub with the post “I try not to start fights, but I do finish them”.

In the ensuing years, Musk has taken a hard turn to the political right. He has turned X into a bastion of far-right influencers, whom he frequently retweets to his more than 200 million followers. He has promoted false theories about Democrats conspiring to get immigrants to illegally vote en masse and embraced far-right political parties around the world. He also became Trump’s most vocal and deep-pocketed supporter, contributing nearly $300m to the re-election campaign and Republican causes.

Musk’s support for Trump placed him in a position of immense power after the president’s inauguration as the tech mogul established and led the so-called “department of government efficiency” and its sweeping dismantling of federal agencies. It also turned him into a prominent guest at political dinners and events, only a year after the British government did not invite him to a major tech summit as he made inflammatory anti-immigrant posts that claimed a “civil war” would take place in the UK.

skip past newsletter promotion

Since Musk and Trump’s relationship imploded in May over policy differences – Musk railed against Trump’s signature One Big Beautiful Bill – which then snowballed into Musk accusing Trump of being in the files pertaining to notorious sex offender Jeffrey Epstein, the xAI CEO has all but vanished from high-profile government events. Although Trump still praises Musk as a “genius”, he told reporters on Wednesday night that Musk has “got some problems” and the two have not been seen together since their public spat.

As Musk has feuded with Trump and ceded his place in the White House, however, rival tech moguls have grown closer with the administration and filled some of the vacuum. Earlier this month, Trump hosted Cook, the Apple CEO, at the White House, who in turn gifted the president a 24-karat-gold souvenir. Meanwhile, Trump aides have discussed cutting Musk’s government contracts, according to the Wall Street Journal, only to find upon review that doing so would endanger too many key operations.

If Musk had attended Thursday’s dinner, it would have created an awkward arrangement as he is suing two of the companies whose leaders were in attendance: Apple and OpenAI, helmed by his former collaborator and now nemesis Altman. As with Trump, Musk has also attacked Gates for his ties to Epstein after the Microsoft founder accused him of “killing children” through Doge’s cuts to foreign aid.



Source link

Continue Reading
Click to comment

Leave a Reply

Your email address will not be published. Required fields are marked *

AI Insights

Top Artificial Intelligence Stocks To Watch Now – September 15th – MarketBeat

Published

on



Top Artificial Intelligence Stocks To Watch Now – September 15th  MarketBeat



Source link

Continue Reading

AI Insights

Full Slate of Hearings This Week on Broadband, Artificial Intelligence and Energy

Published

on


Congress tackles broadband, AI and energy in a busy week of hearings. Broadband Breakfast’s Resilient Critical Infrastructure Summit is on Thursday.

Full Slate of Hearings This Week on Broadband, Artificial Intelligence and Energy
Photo of the Capitol by Mark Fischer used with permission



Source link

Continue Reading

AI Insights

How Math Teachers Are Making Decisions About Using AI

Published

on


Our Findings

Finding 1: Teachers valued many different criteria but placed highest importance on accuracy, inclusiveness, and utility. 

We analyzed 61 rubrics that teachers created to evaluate AI. Teachers generated a diverse set of criteria, which we grouped into ten categories: accuracy, contextual awareness, engagingness, fidelity, inclusiveness, output variety, pedagogical soundness, user agency, and utility. We asked teachers to rank their criteria in order of importance and found a relatively flat distribution, with no single criterion emerging as one that a majority assigned highest importance. Still, our results suggest that teachers placed highest importance on accuracy, inclusiveness, and utility. 13% of teachers listed accuracy (which we defined as mathematically accurate, grounded in facts, and trustworthy) as their top evaluation criterion. Several teachers cited “trustworthiness” and “mathematical correctness” as their most important evaluation criteria, and another teacher described accuracy as a “gateway” for continuing evaluation; in other words, if the tool was not accurate, it would not even be worth further evaluation. Another 13% ranked inclusiveness (which we defined as accessible to diverse cognitive and cultural needs of users) as their top evaluation criterion. Teachers required AI tools to be inclusive to both student and teacher users. With respect to student users, teachers suggested that AI tools must be “accessible,” free of “bias and stereotypes,” and “culturally relevant.” They also wanted AI tools to be adaptable for “all teachers.” One teacher wrote, “Different teachers/scenarios need different levels/styles of support. There is no ‘one size fits all’ when it comes to teacher support!” Additionally, 11% of teachers reported utility as their top evaluation criterion (defined as benefits of using the tool significantly outweigh the costs). Teachers who cited this criterion valued “efficiency” and “feasibility.” One added that AI needed to be “directly useful to me and my students.” 

In addition to accuracy, inclusiveness, and utility, teachers also valued tools that were relevant to their grade level or other context (10%), pedagogically sound (10%), and engaging (7%). Additionally, 8% reported that AI tools should be faithful to their own methods and voice. Several teachers listed “authentic,” “realistic,” and “sounds like me” as top evaluation criteria. One remarked that they wanted ChatGPT to generate questions for coaching colleagues, “in my voice,” adding, “I would only use ChatGPT-generated coaching questions if they felt like they were something I would actually say to that adult.” 

CODE

DESCRIPTION

EXAMPLES

Accuracy

Tool outputs are mathematically accurate, grounded in fact, and trustworthy.

Grounded in actual research and sources ( not hallucinations); mathematical correctness

Adaptability

Tool learns from data and can improve over time or with iterative prompting

Continue to prompt until it fits the needs of the given scenario; continue to tailor it!

Contextual Awareness

Tool is responsive and applicable to specific classroom contexts, including grade level, standards, or teacher-specified goals.

Ability to be specific to a context / grade-level / community

Engagingness

Tool evokes users’ interest, curiosity, or excitement.

A math problem should be interesting or motivate students to engage with the math

Fidelity

Tool outputs are faithful to users’ intent or voice.

In my voice- I would only use chatGPT- generated coaching questions if they felt like they were something I would actually say to that adult

Inclusiveness

Tool is accessible to diverse cognitive and cultural needs of users.

I have to be able to adapt with regard to differentiation and cultural relevance.

Output Variety

Tool can provide a variety of output options for users to evaluate or enhance divergent thinking.

Multiple solutions, not all feedback from chat is useful so providing multiple options is beneficial

Pedagogically Sound

Tool adheres to established pedagogical best practices.

Knowledge about educational lingo and pedagogies

User Agency

Tool promotes users’ control over their own teaching and learning experience.

It is used as a tool that enables student curiosity and advocacy for learning rather than a source to find answers.

Utility

Benefits of using the tool significantly outweigh the costs (e.g., risks, resource and time investment).

Efficiency – will it actually help or is it something I already know

Table 1. Codes for the top criteria, along with definitions and examples. 

Teachers expressed criteria in their own words, which we categorized and quantified via inductive coding.

We have summarized teachers’ evaluation criteria on the chart below:

Finding 2: Teachers’ evaluation criteria revealed important tensions in AI edtech tool design.

In some cases, teachers listed two or more evaluation criteria that were in tension with one another. For example, many teachers emphasized the importance of AI tools that were relevant to their teaching context, grade level, and student population, while also being easy to learn and use. Yet, providing AI tools with adequate context would likely require teachers to invest significant time and effort, compromising efficiency and utility. Additionally, tools with high degrees of context awareness might also pose risks to student privacy, another evaluation criterion some teachers named as important. Teachers could input student demographics, Individualized Education Plans (IEPs), and health records into an AI tool to provide more personalized support for a student. However, the same data could be leaked or misused in a number of ways, including further training of AI models without consent. 

Another tension apparent in our data was the tension between accuracy and creativity. As mentioned above, teachers placed highest importance on mathematical correctness and trustworthiness, with one stating that they would not even consider other criteria if a tool was not reliably accurate or produced hallucinations. However, several teachers also listed creativity as a top criterion – a trait produced by LLMs’ stochasticity, which in turn also leads to hallucinations. The tension here is that while accuracy is paramount for fact-based queries, teachers may want to use AI tools as a creative thought-partner for generating novel, outside-the-box tasks – potentially with mathematical inaccuracies – that motivate student reasoning and discussion. 

Finding 3: A collaborative approach helped teachers quickly arrive at nuanced criteria. 

One important finding we observed is that, when provided time and structure to explore, critique, and design with AI tools in community with peers, teachers develop nuanced ways of evaluating AI – even without having received training in AI. Grounding the summit in both teachers’ own values and concrete problems of practice helped teachers develop specific evaluation criteria tied to realistic classroom scenarios. We used purposeful tactics to organize teachers into groups with peers who held different experiences with and attitudes toward AI than they did, exposing them to diverse perspectives they may not have otherwise considered. Juxtaposing different perspectives informed thoughtful, balanced evaluation criteria, such as, “Teaching students to use AI tools as a resource for curiosity and creativity, not for dependence.” One teacher reflected, “There is so much more to learn outside of where I’m from and it is encouraging to learn from other people from all over.” 

Over the course of the summit, several of our facilitators observed that teachers – even those who arrived with strong positive or strong negative feelings about AI – adopted a stance toward AI that we characterized as “critical but curious.” They moved easily between optimism and pessimism about AI, often in the same sentence. One teacher wrote in her summit reflection, “I’m mostly skeptical about using AI as a teacher for lesson planning, but I’m really excited … it could be used to analyze classroom talk, give students feedback … and help teachers foster a greater sense of community.” Another summed it up well: “We need more people dreaming and creating positive tools to outweigh those that will create tools that will cause challenges to education and our society as a whole.”



Source link

Continue Reading

Trending