Tools & Platforms
Emperor Musk’s AI Clothes – Will Lockett’s Newsletter

Musk has been parading around in his AI clothes for a while now. With the amount he screams and shouts about AI, you’d think he invented it. Of course, like everything else Musk peddles, he had nothing to do with its invention or development, except for underpaying and overworking his engineers and being an awful, overpromising PR man. However, people aren’t just noticing that Musk’s clothes are non-existent — they are also starting to point and laugh at his skid marks and the “I Love the Nazi Man” tattoo down his back. Why? Because he just can’t seem to get his AI up and working. And there is no little blue pill to remedy this situation.
Take, for example, Tesla’s hilariously crap Robotaxi rollout. The media at large is only just cottoning on to it being a huge PR stunt.
I have gone on ad nauseam about why Tesla’s self-driving cars are completely inadequate, so if you want to know the details, read my previous article here. But the helicopter view is that, unlike other autonomous vehicles, Tesla’s system has zero redundancy or safety nets and requires a nearly 100% accurate AI — which categorically can’t exist — to be even remotely safe.
Tesla is painfully aware of this fatal flaw, with Tesla engineers whistleblowing their concerns about it to the media (read more here) and the DOJ opening an investigation (read more here). So I, along with countless other commentators, was pretty damn relieved to find out that Tesla’s Robotaxis had safety drivers. There was even mention of remote workers being able to take control of the car and drive it safely in the case of a critical disengagement.
But this kind of system isn’t impressive enough for Musk. Any Uber or Lyft driver with a Tesla who wastes their money on FSD can do the exact same thing. There is no social or investor kudos to be gained for Tesla or Musk here. And here is a hint: Musk doesn’t make money from Tesla sales. After all, his $50 billion pay packet (which is now less, thanks to Musk tanking Tesla’s valuation) was the equivalent of him getting $10,000 for every Tesla ever sold! Tesla makes substantially less profit from every car sold than that.
So, what do you do if you have bet your entire company’s valuation on autonomous technology that you simply can’t deliver on?
Fudge it.
Tesla put the safety driver in the passenger seat! Because, look, it’s a self-driving car — there is no one in the driver’s seat!
This is a dangerous move that offers no benefit other than optics.
Rather than being able to properly take over the car and drive it to safety, the only thing these safety drivers could do was press a button to bring the vehicle to a stop. Which, as anyone with a driving licence will tell you, is not always the safest option! Particularly when you consider that Robotaxis have been spotted driving into lanes of oncoming traffic.
Yet, this bafflingly shite decision wasn’t really reported on. Or at least it wasn’t until a video surfaced a few days ago that showed FSD failing and a safety driver being forced to exit the vehicle in the middle of traffic to take the driver’s seat and regain control. (watch it here).
This shows just how wildly dangerous Tesla’s Robotaxis are.
The safety driver had to take a serious risk to take control of the car. Not only that, but this incident suggests there are no remote operatives capable of taking over when things go wrong. That has been a core safety feature of all developing self-driving ride-hailing services, such as Waymo and Cruise, since day one and is routinely used to keep passengers safe. The fact that this is absent for Robotaxis, which Tesla already know have a far, far higher critical disengagement rate than any other self-driving ride-hailing service, could easily be seen as insanely negligent.
Musk is comfortable putting other people — not just the safety driver, but paying passengers and the public — in danger, all for a crappy PR stunt to cover up how bad his self-driving system actually is. And the media at large, as well as public consensus, are beginning to catch up to this horrifying fact.
However, Musk’s AI woes go far, far deeper than that.
Tools & Platforms
Meta reportedly explores using rival AI models to enhance its apps

Meta is exploring the use of AI models from Google and OpenAI to enhance its apps while advancing its own Llama AI technology.
Meta is reportedly exploring the use of artificial intelligence models developed by competitors, including Google and OpenAI, to improve AI features across its platforms. According to a report by The Information, executives at the Meta Superintelligence Lab have considered integrating Google’s Gemini model into the company’s Meta AI chatbot. The move would enable Meta to offer a more robust, conversational text-based solution for answering user search queries.
The report also indicated that Meta has held discussions about incorporating OpenAI’s technology into Meta AI and its other AI-powered features. These potential collaborations highlight Meta’s effort to strengthen its AI capabilities while continuing to develop its own large language model, Llama.
Strategic partnerships as a temporary measure
A Meta spokesperson stated that the company is taking an “all-of-the-above approach to building the best AI products,” which includes both building in-house solutions and partnering with external organisations. The report noted that while Meta is exploring external technology, the company’s primary goal is to refine and advance its own AI systems. Leveraging competitor models would only be a temporary measure to accelerate innovation and keep pace with rivals in the rapidly evolving AI market.
Meta’s interest in adopting external AI tools comes at a time when competition in generative AI development is intensifying. By accessing technologies from industry leaders such as Google and OpenAI, Meta aims to enhance user experiences on its apps while gaining insights that can help strengthen future iterations of Llama.
Internal AI adoption and recruitment efforts
The Information reported that Meta employees are already using Anthropic’s AI models to support the company’s internal coding assistant. This indicates that Meta has been integrating third-party AI solutions internally even as it invests heavily in its own research and development.
Additionally, Meta has been actively recruiting AI researchers from Google and OpenAI to enhance expertise at its Superintelligence Lab. These recruitment efforts reportedly include highly competitive compensation packages designed to attract top talent from across the AI sector.
As Meta continues to refine its AI strategy, the company’s willingness to work with external partners shows its commitment to creating cutting-edge products. The temporary reliance on competitor models could help Meta accelerate development and maintain a strong position in the AI race.
Tools & Platforms
Tech expert warns of 'alarming' AI behavior after teen's death – Fox News
Tools & Platforms
Is AI turning your travel experience into a costly trap?

in this commentary
- A look at how travel companies are using AI to automatically bill you for rental car damage, in-room infractions, and higher airfares.
- An analysis of how these automated systems can make mistakes and why the burden of proof is shifting to the consumer.
- Actionable strategies you can use to protect yourself from AI-powered price hikes and false damage claims.
Worried about every little ding on your rental car? Do you always go into “anonymous” mode on your web browser before booking airline tickets?
If you do, then you probably have AI anxiety.
Travel companies are quietly deploying artificial intelligence systems, creating an invisible web of automated billing that can cost you hundreds or thousands of dollars—often without your knowledge or consent. From Hertz’s controversial AI vehicle scanners to hotel vapor detectors that fine guests when their hairdryers overheat, to airline pricing algorithms that jack up fares based on your browsing history, these systems operate in the shadows while your wallet takes a hit.
“Technology can make travelers feel powerless,” says Raymond Yorke, a spokesman for Redpoint Travel Protection. “It’s happening now. We’ve seen everything from automated rental car damage claims to a suspicious surge in airfare driven by dynamic pricing algorithms.”
But it doesn’t have to stay that way.
The technology promises efficiency and fairness, but travelers are discovering that AI often acts more like a digital pickpocket than an impartial assistant. The systems flag false positives, make decisions without human oversight, and shift the burden of proof onto customers who have to defend themselves against algorithmic accusations.
Where are the AI traps?
Rentals have become ground zero for AI overreach. Companies like Hertz are using technology from a company called UVeye that can reportedly detect paint inconsistencies and minor damages down to a millimeter level.
But critics say these systems can’t always distinguish between existing scratches, dirt or lighting changes, and genuine new damage. And car rental companies bill customers automatically, with limited avenues for appeal.
Legal consultant and AI specialist Nicola Cain notes that human intervention only happens when a customer raises a complaint, meaning the AI’s judgment stands unless you fight back. It should be the other way around, she says.
“Human oversight needs to be built into the process,” she adds.
Hotel chains are installing sophisticated sensor networks that go far beyond traditional smoke detectors. These systems monitor vapor particles, noise levels, occupancy counts, and even Wi-Fi usage patterns.
The systems are far from perfect. Ruth Cruz recently got hit with a $250 fee for smoking in her hotel room. She says the AI registered a false positive.
🖐️ Your voice matters
Have you been hit with a surprise charge you suspect was generated by an automated system? Do you think this technology makes travel more efficient, or is it just a new way for companies to make money?
And what are your best tips for protecting yourself from these AI traps?
Share your thoughts in the comments.
“I successfully disputed the charge by explaining the technical limitations of their detection system,” says Cruz, who edits a technology website in San Jose. (These types of errors are easy to find with a little sleuthing. Hers involved a quick online search.)
Airlines are perfecting the art of AI-powered price manipulation. For years, their systems have tracked your search history, location, device type, loyalty status, and dozens of other signals to predict your willingness to pay premium prices. AI is supercharging that practice.
Thomas O’Shaughnessy, a marketing executive from St. Louis, has noticed prices jumping dramatically when he researches flights.
“The price increases weren’t random,” he says. “I believe they were caused by an AI model that changes prices based on demand, the time of booking, and even the user’s search history.”
No wonder travelers have AI anxiety. The question is, what can they do about it?
How to fight the AI
“The key to fighting back is understanding that these systems prioritize speed and automation over accuracy,” explains Frank Harrison, regional security director for the Americas at World Travel Protection. “They’re designed to extract maximum revenue while hoping customers won’t challenge algorithmic decisions. But armed with the right documentation and strategies, travelers can level the playing field.”
Here are some strategies that will help you fight AI:
- Renting a car? Channel your inner Sherlock. Do a comprehensive walk-around and take photos of your car from all angles. Focus on areas AI commonly flags, like bumpers, wheel wells, and roof surfaces. Email these videos to yourself immediately for proof of when they were taken. Document everything—every scratch, every dent, every imperfection—before accepting any rental. And remember, you can always request a different vehicle if the one you’re renting has too many dings or dents.
- Don’t let ’em track you. Use private browsing or incognito mode when you book flights or hotels. Clear your cookies between searches. Use a VPN (Virtual Private Network) to shift your location. “I’ve seen price differences of $200 or more for the same flight just by appearing to browse from different cities,” says Joey Martin, an AI expert. Also, search for fares on multiple devices and compare prices across platforms. AI pricing algorithms often show different rates to smartphone users versus desktop browsers, or to logged-in loyalty members versus anonymous searchers.
- Open your hotel window, if possible. Don’t touch anything with a price tag. It’s true, AI is monitoring the air you breathe and the location of every Coke in your minibar. You already know what to do: Don’t touch the items in your minibar and keep your hotel room ventilated. If a surprise bill arrives, respond immediately and assertively. Ask for the original AI scan data, sensor logs, or algorithmic decision records that supposedly justify the charge. Most companies will struggle to provide concrete evidence that withstands scrutiny.
Bear in mind that these strategies will evolve. AI adjusts to consumer behavior, and you’ll have to make some course corrections along the way, too.
This is the start of an AI arms race
In travel, AI is an imperfect technology, registering false positives and erroneously billing consumers. It raises prices by hundreds of dollars per ticket, believing you’ll happily pay extra for your airfare because of your location. What’s more, these systems are a black box, so when you ask for proof that you damaged a car or removed something from a room, they can’t always provide it.
In short, this is nothing more than a digital money grab, and your AI anxiety is completely justified.
We’re at the beginning of an AI arms race. Travel companies are using machine learning to maximize their revenue. It’s time to fight back.
What happens next? The travel industry is busy deploying AI everywhere. Soon, systems could monitor carry-on luggage to ensure you’re paying for every bag. Hotels could find ways of automatically billing you for every missing towel or bathrobe. Car rental companies could turn their AI resources to car interiors, earning more money from stains or messy upholstery. And don’t even get me started on cruise lines!
Assume AI is tracking your every move — because it probably is.
Rental cars: Document everything
- Take a detailed video walk-around of the car before you leave the lot.
- Photograph every existing scratch, dent, and scuff, inside and out.
- Email the files to yourself immediately to create a timestamped record.
Airfare & hotels: Go undercover
- Use a VPN to mask your location and avoid geographic price targeting.
- Always search in your browser’s private or incognito mode.
- Clear your cookies between searches to prevent tracking.
Hotel rooms: Challenge the charges
- If you get a surprise fee, immediately demand the evidence.
- Ask for the specific sensor logs or AI scan data that triggered the charge.
- Most companies will waive the fee when you challenge them for proof.
-
Tools & Platforms3 weeks ago
Building Trust in Military AI Starts with Opening the Black Box – War on the Rocks
-
Ethics & Policy1 month ago
SDAIA Supports Saudi Arabia’s Leadership in Shaping Global AI Ethics, Policy, and Research – وكالة الأنباء السعودية
-
Business2 days ago
The Guardian view on Trump and the Fed: independence is no substitute for accountability | Editorial
-
Events & Conferences3 months ago
Journey to 1000 models: Scaling Instagram’s recommendation system
-
Jobs & Careers2 months ago
Mumbai-based Perplexity Alternative Has 60k+ Users Without Funding
-
Funding & Business2 months ago
Kayak and Expedia race to build AI travel agents that turn social posts into itineraries
-
Education2 months ago
VEX Robotics launches AI-powered classroom robotics system
-
Podcasts & Talks2 months ago
Happy 4th of July! 🎆 Made with Veo 3 in Gemini
-
Podcasts & Talks2 months ago
OpenAI 🤝 @teamganassi
-
Jobs & Careers2 months ago
Astrophel Aerospace Raises ₹6.84 Crore to Build Reusable Launch Vehicle