News
Google’s DeepMind unit develops AI that predicts 3D layouts from partial images
Google’s DeepMind unit, the same division that created AlphaGo, an AI that outplayed the best Go player in the world, has created a neural network capable of rendering an accurate 3D environment from just a few still images, filling in the gaps with an AI form of perceptual intuition.
According to Google’s official DeepMind blog, the goal of its recent AI project is to make neural networks easier and simpler to train. Today’s most advanced AI-powered visual recognition systems are trained through the use of large datasets comprised of images that are human-annotated. This makes training a very tedious, lengthy, and expensive process, as every aspect of every object in each scene in the dataset has to be labeled by a person.
The DeepMind team’s new AI, dubbed the Generative Query Network (GQN) is designed to remove this dependency on human-annotated data, as the GQN is designed to infer a space’s three-dimensional layout and features despite being provided with only partial images of a space.
Similar to babies and animals, DeepMind’s GQN learns by making observations of the world around it. By doing so, DeepMind’s new AI learns about plausible scenes and their geometrical properties even without human labeling. The GQN is comprised of two parts — a representation network that produces a vector describing a scene and a generation network that “imagines” the scene from a previously unobserved viewpoint. So far, the results of DeepMind’s training for the AI have been encouraging, with the GQN being able to create representations of objects and rooms based on just a single image.
As noted by the DeepMind team, however, the training methods that have been used for the development of the GQN are still limited compared to traditional computer vision techniques. The AI creators, however, remain optimistic that as new sources of data become available and as improvements in hardware get introduced, the applications for the GQN framework could move over to higher-resolution images of real-world scenes. Ultimately, the DeepMind team believes that the GQN could be a useful system in technologies such as augmented reality and self-driving vehicles by giving them a form of perceptual intuition – extremely desirable for companies focused on autonomy, like Tesla.

Google DeepMind’s GQN AI in action. [Credit: Google DeepMind]
In a talk at Train AI 2018 last May, Tesla’s head of AI Andrej Karpathy discussed the challenges involved in training the company’s Autopilot system. Tesla trains Autopilot by feeding the system with massive data sets from the company’s fleet of vehicles. This data is collected through means such as Shadow Mode, which allows the company to gather statistical data to show false positives and false negatives of Autopilot software.
During his talk, Karpathy discussed how features such as blinker detection become challenging for Tesla’s neural network to learn, considering that vehicles on the road have their turn signals off most of the time and blinkers have a high variability from one car brand to another. Karpathy also discussed how Tesla has transitioned a huge portion of its AI team to labeling roles, doing the human annotation that Google DeepMind explicitly wants to avoid with the GQN.
Musk also mentioned that its upcoming all-electric supercar — the next-generation Tesla Roadster — would feature an “Augmented Mode” that would enhance drivers’ capability to operate the high-performance vehicle. With Tesla’s flagship supercar seemingly set on embracing AR technology, the emergence of new techniques for training AI such as Google DeepMind’s GQN would be a perfect fit for the next generation of vehicles about to enter the automotive market.
Cybertruck
Tesla Cybertruck gets long-awaited safety feature
Tesla has announced the rollout of its innovative anti-dooring protection feature to the Cybertruck via the 2026.8 software update.
Tesla is rolling out a new and long-awaited feature to the Cybertruck all-electric pickup, and it is a safety addition geared toward pedestrian and cyclist safety, as well as accidents with other vehicles.
Tesla has announced the rollout of its innovative anti-dooring protection feature to the Cybertruck via the 2026.8 software update.
This safety enhancement uses the vehicle’s existing cameras to detect approaching cyclists, pedestrians, or vehicles in the blind spot while parked. Upon attempting to open a door, if a hazard is detected, the system activates: the blind spot indicator light flashes, an audible chime sounds, and the door will not open on the initial button press.
Drivers must wait briefly and press the button again to override, providing crucial seconds to avoid an accident.
Anti-dooring protection now rolling out to @Cybertruck
This feature comes standard on every new Model 3, Model Y & Cybertruck – using cameras to delay door opening if a cyclist, pedestrian or other vehicle is detected approaching in your blind spot
— Tesla North America (@tesla_na) March 17, 2026
The feature, also known as Blind Spot Warning While Parked, comes standard on every new Model 3 and Model Y, and is now extending to the Cybertruck. Leveraging Tesla’s vision-based system without requiring new hardware, it represents a cost-effective software solution that builds on community suggestions dating back to 2018.
This technology addresses the persistent danger of “dooring,” where a driver opens a car door into the path of a passing cyclist or pedestrian.
Tesla implemented this little-known feature to make its cars even safer
Dooring incidents are alarmingly common in urban environments.
According to Chicago data, in 2011 alone, there were 344 reported dooring crashes, accounting for approximately 20 percent of all bicycle crashes in the city, nearly one incident per day.
While numbers have fluctuated (dropping to 11 percent in 2014 before rising again), dooring consistently represents 10-20 percent of bike-related crashes in major cities.
A national analysis of emergency department data estimates over 17,000 dooring-related injuries treated in the U.S. over a decade, with many involving fractures, contusions, and head trauma, particularly affecting upper extremities.
By automatically intervening, Tesla’s system not only protects vulnerable road users but also safeguards its owners from potential liability and enhances overall road safety.
As cities promote cycling for sustainable transport, features like this demonstrate how advanced driver assistance and camera systems can evolve beyond highway driving to everyday urban scenarios.
Enthusiastic responses on social media highlight appreciation for the proactive safety measure, with some calling for broader rollout to older models where hardware permits. Tesla continues to push the boundaries of vehicle safety through over-the-air updates, making its fleet smarter and safer over time.
Elon Musk
Tesla Roadster is ‘sorcery and magic’ and might be worth the wait, Uber founder says
Perhaps the wait will be worth it, especially according to Uber founder Travis Kalanick, who recently teased the Roadster’s potential capabilities based on what he has heard from internal Tesla sources.
Tesla is planning to unveil the Roadster in late April after years of waiting. But the wait might be worth it, according to Travis Kalanick, the founder of Uber, who recently shed some light on his expectations for the all-electric supercar.
We all know the Roadster is supposed to have some serious capability. CEO Elon Musk has said on numerous occasions that the Roadster will be unlike anything else ever produced. It might go from 0-60 MPH in about a second, it might hover, it might have SpaceX cold gas thrusters.
However, the constant delays in the Roadster program and its unveiling event continue to send Tesla fans into confusion because they’re just not sure when, or if, they’ll ever see the finished product.
Perhaps the wait will be worth it, especially according to Uber founder Travis Kalanick, who recently teased the Roadster’s potential capabilities based on what he has heard from internal Tesla sources.
Kalanick said on X:
When I’ve run into people who are in the know, I inquire, they tell me nothing, but their eyebrows raise and their eyes widen in a way that can only mean something of sorcery and magic is coming…
— travis kalanick (@travisk) March 17, 2026
Musk has said this vehicle is not going to be geared for safety, and that, “If safety is your number one goal, do not buy the Roadster.”
There has been so much hype regarding the Roadster that it is hard to believe the company could not come through on some kind of crazy features for the vehicle.
However, the latest delay that Tesla put on the unveiling event is definitely eye-opening, especially considering it is the latest in a series of pushbacks the company has put on the vehicle for the past several years.
Tesla has made several jumps in the Roadster project over the past few months, as it has ramped up hiring for the vehicle and also applied for a patent for a new seat design.
The car has been a back-burner project for Tesla, as it has been focusing primarily on autonomy and the rollout of Robotaxi and Cybercab. Additionally, its other vehicle projects, like the Model 3 and Model Y refreshes, took precedence.
Tesla still plans to unveil the Roadster next month, so we can hope the company can stick to this timeframe.
Cybertruck
Elon Musk clarifies viral Tesla Cybertruck accident with driver logs
Musk has come out to say that the driver logs have already shown that the driver “disengaged Autopilot four seconds before crashing,” in a post on X.
Tesla CEO Elon Musk has clarified some details regarding the viral Tesla Cybertruck accident with company driver logs, which show various metrics at the time of an incident.
The logs have been used in the past to pull responsibility off of Tesla when the automaker’s Full Self-Driving (Supervised) or Autopilot platforms are blamed for a collision or accident. It appears this will be no different.
On Tuesday, a video of a Cybertruck crashing into an overpass barrier in August 2025 was shared by Fox Business in a story that reported a woman was suing the automaker for $1 million in a liability and negligence case.
In the suit, Justine Saint Amour said that, “Something terrifying happened, without warning, the vehicle attempted to drive straight off an overpass.” Her attorney, Bob Hilliard, said Amour “tried to take control, but crashed into the barrier and was seriously injured (mostly her shoulder, neck, and back).”
The Tesla Model Y is leading China’s electric SUV segment by a wide margin
Tesla vehicle crashes are widely popular to report by mainstream media outlets because of the sensationalism of the event. Oftentimes, these outlets will include Tesla in the headline, especially because it will pique the interest of the masses, as most who read the story are waiting to see the claim that Autopilot or Full Self-Driving was the culprit of the accident.
However, Tesla has access to the logs of every vehicle in its fleet, which will show the various metrics, like whether either FSD or Autopilot was active, if the accelerator was pressed, the speed, and other important factors.
Musk has come out to say that the driver logs have already shown that the driver “disengaged Autopilot four seconds before crashing,” in a post on X.
Logs show driver disengaged Autopilot four seconds before crashing
— Elon Musk (@elonmusk) March 18, 2026
If the logs do show this, which Tesla will likely have to prove in court, the real question would be why did the Amour disengage the suite?
Tesla’s Full Self-Driving suite is still not fully autonomous, meaning the driver cannot pull attention away from the road and must be ready to take over the vehicle at all times.
It will be interesting to see how this particular case pans out, especially considering the clip that was released by the law firm starts at about four seconds before the collision. Tesla logs have dispelled media reports in the past that have accused the company’s suite of being responsible for an accident, so there will be some major attention on what is proven in this particular case.