News
Tesla Autopilot veterans launch company to accelerate self-driving development
After working on Tesla’s Autopilot team for 2.5 years, Andrew Kouri and Erik Reed decided to start their own self-driving, AI-based company rightfully named lvl5. Together with iRobot engineer George Tal, lvl5 aims to develop advanced vision software and HD maps for self-driving cars.
Founded in 2016, lvl5 was incubated at renown Silicon Valley incubator Y Combinator and later raised $2 million in seed funding from investor Paul Buchheit, who’s a partner at Y Combinator and creator of Gmail, and Max Altman’s 9Point Ventures.

In just 3 months, lvl5 racked up almost 500,000 miles of US roadway coverage with Payver. (Photo: lvl5)
“Working with lvl5’s founders while they were at Y Combinator, it was clear they have unmatched expertise in computer vision, which is the secret sauce of their solution,” said Buchheit. “I have no doubt this is the team to make self-driving a reality in the near term.”
At the center of lvl5’s technology is their computer vision algorithms. Founder and CTO George Tall previously specialized in computer vision technology at iRobot. In addition to Tall’s experience at iRobot, Kouri and Reed’s experience at Tesla undoubtedly left them with unparalleled expertise in computer vision.
Instead of turning to expensive LiDAR technology, lvl5’s computer vision analyzes its environment for stoplights, signs, potholes, and other objects. The system can be accurate to 10cm, a notable measure considering it’s derived from simple cameras and smartphones. In comparison, LiDAR systems can cost over $80,000 but are accurate to 3cm.
- Each purple trace through the intersection contributes to building the 3D map from a 2D image. For each frame, lvl5’s computer vision technology computes the position of the vehicle relative to other objects in the intersection and create a point cloud that resembles the output from LiDAR. Each white sideways “pyramid” represents the location of a captured frame in the video trace. (Photo: lvl5)
- This image is taken from one of lvl5’s neural nets, which is designed to draw a box around the position of traffic lights in an image. (Photo: lvl5)
- With only two trips through this intersection, lvl5 can start to extract semantic features such as a stop sign. (Photo: lvl5)
- The three founders of lvl5 in front of their SF home. Left to right: Erik Reed, Andrew Kouri, George Tall (Photo: Lvl5)
So how will lvl5 map roadways in the world using their computer vision technology? Smartphones. Well, for now at least. The company has released an app called Payver that allows anyone’s smartphone to collect data while driving and get paid between $.01-$.05 per mile, depending on a number of factors. Users of the app place their phone in a mount on their dashboard and let the app gather driving data.
The data is sent to lvl5’s central hub and processed by their computer vision technology. “Lvl5 is solving one of the biggest obstacles to widespread availability of self-driving technology,” said Max Altman, one of lvl5’s seed round investors and partner at 9Point Ventures. “Without accurate and efficient HD mapping, as well as the computer vision software that enables it, self-driving vehicles will take much longer to reach mass-market. This will delay everything from safer roads to efficient delivery services.”
GIF: lvl5
“We have to make self-driving available worldwide – not just in California,” Co-Founder and CEO Andrew Kouri said in a company statement. “Our approach, which combines computer vision software, crowdsourcing and widely available, affordable hardware, means our technology is accessible and will make self-driving a reality today, rather than five years from now.”
The company has already established pilot programs with major automakers and both Uber and Lyft. Companies will pay lvl5 an initial fee to use the maps, along with a monthly subscription to keep the maps continuously updated. “Through its OEM-agnostic approach, lvl5 will be able to collect significant amounts of mapping data from millions of cars in order to scale the technology for the benefit of drivers and pedestrians around the world,” the company’s press release states.
Elon Musk
Tesla launches 200mph Model S “Gold” Signature in invite-only purchase
Tesla’s final 350-unit Signature Edition closes the book on two cars that changed everything.
Tesla has announced a super limited Signature Edition run of 250 Model S Plaid and 100 Model X Plaid units as an invite only purchase in a bid to give its original flagship vehicles a proper send-off.
When the Model S first launched in 2012, the first 1,000 units sold were “Signature” editions that required a $40,000 deposit and cost nearly $100,000 each. Those early buyers were Tesla’s first real believers. This new Signature Edition deliberately echoes that moment, bookending a 14-year run with numbered collector hardware.
Both models are finished in an exclusive Garnet Red paint not available on any current Tesla production vehicle, with gold Tesla T badges up front, a gold Plaid badge and Signature badge at the rear, and a white Alcantara interior featuring gold Plaid seat badges, gold piping, Signature-marked door sills, and a numbered dash plate. The Model S adds carbon ceramic brakes with gold calipers. Every unit ships with Tesla’s Luxe Package, bundling Full Self-Driving (Supervised), four years of Premium Service, free lifetime Supercharging, and a Signature Edition key fob. Both are priced at $159,420, a roughly $35,000 premium over standard Plaid inventory.
The discontinuation is part of a broader strategic shift. At Tesla’s Q4 2025 earnings call, Musk described the decision as “slightly sad” but necessary, saying: “It’s time to basically bring the Model S and X programs to an end with an honorable discharge, because we’re really moving into a future that is based on autonomy.”
The Fremont factory floor that built these cars is being converted to manufacture Optimus humanoid robots, with a target of one million units annually.
Elon Musk
Tesla FSD in Europe vs. US: It’s not what you think
Tesla FSD is approved in the Netherlands, but the European version differs from what US drivers use.
On April 10, 2026, the Dutch vehicle authority RDW granted Tesla the first European type approval for Full Self-Driving Supervised, making the Netherlands the first country on the continent to authorize Tesla’s semi-autonomous system for customer use on public roads.
As Teslarati reported, the RDW approval followed 18 months of testing, more than 1.6 million kilometers driven on EU roads, 13,000 customer ride-alongs, and documentation covering over 400 compliance requirements. Tesla Europe had been running public demo drives through cities like Amsterdam and Eindhoven since early 2026, giving passengers their first experience of the system on European streets.
The European version of FSD is not the same software US drivers use. The RDW’s own statement is direct, noting that the software versions and functionalities in the US and Europe “are therefore not comparable one-to-one.” We’ve compile a table below that captures the most significant differences between US-based Tesla FSD vs. European Tesla FSD that’s based on what regulators and Tesla have publicly confirmed.
| Feature | FSD US | FSD Europe (Netherlands) |
| Regulatory framework | Self-certification, post-market oversight | Pre-market type approval required (UN R-171 + Article 39) |
| Hands requirement | Hands-off permitted on highway | Hands must be available to take over immediately |
| Auto turning from stop lights | Available — navigates intersections, turns, and traffic signals autonomously | Available in EU build — confirmed in Amsterdam demo footage handling unprotected turns and signalized intersections |
| Driving modes | Multiple profiles including a more aggressive “Mad Max” mode | EU build is more conservative by default and errs on the side of restraint when it cannot confirm the limit |
| Summon | Available — Smart Summon navigates parking lots to driver | Status unclear — not confirmed as part of the RDW-approved feature set; urban FSD approval targeted separately for 2027 |
| Driver monitoring | Camera-based eye tracking | Stricter continuous monitoring with more frequent intervention alerts |
| Software version | FSD v14.3 | EU-specific builds that must be separately validated by RDW |
| Geographic restriction | US, Canada, China, Mexico, Australia, NZ, South Korea | Netherlands only; EU-wide vote pending summer 2026 |
| Subscription price | $99/month | €99/month |
| Full urban FSD scope | Available | Partial — separate urban application planned for 2027 |
The approval comes as Tesla is under real pressure to grow FSD subscriptions globally. Musk’s 2025 CEO compensation package, approved by shareholders, includes a milestone requiring 10 million active FSD subscriptions as one condition for his stock awards to vest. Tesla hit one million subscriptions during its Q4 2025 earnings call, which is a meaningful start, but still a long way from the target. Opening Europe as a market for subscriptions, rather than just hardware sales, directly accelerates that number.
Tesla has said it anticipates EU-wide recognition of the Dutch approval during summer 2026, which would extend FSD access to Germany, France, and other major markets through a mutual recognition process without each country repeating the full 18-month review. That timeline is Tesla’s projection, not a confirmed regulatory outcome. As Musk acknowledged at Davos in January 2026, “We hope to get Supervised Full Self-Driving approval in Europe, hopefully next month.”
News
Tesla’s troublesome Auto Wipers get a major upgrade
Tesla has quietly deployed a major over-the-air (OTA) update across its entire fleet, implementing a new patent that could finally solve one of the most complained-about features in its vehicles: the Auto Wipers.
One of Tesla’s most complained-about features is that of the Auto Wipers, but they have recently received a major upgrade that impacts every vehicle in the company’s fleet, a company executive confirmed.
Tesla has quietly deployed a major over-the-air (OTA) update across its entire fleet, implementing a new patent that could finally solve one of the most complained-about features in its vehicles: the Auto Wipers.
Confirmed by senior Tesla AI engineer Yun-Ta Tsai on April 10, the improvement is based on patent US 20260097742 A1. It introduces an “energy balance model” that adds a tactile, physics-driven layer to the existing camera-based system—without requiring any new hardware.
🚨 Tesla has already implemented a new patent that improves the accuracy of the Auto Wiper system https://t.co/QjjKHKxSNv pic.twitter.com/mEbd04oJAu
— TESLARATI (@Teslarati) April 10, 2026
Tesla drivers have griped about auto wipers since the company ditched traditional rain sensors in favor of Tesla Vision around 2018.
Owners routinely report the wipers failing to activate in light drizzle or mist, leaving windshields streaked and visibility dangerously reduced. Just as often, they formerly blasted into high-speed mode on dry, sunny days, screeching across glass and risking scratches or premature blade wear.
This is a rare occurrence anymore, but many owners still report the feature having the wipers perform at the incorrect speed or frequency when precipitation is falling.
Tesla has tried repeatedly to fix the problem through software alone.
Early “Deep Rain” initiatives and the 2023 Autowiper v4 update used multi-camera video and refined neural networks, with Elon Musk promising “super good” performance. The 2024.14 update added manual sensitivity boosts, and later FSD versions claimed further gains. Yet complaints persisted.
Elon Musk apologizes for Tesla’s quirky auto wipers, hints at improvements
Vision systems struggle with edge cases—glare, bugs, reflections, or faint mist—because they rely purely on visual inference rather than physical detection
The new patent takes a different approach. The car’s computer constantly measures electrical power delivered to the wiper motor. It subtracts predictable losses—internal motor friction, linkage drag, and aerodynamic resistance—leaving only the friction force between the rubber blade and windshield glass.
Water lubricates the glass, sharply reducing friction; dry or icy surfaces increase it dramatically. This real-time “tactile” data acts as an independent check on the camera’s visual cues, instantly shutting down false triggers on dry glass and fine-tuning speed for actual rain.
The system can also detect ice and auto-activate defrost heaters, while long-term friction trends alert drivers when blades need replacing.
By fusing vision with precise motor-load physics, Tesla has created a hybrid sensor that is both elegant and cost-free. Owners have waited years for reliable auto wipers; this OTA rollout may finally deliver them.















