Connect with us

News

Virtual reality haptic “smart suit” uses AI with biometrics to simulate real-world environments

Published

on

One of the innovative future technologies that was on display at this year’s Consumer Electronics Show (CES 2019) in Las Vegas is the Teslasuit, a full body virtual reality haptic suit that delivers real sensations to users that mimic the environment of a digital simulation or game. In a haptic feedback system, stimuli are used to provoke real sensations involving touch – shocks, punches, pushes, bumps, etc. Teslasuit combines this sort of environmental feedback along with climate control, motion capture, and biometric systems to give a fully responsive, immersive experience to the user. Combined with true-to-life virtual reality graphics, this device could finally provide the experience VR was intended to have since its development began.

The full-body sensations of the Teslasuit seem to indicate a new level of experience for virtual reality users, but the haptic capabilities aren’t the only things making it stand out. The suit’s biometric system is designed to use machine learning to analyze heart rate, stress levels, and overall mental and emotional states to create experiences catered to the user. How this capability will be implemented is really up to game developers, though. As data is gathered from biometrics, capabilities will follow. “This is very important for the gaming industry. We see that in the future, when we come to the end user market, that we will be able to offer a lot of data sensing for the developers to process, for the AI itself to adjust the game to the player,” Dimitri Mikhalchuk, co-founder of Teslasuit, explained in an interview with Digital Trends. The suit was first officially unveiled at CES 2018.

Teslasuit uses 68 channels embedded throughout the suit’s fabric to deliver electrical stimulations, and the company is working to expand on that number. Also included in the suit’s software is a haptic library, enabling game developers to create their own effects to correspond with their virtual worlds. The climate control system adjusts the temperature of the suit, heating and cooling in accordance with the virtual environment. Finally, its motion capture and avatar system track motions from the users entire body to interact with virtual environments (and built with multiple players in mind), something that’s more or less an expansion on the core of VR to begin with.

While the Teslasuit brings virtual worlds closer to a potentially action-packed reality, bruises are not part of the experience, according to Mikhalchuk. The electric stimulation is matched to the users own strength, adapting its responses to the input. In other words, the force a user “hits” with in a game is the amount of force used to “hit” back, i.e., electric pulses make the muscles contract within that range.

Advertisement

With the equivalent of a full body shocking device wrapped around players, the question of security arises, yet another thing Teslasuit has addressed in its design. Its software has military grade encryption (AES 256) built into its wireless control system to ensure users’ bodies are not (literally) hacked into. Among its many features, Teslasuit is also completely wireless, boasts a 10+ hour battery life, and is machine washable (important for sweat-inducing virtual scenarios and general funk from frequent wear). A summary of the device on the Teslasuit website reads like something from a science fiction movie:

A highly developed form of computer modeling allows an operator to immerse in the artificial world. The user can act directly in it with the help of special sensory devices that link movements with audio-visual effects. In this case, the user’s visual, auditory, tactile, and motor sensations are replaced by their imitation, generated by a computer system with full body haptic VR suit. At the same time, biometric system of the Teslasuit analyzes human performance and health (workforce and human performance analytics). – Teslasuit.io

The future of augmented reality and virtual reality applications have been explored in fictional storylines over the years with the most recent and directly relevant one being the haptic system imagined in the movie Ready Player One. The Battlestar Galactica prequel series “Caprica” also imagined an immersive virtual world that users could “feel” experiences in (or otherwise do things they wouldn’t/couldn’t in the real world), but that didn’t require a suit, just a mind. In an episode of Netflix’s Black Mirror series called “Playtest”, an American traveler in the UK signs up to demo an immersive VR horror game to earn some money. Unlike the Teslasuit, however, the experience in that episode involved an implant to the user’s neck called a “mushroom” which linked up the VR experience with a user’s brain. The Matrix also suggested an immersive environment, although people weren’t exactly voluntarily involved, if they ever knew it wasn’t real to begin with.

Other than gaming, Teslasuit is suggestive of several solutions for non-entertainment industries. According to an article on their company website, several VR-training solutions were available for demonstration at CES 2019: Astronaut VR-training in an ISS module, an oil-loading ramp operation VR-simulator, emergency evacuation VR-training, and a powerplant VR training simulator. Teslasuit is currently continuing its outreach and collaboration with game companies to develop content. Its software development kit and corresponding system tools and applications are included with the purchase of the suit which began shipping B2B in the third quarter of 2018.

Watch the below video for a first-hand review of Teslasuit:

Advertisement

Accidental computer geek, fascinated by most history and the multiplanetary future on its way. Quite keen on the democratization of space. | It's pronounced day-sha, but I answer to almost any variation thereof.

Advertisement
Comments

Elon Musk

Tesla Terafab set for launch: Inside the $20B AI chip factory that will reshape the auto industry

Tesla set to launch “Terafab Project: A vertically integrated chip fabrication effort combining logic processing, memory, and advanced packaging.

Published

on

By

Tesla is making one of the boldest bets in its history. On March 14, Elon Musk posted on X that the “Terafab Project launches in 7 days,” pointing to March 21, 2026 as the start date for what he has described as a vertically integrated chip fabrication effort combining logic processing, memory, and advanced packaging.

Tesla first confirmed Terafab on its January 28, 2026 earnings call, where Musk told investors the company needs to build a chip fabrication facility to avoid a supply constraint projected to materialize within three to four years. But the seeds were planted even earlier. At Tesla’s annual general meeting last year, Musk warned that even in the best-case scenario for chip production from their suppliers, it still wouldn’t be enough, and declared that building a “gigantic chip fab” simply had to be done.

While there has been no official announcement on where Tesla plans to break ground on the massive Terafab, all signs point to the North Campus of Giga Texas in Austin.

Months of speculation has surrounded Tesla’s North Campus expansion at Giga Texas, where drone footage captured by observer Joe Tegtmeyer revealed massive construction site preparation just north of the existing factory on a scale that rivals the original Giga Texas footprint itself.

Advertisement

Samsung’s Tesla AI5/AI6 chip factory to start key equipment tests in March: report

The project is projected to produce 100–200 billion AI and memory chips annually, targeting 100,000 wafer starts per month, at an estimated cost of $20 billion. Tesla is targeting 2-nanometre process technology and anticipated to be the most advanced node currently in commercial production. Dubbed the Tesla AI5 chip, the chip will pack 40x–50x more compute performance and 9x more memory than AI4, and will be among the first products Terafab factory is set to produce. This highly optimized, and massively powerful inference chip is designed to make full self-driving (FSD) and Tesla’s Optimus robots faster, safer, and with full autonomy.

tesla-optimus-pilot-production-line

(Credit: Tesla)

This is where Terafab becomes a genuine game-changer. If Tesla successfully builds a 2nm chip fab at scale, it becomes one of only a handful of entities that’s capable of producing AI silicon in-house, with competitive implications that extend far beyond Tesla’s own vehicles, and potentially positioning Tesla as a chip supplier or licensor to other industries.

The next-gen Tesla AI chips will power advancements in Full Self-Driving software, the Cybercab Robotaxi program, and the Optimus humanoid robot line. Musk’s projections for Optimus require chip volumes that no existing external supplier can commit to on Tesla’s timeline.Competitors like Waymo and GM’s Cruise remain dependent on third-party silicon, leaving them exposed to the same supply chain vulnerabilities Tesla is now working to eliminate entirely.

The Terafab launch this week may not mean a factory opens its doors overnight, but it signals Tesla is serious about owning the entire AI stack, from software to silicon.

Advertisement
Continue Reading

Elon Musk

What is Digital Optimus? The new Tesla and xAI project explained

At its core, Digital Optimus operates through a dual-process architecture inspired by human cognition.

Published

on

Credit: Grok

Tesla and xAI announced their groundbreaking joint project, Digital Optimus, also nicknamed “Macrohard” in a humorous jab at Microsoft, earlier this week.

This software-based AI agent is designed to automate complex office workflows by observing and replicating human interactions with computers. As the first major outcome of Tesla’s $2 billion investment in xAI, it represents a powerful fusion of hardware efficiency and advanced reasoning.

Tesla announces massive investment into xAI

At its core, Digital Optimus operates through a dual-process architecture inspired by human cognition.

Advertisement

Tesla’s specialized AI acts as “System 1”—the fast, instinctive executor—processing the past five seconds of real-time computer screen video along with keyboard and mouse actions to perform immediate tasks.

Advertisement

xAI’s Grok model serves as “System 2,” the strategic “master conductor” or navigator, providing high-level reasoning, world understanding, and directional oversight, much like an advanced turn-by-turn navigation system.

When combined, the two can create a powerful AI-based assistant that can complete everything from accounting work to HR tasks.

Will Tesla join the fold? Predicting a triple merger with SpaceX and xAI

The system runs primarily on Tesla’s low-cost AI4 inference chip, minimizing expensive Nvidia resources from xAI for competitive, real-time performance.

Advertisement

Elon Musk described it as “the only real-time smart AI system” capable, in principle, of emulating the functions of entire companies, handling everything from accounting and HR to repetitive digital operations.

Timelines point to swift deployment. Announced just days ago, Musk expects Digital Optimus to be ready for user experience within about six months, targeting rollout around September 2026.

It will integrate into all AI4-equipped Tesla vehicles, enabling parked cars to handle office work during downtime. Millions of dedicated units are also planned for deployment at Supercharger stations, tapping into roughly 7 gigawatts of available power.

Digital Optimus directly supports Tesla’s broader autonomy strategy. It leverages the same end-to-end neural networks, computer vision, and real-time decision-making tech that power Full Self-Driving (FSD) software and the physical Optimus humanoid robot.

By repurposing idle vehicle compute and extending AI4 hardware beyond driving, the project scales Tesla’s autonomy ecosystem from roads to digital workspaces.

Advertisement

As a virtual counterpart to physical Optimus, it divides labor: software agents manage screen-based tasks while humanoid robots tackle physical ones, accelerating Tesla’s vision of general-purpose AI for productivity, Robotaxi fleets, and beyond.

In essence, Digital Optimus bridges Tesla’s vehicle and robotics autonomy with enterprise-scale AI, promising massive efficiency gains. No other company currently matches its real-time capabilities on such accessible hardware.

It really could be one of the most crucial developments Tesla and xAI begin to integrate, as it could revolutionize how people work and travel.

Advertisement
Continue Reading

News

Tesla adds awesome new driving feature to Model Y

Tesla is rolling out a new “Comfort Braking” feature with Software Update 2026.8. The feature is exclusive to the new Model Y, and is currently unavailable for any other vehicle in the Tesla lineup.

Published

on

Credit: Tesla

Tesla is adding an awesome new driving feature to Model Y vehicles, effective on Juniper-updated models considered model year 2026 or newer.

Tesla is rolling out a new “Comfort Braking” feature with Software Update 2026.8. The feature is exclusive to the new Model Y, and is currently unavailable for any other vehicle in the Tesla lineup.

Tesla writes in the release notes for the feature:

“Your Tesla now provides a smoother feel as you come to a complete stop during routine braking.”

Advertisement

Interestingly, we’re not too sure what catalyzed Tesla to try to improve braking smoothness, because it hasn’t seemed overly abrupt or rough from my perspective. Although the brake pedal in my Model Y is rarely used due to Regenerative Braking, it seems Tesla wanted to try to make the ride comfort even smoother for owners.

Advertisement

There is always room for improvement, though, and it seems that there is a way to make braking smoother for passengers while the vehicle is coming to a stop.

This is far from the first time Tesla has attempted to improve its ride comfort through Over-the-Air updates, as it has rolled out updates to improve regenerative braking performance, handling while using Full Self-Driving, improvements to Steer-by-Wire to Cybertruck, and even recent releases that have combatted Active Road Noise.

Tesla set to activate long-awaited Cybertruck feature

Tesla holds a unique ability to change the functionality of its vehicles through software updates, which have come in handy for many things, including remedying certain recalls and shipping new features to the Full Self-Driving suite.

Advertisement

Tesla seems to have the most seamless OTA processes, as many automakers have the ability to ship improvements through a simple software update.

We’re really excited to test the update, so when we get an opportunity to try out Comfort Braking when it makes it to our Model Y.

Continue Reading