Deep Dives Level Up Newsletters Saved Articles Challenges

Elon Musk reveals toilet trouble on SpaceX’s first private mission, promises update, pizza-warming oven and WiFi

By Monica Athnasious

Sep 22, 2021

COPY URL

Elon Musk took to Twitter on Monday 20 September to reveal that there had been some “challenges” for the Inspiration4 crew regarding the toilet used on-board SpaceX’s Crew Dragon spacecraft—which first took off Wednesday 15 September and landed the following Saturday evening. After tweeting that he had met with the crew, the SpaceX CEO began answering questions from users revealing there were issues with the bog and promising updates.

In a response to one user, Musk wrote that for future flights there would “definitely be upgraded toilets. We had some challenges with it this flight.” Also writing that there would be “a little oven for heating food and Starlink WiFi” after details surfaced that the crew ate cold pizza and were unable to send as many photos or videos of Inspiration4 because of lack of accessibility to ground stations where NASA had priority. The WiFi for future flights would be connections made through SpaceX’s Starlink satellites.

One of the civilian passengers on Inspiration4, Jared Isaacman, explained to Insider in July 2021 that Crew Dragon’s—a variant of Dragon 2—toilet facilities were positioned on the spacecraft’s ceiling and encompassed by a glass dome that showcased a 360-degree view for passengers when using the toilet. Basically, you could look out into space while you bake a loaf. “It’s not a ton of privacy. But you do have this privacy curtain that cuts across the top of the spacecraft, so you can kind of separate yourself from everyone else,” Isaacman told Insider.

He also added that part of the civilian astronauts’ intensive training for the flight involved learning how to operate the toilet on board.

The specific “challenges” are currently unclear, but according to SpaceX’s Benji Reed, the problem may be linked to the suction fan—the item responsible for the effective removal of waste. During a post-flight conference, Reed, leader of the human spaceflight programmes for SpaceX, stated, “We had a couple of issues that we worked [on], we did work something [out] on the Waste Management System but that was worked [out] fine and, you know, the crew was happy and healthy.”

Cold pizza and a toilet that doesn’t flush? Sounds like a really expensive easyJet flight to me.

Elon Musk reveals toilet trouble on SpaceX’s first private mission, promises update, pizza-warming oven and WiFi


By Monica Athnasious

Sep 22, 2021

COPY URL


AI

New MIT study confirms Tesla’s autopilot is indeed unsafe

By Alma Fabiani

Sep 21, 2021

COPY URL

A month ago, towards the end of August 2021, the National Highway Transportation Safety Administration (NHTSA) launched an investigation into Tesla’s Autopilot system after it was found responsible for 11 accidents, resulting in 17 injuries and one death. Now a new study, conducted by the Massachusetts Institute of Technology (MIT), has confirmed how unsafe Elon Musk’s infamous autopilot feature actually is.

Titled A model for naturalistic glance behavior around Tesla Autopilot disengagements, the study backs up the idea that the electric vehicle company’s “Full-Self Driving” (FSD) system is in fact—surprise, surprise—not as safe as it claims. After following Tesla Model S and X owners during their daily routine for periods of a year or more throughout the greater Boston area, MIT researchers found that, more often than not, they become inattentive when using partially automated driving systems. Note here that I went from calling the autopilot a Full-Driving system—which is the term Tesla uses to describe it and therefore entails it is fully autonomous—to then qualifying it of an automated driving system, also known as an advanced driver assist system (ADAS), which is what it truly is.

“Visual behavior patterns change before and after [Autopilot] disengagement,” the study reads. “Before disengagement, drivers looked less on road and focused more on non-driving related areas compared to after the transition to manual driving. The higher proportion of off-road glances before disengagement to manual driving were not compensated by longer glances ahead.” To be completely fair, it does make sense that drivers would feel less inclined to be attentive when they think their car’s autopilot is fully in control. Only thing is, it isn’t.

Meanwhile, by the end of this week, Tesla will roll out the newest version of its autopilot beta software, the version 10.0.1 in this case, on public roads—completely ignoring the current federal investigation when it comes to the safety of its system. Billionaire tings, go figure.

Musk has also clarified that not everyone who has paid for the FSD software will be able to access the beta version, which promises more automated driving functions. First things first, Tesla will use telemetry data to capture personal driving metrics over a 7-day period in order to ensure drivers are still remaining attentive enough. “The data might also be used to implement a new safety rating page that tracks the owner’s vehicle, which is linked to their insurance,” added TechCrunch.

In other words, Musk is aware of the risk the current autopilot system represents, and he’s working hard on improving it, or at least making sure he’s not going to be the one to blame if more Tesla-related accidents happen. How do you say your autopilot is not an autopilot without clearly saying it—and therefore risking to hurt your brand? You release a newer version of it that can easily blame drivers for their carelessness, duh.

“The researchers found this type of behavior may be the result of misunderstanding what the [autopilot] feature can do and what its limitations are, which is reinforced when it performs well. Drivers whose tasks are automated for them may naturally become bored after attempting to sustain visual and physical alertness, which researchers say only creates further inattentiveness,” continued TechCrunch.

My opinion on Musk and Tesla aside, the point of the MIT study is not to shame Tesla, but rather to advocate for driver attention management systems that can give drivers feedback in real-time or adapt automation functionality to suit a driver’s level of attention. Currently, Tesla’s autopilot system doesn’t monitor driver attention via eye or head-tracking—two things that researchers deem necessary.

The technology in question—which is a model for glance behaviour—already exists, with automobile manufacturers like Mercedes-Benz and Ford allegedly already working on implementing it. Will Tesla follow suit or will Musk’s ‘only child’ energy rub off on the company?

New MIT study confirms Tesla’s autopilot is indeed unsafe


By Alma Fabiani

Sep 21, 2021

COPY URL


 

×

Emails suck! Ours don't

Sign up to our weekly newsletter

 

Don't show again