There are plenty of older aircraft in service to pick up the slack. This won't have any affect on airlines.
Totally normal I'm sure. /shttps://edition.cnn.com/2019/04/30/politics/boeing-sensor-737-max-faa/index.html
Boeing relied on single sensor for 737 Max that had been flagged 216 times to FAA
It's crazy to me. People say putting profits over safety, but how much are we actually talking about here? Coding to error check an easily damaged sensor against another sensor couldn't have been hard or expensive in the design phase. That just leaves us with ineptness and I'm not sure that's any better than profit making decisions
The FAA has received at least 216 reports of AOA sensors failing or having to be repaired, replaced or adjusted since 2004, according to data from the FAA's Service Difficulty Reporting website.
Those reports, about one-fifth of which involve Boeing planes, include incidents in which AOA sensors were frozen, improperly installed, struck by lightning or even hit by flying birds. In some cases, faulty sensors led to stall warnings, forcing pilots to abort takeoffs or perform emergency landings.
Yup, that's why Airbus installs three of them on each of their planes. One can be ignored when it gets faulty. Curious why Boeing didn't do the same.From the article:
216 refers to total AoA sensor issues reported on all airplanes from any manufacturer. Cited issues include all causes of problems including improper installation, bird strikes and lightning.
Yup, that's why Airbus installs three of them on each of their planes. One can be ignored when it gets faulty. Curious why Boeing didn't do the same.
Yup, that's why Airbus installs three of them on each of their planes. One can be ignored when it gets faulty. Curious why Boeing didn't do the same.
The issue with the way it's presented is 216 reported issues over all flight operations the last 15 years is a very small rate of failure. If it doesn't violate the failure tolerance for the relevant DAL rating I'm not sure what should have been done differently.
well... arguably because Airbus is European and Boeing is american. As far as I am aware, European "public" companies do not have feduciary requirements towards their shareholders, meaning they are not required by law to make decisions which would result in maximized profits. As such, they are more flexible to... you know, spend a few extra cent on making sure stuff doesn't kill people.
what should have been done differently about a sensor that was required for the plane to function that is known to fail sometimes...? really?
so I guess you're still defending that despicable corporation? got it.
It's crazy to me. People say putting profits over safety, but how much are we actually talking about here? Coding to error check an easily damaged sensor against another sensor couldn't have been hard or expensive in the design phase. That just leaves us with ineptness and I'm not sure that's any better than profit making decisions
The tolerance threshold should have been lower. That's what should have been done. If you are talking about policy and ratings the DAL rating, whatever it is, isn't good enough. Simple as that.The issue with the way it's presented is 216 reported issues over all flight operations the last 15 years is a very small rate of failure. If it doesn't violate the failure tolerance for the relevant DAL rating I'm not sure what should have been done differently.
It only didn't do that because Boeing lied to the FAA about the amount of influence MCAS had on the plane's systems. But if you've read the articles published in the last few weeks on that matter, you already know that.The issue with the way it's presented is 216 reported issues over all flight operations the last 15 years is a very small rate of failure. If it doesn't violate the failure tolerance for the relevant DAL rating I'm not sure what should have been done differently.
Uh...all planes rely on systems that fail "sometimes." The goal is knowing how often "sometimes" occurs and categorizing and dealing with it effectively. And "effectively" is also a probabilistic distribution.
The tolerance threshold should have been lower. That's what should have been done. If you are talking about policy and ratings the DAL rating, whatever it is, isn't good enough. Simple as that.
It only didn't do that because Boeing lied to the FAA about the amount of influence MCAS had on the plane's systems. But if you've read the articles published in the last few weeks on that matter, you already know that.
You know how often it happens, more than 10 sensors will fail every year on average.
Now, what is the solution? Install the sole sensor and cross-fingers that it does not fail... or install 3 sensors and have the safeguard software automatically reject one sensor's data should it fail or send faulty data?
And you don't think there was any need for the software to be designed like that in the first place?And to be clear there are 2 AoA sensors already. The software is being updated to reject inputs when they disagree.
But I didn't mention 0. I know you're an engineer from previous discussions, I am as well. And I was a damn good one. Whatever threshold they have is not good enough. There is no getting around that fact. I don't care what it is, it's not low enough.Lower doesn't mean 0, and a failure rate of 0 is impossible. There are so many millions of flight hours that accidents will absolutely occur no matter what, especially when that failure is largely dependent on humans.
But I didn't mention 0. I know you're an engineer from previous discussions, I am as well. And I was a damn good one. Whatever threshold they have is not good enough. There is no getting around that fact. I don't care what it is, it's not low enough.
Better quality control, testing and certification would delay when they could finish the plane, thus losing orders. Proper retraining of pilots to this new plane that does not behave like old 737:s would cost airlines and lead to Boeing losing orders. There's a strong case for greed being the reason for the 737 max crashes.It's crazy to me. People say putting profits over safety, but how much are we actually talking about here? Coding to error check an easily damaged sensor against another sensor couldn't have been hard or expensive in the design phase. That just leaves us with ineptness and I'm not sure that's any better than profit making decisions
But this is based on a single sensor design, why is there no redundancy? And of course the threshold isn't low enough if it ended in two planes crashing without the specific model of plane being grounded previously. There can be arguments based on that, but I will be on the side that says the threshold isn't low enough.I'm not defending Boeing but you are taking a weird stance here and throwing out being an engineer makes it even weirder.
If you were a damn good one, you would have done the math on an average of 14 failures a year on 15.8 million flights a year.
A failure rate of 0.0000088608% is something that will be seen as 0 for most tolerances
I'm not defending Boeing but you are taking a weird stance here and throwing out being an engineer makes it even weirder.
If you were a damn good one, you would have done the math on an average of 14 failures a year on 15.8 million flights a year.
A failure rate of 0.0000088608% is something that will be seen as 0 for most tolerances
I will also go back to your math here a bit. This is executive math, this is what a Boeing Executive would present to FAA based on a 3rd year associates powerpoint deck. 15.8 million toltal flights a year, 14 sensor failures....divide 14/15.8....hey look!!! Almost 0 percent failure, let's all pat each other on the back.I'm not defending Boeing but you are taking a weird stance here and throwing out being an engineer makes it even weirder.
If you were a damn good one, you would have done the math on an average of 14 failures a year on 15.8 million flights a year.
A failure rate of 0.0000088608% is something that will be seen as 0 for most tolerances
I will also go back to your math here a bit. This is executive math, this is what a Boeing Executive would present to FAA based on a 3rd year associates powerpoint deck. 15.8 million toltal flights a year, 14 sensor failures....divide 14/15.8....hey look!!! Almost 0 percent failure, let's all pat each other on the back.
No.
The modeling has to be a whole lot more nuanced and critical than that.
14 failure on a critical sensor that takes control of the plane and pushes its nose down. Based on failure of a single contingency design of sensor,altitude and speed of plane, nose being pushed down, what is the amount of time required to assume control and stabilize the flight?
Then iterate this model for different speeds and altitudes.
Then re model using n-2, n-3, n-4 design contingency scenarios.
It's a shitshow. And at the cost of many, many lives. Not excusable IMO.It also doesn't somehow proof that the design of the flight control & sensors was flawless. The failure rate is just a result of the sensor being robust and study.
Nope. By this time next year people will have forgotten about this whole thing.This plane has to be done now, hasn't it. Their failure to provide a fix in any timely manner just further highlights the systemic nature of the problem, namely that the design of the plane is structurally unsound. Even if they deliver a patch, I still wouldn't book a flight if it said 737MAX in the foreseeable future. If I want a static unstable plane I'll book a fighter jet thank you very much. Maybe the conman in chief was right and the only way to salvage this thing is to give it an unrecognizable name.
Well, in automotive industry (which is far more tolerant than aircraft industry), you'd usually try for at the very least less than one critical error every million hour (I think even more, but I don't remember the exact number) for a system that overrides the driver; the goal for autonomous cars is one in a billion hours, though it's unlikely to ever be reached.I'm not defending Boeing but you are taking a weird stance here and throwing out being an engineer makes it even weirder.
If you were a damn good one, you would have done the math on an average of 14 failures a year on 15.8 million flights a year.
A failure rate of 0.0000088608% is something that will be seen as 0 for most tolerances
The url is delightfully silly.I guess the boeing defenders are in favor of this as well
https://www.theguardian.com/us-news/2019/may/03/boeing-union-workers-fired-south-carolina
But this is based on a single sensor design, why is there no redundancy? And of course the threshold isn't low enough if it ended in two planes crashing without the specific model of plane being grounded previously. There can be arguments based on that, but I will be on the side that says the threshold isn't low enough.
I used to design n-2 substations, but why do you think hospitals have their own power generation even if they are being supplied via n-2 designed substations?
Edit: Also, you are literally defending Boeing by the way.
Boeing has admitted that it knew about a problem with its 737 Max jets a year before the aircraft was involved in two fatal accidents, but took no action.
The US Federal Aviation Administration told Reuters news agency that Boeing had not informed it of the software issue until November 2018, a month after the Lion Air crash.
https://www.theverge.com/2019/5/2/18518176/boeing-737-max-crash-problems-human-error-mcas-faaWhenever it adds a new airplane to a type certificate, the FAA lists where that airplane does or does not differ from other models in the same type. In the case of the 737 Max, the FAA's list extends to 30 pages, reviewing everything from engine noise to de-icing systems, aluminum fatigue to security doors.
Yet this document dedicated to minutiae does not mention MCAS once — not by name, not by description — which is kind of astonishing when you consider that even the seat belts get a mention.
The FAA overlooked MCAS in other places, too.
As part of its certification review, the FAA assigns a "failure condition" to each system, which is basically a guess as to what would happen if that system were to break. The lowest-severity systems should only cause "some inconvenience" to passengers, while the more serious "hazardous" and "catastrophic" failure conditions can endanger the aircraft and its passengers. The more severe the failure condition, the more redundancies that system is supposed to have.
At least, that's the theory. MCAS received a "hazardous failure" designation. This meant that, in the FAA's judgment, any kind of MCAS malfunction would result in, at worst, "a large reduction in safety margins" or "serious or fatal injury to a relatively small number of the occupants." Such systems, therefore, need at least two levels of redundancy, with a chance of failure less than 1 in 10 million.
MCAS, however, does not meet any of these standards.
It has no redundancy: it takes input from just one AoA sensor at a time. That makes MCAS completely unable to cope with a sensor malfunction. It can't "sanity check" its data against a second sensor or switch to a backup if the original source fails. It just believes whatever data it's given, even if that data is bad, which is what happened on Lion Air flight 610 and Ethiopian Airlines flight 302.
It gets worse: over the last five years, 50 flights on US commercial airplanes experienced AoA sensor issues, or about one failure for every 1.7 million commercial flight-hours. Sure, that's a low rate, but it's still nearly six times above what the FAA allows for "hazardous" systems: they're only supposed to fail once every 10 million flight-hours.
Worse still: the FAA did not catch the fact that the version of MCAS actually installed on the 737 Max was much more powerful than the version described in the design specifications. On paper, MCAS was only supposed to move the horizontal stabilizer 0.6 degrees at a time. In reality, it could move the stabilizer as much as 2.5 degrees at a time, making it significantly more powerful when forcing the nose of the airplane down.
"Although officials were aware of the changes," The New York Times reported, "none were fully examined by the FAA."
So had anyone checked, they might have flagged MCAS for one of several reasons, including its lack of redundancy, its unacceptably high risk of failure, or its significant increase in power to the point that it was no longer just a "hazardous failure" kind of system.
There needs to be a serious top-down culture change at Boeing at this point
in short... don't bank on the US industries regulating themselves - ever.
And another example of how the US is falling apart because of un-checked, crony capitalism. I'm glad that the EU is putting people and safety before profit.somewhat of an update:
Looks like the FAA wants to get these planes back in the air, but Easa refuses to accept further FAA certification and issued three demands:
1) design changes must first be approved by the Easa
2) an independent body must test the planes for airworthiness
3) the crew must undergo Max-specfic re-training
https://derstandard.at/2000103645226/Europaeer-bremsen-bei-Neuzulassung-von-Ungluecks-Boeing
in short... don't bank on the US to regulate its own industries - ever.
And another example of how the US is falling apart because of un-checked, crony capitalism. I'm glad that the EU is putting people and safety before profit.