[SystemSafety] Road to Damascus moment in Functional safety engineering - was FOSDEM talk by Paul Sherwood

Derek M Jones derek at knosof.co.uk
Thu Feb 13 14:34:46 CET 2025


Les,

> As far as I know there are no “currently prescribed” measures for evaluating
> the safety integrity of a large language model (LLM). At least nothing that,

The same can be said about humans.

> using the classical EN 50128esque criteria, would justify it’s deployment in a
> safety critical control system. The behaviour of this entity is a mystery even
> to it creators.

The same can be said about humans.

> My point is, “Do you really want this LLM component determining the trajectory
> of your motor vehicle when barrelling down a two lane road at a relative speed
> to the oncoming traffic of 200 km/h - when it’s developer doesn’t understand
> how it works, has not documented how it should work with a validatable
> specification and thinks it’s legitimate to foist it on you with a YOLO
> release under the cover of the name Full Self Driving?”

The same can be said about humans.

> [Phil & Rolf] They have no way to argue the predictive power of their safety
> case for real world safety outcomes other than experts say following
> prescribed engineering rigor requirements should be OK.

Sounds like how things currently work for non-LLM system.

The book "Killed by a Traffic Engineer" by Marshall is packed
with examples of traffic engineering rules that are treated
as having a scientific basis, but are not based on any research.
Traffic engineering, like software engineering, appears to be all
smoke and mirrors.

-- 
Derek M. Jones           Evidence-based software engineering
blog:https://shape-of-code.com



More information about the systemsafety mailing list