This celebrated startup vowed to avoid wasting lives with AI. Now, it’s a cautionary story


Be cautious of any firm that claims to be saving the world utilizing synthetic intelligence.

Final week, the New York Instances printed an investigation of One Concern, a platform designed to assist cities and counties create catastrophe response plans. The corporate claimed to make use of a plethora of information from totally different sources to foretell the way in which that earthquakes and floods would impression a metropolis on a building-by-building foundation with 85% accuracy, inside 15 minutes of a catastrophe hitting a metropolis. However the Instances reviews that San Francisco, one of many first cities that had signed on to make use of One Concern’s platform, is ending its contract with the startup because of issues concerning the accuracy of its predictions.

The Instances paints an image of a slick interface (which was honored in Quick Firm‘s 2018 Innovation by Design awards and 2019 World Altering Concept awards) that hid issues. The warmth map-style interface is meant to indicate metropolis officers near real-time predictions of injury after an earthquake or flood, in addition to run simulations of future earthquakes and supply injury ranges for every block, serving to planners resolve the way to distribute sources to achieve individuals who will likely be most in want of assist.

As I wrote again in November 2018 of One Concern’s interface:

It’s nearly like enjoying SimCity, the place planners click on on a fault, watch what occurs to every constructing, after which add icons like sandbags, shelters, or hearth vehicles to see how these preparation ways affect the simulation. All of this occurs inside a comparatively easy color-coded map interface, the place customers toggle on totally different layers like demographics and demanding infrastructure to know what the injury means in additional depth.

It was this easy-to-use design that satisfied San Francisco’s former emergency administration director to signal on to make use of the platform as a result of it was a lot less complicated and extra intuitive than a free service supplied by FEMA to foretell earthquake injury.

However the technical sophistication simply wasn’t there, in line with the report. An worker in Seattle’s emergency administration division advised the Instances that One Concern’s earthquake simulation map had gaping holes in business neighborhoods, which One Concern stated was as a result of the corporate depends totally on residential census knowledge. He discovered the corporate’s assessments of future earthquake injury unrealistic: The constructing the place the emergency administration division works was designed to be earthquake protected, however One Concern’s algorithms decided that it could have heavy injury, and the corporate confirmed bigger than anticipated numbers of at-risk constructions as a result of it had calculated every condominium in a high-rise as a separate constructing. This worker shared all of those points with the Instances.

One Concern declined to remark publicly on the report. Within the Instances story, One Concern’s CEO and cofounder Ahmad Wani says that the corporate has repeatedly requested cities for extra knowledge to enhance its predictions, and that One Concern just isn’t attempting to interchange the judgment of skilled emergency administration planners.

Many former workers shared misgivings concerning the startup’s claims, and in contrast to rivals like flood prediction startup Fathom, none of its algorithms have been vetted by impartial researchers with the outcomes printed in tutorial journals. The Instances reviews: “Equally, One Concern’s earthquake simulations depend on FEMA’s free damage-prediction methodology referred to as P58, with calculations carried out by one other firm, Haselton Baker Threat Group,” in addition to extensively obtainable free public knowledge, whereas charging cities like San Francisco $148,000 to make use of the platform for 2 years. Moreover, the Instances discovered that One Concern has began to work with insurance coverage corporations, which may use its catastrophe predictions to lift charges, partially as a result of just a few cities have paid for its product thus far—a transfer that triggered some former workers to really feel disillusioned with the corporate’s mission.

Because the Instances investigation reveals, the startup’s early success—with $55 million in enterprise capital funding, advisors akin to a retired common, the previous CIA director David Petraeus, and group members like the previous FEMA head Craig Fugate—was constructed on deceptive claims and modern design.

Pleasure over how synthetic intelligence may repair seemingly intractable issues definitely didn’t assist. I wrote final yr of One Concern’s potential: “As local weather change heralds extra devastating pure disasters, cities might want to rethink how they plan for and reply to disasters. Synthetic intelligence, such because the platform One Concern has developed, gives a tantalizing answer. Nevertheless it’s new and largely untested.”

With defective know-how that’s reportedly not as correct as the corporate says, One Concern could possibly be placing folks’s lives in danger.

Comments are closed.