AI fabricates and hallucinates (or whatever term you want to use) by design - it doesn't provide a neutral record the way audio recordings might.
I think a better question might be what, specifically, does AI bring to the table that justifies its inherent inaccuracy? Why not just use audio recordings, etc?
LLMs, when trained on a language, provide a neutral representation of the statistics of that data set. Used correctly, that makes them perfect for this kind of task. You're reacting to global misuse of llms as databases, but llms as language models is what they are designed for. It's in the name.
Used judiciously as a model, there's no problem. They only become a problem when people try to treat them as a database.
I think a better question might be what, specifically, does AI bring to the table that justifies its inherent inaccuracy? Why not just use audio recordings, etc?