When people hear that AI is being used in a system, one question tends to come up very quickly: “Is our data being used to train the model?”
For Development teams working with sensitive alumni information, the answer needs to be clear: No.
At AlmaReach the rule is simple: client data is never used for model training.
This applies in two directions. First, third-party LLM providers are never permitted to train on client data processed through our systems. The information used in research tasks is handled strictly for that task alone. Second, we do not use client data to train our own internal models. Even when building bespoke tools and automation, training datasets are kept entirely separate from any client information.
Why does this matter?
We want to ensure that our clients’ data remains their data. Personal data must never be absorbed into external AI systems. This prevents potential downstream exposure, safeguarding information from appearing in unrelated outputs or future model behaviour. We have clear boundaries for what data can and cannot be used for training material.
AI tools should process research tasks without turning institutional data into fuel for future models.