Our Approach to AI
Artificial Intelligence can significantly enhance software development and digital platforms. At DrevOps, we believe AI must be used responsibly, transparently, and with strong safeguards to protect client data, intellectual property, and system integrity.
AI systems can be valuable development assistants, but they must be applied carefully and responsibly.
At DrevOps:
- AI tools are used selectively and with human oversight.
- Engineering judgement and accountability always remain with our team.
- Security, confidentiality, and data protection are prioritised above convenience.
AI technologies are treated as supporting tools rather than autonomous decision makers.
AI Usage in Open Source and Internal Work
Many AI-assisted workflows at DrevOps are used in open-source development environments and internal research activities.
These environments typically include:
- publicly available repositories
- internal tooling and experimentation
- documentation and knowledge exploration
- automation of development workflows
Open-source projects and internal tools allow us to explore AI-assisted techniques without exposing confidential client information.
Client Project Safeguards
Client projects require stricter controls.
DrevOps does not use AI tools on client repositories or infrastructure without explicit approval.
Key safeguards include:
- client code is never uploaded to public AI systems without permission
- infrastructure credentials and sensitive data are never shared with AI services
- AI systems are not granted direct repository access without client approval
- all AI-assisted outputs are reviewed by experienced engineers
These safeguards ensure that confidentiality and intellectual property remain fully protected.
Approval Process for AI Usage
If a project could benefit from AI-assisted development or AI-powered functionality, we discuss the proposed approach with the client first.
Before introducing AI tools into a client environment, we define:
- the AI tools involved
- the intended purpose
- potential data exposure risks
- safeguards and access controls
AI technologies are only introduced after written approval and clear documentation.
Human Accountability
Regardless of whether AI tools are used in a workflow, DrevOps engineers remain fully responsible for the quality, security, and correctness of delivered solutions.
All code and infrastructure changes are reviewed, tested, and validated by experienced engineers.
AI systems do not replace engineering judgement or professional responsibility.
Continuous Review
The AI landscape is evolving rapidly. We continuously review our practices to ensure they remain aligned with security best practices, open-source values, and client expectations.
Our goal is to ensure organisations can benefit from AI innovation while maintaining trust, transparency, and control.
Questions About AI Usage
If your organisation has specific requirements or policies regarding AI technologies, our team is happy to discuss them.