Share this
Moving fast and breaking things with open-weight AI
by Deane Sloan on 05 November 2025
Mark Zuckerberg penned a now famous line in Facebook's 2012 SEC filing 'We have a saying: "Move fast and break things". The idea is that if you never break anything, you’re probably not moving fast enough.

AI-generated image (OpenAI DALL-E)
It's an ethos that could describe today’s open-weight generative AI ecosystem, with successive model releases massively accelerating innovation.
Open-weight models are also making real inroads within organisations, with McKinsey’s April 2025 Open source technology in the age of AI survey reporting that 63% of respondent organisations use 'open-source' AI models, thanks in part to their lower operating costs.
Add to this the concerns that Geoffrey Hinton raised, when he argued that open-sourcing big models enables bad actors to fine-tune them for harmful purposes, likening it to 'being able to buy nuclear weapons at RadioShack'.
And then the reality of so called 'abliterated' derivatives of open-weight models that are intentionally bypassing safety refusals (where abliteration refers to the 'ablation' or removal of safety alignment refusals to the point of 'obliterating' them).
While those publishing abliterated models may not have done so with malicious intent, for example citing purposes such as 'role playing', the removal of safety alignment refusals makes these models more prone to malicious use.
Less well understood is that similar issues can arise from well-intentioned modifications. For example, fine-tuned derivatives of open-weight models have been shown to produce more harmful responses than their base models - over 22 times more likely according to research conducted in May 2024 by Robust Intelligence.
This is known as 'safety alignment drift', and it is entirely preventable when you consider and remediate safety alignment across feature, training, and inference concerns. The Model Atlas gives an idea of the scale of published derivative models on Hugging Face - how many of these have undergone safety alignment remediation after modification?
There is an increasing understanding of the need for open-weight model risk management, and a recent post from the (UK) AI Security Institute Managing risks from increasingly capable open-weight AI systems is a good starting point for this. Additionally, mitigation approaches to abliteration are improving, such as the extended-refusal fine-tuning approach published in An Embarrassingly Simple Defense Against LLM Abliteration Attacks.
Given the current state of play in the open-weight space, it's not hard to argue for the kind of oversight the authors of the recent open letter to the New Zealand Parliament (RegulateAI.nz) call for. While open-weight models have resulted in innovation that we wouldn't have otherwise achieved through closed models alone, we need to have a much more robust discussion about the things we're breaking as we move fast with them.
Cloud spending continues to surge globally, but most organisations haven’t made the changes necessary to maximise the value and cost-efficiency benefits of their cloud investments. Download the whitepaper From Overspend to Advantage to learn about our proven approach to optimising cloud value.
Share this
- Agile Development (84)
- Software Development (64)
- Scrum (39)
- Business Analysis (28)
- Agile (27)
- Application Lifecycle Management (26)
- Capability Development (20)
- Requirements (20)
- Solution Architecture (19)
- Lean Software Development (17)
- Digital Disruption (16)
- IT Project (15)
- Project Management (15)
- Coaching (14)
- DevOps (14)
- Equinox IT News (12)
- IT Professional (11)
- Knowledge Sharing (10)
- Strategic Planning (10)
- Agile Transformation (9)
- Digital Transformation (9)
- IT Governance (9)
- International Leaders (9)
- People (9)
- IT Consulting (8)
- AI (7)
- Cloud (7)
- MIT Sloan CISR (7)
- ✨ (7)
- Change Management (6)
- Azure DevOps (5)
- Innovation (5)
- Working from Home (5)
- Business Architecture (4)
- Continuous Integration (4)
- Enterprise Analysis (4)
- FinOps (4)
- Client Briefing Events (3)
- Cloud Value Optimisation (3)
- GitHub (3)
- IT Services (3)
- Business Rules (2)
- Data Visualisation (2)
- Java Development (2)
- Security (2)
- System Performance (2)
- Automation (1)
- Communities of Practice (1)
- Kanban (1)
- Lean Startup (1)
- Microsoft Azure (1)
- Satir Change Model (1)
- Testing (1)
- August 2025 (3)
- July 2025 (3)
- March 2025 (1)
- December 2024 (1)
- August 2024 (1)
- February 2024 (3)
- January 2024 (1)
- September 2023 (2)
- July 2023 (3)
- August 2022 (4)
- July 2021 (1)
- March 2021 (1)
- February 2021 (1)
- November 2020 (2)
- July 2020 (1)
- June 2020 (2)
- May 2020 (2)
- March 2020 (3)
- August 2019 (1)
- July 2019 (2)
- June 2019 (1)
- April 2019 (2)
- October 2018 (1)
- August 2018 (1)
- July 2018 (1)
- April 2018 (2)
- January 2018 (1)
- September 2017 (1)
- July 2017 (1)
- February 2017 (1)
- January 2017 (1)
- October 2016 (2)
- September 2016 (1)
- August 2016 (4)
- July 2016 (3)
- June 2016 (3)
- May 2016 (4)
- April 2016 (5)
- March 2016 (1)
- February 2016 (1)
- January 2016 (1)
- December 2015 (5)
- November 2015 (11)
- October 2015 (3)
- September 2015 (1)
- August 2015 (1)
- July 2015 (7)
- June 2015 (7)
- April 2015 (1)
- March 2015 (2)
- February 2015 (2)
- December 2014 (3)
- September 2014 (2)
- July 2014 (1)
- June 2014 (2)
- May 2014 (8)
- April 2014 (1)
- March 2014 (2)
- February 2014 (2)
- November 2013 (1)
- October 2013 (2)
- September 2013 (2)
- August 2013 (2)
- May 2013 (1)
- April 2013 (3)
- March 2013 (2)
- February 2013 (1)
- January 2013 (1)
- November 2012 (1)
- October 2012 (1)
- September 2012 (1)
- July 2012 (2)
- June 2012 (1)
- May 2012 (1)
- November 2011 (2)
- August 2011 (2)
- July 2011 (3)
- June 2011 (4)
- April 2011 (2)
- February 2011 (1)
- January 2011 (2)
- December 2010 (1)
- November 2010 (1)
- October 2010 (1)
- February 2010 (1)
- July 2009 (1)
- October 2008 (1)
