Reinforcement learning (RL) has shown to be effective for simple automated cyber defence (ACD) type tasks. However, there are limitations to these approaches that prevent them from being deployed onto real-world hardware. Trained policies will often have limited transferability across even small changes to the environment setup. Instability during training can prevent optimal learning, a problem that only increases as the environment scales and grows in complexity. In this work we look at addressing these limitations with a zero-shot transfer approach based on multi-agent reinforcement learning. Our approach partitions up the task into smaller network machine subtasks, where agents learn the solution to the local problem. These local agents are trained in a small-scale network, then transferred to larger networks by mapping the agents to machines in the new network. We have found that this transfer method is effective for direct application to a number of ACD tasks. We show that its performance is robust to changes in network activity, attack scenario and reduces the effects of network scale on performance.