Sort of! But not exactly. This is a topic I've been meaning to write a long post on for ages, and have given a few short impromptu presentations about.
Consequentialism, deontology, and virtue ethics are classifiers over world-histories, actions, and agents, respectively. They're mutually reducible, in that you can take a value system or a value-system fragment in any one of the three forms, and use it to generate a value system or value-system fragment in either of the other two forms. But value-system fragments are not equally naturally expressed in different forms; if you take a value from one and try to reframe it in the others, you sometimes get an explosion of complexity, particularly if you want to reduce value-system fragments which have weights and scaling properties, and have those weights and scaling properties carry through.
I have previously been saying things like "consequentialism is obviously correct". But it occurred to me that this was gibberish this morning.
I maintain that, for any consequentialist goal, you can construct a set of deontological rules which will achieve approximately the same outcome. The more fidelity you require, the more rules you'll have to make (so of course it's only isomorphic in the limit).
Similarly, for any given deontological system, one can construct a set of virtues which will cause the same behavior (e.g., "don't murder" becomes "it is virtuous to be the sort of person who doesn't murder")
The opposite is also true. Given a virtue ethics system, one can construct deontological rules which will cause the same things to happen. And given deontological rules, it's easy to get a consequentialist system by predicting what the rules will cause to happen and then calling that your desired outcome.
Given that you can phrase your desired (outcome, virtues, rules) in any system, it's really silly to argue about which system is the "correct" one.
Instead, recognize that some ethical systems are better for some tasks. Want to compute actions given limited computation? Better use deontological rules or maybe virtue ethics. Want to plan a society that makes everyone "happy" for some value of "happy"? Better use consequentialist reasoning.
Last thought: none of the three frameworks actually gives any insight into morality. Deontology leaves the question of "what rules?", virtue ethics leaves the question of "what virtues?", and consequentialism leaves the question of "what outcome?". The hard part of ethics is answering those questions.
(ducks before accusations of misusing "isomorphic")