Comparing clustering results is unfortunately not trivial. In particular when it comes to overlapping, hierarchical and subspace results. The common measures only work work strict partitioning clusterings. And even then the have different bias and there exists a dozen of quality measures. So your result may well be better on one measure, and worse on the other.
I don't know the details on Carrot, as I am an ELKI user. For comparing clusterings it has amongst others various pair counting measures (Precision, Recall, F1, Jaccard, Rand, Adjusted Rand, Fowlkes-Mallows), entropy based measures ("normalized mutual information"), Bcubed measures (again precision, recall and F1), Set-matching measures (F1, purity and inverse purity), edit-distance based measures and a Gini-based measure. That is a total of like 20 quality measures. I have not yet found a good argument why one or the other is clearly superior, they all have their benefits and drawbacks.
BCubed claims to be superior, but well, who doesn't?
https://en.wikipedia.org/wiki/Cluster_analysis#External_evaluation
gives details on some of these measures, but also no indication of when to use which.
Plus, experiments cannot prove that any algorithm is better than another. You might just have chosen bad parameters for the other! Or you might be using an inappropriate "algorithm 2". There are hundreds of clustering algorithms (ELKI is the largest collection of clustering algorithms I know, which is why I'm currently working with it!) and ideally you should be better than every single one of them. I currently do not think it makes much sense to invent yet another clustering algorithm. You might just be reinventing the wheel, and someone might already have invented exactly this clustering algorithm, or something even superior.