Define capable, and in your definition please address their evaluation of the performance of the 70-200mm f/2.8L IS II, which they score lower than the MkI version of that lens.
DxO have actually responded to this anomaly on their website when questioned by Canon 70-200mm f2.8 IS II users as to why the apparently superior mark II scores so much lower than its predecessor and their reply was that the older lens has better resolving power (@ 62 line pairs per millimeter vs 52 lp/mm for the new lens). But their website also states that there is enormous variation in resolution throughout the focal range, plus their Resolution Index scores are based on averages of multiple shots taken typically between f5 and f8 throughout the zoom range.
Clearly then, one can challenge the validity & reliability of this 'Resolution' test by conducting a similar test (they outline in graphic detail their testing procedure) using say multiple samples (half a dozen) of each of the 2 lenses in question (to also test if copy-to-copy variation exists).
Secondly, irrespective of the first answer, what the results suggest given that the Mark II surpasses the Mark I in each of the other sub-categories (especially CA), is that their arbitrary weighting (Black Box is an apt description as they do not divulge their weighting methodology) system is heavily skewed towards 'Resolution' when computing the Overall Score
The very fact that they place more emphasis on one single category (more so than all other sub-categories combined) is evidence that their overall score is biased & subjective. According to DxO the most impressive Canon lens in the entire EF range is the 85mm f/1.8.
What they should do is post the summary stats of all the resolution test results in a table (freq dist along with standard deviation figures) - that way we can see if it is case of some outlier or skew (to wide or tele) that is causing them to conclude that the mark I is better.
No academic journal would publish a DxO report without both Data (in an Appendix) and Methods (with a clearly defined algorithm stating the parameters for weighting each category). In the world of peer review, Black-Box methodology would simply have REJECTED
stamped on it and returned.