On February 27, 2025, the CJEU delivered an important judgment on the interpretation of Article 15(1)(h) and Article 22 of Regulation (EU) 2016/679 on General Data Protection (GDPR) in C-203/22 CK Magistrat der Stadt Wien v Dun & Bradstreet Austria GmbH.
The facts
The mobile phone operator refused CK’s request
to conclude or extend the mobile telephone contract for a monthly payment of a
mere EUR 10. The refusal was justified with CK not passing a
creditworthiness check with the credit reference agency D & B,
which carried out an automated assessment. Unsurprisingly, CK was unhappy with
the decision; her credit score was good. She brought the matter to the Austrian
data protection authority and, with this, started a long way to the preliminary
reference, going through various instances and avenues for protection.
The referring court raised several questions,
which the CJEU grouped into essentially two questions:
The
first question
Must Article 15(1)(h) be interpreted as
meaning that, in the case of automated decision-making, including profiling,
within the meaning of Article 22(1), the data subject may require the
controller to provide, ‘meaningful information about the logic involved’ in the
decision making, which would mean an exhaustive explanation of the procedure
and principles actually applied in using personal data to obtain a specific
result, in this case, a creditworthiness assessment.
According
to Article 15 (h), the data subject has the right to obtain from the
controller confirmation as to whether his/her personal data is being processed,
information on the use of automated decision-making where applicable, including
profiling, referred to in Article 22(1) and (4), and meaningful
information about the logic involved, as well as the significance and
the envisaged consequences of such processing for the data subject.
Article 22 provides that the data subject shall have the right not to be subject to a decision based solely on automated processing, including profiling, and that certain data enlisted in Article 9(1) GDPR such as racial or ethnic origin, religious beliefs cannot be considered in data processing.
Profiling, in this context, means automated processing of personal data, consisting of using personal data to analyse or predict the consumer's economic situation.
In its analysis, the CJEU first turned to a literal interpretation of the wording of Article 15 (h) and concluded that the concept of ‘meaningful information’ under that provision may have various meanings in different language versions of GDPR, which should be taken to be complementary to each other. In addition, the ‘logic involved’ in automated decision-making, which constitutes the subject matter of ‘meaningful information’ is capable of covering a wide range of ‘logics’ concerning the use of personal data and other data with a view to obtaining a specific result by automated means. The CJEU held, that the provision covers all relevant information concerning the procedure and principles relating to the use, by automated means, of personal data with a view to obtaining a specific result.
The CJEU next turned to contextual analysis of the concept of ‘meaningful information about the logic involved’, within the meaning of Article 15(1)(h). In this analysis the CJEU looked at the Guidelines on automated individual decision-making and profiling for the purposes of Regulation 2016/679 and other provisions of the GDPR providing information duties of data controllers. The CJEU concluded that information duties relate to all relevant information that should be provided in clear, concise, transparent, intelligible and easily accessible form, using plain and clear language
Finally, the CJEU looked at the purpose of the provision, asserting that the purpose of the data subject’s right to obtain the information provided for in Article 15(1)(h) is to enable him or her to effectively exercise the rights conferred on him or her by Article 22(3), namely, the right to express his or her point of view and to contest the relevant decision. This, in turn, requires the right to obtain an explanation of the decision.
The CJEU then concluded that under Article 15(1)(h) the right to obtain ‘meaningful information about the logic involved’ in automated decision-making must be understood as a right to an explanation of the procedure and principles actually applied in order to use, by automated means, the personal data of the data subject with a view to obtaining a specific result, such as a credit profile. In order to enable the data subject to effectively exercise the rights conferred on him/her by the GDPR and, in particular, Article 22(3), that explanation must be provided by means of relevant information in a concise, transparent, intelligible and easily accessible form. Notably, the court further provided guidance on what is considered to be ‘meaningful information about the logic involved’ in automated decision-making. The procedures and principles actually applied must be explained in such a way that the data subject can understand which of his/her personal data have been used in the automated decision-making and the extent to which a variation in the personal data taken into account would have led to a different result. The requirements of Article 15(h) cannot be met by the mere communication of a complex mathematical formula, such as an algorithm, or by the detailed description of all the steps in automated decision-making since neither of those would constitute a sufficiently concise and intelligible explanation.
Second legal question
Another important contribution of the present judgment is the consideration of the relationship between Article 15(1)(h) and Directive 2016/943 on trade secrets, given that D&B argued that the logic of their automated decision-making, including what information is considered in which way, is a trade secret and should, therefore, not be disclosed.
The CJEU highlighted that the protection of personal data is not an absolute right. Restrictions are possible of the scope of the obligations and rights provided for in, inter alia, Article 15 of the GDPR, but only when such a restriction respects the essence of the fundamental rights and freedoms and is a necessary and proportionate to safeguard the protection of the rights and freedoms of others. However, the result of any consideration on the limits of the protection of personal rights should not be a refusal to provide all information to the data subject.
The CJEU concluded that Article 15(1)(h) must be interpreted as meaning that, where the controller takes the view that the information to be provided to the data subject is a trade secrets, within the meaning of point 1 of Article 2 of Directive 2016/943, that controller is required to provide the allegedly protected information to the competent supervisory authority or court, which must balance the rights and interests at issue with a view to determining the extent of the data subject’s right of access provided for in Article 15 of the GDPR.
Our analysis
This decision is significant in addressing the
long-standing problem of the lack of transparency in automated decision-making
by credit reference agencies, an important
problem
in the EU. Given that in most countries we have access to our credit reports we
can know what data is considered in their decision making in producing a credit
score and a credit report, however, credit reference agencies have refused disclosing
the way this data is processed, the logic behind their decision making, in what
way and to what extent various data is considered (weighted) in their decision making.
Although based on this decision, consumers
are still not entitled to get hold of that information directly, but a first
step has been made by mandating disclosure to the relevant authority who then
makes a decision on whether or not to disclose it to the consumer, balancing
the rights and interests of the two parties. This and other judgments of the
CJEU (see C-634/21
SCHUFA Holding) may be gradually bringing transparency into this traditionally
very untransparent area.
As credit reference agencies nowadays use artificial
intelligence for automated decision-making, the judgment is relevant for advancing
transparency considerations of AI systems.
Finally, given that the judgment tackles the
operation of credit reference agencies, which are frequently used by creditors
to assess the affordability of loan applications, it is relevant for
responsible lending rules in Directive 2023/2225 on consumer credit (CCD2),
which in Article 18 refers to creditworthiness assessment based on automated processing
of personal data.