Journal of International Commercial Law and Technology
2026, Volume 7, Issue 1 : 192-214 doi: 10.61336/Jiclt/26-01-18
Research Article
Extracting Key Perceptual Factors Shaping Consumer Attitudes Toward AI-Enabled Advertising: An Exploration
 ,
 ,
1
School of Business, Galgotias University, Greater Noida, India, 203201 Department of Management Studies & Industrial Engineering, Indian Institute of Technology (ISM)-Dhanbad, Dhanbad-826004, Jharkhand, India
2
Associate Professor, Department of Management Studies & Industrial Engineering, Indian Institute of Technology (ISM)-Dhanbad, Dhanbad-826004, Jharkhand, India
3
Assistant Professor, Department of Management Science, Mahatma Gandhi Central University Bihar, Motihari-845401 Bihar, India
Received
Oct. 2, 2025
Revised
Nov. 4, 2025
Accepted
Dec. 13, 2025
Published
Jan. 17, 2026
Abstract

In an ever-evolving digital ecosystem, artificial intelligence (AI) is establishing itself as a disruptive power in advertising, significantly transforming how organisations engage with customers. While programmatic advertising provides unparalleled levels of personalisation and precision in ad-targeting, it also poses challenging issues related to trust among consumers, perceived intrusion, emotional discomfort, and ethical data practice. Understanding the psychological underpinnings of consumer perception in this context is imperative, particularly in emerging markets such as India, where digital literacy is uneven and regulatory frameworks are still maturing. This study aims to extract and interpret the latent factors that shape Indian consumers' perceptions towards relevance of AI-enabled advertising. Employing a structured survey instrument, data were collected from 302 digitally active Indian respondents via a Likert-scale questionnaire. The data were analyzed using Exploratory Factor Analysis (EFA) through Principal Component Analysis (PCA) with Varimax rotation to reveal the underlying constructs governing consumer attitudes towards AI-enabled Advertisements. Nine distinct perceptual factors were identified, namely Perceived Privacy Intrusion, Perceived Ad-Relevance, Perceived Trustworthiness, Perceived Ad-Value, Perceived Privacy Assurance, Purchase Intent, Perceived Information Quality, Information Disclosure Discomfort and Click Willingness. Together, these factors provide a nuanced conceptual framework for understanding how Indian users engage with AI-generated advertisements at cognitive, affective, and behavioural levels. This study adds to the literature by finding and naming the factors for consumer perception in AI advertising, including implications for advertisers, and policymakers looking to create effective AI advertising ecosystem.

Keywords
INTRODUCTION

The advertising industry is going through a paradigm shift, fuelled mostly by the widespread adoption of artificial intelligence (AI) technology. Advertising, formerly controlled by creativity and intuition, has now embraced computational precision, allowing businesses to deliver highly customised messages to specific consumers in real time. These advancements have improved targeting efficiency and transformed advertising into a predictive yet dynamic dialogue between consumers and algorithms. At the heart of this revolution is AI's ability to collect and analyse vast amounts of behavioural, contextual, and transactional data. Based on this information, AI systems may create advertising content that matches a user's tastes and preferences, search patterns, and even inferred emotional states(Forrest & Hoanca, 2015). While this has clearly increased the relevance and ad-engagement, it has also prompted important questions about privacy, data ethics, transparency, and the psychological consequences of algorithmic persuasion(Martin et al., 2017; Taylor, 2019; Vimalkumar et al., 2021).

 

  • The Rise of AI-enabled advertising

The dynamics of customer engagement, media personalization, and marketing analytics have changed in recent years because of the integration of artificial intelligence(AI) into advertising. Initially, started as a rule-based automated system that could send conditional messages or segment customers, these systems have now grown into complex ecosystem of personalization fueled by deep learning. AI is now able to identify patterns in user activity in real time, modify messaging aptly, and present contextually relevant ads on several platforms.

 

The transition from static campaigns to intelligent ad-distribution technologies marks an important milestone in advertising history. Machine Learning, Natural Language Processing(NLP) and predictive analytics enable marketers to move beyond standard demographic segmentation to Individual-level targeting (Rosenkrans & Myers, 2018). For instance, AI can now analyse not just what a consumer clicks on, but also when they are most likely to click, how long they spend viewing a specific content, and which psychological triggers boost the likelihood of purchase. These capabilities have transformed advertising campaigns, making them more responsive, data driven and consumer-focused than ever (Huang & Rust, 2021).

 

One of the most tangible benefits of AI in advertising is its potential to increase engagement through personalisation.  AI-powered platforms analyse massive datasets—ranging from past browsing behaviour and geolocation data to facial expressions and mood indicators—to produce personalised and intuitive advertising experiences.  As a result, customers are more likely to respond to commercials that appear "relevant" or "designed specifically" for them.  This transition is most visible in industries like e-commerce, tourism, financial services, and entertainment, wherein adaptive content is now a norm. Platforms like Google Ads, Facebook Ads, and Amazon's recommendation engine use sophisticated algorithms to not only discover potential client categories, but also to deliver the correct message at the right time and in the appropriate format.

 

In practice, AI-driven ads may appear as personalized banners, chatbot interactions, voice assistant prompts, or content feeds adapted to the user's digital behaviour. These developments guarantee advertisers increased return on investment (ROI) and reduced cost of new customer acquisition; however, from the user's perspective, they raise worries about manipulation, psychological exhaustion, and surveillance capitalism (Zuboff, 2019).

 

  • The Indian Digital Landscape

India provides a very exciting environment that encourages technological growth. The country has experienced exceptional digital growth in the last decade. According to IAMAI and Kantar's Internet in India Report 2024 India has a remarkable internet user base of over 886 million people, expected to cross 900 million in 2025, with 86% those using over-the-top (OTT) music and video services(KANTAR; IAMAI, 2025). This digital expansion is followed by rising usage of social media platforms, e-commerce websites, and OTT services, all of which rely significantly on AI-driven advertising to monetise user engagement.

 

This digital surge has paved the way for AI-driven advertising advancements from global internet giants such as Google and Meta. Both platforms utilise artificial intelligence to maximise ad placement, adjust bidding in real-time, and tailor content based on expected user preferences. In India, for example, Meta's ad delivery system uses artificial intelligence to evaluate consumer sentiment in many languages, whereas Google's responsive search adverts dynamically change headlines and descriptions based on user search intent (Tran, 2024). These systems are efficient and generate billions of impressions per day, influencing customer behaviour on a massive scale.

However, as technology has advanced, several serious concerns have ascended, particularly with data privacy, customer trust, and the ethical limitations of algorithmic persuasion. While AI provides personalisation and efficiency, it also raises concerns about how customer data is acquired, stored, and used, particularly in a regulatory context that is constantly evolving. Many users are unaware of how AI systems work, what data they collect, and how they influence behaviour. In the absence of robust consumer education and comprehensive data protection laws (India's Digital Personal Data Protection Act was only recently introduced in 2023), consumers are often exposed to AI systems without meaningful consent or transparency. The absence of informed engagement poses distinct perceptual and ethical issues that must be discussed in the Indian socio-cultural context.

 

  • Gaps in Existing Research

The prior research on digital advertising has primarily focused on outcome factors like click-through rates, user engagement, and purchase intent. Despite being useful, such research frequently overlooks the underlying cognitive and emotional processes by which people evaluate AI-generated communication. Furthermore, these studies are majorly based on Western societies, where digital literacy is higher and privacy rules are more firmly established.

 

In India, there is a dearth of empirical research exploring how consumers perceive AI-enabled advertisements: what they find appealing, what they distrust, what makes them uncomfortable, and what leads to ad-skepticism. Such insights are critical not only for optimizing marketing strategies but also for informing policy frameworks and ethical standards in deployment of AI for Advertising.

 

1.4 Scope of the Present Study

The present study examines consumer behaviour, digital ethics, and the application of AI technology in advertising. Instead of testing hypotheses or forecasting behaviour using regression-based models, this study takes an exploratory approach, trying to identify the underlying structure of consumer perception towards AI-enabled advertising through statistical analysis of factors.

 

The central question guiding this research is: What are the latent psychological dimensions that shape how Indian consumers perceive and respond to AI-enabled advertising? 

 

By extracting and interpreting these dimensions, the study aims to formulate a conceptually robust framework of consumer perception w.r.t AI-enabled Advertisements, that may support theoretical modelling for future research.

 

LITERATURE REVIEW

2.1 Technology Acceptance and AI Use

The Technology Acceptance Model (TAM), introduced by Fred Davis in 1989 remains one of the most widely applied frameworks in understanding user adoption of new technologies. At its core, TAM postulates that “two cognitive variables namely Perceived Usefulness (PU) and Perceived Ease of Use (PEOU), shape user attitudes and, consequently, behavioural intentions toward using any technology(F. Davis, 1987).” Over the years, TAM has been extended to contexts such as e-commerce, mobile applications, and intelligent systems, often incorporating additional constructs such as trust, perceived risk, and subjective norms (Carlos Martins Rodrigues Pinho & Soares, 2011; Lu et al., 2003; Venkatesh & Bala, 2008; Venkatesh & Davis, 2000). In AI-driven environments, PU manifests through constructs like decision-making support and informational value, while PEOU may be replaced or redefined through dimensions such as system interpretability or user autonomy (Dwivedi et al., 2021).

 

In this study, TAM is used as a theoretical backdrop to identify perceptual constructs such as informativeness and usefulness dimensions that may emerge naturally through exploratory factor analysis.

 

2.2 AI-Driven Personalization in Advertising

Although personalisation has long been a major theme in advertising, the advent of artificial intelligence has significantly altered the scope and level of detail of the idea. These days, AI systems can modify content in real time according to a user's past browsing habits, social media activity, purchasing patterns, and even their implied emotional state(Tucker, 2014). Predictive personalization, driven by machine learning algorithms, optimizes ad delivery by segmenting users not just by demographics but by dynamic behaviour. Research show that consumers are more expected to absorb personalized content, particularly when it reflects perceived relevance(Bleier & Eisenbeiss, 2015; Kim & Huh, 2017). On the other hand, Hyper-personalization might backfire when it exceeds a certain comfort level, resulting in emotions of manipulation or surveillance(Bang et al., 2019). This paradox, the concurrent attraction to and repulsion from AI personalization, highlights the complexities of consumer perception and lends support to an exploratory approach(Aguirre et al., 2016).

 

2.3 Consumer Trust and Perceived Transparency

The effectiveness and adoption of AI-enabled advertising are significantly influenced by trust. AI-enabled advertising forces users to believe the algorithm that creates the ad message in the background, in contrast to traditional advertising, which gets its credibility from message tone or brand reputation (Andrews et al., 2016). This requires faith in algorithmic fairness, system intent, and data management (Bleier & Eisenbeiss, 2015a). Studies show that perceived transparency i.e. “the extent to which a user understand why an ad is being shown, what data is used, and how it is used,” can significantly enhance trust (Bleier & Eisenbeiss, 2015; Sirdeshmukh et al., 2002; Soh et al., 2013). When transparency is lacking, even relevant ads can evoke skepticism. Accordingly, constructs like data clarity, explainability, and control over data use are now integral to understanding consumer trust in AI systems(Diwanji et al., 2022). According to this study, trust and transparency will become crucial perceptual factors, especially in a culture like India where awareness of systems with algorithms is growing yet comparatively scarce.

 

2.4 Privacy Concerns and Data Ethics

Significant ethical concerns about data privacy, informed consent, and digital autonomy have been brought up by the implementation of artificial intelligence (AI) systems into digital advertising(Kumar & Suthar, 2024; Nill & Aalberts, 2014). Asymmetric exchanges of information arise due to the frequent use of opaque methods by AI systems to gather personal data, leaving users in the dark about what information is being gathered and how it is being monetized (Daems et al., 2019). The Consumer Privacy Concerns (CPC) framework (Lee et al., 2011; Malhotra et al., 2004) conceptualizes privacy anxiety as a multi-dimensional construct, encompassing control over personal information, awareness of data collection, and worries of secondary use. In the AI context, these concerns are intensified due to the non-transparent nature of algorithmic logic applied in personalised advertising (Zuboff, 2019). Furthermore, in India, where data protection legislation is still evolving, consumers may lack the legal literacy to evaluate AI practices critically.

 

These issues suggest that privacy concerns are not just regulatory matters; they are psychological dimensions likely to emerge in any factor structure related to consumer perception.

 

2.5 Psychological and Behavioural Reactions to AI enabled Advertising

Recent research has shown that customers’ reactions to AI enabled advertising are not solely rational, but also intensely emotional and psychological. Users frequently express inconsistency, appreciating personalization while being concerned about the system's intelligence and reach (Aguirre et al., 2015; Brinson & Eastin, 2016; Chellappa & Sin, 2005). Emotions like discomfort, worry, and exhaustion are typical in high-frequency, tailored ad settings. These feelings can trigger reactance, a psychological defence mechanism against perceived manipulation. Furthermore, behavioural indications such as click avoidance, ad skipping, and the usage of ad blockers may be motivated by emotion rather than perceived relevance(Goldfarb & Tucker, 2012; Tucker, 2014).

 

These findings suggest that perceptual mapping of AI advertising interactions should incorporate both: emotional constructs such as discomfort or resistance, and cognitive evaluations like perceived usefulness and trust.

 

  1. Research Objectives

Digital consumers' perceptions, responses, and delivery of advertisements are being redefined by artificial intelligence (AI). AI-driven advertising grants advantages like increased ad-personalization and improved decision-making, however, it also presents issues with trust, emotional discomfort, and data ethics(Aguirre et al., 2016). It is crucial to comprehend how customers understand these intricate experiences, especially in India, where digital engagement is growing, but AI system transparency is still low. Instead of relying on preconceived theoretical assumptions, this research uses an exploratory approach to identify the latent perceptual factors that influence consumer attitudes toward AI-enabled advertising.

 

The main objective of the current study is “To extract and interpret the core perceptual factors that influence Indian consumers' responses to AI-enabled advertisements using exploratory factor analysis (EFA).”

 

To fulfil the above-mentioned primary objective, the following are the secondary objectives that need to be achieved.

  1. To identify meaningful clusters of perception (factors) from consumer survey data that reveal how users evaluate AI-enabled advertisements in terms of personalization, usefulness, emotional response, and data concerns.
  2. To build a conceptual model based on extracted factors that may serve as a foundation for future research, ethical advertising strategies, and policy development in the Indian digital advertising ecosystem.

 

  1. Rationale of the Study

The integration of AI in digital advertising has accelerated the transformation of consumer-brand interaction across global markets. However, as technical sophistication increases, so does the complexity of user interpretation, ethical scrutiny, and data governance (Wu & Wen, 2021). While AI-enabled ad-personalization has been lauded for its ability to enhance marketing precision, concerns regarding manipulation, psychological intrusion, and asymmetries in user knowledge is raised simultaneously (Qin & Jiang, 2019). Within this tension lies a critical research question- to examine not what AI systems do technically in the backdrop, but how the AI-enabled advertisements are perceived experientially by end users; particularly in under-researched, high-growth digital economies like India.

4.1 Conceptual Justification

Traditional studies on advertising success have generally focused on visible behavioural outcomes such as click-through rates, purchase intent and conversion metrics(Azimi et al., 2012; Yoo, 2009). Despite their value, these metrics frequently fall short of capturing the complex psychological factors that influences and precedes these behaviours. Particularly in AI-enabled contexts, where content is auto-curated, and decision pathways are algorithmically mediated, consumer perceptions, emotions, and cognitive appraisals play a central role (G. Chen et al., 2019).

 

Moreover, many existing theoretical models, such as TAM, UTAUT, or even more recent extensions integrating trust or privacy concerns, impose predefined constructs on user experience (Wang et al., 2023). While these models have predictive utility, they risk overlooking the emergent and context-dependent dimensions that naturally arise from users' personal interactions with AI-generated advertising. This study adopts an exploratory strategy, diagnosing that consumers construe AI-generated information through complex cognitive-emotional frameworks that shape their perceptions and reactions. Given this sensitive knowledge, exploratory factor analysis (EFA) is utilised as a methodological option that is particularly well-suited to revealing the latent dimensions buried within perceptual data. EFA provides a solid foundation for future theoretical models to be constructed and refined, as well as the ability to empirically detect underlying structures that may not be visible at first glance.

 

4.2 Contextual Relevance: The Indian Digital Landscape

India, with one of the fastest-growing internet user bases, estimated to cross 900 million by 2025, India stands at the forefront of AI implementation in advertising. The country offers a particularly fascinating setting for this kind of research. Domestic and global platforms alike employ AI systems to deliver curated content, drive engagement, and influence consumer behaviour. Nevertheless, this digital advancement is unfolding in a regulatory environment still in transition, where the recently introduced Digital Personal Data Protection Act (2023) is only beginning to formalize user rights, consent protocols, and data usage limits (KANTAR; IAMAI, 2025). Moreover, digital literacy in India remains uneven. Many users interact with AI-driven advertisements without fully understanding the data structures that underlie them, even though urban, educated consumers may have a basic understanding of algorithmic personalization. Perceptual analysis is not only an academic endeavour but also a socio-ethical necessity because of this information asymmetry, which breeds mistrust, confusion, or emotional dissatisfaction.

 

Understanding how Indian consumers internalize and interpret AI-enabled advertising is thus vital for Marketers - who must balance targeting efficiency with user comfort and trust; Policymakers - who require data-driven insights into user sentiment to guide ethical AI deployment; and Technologists and platform designers - who must reconcile system performance with transparency, fairness, and psychological acceptability.

 

4.3 Empirical Gap and Research Contribution

A thorough review of the digital advertising literature reveals a significant dearth of empirical studies concentrating on the latent psychological elements of perceived ad relevance of AI-enabled advertising in India. While numerous conceptual articles and global surveys have raised concerns about personalization (how ads are tailored to individuals) (Chandra et al., 2022; Lee et al., 2011; Sahni et al., 2018); Surveillance (being watched or tracked by technology) (Martin et al., 2017; Rapp et al., 2009; Zuboff, 2019) ; and trust (whether consumers believe or accept the use of AI in privacy advertising) (Culnan & Armstrong, 1999; Huh et al., 2020; Reyes et al., 2025; Sirdeshmukh et al., 2002; Soh et al., 2013); very few have employed rigorous statistical techniques like factor analysis to identify the structural underpinnings of consumer perception in this domain. This study, therefore, makes a methodological and conceptual contribution by (i) applying exploratory factor analysis to extract perceptual factors without imposing pre-existing theoretical constraints and (ii) developing a comprehensive measurement instrument based on validated items and adapting to the Indian context. By doing so, the study provides a first-principles framework upon which future research, regulatory models, and ethical guidelines can be constructed.

 

RESEARCH METHODOLOGY

In keeping with the exploratory and constructivist orientation of the research, a quantitative data collection strategy was employed, supported by exploratory factor analysis (EFA) to uncover latent variable structures.

 

5.1 Research Design

The study is grounded in a quantitative research paradigm, which seeks to derive generalizable insights through objective measurement, statistical analysis, and replicable methods. While the underlying research orientation is exploratory, it emphasizes quantifiable constructs, structured instruments, and statistical rigour in construct identification(Joseph F. Hair Jr. et al., 2014). The exploratory research design is particularly suited to domains where theoretical consensus is absent or where latent constructs are suspected to exist but have not been systematically mapped (Hair et al., 1998). An exploratory research design is methodologically appropriate given the nascent nature of AI-enabled advertising and the insufficient understanding of consumers’ perception towards such intelligent ads in the Indian context.

 

  • Instrument Design

The research instrument consisted of a structured, self-administered questionnaire comprising both demographic questions and perceptual items. The questionnaire was divided into sections comprising Socio-demographic variables (age, gender, education, occupation, marital status, AI exposure frequency) and a set of 55 Likert-scale items measuring perceptual dimensions related to consumers’ perceived ad-relevance towards AI-enabled advertising. These items were adapted from validated scales used in prior studies on digital personalization, consumer trust, privacy concerns, data ethics, emotional discomfort, and behavioural responses (Dinev & Hart, 2006a; Laczniak & Muehling, 1993; MacKenzie & Lutz, 1989; Malhotra et al., 2004; Roy et al., 2017; Soh et al., 2013; Venkatesh et al., 2003; Wang et al., 2023). The Items were adapted and revised to reflect the Indian digital user context and pre-tested for language clarity and interpretability. The questionnaire was rigorously reviewed and validated by experienced researchers and academicians specializing in the fields of marketing and advertising. All perceptual items were measured on a five-point Likert scale ranging from 1 ("Strongly Disagree") to 5 ("Strongly Agree"), enabling interval-level analysis for factor extraction.

 

  • Sampling

Given the exploratory nature of the study, a non-probability convenience sampling strategy was adopted to capture a diverse but accessible cohort of Indian digital users. The inclusion criteria required the participants to be Indian nationals aged 18 years or older; possess regular exposure to AI-enabled advertisements (e.g., Google Ads, YouTube recommendations, Instagram promotions) and be comfortable completing an online survey in English.

 

  • Data Collection Procedure

Platforms like Zoho Survey  and Google Forms was utilised to distribute the survey online, and it was shared via professional forums, university groups, email networks, and social media. All participants gave their informed consent prior to accessing the questionnaire, and participation was entirely voluntary. No personally identifiable information was gathered; even giving their name was optional, and anonymity was guaranteed. Response quality was tracked in real-time during the 12-week data collection period between  April-July 2023. The final analysis did not include the entries with missing values or inconsistent/unengaged responses (e.g., the same ratings for every item). The dataset was screened to ensure compliance with the assumptions of multivariate analysis. Prior to conducting data analysis, the missing values were handled through case wise deletion.

 

 

  1. Data Analysis

The filled-in responses were received from 350 respondents, out of which 48 responses were incomplete, leading to their removal from the study. After elimination of these half-finished or unengaged responses a total of 302 responses were analysed in this pilot study, which aligns with accepted sample size norms for factor analysis, particularly where the variable-to-response ratio is 1:5 or more. As per Hair, “the minimum is to have at least five times as many observations as the number of variables to be analyzed” for Exploratory Factor Analysis (Hair et al., 1998). While this pilot sample does not claim national representativeness, it provides statistically adequate grounds for exploratory factor extraction (Table 1).

 

 

6.1 Suitability of Data for Factor Analysis

To assess whether the dataset is suitable for factor analysis, two preliminary tests are commonly applied, prior to extraction: the Kaiser–Meyer–Olkin (KMO) measure of sampling adequacy and Bartlett’s test of sphericity. These tests help determine whether the relationships among variables are strong enough to justify a factor analysis. The KMO statistic evaluates the proportion of variance among variables that might be common variance. According to Kaiser, a minimum acceptable KMO value is 0.5 (Cerny & Kaiser, 1977). Building on this, (Hutcheson & Sofroniou, 1999) categorized KMO values as follows: 0.5 to 0.7 indicates mediocre adequacy, 0.7 to 0.8 is considered good, 0.8 to 0.9 is regarded as great, and values above 0.9 are deemed excellent. Bartlett’s test, on the other hand, evaluates whether the correlation matrix significantly differs from an identity matrix, implying that variables are interrelated. For this test to support the use of factor analysis, the significance level should be below 0.05, indicating that the data exhibits sufficient correlations to proceed with the analysis (Tobias & Carlson, 1969).

 

In this study, the Kaiser-Meyer-Olkin (KMO) measure of sampling adequacy yielded a value of 0.795 exceeding the minimum threshold of 0.50, This suggests that the responses were sufficiently dense for factor analysis and likely to yield distinct and reliable components(Cerny & Kaiser, 1977). Bartlett's Test of Sphericity was highly significant, indicating that the item correlation matrix was suitable for factor extraction(Tobias & Carlson, 1969) (see Table 2).

 

 

 

6.2 Factor Extraction and Retention Criteria

The primary statistical technique employed was Exploratory Factor Analysis (EFA), which used Principal Component Analysis (PCA) as the extraction method and Varimax rotation as the rotation technique. This approach facilitates the identification of uncorrelated, orthogonal factors that reflect distinct perceptual themes within the data. The initial communalities for all items are set to 1.0 because PCA is predicated on the idea that all variances in the observed variables are initially regarded as common. As presented in Table 3, the communalities for all 51 retained items ranged from 0.501 to 0.865, well within acceptable psychometric thresholds. According to standard guidelines given by Hair et al. (1998), the communality values above 0.50 are considered indicative of sufficient shared variance between observed variables and the underlying factors. This suggests that the extracted components capture a substantial proportion of variance in each individual item.

 

Following factor extraction, the communalities indicate the proportion of variance in each item that is explained by the retained factors. Several items exhibited notably high communalities, reflecting strong representation of their underlying latent constructs. These include Q_25 ("Information conveyed in the AI-enabled Intelligent Advertisements shown to me are truthful") with a communality of 0.864, Q_19 ("Online companies should devote more time and effort into preventing unauthorized access to consumers' personal information") at 0.841, Q_18 (“Computer databases that contain consumers' personal information should be protected from unauthorized access, irrespective of the cost”) at 0.825, Q_41 ("I am willing to consider the ad-conveyed information when making purchase-related decisions") at 0.824, Q_20 ("Online companies should never share consumers' personal information without authorization") at 0.796, Q_31 ("Information conveyed in the AI-enabled Intelligent Advertisements are complete") at 0.790, and. These high values suggest that these items are highly representative of their respective factors.

 

Even the items with relatively lower communalities, such as Q_53 ("I am concerned about what digital advertisers might do with my browsing history") at 0.478, Q_54 ("I am concerned that digital advertisers are collecting too much information about me") at 0.543, and Q_51 ("I believe that my privacy is seriously threatened by use of AIeIA for personalized advertising") at 0.536, were close to or above the commonly accepted minimum threshold of 0.50. Because of their theoretical importance and distinctive contribution to the dimensionality of the construct, these items were kept. The structural integrity and reliability  of the extracted factors are confirmed by the consistently high communalities across most items.

 

Following the evaluation of the total variance explained (Table 4), the analysis revealed that the observed variables could be grouped into nine distinct components, each capturing a unique dimension of consumer perception toward AI-enabled advertising.

 

 

6.3 Rotated Component Matrix:

The Rotated Component Matrix provides detailed insights into the variable loadings on each of the nine extracted factors. Factor loadings above 0.50 are considered strong. This step facilitated clearer identification and labelling of the factors, aiding in the development of a structured and interpretable factor solution. The Rotated Component Matrix Table (Table 5) shows that most items demonstrated high loadings on a single factor with no cross-loading, indicating excellent construct purity and structural clarity.

 

 

 

Since exploratory factor analysis can occasionally yield slightly varied rotated solutions owing to differences in convergence criteria, computational algorithms, or random starting points, the Component Transformation Matrix (Table 6)  ensures methodological transparency by providing a reproducible mathematical record of how the original factor axes were transformed to achieve a simpler, more interpretable structure (Watkins, 2018).

 

 

6.4 Summary of Extracted Factors

The factor analysis produced a robust and interpretable structure, summarizing multi-dimensional consumer perceptions and behavioural orientations toward AI-enabled advertising. These dimensions reflect a nuanced interplay of cognitive, affective, and evaluative factors underlying technology adoption and marketing receptivity in the AI-enabled advertising context. The EFA revealed nine well-defined and interpretable factors, each capturing a unique dimension of consumer perceptions toward AI-enabled advertising. These factors represent the multi-faceted attitudes, emotional responses, and behavioural intentions of Indian consumers in the context of AI-driven marketing communications. This suggests that the constructs identified through Principal Component Analysis (PCA) adequately represent the underlying structure of consumer perceptions toward AI-enabled advertising. The high level of shared variance across the instrument’s items further validates the scale's suitability for subsequent interpretation and naming of latent dimensions, grounded in the Technology Acceptance Model (TAM) and related theories of digital trust, user attitude, and perceived usefulness. Following rotation using the Varimax method with Kaiser Normalization, the retained components were evaluated and named based on theoretical alignment with the Technology Acceptance Model (TAM) and its extensions (e.g., TAM2, UTAUT, Trust-based TAM) and other related theories (Carlos Martins Rodrigues Pinho & Soares, 2011; F. Davis, 1987; Lu et al., 2003; Venkatesh & Bala, 2008; Venkatesh & Davis, 2000).  A comprehensive cross-referencing with existing literature and an iterative content analysis of the items were used to thematically label each factor (Table 7).

 

 

 

Each of these nine factors group together a set of interrelated items with strong internal correlations, thus revealing latent perceptual constructs among Indian digital consumers. The factor labels were developed through a qualitative interpretation of high-loading items and reflect the dominant psychological and behavioural themes underpinning each dimension.

 

6.5 Reliability Analysis

After performing exploratory factor analysis (EFA), reliability analysis is critical for evaluating the internal consistency of the obtained constructs. While EFA helps to reveal the underlying factor structure and group related items, it does not confirm that these items consistently measure the intended latent constructs (Hair et al., 1998; Vaske et al., 2017). Reliability analysis, commonly measured using Cronbach's Alpha, guarantees that the items within each component are coherent and stable, confirming the measurement model's trustworthiness (Tavakol & Dennick, 2011). High reliability shows that the factor is well-defined and the items measure the same underlying concept, which is required to make reliable conclusions from the data (Cortina, 1993; Vaske et al., 2017).

 

 

* The factors "Perceived Information Disclosure Discomfort" (FAC_8) and "Click Willingness" (FAC_9), each comprising only two items, recorded Cronbach’s Alpha values of 0.702 and 0.661, respectively. While the former meets the conventional threshold for internal consistency, the latter, though slightly below 0.70, is considered acceptable in the context of exploratory research.  Despite diverging from the typical recommendation of three items per factor, both constructs are retained based on their strong item loadings, conceptual clarity, and empirical relevance to the research context. Two-item constructs are considered acceptable in exploratory factor analysis when they demonstrate adequate internal correlation and conceptual alignment, especially in early-stage or pilot studies (Eisinga et al., 2013; Worthington & Whittaker, 2006).

 

  1. Interpretation of Results

The results of the exploratory factor analysis (EFA) provided valuable insights into the components influencing Indian consumers' perception towards AI-enabled advertising. The current study discovered nine distinct factors, each indicating different aspects of consumer interaction with AI-enabled advertising. Cronbach's alpha measures the internal reliability of these constructs, which demonstrates excellent consistency among the items within each construct.

 

7.1 Factor Extraction Results and Theoretical Contextualization

7.1.1      Factor 1: Perceived Privacy Intrusion

Perceived Privacy Intrusion, represents reservations regarding privacy and trust in AI-enabled advertising. This factor emerged from items that measure consumer perceptions of how data requirements for AI-enabled advertising might invade personal privacy. With a Cronbach's alpha value of .923, this factor demonstrates high internal consistency, indicating that the items collectively assess a single, coherent construct. This factor is consistent with Privacy Calculus Theory, which postulates that individuals assess the benefits of online services against potential privacy risks (Smith et al., 1996). Similarly, Trust-enhanced TAM (Harrison McKnight et al., 2002) emphasizes the role of trust in moderating consumer attitudes toward digital technologies, which is evident in this construct. Consumers' concerns regarding data privacy are critical for advertisers to address. Privacy assurance in AI-enabled advertising is central to fostering trust, which in turn can impact consumer willingness to engage with ads. This factor's high reliability indicates that privacy concerns are a significant predictor of consumer behaviour in the digital advertising context, aligning with previous research that highlights the centrality of privacy in the acceptance of digital technologies (Featherman & Pavlou, 2003).

 

7.1.2      Factor 2: Perceived Ad-Relevance

Perceived Ad-Relevance represents consumers' perceptions of the relevance of AI-driven advertisements to their needs and preferences. This factor, with a Cronbach’s alpha of .921, also demonstrated excellent reliability. It aligns with the Perceived Usefulness (PU) construct from TAM, which suggests that the more useful individuals perceive a technology to be, the more likely they are to accept and engage with it (F. D. Davis, 1989). In the context of advertising, relevance is a key determinant of perceived usefulness, as personalized and relevant ads are seen as providing more value. This finding confirms the growing importance of personalization in digital marketing, where consumers expect tailored content that speaks directly to their interests. The results underscore the importance for advertisers to develop algorithms that optimize ad relevance, as irrelevant ads lead to consumer disengagement (Bleier & Eisenbeiss, 2015).

 

7.1.3      Factor 3: Perceived Trustworthiness

Perceived Trustworthiness emerged as a significant factor reflecting consumer trust in AI-enabled advertising platforms. With a Cronbach’s alpha of .913, this construct is highly reliable. It is grounded in both the Trust and Perceived Trustworthiness dimensions of TAM extensions, particularly the work of Featherman & Pavlou (2003) and McKnight et al. (2002), which emphasize trust as a key determinant of technology acceptance. This factor highlights the essential role of trust in shaping consumer perceptions about AI-enabled advertising.

To maintain and develop trust, advertisers must ensure that AI systems are transparent, virtuous, and secure. Building trust is critical for driving long-term consumer engagement with AI-powered marketing, especially given the growing worries about data protection and abuse.

 

7.1.4      Factor 4: Perceived Ad-Value

Perceived Ad-Value considers both the hedonic and utilitarian features of AI-enabled ads, representing how consumers perceive the value it brings. With a Cronbach's alpha of.886, this component has excellent internal consistency. It is closely related to TAM's Perceived Usefulness (PU) as well as UTAUT2 components, including Hedonic Motivation and Perceived Usefulness (Venkatesh et al., 2012).

 

Consumers are more inclined to engage with advertisements that they deem useful, whether for amusement, education, or utility. This aspect emphasises that advertisements can bring value in two ways: functionally (useful products or services) and emotionally (entertainment or delight) (Van-Tien Dao et al., 2014). To boost engagement, advertisers should concentrate on developing a system that generates AI-enabled ads which provide both emotional and informational value.

 

7.1.5      Factor 5: Perceived Privacy Assurance

Perceived Privacy Assurance refers to how consumers perceive the protection of their personal information in AI-enabled advertising. This factor had a Cronbach's alpha of.885, indicating a high level of reliability (Bansal et al., 2015). This construct builds on the Perceived Security and Perceived Privacy dimensions in TAM extensions, highlighting the role of privacy assurances in fostering consumer trust and willingness to engage (Featherman & Pavlou, 2003).

 

Privacy concerns remain one of the major impediments to digital advertising acceptability. Thus, marketers must provide unambiguous privacy assurances and data protection procedures. Ensuring privacy is not only a legislative duty but also a strategic advantage in fostering consumer trust and participation (Dinev & Hart, 2006b).

 

7.1.6      Factor 6: Purchase Intent

Purchase Intent reflects consumers' likelihood of purchasing a product or service after being exposed to AI-driven ads. With a Cronbach’s alpha of .869, this factor demonstrates strong reliability. It is grounded in the Behavioural Intentions construct from TAM, which suggests that perceived usefulness and ease of use influence consumers’ intentions to act (Davis, 1989). This component is critical for advertisers since it has a direct correlation with sales performance. AI advertisements that effectively affect buying intentions can yield a high return on investment. This emphasises the necessity of targeting ads that are relevant to consumer tastes and requirements.

 

7.1.7      Factor 7: Perceived Information Quality

Perceived Information Quality refers to how customers assess the relevance, accuracy, clarity, and usefulness of information offered in advertisements powered by artificial intelligence. With a Cronbach's alpha of .712, this component is moderately reliable. It draws on constructs from the Technology Acceptance Model 3 (TAM3), particularly Perceived Information Quality and Output Quality, which emphasize the role of high-quality information in enhancing decision-making and reducing cognitive effort (Venkatesh & Bala, 2008).

 

In the context of AI-enabled advertising, perceived information quality influences customer trust and engagement. Advertisements that convey factual, clear and relevant information are more likely to be perceived as credible and beneficial (Davies, 2001; Lou & Yuan, 2019).

 

7.1.8      Factor 8: Perceived Information Disclosure Discomfort

Perceived Information Disclosure Discomfort reflects consumers’ apprehension regarding the collection, usage, and exposure of personal information in AI-enabled advertising. This factor, with a Cronbach's alpha of .702, indicates moderate reliability. It is informed by Privacy Calculus Theory and the perceived risk dimensions found in extended TAM frameworks, which suggest that consumers weigh the benefits of personalization against the perceived risks to their privacy (Dinev & Hart, 2006b; Xu et al., 2025).

 

Customers experience discomfort when they believe the risks outweigh the advantages, which may make them reluctant to use platforms that use these kinds of data acquisition methods (Meng & Liu, 2025). To enhance consumer receptivity, marketers must address transparency in data use and ensure that AI systems operate within ethically sound and privacy-conscious frameworks to address perceived discomfort with information disclosure (Tadelis & Zettelmeyer, 2015).

 

7.1.9      Factor 9: Click Willingness

Click Willingness reflects consumers' likelihood of clicking on any AI-driven ad. With a Cronbach’s alpha value of .661, this factor shows a relatively lower reliability compared to the other factors, but it is still useful for understanding consumer behavior. This construct is derived from Behavioural Intention in TAM, which suggests that the likelihood of acting (clicking) is influenced by perceived ad effectiveness and trust (Bleier & Eisenbeiss, 2015). Click willingness is a direct measure of ad-effectiveness and an essential metric for digital advertising success. Advertisers can use this factor to assess the engagement potential of AI powered ads and optimize campaigns to encourage more clicks (H. Chen, 2024; Fulgoni & Mörn, 2009).

 

7.2         Conceptual Model Development

Based on the explanatory insights drawn from the extracted factors, a conceptual model is proposed to represent the perceptual framework that underpins Indian consumers’ responses to AI-enabled advertising (Figure 1). This conceptual framework also serves as an initial step toward the development of a broader theoretical understanding and provides a basis for future empirical research in AI-enabled advertising research. Motivated by the Technology Acceptance Model (TAM) and its extensions, the diagram delineates how external variables, such as perceived privacy intrusion, privacy assurance, and information disclosure discomfort, exert positive(+) or negative(-) influences on perceived trustworthiness. Trustworthiness, in turn, along with perceived ad relevance and information quality, drives perceived ad value. These core perceptions collectively inform behavioural intentions, specifically purchase intent and click willingness. This framework visually captures the cognitive, affective, and ethical dimensions underpinning technology acceptance, situating consumer attitudes and actions within a robust theoretical context that has been widely validated in technology adoption research(Carlos Martins Rodrigues Pinho & Soares, 2011; F. Davis, 1987; Koufaris, 2002; Lu et al., 2003; Venkatesh & Bala, 2008; Venkatesh & Davis, 2000).

 

Figure:1 Proposed Conceptual Model “Factors Influencing Consumer Attitudes Toward AI-Enabled Advertising”

 

 

 

  1. Research Implications & Suggestions

The outcomes of this study provide critical managerial insights for advertisers, marketers, and platform developers who want to apply consumer-centric AI-enabled advertising strategies. The identification of nine distinct factors sheds light on the underlying psychological and behavioural mechanisms driving consumers perception regarding AI-generated advertisements in the Indian context. The constructs range from perceived privacy concerns to value perception and behavioural outcomes such as purchase intent and click Willingness. Each of these has actionable implications that can inform the design, deployment, and regulation of AI-driven marketing efforts.

 

8.1 Addressing Privacy Concerns

One of the most striking findings is the presence of Perceived Privacy Intrusion and Perceived Privacy Assurance as distinct but linked concepts. While the former reflects concerns about data misuse, the latter demonstrates customer trust in data protection procedures. Advertisers must go beyond compliance to implement proactive transparency strategies, such as real-time data usage notifications, opt-in consent systems, and streamlined privacy dashboards. These measures can assist to translate privacy concerns into trust, resulting in more positive customer responses to AI-based targeting.

 

8.2 Enhancing Personalization without Compromising Comfort

The survey found that consumers placed a high value on Ad-Relevance, indicating a preference for content suited to their interests and preferences. Over-personalization, on the other hand, may raise concerns about surveillance. Thus, marketers must prioritise contextual targeting over intrusive behavioural profiling. Personalisation tactics should be based on declared data (user choices) rather than inferred behavioural data alone to strike a balance between relevance and perceived intrusiveness.

 

8.3 Strengthening Consumer Trust

Trust was recognised as a key factor in shaping customer perceptions of AI-enabled advertisements. Firms must follow ethical AI norms, which include declaring the involvement of AI in ad generation or targeting and properly communicating the origins and accuracy of recommendations. Third-party audits and certifications of AI systems, as well as the incorporation of trust indicators (such as security badges or certified advertiser tags), can enhance the perception of legitimacy.

 

8.4 Designing Value-Centric Experiences

The study emphasises the dual character of Ad-Value, which includes both utilitarian and hedonic components. Marketers should concentrate on generating ad content that informs, entertains, and emotionally engages the audience. Using storytelling, interactivity, gamification, or immersive formats like augmented reality can improve the perceived value of AI-generated commercials and increase brand affinity.

 

8.5 Enhancing Perceived Information Quality

Customers expect AI-generated commercials to be informative, relevant, and simple to understand without being domineering. The aspect of Perceived Information Quality emphasises the importance of advertisements providing clear, actionable, and quality content that helps people make decisions. Excessively complicated or exaggerated communications might diminish efficacy and raise cognitive load over consumers’ mind. To improve perceived information quality, advertisers should prioritise clarity, practical benefits, and brand differentiation(Davies, 2001). AI technologies can help by using intelligent summarisation tools, natural language generation, and personalisation algorithms to customise the depth of data according to the user's preferences or behavioural patterns. The goal should be to deliver just enough information to inform users but not overwhelming them, hence increasing ad trust and enjoyment.

 

8.6 Addressing Information Disclosure Discomfort

The analysis revealed the persistence of Information Disclosure Discomfort as a unique component, indicating emotional resistance to perceived intrusiveness and overexposure in AI-enabled advertising(Milne et al., 2004; Phelps et al., 2000). Consumers' concerns about the scope of personal data usage, perceived surveillance, and a lack of openness in data handling processes frequently contribute to this uneasiness. To address these issues, marketers could implement techniques that encourage openness and increase consumer autonomy(Baek et al., 2014). Techniques including frequency capping, relevant ad placements, and user-control features (e.g., "hide this ad" or "why am I seeing this?" options) can empower users while reducing perceived coercion(Aguirre et al., 2015).

 

8.7 Encouraging Conversions through Trust and Utility

Purchase Intent and Click Willingness are direct indications of campaign efficacy. Ads must carefully guide customers through the buying funnel, using AI not only to capture their attention, but also to present appealing offers, clear calls-to-action, and seamless transitions to landing sites or in-app payment processes. Trust-enabling and relevance personalisation are critical enablers of these subsequent behaviours (Huh et al., 2020; Sirdeshmukh et al., 2002).

 

It is critical to recognise that some of these elements are both independent and interconnected. For example, Perceived Trustworthiness increases both Purchase Intent and Click Willingness, while Perceived Ad-Relevance improves both Perceived Ad-Value and Perceived Privacy Assurance. As a result, an integrative strategy that targets numerous dimensions at the same time, such as ethical AI, user-centric design, and transparent data practices, can result exponential engagement and hence conversions.

 

  1. Limitations and Future Research Directions

9.1 Limitations

While this study provides useful insights into consumer imprints of AI-enabled advertising in India, it must be noted that there are significant limitations. First, the study is geographically and culturally specific, focussing primarily on Indian consumers. Cultural differences in views towards privacy, trust, and technology use may restrict the findings' generalisability to Western or other Asian cultures, needing appropriate contextualisation. Second, the study used a cross-sectional approach to gather perceptions at a specific point in time. Longitudinal research may provide more dynamic insights as consumer views shift due to technical breakthroughs or strict legislative developments, such as the implementation of stronger data protection regulations. Third, relying on self-reported data has innate drawbacks, including the possibility of social desirability bias, recall problems, and misinterpretation of survey items. While reliability measures were generally acceptable, including behavioural data or observational approaches may improve the robustness of future findings. Finally, certain categories, particularly Click Willingness and Perceived Information Disclosure Discomfort, have lower Cronbach's alpha values, indicating possible weaknesses in measurement design. These constructs could benefit from additional refinement through qualitative research, expert validation, or scale construction to increase their clarity and comprehensiveness.

 

9.2 Future Research Directions

Future research can expand on the findings of this study in a variety of crucial directions. First, cross-cultural comparative research is required to understand how cultural norms, values, and legal frameworks influence consumer impressions of AI-enabled advertising. Such efforts would broaden the global applicability of theoretical models and inform more culturally sensitive marketing techniques. Second, longitudinal and experimental research methodologies can provide useful information about the changing nature of customer sentiments. Tracking changes over time or modifying specific ad attributes—such as transparency cues, personalisation depth, or data-use disclosures—may uncover causal correlations and increase our understanding of how consumers respond to emerging AI capabilities. Third, including behavioural and neuromarketing data—such as click-through rates, purchase behaviour, eye-tracking, or electroencephalography responses—can improve existing models by confirming self-reported notions like Information Disclosure Discomfort or Perceived Ad-Value with objective, real-time measurements. Fourth, future research should try to convert the current exploratory framework into a verified conceptual model through structural equation modelling (SEM). This would allow for the testing of mediating and moderating effects amongst components, resulting in a more sophisticated and predictive understanding of consumer behaviour in AI advertising ecosystems. Finally, considering the importance of characteristics such as Perceived Privacy Intrusion and Trustworthiness, more research into how customers view the ethical features of AI in advertising is required. Studies that look at the impact of data openness, opt-in methods, and computational comprehensibility on trust and regulatory compliance intents are especially promising in this era of increased online accountability.

REFERENCES
  1. Aguirre, E., Mahr, D., Grewal, D., de Ruyter, K., & Wetzels, M. (2015). Unraveling the Personalization Paradox: The Effect of Information Collection and Trust-Building Strategies on Online Advertisement Effectiveness. Journal of Retailing, 91(1), 34–49. https://doi.org/10.1016/j.jretai.2014.09.005
  2. Aguirre, E., Roggeveen, A. L., Grewal, D., & Wetzels, M. (2016). The personalization-privacy paradox: implications for new media. Journal of Consumer Marketing, 33(2), 98–110. https://doi.org/10.1108/JCM-06-2015-1458
  3. Andrews, M., Andrews, M., Luo, X., Fang, Z., & Ghose, A. (2016). Mobile ad effectiveness: Hyper-contextual targeting with crowdedness. Marketing Science, 35(2), 218–233. https://doi.org/10.1287/MKSC.2015.0905
  4. Azimi, J., Zhang, R., Zhou, Y., Navalpakkam, V., Mao, J., & Fern, X. (2012). Visual appearance of display ads and its effect on click through rate. ACM International Conference Proceeding Series, 495–504. https://doi.org/10.1145/2396761.2396826
  5. Baek, Y. M., Bae, Y., Jeong, I., Kim, E., & Rhee, J. W. (2014). Changing the default setting for information privacy protection: What and whose personal information can be better protected? Social Science Journal, 51(4), 523–533. https://doi.org/10.1016/j.soscij.2014.07.002
  6. Bang, H., Choi, D., Wojdynski, B. W., & Lee, Y. I. (2019). How the level of personalization affects the effectiveness of personalized ad messages: the moderating role of narcissism. International Journal of Advertising, 38(8), 1116–1138. https://doi.org/10.1080/02650487.2019.1590069
  7. Bansal, G., Zahedi, F. M., & Gefen, D. (2015). The role of privacy assurance mechanisms in building trust and the moderating role of privacy concern. European Journal of Information Systems, 24(6), 624–644. https://doi.org/10.1057/EJIS.2014.41
  8. Bleier, A., & Eisenbeiss, M. (2015a). Personalized online advertising effectiveness: The interplay of what, when, and where. Marketing Science, 34(5), 669–688. https://doi.org/10.1287/MKSC.2015.0930
  9. Bleier, A., & Eisenbeiss, M. (2015b). The Importance of Trust for Personalized Online Advertising. Journal of Retailing, 91(3), 390–409. https://doi.org/10.1016/j.jretai.2015.04.001
  10. Bleier, A., & Eisenbeiss, M. (2015c). The Importance of Trust for Personalized Online Advertising. Journal of Retailing, 91(3), 390–409. https://doi.org/10.1016/j.jretai.2015.04.001
  11. Brinson, N. H., & Eastin, M. S. (2016). Juxtaposing the persuasion knowledge model and privacy paradox: An experimental look at advertising personalization, public policy and public understanding. Cyberpsychology, 10(1). https://doi.org/10.5817/CP2016-1-7
  12. Carlos Martins Rodrigues Pinho, J., & Soares, A. M. (2011). Examining the technology acceptance model in the adoption of social networks. Journal of Research in Interactive Marketing, 5(2/3), 116–129. https://doi.org/10.1108/17505931111187767
  13. Cerny, B. A., & Kaiser, H. F. (1977). A Study Of A Measure Of Sampling Adequacy For Factor-Analytic Correlation Matrices. Multivariate Behavioral Research, 12(1), 43–47. https://doi.org/10.1207/S15327906MBR1201_3
  14. Chandra, S., Verma, S., Lim, W. M., Kumar, S., & Donthu, N. (2022). Personalization in personalized marketing: Trends and ways forward. In Psychology and Marketing. John Wiley and Sons Inc. https://doi.org/10.1002/mar.21670
  15. Chellappa, R. K., & Sin, R. G. (2005). Personalization versus Privacy: An Empirical Examination of the Online Consumer’s Dilemma. In Information Technology and Management (Vol. 6).
  16. Chen, G., Xie, P., Dong, J., & Wang, T. (2019). Understanding Programmatic Creative: The Role of AI. Journal of Advertising, 48(4), 347–355. https://doi.org/10.1080/00913367.2019.1654421
  17. Chen, H. (2024). The Impact of Personalized Social Media Advertising on Consumers’ Willingness to Click. 2024 IEEE 7th International Conference on Big Data and Artificial Intelligence, BDAI 2024, 225–229. https://doi.org/10.1109/BDAI62182.2024.10692416
  18. Cortina, J. M. (1993). What Is Coefficient Alpha? An Examination of Theory and Applications. Journal of Applied Psychology, 78(1), 98–104. https://doi.org/10.1037/0021-9010.78.1.98
  19. Culnan, M. J., & Armstrong, P. K. (1999). Information Privacy Concerns, Procedural Fairness, and Impersonal Trust: An Empirical Investigation. Organization Science, 10(1), 104–115. https://doi.org/10.1287/ORSC.10.1.104;REQUESTEDJOURNAL:JOURNAL:ORSC;WGROUP:STRING:PUBLICATION
  20. Daems, K., De Pelsmacker, P., & Moons, I. (2019). Advertisers’ perceptions regarding the ethical appropriateness of new advertising formats aimed at minors. Journal of Marketing Communications, 25(4), 438–456. https://doi.org/10.1080/13527266.2017.1409250
  21. Davies, M. A. P. (2001). Perceived information quality:An information processing perspective. Journal of International Consumer Marketing, 13(4), 29–50. https://doi.org/10.1300/J046V13N04_03;REQUESTEDJOURNAL:JOURNAL:WICM20;WGROUP:STRING:PUBLICATION
  22. Davis, F. (1987). User Acceptance of information Systems: The Technology Acceptance Model (TAM).
  23. Davis, F. D. (1989). Perceived Usefulness, Perceived Ease of Use, and User Acceptance of Information Technology. MIS Quarterly, 13(3), 319. https://doi.org/10.2307/249008
  24. Dinev, & Hart. (2006a). Privacy Concerns and Levels of Information Exchange: An Empirical Investigation of Intended e-Services Use. E-Service Journal, 4(3), 25. https://doi.org/10.2979/ESJ.2006.4.3.25
  25. Dinev, T., & Hart, P. (2006b). An extended privacy calculus model for e-commerce transactions. Information Systems Research, 17(1), 61–80. https://doi.org/10.1287/ISRE.1060.0080;WEBSITE:WEBSITE:INFORMS-SITE
  26. Diwanji, V. S., Lee, J., & Cortese, J. (2022). Deconstructing the role of artificial intelligence in programmatic advertising: at the intersection of automation and transparency. Journal of Strategic Marketing. https://doi.org/10.1080/0965254X.2022.2148269
  27. Dwivedi, Y. K., Ismagilova, E., Hughes, D. L., Carlson, J., Filieri, R., Jacobson, J., Jain, V., Karjaluoto, H., Kefi, H., Krishen, A. S., Tran, G. A., & Wang, Y. (2021). Setting the future of digital and social media marketing research: Perspectives and research propositions. International Journal of Information Management, 59. https://doi.org/10.1016/j.ijinfomgt.2020.102168
  28. Eisinga, R., Grotenhuis, M. te, & Pelzer, B. (2013). The reliability of a two-item scale: Pearson, Cronbach, or Spearman-Brown? International Journal of Public Health, 58(4), 637–642. https://doi.org/10.1007/s00038-012-0416-3
  29. Featherman, M. S., & Pavlou, P. A. (2003). Predicting e-services adoption: a perceived risk facets perspective. International Journal of Human-Computer Studies, 59(4), 451–474. https://doi.org/10.1016/S1071-5819(03)00111-3
  30. Forrest, E., & Hoanca, B. (2015). Artificial intelligence: Marketing’s game changer. Trends and Innovations in Marketing Information Systems, 45–64. https://doi.org/10.4018/978-1-4666-8459-1.CH003
  31. Fulgoni, G. M., & Mörn, M. P. (2009). Whither the click: How online advertising works. Journal of Advertising Research, 49(2), 134–142. https://doi.org/10.2501/S0021849909090175
  32. Goldfarb, A., & Tucker, C. (2012). Shifts in privacy concerns. American Economic Review, 102(3), 349–353. https://doi.org/10.1257/aer.102.3.349
  33. Hair, J. F. J., Anderson, R. E., Tatham, R. L., & Black, W. C. (1998). Multivariate Data Analysis (5th ed.). Prentice Hall International.
  34. Harrison McKnight, D., Choudhury, V., & Kacmar, C. (2002). The impact of initial consumer trust on intentions to transact with a web site: a trust building model. The Journal of Strategic Information Systems, 11(3–4), 297–323. https://doi.org/10.1016/S0963-8687(02)00020-3
  35. Huang, M. H., & Rust, R. T. (2021). A strategic framework for artificial intelligence in marketing. Journal of the Academy of Marketing Science, 49(1), 30–50. https://doi.org/10.1007/S11747-020-00749-9
  36. Huh, J., Kim, H., Rath, B., Lu, X., & Srivastava, J. (2020). You reap where you sow: a trust-based approach to initial seeding for viral advertising. International Journal of Advertising, 39(7), 963–989. https://doi.org/10.1080/02650487.2020.1718823
  37. Hutcheson, G. D., & Sofroniou, N. (1999). The Multivariate Social Scientist : Introductory Statistics Using Generalized Linear Models. 1–288.
  38. Joseph F. Hair Jr., William C. Black, Barry J. Babin, & Rolph E. Anderson. (2014). Multivariate Data Analysis (7th ed.). Pearson Education Limited.
  39. KANTAR; IAMAI. (2025). INTERNET IN INDIA 2024. https://www.iamai.in/sites/default/files/research/Kantar_%20IAMAI%20report_2024_.pdf
  40. Kim, H., & Huh, J. (2017). Perceived Relevance and Privacy Concern Regarding Online Behavioral Advertising (OBA) and Their Role in Consumer Responses. Journal of Current Issues and Research in Advertising, 38(1), 92–105. https://doi.org/10.1080/10641734.2016.1233157
  41. Koufaris, M. (2002). Applying the Technology Acceptance Model and flow theory to online Consumer Behavior. Information Systems Research, 13(2), 205–223. https://doi.org/10.1287/ISRE.13.2.205.83
  42. Kumar, D., & Suthar, N. (2024). Ethical and legal challenges of AI in marketing: an exploration of solutions. Journal of Information, Communication and Ethics in Society, 22(1), 124–144. https://doi.org/10.1108/JICES-05-2023-0068/FULL/PDF
  43. Laczniak, R. N., & Muehling, D. D. (1993). Toward a Better Understanding of the Role of Advertising Message Involvement in Ad Processing. In Psychology & Marketing: Vol. lO (Issue 4). John Wiley & Sons, Inc.
  44. Lee, D. J., Ahn, J. H., & Bang, Y. (2011). Managing consumer privacy concerns in personalization: A strategic analysis of privacy protection. MIS Quarterly: Management Information Systems, 35(2), 423–444. https://doi.org/10.2307/23044050
  45. Lou, C., & Yuan, S. (2019). Influencer Marketing: How Message Value and Credibility Affect Consumer Trust of Branded Content on Social Media. Journal of Interactive Advertising, 19(1), 58–73. https://doi.org/10.1080/15252019.2018.1533501
  46. Lu, J., Yu, C. S., Liu, C., & Yao, J. E. (2003). Technology acceptance model for wireless Internet. Internet Research, 13(3), 206–222. https://doi.org/10.1108/10662240310478222
  47. MacKenzie, S. B., & Lutz, R. J. (1989). An Empirical Examination of the Structural Antecedents of Attitude toward the Ad in an Advertising Pretesting Context. Journal of Marketing, 53(2), 48–65. https://doi.org/10.1177/002224298905300204
  48. Malhotra, N. K., Kim, S. S., & Agarwal, J. (2004). Internet Users’ Information Privacy Concerns (IUIPC): The Construct, the Scale, and a Causal Model. Information Systems Research, 15(4), 336–355. https://doi.org/10.1287/isre.1040.0032
  49. Martin, K. D., Borah, A., & Palmatier, R. W. (2017). Data privacy: Effects on customer and firm performance. Journal of Marketing, 81(1), 36–58. https://doi.org/10.1509/JM.15.0497
  50. McKnight, D. H., Choudhury, V., & Kacmar, C. (2002). Developing and validating trust measures for e-commerce: An integrative typology. Information Systems Research, 13(3), 334–359. https://doi.org/10.1287/ISRE.13.3.334.81;REQUESTEDJOURNAL:JOURNAL:ISRE;WGROUP:STRING:PUBLICATION
  51. Meng, X., & Liu, J. (2025). “Talk to me, I’m secure”: investigating information disclosure to AI chatbots in the context of privacy calculus. Online Information Review. https://doi.org/10.1108/OIR-06-2024-0375
  52. Milne, G. R., Rohm, A. J., & Bahl, S. (2004). Consumers’ protection of online privacy and identity. Journal of Consumer Affairs, 38(2), 217–232. https://doi.org/10.1111/J.1745-6606.2004.TB00865.X
  53. Nill, A., & Aalberts, R. J. (2014). Legal and Ethical Challenges of Online Behavioral Targeting in Advertising. Journal of Current Issues and Research in Advertising, 35(2), 126–146. https://doi.org/10.1080/10641734.2014.899529
  54. Phelps, J., Nowak, G., & Ferrell, E. (2000). Privacy concerns and consumer willingness to provide personal information. Journal of Public Policy and Marketing, 19(1), 27–41. https://doi.org/10.1509/JPPM.19.1.27.16941
  55. Qin, X., & Jiang, Z. (2019). The Impact of AI on the Advertising Process: The Chinese Experience. Journal of Advertising, 48(4), 338–346. https://doi.org/10.1080/00913367.2019.1652122
  56. Rapp, J., Hill, R. P., Gaines, J., & Wilson, R. M. (2009). Advertising and consumer privacy: Old practices and new challenges. Journal of Advertising, 38(4), 51–61. https://doi.org/10.2753/JOA0091-3367380404
  57. Reyes, J., Batmaz, A. U., & Kersten-Oertel, M. (2025). Trusting AI: does uncertainty visualization affect decision-making? Frontiers in Computer Science, 7. https://doi.org/10.3389/FCOMP.2025.1464348
  58. Rosenkrans, G., & Myers, K. (2018). Optimizing Location-Based Mobile Advertising Using Predictive Analytics. Journal of Interactive Advertising, 18(1), 43–54. https://doi.org/10.1080/15252019.2018.1441080
  59. Roy, A., Huh, J., Pfeuffer, A., & Srivastava, J. (2017). Development of Trust Scores in Social Media (TSM) Algorithm and Application to Advertising Practice and Research. Journal of Advertising, 46(2), 269–282. https://doi.org/10.1080/00913367.2017.1297272
  60. Sahni, N. S., Wheeler, S. C., & Chintagunta, P. (2018). Personalization in email marketing: The role of noninformative advertising content. Marketing Science, 37(2), 236–258. https://doi.org/10.1287/MKSC.2017.1066
  61. Sirdeshmukh, D., Singh, J., & Sabol, B. (2002). Consumer trust, value, and loyalty in relational exchanges. Journal of Marketing, 66(1), 15–37. https://doi.org/10.1509/jmkg.66.1.15.18449
  62. Smith, H. J., Milberg, S. J., & Burke, S. J. (1996). Information privacy: Measuring individuals’ concerns about organizational practices. MIS Quarterly: Management Information Systems, 20(2), 167–195. https://doi.org/10.2307/249477
  63. Soh, H., Reid, L. N., & King, K. W. (2013). Measuring Trust In Advertising. Http://Dx.Doi.Org/10.2753/JOA0091-3367380206, 38(2), 83–104. https://doi.org/10.2753/JOA0091-3367380206
  64. Tadelis, S., & Zettelmeyer, F. (2015). Information disclosure as a matching mechanism: Theory and evidence from a field experiment. American Economic Review, 105(2), 886–905. https://doi.org/10.1257/AER.20110753
  65. Tavakol, M., & Dennick, R. (2011). Making sense of Cronbach’s alpha. International Journal of Medical Education, 2, 53. https://doi.org/10.5116/IJME.4DFB.8DFD
  66. Taylor, C. R. (2019). Editorial: Artificial intelligence, customized communications, privacy, and the General Data Protection Regulation (GDPR). International Journal of Advertising, 38(5), 649–650. https://doi.org/10.1080/02650487.2019.1618032
  67. Tobias, S., & Carlson, J. E. (1969). BRIEF REPORT: BARTLETT’S TEST OF SPHERICITY AND CHANCE FINDINGS IN FACTOR ANALYSIS. Multivariate Behavioral Research, 4(3), 375–377. https://doi.org/10.1207/S15327906MBR0403_8
  68. Tran, M. T. (2024). A Systematic Analysis of Artificial Intelligence’s Usage in Online Advertising: How Artificial Intelligence Affects Digital Marketing Initiatives. Https://Services.Igi-Global.Com/Resolvedoi/Resolve.Aspx?Doi=10.4018/979-8-3693-4453-8.Ch002, 17–27. https://doi.org/10.4018/979-8-3693-4453-8.CH002
  69. Tucker, C. E. (2014). Social networks, personalized advertising, and privacy controls. Journal of Marketing Research, 51(5), 546–562. https://doi.org/10.1509/jmr.10.0355
  70. Van-Tien Dao, W., Nhat Hanh Le, A., Ming-Sung Cheng, J., & Chao Chen, D. (2014). Social media advertising value. International Journal of Advertising, 33(2), 271–294. https://doi.org/10.2501/IJA-33-2-271-294
  71. Vaske, J. J., Beaman, J., & Sponarski, C. C. (2017). Rethinking Internal Consistency in Cronbach’s Alpha. Leisure Sciences, 39(2), 163–173. https://doi.org/10.1080/01490400.2015.1127189;WEBSITE:WEBSITE:TFOPB;PAGEGROUP:STRING:PUBLICATION
  72. Venkatesh, Morris, Davis, & Davis. (2003). User Acceptance of Information Technology: Toward a Unified View. MIS Quarterly, 27(3), 425. https://doi.org/10.2307/30036540
  73. Venkatesh, V., & Bala, H. (2008). Technology Acceptance Model 3 and a Research Agenda on Interventions. Decision Sciences, 39(2), 273–315. https://doi.org/10.1111/j.1540-5915.2008.00192.x
  74. Venkatesh, V., & Davis, F. D. (2000). A Theoretical Extension of the Technology Acceptance Model: Four Longitudinal Field Studies. Management Science, 46(2), 186–204. https://doi.org/10.1287/mnsc.46.2.186.11926
  75. Venkatesh, V., Thong, J. Y. L., & Xu, X. (2012). Consumer acceptance and use of information technology: Extending the unified theory of acceptance and use of technology. MIS Quarterly: Management Information Systems, 36(1), 157–178. https://doi.org/10.2307/41410412
  76. Vimalkumar, M., Sharma, S. K., Singh, J. B., & Dwivedi, Y. K. (2021). ‘Okay google, what about my privacy?’: User’s privacy perceptions and acceptance of voice based digital assistants. Computers in Human Behavior, 120. https://doi.org/10.1016/J.CHB.2021.106763
  77. Wang, G., Xu, J., & Lin, S. (2023). Influencing factors of using behaviour for computational advertising under the theoretical model of UTAUT. Journal of Control and Decision, 10(1), 19–25. https://doi.org/10.1080/23307706.2022.2082568;CSUBTYPE:STRING:SPECIAL;PAGE:STRING:ARTICLE/CHAPTER
  78. Watkins, M. W. (2018). Exploratory Factor Analysis: A Guide to Best Practice. Journal of Black Psychology, 44(3), 219–246. https://doi.org/10.1177/0095798418771807/ASSET/552AF49C-E36B-43E3-A023-FA2CEEA4A6A2/ASSETS/IMAGES/LARGE/10.1177_0095798418771807-FIG1.JPG
  79. Worthington, R. L., & Whittaker, T. A. (2006). Scale Development Research. The Counseling Psychologist, 34(6), 806–838. https://doi.org/10.1177/0011000006288127
  80. Wu, L., & Wen, T. J. (2021). Understanding ai advertising from the consumer perspective what factors determine consumer appreciation of ai-created advertisements? Journal of Advertising Research, 61(2), 133–146. https://doi.org/10.2501/JAR-2021-004
  81. Xu, C., Gao, F., & Han, L. (2025). Enhancing user information disclosure intention in dynamic conversations of intelligent recommendation systems based on large language models: A perspective of user gratification and privacy calculus. International Journal of Human Computer Studies, 200. https://doi.org/10.1016/J.IJHCS.2025.103511
  82. Yoo, C. Y. (2009). Effects beyond click-through: Incidental exposure to web advertising. Journal of Marketing Communications, 15(4), 227–246. https://doi.org/10.1080/13527260802176419
  83. 83.   Zuboff, S. (2019). Surveillance Capitalism and the Challenge of Collective Action. New Labor Forum, 28(1), 10–29. https://doi.org/10.1177/1095796018819461/ASSET/953342DD-F595-40AA-9440-5E9F95C61AEB/ASSETS/IMAGES/LARGE/10.1177_1095796018819461-FIG1.JPG
Recommended Articles
Research Article
GenAI: A Technological Catalyst for Reshaping India’s Banking Sector
Published: 17/01/2026
Research Article
Shaping the Future of Agritech: Generative AI-Driven Drones and Their Impact on Farming Value Chains
Published: 16/01/2026
Research Article
Impact of Artificial Intelligence–Driven Talent Analytics on Marketing Performance and Financial Outcomes of Organizations
...
Published: 16/01/2026
Research Article
AI Adoption in HRM and Employee Acceptance: A Behavioral Perspective
...
Published: 31/12/2025
Loading Image...
Volume 7, Issue 1
Citations
3 Views
2 Downloads
Share this article
© Copyright Journal of International Commercial Law and Technology