Tag Archives: ChatGPT

AI Is Trending In Canada But Why Isn’t Productivity?

April, 2026 – Capital spending on AI has surged into the hundreds of billions, yet the economic payoff remains elusive. Despite rapid investment and widespread experimentation with generative AI tools, measurable productivity gains have yet to show up clearly in aggregate data across advanced economies – raising an important question: who will be first to turn this momentum into measurable gains?

In “From Hype to Output: How AI Investment Translates to Real Productivity Gains,” I discovered that while AI offers a path to stronger economic performance and could help address Canada’s long-standing productivity challenges, realizing these gains will require policies that accelerate business adoption and diffusion, which remain uneven across most industries and regions. This includes removing barriers, ensuring access to data resources and encouraging integration across firms and sectors.

For the Silo, Rosalie Wyonch/C.D. Howe

  • Artificial intelligence has renewed optimism about improving Canada’s long-standing productivity problem, but history shows that general-purpose technologies often take decades to produce measurable economy-wide gains. Early adoption typically produces modest improvements as firms experiment and invest in complementary assets such as data, skills, and organizational changes.
  • Canada performs reasonably well in international AI rankings and produces strong research output, but it lags leading countries in computing capacity and commercialization. Canadians are among the most frequent users of generative AI tools globally, yet business adoption remains uneven and relatively limited.
  • Canadian data show that about 12 percent of businesses currently use AI, and most report little change in employment following adoption. Many firms say AI is not relevant to their operations or lack the knowledge needed to implement it effectively, suggesting adoption barriers remain significant.
  • Policy should therefore prioritize accelerating AI adoption and diffusion across sectors while maintaining Canada’s research and infrastructure capacity. Governments can support this by improving regulatory clarity, expanding data access, strengthening AI literacy, and encouraging firms to experiment and integrate AI into their operations.

Introduction

Canada faces a persistent productivity challenge.

Growth in output per work-hour has stagnated for decades, and the gap with the United States continues to widen.1 Policymakers have long sought levers to reverse this trend, and the rapid emergence of artificial intelligence (AI) – particularly generative AI since late 2022 – has prompted renewed optimism that a transformative technology might finally deliver the productivity gains that have proved elusive.

This Commentary discusses the impacts of AI technologies, with a focus on more recent generative AI. In particular, these technologies are unique in their low barriers to adoption and use in terms of both skill and cost as generalized tools. With such low barriers to entry, there is potential for rapid adoption and scaling of AI technologies across many sectors of the economy. Yet enthusiasm must be tempered by evidence: general-purpose technologies (such as AI) historically take decades to reshape economies, and the path from adoption to measurable productivity growth is neither linear nor guaranteed.

This Commentary examines AI’s potential contribution to Canadian productivity growth through an evidence-based policy lens. It begins by establishing what AI technologies are and how technological change translates into economic output, drawing on the concept of the “productivity J-curve” to explain why early-stage adoption often fails to register in macroeconomic statistics. The paper then situates the current AI investment boom in a historical and international context and summarizes research quantifying both the infrastructure-driven and adoption-driven channels through which AI may affect GDP. A comparative analysis positions Canada among its international peers, revealing a paradox: world-class research output coexists with middling commercial translation and infrastructure capacity. Detailed examination of Canadian business adoption data shows significant variation across industries and provinces, with early signs that initial experimentation may be giving way to more selective, sustained implementation.

The central argument is that maximizing the likelihood that AI technologies will boost Canada’s productivity growth requires policy focused on accelerating AI adoption and diffusion across sectors. While government investment in computing capacity supports the domestic development of AI technologies (and democratizes access to computing power), government resource constraints, significant uncertainty about AI development trajectories, and the relatively smaller size of the Canadian AI economy suggest that AI infrastructure policies should focus on maintaining and improving Canada’s relative international competitiveness. Policies that encourage firms to experiment, learn, and eventually reimagine their operations around AI capabilities can position Canada to capture productivity gains as the technology matures. AI, as a generalized technology, can be deployed across many industries and is linked to other economic development and industrial policy priorities, including small- and medium-sized business growth and international trade diversification. This paper concludes with policy recommendations that balance the uncertain timeline of AI’s macroeconomic impact against the risk of falling further behind more aggressive international competitors.

The Current AI Boom in Context

The emergence of generative AI since late 2022 has triggered an unprecedented wave of capital investment and rapid enterprise adoption. This technological shift is reshaping economic growth dynamics through two channels: the direct contribution of infrastructure investment to GDP and longer-term productivity gains from AI adoption across industries. Understanding the magnitude, timing, and distribution of these effects is essential for policymakers seeking to position their economies to benefit from AI’s transformative potential. Much of this section refers to the US economy, where more data on AI investment and GDP effects are available.

Capital expenditure on AI infrastructure has reached historic proportions. Hyperscaler companies – Amazon, Google, Meta, Microsoft, and Oracle – allocated about $342 billion to capital expenditures in 2025, a 62 percent increase from the previous year (Aliaga 2025). Estimates suggest AI-related capital expenditures could reach US$527 billion in 2026 (Goldman Sachs 2025). This spending encompasses semiconductors, data centres, networking equipment, and the power infrastructure required to support “compute”-intensive AI workloads.

In Canada, the federal government announced $2 billion over five years starting in 2024-25 for the Canadian Sovereign AI Strategy, including $705 million for “compute” infrastructure. Microsoft has also announced it is spending $19 billion on AI infrastructure investment in Canada between 2023 and 2027, with $7.5 billion of that over the next two years (Smith 2025). Industry categories related to AI2 accounted for $195 billion in 2021 and grew to $229 billion in 2024, representing 9.3 to 10 percent of Canada’s GDP, respectively.

The impact of AI investment on the US economy is significant, though estimates vary widely. In the first nine months of 2025, GDP product categories related to AI investment (such as computing and communications equipment, data centre structures, software, and research and development) represented 37 percent of real GDP growth and 8 percent of real GDP (Levine 2025).3 Other estimates suggest AI contributed roughly 1 percentage point to US GDP growth in 2025 (Boussour 2025; Goldman Sachs 2025; Aliaga 2025), making it the second-largest contributor to growth after consumption (Singh 2026).4 The scale of investment in AI infrastructure in the US has led to debate about whether it is fueling an equity market bubble (BNP Paribas 2025; Barnette and Peterson 2025). There have been multiple “AI winters” in the past – periods of significant decline in enthusiasm, funding, and progress in the field of AI.5

However, much of the spending on AI infrastructure does not translate directly into domestic GDP because many inputs are imported – for example, semiconductors manufactured in Taiwan. This is particularly important for Canada, since both infrastructure inputs and the outputs of dominant AI companies (AI products and software) are predominantly located abroad. Secondary economic effects and labour dynamics associated with the expansion of infrastructure investment are also important to consider. Data centre construction has had significant impacts on the construction industry in the United States. The top 50 contractors by size have doubled revenues within a year, and salaries for trades workers are 32 percent higher for data centre projects compared to other construction (Paoli 2025). Once built, however, data centres require relatively few permanent workers.

There are also counter-balancing factors to consider. Corrado et al. (2025) show that while investment in data as an intangible asset can improve efficiency, the growing importance of proprietary datasets slows the diffusion of innovation. So far, the negative impact on diffusion outweighs the efficiency gains from intangible data capital.

Currently, the main channel through which AI is measurably increasing US GDP is capital expenditures. Over the longer term, however, these investments must translate to products and services that meaningfully increase productivity to have a sustained effect on growth. So far, the acceleration of AI development and diffusion has not been associated with higher productivity growth at the macroeconomic level across G7 countries (Filippucci et al. 2024; Andre and Gal 2024; Goldin et al. 2024). As with earlier general-purpose technologies, widespread productivity gains may take years to materialize.

Estimates of AI’s long-term productivity impact for the US vary widely – from 1.5 percent labour-productivity growth annually (Goldman Sachs 2025) to less than 1 percent cumulatively over a decade (Acemoglu 2024), as shown in Figure 1. Some researchers argue that the long-run impact may instead arise from persistently faster growth rather than a one-time productivity shift (Baily, Brynjolfsson, and Korinek 2023). The timeline and magnitude of significant macroeconomic growth related to AI adoption depend on continued improvements in AI capabilities, widespread adoption, and complementarities with human skills and other technologies (Filippucci et al. 2024).

AI Already Affecting Labour Market

There is some evidence that AI is already having labour market effects: Brynjolfsson, Chander, and Chen (2025) estimate that early career workers in AI-exposed occupations experience a 16 percent relative employment decline while employment remained stable for more experienced workers. Across occupational categories in the US, 2.6-75.5 percent of occupations could have at least 50 percent of their tasks affected by GenAI (Arnon 2025). Research on the UK labour market finds that 11 percent of occupational tasks are exposed to generative AI automation/augmentation as part of “low-hanging fruit” implementation cases, and up to 59 percent of tasks will be exposed to Gen-AI as it develops into more integrated systems (Jung and Desikan, 2024). Heneseke et al. (2025) find that the price premium for AI-exposed tasks declined by 12 percent from 2017 to 2023/24 and that job postings were 5.5 percent lower in 2025-Q2 than if pre-GPT hiring patterns had persisted.

As AI technology is adopted, some labour market disruption is inevitable, but that does not mean that AI necessarily leads to widespread unemployment. Job losses can occur only if innovation outstrips growth in demand for new products and services. Further, the potential for automation does not necessarily translate into actual automation: the decision to automate depends on factors such as firm size, competitive pressure, and the cost of a machine versus the cost of human labour. As technology diffuses and improves productivity, GDP and wages increase, which increases demand for goods and services, creating new employment. Rapid technological progress is regularly accompanied by fears of widespread unemployment. In reality, however, markets adjust dynamically to technology adoption over time, and widespread unemployment is unlikely (Oschinski and Wyonch 2017).6

At the firm level, growing evidence suggests AI can improve labour productivity in many contexts (Figure 2, Table A1). Meta-analysis of research measuring the productivity effects of generative AI shows mixed results across studies and applications. But on average, the use of GenAI tools increased labour productivity (in particular, tasks or departments) by 17 percent (Coupe and Wu 2025).

Despite these gains, there is a notable divide in the wide adoption of tools like ChatGPT and enterprise-grade AI systems (Challapally et al. 2025). Nearly 80 percent of organizations report exploring or piloting large language model (LLM) products, and 40 percent report deployment. These tools primarily enhance individual worker productivity but are not deployed for core business functions.7 By contrast, 60 percent of companies have evaluated enterprise or custom systems, yet only 5 percent reach production.

Tools that succeeded had low configuration barriers and immediately noticeable value, while those that require extensive upfront customization often stalled in the pilot stage. The core barrier to scaling is not infrastructure, regulation, or talent. It is GenAI’s lack of capacity to learn or remember over time (particularly static enterprise tools relative to rapidly evolving consumer-facing tools).8 Aside from enterprise-level adoption and development, there is significant adoption of AI technologies by individual workers to enhance personal productivity. In 2024, 40 percent of the working-age population in the US was using GenAI, 23 percent used it for work, and 9 percent used it every workday (Bick, Blandin, and Deming 2025). The high failure rate of AI pilots can be explained by the productivity J-curve and the current stage of development and adoption (intangible capital accumulation that leads to future productivity improvement). A high failure rate of pilots shows experimentation and contributes to ongoing development on the one hand. On the other, companies that experiment with AI and abandon it might be less likely to adopt it in the future, despite rapid evolution.

Uneven Gains

Productivity gains from new technologies are uneven across sectors and difficult to predict because they depend on where technological change occurs.9 Adoption is also shaped by market competition, regulation, and the current distribution of technology within industries (Howitt 2015). Given the uncertainty around AI’s trajectory and the sectors where it may generate the largest gains, productivity improvements are possible, but expectations may also exceed outcomes. Bridging from the initial development and experimentation phase to widespread adoption, eventual integration and process changes can be accelerated by addressing technical, institutional, and bureaucratic barriers (Ouimetter, Teather, and Allison 2024). Empirical modelling suggests that people, process, and data readiness are required in addition to technology/capital investment to achieve long term operational success (Uren and Edwards 2023). Government policy should be economically broad-based and not narrowly focused on particular industries or applications. It should also be cautious and balance uncertain long-run productivity gains with maintaining competitiveness in a rapidly developing global technology market.

How Canada Compares Internationally

Canada performs reasonably well in international AI rankings, placing eighth overall across five different international AI rankings (Figure 3).10 Rankings vary widely depending on their focus and data sources (Table A3). For example, China’s rank ranges from second in the Global AI Index ranking but 32nd in the Global Index on Responsible AI (GCG 2024). Canada ranks highest in the Global AI Index (eighth) and lowest on the IMF’s AI Preparedness Index (18th). Canada has also fallen in the rankings over time – from fourth in 2021 to eighth in 2025 on the Global AI Index (White and Cesareo 2025) and from fifth in 2022 to 12th in 2025 on the Government AI Readiness ranking (Rogerson et al. 2023; Iida et al. 2026). Notably, Canada’s score improved across categories measured in the Oxford AI Readiness Index, suggesting that advancements in AI adoption, development, and public policy have been made, but they have not kept pace with some international peers.

In terms of development and infrastructure, Canada is a high performer, but not a clear leader among the metrics tracked by the OECD’s AI Policy Observatory. More AI research is produced in Canada than in the US or UK (relative to population size). Canada has higher venture capital investment (as a share of GDP) in AI startups than many peer countries, but falls far behind the US and Israel, the clear leaders in the category (OECD.AI 2026a,b). Canada also maintains a respectable presence in computing infrastructure, with 19 of the world’s top 500 supercomputers (Top500 2025) and public cloud regions (Lehdonvirta et al. 2025). Unfortunately, the processing power of those computers ranks only 18th out of 20 comparator countries (Top500 2025).

Strong Research But Weak Commercialization For Canada

Taken together, these indicators reflect a familiar pattern in Canada’s innovation system: strong research output but weaker commercialization. Continued investment in data and processing infrastructure will help maintain competitiveness, but public policy should also encourage private investment.11 Given the scale of investment in countries such as the United States and China, Canada is unlikely to match their computing capacity regardless of policy choices.

Evaluating the usage of major LLMs by country shows that the US and India account for the largest proportions of total users and interactions. After accounting for population size, however, Canada, Australia, and France show the highest usage rates among individuals (Figure 4).12 While individual use does not directly translate to higher productivity in the production of goods and services, familiarity and comfort with AI tools among the general labour force will improve enterprise adoption prospects. A labour force with AI skills reduces enterprise training costs related to adoption and improves the likelihood of employees generating ideas for reorganizing business processes. Conversely, differences between consumer and enterprise applications can create barriers to workplace adoption (Challapally et al. 2025). Even so, if individual workers are using AI tools to increase their personal productivity, there will be marginal productivity effects from using publicly available tools for work tasks.

High individual adoption in Canada contrasts with lower rates of business adoption.

Based on the most recent comparable data, Canada falls below the OECD average overall and for the proportion of medium and large businesses adopting AI (Figure 5). In Korea, Denmark, and Finland, more than half of large companies are using AI.

International data on the businesses’ adoption of AI is highly variable. The OECD estimates 20.2 percent of businesses (with more than 10 employees) use AI,13 while the McKinsey Global Survey finds that 88 percent of businesses experiment with AI in at least one function (McKinsey 2025; OECD 2025). However, only 10 percent of businesses in G7 countries were using AI for core business functions in 2024. Only 2 to 6 percent of firms across G7 countries have high intensity adoption related to core business functions (Fillipucci et al. 2025). High-intensity adoption is highest in the US, followed by Canada, the UK, and Germany.

There are common features of AI adoption across countries:

  • Firm size: larger businesses are using AI at higher rates than medium or small businesses.14 This could be related to factors including fixed costs of adoption, larger data resources, lower financing constraints, and complementary intangible assets such as information and communication technology (ICT) skills and research capacity (Calvino and Fontanelli 2023).
  • Industry concentration: adoption is highest in ICT, followed by professional, scientific and technical services, and communications and marketing (OECD 2025).15
  • Firm characteristics: higher productivity and younger enterprises are more likely to adopt AI (Calvino and Fontanelli 2023; Calvino, Costa, and Haerie, forthcoming).16

Modelled projections suggest that high-intensity AI adoption among enterprises could reach 13 to 63 percent across G7 countries over the next decade, depending on adoption rates and technological progress (Fillipucci et al. 2025). The US is projected to have the highest adoption rates and the largest annual productivity growth related to AI, while Japan is projected to see the lowest. Canada ranks second in projected adoption across scenarios, but third or fourth in projected labour productivity growth, meaning that productivity gains will likely be relatively modest compared to international peers with similar (and slightly lower) adoption rates (Figure 6).17 Estimates suggest that Canada will reach 32-62 percent enterprise adoption in the next 10 years, contributing 0.35 to 1.13 percentage points to annual labour productivity growth.

Overall, available data show that Canada is about average in terms of adoption and development capacity, and the US is a clear leader. Canada has a lower capacity for development related to AI infrastructure than some peer countries, but remains above average. Canadian individuals have higher adoption rates than those in most other countries. Business adoption is less clear-cut. Some indicators place Canada below the international average, while others rank Canada second in the G7 for high-intensity AI adoption in core business functions. For Canada to close the productivity gap, it will need sufficient AI infrastructure to remain competitive, broad adoption, but most importantly, more rapid or effective AI-enabled innovation to develop adaptations of business processes and new products and services.

Canada’s AI Adoption Landscape

Having compared Canada with international peers, this section examines business adoption within Canada.

In the third quarter of 2025, 14.5 percent of businesses were planning to use AI within the next 12 months, a 3.9 percentage point increase from the previous year. As of the second quarter of 2025, 12.2 percent were already using AI. But two-thirds of businesses report no plans to adopt AI, and 18.9 percent were uncertain (Bryan, Sood, and Johnston 2025a). About six in 10 businesses think that AI investment is not important or not relevant to their business.

Planned adoption is highest in Ontario, British Columbia, Nova Scotia, and New Brunswick, with New Brunswick seeing the largest growth in the proportion of businesses planning to use AI compared to the previous year (Figure 7). Adoption is highest in information and cultural industries; finance and insurance; professional, scientific, and technical services; and healthcare and social services. It is lowest in mining and resources extraction, transportation and warehousing, and construction (Table A1).

Comparing previous and planned AI use shows that the share of businesses planning to adopt AI over the next 12 months is only 2.3 percentage points higher, compared with a 6.1 percentage-point increase from 2024 to 2025. In several sectors – including manufacturing, resource extraction, wholesale and retail trade, and professional services – planned adoption is lower than current usage levels in Q2 2025 (Table 1). Adoption is most likely to increase in accommodation, food services, and real estate over the next year.

Provincial patterns are also uneven. Planned AI use in the information and cultural industries declines in five provinces (including a 39.7-percentage-point drop in New Brunswick) but rises in Ontario, Alberta, and Saskatchewan. In Q2 2025, businesses in Ontario, Quebec, and Manitoba were using AI at higher rates than the national average. By Q3 2026, however, only businesses in Ontario intend to use AI at higher rates than the national average. Businesses in Quebec and Alberta show little appetite for further AI adoption, while those in Prince Edward Island report plans to reduce AI use in 2026 compared with 2025.

Businesses engaging in some international activities are adopting AI at higher rates. Across all provinces, businesses that export services were using AI at higher rates than average. The same is true for importing services in all provinces except Nova Scotia. Businesses that relocated activities outside Canada or made investments outside Canada are also more likely to be using AI at the national level (though these same businesses show the largest declines in intended use over the coming year). Importing and exporting goods has no clear association with AI adoption across provinces.

Overall, businesses involved in international services trade or cross-border investment show higher rates of AI use. This gap will likely close slightly over the year: firms without international activities report higher intentions to adopt AI than the overall business average (3.6 percent compared with 2.3 percent). Meanwhile, firms relocating employees or operations abroad, or investing internationally, report declining intentions to expand AI use.

Canada’s AI adoption landscape (as of Q2 2025) generally mirrors international trends. Younger businesses (10 years or less) adopt AI more frequently than older firms, and large firms adopt more than small ones. The highest adoption rates are concentrated in ICT, finance, and professional, scientific, and technical services. Private enterprises had significantly higher adoption than government agencies and non-profits.

Planned adoption over the next year (reported in Q3 2025) suggests both continuation of these trends and broader diffusion. Government agencies and non-profits serving businesses plan to increase the use of AI at higher proportions than private enterprises. Only 0.1 percent of government agencies reported using AI in early 2025, but 24.3 percent plan on using it within the next year. Younger firms plan to continue to adopt AI at higher rates than older firms.

Planned adoption by firm size varies across provinces. In New Brunswick and Saskatchewan, small (5-19 employees) and medium-sized (20-99 employees) companies report adoption intentions above the national average, while larger firms show declining adoption. In Quebec, by contrast, larger firms continue to adopt AI more rapidly than smaller firms, widening the adoption gap.

Turning to employment effects, most businesses using AI report little change in staffing levels. Among firms that adopted AI in the previous 12 months (12.2 percent of businesses in Q2 2025), 89.4 percent reported no related change in employment (Bryan Sood and Johnston 2025b). Similarly, about 70 percent of firms planning to adopt AI expected no employment change, while another 10 percent are uncertain (Bryan Sood and Johnston 2025a). These results suggest firms may overestimate potential labour savings before implementation. After adoption, 15.1 percent of firms reported no reduction in employee tasks, while 47.2 percent reported only small impacts.

For businesses that do not plan to use AI in the next year, the most common reason is that the technology is not relevant to their products or services (78.1 percent). Only 1.5 percent of businesses said that previous use of AI did not meet expectations. Other common barriers include limited knowledge of AI, privacy, or security concerns, and the technology’s perceived immaturity. Industries with fewer firms reporting AI as irrelevant tend to report lack of knowledge as the primary barrier to adoption.18

Government agencies report the highest rates of disappointment with AI performance (14.8 percent citing unmet expectations as the reason for non-use).19 Regionally, businesses in British Columbia, Alberta, and Ontario report the lowest rates of saying AI is irrelevant to their activities (74.1-76.7 percent), while New Brunswick reports the highest share (88.3 percent). Notably, laws and regulations preventing or restricting the use of AI are one of the least commonly reported reasons for not using AI. This shows that existing industry regulations are not playing a strong role in prohibiting adoption in most industries.

Overall, Canadian data suggest that AI adoption continues to grow but at a slower pace than in 2024. In some sectors, AI use is likely to decline, suggesting that some businesses piloted or implemented AI and did not realize sufficient benefits to continue using the technology. While individuals in Canada use generative AI tools more frequently than those in many peer countries, business adoption remains comparatively limited, and most firms report little measurable labour savings from AI use so far.

Although these trends could indicate declining enthusiasm or unmet expectations, they are also consistent with the early stages of adoption for past general-purpose technologies, where experimentation precedes widespread productivity gains.20

Policy Discussion

Policymakers must recognize that maintaining Canada’s competitive standing requires deliberate, differentiated strategies for AI infrastructure, development, and adoption – objectives that demand distinct policy approaches, skill sets, and resources. While the country currently performs respectably in international rankings, declining relative performance across multiple indices signals cause for concern.

There is significant uncertainty in the trend observations within Canada and across countries – from the beginning of 2024 to the present, we have only two observations of business AI adoption and intended use (from the Canadian Survey of Business Conditions). The OECD international comparisons also depend on this singular survey, and the most recent publicly available Canadian data is for 2023. Given the rapid changes to AI adoption and development, and its potential to have significant growth and productivity effects, we have shockingly little information about how AI is being deployed over time. Statistics Canada is ending the Canadian Survey on Business Conditions (CSBC), with the last scheduled release in August 2026. Budget 2025 allocated $25 million over six years and $4.5 million ongoing for Statistics Canada to implement the AI and Technology Measurement Program (TechStat) to support data and insights to measure how AI is used by organizations and understand its impact on Canadians, the labour force, and the economy. There is significant value and need for national statistics agencies to develop a standardized approach to measuring AI adoption to compare rates across countries, sectors, and over time. This expansion in data monitoring of technology adoption will enable social and economic policy research and inform evidence-based policymaking.21

Temporal considerations must also inform policy design. The substantial productivity gains that AI promises remain, by most credible estimates, a decade or more away. Current measurable benefits, while real, derive primarily from infrastructure buildouts and the “replace” phase of technology adoption, wherein AI-enabled tools substitute for existing technologies within established workflows and processes. These incremental improvements, averaging around 17 percent in labour productivity across various studies, represent meaningful but modest gains. The transformative potential of AI will materialize only when industries advance to the “reimagine” phase, restructuring business models, production processes, and organizational architectures around AI capabilities. History suggests this transition requires substantial complementary investments in training, organizational restructuring, and supporting infrastructure – investments whose returns may not register in productivity statistics for years.

Infrastructure and Data

Data centres have near-immediate positive impacts on GDP and employment while they are being built, but then become an input to production themselves, while requiring significant energy inputs and minimal labour.22 This means that government expenditure directed toward AI development and infrastructure faces inherent limitations as an ongoing economic stimulus. This reality does not argue against infrastructure investment, but rather for calibrated expectations about its short and long term macroeconomic impact and for complementary policies that maximize domestic value creation where possible. Infrastructure represents an investment in the foundational inputs that enable AI and future AI development, with direct GDP impacts occurring in the near term. Data centres themselves do not directly contribute to ongoing productivity growth; they increase the total raw computing capacity. How that capacity is used and deployed determines the longer-term growth and productivity impacts.

The government has a significant role to play in developing its own internal capacity for AI development and adoption, as well as ensuring reasonable access to computing and data resources across economic sectors and business sizes (in particular, for non-profit social enterprises and academic research). The latter objective will be achieved through a combination of policies, including direct spending on infrastructure, funding for low-cost and/or targeted borrowing programs to encourage business investment, and accessibility for small, social and non-profit enterprises. In addition, complementary development of open data assets and investment in initiatives that aggregate high-value administrative and public sector data assets while maintaining appropriate privacy and quality controls would enable AI development activities and improve accessibility for smaller- and lower-resource enterprises.23

Since large companies are already making significant investments in AI infrastructure around the globe, government policy should target improving Canada’s attractiveness for that investment and development. For example, in November 2025, the government tabled Bill C-15. It includes accelerated capital cost allowance and expensing measures. The proposed measures are time-limited and include immediate capital cost deductions for certain productivity-enhancing assets.24 In addition to investment, data centres require land, building/permit approvals, and potentially environmental, energy, and Indigenous impact assessments (and the skilled labour to construct them after approvals). Once they are built, they require significant energy resources for ongoing operations.

A comprehensive strategy will require all levels of government to streamline regulatory approvals and, where possible, implement parallel instead of sequential regulatory steps to speed development timelines. Reducing administrative development costs and shortening approval timelines would benefit general economic development, but could take significant time to implement. In the short term, a targeted strategy could involve identifying suitable development sites at the local or regional level and pre-screening them for energy capacity, required environmental impact assessments, and other considerations.

Adoption and Diffusion

Adoption-focused policy, by contrast, may offer marginal immediate benefits but has higher potential to generate productivity growth in the long term.25 The analysis of business intentions to use AI in the future shows that the majority do not think it is currently relevant to their products or services. Further, the limited data available shows AI adoption across industries is slowing and could possibly decline in five industries over the year.

For companies that do think AI could be relevant to their business but do not plan to use AI, the most commonly cited reasons are a lack of knowledge about the technology, concerns about privacy and security, and that the technology is not yet mature enough. Manufacturing and wholesale trade, in particular, show a lower proportion of firms thinking that AI is not relevant to their products/services, and over 10 percent of firms are not planning to use AI due to a lack of skilled labour.

Given the relatively early stage of adoption and development, these barriers provide a target for government policy intervention. In particular, the government has signalled that it plans to update Canada’s privacy legislation for AI, and it should do so sooner rather than later. An updated privacy policy could have the dual benefit of reducing the uncertainty of potential adopters and providing clarity for companies planning to develop AI technologies using Canadian data.26 Similarly, governments have a role to play in increasing AI literacy and knowledge about the technology (see C.D. Howe Institute, forthcoming). Accelerating the adoption and development of AI in Canadian businesses is also supported by various general and specific tax subsidies, including the Scientific Research and Experimental Development (SR&ED) credit and capital cost allowances discussed above. While employee-training costs are generally tax-deductible, there is a need for a more comprehensive skills investment strategy, particularly in content development and availability for mid-career professionals, with a focus on practical applications. Increasing the labour force capable of implementing AI solutions across different industries is a critical complement to infrastructure investment and to encourage broad adoption and process innovation.

The evidence presented in this analysis shows that adoption in Canada generally follows international patterns – larger and younger firms tend to adopt AI at higher rates. This highlights that targeted policies encouraging technology uptake and growth in small and medium-sized enterprises could speed diffusion across sectors. In addition, it appears that engaging in international trade is associated with higher use of AI, though the intended future use is increasing for companies with no international business activities. In November of 2025, Canada, Australia, and India agreed to enter into a trilateral technology and innovation partnership that includes green energy, resilient supply chains for critical minerals, and the development and mass adoption of AI to improve welfare. Canada should continue to collaboratively secure critical input supply chains and enhance international cooperation on AI development and adoption.

The findings in this Commentary suggest that policies promoting trade diversification and encouraging a broader base of Canadian businesses to engage internationally are interrelated with AI adoption goals. Rather than viewing AI policy in isolation, governments should recognize how existing priorities around business growth, trade promotion, and export development can reinforce technology adoption objectives.

Conclusion

The uncertainty about AI’s development trajectory must be balanced with its potential to be revolutionary and a significant source of ever-elusive productivity growth. The country cannot afford to miss an opportunity that may substantially improve living standards and economic competitiveness over time. On the other hand, the uncertainty surrounding AI’s ultimate productivity impact – credible estimates range from less than 1 percent to 15 percent cumulative GDP gains over the next decade – counsels against excessive concentration of public resources. This suggests a multi-pronged strategic approach that includes investment in infrastructure, enhancing research and development policy,27 and updating privacy legislation. In addition, the government needs to rapidly implement the TechStat initiative to improve monitoring of the diffusion of AI and the associated economic effects. Many policies supporting AI infrastructure, development, and adoption are in the works, while data to monitor their effects is highly limited. The evidence base needs to expand to inform comprehensive and coordinated policy.

Policy should continue to pursue a balanced approach: support for infrastructure development and research capacity to maintain Canada’s standing as a credible AI nation, combined with robust efforts to accelerate adoption across sectors and firm sizes. However, governments should resist the temptation to overemphasize AI at the expense of broader economic priorities.28 Government policy can support AI development and diffusion without direct spending by focusing on a supportive regulatory environment based on principles to manage risk and potential harms, and some stimulus in the form of demand-side policies focusing on potential adopters.29 AI represents one important element of Canada’s productivity agenda, but not its entirety. Maintaining perspective on AI’s current limitations and its uncertain trajectory will help ensure that policy balances risks and limited government resources while remaining responsive to emerging evidence.

The author extends gratitude to Peter MacKenzie, Anindya Sen, Daniel Schwanen, Andrew Sharpe, Tingting Zhang, and several anonymous referees for valuable comments and suggestions. The author retains responsibility for any errors and the views expressed.

Appendix:

REFERENCES

Acemoglu, Daron. 2024. “The Simple Macroeconomics of AI.” NBER Working Paper 32487. MIT. https://economics.mit.edu/sites/default/files/2024-04/The%20Simple%20Macroeconomics%20of%20AI.pdf

Acemoglu, D., and P. Restrepo. 2019. “Automation and New Tasks: How Technology Displaces and Reinstates Labor.” Journal of Economic Perspectives 33(2): 3–30.

Agrawal, A., J. Gans, and A. Goldfarb. 2023. “Similarities and Differences in the Adoption of General Purpose Technologies.” NBER Working Paper 30976.

Aldasoro, Inaki, Leonardo Gambacorta, Rozália Pál, Debora Revoltella, Christoph Weiss, and Marcin Wolski. 2026. “How AI Is Affecting Productivity and Jobs in Europe.” https://cepr.org/voxeu/columns/how-ai-affecting-productivity-and-jobs-europe

Aliaga, S. 2025. “Is AI Already Driving U.S. Growth?” J.P. Morgan Asset Management. https://am.jpmorgan.com/us/en/asset-management/adv/insights/market-insights/market-updates/on-the-minds-of-investors/is-ai-already-driving-us-growth/

Andre, C., and P. Gal. 2024. “Reviving Productivity Growth: A Review of Policies.” OECD Economics Department Working Paper 1822. https://www.oecd.org/content/dam/oecd/en/publications/reports/2024/10/reviving-productivity-growth_936a1da3/61244acd-en.pdf

Appel, R., P. McCrory, A. Tamkin, M. McCain, T. Nayloen, and M. Stern. 2025. “The Anthropic Economic Index Report: Uneven Geographic and Enterprise AI Adoption.” Anthropic. https://www.anthropic.com/research/anthropic-economic-index-september-2025-report

Arnon, A. 2025. “The Projected Impact of Generative AI on Future Productivity Growth.” Penn Wharton Budget Model, University of Pennsylvania. https://budgetmodel.wharton.upenn.edu/p/2025-09-08-the-projected-impact-of-generative-ai-on-future-productivity-growth/

Baily, M., E. Brynjolfsson, and A. Korinek. 2023. “Machines of Mind: The Case for an AI-Powered Productivity Boom.” Brookings. https://www.brookings.edu/articles/machines-of-mind-the-casefor-an-ai-powered-productivity-boo

Barnette, C., and M. Peterson. 2025. “Are We in a Bubble? The AI Boom in Context.” BlackRock Advisor Centre. https://www.blackrock.com/us/financial-professionals/insights/ai-tech-bubble

Bureau of Economic Analysis (BEA). 2025. “Gross Domestic Product, 3rd Quarter 2025 (Initial Estimate) and Corporate Profits (Preliminary).” https://www.bea.gov/index.php/news/2025/gross-domestic-product-3rd-quarter-2025-initial-estimate-and-corporate-profits

Bick, A., A. Blandin, and D. Deming. 2024. “The Rapid Adoption of Generative AI.” NBER Working Paper 32966. https://www.nber.org/system/files/working_papers/w32966/w32966.pdf

Blit, J. 2025. Replace, Reimagine, Recombine: Building an AI Nation to Fix Canada’s Productivity Crisis. https://www.uwaterloo.ca/scholar/sites/ca.scholar/files/jblit/files/blit_ai_3rs.pdf

BNP Paribas. 2025. “The AI Value Gap: Boom or Bust?” https://globalmarkets.cib.bnpparibas/the-ai-value-gap-boom-or-bust/

Boussour, L. 2025. “AI-Powered Growth: GenAI as a New Engine of U.S. Economic Performance.” EY Parthenon. November 17. https://www.ey.com/en_us/insights/ai/ai-powered-growth

Bryan, V., S. Sood, and C. Johnston. 2025. “Analysis on Artificial Intelligence Use by Businesses in Canada, Second Quarter of 2025.” Statistics Canada. https://www150.statcan.gc.ca/n1/pub/11-621-m/11-621-m2025008-eng.htm

————. 2025a. “Analysis on Expected Use of Artificial Intelligence by Businesses in Canada, Third Quarter of 2025.” Statistics Canada. https://www150.statcan.gc.ca/n1/pub/11-621-m/11-621-m2025011-eng.htm

Brynjolfsson, Erik, B. Chandar, and R. Chen. 2025. “Canaries in the Coal Mine? Six Facts about the Recent Employment Effects of Artificial Intelligence.” https://digitaleconomy.stanford.edu/app/uploads/2025/11/CanariesintheCoalMine_Nov25.pdf

Brynjolfsson, E., A. Collis, E. Diewert, F. Eggers, and K. Fox. 2023. “GDP-B: Accounting for the Value of New and Free Goods.” https://datascience.stanford.edu/sites/g/files/sbiybj25376/files/media/file/gdp-b-2023-10.pdf

Brynjolfsson, Erik, Daniel Rock, and Chad Syverson. 2021. “The Productivity J-Curve: How Intangibles Complement General Purpose Technologies.” American Economic Journal: Macroeconomics 13(1): 333–372.

Caamaño-Gordillo, D., J. Mula, and R. de la Torre. 2025. “Impact of Generative Artificial Intelligence on Workload, Efficiency and Labour Productivity.” IFAC-PapersOnline 59(10). https://www.sciencedirect.com/science/article/pii/S240589632500998X

Calvino, F., and L. Fontanelli. 2023. “A Portrait of AI Adopters across Countries: Firm Characteristics, Assets’ Complementarities and Productivity.” OECD Science, Technology and Industry Working Paper. https://www.oecd.org/content/dam/oecd/en/publications/reports/2023/04/a-portrait-of-ai-adopters-across-countries_76004dec/0fb79bb9-en.pdf

Challapally, A., C. Pease, R. Raskar, and P. Chari. 2025. “The GenAI Divide: State of AI in Business 2025.” MIT. https://mlq.ai/media/quarterly_decks/v0.1_State_of_AI_in_Business_2025_Report.pdf

Chatterji, Aaron, Thomas Cunningham, David Deming, Zoe Hitzig, Christopher Ong, Carl Yan Shan, and Kevin Wadman. 2025. “How People Use ChatGPT.” NBER Working Paper 34255. September. https://www.nber.org/papers/w34255

Corrado, C., J. Haskel, M. Iommi, C. Jona-Lasinio, and F. Bontadini. 2025. “Data Intangible Capital and Productivity.” In Technology, Productivity, and Economic Growth. NBER Studies in Income and Wealth, Volume 83. https://www.nber.org/books-and-chapters/technology-productivity-and-economic-growth/data-intangible-capital-and-productivity

Coupé, Tom, and Weilun Wu. 2025. “The Impact of Generative AI on Productivity: Results of an Early Meta-Analysis.” Working Papers in Economics 25/09. University of Canterbury, Department of Economics and Finance. https://ideas.repec.org/p/cbt/econwp/25-09.html

Diewert, E., K. Fox, and P. Schreyer. 2018. “The Digital Economy, New Products and Consumer Welfare.” ESCoE Discussion Paper. https://ideas.repec.org/p/nsr/escoed/escoe-dp-2018-16.html

Eloundou, Tyna, Sam Manning, Pamela Mishkin, and Daniel Rock. 2023. “GPTs are GPTs: An Early Look at the Labor Market Impact Potential of Large Language Models.” arXiv.

EY Global. 2025. “Canada Tables Bill with Accelerated Capital Cost Allowance and Other Immediate Expensing Measures.” December 18. https://www.ey.com/en_gl/technical/tax-alerts/canada-tables-bill-with-accelerated-capital-cost-allowance-and-other-immediate-expensing-measures

Filippucci, Francesco et al. 2024. “The Impact of Artificial Intelligence on Productivity, Distribution and Growth: Key Mechanisms, Initial Evidence and Policy Challenges.” OECD Artificial Intelligence Papers No. 15. Paris: OECD.

Filippucci, Francesco, Peter Gal, Katharina Laengle, and Matthias Schief. 2025. “Macroeconomic Productivity Gains from Artificial Intelligence in G7 Economies.” OECD Working Paper. https://www.oecd.org/en/publications/macroeconomic-productivity-gains-from-artificial-intelligence-in-g7-economies_a5319ab5-en.html

Filippucci, F., P. Gal, K. Laengle, M. Schief, and F. Unsal. 2025. “Opportunities and Risks of Artificial Intelligence for Productivity.” International Productivity Monitor 48. https://www.csls.ca/ipm/48/OECD_Final.pdf

First Page Sage. 2025. “ChatGPT Usage Statistics: December 2025.” https://firstpagesage.com/seo-blog/chatgpt-usage-statistics/

Global Center on AI Governance (GCG). 2024. The Global Index on Responsible AI. https://www.global-index.ai/

Goldin, I. et al. 2024. “Why Is Productivity Slowing Down?” Journal of Economic Literature 62(1): 196–268.

Goldman Sachs. 2023. “Generative AI Could Raise Global GDP by 7 Percent.” https://www.goldmansachs.com/intelligence/pages/generative-ai-could-raise-global-gdp-by-7-percent.html

————. 2025. “Why AI Companies May Invest More Than $500 Billion in 2026.” https://www.goldmansachs.com/insights/articles/why-ai-companies-may-invest-more-than-500-billion-in-2026

Greenwood, Jeremy, and Mehmet Yorukoglu. 1997. “1974.” In Carnegie-Rochester Conference Series on Public Policy 46: 49–95.

Gruenwald, P., and S. Panday. 2025. “How Data Centres and AI Are Becoming a New Engine of Growth.” World Economic Forum. December 17. https://www.weforum.org/stories/2025/12/data-centres-and-ai-new-growth-engine/

Harberger, A.C. 1998. “A Vision of the Growth Process.” American Economic Review 88(1): 1–32.

Heneseke, G., R. Davies, A. Felstead, D. Gallie, F. Green, and Y. Zhou. 2025. “How Exposed Are UK Jobs to Generative AI? Developing and Applying a Novel Task-Based Index.” https://arxiv.org/pdf/2507.22748

Howitt, P. 2015. Mushrooms and Yeast: The Implications of Technological Progress for Canada’s Economic Growth. Commentary 433. Toronto: C.D. Howe Institute. https://cdhowe.org/publication/mushrooms-and-yeast-implications-technological-progress-canadas-economic/

Iida, K., S. Rahim, C. Hutber, G. Grau, C. Moss, and P. Douglas. 2025. Government AI Readiness Index 2025. Oxford Insights. https://oxfordinsights.com/wp-content/uploads/2026/01/2025-Government-AI-Readiness-Index-Report_01_26.pdf

Jung, C., and B.S. Desikan. 2024. “Transformed by AI: How Generative Intelligence Could Affect Work in the UK and How to Manage It.” Institute for Public Policy Research. https://ippr-org.files.svdcdn.com/production/Downloads/Transformed_by_AI_March24_2024-03-27-121003_kxis.pdf

Lehdonvirta, V., B. Wu, Z. Hawkins, C. Caira, and L. Russo. 2025. “Measuring Domestic Public Cloud Compute Availability for Artificial Intelligence.” OECD Artificial Intelligence Papers. October. https://www.oecd.org/content/dam/oecd/en/publications/reports/2025/10/measuring-domestic-public-cloud-compute-availability-for-artificial-intelligence_39fa6b0e/8602a322-en.pdf

Levin, A. 2025. “The AI Investment Boom Continues to Drive GDP Growth.” Barron’s. https://www.barrons.com/articles/duolingo-stock-price-ai-picks-2026-08cab284

Maslej, Nestor, Loredana Fattorini, Raymond Perrault, Yolanda Gil, Vanessa Parli, Njenga Kariuki, Emily Capstick, Anka Reuel, Erik Brynjolfsson, John Etchemendy, Katrina Ligett, Terah Lyons, James Manyika, Juan Carlos Niebles, Yoav Shoham, Russell Wald, Tobi Walsh, Armin Hamrah, Lapo Santarlasci, Julia Betts Lotufo, Alexandra Rome, Andrew Shi, and Sukrut Oak. 2025. The AI Index 2025 Annual Report. Stanford, CA: Institute for Human-Centered AI, Stanford University. April.

McElheran, K. et al. 2023. “AI Adoption in America: Who, What, and Where.” NBER Working Paper 31788.

Narayanan, Arvind, and Sayesh Kapoor. 2024. AI Snake Oil: What Artificial Intelligence Can Do, What It Can’t and How to Tell the Difference. Princeton: Princeton University Press. https://press.princeton.edu/books/ebook/9780691249643/ai-snake-oil-pdf

OECD. 2024. “Recommendations of the Council on Artificial Intelligence.” OECD Legal Instruments. https://legalinstruments.oecd.org/en/instruments/OECD-LEGAL-0449

————. 2025. “AI Adoption by Small and Medium-Sized Enterprises.” OECD Discussion Paper for the G7. https://www.oecd.org/content/dam/oecd/en/publications/reports/2025/12/ai-adoption-by-small-and-medium-sized-enterprises_9c48eae6/426399c1-en.pdf

OECD.AI. 2026a. Data from Preqin, last updated January 5, 2026. https://oecd.ai/

————. 2026b. Data from OpenAlex, last updated September 30, 2025. https://oecd.ai/

OECD/BCG/INSEAD. 2025. The Adoption of Artificial Intelligence in Firms: New Evidence for Policymaking. Paris: OECD Publishing. https://doi.org/10.1787/f9ef33c3-en

OECD. 2026. ICT Access and Usage Database. January.

Oschinski, M., and R. Wyonch. 2017. Future Shock? The Impact of Automation on Canada’s Labour Market. Commentary 472. Toronto: C.D. Howe Institute. https://cdhowe.org/wp-content/uploads/2024/12/Update_Commentary2047220web.pdf

Paoli, N. 2025. “Construction Workers Are Earning Up to 30% More and Some Are Nabbing Six-Figure Salaries in the Data Center Boom.” Fortune. December 5. https://fortune.com/2025/12/05/construction-workers-earning-six-figure-salaries-data-center-boom-ai-tech/

Rogerson, A., E. Hankins, P.F. Nettel, and S. Rahim. 2022. Government AI Readiness Index 2022. Oxford: Oxford Insights. https://oxfordinsights.com/wp-content/uploads/2023/11/Government_AI_Readiness_2022_FV.pdf

Schumpeter, J. 1942. Capitalism, Socialism, and Democracy. New York: Harper and Brothers.

Sen, A. 2026. The Missing Pillar of Canada’s AI Strategy: Data Supply Chains. Commentary 706. Toronto: C.D. Howe Institute. https://cdhowe.org/publication/the-missing-pillar-of-canadas-ai-strategy-data-supply-chains/

Singh, Pia. 2026. “AI Spending Wasn’t the Biggest Engine of U.S. Economic Growth in 2025, Despite Popular Assumptions.” CNBC. https://www.cnbc.com/2026/01/26/ai-wasnt-the-biggest-engine-of-us-gdp-growth-in-2025.html

Smith, Brad. 2025. “Microsoft Deepens Its Commitment to Canada with Landmark $19B AI Investment.” Microsoft. https://blogs.microsoft.com/on-the-issues/2025/12/09/microsoft-deepens-its-commitment-to-canada-with-landmark-19b-ai-investment/

Top500. 2025. “November 2025.” https://top500.org/lists/top500/2025/11/

Uren, V., and J.S. Edwards. 2023. “Technology Readiness and the Organizational Journey towards AI Adoption: An Empirical Study.” International Journal of Information Management 68. https://www.sciencedirect.com/science/article/pii/S0268401222001220

White, J., and S. Cesareo. 2024. The Global Artificial Intelligence Index 2025. https://www.tortoisemedia.com/data/global-ai#rankings

World Bank. 2024. “Population Ages 15–64.” https://data.worldbank.org/indicator/SP.POP.1564.TO

Yotzov, Ivan, Jose Maria Barrero, Nicholas Bloom, Philip Bunn, Steven J. Davis, Kevin M. Foster, Aaron Jalca, Brent H. Meyer, Paul Mizen, Michael A. Navarrete, Pawel Smietanka, Gregory Thwaites, and Ben Zhe Wang. 2026. “Firm Data on AI.” NBER Working Paper 34836. https://www.nber.org/papers/w34836

Avril 2026 – Les investissements en IA ont atteint plusieurs centaines de milliards, mais les retombées économiques se font encore attendre. Malgré des investissements rapides et une expérimentation répandue des outils d’IA générative, les gains de productivité mesurables ne se reflètent pas encore clairement dans les données agrégées des économies avancées, ce qui soulève une question importante : qui sera le premier à transformer cette dynamique en gains tangibles?

Dans « From Hype to Output: How AI Investment Translates to Real Productivity Gains » (De l’engouement à la production : comment les investissements dans l’IA se traduisent par de réels gains de productivité), l’auteure Rosalie Wyonch constate que si l’IA offre une voie vers de meilleures performances économiques et pourrait contribuer à relever les défis de productivité auxquels le Canada est confronté depuis longtemps, la concrétisation de ces gains nécessitera des politiques favorisant son adoption et sa diffusion par les entreprises, lesquelles restent inégales dans la plupart des secteurs et des régions. Cela implique notamment de supprimer les obstacles, de garantir l’accès aux données et d’encourager l’intégration de l’IA au sein des entreprises et par tous les secteurs.

Toronto Based AI Strategist: AI Is Rewriting Executive Decision Making

AI is fundamentally redefining leadership by providing new tools, frameworks, and systems that allow leaders not just to manage complexity, but to see, challenge, and reshape their organizations in ways never before possible. The competitive mandate for leaders is clear: harness AI not merely for efficiency, but as an engine for deeper self-awareness, structured dissent, and proactive sensing that unlocks true organizational agility and resilience.

Strategic Frameworks for Next-Gen AI Leadership

Forward-thinking leaders are moving beyond pilot projects and isolated automation to experiment with new, holistic approaches—many inspired by concepts like the Leadership Mirror, Red-Team Loop, and Organization Pulse Monitor. These paradigms operationalize AI in ways that directly address the perennial blind spots, biases, and inertia that often undermine executive decision-making.

George Yang- helping organizations and executives embrace AI.

The Leadership Mirror: Cultivating Radical Self-Awareness

The Leadership Mirror uses AI to continuously analyze leadership communication, decision rationale, and team interactions, surfacing insights that are often overlooked or difficult for humans to acknowledge. For example, Microsoft has begun leveraging AI tools to track who dominates meetings, which voices get systematically dismissed, and when evidence is overridden by intuition—creating dashboards that encourage leaders to confront uncomfortable patterns.

  • This approach helps leaders challenge their own narrative, improve inclusiveness, and drive more thoughtful debate.
  • With AI’s ability to process language in real time, leaders can receive feedback loops and “reflections” that support a culture of deliberate, transparent leadership.
  • The Leadership Mirror is also a vehicle for mitigating the “competence penalty,” where women and older workers face skepticism for using AI—even when it enhances productivity. By surfacing evidence of expertise and impact, it reduces bias and builds psychological safety.

There are different types of AI including less sophisticated models such as Generative AI. To decide whether to use generative artificial intelligence for a task, ask yourself whether it matters if the output is true and you have the expertise to verify the tool’s output. (Adapted from Aleksandr Tiulkanov‘s LinkedIn post)

The Red-Team Loop: Embedding Structured Dissent

To counter groupthink and executive overconfidence, Red-Team Loop systems employ AI to automate adversarial reviews of strategy and operational decisions. Verizon, for instance, uses an AI framework that captures assumptions, risks, and anticipated outcomes for major decisions, then generates simulated critiques and alternative scenarios—sometimes challenging senior executives on blind spots they themselves hadn’t recognized.

  • By proactively “red-teaming” their own decisions, leaders foster a culture where dissent is routine, rational, and data-driven—not ad hoc or punitive.
  • The approach is especially valuable in M&A, crisis management, and product launches, where high-stakes, high-ambiguity decisions benefit from rigorous challenge.
  • Leading boards now expect Red-Team Loops as part of their fiduciary duty, recognizing that the cost of missed risks is measured not just in dollars, but reputation and long-term viability.

Organization Pulse Monitor: Proactive Sensing for Culture and Risk

The Organization Pulse Monitor uses AI to detect weak signals in organization culture, ethical risk, and operational friction long before traditional metrics or surveys would register them. Some organizations have begun linking AI-powered sentiment analysis of internal communications, workflow behaviors, and network interactions to predict where a culture may be straining, where compliance risks are emerging, or where silent dissent is brewing.

  • When Pulse Monitors flagged drops in engagement and early warning signs of burnout, one multinational fast-tracked well-being interventions, pre-empting attrition.
  • AI-driven pulse scans also help surface ethical risks—such as exclusionary behaviors or data privacy concerns—enabling leaders to respond immediately, not months later.

Actionable Strategies: Bringing AI Experiments to Leadership

How can senior leaders experiment and innovate with these systems while maximizing value and minimizing risk?

  • Map Adoption Hotspots and Blind Spots: Use mirror and pulse data to identify where AI is catalyzing positive behaviors—and where competence penalties or shadow AI usage may be undermining equity or performance. Target interventions accordingly.
  • Mobilize Role Model Leaders: Encourage respected senior leaders, particularly those from underrepresented demographics, to visibly experiment with and champion AI tools. Research shows that when these role models use AI openly, adoption gaps shrink, and psychological safety rises.
  • Redesign Evaluation and Disclosure Policies: Shift performance metrics from subjective ratings of proficiency to objective impact, cycle time, accuracy, and innovation. Blind reviews and private feedback mechanisms can reduce bias against AI users and drive fairer rewards.
  • Embed Structured Red-Teaming in Decision Flows: Institutionalize adversarial testing of key decisions, making AI-enabled dissent a standard step—not a threat or afterthought. Leaders should receive regular “contrarian” insights, not just consensus-building reports.

Common Pitfalls and Human Impact

Despite rising investment, less than one-third of US employers believe staff are equipped for critical thinking in the AI era, and only 16% of American workers use AI on the job despite widespread availability. The main barriers are not just technical, but social: competence penalties, fear of reputation loss, and resistance among influential skeptics.

  • Competence Penalty: AI users, especially women and older employees, may face a perception of diminished competence. This undermines adoption and can exacerbate workplace inequality.
  • Shadow AI and Hidden Risks: Employees sometimes use unauthorized tools to bypass bias, exposing the organization to compliance, reputational, and security risk.
  • Skill Gaps vs. Work Context: Traditional training falls short without tailored, role-specific feedback loops—AI tutors offer scalable, personal learning but must be embedded in daily workflow, not delivered in isolation.

Governance, Ethics, and Sustainable Change

Human-centered leadership isn’t optional—it’s a strategic imperative. Boards and executives must be proactive in:

  • Instituting transparent governance for all AI systems (mirrors, loops, monitors), with clear oversight on privacy, fairness, and impact.
  • Ensuring structured role-modeling and psychological safety—particularly for vulnerable groups confronting competence penalties.
  • Making change management a continuous process, with AI as both coach and sentinel, not just a dashboard.

The call to action for C-suite leaders is urgent and profound: treat responsible, experimental, and self-critical AI adoption as the core discipline of next-generation leadership. Not just for efficiency, but for building organizations where insight, challenge, and well-being are sustainably enabled. Those who master the trifecta of mirror, loop, and pulse will set the new standard for profitable, human-centered growth in the age of AI.

More about:

George Yang is a Toronto-based digital innovator and AI adoption strategist with over 15 years of experience in marketing and digital transformation. As Chair of the AI Working Group at the National Payroll Institute, he helps organizations translate AI strategy into measurable business outcomes. George is passionate about making AI adoption ethical, practical, and impactful, bridging the gap between innovation and implementation across industries. georgeyang.ca

Would You Use AI For Buying A Car? One In Four Buyers Already Do

A recent consumer survey backed by similar results from Elon University reveals that AI adoption for car shopping is skyrocketing, rapidly becoming a standard part of the automobile buying process. This as fully one in four buyers have already used AI tools this year to research, compare prices, negotiate and otherwise outsmart dealerships, and an overwhelming 88% found it helpful. Signaling a seismic shift in the way North Americans are now shopping for cars, nearly half of consumers indicated plans to use AI in their next purchase. Not just for buyer benefits, dealerships are gleaning critical business intelligence from AI to inform sales strategies, train staff and elevate customer engagement. The below  report from our friends at CarEdge, which offers its own AI Negotiator car buying tool saving shoppers thousands, details the first data-backed look at how AI tools are reshaping the car buying experience.

Mornine- AI powered car dealership robot.

Study: 1 in 4 Car Buyers Tap AI for Better Deals


Artificial intelligence is changing the way North Americans buy cars, and it’s a transition that is happening quickly. In the first-ever survey of its kind, CarEdge asked 500 car shoppers if they’re using AI tools like ChatGPT to research, compare, and negotiate during the car buying process. The results confirm a major shift is underway. One in four car buyers in 2025 are already using AI tools to gain an edge, and future buyers are even more likely to embrace these technologies.

Car buyers are finding AI to be a valuable tool. Among those who used tools like ChatGPT, Perplexity, Google Gemini, and others, 88% said it was helpful. AI is quickly becoming a trusted co-pilot for car buyers.

Key Findings: Car Buying Is Changing

The 2025 CarEdge AI & Car Buying Survey reveals a clear and growing trend: AI tools are quickly becoming part of the car buying process for a significant portion of consumers. Here are the standout findings:

1 in 4 Car Buyers Use AI 

25% of car buyers in 2025 say they used or plan to use AI tools like ChatGPT during the shopping or buying process. This contrasts with a recent survey by Elon University that found 52% of Americans now use AI large language models. While signs point towards increased adoption of AI tools, the CarEdge survey found that most car buyers are still in the early stages of integrating these tools into high-stakes decisions like vehicle purchases. This suggests there’s still significant room for growth in AI adoption amongst car buyers.

AI Use Is Accelerating

Among those who haven’t bought a car yet this year, 40% say they are using or plan to use AI tools during their search or deal-making. This is nearly 3x higher than the 14% seen among those who already bought a car earlier in the year.

AI Tools Deliver Results

Among those who used AI:

  • 88% say the tools were helpful
  • 32% found them very helpful
  • 60% used them “a lot” during the process

The AI Holdouts: Drivers Who Lease

Of the respondents who had already leased a car in 2025, none reported using any AI tools.

The AI-Adopting Buyer: Who’s Using It, and How?

AI adoption among car buyers is still in its early stages, but clear trends are beginning to emerge.

Among Buyers Who Already Purchased in 2025:

Just 14% of those who already bought a vehicle this year used AI tools during the process. Adoption rates were nearly identical across new and used buyers, with 14% in each group saying they used AI tools.

Among Future Car Buyers:

The numbers jump significantly when looking at those who haven’t yet bought in 2025. Among this group — who represent 39% of total respondents — 40% say they either already use or plan to use AI tools during their car search and buying process.

That’s more than triple the current usage rate among recent buyers, suggesting AI adoption is accelerating as awareness grows and tools become easier to use.

This group also appears to be more proactive: 60% of those who used AI tools during their buying journey said they used them “a lot,” while 40% used them only occasionally.

What Car Buyers Are Using AI Tools

AI tools are quickly becoming essential research companions for car shoppers looking to make more informed, confident decisions. After all, why go it alone when a wealth of automotive knowledge powered by large language models (LLMs) is right in your pocket?

Among buyers who used AI tools during their car purchase or lease process, here’s how they put them to work:

88% — Researching Vehicles

The most common use by far, AI tools helped buyers learn about different models, trims, features, and reliability. For many, it was like having an always-available expert to explain the pros and cons of their options.

64% — Comparing Prices and Market Values

Buyers used AI to better understand fair pricing, from invoice pricing to out-the-door. 

44% — Learning Negotiation Strategies

Nearly half of AI users leaned on these tools to prepare for conversations with salespeople. Whether role-playing negotiation scenarios or asking how to spot add-on fees, this group used AI to level the playing field at the dealership.

11% — Exploring Finance and Lease Options

A much smaller portion of buyers used these tools to become familiar with leasing vs. financing, how to calculate payments, and similar queries.

Industry Implications

Car buying has always been tilted in favor of the dealership. Information asymmetry — what the dealer knows versus what the customer knows — has long been the source of consumer frustration, confusion, and overpayment.

That dynamic is beginning to shift.

This survey confirms what many in the industry are only starting to realize: AI is giving car buyers the upper hand. Tools like ChatGPT are helping consumers cut through the noise, ask smarter questions, and avoid common dealership traps. Instead of relying on guesswork or scattered advice, buyers are turning to AI for fast, personalized guidance at every step.

But one auto industry veteran has words of caution for buyers relying heavily on AI tools.

It’s both surprising and a little scary to see how quickly people are turning to AI to guide such a major financial decision,” said Ray Shefska, Co-Founder of CarEdge. “While tools like ChatGPT can be powerful, they’re only as good as the data behind them. AI should complement your research, not replace your own critical thinking.

That perspective underscores the real takeaway of this report: AI works best when it’s used thoughtfully as a tool, not as a crutch. In an age where automation raises fears of job loss or decision-making without human oversight, this survey offers a more optimistic view — one where technology helps everyday consumers make smarter choices. Used wisely, AI can help level the playing field and bring more transparency and fairness to the car buying experience.

Methodology

This survey was conducted by CarEdge between June 19 and June 24, 2025. A total of 500 U.S. respondents participated, recruited through the CarEdge email newsletter and social media channels. Questions were tailored based on buying status to better understand how and when AI tools were used in the car shopping process.

For the Silo, Karen Hayhurst.

About CarEdge
Founded in 2019 by father-and-son team Ray and Zach Shefska, CarEdge is a leading platform dedicated to empowering car shoppers with free expert advice, in-depth market insights, and tools to navigate every step of the car-buying journey. From researching vehicles to negotiating deals, CarEdge helps consumers save money, time, and hassle, hundreds of thousands of happy consumers have used CarEdge to buy their car with confidence. With trusted resources like the CarEdge AI Negotiator tool, Research Center, Vehicle Rankings and Reviews, and hundreds of guides on YouTube, CarEdge is redefining transparency and fairness in the automotive industry. Follow them on YouTubeTikTokX,  Facebook, and Instagram for actionable car-buying tips and market insights. Learn more at www.CarEdge.com.

A Pathway To Trusted AI

Artificial Intelligence (AI) has infiltrated our lives for decades, but since the public launch of ChatGPT showcasing generative AI in 2022, society has faced unprecedented technological evolution. 

With digital technology already a constant part of our lives, AI has the potential to alter the way we live, work, and play – but exponentially faster than conventional computers have. With AI comes staggering possibilities for both advancement and threat.

The AI industry creates unique and dangerous opportunities and challenges. AI can do amazing things humans can’t, but in many situations, referred to as the black box problem, experts cannot explain why particular decisions or sources of information are created. These outcomes can, sometimes, be inaccurate because of flawed data, bad decisions or infamous AI hallucinations. There is little regulation or guidance in software and effectively no regulations or guidelines in AI.

How do researchers find a way to build and deploy valuable, trusted AI when there are so many concerns about the technology’s reliability, accuracy and security?

That was the subject of a recent C.D. Howe Institute conference. In my keynote address, I commented that it all comes down to software. Software is already deeply intertwined in our lives, from health, banking, and communications to transportation and entertainment. Along with its benefits, there is huge potential for the disruption and tampering of societal structures: Power grids, airports, hospital systems, private data, trusted sources of information, and more.  

Consumers might not incur great consequences if a shopping application goes awry, but our transportation, financial or medical transactions demand rock-solid technology.

The good news is that experts have the knowledge and expertise to build reliable, secure, high-quality software, as demonstrated across Class A medical devices, airplanes, surgical robots, and more. The bad news is this is rarely standard practice. 

As a society, we have often tolerated compromised software for the sake of convenience. We trade privacy, security, and reliability for ease of use and corporate profitability. We have come to view software crashes, identity theft, cybersecurity breaches and the spread of misinformation as everyday occurrences. We are so used to these trade-offs with software that most users don’t even realize that reliable, secure solutions are possible.

With the expected potential of AI, creating trusted technology becomes ever more crucial. Allowing unverifiable AI in our frameworks is akin to building skyscrapers on silt. Security and functionality by design trump whack-a-mole retrofitting. Data must be accurate, protected, and used in the way it’s intended.

Striking a balance between security, quality, functionality, and profit is a complex dance. The BlackBerry phone, for example, set a standard for secure, trusted devices. Data was kept private, activities and information were secure, and operations were never hacked. Devices were used and trusted by prime ministers, CEOs and presidents worldwide. The security features it pioneered live on and are widely used in the devices that outcompeted Blackberry. 

Innovators have the know-how and expertise to create quality products. But often the drive for profits takes precedence over painstaking design. In the AI universe, however, where issues of data privacy, inaccuracies, generation of harmful content and exposure of vulnerabilities have far-reaching effects, trust is easily lost.

So, how do we build and maintain trust? Educating end-users and leaders is an excellent place to start. They need to be informed enough to demand better, and corporations need to strike a balance between caution and innovation.

Companies can build trust through a strong adherence to safe software practices, education in AI evolution and adherence to evolving regulations. Governments and corporate leaders can keep abreast of how other organizations and countries are enacting policies that support technological evolution, institute accreditation, and financial incentives that support best practices. Across the globe, countries and regions are already developing strategies and laws to encourage responsible use of AI. 

Recent years have seen the creation of codes of conduct and regulatory initiatives such as:

  • Canada’s Voluntary Code of Conduct on the Responsible Development and Management of Advanced Generative AI Systems, September 2023, signed by AI powerhouses such as the Vector Institute, Mila-Quebec Artificial Intelligence Institute and the Alberta Machine Intelligence Institute;
  • The Bletchley Declaration, Nov. 2023, an international agreement to cooperate on the development of safe AI, has been signed by 28 countries;
  • US President Biden’s 2023 executive order on the safe, secure and trustworthy development and use of AI; and
  • Governing AI for Humanity, UN Advisory Body Report, September 2024.

We have the expertise to build solid foundations for AI. It’s now up to leaders and corporations to ensure that much-needed practices, guidelines, policies and regulations are in place and followed. It is also up to end-users to demand quality and accountability. 

Now is the time to take steps to mitigate AI’s potential perils so we can build the trust that is needed to harness AI’s extraordinary potential. For the Silo, Charles Eagan. Charles Eagan is the former CTO of Blackberry and a technical advisor to AIE Inc.

AI Tinkerers Take Note -Effective Prompting Can Build Actual Products

Hello AI Tinkerers and welcome to the latest Sci-Tech article here at The Silo. Get ready, You will want to pay attention because the spotlight is on this Dude because he knows how to get around ‘bad ai prompting’. Just recently, he has helped spin out 40 startups using one core skill. Can you guess which one? Yep. Prompting.

In the One-Shot video below, Kevin Leneway breaks down his real workflow for shipping AI products fast — using markdown checklists, agent coding, rubric-based UI design, and zero Figma.

“I don’t need Figma. I just prompt my way to a working front end.” — Kevin Leneway

While most people are still asking ChatGPT to write code snippets, Kevin is building full-stack products using nothing but prompts. In this One-Shot episode, he reveals the exact system he’s used to launch over 40 startups at Pioneer Square Labs. We break down:

  • How he writes BRDs and PRDs that don’t suck
  • Why vibe coding fails and how to actually use AI agents
  • The markdown checklist that replaces a product team
  • How to go from idea to working app with zero context switching
  • His open-source starter kit that makes Cursor and Claude 3.5 feel like magic

“I’ve helped launch six startups including Singlefile (singlefile.io, $24M raised), Recurrent (recurrentauto.com, $24M raised), Joon (joon.com, $9.5M raised), Gradient (gradient.io, $3.5M raised), Genba (genba.ai, acquired May 2022) and Enzzo (enzzo.ai, $3M raised).”

If you’re a builder, this will change how you work. No gimmicks. Just a ruthless focus on speed, clarity, and shipping. Watch now. Learn the system. Steal it. For the Silo, Joe at aitinkerers.org

Featured image- DALL·E robot dressed like shakespeare – AllAboutLean.com.

AI Aggregates, But Dyslexia Innovates

The rise of AI is truly remarkable. It is transforming the way we work, live, and interact with each other, and with so many other touchpoints of our lives. However, while AI aggregates, dyslexic thinking skills innovate. If used in the right way, AI could be the perfect co-pilot for dyslexics to really move the world forward. In light of this, Virgin and Made By Dyslexia have launched a brilliant campaign to show what is possible if AI and dyslexic thinking come together. The film below says it all.

As the film shows, AI can’t replace the soft skills that index high in dyslexics – such as innovating, lateral thinking, complex problem solving, and communicating.

If you ask AI for advice on how to scale a brand that has a record company – it offers valuable insights, but the solution lacks creative instinct and spontaneous decision making. If I hadn’t relied on my intuition, lateral thinking and willingness to take a risk, I would have never jumped from scaling a record company to launching an airline – which was a move that scaled Virgin into the brand it is today.

Together, dyslexic thinkers and AI are an unstoppable force, so it’s great to see that 72% of dyslexics see AI tools (like ChatGPT) as a vital starting point for their projects and ideas – according to new research by Made By Dyslexia and Randstad Enterprise. With help from AI, dyslexics have limitless power to change the world, but we need everyone to welcome our dyslexic minds. If businesses fail to do this, they risk being left behind. As the Value of Dyslexia report highlighted, dyslexic skillsets will mirror the World Economic Forum’s future skills needs by end of this year (2025). Given the speed at which technology and AI have progressed, this cross-over has arrived two years earlier than predicted.

Image: Sarah Rogers/MITTR

With all of this in mind, it’s concerning to see a big difference between how HR departments think they understand and support dyslexia in the workplace, versus the experience of dyslexic people themselves.

 The new research also shows that 66% of HR professionals believe they have support structures in place for dyslexia, yet only 16% of dyslexics feel supported in the workplace. It’s even sadder to see that only 14% of dyslexic employees believe their workplace understands the value of dyslexic thinking. There is clearly work to be done here.

To empower dyslexic thinking in the workplace (which has the two-fold benefit of bringing out the best in your people and in your business), you need to understand dyslexic thinking skills. To help with this, Made By Dyslexia is launching a workplace training course later this year on LinkedIn Learning – and you can sign up for it now. The course will be free to access, and I’m delighted that Virgin companies from all across the world have signed up for it – from Virgin Australia, to Virgin Active Singapore, to Virgin Plus Canada and Virgin Voyages. It’s such an insightful course, designed by experts at Made By Dyslexia to educate people on how to understand, support, and empower dyslexic thinking in the workplace, and make sure businesses are ready for the future.

It’s always inspiring to see how Made By Dyslexia empowers dyslexics, and shows the world the limitless power of dyslexic thinking. If businesses can harness this power, and if dyslexics can harness the power of AI – we can really drive the future forward.  Richard Branson, Founder at Virgin Group.

OPED: Made by Human: The Threat of Artificial Intelligence on Human Labor

This Article is 95.6% Made by Human / 4.4% by Artificial Intelligence

One of the most concerning uncertainties surrounding the emergence of artificial intelligence is the impact on human jobs.

100% Satisfaction Guarantee

Let us start with a specific example – the customer support specialist. This is a human-facing role. The primary objective of a Customer Support Specialist is to ensure customer satisfaction.

The Gradual Extinction of Customer Support Roles

Within the past decade or so, several milestone transformations have influenced the decline of customer support specialists. Automated responses for customer support telephone lines. Globalization. And chat-bots. 

Chat-bots evolved with the human input of information to service clients. SaaS-based products soon engineered fancy pop-ups for everyone. Just look at Uber if you want a solid case-study – getting through to a person is like trying to contact the King of Thailand. 

The introduction of new artificial intelligence for customer support solutions will make chat-bots look like an AM/FM frequency radio at the antique market. 

The Raging Battle: A Salute to Those on the Front Lines

There are a handful of professions waging a battle against the ominous presence of artificial intelligence. This is a new frontier – not only for technology, but for legal precedent and our appetite for consumption. 

OpenAI is serving our appetite in two fundamental ways: text-based content (i.e. ChatGPT) and visual-based content (i.e. DALL·E). How we consume this content boils down to our own taste-buds, perceptions and individual needs. It is all very human-driven, and it is our degrees of palpable fulfillment that will ultimately dictate how far this penetrates the fate of other professions. 

Sarah Silverman, writer, comedian and actress sued the ChatGPT developer OpenAI and Mark Zuckerberg’s Meta for copyright infringement. 

We need a way to leave a human mark. Literally, a Made by Human insignia that traces origins of our labor, like certifying products as “organic”.

If we’re building the weapon that threatens our very livelihood, we can engineer the solution that safeguards it. 

The Ouroboros Effect

If we seek retribution for labor and the preservation of human work, we need to remain ahead of innovation. There are several action-items that may safeguard human interests:

  • Consolidation of Interest. Concentration of efforts within formal structures or establish new ones tailored to this subject;
  • Litigation. Swift legal action based on existing laws to remedy breaches and establish legal precedents for future litigation;
  • Technological Innovation. Cutting-edge technology that: (a) engineers firewalls for preventing AI scraping technologies; (b) analyzes human work products; and (c) permits tracking of intellectual property.
  • Regulatory Oversight. Formation of a robust framework for monitoring, enforcing and balancing critical issues arising from artificial intelligence. United Nations, but without the thick, glacial layers of bureaucracy.  

These front-line professionals are just the first wave – yet if this front falls, it will be a fatal blow to intellectual property rights. We will have denied ourselves the ideological shields and weapons needed to preserve and protect origins of human creativity

At present, the influence of artificial intelligence on labor markets is in our own hands. If you think this is circular reasoning, like some ouroboros, you would be correct. The very nature of artificial intelligence relies on humans.

Ouroboros expresses the unity of all things, material and spiritual, which never disappear but perpetually change form in an eternal cycle of destruction and re-creation.

Equitable Remuneration 

Human productivity will continue to blend with artificial intelligence. We need to account for what is of human origin versus what has been interwoven with artificial intelligence. Like royalties for streaming music, with the notes of your original melody plucked-out. Even if it’s mashed-up, Mixed by Berry and sold overseas. 

These are complex quantum-powered algorithms. The technology exists. It is along the same lines of code that is empowering artificial intelligence. Consider a brief example: 

A 16-year old boy named Olu decides to write a book about growing-up in a war torn nation. 

 Congratulations on your work, Olu! 

47.893% Human /  52.107% Artificial

Meanwhile, back in London, a 57-year old historian named Elizabeth receives an email:

 Congratulations Elizabeth, your work has been recycled! 

34.546% of your writing on the civil war torn nation has been used in an upcoming book publication. Click here to learn more.

We need a framework that preserves and protects sweat-of-the-brow labor. 

As those on the front-line know: Progress begets progress while flying under the banner of innovation. If we’re going to spill blood to save our income streams – from content writers and hand models to lawyers and software engineers – the fruit of our labor cannot be genetically modified without equitable remuneration. 

Feds False News Checker Tool To Use AI- At Risk Of Language & Political Bias

Ottawa-Funded Misinformation Detection Tool to Rely on Artificial Intelligence

Ottawa-Funded Misinformation Detection Tool to Rely on Artificial Intelligence
Canadian Heritage Minister Pascale St-Onge speaks to reporters on Parliament Hill after Bell Media announces job cuts, in Ottawa on Feb. 8, 2024. (The Canadian Press/Patrick Doyle)

A new federally funded tool being developed with the aim of helping Canadians detect online misinformation will rely on artificial intelligence (AI), Ottawa has announced.

Heritage Minister Pascale St-Onge said on July 29 that Ottawa is providing almost $300,000 cad to researchers at Université de Montréal (UdeM) to develop the tool.

“Polls confirm that most Canadians are very concerned about the rise of mis- and disinformation,” St-Onge wrote on social media. “We’re fighting for Canadians to get the facts” by supporting the university’s independent project, she added.

Canadian Heritage says the project will develop a website and web browser extension dedicated to detecting misinformation.

The department says the project will use large AI language models capable of detecting misinformation across different languages in various formats such as text or video, and contained within different sources of information.

“This technology will help implement effective behavioral nudges to mitigate the proliferation of ‘fake news’ stories in online communities,” says Canadian Heritage.

Related-

OpenAI, Google DeepMind Employees Warn of ‘Serious Risks’ Posed by AI Technology

OpenAI, Google DeepMind Employees Warn of ‘Serious Risks’ Posed by AI Technology

With the browser extension, users will be notified if they come across potential misinformation, which the department says will reduce the likelihood of the content being shared.

Project lead and UdeM professor Jean-François Godbout said in an email that the tool will rely mostly on AI-based systems such as OpenAI’s ChatGPT.

“The system uses mostly a large language model, such as ChatGPT, to verify the validity of a proposition or a statement by relying on its corpus (the data which served for its training),” Godbout wrote in French.

The political science professor added the system will also be able to consult “distinct and reliable external sources.” After considering all the information, the system will produce an evaluation to determine whether the content is true or false, he said, while qualifying its degree of certainty.

Godbout said the reasoning for the decision will be provided to the user, along with the references that were relied upon, and that in some cases the system could say there’s insufficient information to make a judgment.

Asked about concerns that the detection model could be tainted by AI shortcomings such as bias, Godbout said his previous research has demonstrated his sources are “not significantly ideologically biased.”

“That said, our system should rely on a variety of sources, and we continue to explore working with diversified and balanced sources,” he said. “We realize that generative AI models have their limits, but we believe they can be used to help Canadians obtain better information.”

The professor said that the fundamental research behind the project was conducted before receiving the federal grant, which only supports the development of a web application.

Bias Concerns

The reliance on AI to determine what is true or false could have some pitfalls, with large language models being criticized for having political biases.

Such concerns about the neutrality of AI have been raised by billionaire Elon Musk, who owns X and its AI chatbot Grok.

British and Brazilian researchers from the University of East Anglia published a study in January that sought to measure ChatGPT’s political bias.

“We find robust evidence that ChatGPT presents a significant and systematic political bias toward the Democrats in the US, Lula in Brazil, and the Labour Party in the UK,” they wrote. Researchers said there are real concerns that ChatGPT and other large language models in general can “extend or even amplify the existing challenges involving political processes posed by the Internet and social media.”

OpenAI says ChatGPT is “not free from biases and stereotypes, so users and educators should carefully review its content.”

Misinformation and Disinformation

The federal government’s initiatives to tackle misinformation and disinformation have been multifaceted.

The funds provided to the Université de Montréal are part of a larger program to shape online information, the Digital Citizen Initiative. The program supports researchers and civil society organizations that promote a “healthy information ecosystem,” according to Canadian Heritage.

The Liberal government has also passed major bills, such as C-11 and C-18, which impact the information environment.

Bill C-11 has revamped the Broadcasting Act, creating rules for the production and discoverability of Canadian content and giving increased regulatory powers to the CRTC over online content.

Bill C-18 created the obligation for large online platforms to share revenues with news organizations for the display of links. This legislation was promoted by then-Heritage Minister Pablo Rodriguez as a tool to strengthen news media in a “time of greater mistrust and disinformation.”

These two pieces of legislation were followed by Bill C-63 in February to enact the Online Harms Act. Along with seeking to better protect children online, it would create steep penalties for saying things deemed hateful on the web.

There is some confusion about what the latest initiative with UdeM specifically targets. Canadian Heritage says the project aims to counter misinformation, whereas the university says it’s aimed at disinformation. The two concepts are often used in the same sentence when officials signal an intent to crack down on content they deem inappropriate, but a key characteristic distinguishes the two.

The Canadian Centre for Cyber Security defines misinformation as “false information that is not intended to cause harm”—which means it could have been posted inadvertently.

Meanwhile, the Centre defines disinformation as being “intended to manipulate, cause damage and guide people, organizations and countries in the wrong direction.” It can be crafted by sophisticated foreign state actors seeking to gain politically.

Minister St-Onge’s office has not responded to a request for clarification as of this posts publication.

In describing its project to counter disinformation, UdeM said events like the Jan. 6 Capitol breach, the Brexit referendum, and the COVID-19 pandemic have “demonstrated the limits of current methods to detect fake news which have trouble following the volume and rapid evolution of disinformation.” For the Silo, Noe Chartier/ The Epoch Times.

The Canadian Press contributed to this report.

Can C3.AI Stock Keep Rallying with AI in the Spotlight?

The recent rise of Artificial intelligence (AI) programs such as ChatGPT has created a frenzy around AI-related stocks.


C3.AI, a pure play AI stock, is up over 100% since late December.

But is this rally sustainable? After all, the public was already surrounded by AI without realizing it. Almost everything people use in daily life is affected by AI already: 

  • advertising
  • entertainment streaming services
  • social media
  • cars (collision detection and blind spot monitoring)
  • fraud prevention
  • screening job applicants
  • email spam filters
  • many other applications

C3.AI is a company that creates software to help other companies deploy AI projects. C3 software is being used in multiple ways, including managing inventories, monitoring for energy inefficiencies, and predicting system failures. [Of particular note is one new product from C3 called ex machina which allows users to program AI initiatives without using any coding at all but instead via a series of visual programming tools. CP]

AI stocks, and technology stocks as a whole, were a neglected market in 2022. The Nasdaq 100, an index heavy in technology stock, fell more than 30% in 2022. C3.AI fell over 65% in 2022, and is currently down almost 90% from its 2020 high (even after the 100% rally in 2023). All currency quotes that follow are in USD.

C3.AI recently peaked at $30.92 on February 6. It then reached a low of $20.31 on March 1 before rallying back to $29.98. It has since fallen and is back near the $20.33 low.

This puts the stock at a crucial level.

An analyst from SafeTradeBinaryOptions.com had this input: “Right now, the stock is in an uptrend, albeit a precarious one. The price has been making higher swing lows and higher swing highs throughout 2023. But if the price drops much below $20, that will no longer be the case. The price will have made a lower high on March 6 (compared to February 6) and if the price drops below the March 2 low, that is a lower low. These are signs of a downtrend starting — not an uptrend.”

All facets of our modern world are already in the embrace of A.I. whether we know it or not.

This $20 region is important because if the area holds, this indicates the price is moving in a range, with the possibility of the price moving back up to the top of the range near $29. If that happens, there is still hope that the price will eventually break out of the range to upside, continuing its advance to $40, for example. 

However, if the price drops below the $20 region, the range is broken and the uptrend is in jeopardy. 

It’s important to watch C3.AI to see how investors are perceiving the future of AI, and what that may mean for the industry’s future. 

As of March 2023, C3 doesn’t have a lot of direct competition. The company is not yet even profitable. How the stock moves is based on whether investors believe the company can eventually generate profits — and in this case, its profits largely depend on whether AI becomes even more widespread than it already is. For the Silo, Kat Fleischman.