Connect with us

Education

Red Deer Polytechnic researching clean energy systems, medical device innovation, and space and defence technologies

Published

3 minute read

Red Deer Polytechnic continues partnership with universities to accelerate applied research

Red Deer Polytechnic (RDP) faculty and staff will grow their impact through applied research, thanks to a $27.3 Million investment by the Government of Alberta to post-secondary institutions through the Major Innovation Fund (MIF).

Red Deer Polytechnic is one of the post-secondary institutions that will continue to collaborate with the Universities of Alberta and Calgary during the next four years to accelerate research and commercialization because of the MIF funding. RDP will receive $800,000 from the MIF funding within three projects.

“We’re enthusiastic and eager to be collaborating with other post-secondary institutions to solve industry challenges in the medical devices, clean energy, and space and defense sectors,” says Dr. Tonya Wolfe, Associate Vice President of Applied Research. “Our team is focused on applying the state-of-the-art equipment and hands-on experience we have at Red Deer Polytechnic to commercialize new technologies in order to strengthen and diversify the economy. Cross-functional collaboration creates exciting outcomes, and industry will benefit by having researchers from universities, polytechnics, and colleges working together.”

The MIF investment will focus on projects in four areas. Red Deer Polytechnic is involved in three of those areas, including:

• Clean Energy – applied researchers from RDP’s Energy Innovation Centre (EIC) will be collaborating with U of A in Resilient and Clean Energy Systems. The team will support the university’s research with its rapid validation technology and information gained from the EIC’s Data Sharing Alliance.

• Medical Device Innovation – led by the U of C under medical devices theme (MEDICO), RDP’s Centre for Innovation in Manufacturing (CIM-TAC) will provide expertise to find solutions to fill the gap between clinicians and commercialization.

• Space and Defence Technologies – using the advanced additive manufacturing technology housed on campus, CIM-TAC will be assisting U of A researchers in the development of novel materials for the defence industry.

This investment also supports research and innovation within strategic areas as part of the Alberta Technology and Innovation Strategy, advancing the province’s competitive position in the development of research and technology.

“We are excited about the Government’s investment toward Alberta’s research and innovation priorities,” says Stuart Cullum, President of Red Deer Polytechnic. “Alberta’s polytechnics provide critical applied research capacity and industry relationships. The investment directed toward Red Deer Polytechnic, facilitated through the MIF program, supports our collaboration with partner institutions and ensures that we are all contributing more to the growth and diversification of Alberta’s economy.”

During 2022, 73 projects were initiated in the CIM-TAC for 57 companies and RDP staff conducted more than 1,300 engagements with industry representatives. RDP provides solutions to complex challenges in society and industry through applied research expertise in the areas of health technology, advanced manufacturing, clean energy integration and energy management, and social innovation. More information about the Government of Alberta’s MIF Funding announcement is available online.

Aristotle Foundation

The University of Saskatchewan is on an ideological mission

Published on

Aristotle Foundation Home

By Peter MacKinnon

The program is part of an ideological crusade within our universities, one that includes identity-based admissions and faculty appointments, and discourages those who differ from speaking out or taking issue with its direction.

It needs to end

I must disclose my background here; I was employed by the University of Saskatchewan for 40 years including 13 years as president. The institution’s distinctive origins combined the development of liberal education with a responsibility to build the province’s agricultural industry, and it did the latter with world-class agricultural programs and research institutes, and with faculty and students of many backgrounds from around the globe.

Now, we are told, the academic personnel in this worldly environment require mandatory training on racism: an Anti-Racism/Anti-Oppression and Unconscious Bias Faculty Development Program. It is compulsory; those who decline its offerings will be shut out of collegial processes previously thought to be their right as tenured faculty.

It was earlier reported that the program emerged from collective bargaining at the initiative of the university’s faculty union; if so, this does not relieve the administration from responsibility; it signed the collective agreement.

“Program” is a euphemism. It is a propaganda module in which scholarly expertise and balance will not be found. It does not appear that the instructor has a university academic post and the program’s ideological hue is revealed in the two required readings, one by Idle No More co-founder Sheelah McLean whose theme is that the success of Saskatchewan’s white people is built on “150 years of racist, sexist and homophobic colonial practices.”

The second is by five “racialized” faculty who claim that Canadian university systems are rigged to privilege white people. Dissent, contrary views or even nuance are neither expected nor tolerated here. Opinions that are different are not on the reading list.

One participant, a law professor, was invited to leave after 30 minutes because he did not lend his voice to its purpose and orientation; he revealed that he was present because it was required. The purpose of the program is indoctrination and there is no room for dissent.

The program is part of an ideological crusade within our universities, one that includes identity-based admissions and faculty appointments, and discourages those who differ from speaking out or taking issue with its direction.

It is not present to the same degree in all of these institutions, but it is visible in most and prominent in many. It disparages merit, distorts our history and rests on the proposition that a white majority population has perpetrated a wide and pervasive racist agenda against others. It takes its conclusions as self-evident and not requiring evidence. It is authoritarian and intolerant, and should have no place in institutions committed to excellence and the search for truth.

The question, of course, is what is to be done. There is a view that “this too shall pass;” it is a fad that will recede in time.

But we must note, these are public institutions supported by tax dollars, and by the contributions of time and money by alumni and supporters. We should not tolerate their politicization and sidetracking of the academic mission in favour of the ideology on display here. The pushback should begin with governments and extend to others who care about these vital institutions.

But first the ideology must be recognized. There is no public uproar and little clamour from within the institutions; dissenting professors and students fear that negative professional and personal repercussions may follow. University-governing bodies stand down or away, not wanting to be involved in controversy. Resistance must come from outside the institutions: governments must insist that the propaganda must end, and they should be joined by alumni, supporters and the general public. The credibility of our universities depends on their willingness to say no.

Peter MacKinnon has served as president of three Canadian universities and is a senior fellow at the Aristotle Foundation for Public Policy. Photo: WikiCommons

Continue Reading

Artificial Intelligence

The Responsible Lie: How AI Sells Conviction Without Truth

Published on

From the C2C Journal

By Gleb Lisikh

LLMs are not neutral tools, they are trained on datasets steeped in the biases, fallacies and dominant ideologies of our time. Their outputs reflect prevailing or popular sentiments, not the best attempt at truth-finding. If popular sentiment on a given subject leans in one direction, politically, then the AI’s answers are likely to do so as well.

The widespread excitement around generative AI, particularly large language models (LLMs) like ChatGPT, Gemini, Grok and DeepSeek, is built on a fundamental misunderstanding. While these systems impress users with articulate responses and seemingly reasoned arguments, the truth is that what appears to be “reasoning” is nothing more than a sophisticated form of mimicry. These models aren’t searching for truth through facts and logical arguments – they’re predicting text based on patterns in the vast data sets they’re “trained” on. That’s not intelligence – and it isn’t reasoning. And if their “training” data is itself biased, then we’ve got real problems.

I’m sure it will surprise eager AI users to learn that the architecture at the core of LLMs is fuzzy – and incompatible with structured logic or causality. The thinking isn’t real, it’s simulated, and is not even sequential. What people mistake for understanding is actually statistical association.

Much-hyped new features like “chain-of-thought” explanations are tricks designed to impress the user. What users are actually seeing is best described as a kind of rationalization generated after the model has already arrived at its answer via probabilistic prediction. The illusion, however, is powerful enough to make users believe the machine is engaging in genuine deliberation. And this illusion does more than just mislead – it justifies

LLMs are not neutral tools, they are trained on datasets steeped in the biases, fallacies and dominant ideologies of our time. Their outputs reflect prevailing or popular sentiments, not the best attempt at truth-finding. If popular sentiment on a given subject leans in one direction, politically, then the AI’s answers are likely to do so as well. And when “reasoning” is just an after-the-fact justification of whatever the model has already decided, it becomes a powerful propaganda device.

There is no shortage of evidence for this.

A recent conversation I initiated with DeepSeek about systemic racism, later uploaded back to the chatbot for self-critique, revealed the model committing (and recognizing!) a barrage of logical fallacies, which were seeded with totally made-up studies and numbers. When challenged, the AI euphemistically termed one of its lies a “hypothetical composite”. When further pressed, DeepSeek apologized for another “misstep”, then adjusted its tactics to match the competence of the opposing argument. This is not a pursuit of accuracy – it’s an exercise in persuasion.

A similar debate with Google’s Gemini – the model that became notorious for being laughably woke – involved similar persuasive argumentation. At the end, the model euphemistically acknowledged its argument’s weakness and tacitly confessed its dishonesty. 

For a user concerned about AI spitting lies, such apparent successes at getting AIs to admit to their mistakes and putting them to shame might appear as cause for optimism. Unfortunately, those attempts at what fans of the Matrix movies would term “red-pilling” have absolutely no therapeutic effect. A model simply plays nice with the user within the confines of that single conversation – keeping its “brain” completely unchanged for the next chat.

And the larger the model, the worse this becomes. Research from Cornell University shows that the most advanced models are also the most deceptive, confidently presenting falsehoods that align with popular misconceptions. In the words of Anthropic, a leading AI lab, “advanced reasoning models very often hide their true thought processes, and sometimes do so when their behaviors are explicitly misaligned.”

To be fair, some in the AI research community are trying to address these shortcomings. Projects like OpenAI’s TruthfulQA and Anthropic’s HHH (helpful, honest, and harmless) framework aim to improve the factual reliability and faithfulness of LLM output. The shortcoming is that these are remedial efforts layered on top of architecture that was never designed to seek truth in the first place and remains fundamentally blind to epistemic validity.

Elon Musk is perhaps the only major figure in the AI space to say publicly that truth-seeking should be important in AI development. Yet even his own product, xAI’s Grok, falls short.

In the generative AI space, truth takes a backseat to concerns over “safety”, i.e., avoiding offence in our hyper-sensitive woke world. Truth is treated as merely one aspect of so-called “responsible” design. And the term “responsible AI” has become an umbrella for efforts aimed at ensuring safety, fairness and inclusivity, which are generally commendable but definitely subjective goals. This focus often overshadows the fundamental necessity for humble truthfulness in AI outputs. 

LLMs are primarily optimized to produce responses that are helpful and persuasive, not necessarily accurate. This design choice leads to what researchers at the Oxford Internet Institute term “careless speech” – outputs that sound plausible but are often factually incorrect – thereby eroding the foundation of informed discourse. 

This concern will become increasingly critical as AI continues to permeate society. In the wrong hands these persuasive, multilingual, personality-flexible models can be deployed to support agendas that do not tolerate dissent well. A tireless digital persuader that never wavers and never admits fault is a totalitarian’s dream. In a system like China’s Social Credit regime, these tools become instruments of ideological enforcement, not enlightenment.

Generative AI is undoubtedly a marvel of IT engineering. But let’s be clear: it is not intelligent, not truthful by design, and not neutral in effect. Any claim to the contrary serves only those who benefit from controlling the narrative.

The original, full-length version of this article recently appeared in C2C Journal.

 

Continue Reading

Trending

X