Another week has passed, and here is a roundup of the news related to artificial intelligence and education. The biggest news has been the launch of Google's new language model, Gemma, and we delve a bit into what you can use the model for in teaching. At the same time, it has emerged that there are challenges with Google's Gemini, which i.a. has trouble generating historically correct images and is accused of being woke. Meanwhile, Marie Bjerre (V) has rejected the idea that a Danish language model should be created that is trained on Danish data. She believes the development is something that the private business community can do better than another large public IT project. At the same time, Læremidler.dk has published Learning Tech 14, focusing on AI and didactics in education.
Good reading

Google Gemma

This week, Google launched a new language model called Gemma that can run locally on one's computer without using the Internet. Therefore, the model is also attractive for subjects such as digital technology understanding, communication and IT, informatics and programming, since students can run and test the model on their computer. This also means that Google Gemma can be built into students' applications or used to test language models in teaching. These offline language models are not yet on par with what we see in ChatGPT or Google Gemini, and they are very challenged in the Danish language.

Google Gemma comes in 2 different sizes, 2B and 7B. Gemma 7B is the largest model that can solve complex tasks such as mathematics and coding. In contrast, in the smaller model, 2B, there is more focus on accessibility on devices with limited computing capacity - e.g., laptops without a large graphics card. Google Gemma is technically not an open-source language model, and it is impossible to train the models further. You can test Google Gemma via LM-Studio if you're up for it. Read more about Gemma below:

Google’s New Gemma 2B and 7B Open-Source AI Models, But Do They Beat Meta Llama 2 7B and Mistral 7B?
Explore the new Gemma 2B and 7B open-source models released by Google and discover how they stack up against Meta Llama 2 7B and Mistral 7B.
Google Gemma LLMs small enough to run on your computer
Your own personal chatbot awaits

The development of a Danish language model is ruled out by the minister

Digitalization Minister Marie Bjerre (V) will not support the development of a Danish language model, since there is a risk that the model will not be "good enough". Although Norway and Sweden work on national language models, the minister believes private companies do the work better.

In their digitization strategy, the government has allocated DKK 61 million. DKK for "strategic efforts for artificial intelligence", and it is unclear which efforts these funds will be used for. Marie Bjerre tells Børsen that she cannot say exactly what the strategic money will go to, as it must be spent with the parties to the agreement during 2024.

Several people have criticized the government's new digitization strategy for lacking ambitions in artificial intelligence. The criticism is that the allocated funds are insufficient to create significant value, and they call on the government for more excellent investment in specific AI projects. On the contrary, Digitalization Minister Marie Bjerre believes that the strategy defends a foundation for future AI efforts and denies that the approach is unambitious.

Kritikere står i kø: Der mangler AI-ambitioner i ny strategi
Brancheorganisationer, CBS-professor og it-ordfører efterlyser større AI-ambitioner i den nye digitaliseringsstrategi, som digitaliseringsminister Marie Bjerre…

Bag betalingsmur

Minister lægger dansk Chat GPT i graven
Digitaliseringsminister Marie Bjerre (V) var i tvivl, men alligevel ville hun ikke udelukke, at regeringen…

Bag betalingsmur

Google Gemini's image generator is paused

Google Gemini is Google's new language model designed to compete with ChatGPT4. ChatGPT includes an integrated image generator comparable to ChatGPT's Dall-E 3. This image generator has now been paused following allegations of historically incorrect images - some have called it incredibly woke.

@BasedTorba på mediet X

For example, Gemini has created images of a Pope, represented as a dark man or woman, and the astronauts from Apollo 11 are depicted as a black man and a woman. Google has stated that they carefully adjusted the model during development to prevent the problems that have previously arisen with image generation technology, such as bias and inappropriate images, and to achieve more inclusive and versatile images. An overcompensation has likely occurred during this fine-tuning, triggering these new challenges.

We have experienced Google Gemini's attempt to be inclusive, also with texts.

Googles billedgenerator er blevet beskyldt for at være absurd woke, og nu er den sat på pause
Gemini havde især problemer med at lave historisk korrekte billeder.
Google explains Gemini’s “embarrassing” AI pictures of diverse Nazis
Gemini will undergo testing before it can start generating people again.
Gemini image generation got it wrong. We’ll do better.
An explanation of how the issues with Gemini’s image generation of people happened, and what we’re doing to fix it.

Learning Tech 14 released: focus on AI and didactics in education

Læremiddel.dk's newest themed issue explores the role of AI in didactics and education. The articles range from using AI in language teaching to adaptive teaching aids in mathematics, ethical considerations around using AI chatbots and measuring literacy with eye tracking technology.

Read the entire release here:

Learning Tech 14: AI og didaktik i uddannelse
Temaet for og titlen på Learning Tech 14 er AI og didaktik i uddannelse. Nummeret er udkommet i december 2023. Læs gratis med her.

New articles on Viden.AI

When we write articles on the blog, we mostly do it based on our interests and from our point of view in the education sector. We write for pleasure and to inform and pass on knowledge about artificial intelligence in the entire education sector. If anyone wants to write on the blog, you should always write to kontakt@viden.ai.

This week, we at Viden.AI have published two new articles:

Who will take responsibility?

Malte has published an article that asks the big question of who should take responsibility for implementing and managing AI in schools. Here, he points out that there must be clarity about the division of tasks and duties throughout the chain of actors. At the same time, the values ​​we want in our education system with the new technological possibilities must be clearly emphasized.

Hvem skal tage ansvaret?
Med den eksplosive udvikling af systemer, der anvender generativ kunstig intelligens, rejser der sig en masse uddannelsesmæssige spørgsmål. Spørgsmålet om, hvem der skal tage ansvaret for implementeringen og styringen af AI i skolerne, presser sig på. Men har vi brug for at handle? Skal vi hellere v…

Analysis: How does AI affect the work of teachers?

To what extent will language models affect the education sector? Statistics Denmark has published an analysis of how the Danish labor market is affected by large language models such as ChatGPT. Per has written briefly about it focusing on the education sector, but the entire analysis from Statistics Denmark is worth diving into.

If you want to avoid artificial intelligence in your workplace, you should probably become a painter 🤔

Analyse: Hvordan påvirker AI undervisernes arbejde?
Danmarks Statistik har sammen med forskere fra Copenhagen Center for Social Data Science (SODAS) ved Københavns Universitet udgivet en analyse, hvor de har undersøgt, hvordan de store sprogmodeller (LLM) som fx ChatGPT påvirker det danske arbejdsmarked. I denne artikel vil jeg meget kort introducere…

News

Minister lægger dansk Chat GPT i graven
Digitaliseringsminister Marie Bjerre (V) var i tvivl, men alligevel ville hun ikke udelukke, at regeringen…
Vi har brug for en germansk sprogmodel | DataTech (PRO)
Vi ser i øjeblikket en udvikling, hvor amerikanske sprogmodeller dominerer teknologiudviklingen. De fleste kender ChatGPT og Bard, der er sprogmodeller i lukkede systemer, som primært er trænet på store sprog som engelsk og drevet af kommercielle interesser i USA. Men de store systemer har også en s…
China releases list of 184 AI education bases in schools
China’s Ministry of Education (MOE) recently announced a list of 184 primary and secondary schools selected as artificial intelligence (AI) education bases, with the aim of better promoting the development of AI education.
– Et problem at lærere forholder seg så ulikt til kunstig intelligens
Elevene er opptatt av at tilgangen til kunstig intelligens og regler for bruk må være lik ved alle skoler.
Har AI og arbejdsmiljø noget med hinanden at gøre?
Hvordan skal vi bruge chatbots i fremtiden?
Prodekan: ”Vi integrerer kunstig intelligens i fremtidens uddannelser”
AU giver nu grønt lys til, at studerende må benytte generativ kunstig intelligens (GAI). Det er et naturligt skridt, hvis vores uddannelser skal følge med tiden, lyder det fra Lise Wogensen Bach, der er prodekan for uddannelse på Health.


Scientific articles

We have previously written about misinformation and disinformation, which will certainly be something we will focus on for a long time.

In the article: "Comparing the willingness to share for human-generated vs. AI-generated fake news", researchers from the University of Lausanne and the Munich Center for Machine Learning examine users' reactions to fake news created by humans and artificial intelligence, respectively. The study shows that people generally find fake news written by artificial intelligence slightly less credible than those produced by humans. However, it does not matter whether you want to share the news, because according to the study, it makes no difference who wrote the fake news.

They also found in the study that conservative participants were more likely to accept fake news as accurate.

The Decoder has made an excellent review of the article, which can be read here:

Users can’t tell the difference between AI-generated and human fake news, study finds
A study has investigated how users react to fake news generated by humans and AI, and identified socio-economic factors that influence who falls for AI-generated fake news.
Comparing the willingness to share for human-generated vs. AI-generated fake news
Generative artificial intelligence (AI) presents large risks for society when it is used to create fake news. A crucial factor for fake news to go viral on social media is that users share such content. Here, we aim to shed light on the sharing behavior of users across human-generated vs. AI-generat…

How generative artificial intelligence is transforming business and society

Oliver Wyman Forums has just published a report that overviews how generative artificial intelligence will affect companies, employees, and consumers - including a bit about education.

It is a comprehensive report and can be downloaded here.

Rapport: How Generative AI Is Transforming Business And Society