Episode Transcript
[00:00:00] Your guide to 2024's rapid AI developments and their impacts on the language industry By Veronica hylak In just 58 years since the Massachusetts Institute of Technology created the first artificial intelligence chatbot in 1966, the world has seen a series of monumental technological breakthroughs. First personal computers, then email, the Internet, and smartphones.
[00:00:28] Each step forward arrived a few years after the last, unfolding at a quick yet manageable pace that allowed society time to adjust accordingly. But 2024 has felt different, marked by a relentless acceleration driven by AI.
[00:00:42] Some advancements feel like they happened ages ago, yet they're only months old. We're no longer in a steady progression it's now a wave of rapid advancements that arrive almost daily.
[00:00:54] Feeling overwhelmed by the endless AI noise, I started a video blog or vlog called the AI Almanac.
[00:01:01] The project quickly resonated with thousands of people who felt the same way. Each episode I focused only on the most critical new AI advancements.
[00:01:11] But with the year coming to a close, I wanted to choose the top AI models of 2024. But then it hit me. The AI landscape is no longer about isolated product releases, but rather larger changes across many companies that are fundamentally reshaping how we interact with the world. I now believe it's better to evaluate the most important overarching trends that have emerged in the AI space.
[00:01:35] In this article, I highlight the eight biggest takeaways we can glean from the many updates that happened this year. 1. The AGI dream is officially dead for now Artificial General Intelligence AGI the idea that AI models could reason like humans was effectively put to rest after growing suspicions increased over the course of the year. The final nail in the coffin was when Apple released a research paper stating that large language models are excellent at recognizing patterns but show no signs of actual logical reasoning.
[00:02:09] Even chain of thought techniques designed to simulate reasoning used in the OpenAI 01 models have been exposed as merely imitating learned patterns rather than engaging in an actual thought process.
[00:02:22] The revelation aligns with the growing consensus that the tech industry's next hurdle lies in building models that move beyond reproducing trained patterns to achieve genuine logical reasoning. While many people are suddenly emerging from the shadows claiming that they have known this all along, others still hold out hope. Even OpenAI seems to still believe AGI is achievable.
[00:02:44] Personally, I'm open to the idea that they might know something we don't. Regardless, this year marked a major turning point for AI, as the hype finally began to cool.
[00:02:54] Investors who once poured money into speculative projects are now demanding real results, shifting from flashy AI toys to practical applications that prove genuine value. 2. Generative engines changed how we search online In Multilingual Magazine's December 2023 issue, I predicted that generative AI would change how we find information online.
[00:03:18] That quickly became reality in 2024, while perplexity AI was already around. Google's overhaul of its legacy algorithms and launch of AI summaries alongside Bing and OpenAI search GPT pushed generative search into the mainstream. Generative engines synthesize data from multiple sources using LLMs to provide immediate natural language responses.
[00:03:43] This reduces the need to visit individual sites or dig through links like with traditional search engines.
[00:03:49] This shift has spawned a new search optimization method called Generative engine Optimization that focuses on aligning content with generative patterns rather than traditional search engine optimization and SEO tactics like keyword stuffing and backlinks, which reduce visibility in generative environments.
[00:04:09] For example, keyword heavy content, once central to SEO, can lower visibility in generative engines by up to 10%. The rise of generative search engines has raised anti competitive concerns among government bodies across the globe. Generative engines often use content without consent, redirect traffic away from original sources, and cause significant drops in site visits.
[00:04:34] To prevent these effects, publishers must opt out of all search indexing, including Google.
[00:04:40] This means companies are essentially forced to allow AI access or risk being excluded from the digital marketplace. With technology fatigue on the rise, the shift to generative engines feels like a relief for some people, though many others aren't even aware it's happened.
[00:04:57] For me, I'm happy I can now snap a picture of a random mushroom in the forest to ask Google what it is or get a smoothie recommendation without digging through endless links.
[00:05:07] Personally, I'm not going back, but this development has many companies worried websites could become obsolete if users no longer need to visit them.
[00:05:18] I'm constantly pulled into many board conversations where SEO teams are under growing pressure to figure out how to handle optimizing their content for generative environments as traffic continues to drop. 3. Data privacy is at risk and people finally paid attention Genai became embedded in nearly every social media platform, laptop and mobile phone this year, weaving itself into the daily lives of users worldwide.
[00:05:44] Platforms like Facebook, Instagram, WhatsApp, and even LinkedIn introduced AI features that leverage vast amounts of user data, bringing data privacy to the front of the conversion, despite being something we have neglected up until now.
[00:05:58] In particular, Meta's approach to training on user data hit roadblocks in regions like Europe and Brazil, signaling a pushback against tech giant's extensive data usage.
[00:06:08] Meanwhile, consent has become almost meaningless to many people as users navigate hidden policies that are impossible to find or turn off. What many people don't realize is that companies like Meta operate far beyond traditional stateless models. Their AI systems are deep contextual engines designed to analyze decades of user behavior, tailoring responses based on highly personalized insights.
[00:06:33] This goes far beyond surface level personalization. It's an attempt to predict and shape user interactions at an unprecedented scale to execute almost everything.
[00:06:42] Apple took a notably different approach with the launch of Apple Intelligence, which features the most advanced privacy protective AI infrastructure available to everyday users.
[00:06:53] Apple's private cloud compute infrastructure ensures that data remains encrypted and inaccessible even to Apple, gives every user a private cloud instance, and sets a new benchmark for privacy protection.
[00:07:06] Unlike the data driven AI models of Meta and LinkedIn, Apple's approach prioritizes user security over data exploitation, signaling a crucial shift as the demand for privacy in AI grows. Unfortunately, Apple Intelligence was the year's most overlooked AI launch.
[00:07:23] While not as flashy as competitors, Apple's approach set a new privacy standard, standing out as one of the few truly protecting everyday users.
[00:07:32] As a data privacy advocate, I feel reassured using Apple devices, a trust one don't extend to others. Many people believe Apple is falling behind.
[00:07:42] I believe that they know that slow and steady wins the race.
[00:07:46] Apple, I'm keeping my eye on you. 4 the language industry got its first native LLM in 2024. The language industry saw significant advancements as a natural evolution. Given that LLMs were originally designed with translation in mind.
[00:08:03] More companies focused on developing task specific proprietary models to improve precision and quality assurance.
[00:08:11] And DeepL entered the interpreting space with its launch of DeepL Voice, showcasing its potential to break down language barriers in real time and foster communication in virtual settings like Microsoft Teams. One standout release sparked deeper questions about the evolving role of human translators Translate. Its launch of Lara, the industry's first native LLM developed with Nvidia Lara set a new accuracy benchmark, claiming only 2.5 errors per thousand words, 50% fewer than the industry standard of five errors with professional human translators.
[00:08:46] While some fear these advancements will edge out human translators, the idea of humans as essential co pilots in the AI driven translation process is surging powerfully. Articulated by Marina Panchiva of rws, this idea resonates across the industry. Thirty years ago, linguists were at the center of the translation process.
[00:09:07] As technology and automation advanced, they were pushed to the periphery, ending up somewhere in the loop.
[00:09:14] Now it's time for linguists to take their seat in the cockpit from where they guide the development of linguistic AI.
[00:09:21] 5. AI video dubbing and Translation Went Mainstream the video dubbing industry saw significant advancements with the integration of AI to enhance content accessibility and localization.
[00:09:33] Meta, YouTube and TikTok all launched AI driven dubbing features to help creators reach broader audiences by providing multilingual content.
[00:09:42] TikTok's Symphony AI dubbing tool automatically detects the original language in a video, transcribes, translates, and produces a dubbed version in the selected languages, enabling creators and brands to produce content that resonates across cultures.
[00:09:58] Similarly, heygen launched an AI powered video translation feature that clones the user's facial expressions, natural speaking voice and style. This allows seamless delivery in multiple languages and makes content more accessible to a global audience.
[00:10:14] While some fear AI will replace human translators, others believe it will lead to an unprecedented increase in translated content keeping professionals engaged.
[00:10:24] This perspective aligns with RWS Trados's Translate Everything slogan and was echoed by Gabriel Fairman of Bureau Works at the American Translators association conference this year.
[00:10:38] As social media platforms enable creators to reach more audiences and companies find translation more accessible, this trend seems likely. While I can agree AI will boost translation volumes in commercial sectors, I do not believe this surge will extend by default to government or specialized fields where reaching broader audiences doesn't directly drive sales.
[00:10:59] 6. AI in the creative arts fueled copyright battles in 2024, countless new AI powered tools emerged to transform traditionally human driven arts like film, music and visual media.
[00:11:13] AI video generators produced high quality realistic clips that simulate human emotions and animations such as OpenAI's Sora and Runway's Gen 2, creating visuals as if they were straight out of a movie.
[00:11:26] Their potential quickly attracted attention in the film industry through an opportunity to reduce costs and achieve complex visual effects without large budgets.
[00:11:36] On the music side, AI platforms like Suno and Udio launched to allow users to generate music from scratch, replicating popular artists styles and even inventing new genres. While these tools expanded creative possibilities, they also raised significant questions about the legality of their outputs.
[00:11:55] Major record labels like Universal, Sony and Warner filed their first lawsuits against AI music generators for using copyrighted material or artist likeness in outputs. While Hollywood unions went on strike to secure protections against AI generated scripts and digital likenesses.
[00:12:14] Visual artists and authors also filed lawsuits over the unauthorized use of their work in AI training datasets. We all knew this was coming. As a former Sony Music employee who worked in copyright compliance, I weirdly see both sides of the argument.
[00:12:31] While my opinion may not be entirely popular. The truth is we are touching an area of copyright law that hasn't been considered.
[00:12:39] There is currently a gap in copyright law that mainly addresses public distribution, not private use within companies.
[00:12:47] AI companies argue that using copyrighted works privately falls outside traditional copyright rules.
[00:12:54] Similarly, an artist's likeness such as voice and style was never traditionally considered copyrightable, but laws such as Tennessee's Elvis act are starting to change that. The bottom line is AI model creators are currently operating in a legal gray area similar to Spotify's early days before copyright law adapted to streaming.
[00:13:15] I predict this will lead to a lose lose situation with new copyright provisions being established moving forward rather than significant repercussions retroactively.
[00:13:25] Seven AI companies got involved in government and military affairs in January, OpenAI quietly removed its military use ban from its terms of service and confirmed collaboration with the United States US Pentagon, sparking global concerns about private companies collaboration with military AI applications.
[00:13:46] Multiple summits throughout the year continued to address military AI use. Over 90 nations convened in Seoul to establish guidelines for responsible conflict use, especially as AI enabled drones gained prominence in the Russia Ukraine war.
[00:14:01] Many organizations like the North Atlantic Treaty Organization also updated their AI guidelines. The European Union passed the AI act, categorizing AI applications into risk tiers and enforcing strict rules for high risk AI systems, while the Council of Europe introduced an international treaty to protect human rights against AI.
[00:14:24] Microsoft also banned US Law enforcement from using Azure AI for facial recognition, citing risks of bias in high stakes environments. Six years ago, I worked with AI on US Government drone ships, so this isn't exactly new. But with Gen AI as the buzzword, even governments are centering nonstop discussions around it.
[00:14:45] I still find it ironic that model makers emphasize putting humans first while essentially saying our priority is humanity's safety. But let's teach the robot's son tzu, just in case. 8ai's environmental impact drove investment in nuclear power as millions of people interact with AI models regularly, their environmental impact is quickly increasing.
[00:15:08] AI data centers already consume 1.5% of global electricity, according to the International Energy Agency.
[00:15:17] This figure is projected to match the entire energy consumption of the Netherlands by 2027. As noted in research published in Joule, every 10 to 50 prompts on ChatGPT3 requires about half a liter of water just to keep servers cool. As found by University of California researchers, AIs powerful GPUs consume four times more energy than typical cloud servers. According to Reuters, generating a single image with Dall E uses about the same amount of energy as charging a smartphone as shown by researchers at Hugging Face and Carnegie Mellon University Training GPT4 emits approximately 500 metric tons of CO2, equivalent to burning 2.7 million pounds of coal in a single day at a standard coal fired power plant, according to research from Hugging face.
[00:16:06] Companies like OpenAI and Google are realizing they can't keep up with AI's skyrocketing energy demands without serious infrastructure changes.
[00:16:15] These companies heavily invested this year in nuclear power, highlighting a push in the tech industry for sustainable energy.
[00:16:23] Balancing nuclear power with safer renewables like wind and solar and ensuring government oversight will be essential to avoid environmental and safety risks. Conclusion AI has reshaped our world in profound ways this year, sending ripples we're only beginning to grasp.
[00:16:41] From the surge in data privacy concerns to groundbreaking strides in creative fields, translation and everyday tech, the rapid transformation has exposed both immense potential and significant risks that society is racing to keep up with as we look to the future. It's about not only advancing AI, but also ensuring it aligns with values that protect and empower.
[00:17:04] The road ahead will demand careful balance, responsibility and an unwavering commitment to ethical innovation that truly serves humanity.
[00:17:12] This article was written by Veronica Hylak, co CEO of Metalinguist, award winning AI product innovator and host of the AI Almanac flog. With 10 years of experience working with Fortune 500 companies, the US government and startups, she has led many high impact projects and loves to build things that solve problems.
[00:17:33] Originally published in Multilingual Magazine Issue 234 October 2024.