Episodios

  • FIR #508: Inside AI’s Human Raw Material Supply Chain
    Apr 8 2026
    When workers lose their jobs, many turn to gig work to earn income while waiting for new opportunities. Increasingly, companies that hire gig workers are shifting from delivering food or sharing rides to creating content to train AI systems. This raises various communication and ethical issues. Neville and Shel explain what’s happening and discuss the implications in this short midweek episode. Links from this episode: The jobs AI can’t do – and the young adults doing themThousands of people are selling their identities to train AI – but at what cost?The gig workers who are training humanoid robots at homeGig economy becomes new AI training ground The next monthly, long-form episode of FIR will drop on Monday, April 27. We host a Communicators Zoom Chat most Thursdays at 1 p.m. ET. To obtain the credentials needed to participate, contact Shel or Neville directly, request them in our Facebook group, or email fircomments@gmail.com. Special thanks to Jay Moonah for the opening and closing music. You can find the stories from which Shel’s FIR content is selected at Shel’s Link Blog. You can catch up with both co-hosts on Neville’s blog and Shel’s blog. Disclaimer: The opinions expressed in this podcast are Shel’s and Neville’s and do not reflect the views of their employers and/or clients. Raw Transcript Shel Holtz Hi everybody and welcome to episode number 508 of For Immediate Release. I’m Shel Holtz. Neville Hobson And I’m Neville Hobson. Over the past few weeks, I’ve come across a set of stories that all point to something quite striking — not just how AI is evolving, but how it’s being built. Increasingly, the raw material behind AI isn’t just data scraped from the web. It’s us: our voices, our movements, our everyday lives, and increasingly, our identities. There’s a new layer of the gig economy emerging. We’ll explore this in just a minute. People are being paid, typically in small amounts, to record themselves walking down the street, having conversations, folding laundry, even just going about their day. That data is then used to train AI systems because those systems need examples of how people actually speak, move, and interact in the real world. In one case, delivery drivers in the US are being redirected to film tasks for robotics training. Platforms are turning existing gig workers like delivery drivers into distributed data collectors for AI. In another example, people are selling access to their phone conversations through apps that pay contributors to upload voice and text data. And in yet another, workers are strapping phones to their heads to record household chores so humanoid robots can learn how to move. The work is global, fragmented, and often invisible, with workers spanning Nigeria, India, South Africa, the US, and far beyond. Humans are no longer just users of AI — they are raw material suppliers. In China, there are even state-run centers where workers wear virtual reality headsets and exoskeletons to teach robots how to carry out everyday physical tasks. What we’re seeing is the rise of what you might call data labor, where identity itself becomes part of the work. There’s a clear driver behind it. AI companies are running out of high-quality training data. The open web isn’t enough anymore, and synthetic data has its limits. So the industry is turning to something else: real human lived experience. Because if you want a robot to understand how to load a dishwasher, navigate a room, or interact with objects, you need to see humans doing it at scale. But there’s an interesting contrast here. One of the stories highlights a 23-year-old in the US, a guy called Cale Mouser, who earns well into six figures repairing diesel engines. It’s something he’s developed great skill in doing. His work depends on judgment, experience, and problem solving in the real world — things that don’t easily translate into data. So while some people are being paid small amounts to generate data for AI systems, others like Cale Mouser are building highly valuable careers precisely because their skills can’t be reduced to it. And that contrast feels important. Because on one level, this new kind of work does create opportunity. For some people, especially in lower-income regions in the Global South, this is real income — paid in dollars, flexible and accessible. But there’s another side to it. Because what people are actually selling isn’t just time, it’s identity: their voice, their behavior, their presence in the world. And often once that data is handed over, it’s gone — permanently licensed, reused, repurposed, potentially in ways the individual never sees or understands. So you have this asymmetry: individuals earning small immediate payments while companies build long-term, highly valuable AI systems. Perhaps it’s a new version of the Mechanical Turk for the AI era. And that raises a deeper question. What does it mean when ...
    Más Menos
    21 m
  • ALP 300: 300 episodes in: what’s changed, what hasn’t, and what we got wrong
    Apr 6 2026

    Eight years and 300 episodes later, Chip and Gini take stock of what the Agency Leadership Podcast has actually been about and where their thinking has shifted since they sat down for lunch outside Wrigley Field and decided to start a show.

    Chip shares an AI-generated analysis of the 10 most common themes across 300 episodes. Gini distills them into four she considers non-negotiable: communication fixes most problems, know your numbers, focus on particular wins, and the owner sets the temperature. Chip adds that communication doesn’t just solve problems, it prevents them. Ironic, given that probably everyone listening is in the communications business.

    On what’s changed, Gini has moved from annual retainer-focused planning to quarterly reviews that constantly show results and surface what’s working. She also notes that her advice for navigating a tough business environment now mirrors what worked during the pandemic: find the project work, start with an assessment, and build trust before building a retainer.

    The biggest evolution for Chip is his position on AI. While he was skeptical a few years ago about the timeline, now he thinks agencies are under-emphasizing it. He and Gini disagree on AI’s limits. Gini believes critical thinking, emotional intelligence, and crisis work still require human judgment. Chip is less certain those guardrails will hold. What they do agree on: AI is turning everyone into a manager, and that puts a premium on skills that were already in short supply.

    The episode closes with a lightning round covering worst advice agencies still believe, best scary decisions, and prospect red flags including unreasonable expectations and unwillingness to discuss budget. [read the transcript]

    The post ALP 300: 300 episodes in: what’s changed, what hasn’t, and what we got wrong appeared first on FIR Podcast Network.

    Más Menos
    24 m
  • ALP 299: Hire people who understand how to solve problems
    Mar 30 2026

    Most hiring processes obsess over the wrong things. Do they know our project management software? Are they proficient in this specific tool? Meanwhile, the one capability that actually determines whether someone will make your life easier or harder—their ability to solve problems independently—gets a cursory “are you a good problem solver?” question that everyone answers with “yes.”

    In this episode, Chip and Gini break down why problem-solving ability should be the primary hiring criterion, especially as AI makes technical skills easier to acquire and offload. The conversation explores why this matters more now than ever: as AI handles tactical execution, the ability to define problems clearly, break them into components, and figure out solutions becomes the differentiator between humans who add value and humans who get replaced.

    Chip and Gini discuss how problem-solving cuts across every role, even ones you don’t typically think of as problem-solving positions. Designers facing impossible deadlines, account people navigating last-minute client demands, anyone dealing with the reality that things rarely go according to plan. They all need to be able to figure out how to move forward rather than escalating every obstacle upward.

    The episode tackles the mechanics of actually interviewing for this capability. You can’t just ask “are you a good problem solver?”—you need scenario-based questions that reveal how candidates think through challenges. But not hypothetical scenarios you make up; real situations that have happened in your agency. Ask them to walk through how they’ve handled compressed timelines, missing information, conflicting priorities, or last-minute changes in past roles.

    Gini shares how her daughter’s school explicitly focuses on humanities and emotional intelligence rather than technical skills, anticipating that AI will reshape what jobs exist. She connects this to Anthropic’s hiring practice of seeking people with humanities degrees who can absorb information, think critically, and demonstrate emotional intelligence rather than just technical proficiency.

    The episode concludes with an important reminder: if you hire problem solvers but then micromanage how they solve problems, you’ve wasted the hire. You need to let them solve things their way, even if it’s different from how you’d do it, or you’ll end up with everything back on your plate anyway. [read the transcript]

    The post ALP 299: Hire people who understand how to solve problems appeared first on FIR Podcast Network.

    Más Menos
    21 m
  • Circle of Fellows #126: Communicating in the Era of the Polycrisis
    Mar 29 2026
    The days when a crisis communicator could simply reach for a dusty binder and follow a pre-scripted, linear checklist are gone — and they aren’t coming back. In the “good old days,” a crisis was often a contained event with a predictable lifecycle; crisis teams could address them by checking off items on a checklist. Today, we face the era of the polycrisis, where economic instability, geopolitical friction, and a 24/7 social media cycle collide, creating a torrent of simultaneous challenges. This new reality has effectively obliterated the traditional news cycle, replacing it with an always-on environment where a single viral post can tarnish a brand before leadership even knows there is a problem. Thriving in this volatile landscape requires a move away from rigid manuals toward a more fluid, strategic approach. Rather than a step-by-step rulebook, modern practitioners need logical scaffolding — a flexible framework of principles and values that provides a foundation for action while allowing for real-time adaptability. It is about preparation, not just prescription. As the boundaries between internal and external perception continue to erode, the ability to maintain transparency and connection through these multifaceted disruptions is no longer a luxury; it is table stakes for organizational survival. Four Fellows of the International Association of Business Communicators (IABC) shared their perspectives in this episode of IABC’s Circle of Fellows. About the Panel: Edward “Ned” Lundquist is a retired U.S. Navy captain with 43 years of professional public affairs and strategic communications experience. His company, Echo Bridge LLC, which provides outreach and advocacy support to government and commercial clients. He served on active duty for 24 years in the U.S. Navy as a surface warfare officer and public affairs specialist. Captain Lundquist was a Pentagon spokesman with the Office of the Assistant Secretary of Defense for Public Affairs, Director of the Fleet Home Town News Center, and director of public affairs and corporate communications for the Navy Exchange Service Command. His last tour of duty was commanding the 450 men and women of the Naval Media Center. He is an accredited business communicator and award-winning communicator who served as president of IABC/Hampton Roads and IABC/Washington, director of U.S. District 3, and chair of the International Accreditation Council. He was named an IABC Fellow in 2016. Captain Lundquist received the Surface Navy Association’s Special Recognition Award in January of this year, for his service on SNA’s executive committee and chair of the SNA communications committee. He writes for numerous naval, maritime, and defense publications and chairs and presents at communications, naval, and maritime security conferences around the world. Robin McCasland, IABC Fellow, SCMP, is Senior Director of Corporate Communications for Health Care Service Corporation (HCSC). She leads the company’s communications team and the employee listening program, demonstrating to senior leaders how employee and executive communication add value to the business’s bottom line. Previously, Robin excelled in leadership roles in communication for Texas Instruments, Dell, Tenet Healthcare, and Burlington Northern Santa Fe. She has also worked for large and boutique HR consulting firms, leading major communication initiatives for various well-known companies. Robin is a past IABC chairman and has served in numerous association leadership roles for over 30 years. She was honored in 2023 and 2021 by Ragan/PR Daily as one of the Top Women Leaders in Communication. She’s also received IABC Southern Region and IABC Dallas Communicator of the Year honors. Robin is a graduate of The University of Texas at Austin and a Leadership Texas alumnus. Her own podcast, Torpid Liver (and Other Symptoms of Poor Communication), features guest speakers addressing timely topics to help communication professionals become more influential, strategic advisors and leaders. She resides in Dallas, Texas, with her husband, Mitch, and their canine kids, Tank and Petunia. George McGrath is founder and managing principal of McGrath Business Communications, which helps clients build winning corporate reputations, promote their products and services, and advance their views on key issues. George brings more than 25 years in PR and public affairs to his firm. Over the course of his career, he has held senior management positions at leading strategic communications and integrated marketing agencies including Hill and Knowlton, Carl Byoir & Associates, and Brouillard Communications. Caroline Sapriel, founder and Managing Partner of CS&A, brings over 30 years of specialized expertise in risk, crisis, and business continuity management to the table. A Fellow of the International Association of Business Communicators (IABC) and a recipient of the Gold Quill Award for her “10 Commandments of ...
    Más Menos
    1 h y 3 m
  • FIR #507: Should Nobody Really Ever Write with AI?
    Mar 30 2026
    Take a stroll through LinkedIn. You’ll find no shortage of posts stridently deriding the notion that anyone should ever use AI to write for them. While that case isn’t hard to make for professional writers, there are countless professionals in other fields who struggle with writing, never trained to be writers, yet now have to write everything from emails to reports as part of their jobs. Should they really sweat for hours over wording, time they could be devoting to the core areas of subject expertise, when AI can produce content that is cogent, clear, and direct? In this short mid-week episode, Neville and Shel look at the trends in using AI for writing, despite the plethora of opinions from the pundits. Links from this episode: Meet the Tech Reporters Using AI to Help Write and Edit Their StoriesMeet the Journalist Using AI to Write StoriesHow Journalists Feel About AIMuck Rack’s 2026 State of Journalism Report Finds 82% of Journalists Use AIAI Doesn’t Reduce Work—It Intensifies ItIs Writing with AI at Work Undermining Your Credibility?How We’re Using AIReview of ‘Using Artificial Intelligence in Academic Writing’Best Practices for the Effective Use of AI in Business WritingAI Tools for Business Writing5 Ways to Instantly Level Up Your Communication Using AI ToolsCharlene Li and Katia Walsh demonstrate the right way to build a book with AI help – Josh BernoffThe Truth About Writing a Book on AI The next monthly, long-form episode of FIR will drop on Monday, April 27. We host a Communicators Zoom Chat most Thursdays at 1 p.m. ET. To obtain the credentials needed to participate, contact Shel or Neville directly, request them in our Facebook group, or email fircomments@gmail.com. Special thanks to Jay Moonah for the opening and closing music. You can find the stories from which Shel’s FIR content is selected at Shel’s Link Blog. You can catch up with both co-hosts on Neville’s blog and Shel’s blog. Disclaimer: The opinions expressed in this podcast are Shel’s and Neville’s and do not reflect the views of their employers and/or clients. Raw Transcript Neville: Hi everyone and welcome to For Immediate Release episode 507. I’m Neville Hobson. Shel: And I’m Shel Holtz. And if you spend any time at all on LinkedIn, you’ll see the degree to which anti-AI sentiment is ramping up. A lot of it’s aimed at using AI for writing and how absolutely wrong that is. Yet just last week, on the same day, Wired Magazine and The Wall Street Journal both published articles on reporters using AI to help write and edit their stories. So today, let’s talk about using AI to write. Specifically, is it okay for employees to use AI to help them write for work? And my answer is not only is it okay for many employees, it might be one of the most genuinely useful things AI can do. Here’s the framing I would push back on. When we talk about AI writing assistants, we tend to picture a journalist or a marketer or a communications professional, someone whose craft is writing, it’s what they’re paid for, handing their keyboard over to a robot. And for those of us who are professional writers, that raises legitimate professional and ethical questions. But that’s not the population we’re talking about when we’re communicating AI adoption in most organizations. Think about who actually has to write at work. Engineers document processes. Product managers write status updates. Safety officers draft incident reports. Shel: Finance analysts compose budget justifications. Scientists write up findings for non-technical stakeholders. These are not people who chose their careers because they love writing. Writing is a tax they pay to do the work they actually care about. And many of them pay that tax really, really badly. The idea that a structural engineer should produce elegant prose unaided is the same logic as saying a communications director should coordinate the concrete mix for a construction project. We don’t expect that. So why do we expect every knowledge worker to be a competent writer? Muckrack’s 2026 State of Journalism report found that 82% of journalists, professional writers, people whose job this is, are now using at least one AI tool. That’s up from 77% the year before. If the people whose professional identity is tied to their writing are using AI tools, it shouldn’t surprise us that everyone else is too, or that they should. Now the research does tell us something important about how to use these tools. A University of Florida study of 1,100 professionals found that AI tools can make workplace writing more professional. But regular heavy use can undermine trust between managers and employees, particularly for relationship-oriented messages like praise, motivation, or personal feedback. The study found that employees are more skeptical when they perceive a supervisor is leaning heavily on AI for those kinds of communications. Now that’s a meaningful finding and it...
    Más Menos
    26 m
  • ALP 298: Build the business you want to own, not the one you hope to sell
    Mar 23 2026

    Most agency owners have read Built to Sell. But many have internalized the wrong lesson from it—fixating on that final chapter where the protagonist drives off into the sunset with a pile of cash, rather than the actual business-building advice throughout the book. The result is owners spending years building businesses optimized for a sale that may never happen, or that won’t deliver the outcome they’re imagining.

    In this episode, Chip and Gini discuss Chip’s “Build to Own” philosophy as a counterpoint to the built-to-sell mindset. The core principle: focus on creating a business that serves you today, not some hypothetical buyer tomorrow. This doesn’t mean you can’t or won’t sell—it means you stop treating the sale as the primary objective and start treating ownership as the thing you’re optimizing for right now.

    Chip breaks down the TMRW framework for thinking about what you want from your business: Time (how much you spend and what flexibility you have), Meaning (what gives you satisfaction—clients, team, impact), Rewards (financial outcomes that fund your life today and tomorrow), and Work (the actual role you’re crafting for yourself). Gini shares her decision to retire from speaking despite conventional wisdom saying agency owners should be out there raising their profile—because the anxiety wasn’t worth the marginal business benefit.

    The conversation tackles the uncomfortable reality that most agency owners counting on a sale to fund their retirement are likely building businesses that won’t command the multiple they’re hoping for. Meanwhile, owners who build businesses that throw off enough cash to fund retirement directly—while also being enjoyable to run—end up with something far more attractive to buyers when and if they do decide to sell.

    Gini tells the story of a friend who prepared five years in advance for a sale: removing himself from day-to-day operations, hiring a president to build culture, ensuring the business wasn’t founder-dependent. The result? An 18x multiple. But the episode’s point isn’t “here’s how to get a great sale”—it’s that you should make every decision through the lens of “would I still be happy with this if I never sold?” [read the transcript]

    The post ALP 298: Build the business you want to own, not the one you hope to sell appeared first on FIR Podcast Network.

    Más Menos
    20 m
  • FIR #506: Battle of the Bots!
    Mar 23 2026
    In this monthly long-form episode for March, Neville and Shel tackle a trio of interconnected themes reshaping the communications profession in the age of AI. The conversation opens with Anthropic’s top lawyer declaring that AI will destroy the billable hour. That thread leads naturally into JP Morgan’s controversial use of digital monitoring to verify junior bankers’ working hours, where Shel and Neville question whether surveillance technology can substitute for genuine managerial trust and engagement. The episode also examines Gartner’s widely circulated prediction that PR budgets will double by 2027 as AI search engines favor earned media. Shel delivers a detailed report on the escalating misinformation crisis, citing a 900% surge in global deepfake incidents and new research from the C2PA on content provenance standards. The episode closes with a discussion of Cloudflare CEO Matthew Prince’s prediction that bot traffic will exceed human traffic by 2027, and a sobering peer-reviewed study on how social bots hijack organizational messaging — research reported by Bob Pickard, who has experienced bot-driven attacks firsthand. Dan York also contributes a tech report on the state of the Fediverse and Mastodon, as well as on AI developments for WordPress. Links from this episode: AI will destroy the billable hour, says Anthropic’s top lawyerGartner predicts PR budgets will increase 2x by 20275 takes on Gartner’s new optimism for PR and earned media in the age of AIPR is back, baby — Gartner is predicting… [LinkedIn post by Lindsay Bennett]The Gartner claim that public relations and earned media budgets will double by 2027JPMorgan starts programme to monitor junior banker hours [Financial Times]FT Exclusive: The US bank has started to… [Financial Times LinkedIn post]Senator Bernie Sanders Discusses the Impact of AI on Privacy and Democracy with ClaudeLet’s Talk Keyboard Jamming and Why It Might Suggest Bigger Problems at WorkTelling Fact From Fiction With Online MisinformationOnline bot traffic will exceed human traffic by 2027, Cloudflare CEO saysPublic Relations & Organizational Communication [LinkedIn post by Bob Pickard]Social Bots as Agenda-Builders: Evaluating the Impact of Algorithmic Amplification on Organizational Messaging Links from Dan York’s Tech Report: Mastodon post by Eugen Rochko (@Gargron) — mastodon.socialMastodon — Decentralized social mediaHow to Generate a WordPress Theme with TelexTelex — AI-Assisted Authoring Environment for WordPressWordPress.com now lets AI agents write and publish posts, and moreYour AI agent can now create, edit, and manage content on WordPress.comEnable MCP tool access for AI agentsWordPress.com MCP prompt examples The next monthly, long-form episode of FIR will drop on Monday, April 27. We host a Communicators Zoom Chat most Thursdays at 1 p.m. ET. To obtain the credentials needed to participate, contact Shel or Neville directly, request them in our Facebook group, or email fircomments@gmail.com. Special thanks to Jay Moonah for the opening and closing music. You can find the stories from which Shel’s FIR content is selected at Shel’s Link Blog. You can catch up with both co-hosts on Neville’s blog and Shel’s blog. Disclaimer: The opinions expressed in this podcast are Shel’s and Neville’s and do not reflect the views of their employers and/or clients. Raw Transcript Neville:  Hi everyone, and welcome to the Forum Immediate Release podcast, long form episode for March, 2026. I’m Neville Hobson. Shel: And I’m Shel Holtz. Neville: As ever, we have six great stories to discuss and share with you, and we hope you’ll gain insight and enjoyment from our discussion. Perhaps you’ll want to share a comment with us once you’ve had a listen. We’d like that. Our topics this month range from AI in the end of the billable hour to Gartner’s predictions about PR budgets to monitoring work in the age of AI to newsrooms battling AI generated misinformation and more, including Dan York’s tech reports. Before we get into our discussion, let’s begin with a recap of the episodes we’ve published over the past month and some list of comments in the long form. In episode 502 for February, published on the 23rd of that month, we explored how rapidly accelerating technology is reshaping the communication profession from autonomous agents with attitudes to the evolving ROI of podcasting. We led with a chilling milestone moment, an autonomous AI coding agent that publicly shamed a human developer after he rejected its code contribution. A leader can build goodwill for days and lose it in seconds. In FIR 503 on the 2nd of March, we reported on the president of the IOC, that’s the International Olympic Committee, who had no answers to reporters’ questions and suggested on camera that someone on her communications team should be fired. We got comment on this, haven’t we, Shel? Shel: Boy, do we have comments on ...
    Más Menos
    1 h y 43 m
  • FIR #505: Social Media’s Big Shift
    Mar 17 2026
    In FIR #505, Neville and Shel dig into Hootsuite’s Social Media Trends 2026 report, which argues that social media is no longer just a communication channel — it’s morphing into a search engine, cultural radar, and real-time research tool. They explore what it means for communicators when younger audiences treat TikTok and Instagram as their primary discovery platforms, and when Google itself starts indexing social content. The conversation also tackles “fastvertising” — the growing pressure on brands to react to cultural moments within hours — and whether that speed actually translates to bottom-line results or just burnout. The discussion takes a provocative turn when Shel raises Ethan Mollick’s warning that public forums are being systematically overrun by machine-generated content, with research suggesting one in five accounts in public conversations may be automated. They weigh the AI paradox facing communicators: generative AI has become table stakes for social media production, yet 30% of consumers say they’re less likely to choose a brand whose ads they know were AI-created. Neville and Shel agree that social media can serve as both a publishing channel and a listening tool — but only if human-to-human communication can survive the rising tide of bot-generated noise. Links from this episode: Social Media Trends 2026 | HootsuiteThe 18 social media trends to shape your 2026 strategySferra Design video on Social Media Trends report | InstagramWorld-first social media wargame reveals how AI bots can swing electionsAI bot swarms threaten to undermine democracyB2B Social Media Trends and Predictions for 2026 The next monthly, long-form episode of FIR will drop on Monday, March 23. We host a Communicators Zoom Chat most Thursdays at 1 p.m. ET. To obtain the credentials needed to participate, contact Shel or Neville directly, request them in our Facebook group, or email fircomments@gmail.com. Special thanks to Jay Moonah for the opening and closing music. You can find the stories from which Shel’s FIR content is selected at Shel’s Link Blog. You can catch up with both co-hosts on Neville’s blog and Shel’s blog. Disclaimer: The opinions expressed in this podcast are Shel’s and Neville’s and do not reflect the views of their employers and/or clients. Raw Transcript: Shel: Hi everybody, and welcome to episode number 505 of For Immediate Release. I’m Shel Holtz. Neville: And I’m Neville Hobson. Social media might be going through its biggest change since the rise of the news feed, and it’s happening quietly. Platforms that started as places to connect with friends are increasingly acting like search engines, cultural sensors, and even market research tools. It’s been a while since Shel and I talked about social media on the podcast, and frankly, that’s partly because the conversation often feels repetitive. New platforms appear, algorithms change, someone declares the death of Twitter again. That’s the kind of format that we seem to be following. But every now and then, a report comes along that suggests something deeper is happening. Hootsuite’s new Social Media Trends 2026 report published last month argues that social media is no longer just a communication channel. It’s becoming something much broader — part search engine, part cultural radar, and part market research lab. Take search, for example. Younger users increasingly treat platforms like TikTok or Instagram as search tools. Instead of Googling “best coffee shop in London,” they search TikTok and watch short videos from real people recommending places to go. And now Google itself has started indexing Instagram posts and surfacing short-form social video in search results. The line between social media and search is starting to blur. At the same time, we’re seeing a strange tension around artificial intelligence. According to the report, most social media managers now use generative AI tools every day to write captions, brainstorm ideas, edit images or video. But audiences are increasingly suspicious of content that feels automated or synthetic. More than 30% of consumers say they’re less likely to choose a brand if they know its ads were created by AI. So brands are in a curious position. AI is becoming essential behind the scenes, but the content that performs best often needs to feel unmistakably human. And culturally, social media itself is fragmenting. The report points to what it calls Gen Alpha Chaos Culture — absurd memes, distorted audio, and intentionally chaotic editing styles that dominate TikTok among younger audiences. Meanwhile, older audiences — that’s you and me, Shel — are gravitating towards almost the opposite aesthetic: nostalgic references to the ’80s and ’90s, calming, cozy content, and even posts about slow living and digital detox. I do some of that, but I also do the other stuff too. So it’s hard to pigeonhole me, I have to tell you that. So ...
    Más Menos
    21 m