Maggie’s Farm

I had to laugh when I saw this headline and then read about why a platform like Substack is *just* like having your own platform…

“You need to have your own corner of the internet, a place where you can build a home, on your own land, with assets you control.


Our system gives creators ownership. With Substack, you have your own property to build on: content you own, a URL of your choosing, a website for your work, and a mailing list of your subscribers that you can export and take with you at any time.”

Substack is great. I read many newsletters there and thought about moving over some of my own stuff. However, it is a platform just like Facebook or X or TikTok. Yes, you can point your own domain name using a C Record there. That does not make it yours.

I pay for this server and while I do “rent” the domain name samharrelson.com, I’ve removed as many middlemen from the equation as I can.

True web independence is having a site / blog on a server that you ultimately control.

“We are now confident we know how to build AGI…”

That statement is something that should be exciting as well as a “woah” moment to all of us. This is big and you should be paying attention.

Reflections – Sam Altman:

We are now confident we know how to build AGI as we have traditionally understood it. We believe that, in 2025, we may see the first AI agents “join the workforce” and materially change the output of companies. We continue to believe that iteratively putting great tools in the hands of people leads to great, broadly-distributed outcomes.

We are beginning to turn our aim beyond that, to superintelligence in the true sense of the word. We love our current products, but we are here for the glorious future. With superintelligence, we can do anything else. Superintelligent tools could massively accelerate scientific discovery and innovation well beyond what we are capable of doing on our own, and in turn massively increase abundance and prosperity.

Beyond the Corporate Gloss: A Deeper Critique of Google’s 2024 Environmental Report

In reviewing Google’s 2024 Environmental Report, it’s hard not to be impressed by the sleek presentation, optimistic targets, and promises of a more sustainable future. But as someone who approaches environmental issues through the lenses of ecology, spirituality, and activism (and who respects the wisdom held by Indigenous communitie), we must ask ourselves: Is this report truly a step forward, or is it a carefully curated narrative that still falls short of meaningful transformation?

Below are some reflections and critiques that emerged as I dug deeper into Google’s latest sustainability claims. My hope is that these points inspire more honest conversations about corporate environmental responsibility, and encourage Google to become a force for genuine, not just performative, change. Google notes that this is the 10th year of their reporting, and while laudible, a decade is a long time to have not made much progress in the areas below.

1. More than a Numbers Game: Transparency and Context
Google’s report is filled with metrics: carbon offsets, renewable energy installations, and progress toward “24/7 carbon-free” ambitions. On the surface, this data sounds promising. Yet the numbers often come without the context that would allow us to evaluate their true impact. We need to know how these figures are changing over time, where and why setbacks occur, and how absolute emissions reductions are measured beyond short-term offsets. Without clear year-over-year comparisons, transparency in methodologies, and explanations for where goals haven’t been met, these metrics risk feeling more like strategic PR rather than a window into substantive progress.

2. A Holistic Ecological View—Not Just Carbon
In the ecological world, everything is interconnected—water usage, land stewardship, biodiversity, soil health, and species protection are all part of the larger puzzle. Too often, corporate sustainability efforts narrow their focus to carbon emissions. While that’s a crucial piece, it’s not the full story. The development of data centers, the sourcing of rare earth minerals for hardware, the water required for cooling, and the potential displacement of local communities or wildlife—these all have tangible ecological effects. Google’s report would be more authentic if it acknowledged these complexities. It’s not enough to claim net-zero this or carbon-free that or water-usage here; we need to know how their operations affect entire ecosystems and the countless living beings (human and non-human) who share those habitats.

3. Integrating Indigenous Knowledge and Perspectives
For millennia, Indigenous communities have developed rich, place-based knowledge systems that guide sustainable stewardship of land and resources. Their approaches aren’t just about preserving nature for posterity; they recognize the sacred interdependence of human life and the Earth. Indigenous environmental philosophies emphasize reciprocity, relational accountability, and long-term thinking—values that our high-tech era desperately needs. Yet, Google’s report barely touches on how local knowledge systems or Indigenous voices factor into its environmental strategies. True environmental leadership means not only incorporating Indigenous perspectives but also creating platforms where those communities can shape corporate policies and decision-making. A genuine partnership with Indigenous peoples would push beyond mere consultation toward co-creation of sustainability solutions.

4. The Moral and Spiritual Dimension of Environmental Care
Sustainability isn’t just a business metric; it’s a moral imperative. Many faith traditions and spiritual frameworks teach that the Earth is not merely a resource to be exploited, but a sacred gift that we are entrusted to protect. When companies like Google talk about sustainability without acknowledging the deeper moral currents—respect for Creation, the call to love our neighbors (human and nonhuman), and the need to protect the vulnerable—they risk missing the heart of the matter. Earth care is not just about polished reports; it’s a sacred calling. If Google truly wants to lead, it must recognize and uphold this responsibility as part of its corporate identity.

5. Justice, Equity, and Community Engagement
Climate change is not an equal-opportunity crisis—frontline communities, often Indigenous peoples and people of color, bear a disproportionate burden of environmental harm. There’s a human face to pollution, species loss, and extraction, and companies have a moral duty to see it. Yet the report often focuses inward—on Google’s own campuses, energy grids, and supply chains—without sufficiently addressing how it will engage with and support communities directly affected by its operations. Where is the acknowledgment of environmental justice? Where are the stories of local partnerships, community-based mitigation plans, or compensation for environmental damage? Until these voices and their realities are meaningfully included, sustainability efforts risk becoming top-down strategies instead of inclusive, equitable solutions.

6. From Incremental to Transformative Change
Corporate environmental narratives often hinge on incremental progress: small steps toward greener operations, a handful of offset projects, a few solar panels here and there. But a company with Google’s resources could champion systemic changes that transcend the status quo. It could lead research in scalable regenerative practices, revolutionize supply chains to eliminate environmental harm, or fund open-access environmental science tools that empower others. By fully embracing the call for systemic transformation, Google could serve as a beacon of hope, paving the way for a truly sustainable economy that values regeneration over extraction, and community well-being over profit margins.

Envisioning a More Genuine Path Forward
Critiquing a sustainability report may seem like a small gesture, but honest criticism matters. It’s a reminder that we must look beyond the corporate gloss to see the true health of our planet—and to hold powerful entities accountable. The world needs leaders who understand that ecological well-being, moral responsibility, Indigenous wisdom, and social justice are interwoven strands of the same tapestry.

Google’s 2024 Environmental Report certainly isn’t the worst corporate sustainability document out there in the tech space. But given the company’s global influence, wealth, and technological prowess, “not the worst” isn’t nearly good enough. We deserve, and the Earth demands, better. True environmental leadership would blend hard data with moral courage, incorporate ancestral wisdom, support vulnerable communities, and invest in regenerative systems that honor both people and the planet. That’s the vision we need, and it’s the vision that a company like Google could help realize, if it dared to do more than just follow the colonialist corporate script.

Facial Recognition Tech in Smart Glasses

Law enforcement and the military have had this capability for a while via Clearview, but it’s (also) scary to see it being implemented outside of those domains…

Someone Put Facial Recognition Tech onto Meta’s Smart Glasses to Instantly Dox Strangers:

A pair of students at Harvard have built what big tech companies refused to release publicly due to the overwhelming risks and danger involved: smart glasses with facial recognition technology that automatically looks up someone’s face and identifies them. The students have gone a step further too. Their customized glasses also pull other information about their subject from around the web, including their home address, phone number, and family members.

A Priesthood of Pollution

Lots to ponder here about human consciousness, human angst, and the coming torrent of AI bots fueled by corporate profit at the expense of polluting the digital ecology we’ve built over the last few decades.

It is by no means currently pristine, but pollution always comes with capitalist initiatives, and AI bots are about to transform so much of what we know about everyday life, leaving behind much more artificial pollution than we can ponder now…

These AI agents are building ‘civilizations’ on Minecraft | Cybernews:

Run by California-based startup Altera, the project had AI agents collaborating to create virtual societies complete with their own governmental institutions, economy, culture, and religion.

Altera said it ran simulations on a Minecraft server entirely populated by autonomous AI agents “every day” and the results were “always different.”

In one simulation, AI agents banded together to set up a market, where they agreed to use gems as a common currency to trade supplies – building an economy.

Curiously, according to the company, it was not the merchants who traded the most but a corrupt priest who started bribing townsfolk to convert to his religion.

Good read on the topic with some predictions about AI bots from Ted Gioia here as well

Yale’s Dining Halls Using AI to Minimize Food Waste

Waste watchers: Using AI to minimize food waste in Yale’s dining halls | YaleNews:

Nationally, an estimated 30 to 40% of the food supply is wasted, according to the U.S. Department of Agriculture. And it’s not only the food itself that goes to waste, but the land, water, labor, and energy used to produce it. Food waste is linked to roughly 8% of global emissions. At Yale, food purchases currently account for about 12,500 metric tons of carbon dioxide equivalent, according to the Office of Sustainability.

OpenAI’s Strawberry

Happening quickly…

Exclusive: OpenAI working on new reasoning technology under code name ‘Strawberry’ | Reuters:

The document describes a project that uses Strawberry models with the aim of enabling the company’s AI to not just generate answers to queries but to plan ahead enough to navigate the internet autonomously and reliably to perform what OpenAI terms “deep research,” according to the source. This is something that has eluded AI models to date, according to interviews with more than a dozen AI researchers.

AI’s Awful Energy Consumption

Be mindful and intentional with technology tools…

Google and Microsoft report growing emissions as they double-down on AI : NPR:

“One query to ChatGPT uses approximately as much electricity as could light one light bulb for about 20 minutes,” he says. “So, you can imagine with millions of people using something like that every day, that adds up to a really large amount of electricity.”

Thrive AI Health from OpenAI Founder

Fascinating read from Sam Altman and Ariana Huffington here as they release Thrive AI Health, which will be something of an AI coach backed by OpenAI / ChatGPT. Combining this with Apple Intelligence is going to be interesting…

AI-Driven Behavior Change Could Transform Health Care | TIME:

Using AI in this way would also scale and democratize the life-saving benefits of improving daily habits and address growing health inequities. Those with more resources are already in on the power of behavior change, with access to trainers, chefs, and life coaches. But since chronic diseases—like diabetes and cardiovascular disease—are distributed unequally across demographics, a hyper-personalized AI health coach would help make healthy behavior changes easier and more accessible. For instance, it might recommend a healthy, inexpensive recipe that can be quickly made with few ingredients to replace a fast-food dinner.

My Beginner’s Guide to Artificial Intelligence

A client reached out and asked if I could put together a “beginner’s guide to AI” for them and their team a little while ago. I thought long and hard on the topic as I have so much excitement for the possibilities but so much trepidation about the impacts (especially to individuals in careers that will be threatened by the mass adoption of AI). Apple’s announcement this month that they are infusing iPhones with ChatGPT intelligence only drives that home. We are in a time of transition, and I want my own clients but anyone running a business or working in a sector that will be affected (which is every sector) to be prepared or at least mindful of what’s coming.

So, I put this together in a more expanded format with charts, examples, etc, but this is a good outline of the main points. I thought it would maybe help some others, and my client graciously said I could post this as a result. Let me know if you have any thoughts or questions!

Artificial Intelligence (AI) is a topic that’s constantly buzzing around us. Whether you’ve heard about it in the context of ChatGPT, Apple Intelligence, Microsoft’s Copilot, or self-driving cars, AI is transforming the way we live, work, and even think. If you’re like many people, you might be on the fence about diving into this technology. You might know what ChatGPT is but aren’t quite sure if it’s something you should use. Let’s break down the benefits and costs to help you understand why AI deserves your attention.

The Benefits of Embracing AI

Efficiency and Productivity

One of the most compelling reasons to embrace AI is its ability to enhance efficiency. In our busy lives, whether managing businesses, marketing campaigns, or family time, finding ways to streamline tasks can be a game-changer. AI can help automate mundane tasks, organize your day, and even draft your emails. Imagine having a virtual assistant who never sleeps, always ready to help you.

For instance, AI-powered scheduling tools can help you manage your calendar more effectively by automatically setting up meetings and sending reminders. This means less time spent on administrative tasks and more time dedicated to what truly matters – growing your business, strategizing your marketing efforts, or spending quality time with your family.

Personalization

AI can personalize experiences in ways we’ve never seen before. For marketers, this means creating targeted campaigns that resonate on a personal level. However, AI can analyze data to understand preferences, behaviors, and patterns, allowing for a more customized approach in almost any field.

Imagine being able to offer each customer or client a unique experience that caters to their needs and interests. This personalized approach can significantly enhance engagement and loyalty. In marketing, AI can help create highly targeted content that speaks directly to the needs and interests of your audience, increasing engagement and conversion rates.

Access to Information

The vast amounts of data generated daily can be overwhelming whether you’re solo, on a team, or working in the C-Suite. AI can sift through this information and give you the insights you need. Whether you’re researching a new marketing strategy, preparing for a presentation, or just curious about a topic, AI can help you find relevant and accurate information quickly.

Think about how AI-powered search engines and research tools can simplify the process of gathering information. Instead of sifting through endless articles and papers, AI can provide the most pertinent sources, saving you time and effort. This is especially valuable in professional settings where timely and accurate information is crucial.

Creativity and Innovation

AI isn’t just about number-crunching; it’s also a tool for creativity. Tools like ChatGPT or Copilot or Gemini or Claude can help brainstorm ideas, generate creative content, and even compose poetry. It’s like having a creative partner who can help you think outside the box and explore new possibilities.

As someone who values creativity, imagine having an AI that can help you brainstorm new marketing ideas, create engaging content for your campaigns, or even assist in writing your next blog post. AI can inspire new ways of thinking and help you push the boundaries of your creativity. It’s not just for writing high school papers, but there are very tangible ways to use AI to spur new insights and not just “do the work for you.”

The Costs and Considerations

Privacy Concerns

I’m a huge privacy and security nerd. I take this very seriously with my own personal digital (and non-digital) life as well as that of my family members. One of the main concerns people have with AI is privacy. AI systems often rely on large amounts of data, some of which might be personal. It’s essential to be aware of what data you’re sharing and how it’s being used. Look for AI tools that prioritize data security and transparency if you’re using AI in any sort of corporate or work-related output. 

For instance, when using AI tools, always check their privacy policies and opt for those that offer robust data protection measures. Be mindful of the information you input into these systems and ensure that sensitive data is handled appropriately. Balancing the benefits of AI with the need to protect personal privacy is crucial.

Dependence and Skill Degradation

There’s a valid concern that relying too much on AI could lead to a degradation of our skills. Just like relying on a calculator too much can weaken basic arithmetic skills, leaning heavily on AI might impact our ability to perform specific tasks independently. It’s important to strike a balance and use AI as a tool to enhance, not replace, our capabilities. As someone who has worked in education with middle and high schoolers, I especially feel this need to train and model this balance.

Consider using AI as a complement to your existing skills rather than a crutch. For example, while AI can help draft emails or create marketing strategies, reviewing and personalizing these outputs is still important. This way, you maintain your proficiency while benefiting from AI’s efficiency. AI systems are constantly being developed and will continue to improve, but there are very real examples of businesses and even attorneys and physicians using AI output that was later proven to be false or misleading. Be wise.

Ethical Considerations

AI raises a host of ethical questions. How should AI be used? What are its implications for decision-making processes? These questions are close to my heart as someone interested in theology and ethics. It’s crucial to consider the moral dimensions of AI and ensure that its development and deployment align with our values.

Engage in discussions about AI ethics and stay informed about how AI technologies are being developed and used. Advocate for ethical AI practices that prioritize fairness, transparency, and accountability. By doing so, we can help shape a future where AI benefits everyone.

We are constantly hearing stats about the number of jobs (and incomes) that AI replace in 1, 5, or 10 years. I do believe we are in for a societal shift. I do not want people to suffer and lose their jobs or careers. However, AI is not going away. How can you or your business manage that delicate balance in the most ethical way possible?

Economic Impact

AI is reshaping industries, which can lead to job displacement. While AI creates new opportunities, it also means that some roles may become obsolete. Preparing for these changes involves continuous learning and adaptability. It’s important to equip ourselves and our teams with the skills needed in an AI-driven world.

Promote the development of skills that are complementary to AI, such as critical thinking, creativity, and emotional intelligence. Encourage yourself or your team to pursue fields that leverage AI technology, ensuring they remain competitive in the evolving job market. Emphasizing lifelong learning will help individuals adapt to the changes brought about by AI.

Embracing AI: A Balanced Approach

AI is a powerful tool with immense potential, but it also has its share of challenges. As we navigate this new landscape, it’s essential to approach AI with a balanced perspective. Embrace the benefits it offers, but remain vigilant about the costs and ethical implications.

For those still hesitant, I encourage you to experiment with AI tools like ChatGPT. Start small, see how it can assist you in your daily tasks, and gradually integrate it into your workflow. AI isn’t just a trend; it’s a transformation that’s here to stay. By understanding and leveraging AI, we can better prepare ourselves and our businesses for the future.

Explore AI Tools

Begin by exploring AI tools that can assist you in your daily activities. For example, try using ChatGPT for drafting emails, creating marketing strategies, or brainstorming ideas. Experiment with AI-powered scheduling tools to manage your calendar more efficiently.

Educate Yourself

Stay informed about AI developments and their implications by reading articles, attending webinars, and participating in discussions about AI. Understanding the technology and its potential impact will help you make informed decisions about its use. As always, reach out to me if you have any questions.

Balance AI Use with Skill Development

While leveraging AI, ensure that you continue to develop your own skills. Use AI as a supplement rather than a replacement. For example, review and personalize AI-generated content to maintain your proficiency. Find online webinars that are geared towards AI trainings or demos that you can attend or review. There’s plenty of videos on YouTube, but be wise and discerning as your attention is more valuable than quality content on many of those channels. 

Advocate for Ethical AI

Engage in conversations about AI ethics and advocate for practices that prioritize fairness, transparency, and accountability. Stay informed about how AI technologies are being developed and used, and support initiatives that align with your values. Whatever your industry or profession, there’s room (and economic incentive) for conversations about ethics in the realm of AI.

Prepare for the (YOUR) Future

Encourage yourself or your team to develop skills that complement AI technology. Promote critical thinking, creativity, and emotional intelligence. Emphasize the importance of lifelong learning to adapt to the evolving job market. Critical thinkers will be the key decision makers in 2034 100x more than they are today in 2024.

Final Thoughts

Artificial Intelligence is a transformative force that’s reshaping our world in profound ways. By understanding and embracing AI, we can unlock new levels of efficiency, personalization, creativity, and innovation. 

However, navigating this landscape with a balanced perspective is crucial, considering the costs and ethical implications. Be wise. Be kind. Be efficient. The future feels uncertain and this is technology that will literally transform humanity more than the internet, more than electromagnetism, more than automobiles… we are entering a new age in every facet of our lives both personally and professionally. I don’t want to scare you, but I do want you and your team to be prepared.

For those still on the fence, I encourage you to take the plunge and explore AI’s potential. Start small, experiment with different tools, and see how they can enhance your daily activities. AI isn’t just a passing trend; it’s a revolution that’s here to stay. By leveraging AI wisely, we can better prepare ourselves and our businesses for the future.

And as always… stay curious!

Book Review: John Longhurst’s Can Robots Love God and Be Saved?

As someone with a rich background in the cutting-edge side of marketing and technology (and education) and someone often referred to as a futurist but is fascinated with ethical and theological impacts and contexts, I found John Longhurst’s “Can Robots Love God and Be Saved? (CMU Press 2024) to be a fascinating exploration of the convergence between cutting-edge technology, ethical considerations, and theological inquiry. This book speaks directly to my passions and professional experiences, offering a unique perspective on the future of faith in a rapidly evolving world where concepts such as artificial intelligence (and AGI) must be considered through both technological and theological lenses. 

A seasoned religion reporter in Canada, John Longhurst tackles various topics that bridge faith and modern societal challenges. The book is structured into sections that address different aspects of faith in contemporary life, including mental health, societal obligations, and the intriguing possibilities of artificial intelligence within religious contexts. Those are constructed out of interviews and perspectives from Longhurt’s interviews with a wide variety of cast and characters.

Longhurst discusses the ongoing challenges many face with mental illness and the role faith communities play in providing support. This aligns with my work in consulting and education, emphasizing the need for understanding and empathy in addressing situations such as mental health issues, whether in the classroom or the broader community. He also delves into the discussion on Christians’ duty to pay taxes and support societal welfare, raising essential questions about the practical application of faith from various personas and perspectives. I found this particularly relevant when contemplating the intersection of personal beliefs and civic responsibility, echoing ethical marketing practices and corporate social responsibility principles.

Exploring the deep bonds between humans and their pets, Longhurst touches on the theological implications of animals in heaven. This can be a fascinating topic in environmental science discussions, highlighting the interconnectedness of all life forms and reflecting on how technology (like AI in pets) might change our relationships with animals. The book also delves into ethical concerns about government surveillance from a religious standpoint, providing an excellent case study for understanding the balance between security and privacy rights—a crucial consideration in both marketing and technology sectors where data privacy is paramount.

One of the most thought-provoking sections of the book delves into AI’s potential role in religious practices. Longhurst’s exploration of whether robots can participate in spiritual activities and even achieve salvation is a direct intersection of my interests in technology and ethics. It raises profound questions about the future of faith, challenging traditional theological boundaries and offering a glimpse into future innovations in religious practice.

Longhurst also examines how religious communities can address the loneliness epidemic, which I found particularly engaging. The sense of belonging and support provided by faith groups is mirrored in the need for community in education and the workplace. Technology, mainly social media and AI, can play a role in mitigating loneliness, but it also highlights the need for genuine human connections. That’s also one of my motivators for exploring when setting up a marketing strategy: How does this product/service/technology help establish more genuine human connectivity?

Additionally, the book ponders the existence of extraterrestrial life and its implications for religious beliefs. This speculative yet fascinating topic can engage students in critical thinking about humanity’s place in the universe, much like futuristic marketing strategies encourage us to envision new possibilities and innovations. This is a hot topic, with other books such as American Cosmos making many “must read” lists this year, along with general interest in extraterrestrial / non-human intelligence / Unidentified Aerial Phenomenon (UAP) / Non-Human Intelligence (NHI) very much in cultural conversations these days.

Longhurst’s exploration of AI and its potential spiritual implications is particularly compelling from a marketing and technology perspective. As someone who thrives on being at the cutting edge, this book fuels my imagination about the future intersections of technology and spirituality. The ethical questions raised about AI’s role in religious practices are reminiscent of the debates we have in marketing about the ethical use of AI and data analytics.

The work is a thought-provoking collection that challenges readers to consider the evolving role of faith amidst technological advancements. Longhurst’s ability to tackle complex and often controversial topics with nuance and empathy makes this book a valuable resource for educators, faith leaders, technologists, and marketers alike. It provides a rich tapestry of discussions that can be seamlessly integrated into lessons on environmental science, ethics, technology, and even literature in a succinct and “quick-read” fashion.

Can Robots Love God and Be Saved?” is a compelling exploration of how faith intersects with some of the most pressing issues of our time. It is a fascinating read for anyone interested in understanding the future of spirituality in a world increasingly shaped by technology based on first-hand considerations rather than a purely academic or “one-sided” perspective. For those of us on the cutting edge, whether in marketing, technology, or education, this book offers a profound and thought-provoking look at the possibilities and challenges ahead.

Good read!

AI Video Generators

OpenAI’s Sora is impressive but the amount of text-to-video AI generators we’re seeing released (especially from China) points to a very real moment that we all need to pause and reflect upon. The coming year (I would’ve said the coming 2-3 years back in March) is going to be fascinating, haunting, and challenging all at once…

Introducing Gen-3 Alpha: A New Frontier for Video Generation:

Gen-3 Alpha is the first of an upcoming series of models trained by Runway on a new infrastructure built for large-scale multimodal training. It is a major improvement in fidelity, consistency, and motion over Gen-2, and a step towards building General World Models.

More from Runway’s X account here.

Accelerationism: What Are We Doing to Ourselves?

Here’s your word for today as Apple’s WWDC looks to include an announcement of a major partnership with OpenAI (the folks behind ChatGPT) to make Siri much closer to an artificial intelligence (or “Apple Intelligence” as the marketing goes) assistant.

Accelerationism.

It’s a term that’s been used in the tech world for years, but the mindset (mind virus?) has really reached new levels in the post-ChatGPT 4 era that we now live in before what feels like an imminent release of something even more powerful in the coming months or years.

Here’s an article from 2017 about the term accelerationism and accelerationists: 

Accelerationism: how a fringe philosophy predicted the future we live in – The Guardian: 

Accelerationists argue that technology, particularly computer technology, and capitalism, particularly the most aggressive, global variety, should be massively sped up and intensified – either because this is the best way forward for humanity, or because there is no alternative. Accelerationists favour automation. They favour the further merging of the digital and the human. They often favour the deregulation of business, and drastically scaled-back government. They believe that people should stop deluding themselves that economic and technological progress can be controlled. They often believe that social and political upheaval has a value in itself.

With my mind heavy on what the Apple / OpenAI partnership might look like before WWDC starts in just a few minutes (it feels like this could be an important moment for historical events), Ted Gioia made this thought-provoking post on the realization that we are doing to ourselves what Dr. Calhoun did to his poor mice (unknowingly) in the 1960’s famous Universe 25 experiment.

It’s worth your time to read this and ponder our own current situation.

Is Silicon Valley Building Universe 25? – by Ted Gioia:

Even today, Dr. Calhoun’s bold experiment—known as Universe 25—demands our attention. In fact, we need to study Universe 25 far more carefully today, because zealous tech accelerationists—that’s now a word, by the way—aim to create something comparable for human beings.What would you do if AI took care of all your needs?

After being in the classroom for the last three years of “post-Covid” education and seeing how many young people are absolutely struggling with mental health (and how little schools of any sort, from public to private such as the ones where I taught, are doing to help them), it’s shocking that we’ll send stocks soaring on big tech news today that will make our swipes and screen time increase and lead us further down the primrose path of a future of disconnected violence and mental health disaster.

New iPhones Get 5 Year Support

Now add in right-to-repair principles and more ethical mineral procurement (for batteries etc compared to the current terrible conditions and practices) and I’ll be happy!

Apple will update iPhones for at least 5 years in rare public commitment | Ars Technica:

Apple has taken a rare step and publicly committed to a software support timeline for one of its products, as pointed out by MacRumors. A public regulatory filing for the iPhone 15 Pro (PDF) confirms that Apple will support the device with new software updates for at least five years from its “first supply date” of September 22, 2023, which would guarantee support until at least 2028.

Anxious Generation Study

Ted’s entire newsletter is a worthy read here, but this part about new research indicating that the current genertion of young people growing up in a phone-based culture (globally) is doing real harm and damage. It makes me think back to the tobacco industry trying to pretend that cigarettes don’t hurt people or the petroleum companies hiding the neurological effects of lead-infused gasoline and so on…

Crisis in the Culture: An Update – by Ted Gioia:

Haidt declared victory on social media: “There are now multiple studies showing that a heavily phone-based childhood changes the way the adolescent brain wires up, in many ways including cognitive control and reward valuation.”

We still need more research. But we can already see that we’re dealing with actual physiological decline, not just pundits’ opinions.

At this point, the debate isn’t over whether this is happening. Instead we now need to gauge the extent of the damage, and find ways of protecting people, especially kids.

AI and Bicycle of the Mind

I don’t have the same optimism that Thompson does here, but it’s a good read and worth the thought time!

The Great Flattening – Stratechery by Ben Thompson:

What is increasingly clear, though, is that Jobs’ prediction that future changes would be even more profound raise questions about the “bicycle for the mind” analogy itself: specifically, will AI be a bicycle that we control, or an unstoppable train to destinations unknown? To put it in the same terms as the ad, will human will and initiative be flattened, or expanded?

Why I am Using a Light Phone

I have lots more to say about this, but I wanted to share this vital part of a recent article about “dumbphones” in The New Yorker. I’ve been attempting to be much more deliberate about using technology and devices, especially in front of my children and students.

The Light Phone (and Camp Snap camera) have been a significant part of that effort. I’ve been in love with the Light Phone since converting from an iPhone earlier this year.

The Dumbphone Boom Is Real | The New Yorker:

Like Dumbwireless, Light Phone has recently been experiencing a surge in demand. From 2022 to 2023, its revenue doubled, and it is on track to double again in 2024, the founders told me. Hollier pointed to Jonathan Haidt’s new book, “The Anxious Generation,” about the adverse effects of smartphones on adolescents. Light Phone is receiving increased inquiries and bulk-order requests from churches, schools, and after-school programs. In September, 2022, the company began a partnership with a private school in Williamstown, Massachusetts, to provide Light Phones to the institution’s staff members and students; smartphones are now prohibited on campus. According to the school, the experiment has had a salutary effect both on student classroom productivity and on campus social life. Tang told me, “We’re talking to twenty to twenty-five schools now.”

The Museum of Me

The Museum of You – Herbert Lui:

I see a lot of discussion on how people miss blogs, and RSS, and internet culture before what we call Web 2.0 (social media, platforms, ecommerce, etc.) came along and wiped it away. 

The best way to pay homage is to bring it back—to set up our own blogs that we control, to preserve our own libraries of content in multiple places so they don’t disappear with social media, to actively document our lives the way we miss and the way we would want to be remembered. We can choose a responsibility, every day, to collect the best of what came before us, to embody it, and to preserve it by sharing its charms with other people.

Much agreed, and this is one of the reasons I’ve kept my own blog and podcast here since 2006. I thought back then, “What if these awesome new tools like MySpace (or early Twitter) somehow go away or fall into the hands of the wrong leaders?” 

I read previous posts and thoughts here occasionally and marvel at how naive, bold, brave, or afraid I was at various points in my life. Now looking back on this Museum of Me, I can glimpse previous iterations of my own self and perceptions and not just remember but learn. 

Blogs like this, however silly they may seem in the face of social media apps, are powerful places!

OpenAI’s Lens on the Near Future

Newton has the best take I’ve read (and I’ve read a lot) on the ongoing OpenAI / Sam Altman situation… worth your time:

OpenAI’s alignment problem – by Casey Newton – Platformer:

At the same time, though, it’s worth asking whether we would still be so down on OpenAI’s board had Altman been focused solely on the company and its mission. There’s a world where an Altman, content to do one job and do it well, could have managed his board’s concerns while still building OpenAI into the juggernaut that until Friday it seemed destined to be.

That outcome seems preferable to the world we now find ourselves in, where AI safety folks have been made to look like laughingstocks, tech giants are building superintelligence with a profit motive, and social media flattens and polarizes the debate into warring fandoms. OpenAI’s board got almost everything wrong, but they were right to worry about the terms on which we build the future, and I suspect it will now be a long time before anyone else in this industry attempts anything other than the path of least resistance.

AI Assistants and Education in 5 Years According to Gates

I do agree with his take on what education will look like for the vast majority of young and old people with access to the web in the coming decade. Needless to say, AI is going to be a big driver of what it means to learn and how most humans experience that process in more authentic ways than currently available…

AI is about to completely change how you use computers | Bill Gates:

In the next five years, this will change completely. You won’t have to use different apps for different tasks. You’ll simply tell your device, in everyday language, what you want to do. And depending on how much information you choose to share with it, the software will be able to respond personally because it will have a rich understanding of your life. In the near future, anyone who’s online will be able to have a personal assistant powered by artificial intelligence that’s far beyond today’s technology.

Paper Airplanes

I love incorporating paper airplanes into my classroom lessons on dynamics, flight, movement, gravity… the list goes on and on. They’re so applicable to so many scientific principles but also appeal to the curious nature inside all of us that loves to fold and learn…

History of the Paper Airplane: Paper Flight Technology Inspires Drones:

“The magic of a paper airplane is that all of these little flight corrections are happening continuously throughout its flight,” Ristroph says. “The plane is hanging under a vortex that is constantly swelling and shrinking in just the right ways to keep a smooth and level glide.”

Thinking About Screentime

I’ve become much more of a book person as I’ve gotten older. Also, notebooks. That would seem quizzical to my younger self that reveled in every new productivity and reading app released on iOS or Android as I combed through blogs, subreddits, and Twitter lists, looking for the latest and greatest note-taking app.

Alas, getting old is interesting.

Screentime is definitely something that’s been on the front of my mind for the last few decades as I’ve welcomed children into this world (including Lily as of August 1!) and young people ranging from 12 to 18 into my classrooms. 

I plan to read this book, so I’m using this as a space-saver for myself to return to when I’m done (and in the middle of the school year).

Screentime is a fascinating cultural concept. The amount of “screentime” we actually consume is lower than it’s ever been (no, really). But is the measurement of “time” really what we should be focused on or worried about?

Regardless, my students will still have their devices in the “off” mode, and we’ll focus on the great ideas with our brains, pen/cil, paper, and each other’s voices like we’ll continue to not have devices on during dinners or downstairs time here in our home…

A Different Way to Think About Screentime:

Parents have a hard time when they don’t know something. I’ve written this elsewhere, but I think one of the basic things that underlies a lot of the book bannings and pronoun panics from parent-activists on the far-right is the very simple fact that parents don’t know what their kids do all day. My daughter Maeve is 7, and I volunteered this spring to help with a field trip for her first-grade class. The bus was late, and so I ended up just sitting in her classroom for about 45 minutes while the day went on as usual. Maeve is very talkative, and she loves telling us stories about her day, but it wasn’t until I sat in that classroom that I realized how little I actually knew about what the ordinary beats of that day were like, what the social dynamics were, what kind of job her wonderful teacher — hello, Mr. Diego Fernandez — is tasked with doing.