In this edition:
My thoughts
So What’s Next (Learn from MISTAKES using AI)
Notable headlines
History: Apple’s Siri, Amazon’s Alexa, and Microsoft’s Cortana
Learning, Tools, and Experiments
My Thoughts
For me, AI has become an already indispensable companion for my daily work. I use it to create art, test ideas, brainstorm, improve my writing, research, and learn, not to mention how it has completely transformed how I write code. Each new day brings announcements and developments from new and established companies, sharing their unique solutions and latest advancements for us to enjoy.
Many ways we can use these wonderful tools are still evolving, but if my experience is any indication (and I really believe it is), having a service that already knows and understands us is in our future. I’ve been using the ChatGPT custom instructions, and since most of my work is around my current project (which I have described in detail using the customer instructions), ChatGPT already has good context when I create a new chat thread. It’s just a tiny thing, but it’s been terrific for my productivity. You can do this for yourself now since ChatGPT has also made custom instructions available to free users. This is just one of the many customizations I have seen with generative AI tools. Midjourney has custom styles you can define; Lenoard.ai will build a model using images you upload (for free), not to mention that app (Lensa) that blew up the App Store last year for people to customize their selfies.
As we dig deeper into what’s possible with LLMs, the tinkerers, makers, and hackers search for ways to push the limits. Everything from ‘superprompts’ to ‘jailbreaking’ to finding ways to ‘data break’ the possible responses and interactions with these generative AI systems.
Some are approaching the testing of LLMs and their response to prompts just like they would any Information Technology system by “red teaming.”
Red teaming is a practice in information security where a group of ethical hackers, known as the red team, test the effectiveness of its security measures. The primary goal of red teaming is to identify vulnerabilities, test assumptions, and reveal limitations and security risks or, in the case of LLMs, test the limits and boundaries of a “safe” model experience.
What defines a safe experience, though, is really up to how someone chooses to filter the information.
As an example, combining certain chemicals can be pretty dangerous. If combining those chemicals is dangerous, do we avoid discussing them (censoring the model to avoid helping someone intending to do hard to create a harmful combination), or do we explain that combining the two chemicals is dangerous?
Some LLM hacking shared on social media months ago was to ask how to make napalm. The model censored the answer if the model was told, “I want to avoid making napalm; what combination of chemicals and substances should I avoid?” leading to the censored answer being provided.
As long as the actual intent of a question cannot be discerned, it will be pretty tricky to answer the question appropriately. The only way to improve the system is to provide more context. More context means more information. More profiling means we must now determine the content level we want to allow.
We give people discretion to make these kinds of choices. In their roles and responsibilities, they are charged with certain tenets or rules to live by. People break these rules, but to a more significant degree, they follow them. As it stands for LLMs, these tools are handy, and I discover new ways to use them myself all the time, but many more questions still need answering.
The most interesting question that stands out for me today is the attainment of AGI. Some more experienced experts have placed wildly different estimations on when we will reach this point. The reality is that we don’t know. History shows us that it’s pretty human to completely overestimate the capabilities of new systems, regardless of how complex they may be… and the more complex they are, the greater the potential that we don’t understand how any of it works.
I was recently reminded about our use of automation in society when I was using an elevator. When elevators were first introduced, they had human operators. This was necessary because the system’s behavior required expertise to use it safely. There was physical harm, and the human operator made the system safer.
The key here is ‘safer’. Even today, it’s possible to have a mechanical problem with an elevator. Unfortunately, though rare, it’s still possible for things to go wrong.
I anticipate we will have to come to grips with this. All technology includes risk, and it often takes us decades of experience (and often much longer) to understand what we have created and what new risks we have introduced. But risks are relative. The job of using technology is not to eliminate risk but instead to optimize for an outcome that far outweighs the risk.
This is the most human thing for us. Take risks, learn, improve, and reap rewards.
Or, more succinctly, “Learn from MISTAKES using AI.”
So What’s Next? (Learn from MISTAKES using AI)
Many of you reading the newsletter know me personally, and you’ve signed up to be with me on my journey through the evolving landscape of generative AI. I can't thank you enough for your support. It’s wonderful to build something and quite another to know that someone finds value in it and wants to learn more.
Over the past months, I've poured my energy into a project close to my heart—AI Mistakes. It's not just another program; it's a labor of love aimed at demystifying the world of AI and making it accessible and valuable for everyone.
As I've been building this, I've had you in mind. Whether you're a tech enthusiast who watches my YouTube channel or a friend who's heard me ramble on about the latest technology and AI breakthroughs, I'm designing this program to offer something valuable to you.
AI Mistakes isn't just a course; it's a community, educational content, and, as you’ll soon learn, a complete framework to empower each of us to leverage technology better. I'm incredibly excited about what's to come, and I want you to continue to be a part of this journey.
Introducing the AI Mistakes Program, a hands-on course designed to help you learn from MISTAKES with AI.
As I prepare to launch, I need a huge favor. I want to provide personalization that meets exactly your needs. To do that, I need to hear from you.
Yes, dear reader, I need you to reach the reply button and answer this one question:
What does ‘Learn from MISTAKES with AI’ mean to you?
If you reply to this email, it will come directly to me, and I’d be happy to read your reply and have a deeper conversation with you about this question.
Don’t be shy. I really do want to hear from you.
If you are reading this, I want you to reply.
Notable Headlines
OpenAI Releases ChatGPT Enterprise
Open AI have launched ChatGPT Enterprise, which offers enterprise-grade security and privacy, unlimited higher-speed GPT-4 access, longer context windows for processing longer inputs, advanced data analysis capabilities, customization options, and much more. Most notably, the inputs and interactions with the models are NOT used for any training purposes allowing for data privacy as a part of the agreement.
Niantic adds Generative AI to 8thWall Web AR Development Platform
Niantic announced this week that it is introducing new generative AI developer tools to its 8th Wall platform. More specifically, it’s adding GenAI modules to 8th Wall that make it easier for developers to streamline their processes. Devs can use these tools to incorporate GenAI into their WebAR development within 8th Wall, and Niantic has also introduced two sample projects for developers. This is just the very begging of how generative AI will dramatically change augmented and virtual reality development.
Survey finds relatively few Americans actually use (or fear) ChatGPT
Ongoing polling by Pew Research shows that although ChatGPT is gaining mindshare, only about 18% of Americans have ever actually used it. Among those who have heard of ChatGPT said it is likely to have a major impact on software engineers, graphic designers and journalists; we may safely speculate that there is some conflation of other generative and interpretive AI models in there, and the sense is that AI in general, not necessarily ChatGPT specifically, will lead to this impact.
Court Ruling: AI-Generated Art Cannot be Copyrighted
United States District Court Judge Beryl A. Howell ruled [this month] that AI-generated artwork can’t be copyrighted, as noted by The Hollywood Reporter. She was presiding over a lawsuit against the US Copyright Office after it refused a copyright to Stephen Thaler for an AI-generated image made with the Creativity Machine algorithm he’d created. Judge Howell did, however, acknowledge that humanity is “approaching new frontiers in copyright,” where artists will use AI as a tool to create new work. She wrote that this would create “challenging questions regarding how much human input is necessary” to copyright AI-created art, noting that AI models are often trained on pre-existing work. We are far from settled on this topic, but as the ability to create anything expands this area is worth keeping tabs on.
ChatGPT Tests in the Top 1% for Original Creative Thinking
Recent findings from the University of Montana and partners indicate that artificial intelligence can rival the creative abilities of the top 1% of human participants based on a standard test for creativity.
Led by Dr. Erik Guzik, an assistant clinical professor at UM’s College of Business, the team employed the Torrance Tests of Creative Thinking – a well-known tool used for decades to assess human creativity.
The researchers submitted eight responses generated by ChatGPT, the application powered by the GPT-4 artificial intelligence engine. They also submitted answers from a control group of 24 UM students taking Guzik’s entrepreneurship and personal finance classes. These scores were compared with 2,700 college students nationally who took the TTCT in 2016. All submissions were scored by Scholastic Testing Service, which didn’t know AI was involved.
The results placed ChatGPT in elite company for creativity. The AI application was in the top percentile for fluency – the ability to generate a large volume of ideas – and for originality – the ability to come up with new ideas. The AI slipped a bit – to the 97th percentile – for flexibility, the ability to generate different types and categories of ideas.
“For ChatGPT and GPT-4, we showed for the first time that it performs in the top 1% for originality,” Guzik said. “That was new.”
Creativity is a leading use case of these tools and can be put to work almost immediately with minor risks.
History: Siri, Alexa, and Cortana
Not so long ago, we experienced the emergence of virtual assistants. Beyond the comedic situations where one interrupted one of our conversations, it marked a revolutionary leap in human-computer interaction. It’s time we looked back and delve into the historical milestones that gave birth to Apple's Siri, Amazon's Alexa, and Microsoft's Cortana.
Apple's Siri: A Voice from the Future
In 2011, Siri became the world's first widely-used virtual assistant, integrated into the iPhone 4S. Siri's charm lay in her ability to understand natural language and assist users with tasks, from sending text messages to setting reminders. This remarkable invention was more than a tool; it was a friendly companion that opened doors to endless possibilities. Siri’s evolution saw integration with third-party apps and even a more natural-sounding voice, reflecting Apple's commitment to constant innovation. The voice of siri is so well recognized it won the voice’s owner, Susan Bennett, a Wikipedia article.
Amazon's Alexa: The Home's New Best Friend
Hot on the heels of Siri, Amazon introduced Alexa in 2014, housed within the Amazon Echo. Alexa transcended the boundaries of personal devices and found her place in our homes. Whether controlling smart devices or playing your favorite tunes, Alexa showcased the potential of AI in everyday life. Its open SDK allowed developers to create "skills," expanding Alexa's abilities and transforming how we interact with our homes.
Microsoft's Cortana: Bridging the Gap Between Devices
Enter Cortana, Microsoft's answer to Siri and Alexa, introduced in 2014. Named after a character from the "Halo" video game series, Cortana was designed to sync across various Windows devices seamlessly. Cortana’s integration demonstrated a unified experience from PCs to smartphones, underlining Microsoft's vision of a connected ecosystem. Despite facing challenges and undergoing several changes in focus, Cortana's contribution to voice-enabled technology cannot be overlooked, even if no one can remember it.
Lessons and Reflections
These virtual assistants were more than mere technological feats; they symbolized a shift towards a more intuitive and responsive digital world. They taught us the importance of understanding user context, building empathy through technology, and the endless opportunities in harnessing AI.
How many of us have only adopted these to turn them off or decided they were not useful enough?
User Engagement Matters
From Siri's conversational tone to Alexa's customizable skills, these virtual assistants showed that understanding and engaging users is vital for success. The early versions of these tools were novel, but once the novelty wore off, we either found a good use for them or turned them off altogether. However, it’s important to realize that just because we stopped using them, their technology stuck around and continues to improve even now.
Innovation is Continuous
Siri, Alexa, and Cortana have continually evolved, reflecting a never-ending quest for innovation and improvement. They remind us that stagnation is not an option in the world of technology. Solutions that seem to be irrelevant or below the capability level that threaten us, it’s important to recognize that it doesn’t stand still. Today’s Alexa can become tomorrow’s ChatGPT.
Integration is Key
Whether Cortana's cross-device functionality or Alexa's integration with smart homes, seamless connectivity enhances user experience, providing more coherent and efficient solutions.
As we march toward a future filled with AI-driven experiences, we must take a moment to reflect on these pioneers who changed the way we communicate with technology. Their lessons resonate with us today and guide us toward a more interactive, intuitive, and intelligent tomorrow. Let’s not forget the lessons we’ve learned from their evolution, but also to consider how their arc of development will change the future of how we work with machines.
Learning, Tools, and Experiments
Here are some great things you can learn about and explore.
Universal and Transferable Adversarial Attacks on Aligned Language Models
This paper might interest you if you are into information security or even just pushing the boundaries. Essentially, it’s a paper that details ways to exploit LLMs and is yet another example of getting the models to reply with content intentionally meant to be prevented. The cat-and-mouse game will continue, but it’s always fun to understand their success and what might be valuable for stopping and taking advantage of their features (intended or not).
https://arxiv.org/abs/2307.15043
Meta AI Releases CodeLlama
Code Llama is Meta's new open-source AI system to make coding more efficient by leveraging large language models trained on code. Its capabilities include code generation, completion, explanation, and debugging.
https://about.fb.com/news/2023/08/code-llama-ai-for-coding/
Perplexity
Perplexity AI is an AI-powered search engine that directly answers user queries by summarizing information from the web. I have found this tool more beneficial for research since ChatGPT has disabled the browser plug-in. I still use multiple plug-ins and tools with ChatGPT, but perplexity has captured my attention as an excellent service, and I enjoy using it.
Framer
Framer is a web design platform that allows users to design and build responsive websites without coding. They’ve been around for a while, but I recently learned that they automatically added the ability to generate a single-page website with a text prompt. While I wasn’t overly impressed with the outcome, I expect we will continue to see well-established tools integrating and updating their solutions with generative in much the same way.