If you're new here, you may want to try my weekly newsletter. Thanks for visiting!
Hello Reader,
Welcome to Promptcraft, your weekly newsletter on artificial intelligence for education. In this issue:
- OpenAI Leadership in Turmoil as Sam Altman CEO Sacked
- Google expands Bard AI Chatbot access to teens
- Microsoft becomes a Co-pilot company
- 🎄 Share Promptcraft and enter my Christmas giveaway!
Let’s get started!
.: Tom
Latest News.: AI Updates & Developments .: OpenAI Leadership Turmoil ➜ OpenAI, the company and research lab behind ChatGPT, is in flux as the board sacked Sam Altman the CEO on Friday. It helps not to send Promptcraft too soon as this story has been pretty fluid over the weekend!
Melissa Heikkilä at The Algorithm provides a helpful overview to catch up on and understand the next steps in this unfolding situation at OpenAI. .: Google to expand Bard AI chatbot access to teens globally ➜ Google announced it will open up its AI chatbot Bard to teenagers globally in English starting November 16, with more languages to come. Bard aims to provide a helpful, informational tool for teens to learn new skills and find inspiration. Google consulted child safety experts and implemented guardrails to prioritise safety. Features include math equation solving, data visualisation, content policies to avoid unsafe content, and double checking responses to develop critical thinking. .: Microsoft unveils major AI plans and products at Ignite 2023 ➜ At its annual Ignite conference, Microsoft announced significant AI-related products and initiatives. These include rebranding Bing Chat to Microsoft Copilot, a Copilot Studio to allow custom AI bot creation, new AI chips like Azure Maia and Azure Cobalt to power Azure cloud services, adding generative AI capabilities to Teams VR meetings, and more. Key highlights show Microsoft’s continued push to infuse AI across its products and position itself as a leader in enterprise AI. |
.: In New Experiment, Young Children Destroy AI at Basic Tasks ➜ A study found kids age 3-7 greatly outperform AI models at basic problem solving and thinking tasks. Tests of tool innovation and inferring causal relationships showed children’s superior unconventional thinking. Researchers said unlike AIs, curious and motivated kids are intrinsically better at core innovation. The study highlights limitations of current AI versus human cognition and reasoning.
|
.: YouTube will show labels on content that uses AI ➜ YouTube announced it will require creators to disclose use of AI to alter or synthesise realistic content. Labels will indicate to viewers that content uses AI, especially prominently for sensitive topics. This aims to avoid misleading viewers that AI content is real. Failure to properly disclose could lead to removal and suspension. YouTube is also introducing AI music removal requests to address fake songs.
|
.: Chinese startup 01.AI unveils powerful new open source AI models Yi ➜ Chinese company 01.AI has released two new large language models called Yi-6B-200K and Yi-34B-200K. The models are fully open source and can understand English and Mandarin. Yi-34B boasts 200,000 tokens of context, double ChatGPT’s capacity, though long prompts can challenge its recall. Yi benchmarks show strengths in comprehension, reasoning, and standardised AI tests. By being open source, Yi allows full customisability for developers to build local AI apps. .: Alibaba, the major Chinese e-commerce company, open sources AI models Qwen-7B and Qwen-7B-Chat ➜ Alibaba’s cloud unit unveiled two new open source large language models named Qwen-7B and Qwen-7B-Chat with 7 billion parameters each. This positions the models as competitors to Meta’s similarly open sourced Llama 2 model. Alibaba says the move aims to help small and medium businesses adopt AI. The code and models are freely available globally, though licensing is required for large companies. This represents the first time a major Chinese tech company has open sourced a large language model. .: Germany, France and Italy reach agreement on AI regulation in Europe ➜ The governments of Germany, France and Italy have agreed on an approach for regulating AI in Europe. They support mandatory self-regulation through codes of conduct for foundational AI models. The countries oppose unchecked norms and want to focus regulations on AI applications rather than the core technology. Under the proposal, AI developers would use model cards to provide information on capabilities and limitations. An EU AI governance body could help develop guidelines and oversight. The agreement aims to accelerate EU-level negotiations on an AI Act among European Commission, Parliament and Council. |
Reflection.: Why this news matters for education Amidst all of the tumultuous news about OpenAI, I expanded my AI Literacy with two new terms: the “accels” who want to accelerate AI development at any cost, and the “decels” who favour slowing down development to ensure safety. Although binary and reductionist, this philosophical divide over the pace of progress seems to be at the heart of the rift that led to the leadership shakeup at OpenAI. Some have said that Ilya Sutskever, the Chief Scientist for OpenAI and board member, wants to slow down progress. While Sam Altman represents the race for faster development. This tension between accelerating progress and prioritising safety is not new for OpenAI. Dario Amodei, who was Vice President of Research at OpenAI until 2018, left the organisation amidst similar philosophical differences over the responsible pace of AI development. He went on to co-found Anthropic, the creator of the Claude-2 LLM, along with other former OpenAI researchers who were focused on AI alignment and robustness. On the surface, OpenAI’s boardroom turmoil might appear to be just corporate drama with little bearing on education. However, when viewed through an ecosystem lens, this news sends ripples that connect to our work in education in several ways:
Two undeniable facts (i) OpenAI has set the standard for AI research and development and (ii) possesses the most powerful publicly available large language model, GPT-4. This alone is enough to pique the interest of educators curious about the ripple effects of the organisation’s leadership changes. A shift in the AI research and development ecosystem, inevitably translates into a shift in the education ecosystem.
.: ~ Tom |
|
Prompts.: Refine your promptcraft Let’s talk about GPTs. Remember this stands for Generative Pre-trained Transformer, which means:
OpenAI recently announced the capability with a Plus account (paid) to build your own chatbot or what they call GPTs. So, what does this have to do with Promptcraft? Well the process for building GPTs automatically generates prompts. You can simply say what you are looking to build and it writes a prompt for you. This begins to remove the need for writing your own prompts, but it puts up a fee barrier, and not everyone has access. One way to replicate this is to use an instruction in your prompt to trigger automated improvement. Try this: Act as an expert LLM prompt engineer and writer. Rate my LLM prompt below 1-10 and provide kind, specific and helpful feedback. If the rating is 8 or higher, execute the prompt. If it is lower than 8, generate a better prompt and explain how it is better.
My prompt: [add your prompt here] Here is an example in ChatGPT 3.5, Claude-2-100k and Bard. *Remember the scoring is all a bit unreliable, you are just creating an exchange to improve your prompts. **And, as I always say, remember to make this your own, tinker and evaluate the completions. |
Learning.: Boost your AI Literacy RESEARCH TOOL .: OECD AI Incidents Monitor (AIM) ➜ A fascinating analysis tool which documents AI incidents to help policymakers, AI practitioners, and all stakeholders worldwide gain valuable insights into the incidents and hazards that concretise AI risks. RESEARCH INDEX .: Latin American Index of Artificial Intelligence ➜ A comprehensive analysis of the status of AI in twelve countries in Latin America: Argentina, Bolivia, Brazil, Chile, Colombia, Costa Rica, Ecuador, Mexico, Panama, Paraguay, Peru and Uruguay. Each file elaborates on: Enabling Factors, Research, Development and Adoption, and Governance. I was curious about AI news from here as we live in media geo-bubbles, so I was pleased to discover this resource providing insight into what is happening in Latin America.
REPORT .: Colonialism and AI ➜ This report by Anna Gausen and Accessible AI, explores how AI is at risk of repeating the patterns of our colonial history and how we can begin to decolonise AI.
It covers:
|
Ethics.: Provocations for Balance
~ Inspired by this week’s developments. |
.:
That’s all for this week; I hope you enjoyed this issue of Promptcraft. I would love some kind, specific and helpful feedback.
If you have any questions, comments, stories to share or suggestions for future topics, please reply to this email or contact me at tom@dialogiclearning.com
The more we invest in our understanding of AI, the more powerful and effective our education ecosystem becomes. Thanks for being part of our growing community!
|
.: Tom Barrett/Creator /Coach /Consultant |