Hello Reader,

Welcome to Promptcraft, your weekly newsletter on artificial intelligence for education. In this issue:

  • OpenAI Leadership in Turmoil as Sam Altman CEO Sacked
  • Google expands Bard AI Chatbot access to teens
  • Microsoft becomes a Co-pilot company
  • 🎄 Share Promptcraft and enter my Christmas giveaway!

Let’s get started!

.: Tom

Latest News

.: AI Updates & Developments

.: OpenAI Leadership Turmoil ➜ OpenAI, the company and research lab behind ChatGPT, is in flux as the board sacked Sam Altman the CEO on Friday. It helps not to send Promptcraft too soon as this story has been pretty fluid over the weekend!

  • Friday: Sam Altman fired from OpenAI for lack of candour; Greg Brockman and researchers quit in protest.
  • Saturday: Interim OpenAI CEO Mira Murati tries to rehire Altman and Brockman; board looks for permanent CEO.
  • Sunday: Microsoft hires Altman and Brockman; OpenAI hires Twitch’s Emmett Shear as new CEO.
  • Monday: 500+ OpenAI employees threaten to quit unless board steps down; Sutskever expresses regret over Altman’s firing.
  • Latest: According to The Verge Sam Altman and Greg Brockman have expressed openness to coming back to OpenAI, but only if the board members responsible for firing Altman resign their positions.

Melissa Heikkilä at The Algorithm provides a helpful overview to catch up on and understand the next steps in this unfolding situation at OpenAI.

.: Google to expand Bard AI chatbot access to teens globally ➜ Google announced it will open up its AI chatbot Bard to teenagers globally in English starting November 16, with more languages to come. Bard aims to provide a helpful, informational tool for teens to learn new skills and find inspiration. Google consulted child safety experts and implemented guardrails to prioritise safety. Features include math equation solving, data visualisation, content policies to avoid unsafe content, and double checking responses to develop critical thinking.

.: Microsoft unveils major AI plans and products at Ignite 2023 ➜ At its annual Ignite conference, Microsoft announced significant AI-related products and initiatives. These include rebranding Bing Chat to Microsoft Copilot, a Copilot Studio to allow custom AI bot creation, new AI chips like Azure Maia and Azure Cobalt to power Azure cloud services, adding generative AI capabilities to Teams VR meetings, and more. Key highlights show Microsoft’s continued push to infuse AI across its products and position itself as a leader in enterprise AI.

oZGUzvhM4n9AwVredHpopK

.: In New Experiment, Young Children Destroy AI at Basic Tasks ➜ A study found kids age 3-7 greatly outperform AI models at basic problem solving and thinking tasks. Tests of tool innovation and inferring causal relationships showed children’s superior unconventional thinking. Researchers said unlike AIs, curious and motivated kids are intrinsically better at core innovation. The study highlights limitations of current AI versus human cognition and reasoning.

bgpe9sY9jc3QJyQ6X2cwE9

.: YouTube will show labels on content that uses AI ➜ YouTube announced it will require creators to disclose use of AI to alter or synthesise realistic content. Labels will indicate to viewers that content uses AI, especially prominently for sensitive topics. This aims to avoid misleading viewers that AI content is real. Failure to properly disclose could lead to removal and suspension. YouTube is also introducing AI music removal requests to address fake songs.

.: Chinese startup 01.AI unveils powerful new open source AI models Yi ➜ Chinese company 01.AI has released two new large language models called Yi-6B-200K and Yi-34B-200K. The models are fully open source and can understand English and Mandarin. Yi-34B boasts 200,000 tokens of context, double ChatGPT’s capacity, though long prompts can challenge its recall. Yi benchmarks show strengths in comprehension, reasoning, and standardised AI tests. By being open source, Yi allows full customisability for developers to build local AI apps.

.: Alibaba, the major Chinese e-commerce company, open sources AI models Qwen-7B and Qwen-7B-Chat ➜ Alibaba’s cloud unit unveiled two new open source large language models named Qwen-7B and Qwen-7B-Chat with 7 billion parameters each. This positions the models as competitors to Meta’s similarly open sourced Llama 2 model. Alibaba says the move aims to help small and medium businesses adopt AI. The code and models are freely available globally, though licensing is required for large companies. This represents the first time a major Chinese tech company has open sourced a large language model.

.: Germany, France and Italy reach agreement on AI regulation in Europe ➜ The governments of Germany, France and Italy have agreed on an approach for regulating AI in Europe. They support mandatory self-regulation through codes of conduct for foundational AI models. The countries oppose unchecked norms and want to focus regulations on AI applications rather than the core technology. Under the proposal, AI developers would use model cards to provide information on capabilities and limitations. An EU AI governance body could help develop guidelines and oversight. The agreement aims to accelerate EU-level negotiations on an AI Act among European Commission, Parliament and Council.

Reflection

.: Why this news matters for education

Amidst all of the tumultuous news about OpenAI, I expanded my AI Literacy with two new terms: the “accels” who want to accelerate AI development at any cost, and the “decels” who favour slowing down development to ensure safety.

Although binary and reductionist, this philosophical divide over the pace of progress seems to be at the heart of the rift that led to the leadership shakeup at OpenAI.

Some have said that Ilya Sutskever, the Chief Scientist for OpenAI and board member, wants to slow down progress. While Sam Altman represents the race for faster development.

This tension between accelerating progress and prioritising safety is not new for OpenAI.

Dario Amodei, who was Vice President of Research at OpenAI until 2018, left the organisation amidst similar philosophical differences over the responsible pace of AI development. He went on to co-found Anthropic, the creator of the Claude-2 LLM, along with other former OpenAI researchers who were focused on AI alignment and robustness.

On the surface, OpenAI’s boardroom turmoil might appear to be just corporate drama with little bearing on education.

However, when viewed through an ecosystem lens, this news sends ripples that connect to our work in education in several ways:

  1. Focus on safety: The safety of AI products and their underlying architecture must be a top priority.
  2. Reliability of products and their architecture: The reliability of AI products is essential for ensuring their effective integration into educational settings.
  3. Centrality of major research labs and developers: OpenAI and other major AI research labs play a pivotal role in shaping the future of AI for education.
  4. Power shifts between big tech companies: The power dynamics among major tech companies can influence the trajectory of AI development.
  5. Profits over benefits for humanity: The pursuit of profits overshadows the broader societal benefits of AI.
  6. Distracting noise: Energy, effort and time is pulled away from putting powerful AI tools in service of education.

Two undeniable facts (i) OpenAI has set the standard for AI research and development and (ii) possesses the most powerful publicly available large language model, GPT-4.

This alone is enough to pique the interest of educators curious about the ripple effects of the organisation’s leadership changes.

A shift in the AI research and development ecosystem, inevitably translates into a shift in the education ecosystem.

.:

~ Tom

Tom%20Barrett%20Christmas%20Sticker%20(1).

Get Poe AI Access for Free – Refer Friends to Win!

To enter share your unique link below and get an entry for every friend that signs up to the Promptcraft newsletter. The more referrals you get, the higher your chances to win!

Prizes: 10 x 1 month Poe AI access (USD $20 value each)
Draw date: December 20th 2023

[RH_REFLINK_2 GOES HERE]

Facebook Whatsapp Linkedin Email

PS: You have referred [RH_TOTREF_2 GOES HERE] people so far

See how many referrals you have

Prompts

.: Refine your promptcraft

Let’s talk about GPTs.

Remember this stands for Generative Pre-trained Transformer, which means:

  • Generative: GPTs are able to generate new outputs, rather than simply regurgitating text that they have been trained on.
  • Pre-trained: GPTs are trained on a massive dataset of text and code before they are released to the public.
  • Transformer: GPTs use a transformer architecture, which is a type of neural network that is well-suited for natural language processing tasks.

OpenAI recently announced the capability with a Plus account (paid) to build your own chatbot or what they call GPTs.

So, what does this have to do with Promptcraft?

Well the process for building GPTs automatically generates prompts. You can simply say what you are looking to build and it writes a prompt for you.

This begins to remove the need for writing your own prompts, but it puts up a fee barrier, and not everyone has access.

One way to replicate this is to use an instruction in your prompt to trigger automated improvement. Try this:

Act as an expert LLM prompt engineer and writer. Rate my LLM prompt below 1-10 and provide kind, specific and helpful feedback. If the rating is 8 or higher, execute the prompt. If it is lower than 8, generate a better prompt and explain how it is better.

My prompt: [add your prompt here]

Here is an example in ChatGPT 3.5, Claude-2-100k and Bard.

*Remember the scoring is all a bit unreliable, you are just creating an exchange to improve your prompts.

**And, as I always say, remember to make this your own, tinker and evaluate the completions.

Learning

.: Boost your AI Literacy

RESEARCH TOOL .: OECD AI Incidents Monitor (AIM) ➜ A fascinating analysis tool which documents AI incidents to help policymakers, AI practitioners, and all stakeholders worldwide gain valuable insights into the incidents and hazards that concretise AI risks.

RESEARCH INDEX .: Latin American Index of Artificial Intelligence ➜ A comprehensive analysis of the status of AI in twelve countries in Latin America: Argentina, Bolivia, Brazil, Chile, Colombia, Costa Rica, Ecuador, Mexico, Panama, Paraguay, Peru and Uruguay. Each file elaborates on: Enabling Factors, Research, Development and Adoption, and Governance.

I was curious about AI news from here as we live in media geo-bubbles, so I was pleased to discover this resource providing insight into what is happening in Latin America.

REPORT .: Colonialism and AI ➜ This report by Anna Gausen and Accessible AI, explores how AI is at risk of repeating the patterns of our colonial history and how we can begin to decolonise AI.

It covers:

  • A Look Back At Our Past: Society has been shaped by our colonial history.
  • Where We Are Today: The way AI is being deployed by the global west could reinforce colonial power dynamics.
  • A Vision For The Future: How we can rebalance power and diversify voices in AI.

Ethics

.: Provocations for Balance

  • When making decisions about AI progress, whose voices need to be at the table beyond corporate executives?
  • If current AI lacks core elements of human reasoning, when should we be cautious about over-applying it to tasks requiring critical thinking?
  • When AI-generated content crosses ethical lines, how should accountability be determined given the complex web of humans and algorithms involved in systems?

~ Inspired by this week’s developments.

.:

That’s all for this week; I hope you enjoyed this issue of Promptcraft. I would love some kind, specific and helpful feedback.

If you have any questions, comments, stories to share or suggestions for future topics, please reply to this email or contact me at tom@dialogiclearning.com

The more we invest in our understanding of AI, the more powerful and effective our education ecosystem becomes. Thanks for being part of our growing community!

.: Tom Barrett

/Creator /Coach /Consultant