Exciting news! đ
Our Co-Founder/CPO,
@jeanmarie_johnm
, recently shared insights with
@safetydet
in an in-depth interview on #AISafety and Security.
Discover how we're identifying #vulnerabilities, and ensuring the #security of AI #models.
Full article: https://t.co/nr6yJywivB
#aisafety #vulnerabilities #security #models
#UK #AI #AISafety: "The current âAI momentâ is a critical inflection point for the UK and the world. As AI systems become more complex and capable, organisations across all sectors of the global economy are looking to develop, deploy and make use of their potential to benefit people and society.
With this opportunity come considerable risks: ranging from bias, misuse, and system failure to structural harms, such as the concentration of economic power in the hands of a small number of companies.
Without concerted action, we may unwittingly lock ourselves into a set of technologies and economic dynamics that fail to benefit people, communities, or society as a whole.
Enter the UK Government, which is hosting an âAI Safety Summitâ on 1 and 2 November at a venue synonymous with overcoming wicked technical problems: Bletchley Park, where Allied codebreakers deciphered the German âEnigmaâ code during World War II.
The Government has recently set out objectives for the Summit, including reaching a âshared understandingâ of AI risks, agreement on areas of potential international collaboration and showcasing âAI for good.â
Reaching consensus on these topics will not be easy â so what can Government do to improve its chances of success?"
đ Passionate about #AISafety? Join our #Community program!
đ Whether you're a #DataScientist, #ML Engineer, #AIEthicist, or business stakeholder, share your insights through articles & tutorials. You can gain recognition & get paidđ°
Learn more: đ https://www.giskard.ai/write-for-the-community
#aisafety #community #datascientist #ml #aiethicist
It was a pleasure to be part of this event and witness the #AICommunity come together, especially on topics as critical as #AISafety and LLMs' security. đĄ đ We had the opportunity to engage with leading AI teams from
@huggingface, @stabilityai, #Cohere, and #AnthropicAI
[2/5]
#aicommunity #aisafety #cohere #anthropicai
Yannic Kilcher comments on generative recipe AIs making chlorine gas then creates an AI that accepts nails as an ingredient when asked to
https://youtu.be/BMAu7hAcjqU
#generativeAI #AISafety #scicomm #AIEthics #AI #ArtificialIntelligence
#generativeai #aisafety #scicomm #aiethics #ai #artificialintelligence
Amazing 2nd day at @defcon đť
This year we're happily sponsoring the #AIVillage đ and we've contributed to some of the challenges for their traditional #CTF that will start on September 1st.
đ DM to keep the discussion on #AISafety #LLMs #ML #vulnerabilities
#aivillage #ctf #aisafety #llms #ml #vulnerabilities
Greetings from #DEFCON31! đ
đ˘ The Giskard team is now at #DC31 and we'll be happy to meet you. Join us at the #AIVillage for the #GenAI #RedTeam.
đŠ DM us if you want to meet and discuss about #AISafety, #LLMs safety, #AI #Testing and #MLOps.
#Defcon31 #dc31 #aivillage #genai #redteam #aisafety #llms #ai #testing #MLops
AI-powered grocery bot suggests recipe for toxic gas, âpoison bread sandwichâ - Enlarge (credit: PAK'nSAVE)
When given a list of harmful ingre... - https://arstechnica.com/?p=1960122 #largelanguagemodels #machinelearning #newzealand #redteaming #aiethics #aisafety #paknsave #gpt-3.5 #biz⢠#openai #gpt-3 #tech #ai
#ai #tech #openai #biz #gpt #paknsave #aisafety #aiethics #redteaming #newzealand #machinelearning #largelanguagemodels
Ars Technica: AI-powered grocery bot suggests recipe for toxic gas, âpoison bread sandwichâ https://arstechnica.com/?p=1960122 #Tech #arstechnica #IT #Technology #largelanguagemodels #machinelearning #newzealand #redteaming #PAK'nSAVE #AIethics #AIsafety #GPT-3.5 #Biz&IT #openai #GPT-3 #Tech #AI
#Tech #arstechnica #it #technology #largelanguagemodels #machinelearning #newzealand #redteaming #pak #aiethics #aisafety #gpt #biz #openai #ai
In this new talk by Yann Le Cun he floats the idea that "evil AI" will be controlled by the "Good Guysâ AI police".
I assume it will be less fun when someone figures out how to build Nazi AI...
LeCun's link to slides: https://drive.google.com/file/d/1wzHohvoSgKGZvzOWqZybjm4M4veKR6t3/view
This is the true "#AISafety" risk. It's not that a chatbot will become sentient and take over the world - it's that the original artificial lifeform, the limited liability company, will use "AI" to accelerate its murderous shell-game until we can't spot the trick:
https://pluralistic.net/2023/06/10/in-the-dumps-2/
32/
đ˘ At Giskard, we're creating a robust #ML framework for #testing ML #models effectively. We help identify #biases and #errors in AI models, from #tabular to #LLMs. Participating in DEFCON allows us to collaborate with leading experts and share our commitment to #AISafety [3/4]
#ml #testing #models #biases #errors #tabular #llms #aisafety
đWe'll join you at the #AIVillage for the #GenAI #RedTeam! It's a great opportunity to show the potential of GenAI, and emphasize the importance of #AISafetyđĄď¸
We've contributed to some of the challenges for the AIVillage #CTF and canât wait to have you try them out!𤯠[2/4]
#aivillage #genai #redteam #aisafety #ctf
#AI
#AIGenerated
#AIArt
#AIhype
#AIrisk
#AIsafety
#GenerativeAI
#GenerativeArt
#ChatGPT
So Iâm in an ethical quandary.Iâd like to do some simple generative blog layout and art, because I have no money to hire artists and designers. But I wonât use any system that was trained on non consensual content, or violates anyoneâs copyright or other rights.
So is there any system out there I can responsibly use? If not Iâll try to do the work myself.
#ai #AIgenerated #aiart #aihype #airisk #aisafety #generativeAI #generativeart #chatgpt
#EU #USA #BigTech #AI #AISafety #SelfRegulation: "The European Union might be making strides toward regulating artificial intelligence (with passage of the AI Act expected by the end of the year), but the US government has largely failed to keep pace with the global push to put guardrails around the technology.
The White House, which said it âwill continue to take executive action and pursue bipartisan legislation,â introduced an interim measure last week in the form of voluntary commitments for âsafe, secure, and transparent development and use of AI technology.â
Amazon, Anthropic, Google, Inflection, Meta, Microsoft, and OpenAI agreed to âprioritize research on societal risks posed by AI systemsâ and âincent third-party discovery and reporting of issues and vulnerabilities,â among other things.
But according to academic experts, the agreements fall far short."
https://www.emergingtechbrew.com/stories/2023/07/25/us-voluntary-ai-safety-commitment
#eu #usa #bigtech #ai #aisafety #selfregulation
#AI #AISafety #SystemicRisk #Nuclear: "Like Oppenheimer before them, many merchants of AI believe their creations might change the course of history, and so they wrestle with profound moral concerns. Even as they build the technology, they worry about what will happen if AI becomes smarter than humans and goes rogue, a speculative possibility that has morphed into an unshakable neurosis as generative-AI models take in vast quantities of information and appear ever more capable. More than 40 years ago, Rhodes set out to write the definitive account of one of the most consequential achievements in human history. Today, itâs scrutinized like an instruction manual.
Rhodes isnât a doomer himself, but he understands the parallels between the work at Los Alamos in the 1940s and whatâs happening in Silicon Valley today. âOppenheimer talked a lot about how the bomb was both the peril and the hope,â Rhodes told meâit could end the war while simultaneously threatening to end humanity. He has said that AI might be as transformative as nuclear energy, and has watched with interest as Silicon Valleyâs biggest companies have engaged in a frenzied competition to build and deploy it."
#ai #aisafety #systemicrisk #nuclear
#USA #AI #Biden #Algorithms #AISafety: "An executive order can enshrine these best practices in at least four ways. First, it could require all government agencies developing, using, or deploying AI systems that affect peopleâs lives and livelihoods to ensure that these systems comply with best practices. For example, the federal government might make use of AI to determine eligibility for public benefits and identify irregularities that might trigger an investigation. A recent study showed that IRS auditing algorithms might be implicated in disproportionately high audit rates for Black taxpayers. If the IRS were required to comply with these guidelines, it would have to address this issue promptly.
Second, it could instruct any federal agency procuring an AI system that has the potential to âmeaningfully impact [our] rights, opportunities, or access to critical resources or servicesâ to require that the system comply with these practices and that vendors provide evidence of this compliance. This recognizes the federal governmentâs power as a customer to shape business practices. After all, it is the biggest employer in the country and could use its buying power to dictate best practices for the algorithms that are used to, for instance, screen and select candidates for jobs."
https://www.wired.com/story/the-white-house-already-knows-how-to-make-ai-safer/
#usa #ai #biden #algorithms #aisafety
As long as companies like openai, anthropic, gooogle and co don't put out high quality training material explaining to users what LLMs are, how they function, how they can be abused and how to deal with that, it's really hard to take their getting all worked up about "AI safety" seriously.
A decent, level-headed online course with 5 little 5 minute modules would solve so many immediate issues. Every saas company does this.
In reading of the wild hopes, fears, and visions of the future tied to #AISafety, and @emilymbender on how it, effective altruism, and longtermism are tied to racism and pseudoscience, I'm reminded of the ultimate dystopic AI scare, in Francis E. Dec's old rants.
In Dec's view, an ancient Slovene computer encyclopedia became the Worldwide Mad Deadly Communist Gangster Computer God, all humans its remote-controlled Frankenstein Slaves save Dec with his pure Polish genes.
#AI #AISafety #AIEthics: "What does it mean to make AI systems safe, and what values and approaches must be applied to do so? Is it about âalignment,â ensuring that deployment of AI complies with some designersâ intent? Or is it solely about preventing the destruction of humanity by advanced AI? These goals are clearly insufficient. An AI system capable of annihilating humankind, even if we managed to prevent it from doing so, would still be among the most powerful technologies ever created and would need to abide by a much richer set of values and intentions. And long before such powerful ârogueâ AI systems are built, many others will be made that people will use dangerously in their self-interest. Years of sociotechnical research show that advanced digital technologies, left unchecked, are used to pursue power and profit at the expense of human rights, social justice, and democracy. Making advanced AI safe means understanding and mitigating risks to those values, too. And a sociotechnical approach emphasizes that no group of experts (especially not technologists alone) should unilaterally decide what risks count, what harms matter, and to which values safe AI should be aligned. Making AI safe will require urgent public debate on all of these questions and on whether we should be trying to build so-called âgod-likeâ AI systems at all."