How should we navigate the growing role of AI in preprint peer review? In #OAW2025, we highlight insights from the webinar talk of Dr. Jonny Coates, Executive Director and Founder of Rippling Ideas. Dr. Coates proposed 5 guiding principles for responsible AI use: Transparency, Human Oversight, Security & Confidentiality, Quality & Integrity, and Ethical & Unbiased Review. He also outlined 5 key implementation ways to put these principles into action. Hear Dr. Coates' full perspective: https://brnw.ch/21wWPgV #Preprints #OpenScience #Webinar #OAWeek
Navigating AI in preprint peer review with Dr. Coates
More Relevant Posts
-
The rapid advancement of AI brings both unprecedented opportunities and significant risks. Our latest post, 'AI-2027: A Wake-Up Call on Advanced AI and Existential Risk', delves deep into potential scenarios of superintelligent systems and their implications on national security. Engage with us in addressing these pressing challenges! Read more here: https://wix.to/uIaYhHm #ArtificialIntelligence #NationalSecurity #FutureTrends
To view or add a comment, sign in
-
I don’t say this lightly — this might be one of the most important things I post on LinkedIn. We’re approaching a tipping point in AI development that could define our future — or end it. My new blog explores the AI-2027 scenario, existential risk, and what it means for national security and secure facilities like SCIFs and SAPFs. Give it a read, share your thoughts, and let’s start the hard conversations now.
The rapid advancement of AI brings both unprecedented opportunities and significant risks. Our latest post, 'AI-2027: A Wake-Up Call on Advanced AI and Existential Risk', delves deep into potential scenarios of superintelligent systems and their implications on national security. Engage with us in addressing these pressing challenges! Read more here: https://wix.to/uIaYhHm #ArtificialIntelligence #NationalSecurity #FutureTrends
To view or add a comment, sign in
-
AI-powered ransomware is no longer science fiction. The discovery of PromptLock — the first self-adapting ransomware driven by local AI models — marks a turning point in cybersecurity. Curious about what this means for AI safety and defense? 🎧 Listen to the latest edition of The Promptly Times by the AI Security Foundation.
Dear AI community, The long-awaited Newsletter #11 is here! 🎉 Dive into this edition for cutting-edge updates on AI safety, privacy, and emerging risks shaping the future of intelligent systems. In This Issue: 🔍 Detecting and Reducing Scheming in AI Models 🔒 Zero Data Retention: The Next Step in LLM Privacy 🧫 Small Samples Can Poison Large Language Models 🖥️ Gemini 2.5 Computer Use: Agents that Interact with Interfaces 🧠 The First AI-Powered Ransomware Discovered: PromptLock 📩 Read the newsletter: https://lnkd.in/ebM7A3_D 🎧 Listen to the podcast: https://lnkd.in/e7D6pmsC Enjoy the long-awaited drop — packed with insights, research, and the latest AI developments. #ai #artificialinteligence #globalsecurity #aisecfoundation #thepromptlytimes
To view or add a comment, sign in
-
Last week, Nardello & Co.’s Joseph Pochron and Karsten Wilkinson presented at the Techno Security & Digital Forensics Conference about fraud detection and adding open-source AI detection to your digital forensics and incident response toolkit. Here’s a recap of what they discussed: - 78% of organizations use AI in at least one function of their business - The four deepfake detection options: algorithmic or model-based detection; forensic analysis; content authentication; and human-in-the-loop - The pros and cons of leveraging open-source deepfake detection tools on platforms such as HuggingFace - The steps of deepfake detection implementation and how to utilize the results - Current issues in the AI detection landscape, such as watermarking, training limitations, and convergence of LLMs Want to learn more? Reach out to Joseph Pochron, Karsten Wilkinson, or read the resources available here: https://lnkd.in/ey7mNHcu #AI #LLM #Deepfake #HuggingFace #CyberDefense #Investigations
To view or add a comment, sign in
-
-
🔒 AI in healthcare requires trust, transparency, and shared understanding. Our team has released Promise to Map - a practical approach to threat modeling AI and LLM Systems to help healthcare organizations evaluate security and privacy risks in AI implementation. We're sharing this framework freely because we believe building trust in healthcare AI is a collective responsibility. Together, we can create a future where innovation and patient safety go hand in hand. Read the whitepaper: https://lnkd.in/e_T_XQKV #HealthcareAI #PatientSafety #DigitalHealth
To view or add a comment, sign in
-
-
This week, I had the opportunity to attend TASK - Toronto Area Security Klatch and listen to Bruce Schneier speak about "AI and Trust." His insights on the evolution of personal trust into scalable social trust, as well as the roles of corporations and governments in securing AI systems and protecting social trust through reliability and predictability, were enlightening. His emphasis that integrity is the major challenge among the CIA triad when it comes to securing these new and far-reaching technologies alongside the question of INTEGROUS system design was a point that stuck with me after the talk. "Our fears of AI are basically fears of misplaced capitalism."
To view or add a comment, sign in
-
-
SOC operators might soon get assistance from an emerging technology: AI agents leveraging Agentic AI. This advancement could a.o. significantly reshape the way SOCs operate, turning static automation into flexible threat management. Researchers Armin Shokri Kalisa and Robbert Schravendijk outline the developments surrounding this new technology in their latest Journal article 'Smarter Autonomous Security Operations: The Next Evolution, The Role of Agentic AI in a SOC'. Is your security team ready to weigh the risks and benefits of Agentic AI? Read more via: https://lnkd.in/etgAaBFx Many thanks to the authors for sharing their knowledge and views, as well as to the Review Board for their careful revision: dr.Barry Derksen and Eric Mantelaers. #agenticai #ai #security #soc #threatmanagement #IsacaNetherlandsChapter
To view or add a comment, sign in
-
-
🔐 The Next Evolution of Security Operations: Agentic AI in the SOC Security Operations Centers are facing increasing pressure from alert fatigue and data overload to a persistent shortage of skilled analysts. In our latest article, Armin Shokri Kalisa and I look at how Agentic AI could change this landscape. What if AI agents didn’t just assist analysts, but could reason, act, and learn autonomously alongside them? The potential is huge, ranging from faster response times to smarter decision-making, but it also raises important questions: ⚖️ How do we stay in control of autonomous AI agents? 💰 Do the benefits outweigh the implementation costs? 🔒 And how do we ensure new technologies don’t create new vulnerabilities? We’d love to hear your thoughts: ➡️ How will autonomous agents reshape the way SOCs work? Will the SOC of the future rely more on Agentic agents, or will human expertise remain at its core? Thanks to the #isaca review board, dr.Barry Derksen, Eric Mantelaers, Susan Schaeffer Read the full article here: https://lnkd.in/eUZBqSBf #CyberSecurity #AgenticAI #SOC #AI #Innovation #SecurityOperations
SOC operators might soon get assistance from an emerging technology: AI agents leveraging Agentic AI. This advancement could a.o. significantly reshape the way SOCs operate, turning static automation into flexible threat management. Researchers Armin Shokri Kalisa and Robbert Schravendijk outline the developments surrounding this new technology in their latest Journal article 'Smarter Autonomous Security Operations: The Next Evolution, The Role of Agentic AI in a SOC'. Is your security team ready to weigh the risks and benefits of Agentic AI? Read more via: https://lnkd.in/etgAaBFx Many thanks to the authors for sharing their knowledge and views, as well as to the Review Board for their careful revision: dr.Barry Derksen and Eric Mantelaers. #agenticai #ai #security #soc #threatmanagement #IsacaNetherlandsChapter
To view or add a comment, sign in
-
-
💡 It's fantastic to see my Sysdig teammate Crystal Morin leading the discussion on the true essence of AI in security — prioritizing people over technology. ✍️ Her new 𝘊𝘳𝘺𝘴𝘵𝘢𝘭 𝘊𝘭𝘦𝘢𝘳 LinkedIn article series explores the human aspect of AI in the SOC, offering a perspective that's crucial for our industry today. Check it out ➡️ https://okt.to/znejxL
To view or add a comment, sign in
-
-
SmartGateVC partner Ashot Arzumanyan was featured in CIOReview’s CXO Insights. In the interview, Ashot shares how we catch waves early (AI in 2018), why we are now focused on Physical AI, Brain-Computer Interfaces, and Cybersecurity, and how our SoCal ↔ Armenia model (via Hero House hubs) compresses time to validation. He also underscores our builder ethos- “capable, unstoppable, unbreakable” - and Fund I’s top-decile performance with companies like SuperAnnotate, Deep Origin, Grovf and Krisp. Link to the article: https://lnkd.in/dVtWM7Pt
To view or add a comment, sign in