RAG Predictive Coding for AI Alignment Against Prompt Injections and Jailbreaks Post date September 5, 2024 Post author By Stephen Post categories In ai-alignment, ai-chatbot, ai-chatbot-development, ai-safety, predictive-coding, prompt-injection, retrieval-augmented-generation, rlhf
LLMs: Is NIST’s AI Safety Consortium Relevant Amid California’s SB 1047? Post date August 29, 2024 Post author By Stephen Post categories In ai-alignment, ai-governance, ai-regulation, ai-safety, llms, mathematics, neuroscience, nist
California AI Safety|EU Regulation: LLMs Emergent Abilities and Existential Threat Post date August 14, 2024 Post author By Stephen Post categories In ai-alignment, ai-safety, emergent-abilities, eu-ai-act, existential-risks, llms, sb-1047, superintelligence
AI Safety and Alignment: Could LLMs Be Penalized for Deepfakes and Misinformation? Post date July 19, 2024 Post author By Stephen Post categories In agi, ai-alignment, ai-safety, deepfakes, llms, misinformation, neuroscience, superintelligence