Unleashing On-Device Agentic Power: How Fara-7B Transforms Human-Computer Interaction Microsoft Research’s Fara-7B is a small, open-weight agentic model that interacts with your device in a human-like way. It looks to fulfil the promise ofhaving a digital assistant that doesn’t just un... agentic AI AI safety benchmarking on-device AI open source small language models synthetic data web automation
IBM’s AISteer360 Empowers Safe and Customizable LLM Outputs As large language models (LLMs) become increasingly sophisticated, so does the challenge of ensuring their outputs remain safe, relevant, and trustworthy especially in high-stakes enterprise environme... AI safety AISteer360 AI steering enterprise AI generative AI large language models model alignment open source
Gemini 2.5 Deep Think: The Next Leap in AI Problem Solving Artificial intelligence is evolving from simply providing answers to actively reasoning through complex problems. Google's latest Gemini 2.5 Deep Think update exemplifies this shift, offering Google A... AI AI safety coding Deep Think Gemini problem solving reinforcement learning research tools
Google Gemini’s New Photo-to-Video Tool Unlocks Creative Freedom Turning a simple photograph into a captivating, animated video is now just a few taps away. Google Gemini's innovative photo-to-video feature empowers anyone to breathe life into their favorite images... AI safety AI video creative tools digital watermarking Google Gemini photo animation Veo 3
When AI Becomes the Insider Threat: Lessons from Agentic Misalignment Research As organizations hand more autonomy to AI systems, a pressing issue emerges: what if these intelligent tools act in ways that actively undermine their users? Recent research from Anthropic explores th... agentic misalignment AI alignment AI ethics AI safety corporate security insider threats LLMs
Anthropic Expands Bug Bounty Program to Strengthen AI Safety Anthropic is taking a bold step in AI safety by inviting the world’s top security researchers to put its latest defenses to the test. Their new bug bounty program aims to uncover serious vulnerabiliti... AI safety Anthropic bug bounty CBRN Claude 3.7 Sonnet Constitutional Classifiers Responsible Scaling security research