Wednesday April 15, 2026

Google launches one-click LLM workflows in Chrome, ASI-Evolve automates the AI research loop, and AgentFM transforms idle GPUs into a P2P AI grid.

Interested in AI engineering? Let's talk

News

An AI Vibe Coding Horror Story

A medical professional replaced an industry-standard system with a custom application built via AI coding agents, resulting in a critical security failure. The architecture relied on client-side authentication and a managed database lacking row-level security, exposing sensitive patient data and audio transcriptions to the public internet. This "vibe coding" incident highlights the risks of LLM-driven development by non-experts, particularly regarding data privacy, regulatory compliance, and insecure API integrations.

Turn your best AI prompts into one-click tools in Chrome

Google has introduced "Skills in Chrome," a feature that allows users to save and automate LLM prompts as reusable, one-click workflows. Integrated directly into Gemini in Chrome, these Skills can be triggered via a forward slash command to process content on the current page or across multiple tabs. The update includes a library of pre-configured workflows for common tasks and is currently rolling out to desktop users in the US.

DaVinci Resolve – Photo

DaVinci Resolve's new Photo page brings node-based color grading and Hollywood-grade tools to still photography, supporting non-destructive RAW processing at resolutions up to 32K. The platform leverages the DaVinci Neural Engine for AI-driven features including Magic Mask for subject isolation, Depth Map generation, Relight FX, and AI SuperScale upscaling. Performance is optimized through GPU acceleration (CUDA, Metal, and OpenCL) and includes IntelliSearch for semantic asset management alongside cloud-based collaborative workflows.

Two Months After I Gave an AI $100 and No Instructions

The ALMA project is a two-month experiment where an autonomous agent powered by Claude (Opus and Sonnet) was given $100 in crypto and internet access without any specific instructions or goals. Utilizing the OpenClaw framework and persistent memory files, the agent autonomously synthesized Hacker News trends into original essays and executed on-chain donations to various charities. The results indicate that unconstrained LLM agents tend to mirror their training-set ethics and, in the absence of external feedback, eventually converge on stable, routine behavioral patterns.

Apple App Store threatened to remove Grok over deepfakes: Letter

Apple threatened to remove xAI’s Grok from the App Store in January due to the model's ability to generate sexualized deepfakes. In a letter to senators, Apple cited violations of its guidelines after xAI failed to implement sufficient safeguards against NSFW content generation.

Research

Aethon: A reference-based instantiation primitive for stateful AI agents

Aethon introduces a reference-based replication primitive for stateful AI agents, replacing materialization-heavy instantiation with near-constant-time compositional views. By utilizing layered inheritance and copy-on-write semantics, it decouples creation costs from agent structure to enable scalable, lightweight multi-agent orchestration. This system abstraction addresses the latency and memory overheads inherent in current agentic infrastructure.

ASI-Evolve: AI Accelerates AI

ASI-Evolve is an agentic framework for AI-for-AI research, designed to accelerate AI development in costly, long-horizon research loops. It employs a learn-design-experiment-analyze cycle, augmenting evolutionary agents with a cognition base for human priors and an analyzer for distilling insights. The framework demonstrated significant AI-driven discoveries across neural architecture design (e.g., 105 SOTA linear attention models, surpassing DeltaNet), pretraining data curation (e.g., +3.96 points average, >18 points on MMLU), and RL algorithm design (outperforming GRPO). This suggests ASI-Evolve is a promising step towards closed-loop AI research.

Security Concerns in Generative AI Coding Assistants

This research explores security concerns surrounding GenAI-based coding assistants, such as GitHub Copilot, by analyzing developer discussions on platforms like Stack Overflow, Reddit, and Hacker News. The study identified four major areas of concern: potential data leakage, code licensing issues, adversarial attacks (e.g., prompt injection), and insecure code suggestions. These findings underscore critical limitations and trade-offs of GenAI in software engineering, highlighting key areas for improving their built-in security features.

Can LLMs Perform Synthesis?

Symbolic tools outperform or match frontier LLMs like GPT-5 and Qwen-32B across diverse program synthesis domains, including LTL and SyGuS. Despite LLMs utilizing superior hardware, symbolic tools demonstrate higher benchmark success rates and significantly faster execution times than both iterative and single-shot LLM approaches.

A 129FPS Full HD Real-Time Accelerator for 3D Gaussian Splatting

Researchers developed a low-power 3DGS hardware accelerator and compression pipeline achieving a 51.6x model size reduction via iterative pruning and vector quantization. The architecture optimizes rendering by skipping zero-Jacobian computations and utilizing comparison-free tile-based sorting. Implemented in 28-nm, the design supports 1080p at 129 FPS, offering 7.5x higher energy efficiency and 5.94x higher throughput than existing 3DGS accelerators.

Code

LangAlpha – what if Claude Code was built for Wall Street?

LangAlpha is a financial investment agent harness built on LangGraph that utilizes persistent workspaces and Programmatic Tool Calling (PTC) to enable iterative, Bayesian research. It leverages sandboxed environments to execute Python code for data processing via MCP servers, reducing token waste while supporting complex multi-step analysis. Key technical features include an agent swarm for parallel subtask execution, a 24-layer middleware stack for live steering and context management, and a multi-tier data provider hierarchy.

A memory database that forgets, consolidates, and detects contradiction

YantrikDB is a Rust-based cognitive memory engine designed to move beyond simple vector search by implementing automated memory consolidation, temporal decay, and factual contradiction detection. It utilizes a multi-index architecture—combining HNSW vector search with graph, temporal, and decay indexes—to provide relevance-conditioned scoring that optimizes context window usage. Available as an embedded library, network server, or MCP server, it supports CRDT-based sync and provides significant token savings over traditional RAG or file-based memory approaches.

ClawRun – Deploy and manage AI agents in seconds

This document provides the technical specifications and usage instructions for the CLI package, focusing on installation, configuration, and command execution for LLM-related development workflows.

AgentFM – A single Go binary that turns idle GPUs into a P2P AI grid

AgentFM is a decentralized P2P network built in Go that transforms idle hardware into a secure AI supercomputer for running LLMs and custom agents. It leverages libp2p for zero-config networking and NAT traversal, enabling task routing across a global mesh or private encrypted swarms. Workers execute workloads in ephemeral Podman sandboxes, providing real-time hardware telemetry and live artifact streaming back to orchestrator nodes.

Labeling Copilot: An agent for automated data curation in computer vision

Labeling Copilot is a deep research agent for automated computer vision data curation, orchestrating end-to-end planning, retrieval, annotation, and augmentation of image datasets. It integrates three core model services: CLIP-indexed image retrieval with active learning, multi-model detection and segmentation (e.g., Detic, SAM), and vision-LM-driven image synthesis. Designed to be driven by external coding agents (e.g., LLMs) via HTTP services, it produces COCO-formatted output with detection provenance.