New /less-permission-prompts skill scans your Claude Code transcripts for common read-only Bash and MCP tool calls, then proposes a prioritized allowlist for .claude/settings.json.
CLAUDE / BASH
# Run after 5-10 sessions/less-permission-prompts# Review the proposed allowlist it outputs# Accept to write to .claude/settings.json
Claude Code adds xhigh effort level for Opus 4.7 — sits between high and max. Available via /effort, --effort flag, and the model picker. Other models fall back to high.
CLAUDE / BASH
# Launch with xhigh effortclaude --effort xhigh# Or change mid-session/effort xhigh# Interactive slider (new)/effort
Claude Design now exports a handoff bundle that Claude Code can ingest directly. This replaces manual copying of specs, design tokens, and component structure.
CLAUDE / BASH
# In your project directoryunzip ~/Downloads/claude-design-handoff.zip -d ./design-handoffclaude "Implement the design in ./design-handoff using our existing component library. Follow the design tokens in design-handoff/tokens.json. Generate React + Tailwind."
Opus 4.7 uses the model string claude-opus-4-7 and supports 1M token context by default. Vision inputs now accept up to 3.75MP images for pixel-accurate UI reasoning.
Extended thinking in claude-sonnet-4-6 now streams at 2x the previous throughput. Internal reasoning tokens arrive in real-time via the thinking content block, with no additional latency penalty on the first token.
CLAUDE / TYPESCRIPT
import Anthropic from "@anthropic-ai/sdk";const client = new Anthropic();async function streamWithThinking(prompt: string) { const stream = await client.messages.stream({ model: "claude-sonnet-4-6", max_tokens: 16000, thinking: { type: "enabled", budget_tokens: 10000, }, messages: [{ role: "user", content: prompt }], }); let thinkingText = ""; let responseText = ""; for await (const event of stream) { if (event.type === "content_block_delta") { if (event.delta.type === "thinking_delta") { thinkingText += event.delta.thinking; process.stdout.write("\x1b[2m"); // dim process.stdout.write(event.delta.thinking); process.stdout.write("\x1b[0m"); } else if (event.delta.type === "text_delta") { responseText += event.delta.text; process.stdout.write(event.delta.text); } } } return { thinking: thinkingText, response: responseText };}streamWithThinking( "Design a database schema for a multi-tenant SaaS app with row-level security.");
Google has made grounding with Google Search free for up to 1,500 queries per day on Gemini 2.5 Pro via the Gemini API. Previously this was a paid add-on. Beyond 1,500 queries, standard grounding rates apply.
GEMINI / PYTHON
import google.generativeai as genaiimport osgenai.configure(api_key=os.environ["GEMINI_API_KEY"])model = genai.GenerativeModel( model_name="gemini-2.5-pro", tools=[{"google_search": {}}],)response = model.generate_content( "What are the latest updates to Claude's API in 2026?", generation_config=genai.GenerationConfig( temperature=0.1, ),)# Print grounded responseprint(response.text)# Access grounding metadata (sources)if response.candidates[0].grounding_metadata: for chunk in response.candidates[0].grounding_metadata.grounding_chunks: print(f"Source: {chunk.web.uri}") print(f"Title: {chunk.web.title}")
ElevenLabs now supports Instant Voice Cloning from as little as 10 seconds of audio via their API. Upload a clean audio sample, get a voice_id back in under 3 seconds, and immediately use it for text-to-speech generation.
ELEVENLABS / PYTHON
import requestsimport osELEVEN_API_KEY = os.environ["ELEVEN_API_KEY"]# Step 1: Clone a voice from a short audio sampledef clone_voice(name: str, audio_path: str) -> str: with open(audio_path, "rb") as f: response = requests.post( "https://api.elevenlabs.io/v1/voices/add", headers={"xi-api-key": ELEVEN_API_KEY}, data={"name": name}, files={"files": (audio_path, f, "audio/mpeg")}, ) response.raise_for_status() return response.json()["voice_id"]# Step 2: Generate speech with the cloned voicedef speak(voice_id: str, text: str, output_path: str): response = requests.post( f"https://api.elevenlabs.io/v1/text-to-speech/{voice_id}", headers={ "xi-api-key": ELEVEN_API_KEY, "Content-Type": "application/json", }, json={ "text": text, "model_id": "eleven_multilingual_v2", "voice_settings": {"stability": 0.5, "similarity_boost": 0.8}, }, ) response.raise_for_status() with open(output_path, "wb") as f: f.write(response.content)voice_id = clone_voice("My Narrator", "sample.mp3")speak(voice_id, "Welcome to the future of voice AI.", "output.mp3")
Anthropic released Claude Sonnet 4.6, the latest in the Claude 4 family, with improved reasoning, faster response times, and better instruction following compared to Sonnet 3.7.
Claude Code, Anthropic's agentic coding tool that runs in the terminal, exited beta and is now generally available. It can edit files, run tests, commit code, and navigate large codebases autonomously.
CLAUDE / BASH
# Installnpm install -g @anthropic-ai/claude-code# Navigate to your project and startcd your-projectclaude# Give it a task# > Add rate limiting to the /api/auth/login endpoint using Redis
Google made the Search grounding feature free for Gemini API users up to 1,500 queries per day, letting the model cite real-time web sources.
GEMINI / PYTHON
import google.generativeai as genaigenai.configure(api_key=os.environ["GEMINI_API_KEY"])model = genai.GenerativeModel("gemini-2.5-pro")response = model.generate_content( "What happened in AI this week?", tools=[{"google_search_retrieval": {}}])print(response.text)# Access source citationsfor chunk in response.candidates[0].grounding_metadata.grounding_chunks: print(f"Source: {chunk.web.uri}")
Vercel released AI SDK 4.0 with a unified API that works identically across Claude, GPT-4o, Gemini, Mistral, and Llama. Includes streaming, tool use, and structured output.
OTHER / TYPESCRIPT
import { generateText } from "ai";import { anthropic } from "@ai-sdk/anthropic";const { text } = await generateText({ model: anthropic("claude-sonnet-4-6"), prompt: "Explain RAG in one paragraph"});console.log(text);// Swap provider in one line — same API// import { openai } from "@ai-sdk/openai";// model: openai("gpt-4o")
OpenAI's Realtime API now supports mixing text and audio modalities in the same WebSocket session. Send text, receive audio, or switch modes mid-conversation.