π§ Behind the Scenes CRAFT development: Sun-Mon only (full-time job Wed-Sat). 48 projects completed solo. The time constraint is the proofβif the methodology works under pressure, it works. Beta: February 2026 β craftframework.ai
reacted to tsungyi's
post with πabout 4 hours ago
Big news from CES β Cosmos Reason 2 is here β our most advanced reasoning vision-language model for physical AI, now topping the Physical AI Bench leaderboardπ shi-labs/physical-ai-bench-leaderboard
Whatβs new: - Enhanced physical reasoning & spatio-temporal understanding - Flexible deployment with 2B & 8B model sizes - Long-context understanding (up to 256K tokens) - Object detection with 2D/3D point localizations and trajectory data - New Cosmos Cookbook Recipes for faster onboarding
On top of Cosmos Reason 2, we also rolled out other new updates, including: - Cosmos Predict 2.5 β Unified Text2World/Image2World/Video2World model for higher-quality synthetic video worlds - Cosmos Transfer 2.5-2B β Lightweight, high-fidelity world-to-world translation with stronger physics alignment - NVIDIA GR00T N1.6 β Open robot foundation model for general-purpose robotic learning and control, integrated with Cosmos Reason
π Excited to share: The First vLLM container for NVIDIA DGX Spark!
I've been working on getting vLLM to run natively on the new DGX Spark with its GB10 Blackwell GPU (SM121 architecture). The results? 2.5x faster inference compared to llama.cpp!
π§ Technical Challenges Solved: β’ Built PyTorch nightly with CUDA 13.1 + SM121 support β’ Patched vLLM for Blackwell architecture β’ Created custom MoE expert configs for GB10 β’ Implemented TRITON_ATTN backend workaround
The DGX Spark's 119GB unified memory opens up possibilities for running massive models locally. Happy to connect with others working on the DGX Spark Blackwell!