- Singapore
-
12:20
(UTC +08:00) - https://26hzhang.github.io/
- https://orcid.org/0000-0002-2725-6458
- @hzhang26
- in/hzhang26
Highlights
- Pro
Lists (1)
Sort Name ascending (A-Z)
Stars
📚 A curated list of Awesome Efficient dLLMs Papers with Codes
Elevate your AI research writing, no more tedious polishing ✨
An agentic skills framework & software development methodology that works.
Ralph is an autonomous AI agent loop that runs repeatedly until all PRD items are complete.
Claude Code skill implementing Manus-style persistent markdown planning — the workflow pattern behind the $2B acquisition.
A collection of notebooks/recipes showcasing some fun and effective ways of using Claude.
Transforms complex documents like PDFs into LLM-ready markdown/JSON for your Agentic workflows.
Agentic Design Patterns: A Hands-On Guide to Building Intelligent Systems by Antonio Gulli
Official repository for DR Tulu: Reinforcement Learning with Evolving Rubrics for Deep Research
sail-sg / Precision-RL-verl
Forked from verl-project/verlDefeating the Training-Inference Mismatch via FP16
MrlX: A Multi-Agent Reinforcement Learning Framework
LLM Council works together to answer your hardest questions
A paper list for spatial reasoning
This repository is maintained to release dataset and models for multimodal puzzle reasoning.
We introduce 'Thinking with Video', a new paradigm leveraging video generation for multimodal reasoning. Our VideoThinkBench shows that Sora-2 surpasses GPT5 by 10% on eyeballing puzzles and reache…
The absolute trainer to light up AI agents.
OmniGen2: Exploration to Advanced Multimodal Generation. https://arxiv.org/abs/2506.18871
Fully Local Manus AI. No APIs, No $200 monthly bills. Enjoy an autonomous agent that thinks, browses the web, and code for the sole cost of electricity. 🔔 Official updates only via twitter @Martin9…
Code repo for FaStfact: Faster, Stronger Long-Form Factuality Evaluations in LLMs.
A simple, unified multimodal models training engine. Lean, flexible, and built for hacking at scale.
Qwen3-VL is the multimodal large language model series developed by Qwen team, Alibaba Cloud.
[NeurIPS 2025] Scaling Language-centric Omnimodal Representation Learning
[NeurIPS 2025] Implementation for the paper "The Surprising Effectiveness of Negative Reinforcement in LLM Reasoning"