Hi, I'm Daniel Lee . Currently evolving from RAG to Graph RAG and Vision-Language Models (VLM/VLA) to bridge digital reasoning with physical action. Striving to be a Problem Definer who doesn't just solve tasks, but poses the next big challenges.
2025
- Top Excellence Award (1st Prize)
- Top Excellence Award (Institute for Information & Communication Technology Planning & Evaluation Director's Award)
- Silver Medal
- [HCLT 2025] Enhancing Multi-Hop Complex Query Retrieval Efficiency through the Integration of RAG and Graph RAG
IP-to-Portrait - High-Fidelity Face Synthesis Pipeline
- Advanced AI Pipeline: End-to-end face synthesis preserving identity, background, and lighting using SDXL Inpainting & IP-Adapter FaceID Plus v2.
- Multimodal Integration: Auto-prompting via Gemini 2.5 Flash VLM and precision masking with BiSeNet & InsightFace.
- Tech: Next.js, FastAPI, Celery, Redis, PyTorch, Diffusers, ONNX Runtime.
Research Focus
- Vision-Language Models (VLM) & VLA: Focus on Physical AI and autonomous robot control
- Agentic Systems: Developing autonomous decision-making loops and agentic workflows.
- RAG & Graph RAG: Exploring advanced retrieval and knowledge graph integration for agents.
|
AI apps, demos & services |
Impl of Multimodal model |
Scheduling, Logic, Multicycle |




