SeeAct is designed to empower vision-language agents with a two-stage pipeline: a planning module powered by large language models generates subgoals based on observed scenes, and an execution module translates subgoals into environment-specific actions. A perception backbone extracts object and scene features from images or simulations. The modular architecture allows easy replacement of planners or perception networks and supports evaluation on AI2-THOR, Habitat, and custom environments. SeeAct accelerates research on interactive embodied AI by providing end-to-end task decomposition, grounding, and execution.