TLDR: DeepVIS introduces a Chain-of-Thought (CoT) reasoning process into Natural Language to Visualization (NL2VIS) systems. This approach breaks down visualization generation into five transparent steps, enhancing model performance and allowing users to understand, identify errors, and refine AI-generated visualizations through an interactive interface. A new dataset, nvBench-CoT, supports this method, leading to improved accuracy and user trust.
Data visualization is a powerful tool for uncovering patterns and communicating insights from complex datasets. However, creating effective visualizations often requires specialized knowledge of authoring tools, which can interrupt the flow of data analysis. While large language models (LLMs) have shown great promise in automatically converting natural language descriptions into visualizations, existing methods often operate as ‘black boxes’. This means users cannot understand how or why specific visualization choices were made, making it difficult to trust the output or refine suboptimal results.
To address this challenge, a new approach called DeepVIS integrates Chain-of-Thought (CoT) reasoning into the Natural Language to Visualization (NL2VIS) pipeline. CoT reasoning encourages LLMs to break down complex problems into a series of intermediate, transparent steps, enhancing both model performance and user understanding.
The Five-Stage Chain-of-Thought Process
The core of DeepVIS is a comprehensive five-stage CoT reasoning process, designed after reviewing existing literature and conducting interviews with visualization experts. This structured approach mimics how experienced analysts design visualizations:
- S1: Determine chart type: The initial step involves selecting the most appropriate visualization method based on the data and the user’s analysis goal.
- S2: Retrieve relevant data: This stage focuses on identifying and extracting the specific data attributes, tables, and conditions necessary for the visualization.
- S3: Define data granularity: Here, the system determines how to group and aggregate data, for example, by grouping categorical data or binning time-based data.
- S4: Refine data for visualization: This step applies transformations like filtering and sorting to prepare the final data for optimal display.
- S5: Generate visualization: The final stage synthesizes all previous reasoning results to configure and produce the complete visualization, ensuring it effectively communicates the intended insights.
This step-by-step breakdown makes the entire process transparent, allowing users to see the rationale behind each decision.
Building the nvBench-CoT Dataset
To train models to follow this reasoning process, the researchers developed an automatic pipeline to augment existing NL2VIS datasets with structured CoT reasoning steps. This pipeline includes a database description module, which provides essential details like schema and value samples, and a reasoning steps generation module, which uses a powerful LLM (GPT-4o-mini) to create the detailed CoT steps based on the ground truth visualization query language (VQL). The resulting dataset, nvBench-CoT, is crucial for fine-tuning models to learn these reasoning patterns.
DeepVIS: An Interactive Visual Interface
DeepVIS is an interactive visual interface that tightly integrates with the CoT reasoning process. It allows users to inspect each reasoning step, identify potential errors, and make targeted adjustments to improve visualization outcomes. The interface features:
- CoT View: Provides a structured, hierarchical overview of the model’s reasoning process, allowing users to expand or collapse nodes for details.
- Information View: Displays comprehensive reasoning text for any selected step.
- Chart View: Renders the final visualization based on the generated VQL.
- Table View: Shows the underlying dataset, enabling users to verify how raw data translates into visual elements.
DeepVIS also offers interactive refinement mechanisms: ‘Self correction’ prompts the model to reconsider decisions within a selected step, while ‘Manual correction’ allows users to provide specific preferences to steer the regeneration process. After corrections, the tool intelligently regenerates subsequent steps to maintain logical consistency, highlighting changes for easy comparison.
Also Read:
- Improving AI Explanations: CoRGI Introduces Visual Grounding to Chain-of-Thought
- New Benchmark Reveals Visual Language Models Struggle with Complex Graphic Reasoning, But New Methods Show Promise
Evaluation and Impact
Quantitative evaluations show that the NL2VIS-CoT method significantly outperforms traditional black-box methods and other LLM-based approaches in terms of accuracy across various metrics, including chart type, axis configuration, and data accuracy. This highlights the effectiveness of incorporating CoT reasoning. User studies further confirm that the transparent reasoning process in DeepVIS enhances user understanding of model behavior, facilitates error identification, and improves the efficiency of refining visualizations. Users reported increased trust in the model outputs and found the detailed reasoning steps valuable for learning and improving their own visualization expertise.
DeepVIS represents a significant step towards more transparent and controllable AI-powered data visualization tools, fostering better human-AI collaboration in data analysis. For more details, you can refer to the original research paper.


