DATAGEN is a powerful brand name that represents our vision of leveraging artificial intelligence technology for data generation and analysis. The name combines "DATA" and "GEN"(generation), perfectly embodying the core functionality of this project - automated data analysis and research through a multi-agent system.
Visit us at DATAGEN Digital(website under development) to learn more about our vision and services.
DATAGEN is an advanced AI-powered data analysis and research platform that utilizes multiple specialized agents to streamline tasks such as data analysis, visualization, and report generation. Our platform leverages cutting-edge technologies including LangChain, OpenAI's GPT models, and LangGraph to handle complex research processes, integrating diverse AI architectures for optimal performance.
- Advanced Hypothesis Engine
- AI-driven hypothesis generation and validation
- Automated research direction optimization
- Real-time hypothesis refinement
- Enterprise Data Processing
- Robust data cleaning and transformation
- Scalable analysis pipelines
- Automated quality assurance
- Dynamic Visualization Suite
- Interactive data visualization
- Custom report generation
- Automated insight extraction
- Multi-Agent Intelligence
- Specialized agents for diverse tasks
- Intelligent task distribution
- Real-time coordination and optimization
- Smart Memory Management
- State-of-the-art Note Taker agent
- Efficient context retention system
- Seamless workflow integration
- Adaptive Processing Pipeline
- Dynamic workflow adjustment
- Automated resource optimization
- Real-time performance monitoring
DATAGEN revolutionizes data analysis through its innovative multi-agent architecture and intelligent automation capabilities:
-
Advanced Multi-Agent System
- Specialized agents working in harmony
- Intelligent task distribution and coordination
- Real-time adaptation to complex analysis requirements
-
Smart Context Management
- Pioneering Note Taker agent for state tracking
- Efficient memory utilization and context retention
- Seamless integration across analysis phases
-
Enterprise-Grade Performance
- Robust and scalable architecture
- Consistent and reliable outcomes
- Production-ready implementation
- Python 3.10 or higher
- Clone the repository:
git clone https://github.com/starpig1129/DATAGEN.git- Create and activate a Conda virtual environment:
conda create -n datagen python=3.10
conda activate datagen- Install dependencies:
pip install -r requirements.txt- Set up environment variables:
Rename
.env Exampleto.envand fill all the values
# Your data storage path(required)
DATA_STORAGE_PATH =./data/
# Conda environment name(required)
CONDA_ENV = datagen
# ChromeDriver executable path(required)
CHROMEDRIVER_PATH =./chromedriver-linux64/chromedriver
# Firecrawl API key (optional)
# Note: If this key is missing, query capabilities may be reduced
FIRECRAWL_API_KEY = XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX
# OpenAI API key (optional)
OPENAI_API_KEY = XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX
# Anthropic API key (optional)
ANTHROPIC_API_KEY = XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX
# Google API key (optional)
GOOGLE_API_KEY = XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX
# LangChain API key (optional)
# Used for monitoring the processing
LANGCHAIN_API_KEY = XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXYou can run the system using main.py:
-
Place your data file (e.g., YourDataName.csv) in the data directory
-
Modify the user_input variable in the main() function of main.py:
user_input = '''
datapath:YourDataName.csv
Use machine learning to perform data analysis and write complete graphical reports
'''ˇ. Run the script:
python main.pyhypothesis_agent: Generates research hypothesesprocess_agent: Supervises the entire research processvisualization_agent: Creates data visualizationscode_agent: Writes data analysis codesearcher_agent: Conducts literature and web searchesreport_agent: Writes research reportsquality_review_agent: Performs quality reviewsnote_agent: Records the research process
The system uses LangGraph to create a state graph that manages the entire research process. The workflow includes the following steps:
- Hypothesis generation
- Human choice (continue or regenerate hypothesis)
- Processing (including data analysis, visualization, search, and report writing)
- Quality review
- Revision as needed
Users can customize each agent's language model provider and model configuration by editing the config/agent_models.yaml file. This allows for optimization of model selection and parameters based on different agent requirements.
Here's an example structure of config/agent_models.yaml:
agents:
hypothesis_agent:
provider: openai
model_config:
model: gpt-5-nano
temperature: 1.0
note_agent:
provider: google
model_config:
model: gemini-2.5-pro
temperature: 1.0
code_agent:
provider: anthropic
model_config:
model: claude-haiku-4-5
temperature: 1.0- provider: Specifies the language model provider to use (e.g., openai, google, anthropic, ollama, groq)
- model_config: Contains model-specific configuration parameters
model: The specific model name to usetemperature: Controls the randomness of model output (range: 0.0-2.0)
- Ensure you have sufficient API credits, as the system will make multiple API calls.
- The system may take some time to complete the entire research process, depending on the complexity of the task.
- WARNING: The agent system may modify the data being analyzed. It is highly recommended to backup your data before using this system.
- NoteTaker Efficiency Improvement
- Overall Runtime Optimization
- Refiner needs to be better
Pull requests are welcome. For major changes, please open an issue first to discuss what you would like to change.
This project is licensed under the MIT License - see the LICENSE file for details.
Here are some of my other notable projects:
ShareLMAPI is a local language model sharing API that uses FastAPI to provide interfaces, allowing different programs or device to share the same local model, thereby reducing resource consumption. It supports streaming generation and various model configuration methods.
- GitHub: ShareLMAPI
A powerful Discord bot based on multi-modal Large Language Models (LLM), designed to interact with users through natural language. It combines advanced AI capabilities with practical features, offering a rich experience for Discord communities.
- GitHub: ai-discord-bot-PigPig

