Skip to content

Conversation

@patdevinwilson
Copy link
Contributor

  • Problem: No validation for large scale factors
  • Solution: Automatic resource checks + comprehensive docs
  • Testing: Successfully running SF3000 generation
  • Impact: 6 files, +613 lines, fully backward compatible
  • Benefits: Prevents failures, better UX, automation-friendly

- Add SF3000_SUPPORT.md with comprehensive documentation
- Update setup_benchmark_data_and_tables.sh with automatic warnings for large scale factors
- Add validation in generate_data_files.py for SF3000+ generations
- Update README.md files with SF3000 examples and requirements
- Automatic checks for disk space, memory, and time requirements
- Interactive confirmation prompts for SF3000+ generations

Users can now generate SF3000 (~3TB) datasets with appropriate warnings about:
- 4TB+ disk space requirements
- 256GB+ RAM recommendations
- 8-24 hour generation time
- Resource availability checks
- Add SF3000 example command with resource requirements comment
- Add SCALE FACTORS section showing requirements for different scale factors
- Update both setup scripts to display comprehensive scale factor information
- Help text now shows: SF1-SF100, SF1000 (~1TB), SF3000 (~3TB) requirements
- Add sys.stdin.isatty() check to detect non-interactive mode
- Auto-proceed with SF3000 generation when running in background
- Fixes EOFError when running via nohup or automated scripts
- Interactive mode still prompts for confirmation as before
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant