qwen2-vl-7b-pdf-weka.yaml 1.1 KB

1234567891011121314151617181920212223242526272829303132333435363738394041
  1. model:
  2. # full fine tune
  3. name_or_path: weka://oe-data-default/jakep/Qwen_Qwen2-VL-7B-Instruct-e4ecf8-01JAH8GMWHTJ376S2N7ETXRXH4/best_bf16/
  4. #name_or_path: s3://ai2-oe-data/jakep/experiments/qwen2vl-pdf/v1/models/jakep/Qwen_Qwen2-VL-7B-Instruct-e4ecf8-01JAH8GMWHTJ376S2N7ETXRXH4/checkpoint-9500/bf16/
  5. vlm: true
  6. # necessary to prevent random crashes, until vllm fixes some bugs
  7. num_scheduler_steps: 1
  8. format:
  9. add_generation_prompt: true
  10. generate:
  11. # The model's max context length is 8096, but around 1500 tokens are reserved for the image itself
  12. max_context_length: 6500
  13. temperature: 0.8
  14. top_p: 1.0
  15. drop_long_outputs: false
  16. pipeline:
  17. sqs_queue_name: jake-pdf
  18. num_workers: 3
  19. generation_batch_size: 256
  20. tokenization_batch_size: 64
  21. output_serializer: default
  22. target_bucket: ai2-oe-data
  23. target_object_prefix: [your username]/pdfworkspaces/s2orc_3200k_v2/inference_outputs
  24. allowed_restarts_per_predictor: 10
  25. task:
  26. budget: ai2/oe-data
  27. workspace: ai2/oe-data-model-based-cleanup
  28. name: qwen2vl-schedsteps-bg
  29. replicas: 128
  30. priority: LOW
  31. gpu_count: 1
  32. cluster:
  33. - ai2/jupiter-cirrascale-2
  34. - ai2/saturn-cirrascale