Files changed (1) hide show
  1. firered-ocr.py +7 -10
firered-ocr.py CHANGED
@@ -104,10 +104,10 @@ def make_ocr_message(
104
  # Convert to RGB
105
  pil_img = pil_img.convert("RGB")
106
 
107
- # Convert to base64 data URI (JPEG is faster than PNG for encoding)
108
  buf = io.BytesIO()
109
- pil_img.save(buf, format="JPEG", quality=95)
110
- data_uri = f"data:image/jpeg;base64,{base64.b64encode(buf.getvalue()).decode()}"
111
 
112
  # Return message in vLLM format
113
  return [
@@ -228,7 +228,7 @@ def main(
228
  image_column: str = "image",
229
  batch_size: int = 16,
230
  model: str = "FireRedTeam/FireRed-OCR",
231
- max_model_len: int = 32768,
232
  max_tokens: int = 8192,
233
  gpu_memory_utilization: float = 0.8,
234
  hf_token: str = None,
@@ -335,10 +335,7 @@ def main(
335
  processing_duration = datetime.now() - start_time
336
  processing_time_str = f"{processing_duration.total_seconds() / 60:.1f} min"
337
 
338
- # Add output column to dataset (remove existing column if present)
339
- if output_column in dataset.column_names:
340
- logger.info(f"Removing existing '{output_column}' column before adding new results")
341
- dataset = dataset.remove_columns([output_column])
342
  logger.info(f"Adding '{output_column}' column to dataset")
343
  dataset = dataset.add_column(output_column, all_outputs)
344
 
@@ -483,8 +480,8 @@ Examples:
483
  parser.add_argument(
484
  "--max-model-len",
485
  type=int,
486
- default=32768,
487
- help="Maximum model context length (default: 32768)",
488
  )
489
  parser.add_argument(
490
  "--max-tokens",
 
104
  # Convert to RGB
105
  pil_img = pil_img.convert("RGB")
106
 
107
+ # Convert to base64 data URI
108
  buf = io.BytesIO()
109
+ pil_img.save(buf, format="PNG")
110
+ data_uri = f"data:image/png;base64,{base64.b64encode(buf.getvalue()).decode()}"
111
 
112
  # Return message in vLLM format
113
  return [
 
228
  image_column: str = "image",
229
  batch_size: int = 16,
230
  model: str = "FireRedTeam/FireRed-OCR",
231
+ max_model_len: int = 8192,
232
  max_tokens: int = 8192,
233
  gpu_memory_utilization: float = 0.8,
234
  hf_token: str = None,
 
335
  processing_duration = datetime.now() - start_time
336
  processing_time_str = f"{processing_duration.total_seconds() / 60:.1f} min"
337
 
338
+ # Add output column to dataset
 
 
 
339
  logger.info(f"Adding '{output_column}' column to dataset")
340
  dataset = dataset.add_column(output_column, all_outputs)
341
 
 
480
  parser.add_argument(
481
  "--max-model-len",
482
  type=int,
483
+ default=8192,
484
+ help="Maximum model context length (default: 8192)",
485
  )
486
  parser.add_argument(
487
  "--max-tokens",