| #!/usr/bin/env ruby |
| require_relative 'config' |
| require 'open3' |
| require 'json' |
|
|
| |
| |
| |
| |
| |
| def copy_task_without_test_output(source_task_json_path, destination_task_json_path) |
| if source_task_json_path == destination_task_json_path |
| raise "the paths are supposed to be different. #{source_task_json_path}" |
| end |
| json_string = IO.read(source_task_json_path) |
| json = JSON.parse(json_string) |
| test_pairs = json['test'] |
| test_pairs.each do |pair| |
| pair['output'] = [] |
| end |
| File.write(destination_task_json_path, JSON.dump(json)) |
| end |
|
|
| |
| |
| |
| def size_from_json_image(rows) |
| columns_min = 255 |
| columns_max = 0 |
| rows.each do |row| |
| columns_max = [columns_max, row.count].max |
| columns_min = [columns_min, row.count].min |
| end |
| if columns_min != columns_max |
| raise "the columns are supposed to have the same length. #{task_json_path}" |
| end |
| width = columns_min |
| height = rows.count |
| "#{width}x#{height}" |
| end |
|
|
| def same_size_for_input_and_output_in_task(task_json_path) |
| json_string = IO.read(task_json_path) |
| json = JSON.parse(json_string) |
| sizes_input = [] |
| sizes_output = [] |
| json['train'].each do |pair| |
| sizes_input << size_from_json_image(pair['input']) |
| sizes_output << size_from_json_image(pair['output']) |
| end |
| json['test'].each do |pair| |
| sizes_input << size_from_json_image(pair['input']) |
| sizes_output << size_from_json_image(pair['output']) |
| end |
| sizes_input == sizes_output |
| end |
|
|
| |
| |
| |
| def sizes_from_task(task_json_path) |
| json_string = IO.read(task_json_path) |
| json = JSON.parse(json_string) |
| test_pairs = json['test'] |
| sizes = [] |
| test_pairs.each do |pair| |
| rows = pair['output'] |
| sizes << size_from_json_image(rows) |
| end |
| sizes |
| end |
|
|
| |
| |
| |
| def predicted_sizes(json_string) |
| json = JSON.parse(json_string) |
| test_pairs = json['test'] |
| sizes = [] |
| test_pairs.each do |pair| |
| dict = pair['output_size'] |
| width = dict['width'].to_i |
| height = dict['height'].to_i |
| sizes << "#{width}x#{height}" |
| end |
| sizes |
| end |
|
|
| OUTPUT_DIR = File.expand_path("data/arc_size") |
| TEMP_PATH = File.join(OUTPUT_DIR, 'temp.json') |
|
|
| LODA_RUST_EXECUTABLE = Config.instance.loda_rust_executable |
| unless File.executable?(LODA_RUST_EXECUTABLE) |
| raise "No such file #{LODA_RUST_EXECUTABLE}, cannot run script" |
| end |
|
|
| ARC_REPOSITORY_DATA = Config.instance.arc_repository_data |
| unless File.directory?(ARC_REPOSITORY_DATA) |
| raise "No such dir #{ARC_REPOSITORY_DATA}, cannot run script" |
| end |
|
|
| if File.directory?(OUTPUT_DIR) |
| raise "The OUTPUT_DIR #{OUTPUT_DIR} already exist. Please delete it manually, and try again." |
| end |
|
|
| FileUtils.mkdir_p(OUTPUT_DIR) |
| unless File.directory?(OUTPUT_DIR) |
| raise "unable to create dir: #{OUTPUT_DIR}" |
| end |
|
|
| count_tasks = 0 |
| count_ok_predictions = 0 |
| count_bad_predictions = 0 |
| count_cannot_predict = 0 |
| count_other_errors = 0 |
| count_same_size = 0 |
| Dir.chdir(ARC_REPOSITORY_DATA) do |
| paths = Dir.glob("**/*.json") |
|
|
| |
| paths = paths.reject { |path| File.basename(path) == 'solution_notXORdinary.json' } |
| |
| paths.each_with_index do |path, index| |
| if index % 100 == 0 |
| puts "Progress: #{index} of #{paths.count}" |
| end |
| |
| if same_size_for_input_and_output_in_task(path) |
| count_same_size += 1 |
| end |
| |
| |
| expected_sizes = sizes_from_task(path) |
| |
| |
| copy_task_without_test_output(path, TEMP_PATH) |
| |
| |
| output_path = File.join(OUTPUT_DIR, path) |
| output_dirname = File.dirname(output_path) |
| FileUtils.mkdir_p(output_dirname) |
| unless File.directory?(output_dirname) |
| raise "unable to create dir: #{output_dirname}" |
| end |
| |
| |
| command = "#{LODA_RUST_EXECUTABLE} arc-size #{TEMP_PATH}" |
| stdout_and_stderr, status = Open3.capture2e(command) |
| output = stdout_and_stderr |
| count_tasks += 1 |
|
|
| unless status.success? |
| if output.include?('Cannot predict the output sizes') |
| output_path2 = output_path.gsub(/[.]json$/, '-cannot-predict.txt') |
| IO.write(output_path2, stdout_and_stderr) |
| count_cannot_predict += 1 |
| next |
| else |
| output_path2 = output_path.gsub(/[.]json$/, '-error.txt') |
| IO.write(output_path2, stdout_and_stderr) |
| count_other_errors += 1 |
| next |
| end |
| end |
| json = stdout_and_stderr.strip |
| predicted_sizes = predicted_sizes(json) |
| if predicted_sizes != expected_sizes |
| |
| output_path2 = output_path.gsub(/[.]json$/, '-bad-prediction.txt') |
| error_message = stdout_and_stderr + "\n\n--\nThis is a bad prediction!\nPredicted #{predicted_sizes}. But the actual size is #{expected_sizes}" |
| IO.write(output_path2, error_message) |
| count_bad_predictions += 1 |
| next |
| end |
| IO.write(output_path, json) |
| count_ok_predictions += 1 |
| next |
| end |
| end |
|
|
| File.delete(TEMP_PATH) if File.exist?(TEMP_PATH) |
|
|
| puts |
| puts "count_tasks: #{count_tasks} The number of tasks processed." |
| puts "count_ok_predictions: #{count_ok_predictions} Predictions that matches with the actual data." |
| puts "count_bad_predictions: #{count_bad_predictions} Predictions that are different than the actual data." |
| puts "count_cannot_predict: #{count_cannot_predict} Unable to make a prediction. Insufficient data, lack of algorithms for predicting." |
| puts "count_other_errors: #{count_other_errors} Something went wrong." |
| puts "count_same_size: #{count_same_size} Number of tasks where input size and output size are the same." |
|
|