[
  {
    "instance_id": "id1413",
    "train_trial_index": 0,
    "predicted_text": "The school's mascot.",
    "base64_images": [],
    "stats": {
      "num_prompt_tokens": 3511.0,
      "num_output_tokens": 5.0,
      "inference_runtime": 0.9186711311340332,
      "num_train_instances": 5.0,
      "prompt_truncated": 0.0,
      "f1_score": 1.0
    }
  },
  {
    "instance_id": "id1332",
    "train_trial_index": 0,
    "predicted_text": "Niece",
    "base64_images": [],
    "stats": {
      "num_prompt_tokens": 3375.0,
      "num_output_tokens": 2.0,
      "inference_runtime": 0.4032626152038574,
      "num_train_instances": 5.0,
      "prompt_truncated": 0.0,
      "f1_score": 0.6666666666666666
    }
  },
  {
    "instance_id": "id1123",
    "train_trial_index": 0,
    "predicted_text": "Barbarous",
    "base64_images": [],
    "stats": {
      "num_prompt_tokens": 3399.0,
      "num_output_tokens": 3.0,
      "inference_runtime": 0.889366865158081,
      "num_train_instances": 5.0,
      "prompt_truncated": 0.0,
      "f1_score": 0.6666666666666666
    }
  },
  {
    "instance_id": "id1514",
    "train_trial_index": 0,
    "predicted_text": "John Barton",
    "base64_images": [],
    "stats": {
      "num_prompt_tokens": 3496.0,
      "num_output_tokens": 2.0,
      "inference_runtime": 0.9155101776123047,
      "num_train_instances": 5.0,
      "prompt_truncated": 0.0,
      "f1_score": 1.0
    }
  },
  {
    "instance_id": "id1340",
    "train_trial_index": 0,
    "predicted_text": "The high priestess",
    "base64_images": [],
    "stats": {
      "num_prompt_tokens": 3042.0,
      "num_output_tokens": 4.0,
      "inference_runtime": 0.2613513469696045,
      "num_train_instances": 5.0,
      "prompt_truncated": 0.0,
      "f1_score": 1.0
    }
  },
  {
    "instance_id": "id1281",
    "train_trial_index": 0,
    "predicted_text": "Marrying someone while already being legally married to another person.",
    "base64_images": [],
    "stats": {
      "num_prompt_tokens": 3630.0,
      "num_output_tokens": 13.0,
      "inference_runtime": 1.0612504482269287,
      "num_train_instances": 5.0,
      "prompt_truncated": 0.0,
      "f1_score": 0.14285714285714285
    }
  },
  {
    "instance_id": "id1234",
    "train_trial_index": 0,
    "predicted_text": "Inspector Morton",
    "base64_images": [],
    "stats": {
      "num_prompt_tokens": 3321.0,
      "num_output_tokens": 2.0,
      "inference_runtime": 0.835554838180542,
      "num_train_instances": 5.0,
      "prompt_truncated": 0.0,
      "f1_score": 0.4
    }
  },
  {
    "instance_id": "id1237",
    "train_trial_index": 0,
    "predicted_text": "Gabriel",
    "base64_images": [],
    "stats": {
      "num_prompt_tokens": 3662.0,
      "num_output_tokens": 2.0,
      "inference_runtime": 0.25126194953918457,
      "num_train_instances": 5.0,
      "prompt_truncated": 0.0,
      "f1_score": 1.0
    }
  },
  {
    "instance_id": "id1173",
    "train_trial_index": 0,
    "predicted_text": "Mother Firefly's ex-husband",
    "base64_images": [],
    "stats": {
      "num_prompt_tokens": 3428.0,
      "num_output_tokens": 7.0,
      "inference_runtime": 1.0240693092346191,
      "num_train_instances": 5.0,
      "prompt_truncated": 0.0,
      "f1_score": 1.0
    }
  },
  {
    "instance_id": "id1483",
    "train_trial_index": 0,
    "predicted_text": "Sam blackmails David with knowledge of his prostitution attempt.",
    "base64_images": [],
    "stats": {
      "num_prompt_tokens": 3228.0,
      "num_output_tokens": 11.0,
      "inference_runtime": 0.9479615688323975,
      "num_train_instances": 5.0,
      "prompt_truncated": 0.0,
      "f1_score": 0.2
    }
  },
  {
    "instance_id": "id1474",
    "train_trial_index": 0,
    "predicted_text": "She was branded as a traitor to the Ozunu Clan.",
    "base64_images": [],
    "stats": {
      "num_prompt_tokens": 3561.0,
      "num_output_tokens": 13.0,
      "inference_runtime": 1.0144038200378418,
      "num_train_instances": 5.0,
      "prompt_truncated": 0.0,
      "f1_score": 0.7692307692307692
    }
  },
  {
    "instance_id": "id1571",
    "train_trial_index": 0,
    "predicted_text": "She informs him that she will be suing him.",
    "base64_images": [],
    "stats": {
      "num_prompt_tokens": 3706.0,
      "num_output_tokens": 10.0,
      "inference_runtime": 0.8954699039459229,
      "num_train_instances": 5.0,
      "prompt_truncated": 0.0,
      "f1_score": 0.3333333333333333
    }
  },
  {
    "instance_id": "id1366",
    "train_trial_index": 0,
    "predicted_text": "Because he learns that she is only after his money.",
    "base64_images": [],
    "stats": {
      "num_prompt_tokens": 3755.0,
      "num_output_tokens": 11.0,
      "inference_runtime": 0.3721733093261719,
      "num_train_instances": 5.0,
      "prompt_truncated": 0.0,
      "f1_score": 0.8235294117647058
    }
  },
  {
    "instance_id": "id1442",
    "train_trial_index": 0,
    "predicted_text": "By the Mayflower",
    "base64_images": [],
    "stats": {
      "num_prompt_tokens": 3557.0,
      "num_output_tokens": 4.0,
      "inference_runtime": 0.258669376373291,
      "num_train_instances": 5.0,
      "prompt_truncated": 0.0,
      "f1_score": 0.4
    }
  },
  {
    "instance_id": "id1529",
    "train_trial_index": 0,
    "predicted_text": "Charlie",
    "base64_images": [],
    "stats": {
      "num_prompt_tokens": 3308.0,
      "num_output_tokens": 1.0,
      "inference_runtime": 0.2193920612335205,
      "num_train_instances": 5.0,
      "prompt_truncated": 0.0,
      "f1_score": 1.0
    }
  },
  {
    "instance_id": "id1209",
    "train_trial_index": 0,
    "predicted_text": "Reception room",
    "base64_images": [],
    "stats": {
      "num_prompt_tokens": 3520.0,
      "num_output_tokens": 3.0,
      "inference_runtime": 0.2627694606781006,
      "num_train_instances": 5.0,
      "prompt_truncated": 0.0,
      "f1_score": 1.0
    }
  },
  {
    "instance_id": "id1408",
    "train_trial_index": 0,
    "predicted_text": "Derek attends Sara's audition and offers her encouragement and moral support.",
    "base64_images": [],
    "stats": {
      "num_prompt_tokens": 3277.0,
      "num_output_tokens": 14.0,
      "inference_runtime": 0.9844205379486084,
      "num_train_instances": 5.0,
      "prompt_truncated": 0.0,
      "f1_score": 0.3333333333333333
    }
  },
  {
    "instance_id": "id1315",
    "train_trial_index": 0,
    "predicted_text": "1,261 out of 10,000",
    "base64_images": [],
    "stats": {
      "num_prompt_tokens": 3150.0,
      "num_output_tokens": 9.0,
      "inference_runtime": 0.6034114360809326,
      "num_train_instances": 5.0,
      "prompt_truncated": 0.0,
      "f1_score": 1.0
    }
  },
  {
    "instance_id": "id1421",
    "train_trial_index": 0,
    "predicted_text": "Because she saw a poster for the film Gilda starring Rita Hayworth.",
    "base64_images": [],
    "stats": {
      "num_prompt_tokens": 3772.0,
      "num_output_tokens": 15.0,
      "inference_runtime": 0.409564733505249,
      "num_train_instances": 5.0,
      "prompt_truncated": 0.0,
      "f1_score": 0.7058823529411764
    }
  },
  {
    "instance_id": "id1308",
    "train_trial_index": 0,
    "predicted_text": "Fredrick Zoller",
    "base64_images": [],
    "stats": {
      "num_prompt_tokens": 3643.0,
      "num_output_tokens": 4.0,
      "inference_runtime": 0.2699699401855469,
      "num_train_instances": 5.0,
      "prompt_truncated": 0.0,
      "f1_score": 0.6666666666666666
    }
  },
  {
    "instance_id": "id1456",
    "train_trial_index": 0,
    "predicted_text": "Ancient Egyptian architecture",
    "base64_images": [],
    "stats": {
      "num_prompt_tokens": 3453.0,
      "num_output_tokens": 4.0,
      "inference_runtime": 0.8493340015411377,
      "num_train_instances": 5.0,
      "prompt_truncated": 0.0,
      "f1_score": 1.0
    }
  },
  {
    "instance_id": "id1335",
    "train_trial_index": 0,
    "predicted_text": "Falmouth",
    "base64_images": [],
    "stats": {
      "num_prompt_tokens": 3677.0,
      "num_output_tokens": 3.0,
      "inference_runtime": 0.8066091537475586,
      "num_train_instances": 5.0,
      "prompt_truncated": 0.0,
      "f1_score": 1.0
    }
  },
  {
    "instance_id": "id1337",
    "train_trial_index": 0,
    "predicted_text": "Reginald Bliss",
    "base64_images": [],
    "stats": {
      "num_prompt_tokens": 3390.0,
      "num_output_tokens": 4.0,
      "inference_runtime": 0.326526403427124,
      "num_train_instances": 5.0,
      "prompt_truncated": 0.0,
      "f1_score": 1.0
    }
  },
  {
    "instance_id": "id1114",
    "train_trial_index": 0,
    "predicted_text": "Charity",
    "base64_images": [],
    "stats": {
      "num_prompt_tokens": 3165.0,
      "num_output_tokens": 2.0,
      "inference_runtime": 0.23354363441467285,
      "num_train_instances": 5.0,
      "prompt_truncated": 0.0,
      "f1_score": 1.0
    }
  },
  {
    "instance_id": "id1198",
    "train_trial_index": 0,
    "predicted_text": "She is released by the FBI.",
    "base64_images": [],
    "stats": {
      "num_prompt_tokens": 3539.0,
      "num_output_tokens": 7.0,
      "inference_runtime": 0.8684811592102051,
      "num_train_instances": 5.0,
      "prompt_truncated": 0.0,
      "f1_score": 0.5
    }
  },
  {
    "instance_id": "id1178",
    "train_trial_index": 0,
    "predicted_text": "The Acorn Planters",
    "base64_images": [],
    "stats": {
      "num_prompt_tokens": 3100.0,
      "num_output_tokens": 5.0,
      "inference_runtime": 0.8762445449829102,
      "num_train_instances": 5.0,
      "prompt_truncated": 0.0,
      "f1_score": 1.0
    }
  },
  {
    "instance_id": "id1117",
    "train_trial_index": 0,
    "predicted_text": "Shoe-shining",
    "base64_images": [],
    "stats": {
      "num_prompt_tokens": 3494.0,
      "num_output_tokens": 4.0,
      "inference_runtime": 0.5429232120513916,
      "num_train_instances": 5.0,
      "prompt_truncated": 0.0,
      "f1_score": 0.0
    }
  },
  {
    "instance_id": "id1273",
    "train_trial_index": 0,
    "predicted_text": "By stealing all the Glengarry leads and selling them to a competing real estate agency.",
    "base64_images": [],
    "stats": {
      "num_prompt_tokens": 3783.0,
      "num_output_tokens": 18.0,
      "inference_runtime": 1.0977084636688232,
      "num_train_instances": 5.0,
      "prompt_truncated": 0.0,
      "f1_score": 0.8461538461538461
    }
  },
  {
    "instance_id": "id1180",
    "train_trial_index": 0,
    "predicted_text": "Special Agent in Charge Darius Michaud",
    "base64_images": [],
    "stats": {
      "num_prompt_tokens": 3820.0,
      "num_output_tokens": 8.0,
      "inference_runtime": 0.44613194465637207,
      "num_train_instances": 5.0,
      "prompt_truncated": 0.0,
      "f1_score": 0.5
    }
  },
  {
    "instance_id": "id1261",
    "train_trial_index": 0,
    "predicted_text": "Because he is in need of therapy.",
    "base64_images": [],
    "stats": {
      "num_prompt_tokens": 3755.0,
      "num_output_tokens": 8.0,
      "inference_runtime": 0.39412593841552734,
      "num_train_instances": 5.0,
      "prompt_truncated": 0.0,
      "f1_score": 0.14285714285714285
    }
  },
  {
    "instance_id": "id1290",
    "train_trial_index": 0,
    "predicted_text": "To avenge a fellow gangster's death at the hands of Ted and Officer Brazier.",
    "base64_images": [],
    "stats": {
      "num_prompt_tokens": 3616.0,
      "num_output_tokens": 19.0,
      "inference_runtime": 1.0970735549926758,
      "num_train_instances": 5.0,
      "prompt_truncated": 0.0,
      "f1_score": 0.5882352941176471
    }
  },
  {
    "instance_id": "id1202",
    "train_trial_index": 0,
    "predicted_text": "Charity was a librarian.",
    "base64_images": [],
    "stats": {
      "num_prompt_tokens": 3590.0,
      "num_output_tokens": 6.0,
      "inference_runtime": 0.9080181121826172,
      "num_train_instances": 5.0,
      "prompt_truncated": 0.0,
      "f1_score": 0.5454545454545455
    }
  },
  {
    "instance_id": "id1463",
    "train_trial_index": 0,
    "predicted_text": "Lydia and Jack",
    "base64_images": [],
    "stats": {
      "num_prompt_tokens": 3637.0,
      "num_output_tokens": 4.0,
      "inference_runtime": 0.8191604614257812,
      "num_train_instances": 5.0,
      "prompt_truncated": 0.0,
      "f1_score": 1.0
    }
  },
  {
    "instance_id": "id1112",
    "train_trial_index": 0,
    "predicted_text": "Misusing the symbol of Jhebbal Sag",
    "base64_images": [],
    "stats": {
      "num_prompt_tokens": 3563.0,
      "num_output_tokens": 10.0,
      "inference_runtime": 0.9107813835144043,
      "num_train_instances": 5.0,
      "prompt_truncated": 0.0,
      "f1_score": 0.7272727272727273
    }
  },
  {
    "instance_id": "id1293",
    "train_trial_index": 0,
    "predicted_text": "Eroshka",
    "base64_images": [],
    "stats": {
      "num_prompt_tokens": 3589.0,
      "num_output_tokens": 4.0,
      "inference_runtime": 0.27483654022216797,
      "num_train_instances": 5.0,
      "prompt_truncated": 0.0,
      "f1_score": 1.0
    }
  },
  {
    "instance_id": "id1256",
    "train_trial_index": 0,
    "predicted_text": "The Virgin Mary",
    "base64_images": [],
    "stats": {
      "num_prompt_tokens": 3207.0,
      "num_output_tokens": 3.0,
      "inference_runtime": 0.8442568778991699,
      "num_train_instances": 5.0,
      "prompt_truncated": 0.0,
      "f1_score": 1.0
    }
  },
  {
    "instance_id": "id1499",
    "train_trial_index": 0,
    "predicted_text": "Unprotected sex",
    "base64_images": [],
    "stats": {
      "num_prompt_tokens": 3605.0,
      "num_output_tokens": 3.0,
      "inference_runtime": 0.8656706809997559,
      "num_train_instances": 5.0,
      "prompt_truncated": 0.0,
      "f1_score": 1.0
    }
  },
  {
    "instance_id": "id1419",
    "train_trial_index": 0,
    "predicted_text": "Jungle",
    "base64_images": [],
    "stats": {
      "num_prompt_tokens": 3112.0,
      "num_output_tokens": 2.0,
      "inference_runtime": 0.2510643005371094,
      "num_train_instances": 5.0,
      "prompt_truncated": 0.0,
      "f1_score": 0.6666666666666666
    }
  },
  {
    "instance_id": "id1103",
    "train_trial_index": 0,
    "predicted_text": "Crack addiction",
    "base64_images": [],
    "stats": {
      "num_prompt_tokens": 3487.0,
      "num_output_tokens": 3.0,
      "inference_runtime": 0.7961385250091553,
      "num_train_instances": 5.0,
      "prompt_truncated": 0.0,
      "f1_score": 0.6666666666666666
    }
  },
  {
    "instance_id": "id1331",
    "train_trial_index": 0,
    "predicted_text": "Sell it to another interested party.",
    "base64_images": [],
    "stats": {
      "num_prompt_tokens": 3276.0,
      "num_output_tokens": 7.0,
      "inference_runtime": 0.8257331848144531,
      "num_train_instances": 5.0,
      "prompt_truncated": 0.0,
      "f1_score": 0.5454545454545454
    }
  },
  {
    "instance_id": "id1410",
    "train_trial_index": 0,
    "predicted_text": "Octopus",
    "base64_images": [],
    "stats": {
      "num_prompt_tokens": 3174.0,
      "num_output_tokens": 2.0,
      "inference_runtime": 0.7646636962890625,
      "num_train_instances": 5.0,
      "prompt_truncated": 0.0,
      "f1_score": 1.0
    }
  },
  {
    "instance_id": "id1516",
    "train_trial_index": 0,
    "predicted_text": "Thomas",
    "base64_images": [],
    "stats": {
      "num_prompt_tokens": 3500.0,
      "num_output_tokens": 1.0,
      "inference_runtime": 0.22931790351867676,
      "num_train_instances": 5.0,
      "prompt_truncated": 0.0,
      "f1_score": 1.0
    }
  },
  {
    "instance_id": "id1449",
    "train_trial_index": 0,
    "predicted_text": "Thugs",
    "base64_images": [],
    "stats": {
      "num_prompt_tokens": 3153.0,
      "num_output_tokens": 2.0,
      "inference_runtime": 0.7623381614685059,
      "num_train_instances": 5.0,
      "prompt_truncated": 0.0,
      "f1_score": 0.5
    }
  },
  {
    "instance_id": "id1411",
    "train_trial_index": 0,
    "predicted_text": "Sybil Thorndike",
    "base64_images": [],
    "stats": {
      "num_prompt_tokens": 3585.0,
      "num_output_tokens": 5.0,
      "inference_runtime": 0.8195762634277344,
      "num_train_instances": 5.0,
      "prompt_truncated": 0.0,
      "f1_score": 1.0
    }
  },
  {
    "instance_id": "id1495",
    "train_trial_index": 0,
    "predicted_text": "Evil Ed",
    "base64_images": [],
    "stats": {
      "num_prompt_tokens": 3571.0,
      "num_output_tokens": 3.0,
      "inference_runtime": 0.3450310230255127,
      "num_train_instances": 5.0,
      "prompt_truncated": 0.0,
      "f1_score": 0.6666666666666666
    }
  },
  {
    "instance_id": "id1445",
    "train_trial_index": 0,
    "predicted_text": "Clermont commits suicide.",
    "base64_images": [],
    "stats": {
      "num_prompt_tokens": 3053.0,
      "num_output_tokens": 6.0,
      "inference_runtime": 0.8340392112731934,
      "num_train_instances": 5.0,
      "prompt_truncated": 0.0,
      "f1_score": 0.6666666666666666
    }
  },
  {
    "instance_id": "id1507",
    "train_trial_index": 0,
    "predicted_text": "Prowler",
    "base64_images": [],
    "stats": {
      "num_prompt_tokens": 3543.0,
      "num_output_tokens": 3.0,
      "inference_runtime": 0.2490241527557373,
      "num_train_instances": 4.0,
      "prompt_truncated": 0.0,
      "f1_score": 1.0
    }
  },
  {
    "instance_id": "id1414",
    "train_trial_index": 0,
    "predicted_text": "Newgate Prison",
    "base64_images": [],
    "stats": {
      "num_prompt_tokens": 3442.0,
      "num_output_tokens": 3.0,
      "inference_runtime": 0.29587244987487793,
      "num_train_instances": 4.0,
      "prompt_truncated": 0.0,
      "f1_score": 0.6666666666666666
    }
  },
  {
    "instance_id": "id1246",
    "train_trial_index": 0,
    "predicted_text": "Pizzas",
    "base64_images": [],
    "stats": {
      "num_prompt_tokens": 3871.0,
      "num_output_tokens": 2.0,
      "inference_runtime": 0.2909660339355469,
      "num_train_instances": 5.0,
      "prompt_truncated": 0.0,
      "f1_score": 0.0
    }
  },
  {
    "instance_id": "id1555",
    "train_trial_index": 0,
    "predicted_text": "Because the protagonist's ignorance is shattered by the discovery of her husband's affair.",
    "base64_images": [],
    "stats": {
      "num_prompt_tokens": 3132.0,
      "num_output_tokens": 16.0,
      "inference_runtime": 0.38845372200012207,
      "num_train_instances": 5.0,
      "prompt_truncated": 0.0,
      "f1_score": 0.25
    }
  }
]