Gemini Pro vs Gemini Flash: Which is more accurate?

In my Udemy course, I tried to extract the outcome for a VAERS report using both Gemini Flash and Gemini Pro (latest).

class RecoveryStatus(Enum):
    Yes = "Yes"
    Recovered = "Yes"
    Recovered_With_Treatment = "Yes"
    No = "No"
    Recovering = "No"
    Resolving = "No"
    Recovering_With_Treatment = "No"
    Unknown = "Unknown"

class Explanation(typing.TypedDict):
    matching_sentence_if_any: str
    explanation: str

class Symptom(typing.TypedDict):
    symptom_value: str
    symptom_recovery_status: RecoveryStatus

class PatientInfo_v5(typing.TypedDict):
    recovered_from_all_symptoms: RecoveryStatus
    recovered_from_all_symptoms_explanation: Explanation
    symptom_list: list[Symptom]

Here is how it works:

  • the LLM first needs to extract the full list of symptoms that the patient experienced
  • then it has to extract the outcome for each symptom
  • the worst outcome among all the symptoms should be selected
  • Y = Yes, N = No, and U = Unknown

I first tried using Gemini Pro and got pretty poor accuracy on this task.

Surprisingly, even though Gemini Flash is supposed to be a “less powerful” model (which means it is both cheaper and faster!) it comfortably outperformed Gemini Pro for this particular task.

Comparison using https://countless.dev/

So I ended up creating the course chapter using Gemini Flash.

This is why one of my 10 tips for improving LLM accuracy is to also test the less powerful models from the same vendor.

While you may not get the same results, it is still a good idea to set up a quick test to figure out which LLM is most accurate for your specific task.

Similar Posts

Leave a Reply