Skip to content

Reproducing results on Llama-3.1-8B-Inst #8

@chtmp223

Description

@chtmp223

Thank you for your great work!

I tried to reproduce the results for meta-llama/Llama-3.1-8B-Instruct, but I noticed several discrepancies between my outcomes and those reported in the public result file under Llama-3.1-8B-Inst.

You can view my results here: Results. I obtained these results by running scripts/run_eval_slurm.sh on both short and long configs for some benchmarks, and I compiled all the results using scripts/collect_results.py.

Do you have any suggestion on specific things I could adjust to align my results with yours?

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions