| Benchmark | Expected Score (1.3B) | Mila AI -v1.3.7b- -aDDont- (speculative) | |-----------|----------------------|-------------------------------------------| | HellaSwag (0-shot) | ~45% | ~48% (if well-tuned) | | MMLU (5-shot) | ~25% | ~27% | | HumanEval (pass@1) | ~4% | ~5.5% | | French GLUE (FLeX) | N/A | Could excel (bilingual) |
prompt = "Explain the significance of the -aDDont- flag in attention mechanisms." inputs = tokenizer(prompt, return_tensors="pt").to("cuda") output = model.generate(**inputs, max_new_tokens=200) print(tokenizer.decode(output[0])) Mila AI -v1.3.7b- -aDDont-
If you have access to this model or are its creator, please share a link in the discussion section below so this article can be updated with real benchmarks and usage examples. | Benchmark | Expected Score (1
However, a quick check shows that this exact string does not correspond to any widely known or documented AI model, software release, or open-source project on platforms like Hugging Face, GitHub, or official AI research pages. or official AI research pages.
July 25th, 2023
July 25th, 2023
March 10th, 2023