Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Add Orca if it ever releases #47

Open
timjzee opened this issue Jul 20, 2023 · 4 comments
Open

Add Orca if it ever releases #47

timjzee opened this issue Jul 20, 2023 · 4 comments

Comments

@timjzee
Copy link
Collaborator

timjzee commented Jul 20, 2023

Whitepaper: https://arxiv.org/pdf/2306.02707.pdf

Will be released here: https://aka.ms/orca-lm

Summary: https://www.youtube.com/watch?v=Dt_UNg7Mchg

@mdingemanse
Copy link
Collaborator

Interesting find. And more evidence of the growing importance of synthetic instruction tuning data.

@mdingemanse
Copy link
Collaborator

Looks like there may be some version of it here: https://huggingface.co/yhyhy3/med-orca-instruct-33b-GPTQ

@timjzee
Copy link
Collaborator Author

timjzee commented Jul 26, 2023

I think most of the "Orca" models on Huggingface are projects which used a similar approach to the one described in the Microsoft paper. AFAIK they are not actual Orca releases.

@mdingemanse
Copy link
Collaborator

mdingemanse commented Nov 22, 2023

Nah. There is a new preprint that says

We open-source Orca 2 to encourage further research on the development, evaluation, and alignment of smaller LMs.

But nothing is open-sourced; this is a Llama2 finetune where only the instruction-tuned (or what they call explanation-tuned) model weights are made available, but none of instruction/explanation datasets and none of the source code is made available.

Thanks Meta for thoroughly diluting the term open source and thanks Microsoft for further contributing to it.

mdingemanse added a commit that referenced this issue Nov 22, 2023
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants