-
Notifications
You must be signed in to change notification settings - Fork 11.2k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
WHY WHY WHY ????? #1572
Comments
There is no ISO standard for the model file formats, there is no standard at all. There is no ML framework that produces ggml models, only converted from other formats which means they can always be converted again. Adding support for all models is a lot of effort to maintain. |
@raymerjacque If you don't like what is being built here, then don't use it. :) |
@raymerjacque if it's any consolation, I am (for now) still maintaining backwards compatibility with all prior ggml models in my KoboldCpp fork, though you will not have CUDA support. |
You can always like.. Use a previous version... |
"I want quality to improve, but I don't want updates to the model, but I want all the latest features." Recognize that you're being contradictory and selfish. |
"And everything for free" |
Just get KoboldCPP. It is free, simple and powerful. It supports all of the past models too. |
hey, I know this feeling, but we don't have to be lazy llama pirates, we can convert hf to ggml, and self-quantize, just need to raise your pagefile/swapfile |
actually i havnt used llama in a while, else i would have known about this issue sooner. it was my backup model and now it doesnt work. this is why i HATE open source... give me a paid version of anything and id be much happier knowing it wont be screwed with in unexpected ways... I think ill skip this project. ill just use Bard and Open AI, at least there i dont have to worry about things suddenly being broken without warning or explanation. |
One word.. VELOCITY |
FWIW the following is generated with llama.cpp/Vicuna 13B V1.0 quantized to Q5_1, with parameters
|
Well, that's easy - I will guarantee that i will never do any updates and that with this version, your model will always run. P.S. i give you a discount of 20 % if you decide to buy within 48 hours. |
You can pay if you want: https://github.com/sponsors/ggerganov |
but those online products can change the seed values of your results and you cant rollback, and for this just type (an example) see the versions and dates here: https://github.com/ggerganov/llama.cpp/commits/master once you know how to jump to a previous version = you can do it for all the github projects you come across. Those products don't allow you to use finetune models+replicate the seed. Non-open source products like windows also release updates that change many things and break your programs in unexpected ways. |
Actually windows is a bit of a counterexample. It has an absolutely arcane convoluted codebase, parts of which solely exists to ensure backwards compatibility with some decision made two decades ago. Which is why you have beautiful stuff like WoW64 allowing you to run decade old 32 bit programs on your current system. I do wish they had kept NTVDM. How else are you going to run stuff like chips challenge? Sigh. |
Check PR #1504 which you can easily to convert your existing model to latest. |
@JWNoctis that's very interesting, but could you please share your complete command line ? I'm still struggling to get vicuna to work as anything but a chat bot and it seems you managed to pass it instructions. But whatever I try constantly ends up with an empty response. I've pasted your instructions into a file loaded by -f, or using --prompt "$(cat thisfile.txt)", tried with/without --instruct, with/without --interactive or --interactive-first, and I never get any usable result. I know I'm missing something but previously I was convinced it was related to the models. Now that I'm seeing you're using the exact same as I am, I'm convinced the problem is somewhere else. |
I mean he's right in a way .. we've had a ton of trouble from changes and wanted to keep backward compatibility (mmap etc) I believe one day this project, or a close fork of it, will do that. A plugin type of loaders and evaluation code that adapts on the model type and version number. But .. as long as we do not have a adaptive backend the code should be kept tight and focused. So backward compaibility is going to break over and over again. That's just my personal view on this. |
I understand why changes are needed to make the project progress faster and to get rid of possible initial design mistakes. However that's exactly what tags and branches are made for. It would tremendously help users if just before changing the format, at least one tag was set to mark the latest release supporting this or that format. For example I had to regenerate vicuna-13b-q5 and didn't have enough FS space anymore, so I faced a dilemma: delete the old one and possibly have to download it again if the new version didn't work, or stay on the old version. I finally found a pre-generated version of it on hf.co so I didn't have to delete the old one yet. But these are concrete issues caused by the lack of tag. |
@wtarreau My full command was You can add |
Thanks! So you mean that with 1.1 I should use |
For 1.1 I used
Most of these came from referencing how FastChat, which is from the same team that finetuned Vicuna, handled the prompts. I haven't found a way to make it work nearly as reliably as 1.0 did, unfortunately. |
Interesting, thanks! I'm personally proceeding a different way by naming the participants, like this, which is probably why I never had to know the prompt names:
In such a case I'm using -r 'Usr:' and for me it gives amazingly good results. Example:
However I can't reach anything usable using '' nor '###' nor anything like this (with or without --ignore-eos). That's why I end up using it exclusively in chat mode with "usr" and "bot" discussing. I had hoped to be able to directly ask it to complete text without faking a conversation. But that's not that big of a problem :-) |
My problem is not that you guys broke backward compatibility... but that you gave NO WARNING.... if something like that happens you should slap a BIG FAT warning on there before users update. that's the reason i was angry... anyway, thanx to the screw up with breaking models, it taught me not to be so dependent on this project, so i have shifted my focus onto making my own scripts. however, i will be keeping an eye on this project. you can see my work in detail here ( linux version ) : https://youtu.be/71DSwIk3u6Y and the Windows version here : https://youtu.be/JTz9Hh_0PTE Android version here : https://www.youtube.com/shorts/UqE-0IHjQI4 |
So basically you're shouting at those who give you all the work for free when what you're just left to do is a bunch of scripts. I think that an alternate approach consisting in asking if in the future it would be possible to add a tag before each breaking change would have been much better received. The amount of work we all have to do to integrate llama into our respective projects is nothing compared to all that was done, so I think that a little bit of humility and respect would be deserved. I too was a little bit annoyed by the format change though I saw it mentioned in the various PRs because I'm following the project, and google gave me a URL with an updated one and 10 minutes later it was downloaded, up and running. I accepted that it was for the good cause. If you don't want to risk this, just refrain from updating and keep it frozen on your side, but please, do not complain. |
dont give me that crap, I give to the community just as much as everyone else... in a few weeks ill release a voice capable AI linux distro FREE for everyone to use, thousands, millions... I've been contributing to the open source community for many many many years. right now you can download my android AI app in the google playstore and use it FREE, no ads, no registration, no fees... and all at my expense and costs. We all use whatever is available and we all pay our dues in giving back. everything works like a chain... you break one link and it messes up someone elses day. I guess i may have gone a little far with my rant in the first post, but that update really really messed up my week... and it could easily have been avoided had there been a clear notice about such a big change. |
OK then you understand very well how they came to that decision. For now they're not trying to maintain that big a backwards compatibility but to be compatible with various models and quantization levels. This does require breaking changes. The project doesn't advertise any form of stability at the moment, there's not even a single non-dev branch. Like any such projects you know you're walking on eggs when using them and that you have to keep your own working tags/branches/binaries before pulling. It's not in a form that's suitable for being used as-is beyond testing and that's fine. The feedback, ideas and optimizations from the community are more valuable at this point than the stability. Would it be my project, I would do my best to maintain compatibility and would create branches when it breaks. But I perfectly understand the current choices and your attitude is not welcome, really. In addition what prevents you from uploading your own fork at the last working commit and referencing it ? Hmmm ? |
You can still put a CLEAR notice on a commit when you know its going to break with virtually half the models on hugging face. this was no small change... no matter how you look at this. it was poorly implemented. it literally took people to come in here to complain and them to go and trace back commits to even realize what they had done... |
I'd recommend lowering your expectations for open source support/documentation. $Free comes with no promise of any of that. |
@raymerjacque ah OK you just mean like this: 56551bc That's where I saw it, in the commit log and it appeared in the project's readme. Sure it didn't last very long, but that was probably deemed sufficient for all those following the project closely. I'm sorry about what you say "people to come in here to complain" but for now you are the only one. Maybe you were very frustrated or it wasted you some time at a critical moment but again you are the one who knows what you're doing with the code and how you should keep track of working versions. Even if the format didn't change, the code could have introduced a big bug making it not work anymore for you. They're not going to notify you after every single risky commit gets merge. So now please stop this, you're just fooling yourself and your arguments don't stand given the project's situation. |
Wait a minute...why are you just releasing the app for the Android environment? Why aren't you spending the time and resources to support IOS as well? Unbelievable. |
@cmp-nct i am literally already doing it. |
because they charge 100usd a year to add something to the apple store... f@ck apple, they have always been greedy... if you can afford their over bloated prices and products then you clearly dont have a brain and don't need free stuff, you can go pay for chatgpt plus ... |
🤔 |
difference between a paid version and 100 usd a year to upload a free app, dont you think ? or are you one of those apple users with no brain ? |
I think you're too much in the "I know everything and all other ones are both wrong and stupid" mood and this is not contributing any value here. Better stop here or have an admin close this non-sensival issue. |
can i remind you, you guys keep throwing insults my way. i'm just returning the favor ... |
WHY did you guys end support for older Llama models ? why is backwards compatibility not added when you change formats ? This is what pisses me off about open source, its absolute fraken chaos, things get added and there is no standard that is kept. you guys do realise this project filters down to additional projects, so its not as simple as rolling back or using an earlier commit, the other projects now also have this b@llsh!t added to their githubs. what a fraken mess you guys have created with these latest updates.
youve killed support for half the models at huggingface... i am so angry right now.
The text was updated successfully, but these errors were encountered: