-
Notifications
You must be signed in to change notification settings - Fork 9.7k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
sync : ggml #9567
sync : ggml #9567
Conversation
* CUDA eval works * stochastic gradient descent op * Adam except decay * CUDA CROSS_ENTROPY_LOSS_BACK * CUDA mnist-fc training works * backend CLI arg * refactor gguf load * remove sched from opt_step_adam * implement l1 regularization (weight decay) * extra call to add optimizer * initialize gradients with ggml_graph_reset * gradient accumulation * increment iter per eval instead of epoch * adjust backend interfaces * fix ggml_graph_reset without backend * fix ggml graph export/import * fixup * rename * revert ggml_opt changes * more general CUDA repeat_back * update documentation, fix CNN * validation split * add clarifying comment * optimize PyTorch training * adjust buffer size, thread count * fix 0.0f validation split * Update examples/mnist/mnist-common.cpp Co-authored-by: Georgi Gerganov <ggerganov@gmail.com> * fix gradient accumulation * tensor flag for accumulators -> tensor hash set * Update include/ggml.h Co-authored-by: slaren <slarengh@gmail.com> * Update tests/test-backend-ops.cpp Co-authored-by: slaren <slarengh@gmail.com> * Update tests/test-backend-ops.cpp Co-authored-by: slaren <slarengh@gmail.com> * fix test prints * Update src/ggml-backend.c Co-authored-by: Georgi Gerganov <ggerganov@gmail.com> * better CUDA support for noncontiguous out_prod * add comment --------- Co-authored-by: Georgi Gerganov <ggerganov@gmail.com> Co-authored-by: slaren <slarengh@gmail.com>
ggml-ci
@JohannesGaessler @slaren How to fix the HIP build: [ 11%] Building HIP object ggml/src/CMakeFiles/ggml.dir/ggml-cuda/out-prod.cu.o
In file included from /__w/llama.cpp/llama.cpp/ggml/src/ggml-cuda/out-prod.cu:2:
/__w/llama.cpp/llama.cpp/ggml/src/ggml-cuda/vendors/cuda.h:3:10: fatal error: 'cuda_runtime.h' file not found
#include <cuda_runtime.h>
^~~~~~~~~~~~~~~~
1 error generated when compiling for gfx906.
gmake[2]: *** [ggml/src/CMakeFiles/ggml.dir/build.make:404: ggml/src/CMakeFiles/ggml.dir/ggml-cuda/out-prod.cu.o] Error 1 Edit: I think I just have to remove the include from |
ggml/src/ggml-cuda/out-prod.cu
Outdated
@@ -0,0 +1,52 @@ | |||
#include "out-prod.cuh" | |||
#include "vendors/cuda.h" |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
#include "vendors/cuda.h" |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
This should do it, this file cannot be included directly, and it is already included in common.cuh
.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I can confirm that this fixes compilation with GGML_HIPBLAS
.
ggml-ci
No description provided.