| | #include "ggml.h" |
| |
|
| | #include <cmath> |
| | #include <cstdio> |
| | #include <cstdlib> |
| | #include <cassert> |
| |
|
| | #define MAX_NARGS 2 |
| |
|
| | #if defined(__GNUC__) |
| | #pragma GCC diagnostic ignored "-Wdouble-promotion" |
| | #endif |
| |
|
| | |
| | |
| | |
| | #define GGML_DEBUG 0 |
| | #if (GGML_DEBUG >= 1) |
| | #define GGML_PRINT_DEBUG(...) printf(__VA_ARGS__) |
| | #else |
| | #define GGML_PRINT_DEBUG(...) |
| | #endif |
| |
|
| | #if (GGML_DEBUG >= 5) |
| | #define GGML_PRINT_DEBUG_5(...) printf(__VA_ARGS__) |
| | #else |
| | #define GGML_PRINT_DEBUG_5(...) |
| | #endif |
| |
|
| | #if (GGML_DEBUG >= 10) |
| | #define GGML_PRINT_DEBUG_10(...) printf(__VA_ARGS__) |
| | #else |
| | #define GGML_PRINT_DEBUG_10(...) |
| | #endif |
| |
|
| | #define GGML_PRINT(...) printf(__VA_ARGS__) |
| |
|
| |
|
| | static float frand(void) { |
| | return (float)rand()/(float)RAND_MAX; |
| | } |
| |
|
| | static struct ggml_tensor * get_random_tensor( |
| | struct ggml_context * ctx0, int ndims, int64_t ne[], float fmin, float fmax |
| | ) { |
| | struct ggml_tensor * result = ggml_new_tensor(ctx0, GGML_TYPE_F32, ndims, ne); |
| |
|
| | switch (ndims) { |
| | case 1: |
| | for (int i0 = 0; i0 < ne[0]; i0++) { |
| | ((float *)result->data)[i0] = frand()*(fmax - fmin) + fmin; |
| | } |
| | break; |
| | case 2: |
| | for (int i1 = 0; i1 < ne[1]; i1++) { |
| | for (int i0 = 0; i0 < ne[0]; i0++) { |
| | ((float *)result->data)[i1*ne[0] + i0] = frand()*(fmax - fmin) + fmin; |
| | } |
| | } |
| | break; |
| | case 3: |
| | for (int i2 = 0; i2 < ne[2]; i2++) { |
| | for (int i1 = 0; i1 < ne[1]; i1++) { |
| | for (int i0 = 0; i0 < ne[0]; i0++) { |
| | ((float *)result->data)[i2*ne[1]*ne[0] + i1*ne[0] + i0] = frand()*(fmax - fmin) + fmin; |
| | } |
| | } |
| | } |
| | break; |
| | case 4: |
| | for (int i3 = 0; i3 < ne[3]; i3++) { |
| | for (int i2 = 0; i2 < ne[2]; i2++) { |
| | for (int i1 = 0; i1 < ne[1]; i1++) { |
| | for (int i0 = 0; i0 < ne[0]; i0++) { |
| | ((float *)result->data)[i3*ne[2]*ne[1]*ne[0] + i2*ne[1]*ne[0] + i1*ne[0] + i0] = frand()*(fmax - fmin) + fmin; |
| | } |
| | } |
| | } |
| | } |
| | break; |
| | default: |
| | assert(false); |
| | } |
| |
|
| | return result; |
| | } |
| |
|
| | int main(void) { |
| | struct ggml_init_params params = { |
| | 1024*1024*1024, |
| | NULL, |
| | false, |
| | }; |
| |
|
| | struct ggml_context * ctx = ggml_init(params); |
| |
|
| | int64_t ne1[4] = {4, 128, 1, 1}; |
| | int64_t ne2[4] = {4, 256, 1, 1}; |
| | int64_t ne3[4] = {128, 256, 1, 1}; |
| |
|
| | struct ggml_tensor * a = get_random_tensor(ctx, 2, ne1, -1, +1); |
| | struct ggml_tensor * b = get_random_tensor(ctx, 2, ne2, -1, +1); |
| | ggml_set_param(ctx, a); |
| | ggml_set_param(ctx, b); |
| |
|
| | struct ggml_tensor * c = get_random_tensor(ctx, 2, ne3, -1, +1); |
| |
|
| | struct ggml_tensor * ab = ggml_mul_mat(ctx, a, b); |
| | struct ggml_tensor * d = ggml_sub(ctx, c, ab); |
| | struct ggml_tensor * e = ggml_sum(ctx, ggml_sqr(ctx, d)); |
| |
|
| | struct ggml_cgraph ge = ggml_build_forward(e); |
| | ggml_graph_reset(&ge); |
| |
|
| | ggml_graph_compute_with_ctx(ctx, &ge, 1); |
| |
|
| | const float fe = ggml_get_f32_1d(e, 0); |
| | printf("%s: e = %.4f\n", __func__, fe); |
| |
|
| | struct ggml_opt_params opt_params = ggml_opt_default_params(GGML_OPT_ADAM); |
| |
|
| | ggml_opt(ctx, opt_params, e); |
| |
|
| | ggml_graph_reset(&ge); |
| |
|
| | ggml_graph_compute_with_ctx(ctx, &ge, 1); |
| |
|
| | const float fe_opt = ggml_get_f32_1d(e, 0); |
| | printf("%s: original e = %.4f\n", __func__, fe); |
| | printf("%s: optimized e = %.4f\n", __func__, fe_opt); |
| |
|
| | const bool success = (fe_opt <= fe); |
| | assert(success); |
| |
|
| | ggml_free(ctx); |
| | return success ? 0 : -1; |
| | } |
| | |
| | |
| | |
| | |
| | |
| |
|
| | |
| | |
| | |
| | |
| | |
| |
|
| | |
| | |
| | |
| | |
| | |
| |
|
| |
|
| | |
| | |
| | |
| | |
| | |
| |
|
| | |
| | |
| | |
| | |
| | |
| |
|
| | |
| | |
| | |
| | |
| | |
| | |
| |
|
| | |
| | |
| | |
| | |
| | |
| |
|