Skip to content

Code showing when running. #717

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Closed
betolley opened this issue Apr 2, 2023 · 4 comments
Closed

Code showing when running. #717

betolley opened this issue Apr 2, 2023 · 4 comments

Comments

@betolley
Copy link

betolley commented Apr 2, 2023

When I start chat.exe with a alpaca bin I get.
main: seed = 1680456908
llama_model_load: loading model from 'models/llama-7B/ggml-model.bin' - please wait ...
llama_model_load: n_vocab = 32000
llama_model_load: n_ctx = 512
llama_model_load: n_embd = 4096
llama_model_load: n_mult = 256
llama_model_load: n_head = 32
llama_model_load: n_layer = 32
llama_model_load: n_rot = 128
llama_model_load: f16 = 3
llama_model_load: n_ff = 11008
llama_model_load: n_parts = 1
llama_model_load: type = 1
llama_model_load: ggml map size = 4820.95 MB
llama_model_load: ggml ctx size = 81.25 KB
llama_model_load: mem required = 6613.03 MB (+ 1026.00 MB per state)
llama_model_load: loading tensors from 'models/llama-7B/ggml-model.bin'
llama_model_load: model size = 4820.52 MB / num tensors = 291
llama_init_from_file: kv self size = 256.00 MB

system_info: n_threads = 4 / 8 | AVX = 1 | AVX2 = 0 | AVX512 = 0 | FMA = 0 | NEON = 0 | ARM_FMA = 0 | F16C = 0 | FP16_VA = 0 | WASM_SIMD = 0 | BLAS = 0 | SSE3 = 0 | VSX = 0 |
main: interactive mode on.
sampling: temp = 0.800000, top_k = 40, top_p = 0.950000, repeat_last_n = 64, repeat_penalty = 1.100000
generate: n_ctx = 512, n_batch = 8, n_predict = 128, n_keep = 0

== Running in interactive mode. ==

  • Press Ctrl+C to interject at any time.
  • Press Return to return control to LLaMa.
  • If you want to submit another line, end your input in ''.

using System;

using System.Collections.Generic;
using System.Linq;
using System.Text;
using System.Threading.Tasks;

namespace _1.Write_a_program_to_find_the_minimum_element_in_an_array
{
class Program
{
static void Main()
{
//Create an array of five elements
int[] arr = new int[5];

        //Fill the array with random values between 0 and 10
@Belluxx
Copy link

Belluxx commented Apr 2, 2023

What are you trying to achieve? If you want to chat you must provide an example/context for that. If you want a story you must introduce it before starting the inference.

For example try adding this argument: -p "Once upon a time"

@prusnak
Copy link
Collaborator

prusnak commented Apr 2, 2023

Yes, you forgot to set a prompt so Llama just came up with its own text completely on her own! :-)

@betolley
Copy link
Author

betolley commented Apr 2, 2023

I was used to alpaca I didn't have to. WHen I run llama with -i it did it also.

@Belluxx
Copy link

Belluxx commented Apr 2, 2023

Because alpaca.cpp adds context without showing you, it's a "wrapper" for alpaca finetuning

@prusnak prusnak closed this as not planned Won't fix, can't repro, duplicate, stale Apr 2, 2023
Deadsg pushed a commit to Deadsg/llama.cpp that referenced this issue Dec 19, 2023
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants