|
13 | 13 | // |
14 | 14 | //$generator = pipeline('text-generation', 'Xenova/gpt2'); |
15 | 15 | //$generator = pipeline('text-generation', 'Xenova/Qwen1.5-0.5B-Chat'); |
16 | | -$generator = pipeline('text-generation', 'Xenova/TinyLlama-1.1B-Chat-v1.0'); |
17 | | - |
18 | | -$streamer = StdOutStreamer::make(); |
19 | | - |
20 | | -$messages = [ |
21 | | - ['role' => 'system', 'content' => 'You are a helpful assistant.'], |
22 | | - ['role' => 'user', 'content' => 'What is diffusion in chemistry?'], |
23 | | -]; |
| 16 | +//$generator = pipeline('text-generation', 'Xenova/TinyLlama-1.1B-Chat-v1.0'); |
| 17 | +// |
| 18 | +//$streamer = StdOutStreamer::make(); |
| 19 | +// |
| 20 | +//$messages = [ |
| 21 | +// ['role' => 'system', 'content' => 'You are a helpful assistant.'], |
| 22 | +// ['role' => 'user', 'content' => 'What is diffusion in chemistry?'], |
| 23 | +//]; |
| 24 | +// |
| 25 | +//$input = $generator->tokenizer->applyChatTemplate($messages, addGenerationPrompt: true, tokenize: false); |
| 26 | +// |
| 27 | +//$output = $generator($input, |
| 28 | +// streamer: $streamer, |
| 29 | +// maxNewTokens: 128, |
| 30 | +// doSample: true, |
| 31 | +// returnFullText: false, |
| 32 | +//// temperature: 0.7, |
| 33 | +//// repetitionPenalty: 1.3, |
| 34 | +//// earlyStopping: true |
| 35 | +//); |
24 | 36 |
|
25 | | -$input = $generator->tokenizer->applyChatTemplate($messages, addGenerationPrompt: true, tokenize: false); |
| 37 | +$generator = pipeline('text-generation', 'Xenova/codegen-350M-mono'); |
| 38 | +$streamer = StdOutStreamer::make($generator->tokenizer); |
26 | 39 |
|
27 | | -$output = $generator($input, |
| 40 | +$output = $generator( |
| 41 | + 'def fib(n):', |
28 | 42 | streamer: $streamer, |
29 | | - maxNewTokens: 128, |
30 | | - doSample: true, |
31 | | - returnFullText: false, |
32 | | -// temperature: 0.7, |
33 | | -// repetitionPenalty: 1.3, |
34 | | -// earlyStopping: true |
| 43 | + maxNewTokens: 100, |
| 44 | + doSample: true |
35 | 45 | ); |
36 | 46 |
|
37 | | -//$generator = pipeline('text-generation', 'Xenova/codegen-350M-mono'); |
38 | | -//$streamer = StdOutStreamer::make($generator->tokenizer); |
39 | | -// |
40 | | -//$output = $generator( |
41 | | -// 'def fib(n):', |
42 | | -// streamer: $streamer, |
43 | | -// maxNewTokens: 100, |
44 | | -// doSample: true |
45 | | -//); |
46 | | - |
47 | 47 | dd($output[0]['generated_text'], timeUsage(), memoryUsage()); |
0 commit comments