@@ -12,20 +12,18 @@ describe('ai', () => {
|
12 | 12 | spans: expect.arrayContaining([
|
13 | 13 | expect.objectContaining({
|
14 | 14 | data: expect.objectContaining({
|
15 |
| -'ai.completion_tokens.used': 20, |
16 | 15 | 'ai.model.id': 'mock-model-id',
|
17 | 16 | 'ai.model.provider': 'mock-provider',
|
18 |
| -'ai.model_id': 'mock-model-id', |
19 | 17 | 'ai.operationId': 'ai.generateText',
|
20 | 18 | 'ai.pipeline.name': 'generateText',
|
21 |
| -'ai.prompt_tokens.used': 10, |
22 | 19 | 'ai.response.finishReason': 'stop',
|
23 | 20 | 'ai.settings.maxRetries': 2,
|
24 | 21 | 'ai.settings.maxSteps': 1,
|
25 | 22 | 'ai.': false,
|
26 |
| -'ai.total_tokens.used': 30, |
27 |
| -'ai.usage.completionTokens': 20, |
28 |
| -'ai.usage.promptTokens': 10, |
| 23 | +'gen_ai.response.model': 'mock-model-id', |
| 24 | +'gen_ai.usage.input_tokens': 10, |
| 25 | +'gen_ai.usage.output_tokens': 20, |
| 26 | +'gen_ai.usage.total_tokens': 30, |
29 | 27 | 'operation.name': 'ai.generateText',
|
30 | 28 | 'sentry.op': 'ai.pipeline.generateText',
|
31 | 29 | 'sentry.origin': 'auto.vercelai.otel',
|
@@ -47,18 +45,17 @@ describe('ai', () => {
|
47 | 45 | 'gen_ai.system': 'mock-provider',
|
48 | 46 | 'gen_ai.request.model': 'mock-model-id',
|
49 | 47 | 'ai.pipeline.name': 'generateText.doGenerate',
|
50 |
| -'ai.model_id': 'mock-model-id', |
51 | 48 | 'ai.': false,
|
52 | 49 | 'ai.response.finishReason': 'stop',
|
53 | 50 | 'ai.response.model': 'mock-model-id',
|
54 |
| -'ai.usage.promptTokens': 10, |
55 |
| -'ai.usage.completionTokens': 20, |
| 51 | +'ai.response.id': expect.any(String), |
| 52 | +'ai.response.timestamp': expect.any(String), |
56 | 53 | 'gen_ai.response.finish_reasons': ['stop'],
|
57 | 54 | 'gen_ai.usage.input_tokens': 10,
|
58 | 55 | 'gen_ai.usage.output_tokens': 20,
|
59 |
| -'ai.completion_tokens.used': 20, |
60 |
| -'ai.prompt_tokens.used': 10, |
61 |
| -'ai.total_tokens.used': 30, |
| 56 | +'gen_ai.response.id': expect.any(String), |
| 57 | +'gen_ai.response.model': 'mock-model-id', |
| 58 | +'gen_ai.usage.total_tokens': 30, |
62 | 59 | }),
|
63 | 60 | description: 'generateText.doGenerate',
|
64 | 61 | op: 'ai.run.doGenerate',
|
@@ -67,22 +64,21 @@ describe('ai', () => {
|
67 | 64 | }),
|
68 | 65 | expect.objectContaining({
|
69 | 66 | data: expect.objectContaining({
|
70 |
| -'ai.completion_tokens.used': 20, |
71 | 67 | 'ai.model.id': 'mock-model-id',
|
72 | 68 | 'ai.model.provider': 'mock-provider',
|
73 |
| -'ai.model_id': 'mock-model-id', |
74 |
| -'ai.prompt': '{"prompt":"Where is the second span?"}', |
75 | 69 | 'ai.operationId': 'ai.generateText',
|
76 | 70 | 'ai.pipeline.name': 'generateText',
|
77 |
| -'ai.prompt_tokens.used': 10, |
| 71 | +'ai.prompt': '{"prompt":"Where is the second span?"}', |
78 | 72 | 'ai.response.finishReason': 'stop',
|
79 |
| -'ai.input_messages': '{"prompt":"Where is the second span?"}', |
| 73 | +'ai.response.text': expect.any(String), |
80 | 74 | 'ai.settings.maxRetries': 2,
|
81 | 75 | 'ai.settings.maxSteps': 1,
|
82 | 76 | 'ai.': false,
|
83 |
| -'ai.total_tokens.used': 30, |
84 |
| -'ai.usage.completionTokens': 20, |
85 |
| -'ai.usage.promptTokens': 10, |
| 77 | +'gen_ai.prompt': '{"prompt":"Where is the second span?"}', |
| 78 | +'gen_ai.response.model': 'mock-model-id', |
| 79 | +'gen_ai.usage.input_tokens': 10, |
| 80 | +'gen_ai.usage.output_tokens': 20, |
| 81 | +'gen_ai.usage.total_tokens': 30, |
86 | 82 | 'operation.name': 'ai.generateText',
|
87 | 83 | 'sentry.op': 'ai.pipeline.generateText',
|
88 | 84 | 'sentry.origin': 'auto.vercelai.otel',
|
@@ -104,18 +100,20 @@ describe('ai', () => {
|
104 | 100 | 'gen_ai.system': 'mock-provider',
|
105 | 101 | 'gen_ai.request.model': 'mock-model-id',
|
106 | 102 | 'ai.pipeline.name': 'generateText.doGenerate',
|
107 |
| -'ai.model_id': 'mock-model-id', |
108 | 103 | 'ai.': false,
|
109 | 104 | 'ai.response.finishReason': 'stop',
|
110 | 105 | 'ai.response.model': 'mock-model-id',
|
111 |
| -'ai.usage.promptTokens': 10, |
112 |
| -'ai.usage.completionTokens': 20, |
| 106 | +'ai.response.id': expect.any(String), |
| 107 | +'ai.response.text': expect.any(String), |
| 108 | +'ai.response.timestamp': expect.any(String), |
| 109 | +'ai.prompt.format': expect.any(String), |
| 110 | +'ai.prompt.messages': expect.any(String), |
113 | 111 | 'gen_ai.response.finish_reasons': ['stop'],
|
114 | 112 | 'gen_ai.usage.input_tokens': 10,
|
115 | 113 | 'gen_ai.usage.output_tokens': 20,
|
116 |
| -'ai.completion_tokens.used': 20, |
117 |
| -'ai.prompt_tokens.used': 10, |
118 |
| -'ai.total_tokens.used': 30, |
| 114 | +'gen_ai.response.id': expect.any(String), |
| 115 | +'gen_ai.response.model': 'mock-model-id', |
| 116 | +'gen_ai.usage.total_tokens': 30, |
119 | 117 | }),
|
120 | 118 | description: 'generateText.doGenerate',
|
121 | 119 | op: 'ai.run.doGenerate',
|
|
0 commit comments