Nh/unify request attribute extraction #45314
Annotations
9 errors and 6 warnings
|
Run integration tests
Process completed with exit code 1.
|
|
Run integration tests:
dev-packages/node-integration-tests/utils/assertions.ts#L35
AssertionError: expected { contexts: { …(8) }, …(14) } to match object { event_id: Any<String>, …(5) }
(92 matching properties omitted from actual)
- Expected
+ Received
@@ -1,12 +1,13 @@
{
"event_id": Any<String>,
- "spans": ArrayContaining [
- ObjectContaining {
+ "spans": [
+ {
"data": {
"gen_ai.input.messages": "[{\"role\":\"user\",\"content\":\"What is the capital of France?\"}]",
"gen_ai.operation.name": "chat",
+ "gen_ai.request.max_tokens": 100,
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.request.temperature": 0.7,
"gen_ai.response.finish_reasons": "[\"stop\"]",
"gen_ai.response.id": "chatcmpl-mock123",
"gen_ai.response.model": "gpt-3.5-turbo",
@@ -21,13 +22,47 @@
"sentry.sdk_meta.gen_ai.input.messages.original_length": 1,
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "53349329f1560342",
+ "span_id": "12977ce11c301c03",
+ "start_timestamp": 1775213321.614,
"status": "ok",
+ "timestamp": 1775213321.7244709,
+ "trace_id": "c97758d6afcc832f79e7463ef56eea46",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 38767,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 38767,
+ "url": "http://localhost:38767/openai/chat/completions",
+ "url.full": "http://localhost:38767/openai/chat/completions",
+ "url.path": "/openai/chat/completions",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 6.0.0",
+ },
+ "description": "POST http://localhost:38767/openai/chat/completions",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "12977ce11c301c03",
+ "span_id": "c403df8cc2e1c799",
+ "start_timestamp": 1775213321.654,
+ "status": "ok",
+ "timestamp": 1775213321.7189348,
+ "trace_id": "c97758d6afcc832f79e7463ef56eea46",
+ },
+ {
"data": {
"gen_ai.input.messages": "Translate this to French: Hello",
"gen_ai.operation.name": "chat",
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.response.finish_reasons": "[\"completed\"]",
@@ -43,13 +78,47 @@
"sentry.sdk_meta.gen_ai.input.messages.original_length": 1,
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "53349329f1560342",
+ "span_id": "117439780973bb27",
+ "start_timestamp": 1775213321.725,
"status": "ok",
+ "timestamp": 1775213321.743855,
+ "trace_id": "c97758d6afcc832f79e7463ef56eea46",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 38767,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 38767,
+ "url": "http://localhost:38767/openai/responses",
+ "url.full": "http://localhost:38767/openai/responses",
+ "url.path": "/openai/responses",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 6.0.0",
+ },
+ "description": "POST http://localhost:38767/openai/responses",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "11743978097
|
|
Run integration tests:
dev-packages/node-integration-tests/utils/assertions.ts#L35
AssertionError: expected { contexts: { …(8) }, …(14) } to match object { event_id: Any<String>, …(5) }
(92 matching properties omitted from actual)
- Expected
+ Received
@@ -1,12 +1,13 @@
{
"event_id": Any<String>,
- "spans": ArrayContaining [
- ObjectContaining {
+ "spans": [
+ {
"data": {
"gen_ai.input.messages": "[{\"role\":\"user\",\"content\":\"What is the capital of France?\"}]",
"gen_ai.operation.name": "chat",
+ "gen_ai.request.max_tokens": 100,
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.request.temperature": 0.7,
"gen_ai.response.finish_reasons": "[\"stop\"]",
"gen_ai.response.id": "chatcmpl-mock123",
"gen_ai.response.model": "gpt-3.5-turbo",
@@ -21,13 +22,47 @@
"sentry.sdk_meta.gen_ai.input.messages.original_length": 1,
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "f003db990997985f",
+ "span_id": "484777a4bcf06779",
+ "start_timestamp": 1775213318.227,
"status": "ok",
+ "timestamp": 1775213318.3155127,
+ "trace_id": "43a169740f9ca88b186220a856c44a82",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 40737,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 40737,
+ "url": "http://localhost:40737/openai/chat/completions",
+ "url.full": "http://localhost:40737/openai/chat/completions",
+ "url.path": "/openai/chat/completions",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 6.0.0",
+ },
+ "description": "POST http://localhost:40737/openai/chat/completions",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "484777a4bcf06779",
+ "span_id": "32880761ed5d202d",
+ "start_timestamp": 1775213318.243,
+ "status": "ok",
+ "timestamp": 1775213318.3111184,
+ "trace_id": "43a169740f9ca88b186220a856c44a82",
+ },
+ {
"data": {
"gen_ai.input.messages": "Translate this to French: Hello",
"gen_ai.operation.name": "chat",
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.response.finish_reasons": "[\"completed\"]",
@@ -43,13 +78,47 @@
"sentry.sdk_meta.gen_ai.input.messages.original_length": 1,
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "f003db990997985f",
+ "span_id": "db2470c4bff1af37",
+ "start_timestamp": 1775213318.316,
"status": "ok",
+ "timestamp": 1775213318.3370628,
+ "trace_id": "43a169740f9ca88b186220a856c44a82",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 40737,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 40737,
+ "url": "http://localhost:40737/openai/responses",
+ "url.full": "http://localhost:40737/openai/responses",
+ "url.path": "/openai/responses",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 6.0.0",
+ },
+ "description": "POST http://localhost:40737/openai/responses",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "db2470c4bf
|
|
Run integration tests:
dev-packages/node-integration-tests/utils/assertions.ts#L35
AssertionError: expected { contexts: { …(8) }, …(14) } to match object { event_id: Any<String>, …(5) }
(92 matching properties omitted from actual)
- Expected
+ Received
@@ -1,11 +1,12 @@
{
"event_id": Any<String>,
- "spans": ArrayContaining [
- ObjectContaining {
+ "spans": [
+ {
"data": {
"gen_ai.operation.name": "chat",
+ "gen_ai.request.max_tokens": 100,
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.request.temperature": 0.7,
"gen_ai.response.finish_reasons": "[\"stop\"]",
"gen_ai.response.id": "chatcmpl-mock123",
"gen_ai.response.model": "gpt-3.5-turbo",
@@ -17,13 +18,47 @@
"sentry.origin": "auto.ai.openai",
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "09f4dbe679923a3b",
+ "span_id": "de6532221288b996",
+ "start_timestamp": 1775213312.523,
"status": "ok",
+ "timestamp": 1775213312.6392276,
+ "trace_id": "51622c420a774445ce997d2baa9d80dc",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 46689,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 46689,
+ "url": "http://localhost:46689/openai/chat/completions",
+ "url.full": "http://localhost:46689/openai/chat/completions",
+ "url.path": "/openai/chat/completions",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 6.0.0",
+ },
+ "description": "POST http://localhost:46689/openai/chat/completions",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "de6532221288b996",
+ "span_id": "472eb014494223f1",
+ "start_timestamp": 1775213312.565,
+ "status": "ok",
+ "timestamp": 1775213312.6344016,
+ "trace_id": "51622c420a774445ce997d2baa9d80dc",
+ },
+ {
"data": {
"gen_ai.operation.name": "chat",
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.response.finish_reasons": "[\"completed\"]",
"gen_ai.response.id": "resp_mock456",
@@ -36,26 +71,152 @@
"sentry.origin": "auto.ai.openai",
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "09f4dbe679923a3b",
+ "span_id": "c5183479177b239e",
+ "start_timestamp": 1775213312.64,
"status": "ok",
+ "timestamp": 1775213312.661619,
+ "trace_id": "51622c420a774445ce997d2baa9d80dc",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 46689,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 46689,
+ "url": "http://localhost:46689/openai/responses",
+ "url.full": "http://localhost:46689/openai/responses",
+ "url.path": "/openai/responses",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 6.0.0",
+ },
+ "description": "POST http://localhost:46689/openai/responses",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "c5183479177b239e",
+ "span_id": "d1ff1f051a3cb1f6",
+ "start_timestamp": 1775213312.642,
+ "status": "ok",
+ "timestamp": 1775213312.6597168,
+ "trace_id": "516
|
|
Run integration tests:
dev-packages/node-integration-tests/utils/assertions.ts#L35
AssertionError: expected { contexts: { …(8) }, …(14) } to match object { event_id: Any<String>, …(5) }
(92 matching properties omitted from actual)
- Expected
+ Received
@@ -1,11 +1,12 @@
{
"event_id": Any<String>,
- "spans": ArrayContaining [
- ObjectContaining {
+ "spans": [
+ {
"data": {
"gen_ai.operation.name": "chat",
+ "gen_ai.request.max_tokens": 100,
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.request.temperature": 0.7,
"gen_ai.response.finish_reasons": "[\"stop\"]",
"gen_ai.response.id": "chatcmpl-mock123",
"gen_ai.response.model": "gpt-3.5-turbo",
@@ -17,13 +18,47 @@
"sentry.origin": "auto.ai.openai",
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "c7b7c63769d1c54e",
+ "span_id": "387c0ae80ad99234",
+ "start_timestamp": 1775213308.877,
"status": "ok",
+ "timestamp": 1775213308.9499173,
+ "trace_id": "cb469cca56d941e8accfc462e0bc7435",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 41717,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 41717,
+ "url": "http://localhost:41717/openai/chat/completions",
+ "url.full": "http://localhost:41717/openai/chat/completions",
+ "url.path": "/openai/chat/completions",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 6.0.0",
+ },
+ "description": "POST http://localhost:41717/openai/chat/completions",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "387c0ae80ad99234",
+ "span_id": "f0fe79fd26116b30",
+ "start_timestamp": 1775213308.891,
+ "status": "ok",
+ "timestamp": 1775213308.94572,
+ "trace_id": "cb469cca56d941e8accfc462e0bc7435",
+ },
+ {
"data": {
"gen_ai.operation.name": "chat",
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.response.finish_reasons": "[\"completed\"]",
"gen_ai.response.id": "resp_mock456",
@@ -36,26 +71,152 @@
"sentry.origin": "auto.ai.openai",
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "c7b7c63769d1c54e",
+ "span_id": "810d98479583b3b4",
+ "start_timestamp": 1775213308.951,
"status": "ok",
+ "timestamp": 1775213308.9703977,
+ "trace_id": "cb469cca56d941e8accfc462e0bc7435",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 41717,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 41717,
+ "url": "http://localhost:41717/openai/responses",
+ "url.full": "http://localhost:41717/openai/responses",
+ "url.path": "/openai/responses",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 6.0.0",
+ },
+ "description": "POST http://localhost:41717/openai/responses",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "810d98479583b3b4",
+ "span_id": "0d6c9e8fb87f13f3",
+ "start_timestamp": 1775213308.954,
+ "status": "ok",
+ "timestamp": 1775213308.9692,
+ "trace_id": "cb469c
|
|
Run integration tests:
dev-packages/node-integration-tests/utils/assertions.ts#L35
AssertionError: expected { contexts: { …(8) }, …(14) } to match object { event_id: Any<String>, …(5) }
(92 matching properties omitted from actual)
- Expected
+ Received
@@ -1,12 +1,13 @@
{
"event_id": Any<String>,
- "spans": ArrayContaining [
- ObjectContaining {
+ "spans": [
+ {
"data": {
"gen_ai.input.messages": "[{\"role\":\"user\",\"content\":\"What is the capital of France?\"}]",
"gen_ai.operation.name": "chat",
+ "gen_ai.request.max_tokens": 100,
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.request.temperature": 0.7,
"gen_ai.response.finish_reasons": "[\"stop\"]",
"gen_ai.response.id": "chatcmpl-mock123",
"gen_ai.response.model": "gpt-3.5-turbo",
@@ -21,13 +22,47 @@
"sentry.sdk_meta.gen_ai.input.messages.original_length": 1,
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "3b454d47611a1735",
+ "span_id": "26c215bc3d2396f6",
+ "start_timestamp": 1775213261.162,
"status": "ok",
+ "timestamp": 1775213261.302652,
+ "trace_id": "4c5af42c2b6f400cd1f4b56d111ddfd7",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 39235,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 39235,
+ "url": "http://localhost:39235/openai/chat/completions",
+ "url.full": "http://localhost:39235/openai/chat/completions",
+ "url.path": "/openai/chat/completions",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
+ },
+ "description": "POST http://localhost:39235/openai/chat/completions",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "26c215bc3d2396f6",
+ "span_id": "219e81f2f705cc05",
+ "start_timestamp": 1775213261.202,
+ "status": "ok",
+ "timestamp": 1775213261.2987921,
+ "trace_id": "4c5af42c2b6f400cd1f4b56d111ddfd7",
+ },
+ {
"data": {
"gen_ai.input.messages": "Translate this to French: Hello",
"gen_ai.operation.name": "chat",
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.response.finish_reasons": "[\"completed\"]",
@@ -43,13 +78,47 @@
"sentry.sdk_meta.gen_ai.input.messages.original_length": 1,
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "3b454d47611a1735",
+ "span_id": "9ddba4044f23628e",
+ "start_timestamp": 1775213261.303,
"status": "ok",
+ "timestamp": 1775213261.3161745,
+ "trace_id": "4c5af42c2b6f400cd1f4b56d111ddfd7",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 39235,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 39235,
+ "url": "http://localhost:39235/openai/responses",
+ "url.full": "http://localhost:39235/openai/responses",
+ "url.path": "/openai/responses",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
+ },
+ "description": "POST http://localhost:39235/openai/responses",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "9ddba4044
|
|
Run integration tests:
dev-packages/node-integration-tests/utils/assertions.ts#L35
AssertionError: expected { contexts: { …(8) }, …(14) } to match object { event_id: Any<String>, …(5) }
(92 matching properties omitted from actual)
- Expected
+ Received
@@ -1,12 +1,13 @@
{
"event_id": Any<String>,
- "spans": ArrayContaining [
- ObjectContaining {
+ "spans": [
+ {
"data": {
"gen_ai.input.messages": "[{\"role\":\"user\",\"content\":\"What is the capital of France?\"}]",
"gen_ai.operation.name": "chat",
+ "gen_ai.request.max_tokens": 100,
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.request.temperature": 0.7,
"gen_ai.response.finish_reasons": "[\"stop\"]",
"gen_ai.response.id": "chatcmpl-mock123",
"gen_ai.response.model": "gpt-3.5-turbo",
@@ -21,13 +22,47 @@
"sentry.sdk_meta.gen_ai.input.messages.original_length": 1,
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "e23b7ec58542b494",
+ "span_id": "4f1d0e3bb5ccf4af",
+ "start_timestamp": 1775213257.388,
"status": "ok",
+ "timestamp": 1775213257.4592996,
+ "trace_id": "2f82e09198a280645be4e3d9243085ce",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 40131,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 40131,
+ "url": "http://localhost:40131/openai/chat/completions",
+ "url.full": "http://localhost:40131/openai/chat/completions",
+ "url.path": "/openai/chat/completions",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
+ },
+ "description": "POST http://localhost:40131/openai/chat/completions",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "4f1d0e3bb5ccf4af",
+ "span_id": "544400f91e4b89e7",
+ "start_timestamp": 1775213257.399,
+ "status": "ok",
+ "timestamp": 1775213257.4534042,
+ "trace_id": "2f82e09198a280645be4e3d9243085ce",
+ },
+ {
"data": {
"gen_ai.input.messages": "Translate this to French: Hello",
"gen_ai.operation.name": "chat",
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.response.finish_reasons": "[\"completed\"]",
@@ -43,13 +78,47 @@
"sentry.sdk_meta.gen_ai.input.messages.original_length": 1,
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "e23b7ec58542b494",
+ "span_id": "11d039a7aa09bf0a",
+ "start_timestamp": 1775213257.46,
"status": "ok",
+ "timestamp": 1775213257.4839168,
+ "trace_id": "2f82e09198a280645be4e3d9243085ce",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 40131,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 40131,
+ "url": "http://localhost:40131/openai/responses",
+ "url.full": "http://localhost:40131/openai/responses",
+ "url.path": "/openai/responses",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
+ },
+ "description": "POST http://localhost:40131/openai/responses",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "11d039a7a
|
|
Run integration tests:
dev-packages/node-integration-tests/utils/assertions.ts#L35
AssertionError: expected { contexts: { …(8) }, …(14) } to match object { event_id: Any<String>, …(5) }
(92 matching properties omitted from actual)
- Expected
+ Received
@@ -1,11 +1,12 @@
{
"event_id": Any<String>,
- "spans": ArrayContaining [
- ObjectContaining {
+ "spans": [
+ {
"data": {
"gen_ai.operation.name": "chat",
+ "gen_ai.request.max_tokens": 100,
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.request.temperature": 0.7,
"gen_ai.response.finish_reasons": "[\"stop\"]",
"gen_ai.response.id": "chatcmpl-mock123",
"gen_ai.response.model": "gpt-3.5-turbo",
@@ -17,13 +18,47 @@
"sentry.origin": "auto.ai.openai",
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "552cdac6b2420195",
+ "span_id": "3b7a048cc411ed59",
+ "start_timestamp": 1775213253.21,
"status": "ok",
+ "timestamp": 1775213253.3301497,
+ "trace_id": "e0cef40ff30f5f431548b20d7e838131",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 38991,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 38991,
+ "url": "http://localhost:38991/openai/chat/completions",
+ "url.full": "http://localhost:38991/openai/chat/completions",
+ "url.path": "/openai/chat/completions",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
+ },
+ "description": "POST http://localhost:38991/openai/chat/completions",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "3b7a048cc411ed59",
+ "span_id": "4cb0cfd7b34927fa",
+ "start_timestamp": 1775213253.245,
+ "status": "ok",
+ "timestamp": 1775213253.325248,
+ "trace_id": "e0cef40ff30f5f431548b20d7e838131",
+ },
+ {
"data": {
"gen_ai.operation.name": "chat",
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.response.finish_reasons": "[\"completed\"]",
"gen_ai.response.id": "resp_mock456",
@@ -36,26 +71,152 @@
"sentry.origin": "auto.ai.openai",
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "552cdac6b2420195",
+ "span_id": "3721332fb59c2dbb",
+ "start_timestamp": 1775213253.33,
"status": "ok",
+ "timestamp": 1775213253.3558068,
+ "trace_id": "e0cef40ff30f5f431548b20d7e838131",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 38991,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 38991,
+ "url": "http://localhost:38991/openai/responses",
+ "url.full": "http://localhost:38991/openai/responses",
+ "url.path": "/openai/responses",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
+ },
+ "description": "POST http://localhost:38991/openai/responses",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "3721332fb59c2dbb",
+ "span_id": "0f55ac50828cdd7a",
+ "start_timestamp": 1775213253.333,
+ "status": "ok",
+ "timestamp": 1775213253.3533673,
+ "trace_id": "e0
|
|
Run integration tests:
dev-packages/node-integration-tests/utils/assertions.ts#L35
AssertionError: expected { contexts: { …(8) }, …(14) } to match object { event_id: Any<String>, …(5) }
(92 matching properties omitted from actual)
- Expected
+ Received
@@ -1,11 +1,12 @@
{
"event_id": Any<String>,
- "spans": ArrayContaining [
- ObjectContaining {
+ "spans": [
+ {
"data": {
"gen_ai.operation.name": "chat",
+ "gen_ai.request.max_tokens": 100,
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.request.temperature": 0.7,
"gen_ai.response.finish_reasons": "[\"stop\"]",
"gen_ai.response.id": "chatcmpl-mock123",
"gen_ai.response.model": "gpt-3.5-turbo",
@@ -17,13 +18,47 @@
"sentry.origin": "auto.ai.openai",
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "669cacdcc907ae9c",
+ "span_id": "54c3576aeaa6bb49",
+ "start_timestamp": 1775213249.45,
"status": "ok",
+ "timestamp": 1775213249.6720672,
+ "trace_id": "555c99f57a80a3cd8db25d44da38285a",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 34121,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 34121,
+ "url": "http://localhost:34121/openai/chat/completions",
+ "url.full": "http://localhost:34121/openai/chat/completions",
+ "url.path": "/openai/chat/completions",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
+ },
+ "description": "POST http://localhost:34121/openai/chat/completions",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "54c3576aeaa6bb49",
+ "span_id": "4fb358f9cdc65971",
+ "start_timestamp": 1775213249.46,
+ "status": "ok",
+ "timestamp": 1775213249.668946,
+ "trace_id": "555c99f57a80a3cd8db25d44da38285a",
+ },
+ {
"data": {
"gen_ai.operation.name": "chat",
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.response.finish_reasons": "[\"completed\"]",
"gen_ai.response.id": "resp_mock456",
@@ -36,26 +71,152 @@
"sentry.origin": "auto.ai.openai",
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "669cacdcc907ae9c",
+ "span_id": "dbca9a64d3623ab9",
+ "start_timestamp": 1775213249.672,
"status": "ok",
+ "timestamp": 1775213249.6835241,
+ "trace_id": "555c99f57a80a3cd8db25d44da38285a",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 34121,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 34121,
+ "url": "http://localhost:34121/openai/responses",
+ "url.full": "http://localhost:34121/openai/responses",
+ "url.path": "/openai/responses",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
+ },
+ "description": "POST http://localhost:34121/openai/responses",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "dbca9a64d3623ab9",
+ "span_id": "c79d6693715c94af",
+ "start_timestamp": 1775213249.674,
+ "status": "ok",
+ "timestamp": 1775213249.682799,
+ "trace_id": "555
|
|
Complete job
Node.js 20 actions are deprecated. The following actions are running on Node.js 20 and may not work as expected: actions/cache/restore@v4, actions/download-artifact@v4. Actions will be forced to run with Node.js 24 by default starting June 2nd, 2026. Node.js 20 will be removed from the runner on September 16th, 2026. Please check if updated versions of these actions are available that support Node.js 24. To opt into Node.js 24 now, set the FORCE_JAVASCRIPT_ACTIONS_TO_NODE24=true environment variable on the runner or in your workflow file. Once Node.js 24 becomes the default, you can temporarily opt out by setting ACTIONS_ALLOW_USE_UNSECURE_NODE_VERSION=true. For more information see: https://github.blog/changelog/2025-09-19-deprecation-of-node-20-on-github-actions-runners/
|
|
Parse and Upload Coverage
Supported formats: clover, cobertura, jacoco, lcov, istanbul, go, codecov
|
|
Parse and Upload Coverage
No coverage files found
|
|
Parse and Upload Coverage
Please ensure your test framework is generating JUnit XML output.
|
|
Parse and Upload Coverage
No JUnit XML files found matching pattern: dev-packages/node-integration-tests/**/*.junit.xml
|
|
Parse and Upload Coverage
No files found matching pattern: dev-packages/node-integration-tests/**/*.junit.xml
|
Loading