Nh/unify request attribute extraction #45314
Annotations
9 errors and 6 warnings
|
Run integration tests
Process completed with exit code 1.
|
|
Run integration tests:
dev-packages/node-integration-tests/utils/assertions.ts#L35
AssertionError: expected { contexts: { …(8) }, …(14) } to match object { event_id: Any<String>, …(5) }
(92 matching properties omitted from actual)
- Expected
+ Received
@@ -1,12 +1,13 @@
{
"event_id": Any<String>,
- "spans": ArrayContaining [
- ObjectContaining {
+ "spans": [
+ {
"data": {
"gen_ai.input.messages": "[{\"role\":\"user\",\"content\":\"What is the capital of France?\"}]",
"gen_ai.operation.name": "chat",
+ "gen_ai.request.max_tokens": 100,
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.request.temperature": 0.7,
"gen_ai.response.finish_reasons": "[\"stop\"]",
"gen_ai.response.id": "chatcmpl-mock123",
"gen_ai.response.model": "gpt-3.5-turbo",
@@ -21,13 +22,47 @@
"sentry.sdk_meta.gen_ai.input.messages.original_length": 1,
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "1e1e98a04043ff59",
+ "span_id": "be45418213235b00",
+ "start_timestamp": 1775213157.686,
"status": "ok",
+ "timestamp": 1775213157.7765694,
+ "trace_id": "5fd99603b4980d56c0f9ac57ad9807b3",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 33725,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 33725,
+ "url": "http://localhost:33725/openai/chat/completions",
+ "url.full": "http://localhost:33725/openai/chat/completions",
+ "url.path": "/openai/chat/completions",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 6.0.0",
+ },
+ "description": "POST http://localhost:33725/openai/chat/completions",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "be45418213235b00",
+ "span_id": "ea047112837e9a09",
+ "start_timestamp": 1775213157.719,
+ "status": "ok",
+ "timestamp": 1775213157.771571,
+ "trace_id": "5fd99603b4980d56c0f9ac57ad9807b3",
+ },
+ {
"data": {
"gen_ai.input.messages": "Translate this to French: Hello",
"gen_ai.operation.name": "chat",
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.response.finish_reasons": "[\"completed\"]",
@@ -43,13 +78,47 @@
"sentry.sdk_meta.gen_ai.input.messages.original_length": 1,
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "1e1e98a04043ff59",
+ "span_id": "15b2191c68a58a00",
+ "start_timestamp": 1775213157.777,
"status": "ok",
+ "timestamp": 1775213157.7995145,
+ "trace_id": "5fd99603b4980d56c0f9ac57ad9807b3",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 33725,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 33725,
+ "url": "http://localhost:33725/openai/responses",
+ "url.full": "http://localhost:33725/openai/responses",
+ "url.path": "/openai/responses",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 6.0.0",
+ },
+ "description": "POST http://localhost:33725/openai/responses",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "15b2191c68a
|
|
Run integration tests:
dev-packages/node-integration-tests/utils/assertions.ts#L35
AssertionError: expected { contexts: { …(8) }, …(14) } to match object { event_id: Any<String>, …(5) }
(92 matching properties omitted from actual)
- Expected
+ Received
@@ -1,12 +1,13 @@
{
"event_id": Any<String>,
- "spans": ArrayContaining [
- ObjectContaining {
+ "spans": [
+ {
"data": {
"gen_ai.input.messages": "[{\"role\":\"user\",\"content\":\"What is the capital of France?\"}]",
"gen_ai.operation.name": "chat",
+ "gen_ai.request.max_tokens": 100,
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.request.temperature": 0.7,
"gen_ai.response.finish_reasons": "[\"stop\"]",
"gen_ai.response.id": "chatcmpl-mock123",
"gen_ai.response.model": "gpt-3.5-turbo",
@@ -21,13 +22,47 @@
"sentry.sdk_meta.gen_ai.input.messages.original_length": 1,
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "3bc2b8b6d27ec3d8",
+ "span_id": "7cd82a8bdbdbfad1",
+ "start_timestamp": 1775213153.933,
"status": "ok",
+ "timestamp": 1775213154.033139,
+ "trace_id": "52e641767af549f4ae36550cc45c5221",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 42013,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 42013,
+ "url": "http://localhost:42013/openai/chat/completions",
+ "url.full": "http://localhost:42013/openai/chat/completions",
+ "url.path": "/openai/chat/completions",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 6.0.0",
+ },
+ "description": "POST http://localhost:42013/openai/chat/completions",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "7cd82a8bdbdbfad1",
+ "span_id": "67e2686d34682474",
+ "start_timestamp": 1775213153.968,
+ "status": "ok",
+ "timestamp": 1775213154.0273008,
+ "trace_id": "52e641767af549f4ae36550cc45c5221",
+ },
+ {
"data": {
"gen_ai.input.messages": "Translate this to French: Hello",
"gen_ai.operation.name": "chat",
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.response.finish_reasons": "[\"completed\"]",
@@ -43,13 +78,47 @@
"sentry.sdk_meta.gen_ai.input.messages.original_length": 1,
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "3bc2b8b6d27ec3d8",
+ "span_id": "fa9037b75ee25be5",
+ "start_timestamp": 1775213154.033,
"status": "ok",
+ "timestamp": 1775213154.0524077,
+ "trace_id": "52e641767af549f4ae36550cc45c5221",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 42013,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 42013,
+ "url": "http://localhost:42013/openai/responses",
+ "url.full": "http://localhost:42013/openai/responses",
+ "url.path": "/openai/responses",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 6.0.0",
+ },
+ "description": "POST http://localhost:42013/openai/responses",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "fa9037b75ee
|
|
Run integration tests:
dev-packages/node-integration-tests/utils/assertions.ts#L35
AssertionError: expected { contexts: { …(8) }, …(14) } to match object { event_id: Any<String>, …(5) }
(92 matching properties omitted from actual)
- Expected
+ Received
@@ -1,11 +1,12 @@
{
"event_id": Any<String>,
- "spans": ArrayContaining [
- ObjectContaining {
+ "spans": [
+ {
"data": {
"gen_ai.operation.name": "chat",
+ "gen_ai.request.max_tokens": 100,
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.request.temperature": 0.7,
"gen_ai.response.finish_reasons": "[\"stop\"]",
"gen_ai.response.id": "chatcmpl-mock123",
"gen_ai.response.model": "gpt-3.5-turbo",
@@ -17,13 +18,47 @@
"sentry.origin": "auto.ai.openai",
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "4fb09e202a8299c3",
+ "span_id": "83a8cdce74f0b52b",
+ "start_timestamp": 1775213148.232,
"status": "ok",
+ "timestamp": 1775213148.3856585,
+ "trace_id": "74262e228c4f55a54b0b5f660ada4b9b",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 37329,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 37329,
+ "url": "http://localhost:37329/openai/chat/completions",
+ "url.full": "http://localhost:37329/openai/chat/completions",
+ "url.path": "/openai/chat/completions",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 6.0.0",
+ },
+ "description": "POST http://localhost:37329/openai/chat/completions",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "83a8cdce74f0b52b",
+ "span_id": "3853f1c2c3b3fa25",
+ "start_timestamp": 1775213148.29,
+ "status": "ok",
+ "timestamp": 1775213148.3802154,
+ "trace_id": "74262e228c4f55a54b0b5f660ada4b9b",
+ },
+ {
"data": {
"gen_ai.operation.name": "chat",
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.response.finish_reasons": "[\"completed\"]",
"gen_ai.response.id": "resp_mock456",
@@ -36,26 +71,152 @@
"sentry.origin": "auto.ai.openai",
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "4fb09e202a8299c3",
+ "span_id": "800e2b0c7de8df1b",
+ "start_timestamp": 1775213148.386,
"status": "ok",
+ "timestamp": 1775213148.4105349,
+ "trace_id": "74262e228c4f55a54b0b5f660ada4b9b",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 37329,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 37329,
+ "url": "http://localhost:37329/openai/responses",
+ "url.full": "http://localhost:37329/openai/responses",
+ "url.path": "/openai/responses",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 6.0.0",
+ },
+ "description": "POST http://localhost:37329/openai/responses",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "800e2b0c7de8df1b",
+ "span_id": "9fe793ea06aa2467",
+ "start_timestamp": 1775213148.392,
+ "status": "ok",
+ "timestamp": 1775213148.4094372,
+ "trace_id": "74
|
|
Run integration tests:
dev-packages/node-integration-tests/utils/assertions.ts#L35
AssertionError: expected { contexts: { …(8) }, …(14) } to match object { event_id: Any<String>, …(5) }
(92 matching properties omitted from actual)
- Expected
+ Received
@@ -1,11 +1,12 @@
{
"event_id": Any<String>,
- "spans": ArrayContaining [
- ObjectContaining {
+ "spans": [
+ {
"data": {
"gen_ai.operation.name": "chat",
+ "gen_ai.request.max_tokens": 100,
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.request.temperature": 0.7,
"gen_ai.response.finish_reasons": "[\"stop\"]",
"gen_ai.response.id": "chatcmpl-mock123",
"gen_ai.response.model": "gpt-3.5-turbo",
@@ -17,13 +18,47 @@
"sentry.origin": "auto.ai.openai",
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "bf5fdc0ebe15bede",
+ "span_id": "176d7787da9977eb",
+ "start_timestamp": 1775213144.226,
"status": "ok",
+ "timestamp": 1775213144.3489854,
+ "trace_id": "2b950455e614f8e74905390962f4c79b",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 35093,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 35093,
+ "url": "http://localhost:35093/openai/chat/completions",
+ "url.full": "http://localhost:35093/openai/chat/completions",
+ "url.path": "/openai/chat/completions",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 6.0.0",
+ },
+ "description": "POST http://localhost:35093/openai/chat/completions",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "176d7787da9977eb",
+ "span_id": "fd629183eb81da23",
+ "start_timestamp": 1775213144.276,
+ "status": "ok",
+ "timestamp": 1775213144.344187,
+ "trace_id": "2b950455e614f8e74905390962f4c79b",
+ },
+ {
"data": {
"gen_ai.operation.name": "chat",
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.response.finish_reasons": "[\"completed\"]",
"gen_ai.response.id": "resp_mock456",
@@ -36,26 +71,152 @@
"sentry.origin": "auto.ai.openai",
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "bf5fdc0ebe15bede",
+ "span_id": "77fbe11f8f5f3fd9",
+ "start_timestamp": 1775213144.35,
"status": "ok",
+ "timestamp": 1775213144.368448,
+ "trace_id": "2b950455e614f8e74905390962f4c79b",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 35093,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 35093,
+ "url": "http://localhost:35093/openai/responses",
+ "url.full": "http://localhost:35093/openai/responses",
+ "url.path": "/openai/responses",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 6.0.0",
+ },
+ "description": "POST http://localhost:35093/openai/responses",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "77fbe11f8f5f3fd9",
+ "span_id": "5317e1002b53ccf4",
+ "start_timestamp": 1775213144.356,
+ "status": "ok",
+ "timestamp": 1775213144.3668969,
+ "trace_id": "2b95
|
|
Run integration tests:
dev-packages/node-integration-tests/utils/assertions.ts#L35
AssertionError: expected { contexts: { …(8) }, …(14) } to match object { event_id: Any<String>, …(5) }
(92 matching properties omitted from actual)
- Expected
+ Received
@@ -1,12 +1,13 @@
{
"event_id": Any<String>,
- "spans": ArrayContaining [
- ObjectContaining {
+ "spans": [
+ {
"data": {
"gen_ai.input.messages": "[{\"role\":\"user\",\"content\":\"What is the capital of France?\"}]",
"gen_ai.operation.name": "chat",
+ "gen_ai.request.max_tokens": 100,
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.request.temperature": 0.7,
"gen_ai.response.finish_reasons": "[\"stop\"]",
"gen_ai.response.id": "chatcmpl-mock123",
"gen_ai.response.model": "gpt-3.5-turbo",
@@ -21,13 +22,47 @@
"sentry.sdk_meta.gen_ai.input.messages.original_length": 1,
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "2bc3363ecff0df98",
+ "span_id": "e591dfbae4c7bff3",
+ "start_timestamp": 1775213089.037,
"status": "ok",
+ "timestamp": 1775213089.1335635,
+ "trace_id": "67037b0695dcea7a9ebe42d8053fd81c",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 46325,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 46325,
+ "url": "http://localhost:46325/openai/chat/completions",
+ "url.full": "http://localhost:46325/openai/chat/completions",
+ "url.path": "/openai/chat/completions",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
+ },
+ "description": "POST http://localhost:46325/openai/chat/completions",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "e591dfbae4c7bff3",
+ "span_id": "56d6f758d07c4ab7",
+ "start_timestamp": 1775213089.073,
+ "status": "ok",
+ "timestamp": 1775213089.12776,
+ "trace_id": "67037b0695dcea7a9ebe42d8053fd81c",
+ },
+ {
"data": {
"gen_ai.input.messages": "Translate this to French: Hello",
"gen_ai.operation.name": "chat",
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.response.finish_reasons": "[\"completed\"]",
@@ -43,13 +78,47 @@
"sentry.sdk_meta.gen_ai.input.messages.original_length": 1,
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "2bc3363ecff0df98",
+ "span_id": "407e5e9408ba2d0f",
+ "start_timestamp": 1775213089.134,
"status": "ok",
+ "timestamp": 1775213089.154512,
+ "trace_id": "67037b0695dcea7a9ebe42d8053fd81c",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 46325,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 46325,
+ "url": "http://localhost:46325/openai/responses",
+ "url.full": "http://localhost:46325/openai/responses",
+ "url.path": "/openai/responses",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
+ },
+ "description": "POST http://localhost:46325/openai/responses",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "407e5e9408b
|
|
Run integration tests:
dev-packages/node-integration-tests/utils/assertions.ts#L35
AssertionError: expected { contexts: { …(8) }, …(14) } to match object { event_id: Any<String>, …(5) }
(92 matching properties omitted from actual)
- Expected
+ Received
@@ -1,12 +1,13 @@
{
"event_id": Any<String>,
- "spans": ArrayContaining [
- ObjectContaining {
+ "spans": [
+ {
"data": {
"gen_ai.input.messages": "[{\"role\":\"user\",\"content\":\"What is the capital of France?\"}]",
"gen_ai.operation.name": "chat",
+ "gen_ai.request.max_tokens": 100,
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.request.temperature": 0.7,
"gen_ai.response.finish_reasons": "[\"stop\"]",
"gen_ai.response.id": "chatcmpl-mock123",
"gen_ai.response.model": "gpt-3.5-turbo",
@@ -21,13 +22,47 @@
"sentry.sdk_meta.gen_ai.input.messages.original_length": 1,
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "9fc5c069b88a6dd3",
+ "span_id": "f65269fa02741f15",
+ "start_timestamp": 1775213085.503,
"status": "ok",
+ "timestamp": 1775213085.6052125,
+ "trace_id": "a8612d1e18713c110c8e8979ea5fdffd",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 38481,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 38481,
+ "url": "http://localhost:38481/openai/chat/completions",
+ "url.full": "http://localhost:38481/openai/chat/completions",
+ "url.path": "/openai/chat/completions",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
+ },
+ "description": "POST http://localhost:38481/openai/chat/completions",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "f65269fa02741f15",
+ "span_id": "ae20448be7fcf156",
+ "start_timestamp": 1775213085.535,
+ "status": "ok",
+ "timestamp": 1775213085.6015606,
+ "trace_id": "a8612d1e18713c110c8e8979ea5fdffd",
+ },
+ {
"data": {
"gen_ai.input.messages": "Translate this to French: Hello",
"gen_ai.operation.name": "chat",
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.response.finish_reasons": "[\"completed\"]",
@@ -43,13 +78,47 @@
"sentry.sdk_meta.gen_ai.input.messages.original_length": 1,
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "9fc5c069b88a6dd3",
+ "span_id": "8a652653a25f614a",
+ "start_timestamp": 1775213085.606,
"status": "ok",
+ "timestamp": 1775213085.622185,
+ "trace_id": "a8612d1e18713c110c8e8979ea5fdffd",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 38481,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 38481,
+ "url": "http://localhost:38481/openai/responses",
+ "url.full": "http://localhost:38481/openai/responses",
+ "url.path": "/openai/responses",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
+ },
+ "description": "POST http://localhost:38481/openai/responses",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "8a652653a
|
|
Run integration tests:
dev-packages/node-integration-tests/utils/assertions.ts#L35
AssertionError: expected { contexts: { …(8) }, …(14) } to match object { event_id: Any<String>, …(5) }
(92 matching properties omitted from actual)
- Expected
+ Received
@@ -1,11 +1,12 @@
{
"event_id": Any<String>,
- "spans": ArrayContaining [
- ObjectContaining {
+ "spans": [
+ {
"data": {
"gen_ai.operation.name": "chat",
+ "gen_ai.request.max_tokens": 100,
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.request.temperature": 0.7,
"gen_ai.response.finish_reasons": "[\"stop\"]",
"gen_ai.response.id": "chatcmpl-mock123",
"gen_ai.response.model": "gpt-3.5-turbo",
@@ -17,13 +18,47 @@
"sentry.origin": "auto.ai.openai",
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "fb26fbc921e5d139",
+ "span_id": "d46643f457d5a500",
+ "start_timestamp": 1775213081.143,
"status": "ok",
+ "timestamp": 1775213081.249633,
+ "trace_id": "db9a4202b3fe069335003ebafc39ccf5",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 37245,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 37245,
+ "url": "http://localhost:37245/openai/chat/completions",
+ "url.full": "http://localhost:37245/openai/chat/completions",
+ "url.path": "/openai/chat/completions",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
+ },
+ "description": "POST http://localhost:37245/openai/chat/completions",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "d46643f457d5a500",
+ "span_id": "9aff52d75c4e36dc",
+ "start_timestamp": 1775213081.181,
+ "status": "ok",
+ "timestamp": 1775213081.245049,
+ "trace_id": "db9a4202b3fe069335003ebafc39ccf5",
+ },
+ {
"data": {
"gen_ai.operation.name": "chat",
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.response.finish_reasons": "[\"completed\"]",
"gen_ai.response.id": "resp_mock456",
@@ -36,26 +71,152 @@
"sentry.origin": "auto.ai.openai",
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "fb26fbc921e5d139",
+ "span_id": "e20a361aec47ef8d",
+ "start_timestamp": 1775213081.251,
"status": "ok",
+ "timestamp": 1775213081.2725525,
+ "trace_id": "db9a4202b3fe069335003ebafc39ccf5",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 37245,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 37245,
+ "url": "http://localhost:37245/openai/responses",
+ "url.full": "http://localhost:37245/openai/responses",
+ "url.path": "/openai/responses",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
+ },
+ "description": "POST http://localhost:37245/openai/responses",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "e20a361aec47ef8d",
+ "span_id": "3d82cebe85fd3bd0",
+ "start_timestamp": 1775213081.254,
+ "status": "ok",
+ "timestamp": 1775213081.2706766,
+ "trace_id": "d
|
|
Run integration tests:
dev-packages/node-integration-tests/utils/assertions.ts#L35
AssertionError: expected { contexts: { …(8) }, …(14) } to match object { event_id: Any<String>, …(5) }
(92 matching properties omitted from actual)
- Expected
+ Received
@@ -1,11 +1,12 @@
{
"event_id": Any<String>,
- "spans": ArrayContaining [
- ObjectContaining {
+ "spans": [
+ {
"data": {
"gen_ai.operation.name": "chat",
+ "gen_ai.request.max_tokens": 100,
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.request.temperature": 0.7,
"gen_ai.response.finish_reasons": "[\"stop\"]",
"gen_ai.response.id": "chatcmpl-mock123",
"gen_ai.response.model": "gpt-3.5-turbo",
@@ -17,13 +18,47 @@
"sentry.origin": "auto.ai.openai",
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "f89f782fd71ce252",
+ "span_id": "7f1726d7f7dc0454",
+ "start_timestamp": 1775213077.139,
"status": "ok",
+ "timestamp": 1775213077.2498863,
+ "trace_id": "1e6c7a633a0279e6b314832e8004a212",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 33495,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 33495,
+ "url": "http://localhost:33495/openai/chat/completions",
+ "url.full": "http://localhost:33495/openai/chat/completions",
+ "url.path": "/openai/chat/completions",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
+ },
+ "description": "POST http://localhost:33495/openai/chat/completions",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "7f1726d7f7dc0454",
+ "span_id": "8c2dde810f7682c2",
+ "start_timestamp": 1775213077.174,
+ "status": "ok",
+ "timestamp": 1775213077.2433662,
+ "trace_id": "1e6c7a633a0279e6b314832e8004a212",
+ },
+ {
"data": {
"gen_ai.operation.name": "chat",
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.response.finish_reasons": "[\"completed\"]",
"gen_ai.response.id": "resp_mock456",
@@ -36,26 +71,152 @@
"sentry.origin": "auto.ai.openai",
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "f89f782fd71ce252",
+ "span_id": "8d177d1c2b5a4fd0",
+ "start_timestamp": 1775213077.25,
"status": "ok",
+ "timestamp": 1775213077.281431,
+ "trace_id": "1e6c7a633a0279e6b314832e8004a212",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 33495,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 33495,
+ "url": "http://localhost:33495/openai/responses",
+ "url.full": "http://localhost:33495/openai/responses",
+ "url.path": "/openai/responses",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
+ },
+ "description": "POST http://localhost:33495/openai/responses",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "8d177d1c2b5a4fd0",
+ "span_id": "199af88ff8394ce5",
+ "start_timestamp": 1775213077.255,
+ "status": "ok",
+ "timestamp": 1775213077.280139,
+ "trace_id": "1e
|
|
Complete job
Node.js 20 actions are deprecated. The following actions are running on Node.js 20 and may not work as expected: actions/cache/restore@v4, actions/download-artifact@v4. Actions will be forced to run with Node.js 24 by default starting June 2nd, 2026. Node.js 20 will be removed from the runner on September 16th, 2026. Please check if updated versions of these actions are available that support Node.js 24. To opt into Node.js 24 now, set the FORCE_JAVASCRIPT_ACTIONS_TO_NODE24=true environment variable on the runner or in your workflow file. Once Node.js 24 becomes the default, you can temporarily opt out by setting ACTIONS_ALLOW_USE_UNSECURE_NODE_VERSION=true. For more information see: https://github.blog/changelog/2025-09-19-deprecation-of-node-20-on-github-actions-runners/
|
|
Parse and Upload Coverage
Supported formats: clover, cobertura, jacoco, lcov, istanbul, go, codecov
|
|
Parse and Upload Coverage
No coverage files found
|
|
Parse and Upload Coverage
Please ensure your test framework is generating JUnit XML output.
|
|
Parse and Upload Coverage
No JUnit XML files found matching pattern: dev-packages/node-integration-tests/**/*.junit.xml
|
|
Parse and Upload Coverage
No files found matching pattern: dev-packages/node-integration-tests/**/*.junit.xml
|
Loading