@@ -171,8 +171,7 @@ def traced_llm_generate(langchain, pin, func, instance, args, kwargs):
171
171
span = integration .trace (
172
172
pin ,
173
173
"%s.%s" % (instance .__module__ , instance .__class__ .__name__ ),
174
- # only report LLM Obs spans if base_url has not been changed
175
- submit_to_llmobs = integration .has_default_base_url (instance ),
174
+ submit_to_llmobs = True ,
176
175
interface_type = "llm" ,
177
176
provider = llm_provider ,
178
177
model = model ,
@@ -230,8 +229,7 @@ async def traced_llm_agenerate(langchain, pin, func, instance, args, kwargs):
230
229
span = integration .trace (
231
230
pin ,
232
231
"%s.%s" % (instance .__module__ , instance .__class__ .__name__ ),
233
- # only report LLM Obs spans if base_url has not been changed
234
- submit_to_llmobs = integration .has_default_base_url (instance ),
232
+ submit_to_llmobs = True ,
235
233
interface_type = "llm" ,
236
234
provider = llm_provider ,
237
235
model = model ,
@@ -288,8 +286,7 @@ def traced_chat_model_generate(langchain, pin, func, instance, args, kwargs):
288
286
span = integration .trace (
289
287
pin ,
290
288
"%s.%s" % (instance .__module__ , instance .__class__ .__name__ ),
291
- # only report LLM Obs spans if base_url has not been changed
292
- submit_to_llmobs = integration .has_default_base_url (instance ),
289
+ submit_to_llmobs = True ,
293
290
interface_type = "chat_model" ,
294
291
provider = llm_provider ,
295
292
model = _extract_model_name (instance ),
@@ -385,8 +382,7 @@ async def traced_chat_model_agenerate(langchain, pin, func, instance, args, kwar
385
382
span = integration .trace (
386
383
pin ,
387
384
"%s.%s" % (instance .__module__ , instance .__class__ .__name__ ),
388
- # only report LLM Obs spans if base_url has not been changed
389
- submit_to_llmobs = integration .has_default_base_url (instance ),
385
+ submit_to_llmobs = True ,
390
386
interface_type = "chat_model" ,
391
387
provider = llm_provider ,
392
388
model = _extract_model_name (instance ),
0 commit comments