Some checks failed
Lock Threads / action (push) Has been cancelled
Mark stale issues and pull requests / stale (push) Has been cancelled
Publish Chatwoot EE docker images / build (linux/amd64, ubuntu-latest) (push) Has been cancelled
Publish Chatwoot EE docker images / build (linux/arm64, ubuntu-22.04-arm) (push) Has been cancelled
Publish Chatwoot EE docker images / merge (push) Has been cancelled
Publish Chatwoot CE docker images / build (linux/amd64, ubuntu-latest) (push) Has been cancelled
Publish Chatwoot CE docker images / build (linux/arm64, ubuntu-22.04-arm) (push) Has been cancelled
Publish Chatwoot CE docker images / merge (push) Has been cancelled
Run Chatwoot CE spec / lint-backend (push) Has been cancelled
Run Chatwoot CE spec / lint-frontend (push) Has been cancelled
Run Chatwoot CE spec / frontend-tests (push) Has been cancelled
Run Chatwoot CE spec / backend-tests (0, 16) (push) Has been cancelled
Run Chatwoot CE spec / backend-tests (1, 16) (push) Has been cancelled
Run Chatwoot CE spec / backend-tests (10, 16) (push) Has been cancelled
Run Chatwoot CE spec / backend-tests (11, 16) (push) Has been cancelled
Run Chatwoot CE spec / backend-tests (12, 16) (push) Has been cancelled
Run Chatwoot CE spec / backend-tests (13, 16) (push) Has been cancelled
Run Chatwoot CE spec / backend-tests (14, 16) (push) Has been cancelled
Run Chatwoot CE spec / backend-tests (15, 16) (push) Has been cancelled
Run Chatwoot CE spec / backend-tests (2, 16) (push) Has been cancelled
Run Chatwoot CE spec / backend-tests (3, 16) (push) Has been cancelled
Run Chatwoot CE spec / backend-tests (4, 16) (push) Has been cancelled
Run Chatwoot CE spec / backend-tests (5, 16) (push) Has been cancelled
Run Chatwoot CE spec / backend-tests (6, 16) (push) Has been cancelled
Run Chatwoot CE spec / backend-tests (7, 16) (push) Has been cancelled
Run Chatwoot CE spec / backend-tests (8, 16) (push) Has been cancelled
Run Chatwoot CE spec / backend-tests (9, 16) (push) Has been cancelled
Run Linux nightly installer / nightly (push) Has been cancelled
- Add Logistics component with progress tracking - Add OrderDetail component for order information - Support data-driven steps and actions - Add blue color scale to widget SCSS - Fix node overflow and progress bar rendering issues - Add English translations for dashboard components Co-Authored-By: Claude Sonnet 4.5 <noreply@anthropic.com>
132 lines
3.9 KiB
Ruby
132 lines
3.9 KiB
Ruby
module Captain::ChatHelper
|
|
include Integrations::LlmInstrumentation
|
|
include Captain::ChatResponseHelper
|
|
include Captain::ChatGenerationRecorder
|
|
|
|
def request_chat_completion
|
|
log_chat_completion_request
|
|
|
|
chat = build_chat
|
|
|
|
add_messages_to_chat(chat)
|
|
with_agent_session do
|
|
response = chat.ask(conversation_messages.last[:content])
|
|
build_response(response)
|
|
end
|
|
rescue StandardError => e
|
|
Rails.logger.error "#{self.class.name} Assistant: #{@assistant.id}, Error in chat completion: #{e}"
|
|
raise e
|
|
end
|
|
|
|
private
|
|
|
|
def build_chat
|
|
llm_chat = chat(model: @model, temperature: temperature)
|
|
llm_chat = llm_chat.with_params(response_format: { type: 'json_object' })
|
|
|
|
llm_chat = setup_tools(llm_chat)
|
|
llm_chat = setup_system_instructions(llm_chat)
|
|
setup_event_handlers(llm_chat)
|
|
end
|
|
|
|
def setup_tools(llm_chat)
|
|
@tools&.each do |tool|
|
|
llm_chat = llm_chat.with_tool(tool)
|
|
end
|
|
llm_chat
|
|
end
|
|
|
|
def setup_system_instructions(chat)
|
|
system_messages = @messages.select { |m| m[:role] == 'system' || m[:role] == :system }
|
|
combined_instructions = system_messages.pluck(:content).join("\n\n")
|
|
chat.with_instructions(combined_instructions)
|
|
end
|
|
|
|
def setup_event_handlers(chat)
|
|
# NOTE: We only use on_end_message to record the generation with token counts.
|
|
# RubyLLM callbacks fire after chunks arrive, not around the API call, so
|
|
# span timing won't reflect actual API latency. But Langfuse calculates costs
|
|
# from model + token counts, so this is sufficient for cost tracking.
|
|
chat.on_end_message { |message| record_llm_generation(chat, message) }
|
|
chat.on_tool_call { |tool_call| handle_tool_call(tool_call) }
|
|
chat.on_tool_result { |result| handle_tool_result(result) }
|
|
|
|
chat
|
|
end
|
|
|
|
def handle_tool_call(tool_call)
|
|
persist_thinking_message(tool_call)
|
|
start_tool_span(tool_call)
|
|
@pending_tool_calls ||= []
|
|
@pending_tool_calls.push(tool_call)
|
|
end
|
|
|
|
def handle_tool_result(result)
|
|
end_tool_span(result)
|
|
persist_tool_completion
|
|
end
|
|
|
|
def add_messages_to_chat(chat)
|
|
conversation_messages[0...-1].each do |msg|
|
|
chat.add_message(role: msg[:role].to_sym, content: msg[:content])
|
|
end
|
|
end
|
|
|
|
def instrumentation_params(chat = nil)
|
|
{
|
|
span_name: "llm.captain.#{feature_name}",
|
|
account_id: resolved_account_id,
|
|
conversation_id: @conversation_id,
|
|
feature_name: feature_name,
|
|
model: @model,
|
|
messages: chat ? chat.messages.map { |m| { role: m.role.to_s, content: m.content.to_s } } : @messages,
|
|
temperature: temperature,
|
|
metadata: {
|
|
assistant_id: @assistant&.id
|
|
}
|
|
}
|
|
end
|
|
|
|
def conversation_messages
|
|
@messages.reject { |m| m[:role] == 'system' || m[:role] == :system }
|
|
end
|
|
|
|
def temperature
|
|
@assistant&.config&.[]('temperature').to_f || 1
|
|
end
|
|
|
|
def resolved_account_id
|
|
@account&.id || @assistant&.account_id
|
|
end
|
|
|
|
# Ensures all LLM calls and tool executions within an agentic loop
|
|
# are grouped under a single trace/session in Langfuse.
|
|
#
|
|
# Without this guard, each recursive call to request_chat_completion
|
|
# (triggered by tool calls) would create a separate trace instead of
|
|
# nesting within the existing session span.
|
|
def with_agent_session(&)
|
|
already_active = @agent_session_active
|
|
return yield if already_active
|
|
|
|
@agent_session_active = true
|
|
instrument_agent_session(instrumentation_params, &)
|
|
ensure
|
|
@agent_session_active = false unless already_active
|
|
end
|
|
|
|
# Must be implemented by including class to identify the feature for instrumentation.
|
|
# Used for Langfuse tagging and span naming.
|
|
def feature_name
|
|
raise NotImplementedError, "#{self.class.name} must implement #feature_name"
|
|
end
|
|
|
|
def log_chat_completion_request
|
|
Rails.logger.info(
|
|
"#{self.class.name} Assistant: #{@assistant.id}, Requesting chat completion
|
|
for messages #{@messages} with #{@tools&.length || 0} tools
|
|
"
|
|
)
|
|
end
|
|
end
|