Files
laravel-llm-gateway/laravel-app/app/Services/LLM/RequestLogger.php
wtrinkl 6573e15ba4 Add complete Laravel LLM Gateway implementation
Core Features:
- Multi-provider support (OpenAI, Anthropic, DeepSeek, Gemini, Mistral)
- Provider service architecture with abstract base class
- Dynamic model discovery from provider APIs
- Encrypted per-user provider credentials storage

Admin Interface:
- Complete admin panel with Livewire components
- User management with CRUD operations
- API key management with testing capabilities
- Budget system with limits and reset schedules
- Usage logs with filtering and CSV export
- Model pricing management with cost calculator
- Dashboard with Chart.js visualizations

Database Schema:
- MariaDB migrations for all tables
- User provider credentials (encrypted)
- LLM request logging
- Budget tracking and rate limiting
- Model pricing configuration

API Implementation:
- OpenAI-compatible endpoints
- Budget checking middleware
- Rate limit enforcement
- Request logging jobs
- Cost calculation service

Testing:
- Unit tests for all provider services
- Provider factory tests
- Cost calculator tests

Documentation:
- Admin user seeder
- Model pricing seeder
- Configuration files
2025-11-18 22:18:36 +01:00

97 lines
2.5 KiB
PHP

<?php
namespace App\Services\LLM;
use App\Jobs\LogLlmRequest;
use Illuminate\Support\Str;
class RequestLogger
{
/**
* Log a successful LLM request
*/
public function logSuccess(
int $userId,
string $provider,
string $model,
array $requestPayload,
array $responsePayload,
array $costs,
int $responseTimeMs,
?string $ipAddress = null,
?string $userAgent = null
): string {
$requestId = $this->generateRequestId();
LogLlmRequest::dispatch(
userId: $userId,
provider: $provider,
model: $model,
requestPayload: $requestPayload,
responsePayload: $responsePayload,
promptTokens: $responsePayload['usage']['prompt_tokens'] ?? 0,
completionTokens: $responsePayload['usage']['completion_tokens'] ?? 0,
totalTokens: $responsePayload['usage']['total_tokens'] ?? 0,
responseTimeMs: $responseTimeMs,
promptCost: $costs['prompt_cost'],
completionCost: $costs['completion_cost'],
totalCost: $costs['total_cost'],
status: 'success',
errorMessage: null,
httpStatus: 200,
ipAddress: $ipAddress,
userAgent: $userAgent,
requestId: $requestId
);
return $requestId;
}
/**
* Log a failed LLM request
*/
public function logFailure(
int $userId,
string $provider,
string $model,
array $requestPayload,
string $errorMessage,
int $httpStatus,
?string $ipAddress = null,
?string $userAgent = null
): string {
$requestId = $this->generateRequestId();
LogLlmRequest::dispatch(
userId: $userId,
provider: $provider,
model: $model,
requestPayload: $requestPayload,
responsePayload: null,
promptTokens: 0,
completionTokens: 0,
totalTokens: 0,
responseTimeMs: null,
promptCost: 0.0,
completionCost: 0.0,
totalCost: 0.0,
status: 'failed',
errorMessage: $errorMessage,
httpStatus: $httpStatus,
ipAddress: $ipAddress,
userAgent: $userAgent,
requestId: $requestId
);
return $requestId;
}
/**
* Generate unique request ID
*/
private function generateRequestId(): string
{
return 'req_' . Str::random(24);
}
}