Decouple title from title block (#696)
* Decouple title from title block * fix lint * remove dry run for FormTitleMigration * Skip form title migration for forms with hidden titles * Refactor AI Form Generation with Dedicated Prompt Services - Extract AI form generation logic from GenerateTemplate command into dedicated prompt service classes - Update GenerateAiForm job to use new prompt generation services - Improve GptCompleter with more robust error handling and token tracking - Add error field to AiFormCompletion model for better error logging - Simplify command signature from 'ai:make-form-template' to 'form:generate' * Consolidate Template Metadata Generation with Unified Prompt Service - Create GenerateTemplateMetadataPrompt to centralize template metadata generation - Update GenerateTemplate command to use new consolidated metadata generation approach - Enhance GptCompleter to support strict JSON schema validation - Increase form prompt max length to support more complex form descriptions - Refactor form generation to simplify metadata extraction and processing * Implement Form Mode Strategy for Flexible Form Rendering - Introduce FormModeStrategy to centralize form rendering logic - Add support for different form modes: LIVE, PREVIEW, TEST, EDIT, PREFILL - Refactor components to use mode-based rendering strategy - Remove legacy boolean props like adminPreview and creating - Enhance form component flexibility and reusability * Refine Form Mode Strategy Display Behavior - Update FormModeStrategy to hide hidden fields in PREVIEW mode - Add FormMode getter in UrlFormPrefill component for mode-specific rendering - Clarify mode-specific validation and display logic in form strategies * Enhance Form Template Generation with Advanced Field Options - Update GenerateTemplate command to use front_url for template URL output - Expand GenerateFormPrompt with comprehensive field configuration options - Add support for advanced field types: date with time, toggle switches, radio/checkbox selections - Introduce field width configuration and HTML formatting for text elements - Re-enable select, multi-select, and matrix field type definitions with enhanced configurations * Remove Deprecated Template Metadata Generation Services - Delete multiple AI prompt services related to template metadata generation - Simplify GenerateTemplate command to use default values instead of complex metadata generation - Remove GenerateTemplateMetadataPrompt and related classes like GenerateTemplateDescriptionPrompt, GenerateTemplateImageKeywordsPrompt, etc. - Update form template generation to use basic fallback metadata generation approach * Restore GenerateTemplateMetadataPrompt for Comprehensive Template Generation - Reintroduce GenerateTemplateMetadataPrompt to replace default metadata generation - Update GenerateTemplate command to use consolidated metadata generation approach - Extract detailed metadata components including title, description, industries, and image search query - Improve template generation with more dynamic and AI-generated metadata * Refactor Template Preview Section Layout - Remove unnecessary nested div in template preview section - Simplify HTML structure for the template preview component - Maintain existing styling and functionality while improving code readability * Refactor Constructor and Code Formatting in AI Form Generation and Prompt Classes - Updated the constructor in GenerateAiForm.php to use a block structure for improved readability and consistency. - Added a blank line in the Prompt.php file to enhance code formatting and maintain consistency with PHP coding standards. - Modified the migration file to use a more concise class declaration syntax, improving clarity. These changes aim to enhance code readability and maintainability across the affected files. --------- Co-authored-by: Julien Nahum <julien@nahum.net>
This commit is contained in:
@@ -26,13 +26,14 @@ class GptCompleter
|
||||
|
||||
protected bool $expectsJson = false;
|
||||
|
||||
protected int $tokenUsed = 0;
|
||||
protected int $inputTokens = 0;
|
||||
protected int $outputTokens = 0;
|
||||
|
||||
protected bool $useStreaming = false;
|
||||
|
||||
public function __construct(string $apiKey, protected int $retries = 2, protected string $model = self::AI_MODEL)
|
||||
public function __construct(?string $apiKey = null, protected int $retries = 2, protected string $model = self::AI_MODEL)
|
||||
{
|
||||
$this->openAi = \OpenAI::client($apiKey);
|
||||
$this->openAi = \OpenAI::client($apiKey ?? config('services.openai.api_key'));
|
||||
}
|
||||
|
||||
public function setAiModel(string $model)
|
||||
@@ -70,6 +71,20 @@ class GptCompleter
|
||||
return $this;
|
||||
}
|
||||
|
||||
public function setJsonSchema(array $schema): self
|
||||
{
|
||||
$this->completionInput['response_format'] = [
|
||||
'type' => 'json_schema',
|
||||
'json_schema' => [
|
||||
'name' => 'response_schema',
|
||||
'strict' => true,
|
||||
'schema' => $schema
|
||||
]
|
||||
];
|
||||
|
||||
return $this;
|
||||
}
|
||||
|
||||
public function completeChat(array $messages, int $maxTokens = 4096, float $temperature = 0.81, ?bool $exceptJson = null): self
|
||||
{
|
||||
if (! is_null($exceptJson)) {
|
||||
@@ -136,9 +151,14 @@ class GptCompleter
|
||||
return trim($this->result);
|
||||
}
|
||||
|
||||
public function getTokenUsed(): int
|
||||
public function getInputTokens(): int
|
||||
{
|
||||
return $this->tokenUsed;
|
||||
return $this->inputTokens;
|
||||
}
|
||||
|
||||
public function getOutputTokens(): int
|
||||
{
|
||||
return $this->outputTokens;
|
||||
}
|
||||
|
||||
protected function computeChatCompletion(array $messages, int $maxTokens = 4096, float $temperature = 0.81): self
|
||||
@@ -157,10 +177,12 @@ class GptCompleter
|
||||
'temperature' => $temperature,
|
||||
];
|
||||
|
||||
if ($this->expectsJson) {
|
||||
if ($this->expectsJson && !isset($this->completionInput['response_format'])) {
|
||||
$completionInput['response_format'] = [
|
||||
'type' => 'json_object',
|
||||
];
|
||||
} elseif (isset($this->completionInput['response_format'])) {
|
||||
$completionInput['response_format'] = $this->completionInput['response_format'];
|
||||
}
|
||||
|
||||
$this->completionInput = $completionInput;
|
||||
@@ -174,23 +196,30 @@ class GptCompleter
|
||||
return $this->queryStreamedCompletion();
|
||||
}
|
||||
|
||||
try {
|
||||
Log::debug('Open AI query: '.json_encode($this->completionInput));
|
||||
$response = $this->openAi->chat()->create($this->completionInput);
|
||||
} catch (ErrorException $errorException) {
|
||||
// Retry once
|
||||
Log::warning("Open AI error, retrying: {$errorException->getMessage()}");
|
||||
$response = $this->openAi->chat()->create($this->completionInput);
|
||||
}
|
||||
$this->tokenUsed += $response->usage->totalTokens;
|
||||
$this->result = $response->choices[0]->message->content;
|
||||
$attempt = 1;
|
||||
$lastError = null;
|
||||
|
||||
return $this;
|
||||
while ($attempt <= $this->retries) {
|
||||
try {
|
||||
Log::debug('Open AI query: ' . json_encode($this->completionInput));
|
||||
$response = $this->openAi->chat()->create($this->completionInput);
|
||||
$this->inputTokens = $response->usage->promptTokens;
|
||||
$this->outputTokens = $response->usage->completionTokens;
|
||||
$this->result = $response->choices[0]->message->content;
|
||||
return $this;
|
||||
} catch (ErrorException $errorException) {
|
||||
$lastError = $errorException;
|
||||
Log::warning("Open AI error, retrying: {$errorException->getMessage()}");
|
||||
$attempt++;
|
||||
}
|
||||
}
|
||||
|
||||
throw $lastError ?? new \Exception('Failed to complete OpenAI request after multiple attempts');
|
||||
}
|
||||
|
||||
protected function queryStreamedCompletion(): self
|
||||
{
|
||||
Log::debug('Open AI query: '.json_encode($this->completionInput));
|
||||
Log::debug('Open AI query: ' . json_encode($this->completionInput));
|
||||
$this->result = '';
|
||||
$response = $this->openAi->chat()->createStreamed($this->completionInput);
|
||||
foreach ($response as $chunk) {
|
||||
|
||||
Reference in New Issue
Block a user