botserver/src/designer/designer_api/llm_integration.rs
Rodrigo Rodriguez (Pragmatismo) 7d8f141fc2
Some checks failed
BotServer CI/CD / build (push) Failing after 1m28s
refactor: Replace all hardcoded ./botserver-stack paths with get_stack_path()/get_work_path()
- Adds get_stack_path() helper: returns /opt/gbo in production (.env without botserver-stack), ./botserver-stack in dev
- Adds get_work_path() helper: returns /opt/gbo/work in production, ./botserver-stack/data/system/work in dev
- Updated 35+ files to use dynamic path resolution
- Production system container no longer needs botserver-stack directory
- Work files go to /opt/gbo/work instead of /opt/gbo/bin/botserver-stack
2026-04-04 09:24:44 -03:00

497 lines
17 KiB
Rust

use super::types::*;
use crate::auto_task::get_designer_error_context;
use crate::core::shared::state::AppState;
use crate::core::shared::get_content_type;
use axum::{extract::State, response::IntoResponse, Json};
use std::fmt::Write;
use std::sync::Arc;
use uuid::Uuid;
pub async fn handle_editor_magic(
State(state): State<Arc<AppState>>,
Json(request): Json<EditorMagicRequest>,
) -> impl IntoResponse {
let code = request.code;
if code.trim().is_empty() {
return Json(EditorMagicResponse {
improved_code: None,
explanation: Some("No code provided".to_string()),
suggestions: None,
});
}
let prompt = format!(
r#"You are reviewing this HTMX application code. Analyze and improve it.
Focus on:
- Better HTMX patterns (reduce JS, use hx-* attributes properly)
- Accessibility (ARIA labels, keyboard navigation, semantic HTML)
- Performance (lazy loading, efficient selectors)
- UX (loading states, error handling, user feedback)
- Code organization (clean structure, no comments needed)
Current code:
```
{code}
```
Respond with JSON only:
{{
"improved_code": "the improved code here",
"explanation": "brief explanation of changes made"
}}
If the code is already good, respond with:
{{
"improved_code": null,
"explanation": "Code looks good, no improvements needed"
}}"#
);
#[cfg(feature = "llm")]
{
let config = serde_json::json!({
"temperature": 0.3,
"max_tokens": 4000
});
match state
.llm_provider
.generate(&prompt, &config, "gpt-4", "")
.await
{
Ok(response) => {
if let Ok(result) = serde_json::from_str::<EditorMagicResponse>(&response) {
return Json(result);
}
return Json(EditorMagicResponse {
improved_code: Some(response),
explanation: Some("AI suggestions".to_string()),
suggestions: None,
});
}
Err(e) => {
log::warn!("LLM call failed: {e}");
}
}
}
let _ = state;
let mut suggestions = Vec::new();
if !code.contains("hx-") {
suggestions.push(MagicSuggestion {
suggestion_type: "ux".to_string(),
title: "Use HTMX attributes".to_string(),
description: "Consider using hx-get, hx-post instead of JavaScript fetch calls."
.to_string(),
});
}
if !code.contains("hx-indicator") {
suggestions.push(MagicSuggestion {
suggestion_type: "ux".to_string(),
title: "Add loading indicators".to_string(),
description: "Use hx-indicator to show loading state during requests.".to_string(),
});
}
if !code.contains("aria-") && !code.contains("role=") {
suggestions.push(MagicSuggestion {
suggestion_type: "a11y".to_string(),
title: "Improve accessibility".to_string(),
description: "Add ARIA labels and roles for screen reader support.".to_string(),
});
}
if code.contains("onclick=") || code.contains("addEventListener") {
suggestions.push(MagicSuggestion {
suggestion_type: "perf".to_string(),
title: "Replace JS with HTMX".to_string(),
description: "HTMX can handle most interactions without custom JavaScript.".to_string(),
});
}
Json(EditorMagicResponse {
improved_code: None,
explanation: None,
suggestions: if suggestions.is_empty() {
None
} else {
Some(suggestions)
},
})
}
pub async fn handle_designer_modify(
State(state): State<Arc<AppState>>,
Json(request): Json<DesignerModifyRequest>,
) -> impl IntoResponse {
let app = &request.app_name;
let msg_preview = &request.message[..request.message.len().min(100)];
log::info!("Designer modify request for app '{app}': {msg_preview}");
let session = match get_designer_session(&state) {
Ok(s) => s,
Err(e) => {
return (
axum::http::StatusCode::UNAUTHORIZED,
Json(DesignerModifyResponse {
success: false,
message: "Authentication required".to_string(),
changes: Vec::new(),
suggestions: Vec::new(),
error: Some(e.to_string()),
}),
);
}
};
match process_designer_modification(&state, &request, &session).await {
Ok(response) => (axum::http::StatusCode::OK, Json(response)),
Err(e) => {
log::error!("Designer modification failed: {e}");
(
axum::http::StatusCode::INTERNAL_SERVER_ERROR,
Json(DesignerModifyResponse {
success: false,
message: "Failed to process modification".to_string(),
changes: Vec::new(),
suggestions: Vec::new(),
error: Some(e.to_string()),
}),
)
}
}
}
pub fn get_designer_session(
state: &AppState,
) -> Result<crate::core::shared::models::UserSession, Box<dyn std::error::Error + Send + Sync>> {
use crate::core::shared::models::schema::bots::dsl::*;
use crate::core::shared::models::UserSession;
use diesel::prelude::*;
let mut conn = state.conn.get()?;
let bot_result: Result<(Uuid, String), _> = bots.select((id, name)).first(&mut conn);
match bot_result {
Ok((bot_id_val, _bot_name_val)) => Ok(UserSession {
id: Uuid::new_v4(),
user_id: Uuid::nil(),
bot_id: bot_id_val,
title: "designer".to_string(),
context_data: serde_json::json!({}),
current_tool: None,
created_at: chrono::Utc::now(),
updated_at: chrono::Utc::now(),
}),
Err(_) => Err("No bot found for designer session".into()),
}
}
async fn process_designer_modification(
state: &AppState,
request: &DesignerModifyRequest,
session: &crate::core::shared::models::UserSession,
) -> Result<DesignerModifyResponse, Box<dyn std::error::Error + Send + Sync>> {
let prompt = build_designer_prompt(request);
let llm_response = call_designer_llm(state, &prompt).await?;
let (changes, message, suggestions) =
parse_and_apply_changes(state, request, &llm_response, session).await?;
Ok(DesignerModifyResponse {
success: true,
message,
changes,
suggestions,
error: None,
})
}
fn build_designer_prompt(request: &DesignerModifyRequest) -> String {
let context_info = request
.context
.as_ref()
.map(|ctx| {
let mut info = String::new();
if let Some(ref html) = ctx.page_html {
let _ = writeln!(
info,
"\nCurrent page HTML (first 500 chars):\n{}",
&html[..html.len().min(500)]
);
}
if let Some(ref tables) = ctx.tables {
let _ = writeln!(info, "\nAvailable tables: {}", tables.join(", "));
}
info
})
.unwrap_or_default();
let error_context = get_designer_error_context(&request.app_name).unwrap_or_default();
format!(
r#"You are a Designer AI assistant helping modify an HTMX-based application.
App Name: {}
Current Page: {}
{}
{}
User Request: "{}"
Analyze the request and respond with JSON describing the changes needed:
{{
"understanding": "brief description of what user wants",
"changes": [
{{
"type": "modify_html|add_field|remove_field|add_table|modify_style|add_page",
"file": "filename.html or styles.css",
"description": "what this change does",
"code": "the new/modified code snippet"
}}
],
"message": "friendly response to user explaining what was done",
"suggestions": ["optional follow-up suggestions"]
}}
Guidelines:
- Use HTMX attributes (hx-get, hx-post, hx-target, hx-swap, hx-trigger)
- Keep styling minimal and consistent
- API endpoints follow pattern: /api/db/{{table_name}}
- Forms should use hx-post for submissions
- Lists should use hx-get with pagination
- IMPORTANT: Use RELATIVE paths for app assets (styles.css, app.js, NOT /static/styles.css)
- For HTMX, use LOCAL: <script src="/js/vendor/htmx.min.js"></script> (NO external CDN)
- CSS link should be: <link rel="stylesheet" href="styles.css">
Respond with valid JSON only."#,
request.app_name,
request.current_page.as_deref().unwrap_or("index.html"),
context_info,
error_context,
request.message
)
}
async fn call_designer_llm(
state: &AppState,
prompt: &str,
) -> Result<String, Box<dyn std::error::Error + Send + Sync>> {
use crate::core::config::ConfigManager;
let config_manager = ConfigManager::new(state.conn.clone());
// Get LLM configuration from bot config or use defaults
let model = config_manager
.get_config(&uuid::Uuid::nil(), "llm-model", Some("claude-sonnet-4-20250514"))
.unwrap_or_else(|_| "claude-sonnet-4-20250514".to_string());
let api_key = config_manager
.get_config(&uuid::Uuid::nil(), "llm-key", None)
.unwrap_or_default();
#[cfg(feature = "llm")]
let response_text = {
let system_prompt = "You are a web designer AI. Respond only with valid JSON.";
let messages = serde_json::json!({
"messages": [
{"role": "system", "content": system_prompt},
{"role": "user", "content": prompt}
]
});
state.llm_provider.generate(prompt, &messages, &model, &api_key).await?
};
#[cfg(not(feature = "llm"))]
let response_text = String::from("{}"); // Fallback or handling for when LLM is missing
let json_text = if response_text.contains("```json") {
response_text
.split("```json")
.nth(1)
.and_then(|s| s.split("```").next())
.unwrap_or(&response_text)
.trim()
.to_string()
} else if response_text.contains("```") {
response_text
.split("```")
.nth(1)
.unwrap_or(&response_text)
.trim()
.to_string()
} else {
response_text
};
Ok(json_text)
}
async fn parse_and_apply_changes(
state: &AppState,
request: &DesignerModifyRequest,
llm_response: &str,
session: &crate::core::shared::models::UserSession,
) -> Result<(Vec<DesignerChange>, String, Vec<String>), Box<dyn std::error::Error + Send + Sync>> {
#[derive(serde::Deserialize)]
struct LlmChangeResponse {
_understanding: Option<String>,
changes: Option<Vec<LlmChange>>,
message: Option<String>,
suggestions: Option<Vec<String>>,
}
#[derive(serde::Deserialize)]
struct LlmChange {
#[serde(rename = "type")]
change_type: String,
file: String,
description: String,
code: Option<String>,
}
let parsed: LlmChangeResponse = serde_json::from_str(llm_response).unwrap_or_else(|_| LlmChangeResponse {
_understanding: Some("Could not parse LLM response".to_string()),
changes: None,
message: Some("I understood your request but encountered an issue processing it. Could you try rephrasing?".to_string()),
suggestions: Some(vec!["Try being more specific".to_string()]),
});
let mut applied_changes = Vec::new();
if let Some(changes) = parsed.changes {
for change in changes {
if let Some(ref code) = change.code {
match apply_file_change(state, &request.app_name, &change.file, code, session).await
{
Ok(()) => {
applied_changes.push(DesignerChange {
change_type: change.change_type,
file_path: change.file,
description: change.description,
preview: Some(code[..code.len().min(200)].to_string()),
});
}
Err(e) => {
let file = &change.file;
log::warn!("Failed to apply change to {file}: {e}");
}
}
}
}
}
let message = parsed.message.unwrap_or_else(|| {
if applied_changes.is_empty() {
"I couldn't make any changes. Could you provide more details?".to_string()
} else {
format!(
"Done! I made {} change(s) to your app.",
applied_changes.len()
)
}
});
let suggestions = parsed.suggestions.unwrap_or_default();
Ok((applied_changes, message, suggestions))
}
pub async fn apply_file_change(
state: &AppState,
app_name: &str,
file_name: &str,
content: &str,
_session: &crate::core::shared::models::UserSession,
) -> Result<(), Box<dyn std::error::Error + Send + Sync>> {
// Use bucket_name from state (like app_generator) - e.g., "default.gbai"
let bucket_name = state.bucket_name.clone();
let sanitized_name = bucket_name.trim_end_matches(".gbai").to_string();
// Always write to local disk first (primary storage, like import templates)
// Match app_server filesystem fallback path: {site_path}/{bot}.gbai/{bot}.gbapp/{app_name}/{file}
let site_path = state
.config
.as_ref()
.map(|c| c.site_path.clone())
.unwrap_or_else(|| format!("{}/sites", crate::core::shared::utils::get_stack_path()));
let local_path = format!("{site_path}/{}.gbai/{}.gbapp/{app_name}/{file_name}", sanitized_name, sanitized_name);
if let Some(parent) = std::path::Path::new(&local_path).parent() {
std::fs::create_dir_all(parent)?;
}
std::fs::write(&local_path, content)?;
log::info!("Designer updated local file: {local_path}");
// Also sync to S3/MinIO if available (with bucket creation retry like app_generator)
if let Some(ref s3_client) = state.drive {
use aws_sdk_s3::primitives::ByteStream;
// Use same path pattern as app_server/app_generator: {sanitized_name}.gbapp/{app_name}/{file}
let file_path = format!("{}.gbapp/{}/{}", sanitized_name, app_name, file_name);
log::info!("Designer syncing to S3: bucket={}, key={}", bucket_name, file_path);
match s3_client
.put_object()
.bucket(&bucket_name)
.key(&file_path)
.body(ByteStream::from(content.as_bytes().to_vec()))
.content_type(get_content_type(file_name))
.send()
.await
{
Ok(_) => {
log::info!("Designer synced to S3: s3://{bucket_name}/{file_path}");
}
Err(e) => {
// Check if bucket doesn't exist and try to create it (like app_generator)
let err_str = format!("{:?}", e);
if err_str.contains("NoSuchBucket") || err_str.contains("NotFound") {
log::warn!("Bucket {} not found, attempting to create...", bucket_name);
// Try to create the bucket
match s3_client.create_bucket().bucket(&bucket_name).send().await {
Ok(_) => {
log::info!("Created bucket: {}", bucket_name);
}
Err(create_err) => {
let create_err_str = format!("{:?}", create_err);
// Ignore if bucket already exists (race condition)
if !create_err_str.contains("BucketAlreadyExists")
&& !create_err_str.contains("BucketAlreadyOwnedByYou") {
log::warn!("Failed to create bucket {}: {}", bucket_name, create_err);
}
}
}
// Retry the write after bucket creation
match s3_client
.put_object()
.bucket(&bucket_name)
.key(&file_path)
.body(ByteStream::from(content.as_bytes().to_vec()))
.content_type(get_content_type(file_name))
.send()
.await
{
Ok(_) => {
log::info!("Designer synced to S3 after bucket creation: s3://{bucket_name}/{file_path}");
}
Err(retry_err) => {
log::warn!("Designer S3 retry failed (local write succeeded): {retry_err}");
}
}
} else {
// S3 sync is optional - local write already succeeded
log::warn!("Designer S3 sync failed (local write succeeded): {e}");
}
}
}
}
Ok(())
}