Skip to content

Commit

Permalink
New structure
Browse files Browse the repository at this point in the history
  • Loading branch information
ParisNeo committed Aug 29, 2024
1 parent 32fa395 commit 9d19668
Show file tree
Hide file tree
Showing 12 changed files with 62 additions and 60 deletions.
7 changes: 6 additions & 1 deletion configs/config.yaml
Original file line number Diff line number Diff line change
@@ -1,5 +1,5 @@
# =================== Lord Of Large Language Multimodal Systems Configuration file ===========================
version: 133
version: 134
binding_name: null
model_name: null
model_variant: null
Expand Down Expand Up @@ -105,6 +105,7 @@ active_tts_service: "None" # xtts (offline), openai_tts (API key required), elev
active_tti_service: "None" # autosd (offline), dall-e (online)
active_stt_service: "None" # whisper (offline), asr (offline or online), openai_whiosper (API key required)
active_ttm_service: "None" # musicgen (offline)
active_ttv_service: "None" # cog_video_x (offline)
# -------------------- Services --------------------------

# ***************** STT *****************
Expand Down Expand Up @@ -204,6 +205,10 @@ comfyui_model: v1-5-pruned-emaonly.ckpt
enable_motion_ctrl_service: false
motion_ctrl_base_url: http://localhost:7861


# ***************** TTV *****************
cog_video_x_model: "THUDM/CogVideoX-5b"

# ***************** TTT *****************

# ollama service
Expand Down
68 changes: 30 additions & 38 deletions lollms/app.py
Original file line number Diff line number Diff line change
Expand Up @@ -339,7 +339,7 @@ def start_servers(self):
def start_ttt(*args, **kwargs):
if self.config.enable_ollama_service:
try:
from lollms.services.ollama.lollms_ollama import Service
from lollms.services.ttt.ollama.lollms_ollama import Service
self.ollama = Service(self, base_url=self.config.ollama_base_url)
tts_services.append("ollama")

Expand All @@ -349,7 +349,7 @@ def start_ttt(*args, **kwargs):

if self.config.enable_vllm_service:
try:
from lollms.services.vllm.lollms_vllm import Service
from lollms.services.ttt.vllm.lollms_vllm import Service
self.vllm = Service(self, base_url=self.config.vllm_url)
tts_services.append("vllm")
except Exception as ex:
Expand All @@ -360,16 +360,16 @@ def start_ttt(*args, **kwargs):
def start_stt(*args, **kwargs):
if self.config.whisper_activate or self.config.active_stt_service == "whisper":
try:
from lollms.services.whisper.lollms_whisper import LollmsWhisper
from lollms.services.stt.whisper.lollms_whisper import LollmsWhisper
self.whisper = LollmsWhisper(self, self.config.whisper_model, self.lollms_paths.personal_outputs_path)
stt_services.append("whisper")
except Exception as ex:
trace_exception(ex)
if self.config.active_stt_service == "openai_whisper":
from lollms.services.openai_whisper.lollms_openai_whisper import LollmsOpenAIWhisper
from lollms.services.stt.openai_whisper.lollms_openai_whisper import LollmsOpenAIWhisper
self.stt = LollmsOpenAIWhisper(self, self.config.openai_whisper_model, self.config.openai_whisper_key)
elif self.config.active_stt_service == "whisper":
from lollms.services.whisper.lollms_whisper import LollmsWhisper
from lollms.services.stt.whisper.lollms_whisper import LollmsWhisper
self.stt = LollmsWhisper(self, self.config.whisper_model)

ASCIIColors.execute_with_animation("Loading loacal STT services", start_stt, ASCIIColors.color_blue)
Expand All @@ -379,7 +379,7 @@ def start_tts(*args, **kwargs):
if self.config.active_tts_service == "xtts":
ASCIIColors.yellow("Loading XTTS")
try:
from lollms.services.xtts.lollms_xtts import LollmsXTTS
from lollms.services.tts.xtts.lollms_xtts import LollmsXTTS
voice=self.config.xtts_current_voice
if voice!="main_voice":
voices_folder = self.lollms_paths.custom_voices_path
Expand All @@ -395,10 +395,10 @@ def start_tts(*args, **kwargs):
trace_exception(ex)
self.warning(f"Couldn't load XTTS")
if self.config.active_tts_service == "eleven_labs_tts":
from lollms.services.eleven_labs_tts.lollms_eleven_labs_tts import LollmsElevenLabsTTS
from lollms.services.tts.eleven_labs_tts.lollms_eleven_labs_tts import LollmsElevenLabsTTS
self.tts = LollmsElevenLabsTTS(self, self.config.elevenlabs_tts_model_id, self.config.elevenlabs_tts_voice_id, self.config.elevenlabs_tts_key, stability=self.config.elevenlabs_tts_voice_stability, similarity_boost=self.config.elevenlabs_tts_voice_boost)
elif self.config.active_tts_service == "openai_tts":
from lollms.services.open_ai_tts.lollms_openai_tts import LollmsOpenAITTS
from lollms.services.tts.open_ai_tts.lollms_openai_tts import LollmsOpenAITTS
self.tts = LollmsOpenAITTS(self, self.config.openai_tts_model, self.config.openai_tts_voice, self.config.openai_tts_key)
elif self.config.active_tts_service == "xtts" and self.xtts:
self.tts = self.xtts
Expand All @@ -409,46 +409,38 @@ def start_tts(*args, **kwargs):
def start_tti(*args, **kwargs):
if self.config.enable_sd_service:
try:
from lollms.services.sd.lollms_sd import LollmsSD
from lollms.services.tti.sd.lollms_sd import LollmsSD
self.sd = LollmsSD(self, auto_sd_base_url=self.config.sd_base_url)
except:
self.warning(f"Couldn't load SD")

if self.config.enable_comfyui_service:
try:
from lollms.services.comfyui.lollms_comfyui import LollmsComfyUI
from lollms.services.tti.comfyui.lollms_comfyui import LollmsComfyUI
self.comfyui = LollmsComfyUI(self, comfyui_base_url=self.config.comfyui_base_url)
except:
self.warning(f"Couldn't load SD")

if self.config.enable_motion_ctrl_service:
try:
from lollms.services.motion_ctrl.lollms_motion_ctrl import Service
self.motion_ctrl = Service(self, base_url=self.config.motion_ctrl_base_url)
except Exception as ex:
trace_exception(ex)
self.warning(f"Couldn't load Motion control")

if self.config.active_tti_service == "diffusers":
from lollms.services.diffusers.lollms_diffusers import LollmsDiffusers
from lollms.services.tti.diffusers.lollms_diffusers import LollmsDiffusers
self.tti = LollmsDiffusers(self)
elif self.config.active_tti_service == "autosd":
if self.sd:
self.tti = self.sd
else:
from lollms.services.sd.lollms_sd import LollmsSD
from lollms.services.tti.sd.lollms_sd import LollmsSD
self.tti = LollmsSD(self)
elif self.config.active_tti_service == "dall-e":
from lollms.services.dalle.lollms_dalle import LollmsDalle
from lollms.services.tti.dalle.lollms_dalle import LollmsDalle
self.tti = LollmsDalle(self, self.config.dall_e_key)
elif self.config.active_tti_service == "midjourney":
from lollms.services.midjourney.lollms_midjourney import LollmsMidjourney
from lollms.services.tti.midjourney.lollms_midjourney import LollmsMidjourney
self.tti = LollmsMidjourney(self, self.config.midjourney_key, self.config.midjourney_timeout, self.config.midjourney_retries)
elif self.config.active_tti_service == "comfyui" and (self.tti is None or self.tti.name!="comfyui"):
if self.comfyui:
self.tti = self.comfyui
else:
from lollms.services.comfyui.lollms_comfyui import LollmsComfyUI
from lollms.services.tti.comfyui.lollms_comfyui import LollmsComfyUI
self.tti = LollmsComfyUI(self, comfyui_base_url=self.config.comfyui_base_url)

ASCIIColors.execute_with_animation("Loading loacal TTI services", start_tti, ASCIIColors.color_blue)
Expand All @@ -463,15 +455,15 @@ def verify_servers(self, reload_all=False):

if self.config.enable_ollama_service and self.ollama is None:
try:
from lollms.services.ollama.lollms_ollama import Service
from lollms.services.ttt.ollama.lollms_ollama import Service
self.ollama = Service(self, base_url=self.config.ollama_base_url)
except Exception as ex:
trace_exception(ex)
self.warning(f"Couldn't load Ollama")

if self.config.enable_vllm_service and self.vllm is None:
try:
from lollms.services.vllm.lollms_vllm import Service
from lollms.services.ttt.vllm.lollms_vllm import Service
self.vllm = Service(self, base_url=self.config.vllm_url)
except Exception as ex:
trace_exception(ex)
Expand All @@ -481,7 +473,7 @@ def verify_servers(self, reload_all=False):

if self.config.whisper_activate and self.whisper is None:
try:
from lollms.services.whisper.lollms_whisper import LollmsWhisper
from lollms.services.stt.whisper.lollms_whisper import LollmsWhisper
self.whisper = LollmsWhisper(self, self.config.whisper_model, self.lollms_paths.personal_outputs_path)
except Exception as ex:
trace_exception(ex)
Expand All @@ -490,7 +482,7 @@ def verify_servers(self, reload_all=False):
if self.config.active_tts_service == "xtts" and self.xtts is None:
ASCIIColors.yellow("Loading XTTS")
try:
from lollms.services.xtts.lollms_xtts import LollmsXTTS
from lollms.services.tts.xtts.lollms_xtts import LollmsXTTS
voice=self.config.xtts_current_voice
if voice!="main_voice":
voices_folder = self.lollms_paths.custom_voices_path
Expand All @@ -509,14 +501,14 @@ def verify_servers(self, reload_all=False):
ASCIIColors.blue("Loading local TTI services")
if self.config.enable_sd_service and self.sd is None:
try:
from lollms.services.sd.lollms_sd import LollmsSD
from lollms.services.tti.sd.lollms_sd import LollmsSD
self.sd = LollmsSD(self, auto_sd_base_url=self.config.sd_base_url)
except:
self.warning(f"Couldn't load SD")

if self.config.enable_comfyui_service and self.comfyui is None:
try:
from lollms.services.comfyui.lollms_comfyui import LollmsComfyUI
from lollms.services.tti.comfyui.lollms_comfyui import LollmsComfyUI
self.comfyui = LollmsComfyUI(self, comfyui_base_url=self.config.comfyui_base_url)
except:
self.warning(f"Couldn't load Comfyui")
Expand All @@ -532,43 +524,43 @@ def verify_servers(self, reload_all=False):

ASCIIColors.blue("Activating TTI service")
if self.config.active_tti_service == "diffusers" and (self.tti is None or self.tti.name!="diffusers"):
from lollms.services.diffusers.lollms_diffusers import LollmsDiffusers
from lollms.services.tti.diffusers.lollms_diffusers import LollmsDiffusers
self.tti = LollmsDiffusers(self)
elif self.config.active_tti_service == "autosd" and (self.tti is None or self.tti.name!="stable_diffusion"):
if self.sd:
self.tti = self.sd
else:
from lollms.services.sd.lollms_sd import LollmsSD
from lollms.services.tti.sd.lollms_sd import LollmsSD
self.tti = LollmsSD(self)
elif self.config.active_tti_service == "dall-e" and (self.tti is None or self.tti.name!="dall-e-2" or type(self.tti.name)!="dall-e-3"):
from lollms.services.dalle.lollms_dalle import LollmsDalle
from lollms.services.tti.dalle.lollms_dalle import LollmsDalle
self.tti = LollmsDalle(self, self.config.dall_e_key)
elif self.config.active_tti_service == "midjourney" and (self.tti is None or self.tti.name!="midjourney"):
from lollms.services.midjourney.lollms_midjourney import LollmsMidjourney
from lollms.services.tti.midjourney.lollms_midjourney import LollmsMidjourney
self.tti = LollmsMidjourney(self, self.config.midjourney_key, self.config.midjourney_timeout, self.config.midjourney_retries)
elif self.config.active_tti_service == "comfyui" and (self.tti is None or self.tti.name!="comfyui"):
if self.comfyui:
self.tti = self.comfyui
else:
from lollms.services.comfyui.lollms_comfyui import LollmsComfyUI
from lollms.services.tti.comfyui.lollms_comfyui import LollmsComfyUI
self.tti = LollmsComfyUI(self, comfyui_base_url=self.config.comfyui_base_url)

ASCIIColors.blue("Activating TTS service")
if self.config.active_tts_service == "eleven_labs_tts":
from lollms.services.eleven_labs_tts.lollms_eleven_labs_tts import LollmsElevenLabsTTS
from lollms.services.tts.eleven_labs_tts.lollms_eleven_labs_tts import LollmsElevenLabsTTS
self.tts = LollmsElevenLabsTTS(self, self.config.elevenlabs_tts_model_id, self.config.elevenlabs_tts_voice_id, self.config.elevenlabs_tts_key, stability=self.config.elevenlabs_tts_voice_stability, similarity_boost=self.config.elevenlabs_tts_voice_boost)
elif self.config.active_tts_service == "openai_tts" and (self.tts is None or self.tts.name!="openai_tts"):
from lollms.services.open_ai_tts.lollms_openai_tts import LollmsOpenAITTS
from lollms.services.tts.open_ai_tts.lollms_openai_tts import LollmsOpenAITTS
self.tts = LollmsOpenAITTS(self, self.config.openai_tts_model, self.config.openai_tts_voice, self.config.openai_tts_key)
elif self.config.active_tts_service == "xtts" and self.xtts:
self.tts = self.xtts

ASCIIColors.blue("Activating STT service")
if self.config.active_stt_service == "openai_whisper" and (self.tts is None or self.tts.name!="openai_whisper"):
from lollms.services.openai_whisper.lollms_openai_whisper import LollmsOpenAIWhisper
from lollms.services.stt.openai_whisper.lollms_openai_whisper import LollmsOpenAIWhisper
self.stt = LollmsOpenAIWhisper(self, self.config.openai_whisper_model, self.config.openai_whisper_key)
elif self.config.active_stt_service == "whisper" and (self.tts is None or self.tts.name!="whisper") :
from lollms.services.whisper.lollms_whisper import LollmsWhisper
from lollms.services.stt.whisper.lollms_whisper import LollmsWhisper
self.stt = LollmsWhisper(self, self.config.whisper_model)

except Exception as ex:
Expand Down
7 changes: 6 additions & 1 deletion lollms/configs/config.yaml
Original file line number Diff line number Diff line change
@@ -1,5 +1,5 @@
# =================== Lord Of Large Language Multimodal Systems Configuration file ===========================
version: 133
version: 134
binding_name: null
model_name: null
model_variant: null
Expand Down Expand Up @@ -105,6 +105,7 @@ active_tts_service: "None" # xtts (offline), openai_tts (API key required), elev
active_tti_service: "None" # autosd (offline), dall-e (online)
active_stt_service: "None" # whisper (offline), asr (offline or online), openai_whiosper (API key required)
active_ttm_service: "None" # musicgen (offline)
active_ttv_service: "None" # cog_video_x (offline)
# -------------------- Services --------------------------

# ***************** STT *****************
Expand Down Expand Up @@ -204,6 +205,10 @@ comfyui_model: v1-5-pruned-emaonly.ckpt
enable_motion_ctrl_service: false
motion_ctrl_base_url: http://localhost:7861


# ***************** TTV *****************
cog_video_x_model: "THUDM/CogVideoX-5b"

# ***************** TTT *****************

# ollama service
Expand Down
8 changes: 4 additions & 4 deletions lollms/functions/memes/memes_builder.py
Original file line number Diff line number Diff line change
Expand Up @@ -251,7 +251,7 @@ def build_meme_image_with_text_overlay(prompt, negative_prompt, width, height, t
try:
if processor.personality.config.active_tti_service == "diffusers":
if not processor.personality.app.tti:
from lollms.services.diffusers.lollms_diffusers import LollmsDiffusers
from lollms.services.tti.diffusers.lollms_diffusers import LollmsDiffusers
processor.step_start("Loading ParisNeo's fork of AUTOMATIC1111's stable diffusion service")
processor.personality.app.tti = LollmsDiffusers(processor.personality.app, processor.personality.name)
processor.personality.app.sd = processor.personality.app.tti
Expand All @@ -265,7 +265,7 @@ def build_meme_image_with_text_overlay(prompt, negative_prompt, width, height, t
)
elif processor.personality.config.active_tti_service == "autosd":
if not processor.personality.app.tti:
from lollms.services.sd.lollms_sd import LollmsSD
from lollms.services.tti.sd.lollms_sd import LollmsSD
processor.step_start("Loading ParisNeo's fork of AUTOMATIC1111's stable diffusion service")
processor.personality.app.tti = LollmsSD(processor.personality.app, processor.personality.name, max_retries=-1, auto_sd_base_url=processor.personality.config.sd_base_url)
processor.personality.app.sd = processor.personality.app.tti
Expand All @@ -279,7 +279,7 @@ def build_meme_image_with_text_overlay(prompt, negative_prompt, width, height, t
)
elif processor.personality.config.active_tti_service == "dall-e":
if not processor.personality.app.tti:
from lollms.services.dalle.lollms_dalle import LollmsDalle
from lollms.services.tti.dalle.lollms_dalle import LollmsDalle
processor.step_start("Loading dalle service")
processor.personality.app.tti = LollmsDalle(processor.personality.app, processor.personality.config.dall_e_key, processor.personality.config.dall_e_generation_engine)
processor.personality.app.dalle = processor.personality.app.tti
Expand All @@ -295,7 +295,7 @@ def build_meme_image_with_text_overlay(prompt, negative_prompt, width, height, t
processor.step_end("Painting")
elif processor.personality.config.active_tti_service == "comfyui":
if not processor.personality.app.tti:
from lollms.services.comfyui.lollms_comfyui import LollmsComfyUI
from lollms.services.tti.comfyui.lollms_comfyui import LollmsComfyUI
processor.step_start("Loading comfyui service")
processor.personality.app.tti = LollmsComfyUI(
processor.personality.app,
Expand Down
Loading

0 comments on commit 9d19668

Please sign in to comment.