Commit graph

  • 9b8a6077bf Switched bytes to tokens to be consistent with output. Allowed much larger outputs and inputs to support long context models. main Pat Wendorf 2025-07-02 12:50:24 -04:00
  • 6e2fa21b62 Removed temp and top_p, that should be controlled by the hosted model now. Pat Wendorf 2025-05-13 08:38:31 -04:00
  • aec8919bf6 fix: join paths when deleting files in clean() Pat Wendorf (aider) 2025-05-09 11:35:35 -04:00
  • a2a3f900d2 fix: join output dir path when opening transcript files Pat Wendorf (aider) 2025-05-09 11:34:23 -04:00
  • eca331e773 fix: use full path when checking for existing transcripts Pat Wendorf (aider) 2025-05-09 11:31:16 -04:00
  • d9de639553 fix: use full paths for wav file operations Pat Wendorf (aider) 2025-05-09 11:26:45 -04:00
  • 23d2e3715c Line spacing in config dialog was too much Pat Wendorf 2025-05-09 10:56:04 -04:00
  • 60215cc0c4 feat: add configurable output directory for transcripts Pat Wendorf (aider) 2025-05-09 10:51:15 -04:00
  • b90b651f14 Forgot the v1 endpoint for llama.cpp/ollama Pat Wendorf 2025-05-09 10:49:11 -04:00
  • b754738146 feat: replace input fields with text areas in config dialog Pat Wendorf (aider) 2025-05-09 10:48:22 -04:00
  • 73279657da feat: increase settings dialog size and field heights Pat Wendorf (aider) 2025-05-09 10:46:00 -04:00
  • f1fb9f0a2b Removed the old sample.env file and updated readme. Pat Wendorf 2025-05-09 10:44:52 -04:00
  • 80101d577c More refined prompts Pat Wendorf 2025-05-09 10:41:49 -04:00
  • 150d910ddb more better defaults Pat Wendorf 2025-05-09 10:15:47 -04:00
  • 40de47afe2 More reasonable defaults from the working system Pat Wendorf 2025-05-09 10:07:29 -04:00
  • 08b2a2b627 feat: add config dialog and move settings from .env to config.json Pat Wendorf (aider) 2025-05-09 09:55:03 -04:00
  • b9963599e3 docs: add macOS homebrew requirements to README Pat Wendorf (aider) 2025-05-09 09:47:16 -04:00
  • 2fe4893296 After merging summarize into meetings removed meetings.py Pat Wendorf 2025-05-09 09:37:26 -04:00
  • 04659ed489 refactor: merge summarize.py functionality into meetings.py Pat Wendorf (aider) 2025-05-09 09:30:36 -04:00
  • b9533bf503 Sample env was missing the fact prompt Pat Wendorf 2025-02-20 19:18:15 -05:00
  • 70a8816e8b Forgot to update sample file. Pat Wendorf 2025-02-08 08:54:58 -05:00
  • c947e08cdc Allow for adjustment of temp, top_p and max tokens for models that have issues with repeating. Pat Wendorf 2025-02-08 08:54:29 -05:00
  • 933b9e24d4 Skip transcribe step if it's already done. Show each part being summarized. Pat Wendorf 2025-02-03 16:30:52 -05:00
  • 2d329d4806 Updated instructions for new .env settings Pat Wendorf 2024-11-29 09:08:24 -05:00
  • 80637cc26c Simplified the transcriber by using the openai library to access llama.cpp server or ollama server mode. Pat Wendorf 2024-11-29 09:06:22 -05:00
  • 77b7f440a0 Replaced the screenshot to show the new recording device selector Pat Wendorf 2024-08-19 08:33:29 -04:00
  • 6128f02615 Added drop down for selecting the recording device. Pat Wendorf 2024-08-19 08:32:24 -04:00
  • ac27eaf826 Treat the summarizer as a module now, and call it with threading instead of subprocess Pat Wendorf 2024-08-19 07:55:46 -04:00
  • 3c05cd78dd Used mistral large to re-write the summarize process. Pat Wendorf 2024-08-19 07:41:32 -04:00
  • 81e4e165bc Added some silence trimming Pat Wendorf 2024-07-30 11:56:40 -04:00
  • f5af4960fc Extract some fact based summaries now as well Pat Wendorf 2024-06-24 09:52:38 -04:00
  • cb00762378 UI looks a bit better now Pat Wendorf 2024-06-24 09:38:54 -04:00
  • fe50d40fce Mention you need llama.cpp and whisper.cpp running in server mode Pat Wendorf 2024-06-21 12:56:41 -04:00
  • f8c7d09a59 Accidentally removed MD files Pat Wendorf 2024-06-21 12:53:08 -04:00
  • 69d269d2f0 first commit Pat Wendorf 2024-06-21 12:52:10 -04:00