Skip to content
Draft
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
11 changes: 7 additions & 4 deletions apps/landing/src/components/Header.svelte
Original file line number Diff line number Diff line change
Expand Up @@ -9,7 +9,9 @@

import { SITE_URLS } from "@bashbuddy/consts";

const BANNER_STORAGE_KEY = "bashbuddy-qwen-banner-closed";
// const BANNER_STORAGE_KEY = "bashbuddy-qwen-banner-closed";
const BANNER_STORAGE_KEY =
"bashbuddy-banner-better-cloud-models-and-upcoming-agent-mode";

let isMenuOpen = $state(false);
let isBannerVisible = $state(false);
Expand Down Expand Up @@ -99,16 +101,17 @@
{#if isBrowserEnv && isBannerVisible}
<div
bind:this={bannerElement}
class="fixed top-0 left-0 right-0 z-50 flex items-center justify-center bg-primary w-full h-min py-2 px-2 text-center"
class="fixed top-0 left-0 right-0 z-50 flex items-center justify-center bg-gradient-to-r from-orange-500 to-red-500 w-full h-min py-2 px-2 text-center"
transition:slide={{ duration: 200, axis: "y" }}
>
<p>
<span class="font-medium">
We've added Qwen 2.5 models to BashBuddy Local!
We're introducing better AI models on BashBuddy Cloud and upcoming agent
mode.
</span>

<a
href="/blog/qwen-2.5-models-arrive-to-bashbuddy"
href="/blog/better-cloud-models-and-upcoming-agent-mode"
class="underline font-bold">Read more ➡️</a
>
</p>
Expand Down
4 changes: 2 additions & 2 deletions packages/api/src/router/chat.ts
Original file line number Diff line number Diff line change
Expand Up @@ -76,7 +76,7 @@ class GroqLLM implements LLM {

const generation = trace?.generation({
name: "chat.ask",
model: "llama-3.1-8b-instant",
model: "llama-3.3-70b-versatile",
input: messages,
modelParameters: {
temperature,
Expand All @@ -87,7 +87,7 @@ class GroqLLM implements LLM {

const chatCompletion = await this.groq.chat.completions.create({
messages,
model: "llama-3.1-8b-instant",
model: "llama-3.3-70b-versatile",
temperature,
max_completion_tokens: maxCompletionTokens,
top_p: topP,
Expand Down