Skip to content

Improve observability (feedback from a new user) #6052

@var1ableX

Description

@var1ableX

Please explain the motivation behind the feature request.
I think goose is a great idea, but after a few days of messing around with it I think it's expensive and I think that's not great news for you. Helping me dispel this belief cheaply might be beneficial. Right now I think you're not prioritizing observability, but lets face it. Novices would not be able to use goose right now. It's too early and too fraught with weird problems that a novice wouldn't be able to figure out. So right now your market are power users and developers. I fall into the latter two brackets. What I struggle with is "is this actually working syndrome"? Is there something i could do differently that would make goose viable vs. cursor. But it's quite difficult to figure out whether any of my changes are doing anything. And at this point the only thing keeping me hanging on are my 3 days investment and excitement at a reputable brand creating a LLM provider agnostic solution to building your on PAI (a la Daniel Miessler).

Expensive because of context poisoning - it's hard to cut down input tokens when you can't easily see when and where they're being injected. It's like staring at a wall of logged jsonl OR hard to understand langfuse traces. But net you're watching your budget go down with every experiment. So far I've found the best experience with goose is with Sonnet 4.5, and that's what I'm trying to get away from (cost-wise).

Describe the solution you'd like

  • Improve langfuse integration OR create a simple visualization alternative so we can see context switching between models; if I'm going to switch from cursor to goose but am struggling with output. So I'm switching to lead / worker mode OR /plan mode in cli, then I need to know it's working
  • Show the active model in the cli e.g.
    ─── llm_search | router ──────────────────────────
    model: lead | worker (model name) <--- here
    extension_name: todo
    query: create a detailed todo list for porting the Python MCP server to TypeScript in a new directory mcp-server-openmeteots
  • Please UPDATE the existing documentation to clearly state that you could waste many hours figuring out that you need LANGFUSE_URL NOT LANGFUSE_BASE_URL if you want your traces to show up.
  • I think you need slightly better examples of what langfuse can do out of the box vs. what you can add to your code to make its traces better. Right now I don't think the latter is relevant tbh. But the native traces in langfuse from goose are pretty bad -- not worth the time I invested to figure out the LANGFUSE_URL problem. This has to do with something called "limited to SPAN" traces. My compare is LangSmith which I have been using with langgraph for a long time and it's much clearer what the inputs and outputs are.

Describe alternatives you've considered

  • Build my own visualization solution on top of the jsonl logs you output for tracing.
  • Stop using goose
  • And consider my plan b options like Kilo, etc ... I short listed a bunch and ultimately settled on you guys.

Additional context
Add any other context or screenshots about the feature request here.

  • I have verified this does not duplicate an existing feature request

Metadata

Metadata

Assignees

No one assigned

    Labels

    enhancementNew feature or request

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions