Yes, you can give the Big Brain Thing a vague task and expect results, some times it'll do it right
But if you want repeatability, give it tools to determine what is a "disruption".
I have this exact system running on a WhateverClaw with a few simple tools: weather, train time tables (I commute via train) and a read-only version of my "will I be at the office or remote today" -calendar via a tool. Oh and a tool to notify me via Telegram.
It gets this information using the given tools and determines if it's worth notifying me.
TBH this doesn't need a "claw", I could just run the tools in cron, construct a prompt with that data and run any LLM on it.
You need to write a clearer prompt.
Your AI assistant orders an experimental jetpack from a random startup lab. Would you have honestly guessed that the prompt was "ambiguous" before you knew how the AI was going to act on it ?
You'll define exactly what good looks like.
"Generate the following JSON formatted object array representing the interruptions in my daily traffic. If no results, emit []. Send this at 8am every morning. {some schema}. Then run jsonreporter.py"
Then just let jsonreporter.py discriminate however it likes. Keep the LLMs doing what they are good at, and keep hard code doing what it's good at.