Steps in this run
| Step |
Calls |
Tokens in |
Cache hit |
Cost |
|
ranking
|
2 |
115,271 |
|
$0.73088 |
|
response generation
|
3 |
7,157 |
|
$0.03454 |
|
learning engine pattern analysis
|
1 |
13,656 |
|
$0.01595 |
|
haiku prescreen
|
2 |
10,733 |
|
$0.01147 |
|
learning engine self eval
|
1 |
4,682 |
|
$0.00935 |
All 9 API calls — tap to expand
$0.009504
Est. cost (USD)
Result preview
```json
[
{
"post_index": 14,
"cluster_ids": [1, 12],
"claim": "Clinical AI adoption requires integration into primary care workflows, not isolated demonstrations",
"argument_type": "mechanism_explanation",
"stance": "challenges_status_quo",
"hyde_excerpt": "The deployment of clinical AI tools has historically prioritized algorithmic performance in controlled settings ove
81,299
Tokens in (billed)
$0.460152
Est. cost (USD)
Result preview
```json
[
{
"post_index": 17,
"matched_article_id": 23,
"match_confidence": 87,
"match_reason": "The tweet claims AI beat emergency physicians 67% vs ~50% accuracy and argues AI diagnostic superiority is large and real under chaotic, incomplete-data conditions — directly engaging the article's central thesis about AI diagnostic tools matching or exceeding physician performance an
$0.014520
Est. cost (USD)
Result preview
The accuracy gap is real and worth taking seriously. But the performance question and the liability question are running on completely separate tracks right now, and that separation is where things get dangerous.
66% of U.S. physicians were using AI clinically in 2024, up from 38% the year before. Adoption is accelerating fast. What isn't accelerating is any legal clarity about what happens when
$0.010611
Est. cost (USD)
Result preview
The question this raises but sidesteps: accountable to whom, and through what paper trail?
Because the workflow problem in primary care isn't speed, it's that PCPs over-refer defensively as a legal default, not because they can't manage the case. I dug into this at https://www.onhealthcare.tech/p/the-pcp-as-specialist-how-ai-and?utm_source=x&utm_medium=reply&utm_content=2050792693240856913&utm_ca
$0.001964
Est. cost (USD)
Result preview
```json
[]
```
30,617
Tokens in (billed)
$0.270731
Est. cost (USD)
Result preview
Looking at post [0], the author argues that software development costs will NOT collapse — instead, quality code review will become expensive because LLMs searching through large solution spaces cost as much as senior developers, and that software quality will remain poor. This is a direct counterpoint to article ID:20, which argues development costs are collapsing 50-90% and that this will disrup
$0.009405
Est. cost (USD)
Result preview
Build cost compression and quality are two different problems, and this post is conflating them.
The prior auth example I modeled out, $4M and eighteen months down to $300K and six weeks, that's real. But cheap to build doesn't mean good. What changes is who can afford to build something mediocre internally instead of buying someone else's mediocre product.
That shift alone breaks a lot of healt
$0.009350
Est. cost (USD)
Result preview
```json
[
{"post_index": 0, "prediction": "reject", "confidence": 72, "reason": "AI infrastructure/engineering hype without healthcare-specific application. Discusses software engineering practices and LLM code review as general tech trend, not healthcare delivery or clinical workflow context."},
13,656
Tokens in (billed)
$0.015949
Est. cost (USD)
Result preview
```json
[
{
"category": "ai_safety_vulnerability_incident_tangent",
"summary": "Posts about AI safety vulnerabilities, data breaches, or security incidents that are tangential to healthcare systems.",
"exclusion_rule": "Exclude posts that focus primarily on AI safety vulnerabilities, s