10 comments

  • bflesch14 hours ago
    Somehow &quot;clerk&quot; is on my ublock origin blocklist and therefore the whole website is not loading. I didn&#x27;t add &quot;clerk&quot; to the blocklist so it must&#x27;ve been added by one of the blocklists that ublock origin is subscribed to, so there must be a good reason why &quot;clerk&quot; is on that blocklist.<p>When building a product for medical audience which might care a lot about privacy maybe don&#x27;t use components which are shady enough that they end up on blocklists.<p>Edit:<p>&gt; Why no Vector DB? In medicine, &quot;freshness&quot; is critical. If a new trial drops today, a pre-indexed vector store might miss it. My real-time approach ensures the answer includes papers published today.<p>This is total rubbish - did you talk to a single medical practitioner when building this? Nobody will do new treatments on their patients if a new paper was &quot;published&quot; (whatever that means, just being added to some search index). These people require trusted source, experimental treatment is only done for private clients who have tried all other options.
    • amber_raza13 hours ago
      Thanks for the feedback—this is helpful.<p>1. Re: Clerk&#x2F;uBlock: You were spot on. The default Clerk domain often gets flagged by strict blocklists. I just updated the DNS records to serve auth from a first-party subdomain (clerk.getevidex.com) to resolve this. It should be working now.<p>2. Re: Freshness &amp; &#x27;Rubbish&#x27;: You are absolutely right that standard of care doesn&#x27;t (and shouldn&#x27;t) change overnight based on one new paper.<p>However, the decision to ditch the Vector DB for Live Search wasn&#x27;t about pushing &#x27;experimental treatments&#x27;—it was about Safety and Engineering constraints:<p>Retractions &amp; Safety Alerts: A stale vector index is a safety risk. If a major paper is retracted or a drug gets a black-box warning today, a live API call to PubMed&#x2F;EuropePMC reflects that immediately. A vector store is only as good as its last re-index.<p>The &#x27;Long Tail&#x27;: Vectorizing the entire PubMed corpus (35M+ citations) is expensive and hard to keep in sync. By using the search APIs directly, we get the full breadth of the database (including older, obscure case reports for rare diseases) without maintaining a massive, potentially stale index.<p>The goal isn&#x27;t to be &#x27;bleeding edge&#x27;—it&#x27;s to be &#x27;currently accurate&#x27;.
      • breadislove13 hours ago
        a good system (like openevidence) indexes every paper released and semantic search can incredible helpful since the the search api of all those providers are extremely limited in terms of quality.<p>now you get why those system are not cheap. keeping indexes fresh, maintaining high quality at large scale and being extremely precise is challenging. by having distributed indexes you are at the mercy of the api providers and i can tell you from previous experience that it won&#x27;t be &#x27;currently accurate&#x27;.<p>for transparency: i am building a search api, so i am biased. but i also build medical retrieval systems for some time.
        • amber_raza13 hours ago
          Appreciate the transparency and the insight from a fellow builder.<p>You are spot on that maintaining a fresh, high-quality index at scale is the &#x27;hard problem&#x27; (and why tools like OpenEvidence are expensive).<p>However, I found that for clinical queries, Vector&#x2F;Semantic Search often suffers from &#x27;Semantic Drift&#x27;—fuzzily matching concepts that sound similar but are medically distinct.<p>My architectural bet is on Hybrid RAG:<p>Trust the MeSH: I rely on PubMed&#x27;s strict Boolean&#x2F;MeSH search for the retrieval because for specific drug names or gene variants, exact keyword matching beats vector cosine similarity.<p>LLM as the Reranker: Since API search relevance can indeed be noisy, I fetch a wider net (top ~30-50 abstracts) and use the LLM&#x27;s context window to &#x27;rerank&#x27; and filter them before synthesis.<p>It&#x27;s definitely a trade-off (latency vs. index freshness), but for a bootstrapped tool, leveraging the NLM&#x27;s billions of dollars in indexing infrastructure feels like the right lever to pull vs. trying to out-index them.
      • jyscao9 hours ago
        This sounds like a cookie cutter ChatGPT reply.
        • amber_raza8 hours ago
          Haha, ouch. I promise it’s just me—I just spent 20 minutes rewriting that comment because I didn&#x27;t want to sound like an idiot explaining search to a search engineer. I&#x27;ll take it as a sign to dial back the formatting next time.
          • bflesch8 hours ago
            That emdash in your reply is so in-your-face &quot;—&quot;.
      • bflesch8 hours ago
        Now it is loading. You are still in violation of GDPR rules by including a SVG file with the google logo from the clerk.com domain and a css file from tailwindcss.com - both are tracking users. There is no privacy policy on your page. The privacy policy should include a list of companies you share my visitor data with and what kind of data is shared, and how I can deny sharing that data.
        • amber_raza8 hours ago
          Fair point on the Privacy Policy link. That definitely slipped through the cracks in the launch rush. I just pushed a fix to add it to the footer now.<p>Re: the trackers: The SVG is just the icon inside the Clerk login button, but you&#x27;re right that loading Tailwind via CDN isn&#x27;t ideal for strict GDPR IP-masking. I&#x27;ll look into self-hosting the assets to clean that up.
  • pdyc1 hour ago
    I like your approach of &quot;smart routing&quot; but using regex&#x2F;keywords based approach has a problem that it does not captures semantic similarity of keywords so search with similar intents are missed, how are you handling it? or you dont need to handle it since it is for domain experts and they are likely to search based on keywords(dictionary)?
  • dataviz100013 hours ago
    I&#x27;m working on building an AI agent that creates queries over a time-series database focused on financial data. For example, it can quantify Federal Reserve reports and generate a table showing how SPY reacted 30 minutes after, at EoD, at the next day’s open, and at the next day’s EoD. It will plan the database query and then query the data from a materialized view. It is magic!<p>How would biomedical researchers use tons of time-series data? A better question is: what questions are biomedical researchers asking with time-series data? I&#x27;m a lot more interested in generalized querying over time-series data than just financial data. What would be a great proof of concept?
    • amber_raza13 hours ago
      That sounds like a fascinating project.<p>To answer your question: In the biomedical world, the &#x27;Time-Series&#x27; equivalent is Patient Telemetry (Continuous Glucose Monitors, ICU Vitals, Wearables).<p>The Question Researchers Ask: &#x27;Can we predict sepsis&#x2F;stroke 4 hours before it happens based on the velocity of change in Heart Rate + BP?&#x27;<p>Right now, Evidex is focused on the Unstructured Text (Literature&#x2F;Guidelines) rather than the structured time-series data, but the &#x27;Holy Grail&#x27; of medical AI is eventually combining them: Using the Literature to interpret the Live Vitals in real-time.
  • adit_ya114 hours ago
    Out of curiosity, what&#x27;s the prioritization of evidence (RTC Metanalysis &gt; RTC &gt; observational ) etc, and what&#x27;s the end user benefit over a tool like OpenEvidence? You mention that other tools are expensive, slow, or increasingly heavy with pharma ads, but OpenEvidence for now seems to be pretty similiar with offerings, speed, and responses. What&#x27;s your pitch as to why one should prefer this?
    • amber_raza13 hours ago
      Great questions.<p>1. Prioritization: I instruct the model to prioritize evidence in this hierarchy: Meta-Analyses &amp; Systematic Reviews &gt; RCTs &gt; Observational Studies &gt; Case Reports. It explicitly deprioritizes non-human studies unless specified.<p>2. Why not OpenEvidence? OE is excellent! But we made two architectural choices to solve different problems:<p>&#x27;Long Tail&#x27; Coverage: OE relies on a pre-indexed vector store, which often creates a blind spot for niche&#x2F;rare diseases where papers aren&#x27;t in the &#x27;Top 1% of Journals.&#x27; Because Evidex queries live APIs, we catch the obscure case reports that static indexes often prune out.<p>Workflow: OE is a &#x27;Consultant&#x27; (Q&amp;A). Evidex is a &#x27;Resident&#x27; (Grunt work). The &#x27;Case Mode&#x27; is built to take messy patient histories and draft the actual documentation (SOAP Notes&#x2F;Appeals) you have to write after finding the answer.
  • neil_naveen14 hours ago
    FYI, You are using Clerk in development mode
    • amber_raza14 hours ago
      Oof, good catch! I must have left the test keys active in the deployment config.<p>Swapping them to production keys right now. Thanks for the heads up!
  • jph12 hours ago
    Great project. Want to contact me when you&#x27;d like to talk? I do software engineering for clinicians at a health care organization, and I&#x27;d love to have my teams try your work in their own contexts. Email joel@joelparkerhenderson.com.
    • amber_raza11 hours ago
      Thanks, Joel! This is exactly the kind of clinical workflow I built &#x27;Case Mode&#x27; for.<p>I will send you an email shortly to get connected. I&#x27;d love to get your teams set up with a pilot instance. Appreciate the reach out.
  • eoravkin14 hours ago
    Out of curiosity, did you actually see any pharma ads on OpenEvidence?
    • amber_raza13 hours ago
      Great question. I haven&#x27;t seen banner ads on OpenEvidence yet, but the &#x27;hidden tax&#x27; of free tools is often Publisher Bias.<p>Users have noted that some current tools heavily overweight citations from &#x27;Partner Journals&#x27; (like NEJM&#x2F;JAMA) because they index the full text, effectively burying better papers from non-partner journals in the vector retrieval.<p>My goal is strictly Neutral Retrieval. By hitting the PubMed&#x2F;OpenAlex APIs live, Evidex treats a niche pediatric journal with the same relevance weight as a major publisher, ensuring the &#x27;Long Tail&#x27; of evidence isn&#x27;t drowned out by business partnerships.
    • breadislove13 hours ago
      this might be interesting: <a href="https:&#x2F;&#x2F;www.theinformation.com&#x2F;articles&#x2F;chatgpt-doctors-startup-doubles-valuation-12-billion-revenue-surges" rel="nofollow">https:&#x2F;&#x2F;www.theinformation.com&#x2F;articles&#x2F;chatgpt-doctors-star...</a><p>&gt; $150M RR on just ads, +3x from August. On &lt;1M users.<p>source: <a href="https:&#x2F;&#x2F;x.com&#x2F;ArfurRock&#x2F;status&#x2F;1999618200024076620" rel="nofollow">https:&#x2F;&#x2F;x.com&#x2F;ArfurRock&#x2F;status&#x2F;1999618200024076620</a>
      • amber_raza12 hours ago
        Whoa. $150M ARR on ads is a wild stat.<p>Thanks for sharing that source. It really validates the thesis that unless the user pays (SaaS), the Pharma companies are the real customers.
        • eoravkin3 hours ago
          You built a cool product. I&#x27;m actually one of the founders of <a href="https:&#x2F;&#x2F;medisearch.io">https:&#x2F;&#x2F;medisearch.io</a> which is similar to what you are building. I think the long-tail problem that you describe can be solved in other ways than with live APIs and you may find other problems with using live APIs.
  • OutOfHere7 hours ago
    All such custom sites are increasingly unnecessary since modern thinking AIs like ChatGPT 5.2 Extended and Gemini 3 Pro do an incredible job surfacing good papers. In my experience, the benefit comes from using multiple AIs because they all have blind spots, and none is pareto optimal.<p>As a patient, sometimes I don&#x27;t want the AI to have my entire medical history, as this lets me consider things from different angles. For each chat, I give it the reconstructed history that I think is sufficient. I want it to be an explorer more than a doctor.
    • amber_raza7 hours ago
      That is a fair critique. The frontier models are getting incredible at general reasoning.<p>The gap Evidex fills isn&#x27;t &#x27;Intelligence&#x27;. It is Provenance and Liability.<p>Strict Sourcing: Even advanced models can hallucinate a plausible-sounding study. Evidex constrains the model to answer only using the abstracts returned by the API. This reduces the risk of a &#x27;creative&#x27; citation.<p>Explorer vs. Operator: You mentioned using AI as an &#x27;explorer&#x27; (Patient use case). Doctors are usually &#x27;operators&#x27;. They need to find the specific dosage or guideline quickly to close a chart.<p>I view this less as replacing Gemini&#x2F;GPT. It is more of a &#x27;Safety Wrapper&#x27; around them for a high-stakes environment.
      • OutOfHere6 hours ago
        The problem is that doctors almost always, except perhaps in the emergency department, are currently too full of themselves, and are not open to reading relevant research unless a patient like me forces it upon the doctor. Maybe they are busy but that doesn&#x27;t work for the patient. Even upon such forcing of the patient sharing research, the doctor will often read only a single line from an entire paper. How do you change this culture? It doesn&#x27;t serve the patient too well to get an inaccurate root cause diagnosis from the doctors as I often do. It comes upon the patient to really spend the time investigating and testing hypotheses and theories, failing which the root causes go ignored, and one ends up taking too many unnecessary or even harmful pharmaceuticals.
  • vikas-sharma13 hours ago
    [dead]