Hacker News new | past | comments | ask | show | jobs | submit | jamesbrady's comments login

Elicit (https://elicit.com/careers) | Oakland, CA and Hybrid | Frontend, AI, and Full-stack software engineering roles.

Elicit is automating high-quality reasoning so that we can help the world make more breakthroughs in every domain: from climate change to the gut microbiome to longevity and economic policy.

We’ve scaled to over 200,000 monthly users purely by word of mouth and recently crossed $1.5MM in annual revenue, 7 months after launching subscriptions.

We’re now building out our software engineering team, and hiring across several technical roles.

If you'd like to know more about some of the work we're doing you could check: - A recent blog post current UX work: https://blog.elicit.com/living-documents-ai-ux/ - Me talking about our "AI engineer" role on a podcast a couple of weeks ago: https://www.latent.space/p/hiring

Check out the careers page linked above, or email me at james@elicit.com


Elician here!

Our main focus is a little different to SciSummary actually. We're focussed on understanding researchers broader workflows, and providing a research assistant (i.e. rather than a particular narrow tool for summarisation or search).

The workflows we're most excited about at the moment are literature and systematic reviews: we think we can make these orders of magnitude faster and higher quality.


Actually, it's not an LLM!

We do use LLMs, but the secret sauce is an approach we call Factored Cognition which we wrote about here: https://ought.org/research/factored-cognition

(Elicit the company and app was spun out from Ought the research lab).

We do joke internally about the homophone (in fact, IIRC we did a little joke on our CEO by rebranding for his birthday in 2022) but I'm sorry to report that we're all careful, ethical, and well-behaved people :(


Cool, thanks for more info, nice to see other approaches. What data is used for training?


Elician here: thank you for sharing our tool and for this praise!

We're glad you're enjoying it.


Elician here.

This is a good point! (Hopefully) obviously, if we knew a particular claim was fishy, we wouldn't make it in the app in the first place.

However, we do do a couple of things which go towards addressing your concern:

1. We can be more or less confident in the answers we're giving in the app, and if that confidence dips below a threshold we mark that particular cell in the results table with a red warning icon which encourages caution and user verification. This confidence level isn't perfectly calibrated, of course, but we are trying to engender a healthy, active, wariness in our users so that they don't take Elicit results as gospel. 2. We provide sources for all of the claims made in the app. You can see these by clicking on any cell in the results table. We encourage users to check—or at least spot-check—the results which they are updating on. This verification is generally much faster than doing the generation of the answer in the first place.


Elician here! Thanks for your comment.

I'm not sure I agree that those rule-of-thumb statistics are "arbitrary" or "fictional"… I guess it depends on what you mean by that. I can say that on our part they're a good faith attempt to help users calibrate how best to use the tool, using evaluations of Elicit based on real usage.

Definitely accept that the tool can work better or worse depending on your domain or workflow though!

One way we do try to distinguish ourselves from vanilla LLMs is that we provide sources for all of the claims made. I mention this because we hope our users can approach the falsification process you mention for Google. We want to show people where particular claims come from such that we earn their trust.

Walking citation trails and verifying transitive claims is something we've talked about but need more people to implement! (https://elicit.com/careers)


> I'm not sure I agree that those rule-of-thumb statistics are "arbitrary" or "fictional"… I guess it depends on what you mean by that.

Sorry for the confusion: I meant that fragmede's comment was arbitrary and fictional, not the 90% figure. I was talking about these numbers:

  if it takes 1 hour to get one answer by hand, but only 20 minutes for the machine, and 20 minutes to check the answer, the user still comes out ahead


Oh, my bad—I misunderstood, thanks for the clarification


Elician here!

Accuracy and supportedness of the claims made in Elicit are two of the most central things we focus on—it's a shame it didn't work as well as we'd like in this case.

I'd appreciate knowing more about the specifics so we can understand and improve


https://analyticalsciencejournals.onlinelibrary.wiley.com/do... elicit summarises this paper abstract: “Psilocybin was present at 0.47 wt% in the mycelium.”

Actual quote from the abstract: “ No tryptamines were detected in the basidiospores, and only psilocin was present at 0.47 wt.% in the mycelium.”

It does not differentiate between psilocin and psilocybin, those are two different molecules.


Elician here.

The two main problems we're addressing now are:

1. Finding papers / claims / data across an academic literature which is ballooning in size.

2. Using these raw materials to to answer questions in a reliable manner.

#2 is where the bulk of the tricky ML work is, and where vanilla language models often fall short because of limited context windows and hallucination.

We're also working to expand Elicit to help academics with other parts of their research, like surfacing critiques, suggesting related prior work, brainstorming related research questions, identifying risks of bias, …


This is not true.


Agree, way too long!

I gave that much of a heads-up of my leaving (not a formal resignation, as it happens) for a few reasons including:

1. It's a C-level replacement, which takes _months_

2. I cared deeply (and still do) about my team and the company mission, and wanted them to succeed: "the more notice the better" seemed right to me (it wasn't)

3. My career plans were to move into a new field which would require various courses, lots of reading, conferences, etc. – although I knew I was leaving, I hadn't even started speaking to potential next employers or investors when I told my CEO I was leaving

Putting all of that together is how I ended up with the 10 month blunder, but the post is less against 10 month notice periods, and more for matching the notice period to your handover, whatever that would be.

I feel your pain on working out those 5 weeks you mention! Awkward for everyone involved.


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: