Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Check if summary at start works #96

Closed
koppor opened this issue Aug 2, 2024 · 5 comments
Closed

Check if summary at start works #96

koppor opened this issue Aug 2, 2024 · 5 comments
Milestone

Comments

@koppor
Copy link
Collaborator

koppor commented Aug 2, 2024

  1. User has new JabRef. This can be achieved with a virtual vm (see https://github.com/JabRef/jabref/tree/main/scripts/vms) or Windows Sandbox
  2. User opens Chocolate.bib from our OneDrive (which has some PDFs attached)
  3. User clicks on "Corti_2009"
  4. User agrees on Privacy Policy (see Fix privacy policy for AI summarization #92)
  5. User waits
  6. User clicks on "AI chat"
  7. User sees "Please wait", too

I think, the user should not wait for more than 5 minutes - there is something wrong.

@koppor koppor added this to the Week 3 milestone Aug 2, 2024
@ThiloteE
Copy link
Collaborator

ThiloteE commented Aug 2, 2024

Just for context: Corti_2009 has 33 pages and that's roughly 100 000 characters. On reddit somebody said 4 tokens for 3 words on average. That would mean roughly 75 000 tokens, but ultimately the token to word ratio depends on the tokenizer of the model that is used, so "it depends". In any case, we can expect that this pdf file actually has quite a lot of tokens.

More information is needed: What AI service are you subscribing to? What model?
What commit in what pull-request?

Some hypotheses about why waiting time is so long:

  • LLM service needs to re-calculate context alot, because the paper is above its supported context window size.
  • The LLM is a really large model and therefore is slow.
  • Something during sending takes too long.
  • Does the API service have an outage or rate-limit?
  • JabRef tries to do embeddings at the same time. Once for ai chat and once for ai summarization.
  • JabRef's embedding model still runs on CPU and creating embeddings is very slow on CPU.

@koppor
Copy link
Collaborator Author

koppor commented Aug 3, 2024

It was OpenAI and gpt4o.

Independent of that, it is good UI practice to provide progress. German explanation: http://www.usabilitypatterns.info/catalog/patterns/fortschrittsanzeige.html

If one cannot report progress, one should output the expected waiting time.

In case the http connection is lost (typical timeout 1 minute) should also be handled. (I assume, it is, but we need to check)

@InAnYan
Copy link
Owner

InAnYan commented Aug 4, 2024

I'm not sure, what this issue talks about?

Is this a general issue for progress indication?

@koppor
Copy link
Collaborator Author

koppor commented Aug 4, 2024

The issue covers multiple things.

  1. progress indication -> I created a new issue Clean progress indication #115.
  2. A systematic final check before merging the feature.

Regarding 2:

One can approach this very structured using System Test Portal or other software. I think, this is too much for our team size. Thus, I created "final-test" label to make it clear, that this should happen if everything is settled. - I also moved the issue as last in the milestone view https://github.com/InAnYan/jabref/milestone/3

@koppor
Copy link
Collaborator Author

koppor commented Aug 7, 2024

I think, we are at a stage, where we regular check from fresh machines (https://github.com/JabRef/jabref/tree/main/scripts/vms) and do not need a separate tracking issue.

@koppor koppor closed this as completed Aug 7, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests

3 participants