kalle07 commited on
Commit
28cd386
·
verified ·
1 Parent(s): 0a321d0

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -2
README.md CHANGED
@@ -24,7 +24,7 @@ better input = better output<br>
24
 
25
  ...
26
 
27
- Most LLM applications only convert your PDF simple to txt, nothing more, its like you save your PDF as txt file. Blocks of text that are close together are often mixed up and tables cannot be read logically.
28
  Therefore its better to convert it with some help of a <b>"Parser"</b>. The embedder can now find a better context.<br>
29
  I work with "<b>pdfplumber/pdfminer</b>" none OCR, so its very fast!<br>
30
  <ul style="line-height: 1.05;">
@@ -34,7 +34,7 @@ I work with "<b>pdfplumber/pdfminer</b>" none OCR, so its very fast!<br>
34
  <li>Instant view of the result, hit one pdf on top of the list</li>
35
  <li>Converts some common tables as json-foramt inside the txt file, readable for embedder</li>
36
  <li>Adds the absolute PAGE number to each page</li>
37
- <li>Adds the label “Chapter” for large font and/or “important” for bold font</li>
38
  <li>All txt files will be created in original folder of PDF</li>
39
  <li>All previous txt files are overwritten</li>
40
  <li>aprox 5 to 20 Pages/sec - depends on complexity and system-power</li>
@@ -46,6 +46,8 @@ It is really hard for me with GUI and the Function and in addition to compile it
46
  For the python-file you need to import missing libraries.<br>
47
  Of course there is a lot of need for optimization(save/error-handling) or the use of other parser libraries, but it's a start.
48
  <br><br>
 
 
49
  ...
50
  <br>
51
  I also have a "<b>docling</b>" parser with OCR (GPU is need for fast processing), its only be a python-file, not compiled.<br>
 
24
 
25
  ...
26
 
27
+ Most LLM applications only convert your PDF simple to txt, nothing more, its like you save your PDF as txt file. For usual flowing text Books its quite okay! But often blocks of text that are close together are mixed up and tables cannot be read logically.
28
  Therefore its better to convert it with some help of a <b>"Parser"</b>. The embedder can now find a better context.<br>
29
  I work with "<b>pdfplumber/pdfminer</b>" none OCR, so its very fast!<br>
30
  <ul style="line-height: 1.05;">
 
34
  <li>Instant view of the result, hit one pdf on top of the list</li>
35
  <li>Converts some common tables as json-foramt inside the txt file, readable for embedder</li>
36
  <li>Adds the absolute PAGE number to each page</li>
37
+ <li>Adds the label “chapter” for large font and/or “important” for bold font</li>
38
  <li>All txt files will be created in original folder of PDF</li>
39
  <li>All previous txt files are overwritten</li>
40
  <li>aprox 5 to 20 Pages/sec - depends on complexity and system-power</li>
 
46
  For the python-file you need to import missing libraries.<br>
47
  Of course there is a lot of need for optimization(save/error-handling) or the use of other parser libraries, but it's a start.
48
  <br><br>
49
+ i am working on a 50% faster version. in addition, the GUI should allow more influence on the processing, e.g. faster raw text, trim margins (yes/no) and set % yourself, set unimportant text block size, layout with line breaks or force more continuous text.<br>
50
+ give me a hand to help ;) <br>
51
  ...
52
  <br>
53
  I also have a "<b>docling</b>" parser with OCR (GPU is need for fast processing), its only be a python-file, not compiled.<br>