mirror of
https://github.com/infiniflow/ragflow.git
synced 2025-12-08 20:42:30 +08:00
Feat: Upgrade html parser (#9675)
### What problem does this PR solve? parse more html content. ### Type of change - [x] Other (please describe):
This commit is contained in:
@ -15,35 +15,200 @@
|
|||||||
# limitations under the License.
|
# limitations under the License.
|
||||||
#
|
#
|
||||||
|
|
||||||
from rag.nlp import find_codec
|
from rag.nlp import find_codec, rag_tokenizer
|
||||||
import readability
|
import uuid
|
||||||
import html_text
|
|
||||||
import chardet
|
import chardet
|
||||||
|
from bs4 import BeautifulSoup, NavigableString, Tag, Comment
|
||||||
|
import html
|
||||||
|
|
||||||
def get_encoding(file):
|
def get_encoding(file):
|
||||||
with open(file,'rb') as f:
|
with open(file,'rb') as f:
|
||||||
tmp = chardet.detect(f.read())
|
tmp = chardet.detect(f.read())
|
||||||
return tmp['encoding']
|
return tmp['encoding']
|
||||||
|
|
||||||
|
BLOCK_TAGS = [
|
||||||
|
"h1", "h2", "h3", "h4", "h5", "h6",
|
||||||
|
"p", "div", "article", "section", "aside",
|
||||||
|
"ul", "ol", "li",
|
||||||
|
"table", "pre", "code", "blockquote",
|
||||||
|
"figure", "figcaption"
|
||||||
|
]
|
||||||
|
TITLE_TAGS = {"h1": "#", "h2": "##", "h3": "###", "h4": "#####", "h5": "#####", "h6": "######"}
|
||||||
|
|
||||||
|
|
||||||
class RAGFlowHtmlParser:
|
class RAGFlowHtmlParser:
|
||||||
def __call__(self, fnm, binary=None):
|
def __call__(self, fnm, binary=None, chunk_token_num=None):
|
||||||
if binary:
|
if binary:
|
||||||
encoding = find_codec(binary)
|
encoding = find_codec(binary)
|
||||||
txt = binary.decode(encoding, errors="ignore")
|
txt = binary.decode(encoding, errors="ignore")
|
||||||
else:
|
else:
|
||||||
with open(fnm, "r",encoding=get_encoding(fnm)) as f:
|
with open(fnm, "r",encoding=get_encoding(fnm)) as f:
|
||||||
txt = f.read()
|
txt = f.read()
|
||||||
return self.parser_txt(txt)
|
return self.parser_txt(txt, chunk_token_num)
|
||||||
|
|
||||||
@classmethod
|
@classmethod
|
||||||
def parser_txt(cls, txt):
|
def parser_txt(cls, txt, chunk_token_num):
|
||||||
if not isinstance(txt, str):
|
if not isinstance(txt, str):
|
||||||
raise TypeError("txt type should be string!")
|
raise TypeError("txt type should be string!")
|
||||||
html_doc = readability.Document(txt)
|
|
||||||
title = html_doc.title()
|
temp_sections = []
|
||||||
content = html_text.extract_text(html_doc.summary(html_partial=True))
|
soup = BeautifulSoup(txt, "html5lib")
|
||||||
txt = f"{title}\n{content}"
|
# delete <style> tag
|
||||||
sections = txt.split("\n")
|
for style_tag in soup.find_all(["style", "script"]):
|
||||||
|
style_tag.decompose()
|
||||||
|
# delete <script> tag in <div>
|
||||||
|
for div_tag in soup.find_all("div"):
|
||||||
|
for script_tag in div_tag.find_all("script"):
|
||||||
|
script_tag.decompose()
|
||||||
|
# delete inline style
|
||||||
|
for tag in soup.find_all(True):
|
||||||
|
if 'style' in tag.attrs:
|
||||||
|
del tag.attrs['style']
|
||||||
|
# delete HTML comment
|
||||||
|
for comment in soup.find_all(string=lambda text: isinstance(text, Comment)):
|
||||||
|
comment.extract()
|
||||||
|
|
||||||
|
cls.read_text_recursively(soup.body, temp_sections, chunk_token_num=chunk_token_num)
|
||||||
|
block_txt_list, table_list = cls.merge_block_text(temp_sections)
|
||||||
|
sections = cls.chunk_block(block_txt_list, chunk_token_num=chunk_token_num)
|
||||||
|
for table in table_list:
|
||||||
|
sections.append(table.get("content", ""))
|
||||||
return sections
|
return sections
|
||||||
|
|
||||||
|
@classmethod
|
||||||
|
def split_table(cls, html_table, chunk_token_num=512):
|
||||||
|
soup = BeautifulSoup(html_table, "html.parser")
|
||||||
|
rows = soup.find_all("tr")
|
||||||
|
tables = []
|
||||||
|
current_table = []
|
||||||
|
current_count = 0
|
||||||
|
table_str_list = []
|
||||||
|
for row in rows:
|
||||||
|
tks_str = rag_tokenizer.tokenize(str(row))
|
||||||
|
token_count = len(tks_str.split(" ")) if tks_str else 0
|
||||||
|
if current_count + token_count > chunk_token_num:
|
||||||
|
tables.append(current_table)
|
||||||
|
current_table = []
|
||||||
|
current_count = 0
|
||||||
|
current_table.append(row)
|
||||||
|
current_count += token_count
|
||||||
|
if current_table:
|
||||||
|
tables.append(current_table)
|
||||||
|
|
||||||
|
for table_rows in tables:
|
||||||
|
new_table = soup.new_tag("table")
|
||||||
|
for row in table_rows:
|
||||||
|
new_table.append(row)
|
||||||
|
table_str_list.append(str(new_table))
|
||||||
|
|
||||||
|
return table_str_list
|
||||||
|
|
||||||
|
@classmethod
|
||||||
|
def read_text_recursively(cls, element, parser_result, chunk_token_num=512, parent_name=None, block_id=None):
|
||||||
|
if isinstance(element, NavigableString):
|
||||||
|
content = element.strip()
|
||||||
|
|
||||||
|
def is_valid_html(content):
|
||||||
|
try:
|
||||||
|
soup = BeautifulSoup(content, "html.parser")
|
||||||
|
return bool(soup.find())
|
||||||
|
except Exception:
|
||||||
|
return False
|
||||||
|
|
||||||
|
return_info = []
|
||||||
|
if content:
|
||||||
|
if is_valid_html(content):
|
||||||
|
soup = BeautifulSoup(content, "html.parser")
|
||||||
|
child_info = cls.read_text_recursively(soup, parser_result, chunk_token_num, element.name, block_id)
|
||||||
|
parser_result.extend(child_info)
|
||||||
|
else:
|
||||||
|
info = {"content": element.strip(), "tag_name": "inner_text", "metadata": {"block_id": block_id}}
|
||||||
|
if parent_name:
|
||||||
|
info["tag_name"] = parent_name
|
||||||
|
return_info.append(info)
|
||||||
|
return return_info
|
||||||
|
elif isinstance(element, Tag):
|
||||||
|
|
||||||
|
if str.lower(element.name) == "table":
|
||||||
|
table_info_list = []
|
||||||
|
table_id = str(uuid.uuid1())
|
||||||
|
table_list = [html.unescape(str(element))]
|
||||||
|
for t in table_list:
|
||||||
|
table_info_list.append({"content": t, "tag_name": "table",
|
||||||
|
"metadata": {"table_id": table_id, "index": table_list.index(t)}})
|
||||||
|
return table_info_list
|
||||||
|
else:
|
||||||
|
block_id = None
|
||||||
|
if str.lower(element.name) in BLOCK_TAGS:
|
||||||
|
block_id = str(uuid.uuid1())
|
||||||
|
for child in element.children:
|
||||||
|
child_info = cls.read_text_recursively(child, parser_result, chunk_token_num, element.name,
|
||||||
|
block_id)
|
||||||
|
parser_result.extend(child_info)
|
||||||
|
return []
|
||||||
|
|
||||||
|
@classmethod
|
||||||
|
def merge_block_text(cls, parser_result):
|
||||||
|
block_content = []
|
||||||
|
current_content = ""
|
||||||
|
table_info_list = []
|
||||||
|
lask_block_id = None
|
||||||
|
for item in parser_result:
|
||||||
|
content = item.get("content")
|
||||||
|
tag_name = item.get("tag_name")
|
||||||
|
title_flag = tag_name in TITLE_TAGS
|
||||||
|
block_id = item.get("metadata", {}).get("block_id")
|
||||||
|
if block_id:
|
||||||
|
if title_flag:
|
||||||
|
content = f"{TITLE_TAGS[tag_name]} {content}"
|
||||||
|
if lask_block_id != block_id:
|
||||||
|
if lask_block_id is not None:
|
||||||
|
block_content.append(current_content)
|
||||||
|
current_content = content
|
||||||
|
lask_block_id = block_id
|
||||||
|
else:
|
||||||
|
current_content += (" " if current_content else "") + content
|
||||||
|
else:
|
||||||
|
if tag_name == "table":
|
||||||
|
table_info_list.append(item)
|
||||||
|
else:
|
||||||
|
current_content += (" " if current_content else "" + content)
|
||||||
|
if current_content:
|
||||||
|
block_content.append(current_content)
|
||||||
|
return block_content, table_info_list
|
||||||
|
|
||||||
|
@classmethod
|
||||||
|
def chunk_block(cls, block_txt_list, chunk_token_num=512):
|
||||||
|
chunks = []
|
||||||
|
current_block = ""
|
||||||
|
current_token_count = 0
|
||||||
|
|
||||||
|
for block in block_txt_list:
|
||||||
|
tks_str = rag_tokenizer.tokenize(block)
|
||||||
|
block_token_count = len(tks_str.split(" ")) if tks_str else 0
|
||||||
|
if block_token_count > chunk_token_num:
|
||||||
|
if current_block:
|
||||||
|
chunks.append(current_block)
|
||||||
|
start = 0
|
||||||
|
tokens = tks_str.split(" ")
|
||||||
|
while start < len(tokens):
|
||||||
|
end = start + chunk_token_num
|
||||||
|
split_tokens = tokens[start:end]
|
||||||
|
chunks.append(" ".join(split_tokens))
|
||||||
|
start = end
|
||||||
|
current_block = ""
|
||||||
|
current_token_count = 0
|
||||||
|
else:
|
||||||
|
if current_token_count + block_token_count <= chunk_token_num:
|
||||||
|
current_block += ("\n" if current_block else "") + block
|
||||||
|
current_token_count += block_token_count
|
||||||
|
else:
|
||||||
|
chunks.append(current_block)
|
||||||
|
current_block = block
|
||||||
|
current_token_count = block_token_count
|
||||||
|
|
||||||
|
if current_block:
|
||||||
|
chunks.append(current_block)
|
||||||
|
|
||||||
|
return chunks
|
||||||
|
|
||||||
|
|||||||
@ -517,7 +517,8 @@ def chunk(filename, binary=None, from_page=0, to_page=100000,
|
|||||||
|
|
||||||
elif re.search(r"\.(htm|html)$", filename, re.IGNORECASE):
|
elif re.search(r"\.(htm|html)$", filename, re.IGNORECASE):
|
||||||
callback(0.1, "Start to parse.")
|
callback(0.1, "Start to parse.")
|
||||||
sections = HtmlParser()(filename, binary)
|
chunk_token_num = int(parser_config.get("chunk_token_num", 128))
|
||||||
|
sections = HtmlParser()(filename, binary, chunk_token_num)
|
||||||
sections = [(_, "") for _ in sections if _]
|
sections = [(_, "") for _ in sections if _]
|
||||||
callback(0.8, "Finish parsing.")
|
callback(0.8, "Finish parsing.")
|
||||||
|
|
||||||
|
|||||||
Reference in New Issue
Block a user